2
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Local minimization of prediction errors drives learning of invariant object representations in a generative network model of visual perception

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          The ventral visual processing hierarchy of the cortex needs to fulfill at least two key functions: perceived objects must be mapped to high-level representations invariantly of the precise viewing conditions, and a generative model must be learned that allows, for instance, to fill in occluded information guided by visual experience. Here, we show how a multilayered predictive coding network can learn to recognize objects from the bottom up and to generate specific representations via a top-down pathway through a single learning rule: the local minimization of prediction errors. Trained on sequences of continuously transformed objects, neurons in the highest network area become tuned to object identity invariant of precise position, comparable to inferotemporal neurons in macaques. Drawing on this, the dynamic properties of invariant object representations reproduce experimentally observed hierarchies of timescales from low to high levels of the ventral processing stream. The predicted faster decorrelation of error-neuron activity compared to representation neurons is of relevance for the experimental search for neural correlates of prediction errors. Lastly, the generative capacity of the network is confirmed by reconstructing specific object images, robust to partial occlusion of the inputs. By learning invariance from temporal continuity within a generative model, the approach generalizes the predictive coding framework to dynamic inputs in a more biologically plausible way than self-supervised networks with non-local error-backpropagation. This was achieved simply by shifting the training paradigm to dynamic inputs, with little change in architecture and learning rule from static input-reconstructing Hebbian predictive coding networks.

          Related collections

          Most cited references76

          • Record: found
          • Abstract: found
          • Article: not found

          The free-energy principle: a unified brain theory?

          A free-energy principle has been proposed recently that accounts for action, perception and learning. This Review looks at some key brain theories in the biological (for example, neural Darwinism) and physical (for example, information theory and optimal control theory) sciences from the free-energy perspective. Crucially, one key theme runs through each of these theories - optimization. Furthermore, if we look closely at what is optimized, the same quantity keeps emerging, namely value (expected reward, expected utility) or its complement, surprise (prediction error, expected cost). This is the quantity that is optimized under the free-energy principle, which suggests that several global brain theories might be unified within a free-energy framework.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Backpropagation Applied to Handwritten Zip Code Recognition

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Distributed and overlapping representations of faces and objects in ventral temporal cortex.

              The functional architecture of the object vision pathway in the human brain was investigated using functional magnetic resonance imaging to measure patterns of response in ventral temporal cortex while subjects viewed faces, cats, five categories of man-made objects, and nonsense pictures. A distinct pattern of response was found for each stimulus category. The distinctiveness of the response to a given category was not due simply to the regions that responded maximally to that category, because the category being viewed also could be identified on the basis of the pattern of response when those regions were excluded from the analysis. Patterns of response that discriminated among all categories were found even within cortical regions that responded maximally to only one category. These results indicate that the representations of faces and objects in ventral temporal cortex are widely distributed and overlapping.
                Bookmark

                Author and article information

                Contributors
                Journal
                Front Comput Neurosci
                Front Comput Neurosci
                Front. Comput. Neurosci.
                Frontiers in Computational Neuroscience
                Frontiers Media S.A.
                1662-5188
                25 September 2023
                2023
                : 17
                : 1207361
                Affiliations
                [1] 1Cognitive and Systems Neuroscience Group, Swammerdam Institute for Life Sciences, University of Amsterdam , Amsterdam, Netherlands
                [2] 2Machine Learning Group, Centrum Wiskunde & Informatica , Amsterdam, Netherlands
                Author notes

                Edited by: Arpan Banerjee, National Brain Research Centre (NBRC), India

                Reviewed by: John Magnotti, University of Pennsylvania, United States; Vignesh Muralidharan, Indian Institute of Technology Jodhpur, India

                *Correspondence: Matthias Brucklacher, m.m.brucklacher@ 123456uva.nl
                Article
                10.3389/fncom.2023.1207361
                10561268
                37818157
                2faca9e8-62e5-4b6c-a42c-b6a92ac1f6bf
                Copyright © 2023 Brucklacher, Bohté, Mejias and Pennartz.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 17 April 2023
                : 31 August 2023
                Page count
                Figures: 8, Tables: 1, Equations: 6, References: 83, Pages: 15, Words: 11083
                Categories
                Neuroscience
                Original Research

                Neurosciences
                self-supervised learning,predictive coding,generative model,vision,hierarchy,representation learning,hebbian learning,video

                Comments

                Comment on this article