27
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Limits to visual representational correspondence between convolutional neural networks and the human brain

      research-article
      1 , , 2
      Nature Communications
      Nature Publishing Group UK
      Neural decoding, Object vision

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Convolutional neural networks (CNNs) are increasingly used to model human vision due to their high object categorization capabilities and general correspondence with human brain responses. Here we evaluate the performance of 14 different CNNs compared with human fMRI responses to natural and artificial images using representational similarity analysis. Despite the presence of some CNN-brain correspondence and CNNs’ impressive ability to fully capture lower level visual representation of real-world objects, we show that CNNs do not fully capture higher level visual representations of real-world objects, nor those of artificial objects, either at lower or higher levels of visual representations. The latter is particularly critical, as the processing of both real-world and artificial visual stimuli engages the same neural circuits. We report similar results regardless of differences in CNN architecture, training, or the presence of recurrent processing. This indicates some fundamental differences exist in how the brain and CNNs represent visual information.

          Abstract

          Convolutional neural networks are increasingly used to model human vision. Here, the authors compare the performance of 14 different CNNs and human fMRI responses to real-world and artificial objects to show some fundamental differences exist between them.

          Related collections

          Most cited references55

          • Record: found
          • Abstract: not found
          • Article: not found

          Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Deep learning.

            Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Cortical surface-based analysis. I. Segmentation and surface reconstruction.

              Several properties of the cerebral cortex, including its columnar and laminar organization, as well as the topographic organization of cortical areas, can only be properly understood in the context of the intrinsic two-dimensional structure of the cortical surface. In order to study such cortical properties in humans, it is necessary to obtain an accurate and explicit representation of the cortical surface in individual subjects. Here we describe a set of automated procedures for obtaining accurate reconstructions of the cortical surface, which have been applied to data from more than 100 subjects, requiring little or no manual intervention. Automated routines for unfolding and flattening the cortical surface are described in a companion paper. These procedures allow for the routine use of cortical surface-based analysis and visualization methods in functional brain imaging. Copyright 1999 Academic Press.
                Bookmark

                Author and article information

                Contributors
                xucogneuro@gmail.com
                Journal
                Nat Commun
                Nat Commun
                Nature Communications
                Nature Publishing Group UK (London )
                2041-1723
                6 April 2021
                6 April 2021
                2021
                : 12
                : 2065
                Affiliations
                [1 ]GRID grid.47100.32, ISNI 0000000419368710, Psychology Department, , Yale University, ; New Haven, CT USA
                [2 ]GRID grid.416868.5, ISNI 0000 0004 0464 0574, Laboratory of Brain and Cognition, National Institute of Mental Health, ; Bethesda, MD USA
                Author information
                http://orcid.org/0000-0002-8697-314X
                Article
                22244
                10.1038/s41467-021-22244-7
                8024324
                33824315
                f5d236b9-788d-429a-9973-f9552107344f
                © The Author(s) 2021

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 30 March 2020
                : 5 March 2021
                Funding
                Funded by: NIH grant 1R01EY022355 and NIH grant 1R01EY030854 to YX.
                Funded by: MVP was supported in part by NIH Intramural Research Program ZIA MH002035.
                Categories
                Article
                Custom metadata
                © The Author(s) 2021

                Uncategorized
                neural decoding,object vision
                Uncategorized
                neural decoding, object vision

                Comments

                Comment on this article