40
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Unsupervised deep learning identifies semantic disentanglement in single inferotemporal face patch neurons

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          In order to better understand how the brain perceives faces, it is important to know what objective drives learning in the ventral visual stream. To answer this question, we model neural responses to faces in the macaque inferotemporal (IT) cortex with a deep self-supervised generative model, β-VAE, which disentangles sensory data into interpretable latent factors, such as gender or age. Our results demonstrate a strong correspondence between the generative factors discovered by β-VAE and those coded by single IT neurons, beyond that found for the baselines, including the handcrafted state-of-the-art model of face perception, the Active Appearance Model, and deep classifiers. Moreover, β-VAE is able to reconstruct novel face images using signals from just a handful of cells. Together our results imply that optimising the disentangling objective leads to representations that closely resemble those in the IT at the single unit level. This points at disentangling as a plausible learning objective for the visual brain.

          Abstract

          Little is known about the brain’s computations that enable the recognition of faces. Here, the authors use unsupervised deep learning to show that the brain disentangles faces into semantically meaningful factors, like age or the presence of a smile, at the single neuron level.

          Related collections

          Most cited references48

          • Record: found
          • Abstract: found
          • Article: not found

          Reducing the dimensionality of data with neural networks.

          High-dimensional data can be converted to low-dimensional codes by training a multilayer neural network with a small central layer to reconstruct high-dimensional input vectors. Gradient descent can be used for fine-tuning the weights in such "autoencoder" networks, but this works well only if the initial weights are close to a good solution. We describe an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data.
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            The free-energy principle: a unified brain theory?

            A free-energy principle has been proposed recently that accounts for action, perception and learning. This Review looks at some key brain theories in the biological (for example, neural Darwinism) and physical (for example, information theory and optimal control theory) sciences from the free-energy perspective. Crucially, one key theme runs through each of these theories - optimization. Furthermore, if we look closely at what is optimized, the same quantity keeps emerging, namely value (expected reward, expected utility) or its complement, surprise (prediction error, expected cost). This is the quantity that is optimized under the free-energy principle, which suggests that several global brain theories might be unified within a free-energy framework.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Representation learning: a review and new perspectives.

              The success of machine learning algorithms generally depends on data representation, and we hypothesize that this is because different representations can entangle and hide more or less the different explanatory factors of variation behind the data. Although specific domain knowledge can be used to help design representations, learning with generic priors can also be used, and the quest for AI is motivating the design of more powerful representation-learning algorithms implementing such priors. This paper reviews recent work in the area of unsupervised feature learning and deep learning, covering advances in probabilistic models, autoencoders, manifold learning, and deep networks. This motivates longer term unanswered questions about the appropriate objectives for learning good representations, for computing representations (i.e., inference), and the geometrical connections between representation learning, density estimation, and manifold learning.
                Bookmark

                Author and article information

                Contributors
                irinah@google.com
                Journal
                Nat Commun
                Nat Commun
                Nature Communications
                Nature Publishing Group UK (London )
                2041-1723
                9 November 2021
                9 November 2021
                2021
                : 12
                : 6456
                Affiliations
                [1 ]GRID grid.498210.6, ISNI 0000 0004 5999 1726, DeepMind, ; London, UK
                [2 ]GRID grid.20861.3d, ISNI 0000000107068890, Caltech, ; Pasadena, USA
                [3 ]GRID grid.9227.e, ISNI 0000000119573309, Institute of Neuroscience, CAS Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, ; Shanghai, China
                [4 ]GRID grid.83440.3b, ISNI 0000000121901201, University College London, ; London, UK
                [5 ]GRID grid.4991.5, ISNI 0000 0004 1936 8948, University of Oxford, ; Oxford, UK
                [6 ]GRID grid.413575.1, ISNI 0000 0001 2167 1581, Howard Hughes Medical Institute, ; Pasadena, USA
                Author information
                http://orcid.org/0000-0002-1890-2091
                http://orcid.org/0000-0003-2812-9917
                http://orcid.org/0000-0003-1083-1919
                http://orcid.org/0000-0001-7758-6896
                Article
                26751
                10.1038/s41467-021-26751-5
                8578601
                34753913
                90e96d16-d6f2-4d5e-9aee-c0c86adb16e2
                © The Author(s) 2021

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 15 December 2020
                : 22 October 2021
                Categories
                Article
                Custom metadata
                © The Author(s) 2021

                Uncategorized
                neuroscience,computational neuroscience,visual system,object vision
                Uncategorized
                neuroscience, computational neuroscience, visual system, object vision

                Comments

                Comment on this article