Literature DB >> 30006365

Large-Scale, High-Resolution Comparison of the Core Visual Object Recognition Behavior of Humans, Monkeys, and State-of-the-Art Deep Artificial Neural Networks.

Rishi Rajalingham1, Elias B Issa1, Pouya Bashivan1, Kohitij Kar1, Kailyn Schmidt1, James J DiCarlo2.   

Abstract

Primates, including humans, can typically recognize objects in visual images at a glance despite naturally occurring identity-preserving image transformations (e.g., changes in viewpoint). A primary neuroscience goal is to uncover neuron-level mechanistic models that quantitatively explain this behavior by predicting primate performance for each and every image. Here, we applied this stringent behavioral prediction test to the leading mechanistic models of primate vision (specifically, deep, convolutional, artificial neural networks; ANNs) by directly comparing their behavioral signatures against those of humans and rhesus macaque monkeys. Using high-throughput data collection systems for human and monkey psychophysics, we collected more than one million behavioral trials from 1472 anonymous humans and five male macaque monkeys for 2400 images over 276 binary object discrimination tasks. Consistent with previous work, we observed that state-of-the-art deep, feedforward convolutional ANNs trained for visual categorization (termed DCNNIC models) accurately predicted primate patterns of object-level confusion. However, when we examined behavioral performance for individual images within each object discrimination task, we found that all tested DCNNIC models were significantly nonpredictive of primate performance and that this prediction failure was not accounted for by simple image attributes nor rescued by simple model modifications. These results show that current DCNNIC models cannot account for the image-level behavioral patterns of primates and that new ANN models are needed to more precisely capture the neural mechanisms underlying primate object vision. To this end, large-scale, high-resolution primate behavioral benchmarks such as those obtained here could serve as direct guides for discovering such models.SIGNIFICANCE STATEMENT Recently, specific feedforward deep convolutional artificial neural networks (ANNs) models have dramatically advanced our quantitative understanding of the neural mechanisms underlying primate core object recognition. In this work, we tested the limits of those ANNs by systematically comparing the behavioral responses of these models with the behavioral responses of humans and monkeys at the resolution of individual images. Using these high-resolution metrics, we found that all tested ANN models significantly diverged from primate behavior. Going forward, these high-resolution, large-scale primate behavioral benchmarks could serve as direct guides for discovering better ANN models of the primate visual system.
Copyright © 2018 the authors 0270-6474/18/387255-15$15.00/0.

Entities:  

Keywords:  deep neural network; human; monkey; object recognition; vision

Mesh:

Year:  2018        PMID: 30006365      PMCID: PMC6096043          DOI: 10.1523/JNEUROSCI.0388-18.2018

Source DB:  PubMed          Journal:  J Neurosci        ISSN: 0270-6474            Impact factor:   6.167


  30 in total

Review 1.  Neural coding and the basic law of psychophysics.

Authors:  Kenneth O Johnson; Steven S Hsiao; Takashi Yoshioka
Journal:  Neuroscientist       Date:  2002-04       Impact factor: 7.519

2.  Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream.

Authors:  Umut Güçlü; Marcel A J van Gerven
Journal:  J Neurosci       Date:  2015-07-08       Impact factor: 6.167

3.  Performance-optimized hierarchical models predict neural responses in higher visual cortex.

Authors:  Daniel L K Yamins; Ha Hong; Charles F Cadieu; Ethan A Solomon; Darren Seibert; James J DiCarlo
Journal:  Proc Natl Acad Sci U S A       Date:  2014-05-08       Impact factor: 11.205

4.  Explicit information for category-orthogonal object properties increases along the ventral stream.

Authors:  Ha Hong; Daniel L K Yamins; Najib J Majaj; James J DiCarlo
Journal:  Nat Neurosci       Date:  2016-02-22       Impact factor: 24.884

5.  Seeing it all: Convolutional network layers map the function of the human visual system.

Authors:  Michael Eickenberg; Alexandre Gramfort; Gaël Varoquaux; Bertrand Thirion
Journal:  Neuroimage       Date:  2016-10-21       Impact factor: 6.556

6.  Deep convolutional models improve predictions of macaque V1 responses to natural images.

Authors:  Santiago A Cadena; George H Denfield; Edgar Y Walker; Leon A Gatys; Andreas S Tolias; Matthias Bethge; Alexander S Ecker
Journal:  PLoS Comput Biol       Date:  2019-04-23       Impact factor: 4.475

7.  Convolutional neural network-based encoding and decoding of visual object recognition in space and time.

Authors:  K Seeliger; M Fritsche; U Güçlü; S Schoenmakers; J-M Schoffelen; S E Bosch; M A J van Gerven
Journal:  Neuroimage       Date:  2017-07-16       Impact factor: 6.556

8.  Deep Networks Can Resemble Human Feed-forward Vision in Invariant Object Recognition.

Authors:  Saeed Reza Kheradpisheh; Masoud Ghodrati; Mohammad Ganjtabesh; Timothée Masquelier
Journal:  Sci Rep       Date:  2016-09-07       Impact factor: 4.379

9.  Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence.

Authors:  Radoslaw Martin Cichy; Aditya Khosla; Dimitrios Pantazis; Antonio Torralba; Aude Oliva
Journal:  Sci Rep       Date:  2016-06-10       Impact factor: 4.379

10.  Dynamics of scene representations in the human brain revealed by magnetoencephalography and deep neural networks.

Authors:  Radoslaw Martin Cichy; Aditya Khosla; Dimitrios Pantazis; Aude Oliva
Journal:  Neuroimage       Date:  2016-04-01       Impact factor: 6.556

View more
  57 in total

1.  Examining the Coding Strength of Object Identity and Nonidentity Features in Human Occipito-Temporal Cortex and Convolutional Neural Networks.

Authors:  Yaoda Xu; Maryam Vaziri-Pashkam
Journal:  J Neurosci       Date:  2021-03-31       Impact factor: 6.167

2.  Do Primates and Deep Artificial Neural Networks Perform Object Categorization in a Similar Manner?

Authors:  Prabaha Gangopadhyay; Jhilik Das
Journal:  J Neurosci       Date:  2019-02-06       Impact factor: 6.167

Review 3.  Artificial cognition: How experimental psychology can help generate explainable artificial intelligence.

Authors:  J Eric T Taylor; Graham W Taylor
Journal:  Psychon Bull Rev       Date:  2020-11-06

Review 4.  If deep learning is the answer, what is the question?

Authors:  Andrew Saxe; Stephanie Nelli; Christopher Summerfield
Journal:  Nat Rev Neurosci       Date:  2020-11-16       Impact factor: 34.870

5.  Controversial stimuli: Pitting neural networks against each other as models of human cognition.

Authors:  Tal Golan; Prashant C Raju; Nikolaus Kriegeskorte
Journal:  Proc Natl Acad Sci U S A       Date:  2020-11-24       Impact factor: 11.205

6.  Unsupervised changes in core object recognition behavior are predicted by neural plasticity in inferior temporal cortex.

Authors:  Xiaoxuan Jia; Ha Hong; James J DiCarlo
Journal:  Elife       Date:  2021-06-11       Impact factor: 8.140

7.  What do adversarial images tell us about human vision?

Authors:  Marin Dujmović; Gaurav Malhotra; Jeffrey S Bowers
Journal:  Elife       Date:  2020-09-02       Impact factor: 8.140

8.  Performance vs. competence in human-machine comparisons.

Authors:  Chaz Firestone
Journal:  Proc Natl Acad Sci U S A       Date:  2020-10-13       Impact factor: 11.205

Review 9.  Beyond the feedforward sweep: feedback computations in the visual cortex.

Authors:  Gabriel Kreiman; Thomas Serre
Journal:  Ann N Y Acad Sci       Date:  2020-02-28       Impact factor: 5.691

10.  Bridging the gap between artificial vision and touch.

Authors:  Giulia Pasquale
Journal:  Nature       Date:  2019-05       Impact factor: 49.962

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.