Literature DB >> 22006295

Comparing machines and humans on a visual categorization test.

François Fleuret1, Ting Li, Charles Dubout, Emma K Wampler, Steven Yantis, Donald Geman.   

Abstract

Automated scene interpretation has benefited from advances in machine learning, and restricted tasks, such as face detection, have been solved with sufficient accuracy for restricted settings. However, the performance of machines in providing rich semantic descriptions of natural scenes from digital images remains highly limited and hugely inferior to that of humans. Here we quantify this "semantic gap" in a particular setting: We compare the efficiency of human and machine learning in assigning an image to one of two categories determined by the spatial arrangement of constituent parts. The images are not real, but the category-defining rules reflect the compositional structure of real images and the type of "reasoning" that appears to be necessary for semantic parsing. Experiments demonstrate that human subjects grasp the separating principles from a handful of examples, whereas the error rates of computer programs fluctuate wildly and remain far behind that of humans even after exposure to thousands of examples. These observations lend support to current trends in computer vision such as integrating machine learning with parts-based modeling.

Entities:  

Mesh:

Year:  2011        PMID: 22006295      PMCID: PMC3203755          DOI: 10.1073/pnas.1109168108

Source DB:  PubMed          Journal:  Proc Natl Acad Sci U S A        ISSN: 0027-8424            Impact factor:   11.205


  4 in total

1.  Hierarchical models of object recognition in cortex.

Authors:  M Riesenhuber; T Poggio
Journal:  Nat Neurosci       Date:  1999-11       Impact factor: 24.884

2.  Object detection with discriminatively trained part-based models.

Authors:  Pedro F Felzenszwalb; Ross B Girshick; David McAllester; Deva Ramanan
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-09       Impact factor: 6.226

3.  A fast learning algorithm for deep belief nets.

Authors:  Geoffrey E Hinton; Simon Osindero; Yee-Whye Teh
Journal:  Neural Comput       Date:  2006-07       Impact factor: 2.026

4.  Learning the compositional nature of visual object categories for recognition.

Authors:  Björn Ommer; Joachim M Buhmann
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-03       Impact factor: 6.226

  4 in total
  18 in total

1.  Not-So-CLEVR: learning same-different relations strains feedforward neural networks.

Authors:  Junkyung Kim; Matthew Ricci; Thomas Serre
Journal:  Interface Focus       Date:  2018-06-15       Impact factor: 3.906

2.  Performance vs. competence in human-machine comparisons.

Authors:  Chaz Firestone
Journal:  Proc Natl Acad Sci U S A       Date:  2020-10-13       Impact factor: 11.205

Review 3.  How Memory Replay in Sleep Boosts Creative Problem-Solving.

Authors:  Penelope A Lewis; Günther Knoblich; Gina Poe
Journal:  Trends Cogn Sci       Date:  2018-06       Impact factor: 20.229

Review 4.  Beyond the feedforward sweep: feedback computations in the visual cortex.

Authors:  Gabriel Kreiman; Thomas Serre
Journal:  Ann N Y Acad Sci       Date:  2020-02-28       Impact factor: 5.691

5.  Modeling invariant object processing based on tight integration of simulated and empirical data in a Common Brain Space.

Authors:  Judith C Peters; Joel Reithler; Rainer Goebel
Journal:  Front Comput Neurosci       Date:  2012-03-09       Impact factor: 2.380

6.  The Microphenotron: a robotic miniaturized plant phenotyping platform with diverse applications in chemical biology.

Authors:  Thomas Burrell; Susan Fozard; Geoff H Holroyd; Andrew P French; Michael P Pound; Christopher J Bigley; C James Taylor; Brian G Forde
Journal:  Plant Methods       Date:  2017-03-01       Impact factor: 4.993

7.  Invariant object recognition is a personalized selection of invariant features in humans, not simply explained by hierarchical feed-forward vision models.

Authors:  Hamid Karimi-Rouzbahani; Nasour Bagheri; Reza Ebrahimpour
Journal:  Sci Rep       Date:  2017-10-31       Impact factor: 4.379

8.  Ultra-rapid object categorization in real-world scenes with top-down manipulations.

Authors:  Bingjie Xu; Mohan S Kankanhalli; Qi Zhao
Journal:  PLoS One       Date:  2019-04-10       Impact factor: 3.240

9.  I spy with my little eye … a knee about to go 'pop'? Can coaches and sports medicine professionals predict who is at greater risk of ACL rupture?

Authors:  Anne Inger Mørtvedt; Tron Krosshaug; Roald Bahr; Erich Petushek
Journal:  Br J Sports Med       Date:  2019-10-14       Impact factor: 13.800

10.  Explaining the Timing of Natural Scene Understanding with a Computational Model of Perceptual Categorization.

Authors:  Imri Sofer; Sébastien M Crouzet; Thomas Serre
Journal:  PLoS Comput Biol       Date:  2015-09-03       Impact factor: 4.475

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.