Literature DB >> 28889976

Humans, but Not Deep Neural Networks, Often Miss Giant Targets in Scenes.

Miguel P Eckstein1, Kathryn Koehler2, Lauren E Welbourne3, Emre Akbas4.   

Abstract

Even with great advances in machine vision, animals are still unmatched in their ability to visually search complex scenes. Animals from bees [1, 2] to birds [3] to humans [4-12] learn about the statistical relations in visual environments to guide and aid their search for targets. Here, we investigate a novel manner in which humans utilize rapidly acquired information about scenes by guiding search toward likely target sizes. We show that humans often miss targets when their size is inconsistent with the rest of the scene, even when the targets were made larger and more salient and observers fixated the target. In contrast, we show that state-of-the-art deep neural networks do not exhibit such deficits in finding mis-scaled targets but, unlike humans, can be fooled by target-shaped distractors that are inconsistent with the expected target's size within the scene. Thus, it is not a human deficiency to miss targets when they are inconsistent in size with the scene; instead, it is a byproduct of a useful strategy that the brain has implemented to rapidly discount potential distractors.
Copyright © 2017 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  computer vision; convolutional neural networks; deep neural networks; guided search; object detection; perception; scene context; search errors; visual attention; visual search

Mesh:

Year:  2017        PMID: 28889976     DOI: 10.1016/j.cub.2017.07.068

Source DB:  PubMed          Journal:  Curr Biol        ISSN: 0960-9822            Impact factor:   10.834


  13 in total

1.  Real-world size coding of solid objects, but not 2-D or 3-D images, in visual agnosia patients with bilateral ventral lesions.

Authors:  Desiree E Holler; Marlene Behrmann; Jacqueline C Snow
Journal:  Cortex       Date:  2019-03-09       Impact factor: 4.027

Review 2.  Beyond the feedforward sweep: feedback computations in the visual cortex.

Authors:  Gabriel Kreiman; Thomas Serre
Journal:  Ann N Y Acad Sci       Date:  2020-02-28       Impact factor: 5.691

3.  Local features and global shape information in object classification by deep convolutional neural networks.

Authors:  Nicholas Baker; Hongjing Lu; Gennady Erlikhman; Philip J Kellman
Journal:  Vision Res       Date:  2020-05-12       Impact factor: 1.886

4.  Saliency-Aware Subtle Augmentation Improves Human Visual Search Performance in VR.

Authors:  Olga Lukashova-Sanz; Siegfried Wahl
Journal:  Brain Sci       Date:  2021-02-25

5.  Cyborg groups enhance face recognition in crowded environments.

Authors:  Davide Valeriani; Riccardo Poli
Journal:  PLoS One       Date:  2019-03-06       Impact factor: 3.240

6.  Qualitative similarities and differences in visual object representations between brains and deep networks.

Authors:  Georgin Jacob; R T Pramod; Harish Katti; S P Arun
Journal:  Nat Commun       Date:  2021-03-25       Impact factor: 14.919

Review 7.  How context changes the neural basis of perception and language.

Authors:  Roel M Willems; Marius V Peelen
Journal:  iScience       Date:  2021-04-02

8.  Under-exploration of Three-Dimensional Images Leads to Search Errors for Small Salient Targets.

Authors:  Miguel A Lago; Aditya Jonnalagadda; Craig K Abbey; Bruno B Barufaldi; Predrag R Bakic; Andrew D A Maidment; Winifred K Leung; Susan P Weinstein; Brian S Englander; Miguel P Eckstein
Journal:  Curr Biol       Date:  2021-01-19       Impact factor: 10.834

9.  Object detection through search with a foveated visual system.

Authors:  Emre Akbas; Miguel P Eckstein
Journal:  PLoS Comput Biol       Date:  2017-10-09       Impact factor: 4.475

10.  Scenes Modulate Object Processing Before Interacting With Memory Templates.

Authors:  Surya Gayet; Marius V Peelen
Journal:  Psychol Sci       Date:  2019-09-16
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.