Literature DB >> 9798003

The viewpoint complexity of an object-recognition task.

B S Tjan1, G E Legge.   

Abstract

There is an ongoing debate about the nature of perceptual representation in human object recognition. Resolution of this debate has been hampered by the lack of a metric for assessing the representational requirements of a recognition task. To recognize a member of a given set of 3-D objects, how much detail must the objects' representations contain in order to achieve a specific accuracy criterion? From the performance of an ideal observer, we derived a quantity called the view complexity (VX) to measure the required granularity of representation. VX is an intrinsic property of the object-recognition task, taking into account both the object ensemble and the type of decision required of an observer. It does not depend on the visual representation or processing used by the observer. VX can be interpreted as the number of randomly selected 2-D images needed to represent the decision boundaries in the image space of a 3-D object-recognition task. A low VX means the task is inherently more viewpoint invariant and a high VX means it is inherently more viewpoint dependent. By measuring the VX of recognition tasks with different object sets, we show that the current confusion about the nature of human perceptual representation is partly due to a failure in distinguishing between human visual processing and the properties of a task and its stimuli. We find general correspondence between the VX of a recognition task and the published human data on viewpoint dependence. Exceptions in this relationship motivated us to propose the view-rate hypothesis: human visual performance is limited by the equivalent number of 2-D image views that can be processed per unit time.

Entities:  

Mesh:

Year:  1998        PMID: 9798003     DOI: 10.1016/s0042-6989(97)00255-1

Source DB:  PubMed          Journal:  Vision Res        ISSN: 0042-6989            Impact factor:   1.886


  15 in total

1.  Classification images with uncertainty.

Authors:  Bosco S Tjan; Anirvan S Nandy
Journal:  J Vis       Date:  2006-04-04       Impact factor: 2.240

2.  Shift in spatial scale in identifying crowded letters.

Authors:  Susana T L Chung; Bosco S Tjan
Journal:  Vision Res       Date:  2007-01-16       Impact factor: 1.886

3.  Do image descriptions underlie word recognition in reading?

Authors:  Gordon E Legge; Bosco S Tjan; Susana T L Chung; Charles Bigelow
Journal:  Br J Psychol       Date:  2009-10-23

4.  Generalization between canonical and non-canonical views in object recognition.

Authors:  Tandra Ghose; Zili Liu
Journal:  J Vis       Date:  2013-01-02       Impact factor: 2.240

5.  Template changes with perceptual learning are driven by feature informativeness.

Authors:  Ilmari Kurki; Miguel P Eckstein
Journal:  J Vis       Date:  2014-09-05       Impact factor: 2.240

6.  Ideal observer analysis of crowding and the reduction of crowding through learning.

Authors:  Gerald J Sun; Susana T L Chung; Bosco S Tjan
Journal:  J Vis       Date:  2010-05-01       Impact factor: 2.240

7.  Simple Learned Weighted Sums of Inferior Temporal Neuronal Firing Rates Accurately Predict Human Core Object Recognition Performance.

Authors:  Najib J Majaj; Ha Hong; Ethan A Solomon; James J DiCarlo
Journal:  J Neurosci       Date:  2015-09-30       Impact factor: 6.167

8.  Using geometric moments to explain human letter recognition near the acuity limit.

Authors:  Lei Liu; Stanley A Klein; Feng Xue; Jun-Yun Zhang; Cong Yu
Journal:  J Vis       Date:  2009-01-21       Impact factor: 2.240

9.  Dynamics of 3D view invariance in monkey inferotemporal cortex.

Authors:  N Apurva Ratan Murty; Sripati P Arun
Journal:  J Neurophysiol       Date:  2015-01-21       Impact factor: 2.714

10.  Parts, wholes, and context in reading: a triple dissociation.

Authors:  Denis G Pelli; Katharine A Tillman
Journal:  PLoS One       Date:  2007-08-01       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.