Literature DB >> 25755262

Visual Turing test for computer vision systems.

Donald Geman1, Stuart Geman2, Neil Hallonquist1, Laurent Younes1.   

Abstract

Today, computer vision systems are tested by their accuracy in detecting and localizing instances of objects. As an alternative, and motivated by the ability of humans to provide far richer descriptions and even tell a story about an image, we construct a "visual Turing test": an operator-assisted device that produces a stochastic sequence of binary questions from a given test image. The query engine proposes a question; the operator either provides the correct answer or rejects the question as ambiguous; the engine proposes the next question ("just-in-time truthing"). The test is then administered to the computer-vision system, one question at a time. After the system's answer is recorded, the system is provided the correct answer and the next question. Parsing is trivial and deterministic; the system being tested requires no natural language processing. The query engine employs statistical constraints, learned from a training set, to produce questions with essentially unpredictable answers-the answer to a question, given the history of questions and their correct answers, is nearly equally likely to be positive or negative. In this sense, the test is only about vision. The system is designed to produce streams of questions that follow natural story lines, from the instantiation of a unique object, through an exploration of its properties, and on to its relationships with other uniquely instantiated objects.

Entities:  

Keywords:  Turing test; binary questions; computer vision; scene interpretation; unpredictable answers

Mesh:

Year:  2015        PMID: 25755262      PMCID: PMC4378453          DOI: 10.1073/pnas.1422953112

Source DB:  PubMed          Journal:  Proc Natl Acad Sci U S A        ISSN: 0027-8424            Impact factor:   11.205


  2 in total

1.  Object detection with discriminatively trained part-based models.

Authors:  Pedro F Felzenszwalb; Ross B Girshick; David McAllester; Deva Ramanan
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-09       Impact factor: 6.226

2.  A fast learning algorithm for deep belief nets.

Authors:  Geoffrey E Hinton; Simon Osindero; Yee-Whye Teh
Journal:  Neural Comput       Date:  2006-07       Impact factor: 2.026

  2 in total
  14 in total

1.  Brain-inspired automated visual object discovery and detection.

Authors:  Lichao Chen; Sudhir Singh; Thomas Kailath; Vwani Roychowdhury
Journal:  Proc Natl Acad Sci U S A       Date:  2018-12-17       Impact factor: 11.205

2.  Not-So-CLEVR: learning same-different relations strains feedforward neural networks.

Authors:  Junkyung Kim; Matthew Ricci; Thomas Serre
Journal:  Interface Focus       Date:  2018-06-15       Impact factor: 3.906

3.  Performance vs. competence in human-machine comparisons.

Authors:  Chaz Firestone
Journal:  Proc Natl Acad Sci U S A       Date:  2020-10-13       Impact factor: 11.205

Review 4.  Beyond the feedforward sweep: feedback computations in the visual cortex.

Authors:  Gabriel Kreiman; Thomas Serre
Journal:  Ann N Y Acad Sci       Date:  2020-02-28       Impact factor: 5.691

5.  Constrained generative adversarial network ensembles for sharable synthetic medical images.

Authors:  Engin Dikici; Matthew Bigelow; Richard D White; Barbaros S Erdal; Luciano M Prevedello
Journal:  J Med Imaging (Bellingham)       Date:  2021-04-10

6.  Deep convolutional networks do not classify based on global object shape.

Authors:  Nicholas Baker; Hongjing Lu; Gennady Erlikhman; Philip J Kellman
Journal:  PLoS Comput Biol       Date:  2018-12-07       Impact factor: 4.475

7.  HEMIGEN: Human Embryo Image Generator Based on Generative Adversarial Networks.

Authors:  Darius Dirvanauskas; Rytis Maskeliūnas; Vidas Raudonis; Robertas Damaševičius; Rafal Scherer
Journal:  Sensors (Basel)       Date:  2019-08-16       Impact factor: 3.576

8.  Automated interpretation of the coronary angioscopy with deep convolutional neural networks.

Authors:  Toru Miyoshi; Akinori Higaki; Hideo Kawakami; Osamu Yamaguchi
Journal:  Open Heart       Date:  2020-05

9.  Linguistic issues behind visual question answering.

Authors:  Raffaella Bernardi; Sandro Pezzelle
Journal:  Lang Linguist Compass       Date:  2021-06-04

10.  Facial color is an efficient mechanism to visually transmit emotion.

Authors:  Carlos F Benitez-Quiroz; Ramprakash Srinivasan; Aleix M Martinez
Journal:  Proc Natl Acad Sci U S A       Date:  2018-03-19       Impact factor: 11.205

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.