| Literature DB >> 32140517 |
Yaniv Morgenstern1, Filipp Schmidt1, Roland W Fleming1.
Abstract
With the advent of deep convolutional neural networks, machines now rival humans in terms of object categorization. The neural networks solve categorization with a hierarchical organization that shares a striking resemblance to their biological counterpart, leading to their status as a standard model of object recognition in biological vision. Despite training on thousands of images of object categories, however, machine-learning networks are poorer generalizers, often fooled by adversarial images with very simple image manipulations that humans easily distinguish as a false image. Humans, on the other hand, can generalize object classes from very few samples. Here we provide a dataset of novel object classifications in humans. We gathered thousands of crowd-sourced human responses to novel objects embedded either with 1 or 16 context sample(s). Human decisions and stimuli together have the potential to be re-used (1) as a tool to better understand the nature of the gap in category learning from few samples between human and machine, and (2) as a benchmark of generalization across machine learning networks.Entities:
Keywords: Abstraction; Categorization; Generalization; Machine vision; Objects; One-shot learning; Shape; Visual perception
Year: 2020 PMID: 32140517 PMCID: PMC7044642 DOI: 10.1016/j.dib.2020.105302
Source DB: PubMed Journal: Data Brief ISSN: 2352-3409
Fig. 1Task and Example Stimuli. Crowdsourced observers judged whether the test (embedded in circle) was in the same categorical class as the surrounding sample(s). In 20 trials, each observer (500 total) judged unique combinations of test and sample(s) to 20 novel object classes (only 4 shown). On each trial, the stimulus frame showed the test embedded in 1 (A, B) or 16 (C,D) surrounding samples in object classes that tended to vary more (C,B) or less (A,D). The test itself varied in its appearance from the samples in 1 of 25 bins based on its skeletal similarity to an original base shape (see Stimulus Frame in Experimental Design, Materials, and Methods). The stimulus frames and human responses are available through Zenedo. By varying data collection parameters listed in the Specifications Table (obj_num, cont_num, distbin_num, and shape_num), the MATLAB code oneSHOTdata_DEMO.m shows how to generate a stimulus frame and recover the corresponding human response from the raw data for an individual trial number. As examples, stimulus frame (A) was rendered with object number 2 (obj_num = 2, which had surrounding samples that tended to appear more similar, lower variability), with only 1 surrounding sample (cont_num = 1), with a test that had a smaller skeletal parameter distance from the base shape (distbin_num = 8), and a shape 3 of 10 for this condition (shape_num = 3) while stimulus frame (B) was rendered with object number 1 (obj_num = 1, with higher variability across samples), with 16 surrounding samples (cont_num = 16), with a test that had a large skeletal parameter distance from the base shapes (distbin_num = 25), and was shape 2 out 10 for this condition (shape_num = 2). In (C) obj_num = 1, cont_num = 16, distbin_num = 25, and shape_num = 1. In (D) obj_num = 5, cont_num = 1, distbin_num = 12, and shape_num = 6.
Specifications Table
| Subject | Experimental and Cognitive Psychology |
| Specific subject area | Vision and Perception |
| Type of data | MATLAB code (that analyses raw data) |
| How data were acquired | Clickworker (crowd-sourcing platform) |
| Data format | Raw data (*.mat file) and MATLAB code to analyse raw data |
| Parameters for data collection | Independent factors (see also |
| Description of data collection | Data were collected with an online crowd-sourcing platform (clickworker). In the main experiment, 500 observers responded to 22 trials indicating whether a central test was in the same class as the samples, which consisted of 1 or 16 sample shapes that varied in their similarity to the test. |
| Data source location | Giessen, Hessen |
| Data accessibility | Repository name: |
| Related research article | Morgenstern, Y., Schmidt, F., & Fleming, R. W. (2019). One-shot categorization of novel object classes in humans. |
From just a single example, we can derive quite precise intuitions about what other class members look like. This stands in stark contrast to machine learning algorithms, which typically require tens or even hundreds of thousands of examples to learn a new category. One of the most important open questions in our field is: How do humans achieve this? The stimuli and data provided in this paper can be used to test machine learning generalization as compared to human and also can be used as a test bed for various kinds of category learning models. These data will benefit cognitive and machine learning scientists interested in testing how their category learning theories and algorithms transfer to the human perception of novel object classes from few samples. Specifically, these stimuli can be analyzed with a model according to an experimenter's preference (e.g., by calculated similarity between the test and sample(s) in terms of image computable features) to produce responses that can be compared to human responses. In this way, the data can be used to test an experimenter's theoretical ideas on how humans learn from few samples. These data can also be used as stimuli to examine category learning in the human brain (e.g., using fMRI to investigate brain activity changes as a function of test similarity to the samples) |