| Literature DB >> 35169072 |
Céline Spriet1, Etienne Abassi1, Jean-Rémy Hochmann1, Liuba Papeo2.
Abstract
Humans make sense of the world by organizing things into categories. When and how does this process begin? We investigated whether real-world object categories that spontaneously emerge in the first months of life match categorical representations of objects in the human visual cortex. Using eye tracking, we measured the differential looking time of 4-, 10-, and 19-mo-olds as they looked at pairs of pictures belonging to eight animate or inanimate categories (human/nonhuman, faces/bodies, real-world size big/small, natural/artificial). Taking infants' looking times as a measure of similarity, for each age group, we defined a representational space where each object was defined in relation to others of the same or of a different category. This space was compared with hypothesis-based and functional MRI-based models of visual object categorization in the adults' visual cortex. Analyses across different age groups showed that, as infants grow older, their looking behavior matches neural representations in ever-larger portions of the adult visual cortex, suggesting progressive recruitment and integration of more and more feature spaces distributed over the visual cortex. Moreover, the results characterize infants' visual categorization as an incremental process with two milestones. Between 4 and 10 mo, visual exploration guided by saliency gives way to an organization according to the animate-inanimate distinction. Between 10 and 19 mo, a category spurt leads toward a mature organization. We propose that these changes underlie the coupling between seeing and thinking in the developing mind.Entities:
Keywords: categorization; cognitive development; fMRI; looking times; visual system
Mesh:
Year: 2022 PMID: 35169072 PMCID: PMC8872728 DOI: 10.1073/pnas.2105866119
Source DB: PubMed Journal: Proc Natl Acad Sci U S A ISSN: 0027-8424 Impact factor: 12.779
Fig. 1.Stimuli, trials, and hypothesis-based models of categorization considered in the design of Exps. 1 and 2. (A) Stimuli were 72 images depicting 9 objects from each of 8 different categories. Silhouettes instead of the actual colorful female human faces used in the experiments, are shown for illustration. (B) In each trial of Exp. 1, two images were presented within two gray frames of identical size, on the right and on the left, equally distant from the center of the screen. (C) In each trial of Exp. 2, the image frame was removed and the image size was modified so that each object had the same number of pixels. (D) Hypothesis-driven (synthetic) models reflecting the categorical object representations tested in the current design. (E) The composite model reflecting the mean of the six synthetic models.
Fig. 2.Results of representational similarity analysis and of the pairwise comparisons of MLT between- and within-categories for each age group in Exps. 1 and 2. (Left) Mean RDM reflecting dissimilarities between- and within-categories in terms of DLTs. Black squares in the RDMs highlight categorization by animacy, humanness, and by the eight categories in 19-mo-olds (A), categorization by animacy in 10-mo-olds (B), and in 4-mo-olds of Exp. 2 (D). (Center) Matrix of t-values for each pairwise comparison between MLTs of the individual categories for 19- (A), 10- (B), and 4-mo-olds (C) in Exp. 1 and 4-mo-olds in Exp. 2 (D). Squares in dark blue denote significant effects; squares in lighter blue denote effects that did not survive the multiple comparison correction (trends); red squares denote nonsignificant (n.s.) or nontested comparisons. (Right) Distribution of MLTs in 19- (A), 10- (B), and 4-mo-olds (C) of Exp. 1 and of 4-mo-olds of Exp. 2 (D). Box-plots represent the minimum, the first quartile, the median, the third quartile and the maximum of the population distribution; outliers are denoted by dots (one in the 19-mo-old group).
Fig. 3.Relationship between infants’ looking behavior and visual object representation in the adults’ visual cortex. (A) ROIs in the adults’ brain: EVC, VOTC, and LOTC. (B) Mean RDMs reflecting relationships (i.e., dissimilarities) between object categories in terms of dissimilarities in the neural activity patterns evoked by viewing objects in the EVC, VOTC, and LOTC of adults (fMRI-based RDMs). (C) Results of the representational similarity analysis between the mean fMRI-based RDM in each ROI and the DLT-RDM of each infant in each age group of Exps. 1 and 2. Box-plots represent the minimum, the first quartile, the median, the third quartile, and the maximum of the population distribution as well as outliers (dots); *P < 0.017; ***P < 0.0003. (D) Results of the representational similarity analysis between the infants’ DLT-RDMs and the fMRI-based RDM derived from each partition along the ventral visual stream. Solid bars represent clusters with significant correlation (above 0) for each age group of Exps. 1 and 2.
Results of the representational similarity analysis reflecting the relationship of the infants’ DLT-RDMs with the fMRI-based RDMs, the RDM for the synthetic composite model of categorization, and the RDMs based on size, elongation, compactness, and color differences
| Exp. | Age (mo) | Model | Mean ρ (SD) | CI (minimum to maximum) |
| Cohen’s | |
| 1 | 19 |
|
|
|
|
|
|
|
|
|
|
|
|
| ||
|
|
|
|
|
|
| ||
|
|
|
|
|
|
| ||
| Size | −0.034 (0.194) | −0.139 to 0.071 | −0.874 (24) | n.s. | 0.175 | ||
| Elongation | −0.010 (0.151) | −0.092 to 0.071 | −0.333 (24) | n.s. | 0.066 | ||
| Compactness | −0.074 (0.166) | −0.163 to 0.016 | −2.218 (24) | 0.036 | 0.444 | ||
| Color | 0.009 (0.186) | −0.092 to 0.109 | 0.234 (24) | n.s. | 0.048 | ||
| 1 | 10 | EVC | 0.053 (0.252) | −0.079 to 0.185 | 1.029 (23) | 0.314 | 0.210 |
| VOTC | 0.077 (0.179) | −0.017 to 0.171 | 2.112 (23) | 0.046 | 0.431 | ||
| LOTC | 0.086 (0.200) | −0.019 to 0.191 | 2.102 (23) | 0.047 | 0.429 | ||
|
|
|
|
|
|
| ||
| Size | 0.068 (0.208) | −0.048 to 0.183 | 1.592 (23) | 0.125 | 0.327 | ||
| Elongation | −0.041 (0.197) | −0.150 to 0.068 | −1.013 (23) | n.s. | 0.208 | ||
| Compactness | 0.054 (0.190) | −0.052 to 0.159 | 1.379 (23) | 0.181 | 0.281 | ||
| Color | 0.044 (0.253) | −0.096 to 0.183 | 0.843 (23) | n.s. | 0.174 | ||
| 1 | 4 | EVC | −0.067 (0.201) | −0.172 to 0.039 | −1.629 (23) | 0.117 | 0.333 |
| VOTC | 0.033 (0.174) | −0.058 to 0.124 | 0.929 (23) | n.s. | 0.190 | ||
| LOTC | −0.025 (0.184) | −0.122 to 0.071 | −0.673 (23) | n.s. | 0.137 | ||
| CM | 0.049 (0.182) | −0.028 to 0.126 | 1.320 (23) | 0.200 | 0.269 | ||
|
|
|
|
|
|
| ||
|
|
|
|
|
| |||
|
|
|
|
|
|
| ||
| Color | 0.035 (0.204) | −0.078 to 0.148 | 0.836 (23) | n.s. | 0.172 | ||
| 2 | 4 m | EVC | −0.026 (0.199) | −0.131 to 0.079 | −0.642 (23) | n.s. | 0.131 |
| VOTC | 0.051 (0.191) | −0.050 to 0.151 | 1.295 (23) | 0.208 | 0.264 | ||
| LOTC | 0.017 (0.184) | −0.080 to 0.114 | 0.453 (23) | n.s. | 0.092 | ||
|
|
|
|
|
|
| ||
|
|
|
|
|
| |||
|
|
|
|
|
|
| ||
| Color | 0.068 (0.182) | −0.033 to 0.169 | 1.837 (23) | 0.079 | 0.374 |
CM, composite model; mean ρ are the Fisher-transformed ρ; CI, 98.3% confidence interval for EVC, VOTC and LOTC; 95% CI for CM; 98.8% CI for size, elongation, compactness, and color. Highlighted in bold are the significant results; α = 0.017 for EVC, VOTC, and LOTC; α = 0.05 for CM; α = 0.0125 for size, elongation, compactness, and color; n.s.= nonsignificant results with P > 0.250.