| Literature DB >> 33723064 |
Rahma Chaabouni1,2, Eugene Kharitonov3, Emmanuel Dupoux3,2, Marco Baroni3,4.
Abstract
Words categorize the semantic fields they refer to in ways that maximize communication accuracy while minimizing complexity. Focusing on the well-studied color domain, we show that artificial neural networks trained with deep-learning techniques to play a discrimination game develop communication systems whose distribution on the accuracy/complexity plane closely matches that of human languages. The observed variation among emergent color-naming systems is explained by different degrees of discriminative need, of the sort that might also characterize different human communities. Like human languages, emergent systems show a preference for relatively low-complexity solutions, even at the cost of imperfect communication. We demonstrate next that the nature of the emergent systems crucially depends on communication being discrete (as is human word usage). When continuous message passing is allowed, emergent systems become more complex and eventually less efficient. Our study suggests that efficient semantic categorization is a general property of discrete communication systems, not limited to human language. It suggests moreover that it is exactly the discrete nature of such systems that, acting as a bottleneck, pushes them toward low complexity and optimal efficiency.Entities:
Keywords: color-naming systems; efficiency of human language; language emergence in artificial neural networks
Year: 2021 PMID: 33723064 PMCID: PMC8000426 DOI: 10.1073/pnas.2016569118
Source DB: PubMed Journal: Proc Natl Acad Sci U S A ISSN: 0027-8424 Impact factor: 12.779
Fig. 3.Human (blue circles) and NN (orange circles) color-naming systems on the information plane. English (light blue circle) is not in WCS, but it is approximated relying on Zaslavsky et al. (SI Appendix, figure S7 in ref. 7). The IB curve (black line) defines the theoretical limit on accuracy given complexity. All color-naming systems achieve near-optimal efficiency.
Fig. 1.The 330 WCS color chips. Rows correspond to equally spaced lightness values and columns to equally spaced Munsell hues. Each stimulus is at the maximum available saturation for that hue/lightness combination.
Fig. 2.A successful round of the discrimination game. A chip is drawn from a uniform distribution and fed to Speaker. Speaker outputs a probability distribution over its vocabulary of size . Here, a probability is mapped to a color according to a gray gradient (with darker colors representing higher probabilities). A word is sampled from and fed to Listener. Finally, Listener—given , the target chip (in position 1 in this illustration), and a distractor chip (in position 2 in this illustration)—assigns a probability to both positions, representing its guess about the position of the target (in this illustration, Listener correctly assigns a higher probability to the target position).
Fig. 4.Complexity distributions of NN systems across different discriminative needs (human distribution included for comparison). There is a decreasing trend in complexity when increasing percentile (P = 0.004; Kruskal–Wallis). Pairwise differences are not significant when evaluated with Bonferroni-corrected Mann–Whitney–Wilcoxon.
Complexity and success rate (game accuracy after training) of FCM-based and NN systems in function of the game percentile parameter
| min complexity | Complexity | Success rate | |
| Percentile | FCM | Best NN | Best NN |
| 20 | 5.39 | 2.50 | 95.45% |
| 30 | 4.34 | 2.28 | 96.97% |
| 40 | 4.01 | 2.23 | 95.76% |
| 50 | 3.75 | 2.68 | 98.79% |
| 60 | 3.44 | 2.17 | 96.97% |
| 70 | 3.39 | 2.30 | 97.56% |
| 80 | 3.12 | 2.24 | 98.78% |
FCM success rate is always 100%. For FCM, we report minimal complexity among fully successful solutions. For NN, we report complexity and success rate of the system achieving highest success rate.
Fig. 5.Complexity and inefficiency of NN color-naming systems trained with REINFORCE or GS with different s. Pairwise differences evaluated with Bonferroni-corrected Mann–Whitney–Wilcoxon. *P < 0.05; **P < 0.01; ***P < 0.001; ****P < 0.0001. Differences that are not significant are not marked.