Literature DB >> 32493748

Emergent linguistic structure in artificial neural networks trained by self-supervision.

Christopher D Manning1, Kevin Clark2, John Hewitt2, Urvashi Khandelwal2, Omer Levy3.   

Abstract

This paper explores the knowledge of linguistic structure learned by large artificial neural networks, trained via self-supervision, whereby the model simply tries to predict a masked word in a given context. Human language communication is via sequences of words, but language understanding requires constructing rich hierarchical structures that are never observed explicitly. The mechanisms for this have been a prime mystery of human language acquisition, while engineering work has mainly proceeded by supervised learning on treebanks of sentences hand labeled for this latent structure. However, we demonstrate that modern deep contextual language models learn major aspects of this structure, without any explicit supervision. We develop methods for identifying linguistic hierarchical structure emergent in artificial neural networks and demonstrate that components in these models focus on syntactic grammatical relationships and anaphoric coreference. Indeed, we show that a linear transformation of learned embeddings in these models captures parse tree distances to a surprising degree, allowing approximate reconstruction of the sentence tree structures normally assumed by linguists. These results help explain why these models have brought such large improvements across many language-understanding tasks.

Entities:  

Keywords:  artificial neural netwok; learning; self-supervision; syntax

Year:  2020        PMID: 32493748      PMCID: PMC7720155          DOI: 10.1073/pnas.1907367117

Source DB:  PubMed          Journal:  Proc Natl Acad Sci U S A        ISSN: 0027-8424            Impact factor:   11.205


  5 in total

Review 1.  Perception viewed as an inverse problem.

Authors:  Z Pizlo
Journal:  Vision Res       Date:  2001-11       Impact factor: 1.886

2.  Broken agreement.

Authors:  K Bock; C A Miller
Journal:  Cogn Psychol       Date:  1991-01       Impact factor: 3.468

3.  Rethinking language: how probabilities shape the words we use.

Authors:  Thomas L Griffiths
Journal:  Proc Natl Acad Sci U S A       Date:  2011-02-23       Impact factor: 11.205

Review 4.  Poverty of the stimulus revisited.

Authors:  Robert C Berwick; Paul Pietroski; Beracah Yankama; Noam Chomsky
Journal:  Cogn Sci       Date:  2011-08-08

Review 5.  Early language acquisition: cracking the speech code.

Authors:  Patricia K Kuhl
Journal:  Nat Rev Neurosci       Date:  2004-11       Impact factor: 34.870

  5 in total
  17 in total

1.  The science of deep learning.

Authors:  Richard Baraniuk; David Donoho; Matan Gavish
Journal:  Proc Natl Acad Sci U S A       Date:  2020-11-23       Impact factor: 11.205

2.  A hierarchy of linguistic predictions during natural language comprehension.

Authors:  Micha Heilbron; Kristijan Armeni; Jan-Mathijs Schoffelen; Peter Hagoort; Floris P de Lange
Journal:  Proc Natl Acad Sci U S A       Date:  2022-08-03       Impact factor: 12.779

3.  A weighted constraint satisfaction approach to human goal-directed decision making.

Authors:  Yuxuan Li; James L McClelland
Journal:  PLoS Comput Biol       Date:  2022-06-16       Impact factor: 4.779

4.  Construction of English Translation Model Based on Neural Network Fuzzy Semantic Optimal Control.

Authors:  Bingjie Zhang; Yiming Liu
Journal:  Comput Intell Neurosci       Date:  2022-05-02

5.  Compositional Processing Emerges in Neural Networks Solving Math Problems.

Authors:  Jacob Russin; Roland Fernandez; Hamid Palangi; Eric Rosen; Nebojsa Jojic; Paul Smolensky; Jianfeng Gao
Journal:  Cogsci       Date:  2021-07

6.  Placing language in an integrated understanding system: Next steps toward human-level performance in neural language models.

Authors:  James L McClelland; Felix Hill; Maja Rudolph; Jason Baldridge; Hinrich Schütze
Journal:  Proc Natl Acad Sci U S A       Date:  2020-09-28       Impact factor: 11.205

7.  Emerging Grounded Shared Vocabularies Between Human and Machine, Inspired by Human Language Evolution.

Authors:  Tom Kouwenhoven; Tessa Verhoef; Roy de Kleijn; Stephan Raaijmakers
Journal:  Front Artif Intell       Date:  2022-04-26

8.  Communicating artificial neural networks develop efficient color-naming systems.

Authors:  Rahma Chaabouni; Eugene Kharitonov; Emmanuel Dupoux; Marco Baroni
Journal:  Proc Natl Acad Sci U S A       Date:  2021-03-23       Impact factor: 12.779

9.  Developing and testing an automated qualitative assistant (AQUA) to support qualitative analysis.

Authors:  Robert P Lennon; Robbie Fraleigh; Lauren J Van Scoy; Aparna Keshaviah; Xindi C Hu; Bethany L Snyder; Erin L Miller; William A Calo; Aleksandra E Zgierska; Christopher Griffin
Journal:  Fam Med Community Health       Date:  2021-11

10.  Measuring context dependency in birdsong using artificial neural networks.

Authors:  Takashi Morita; Hiroki Koda; Kazuo Okanoya; Ryosuke O Tachibana
Journal:  PLoS Comput Biol       Date:  2021-12-28       Impact factor: 4.475

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.