1
|
Hosseini EA, Schrimpf M, Zhang Y, Bowman S, Zaslavsky N, Fedorenko E. Artificial Neural Network Language Models Predict Human Brain Responses to Language Even After a Developmentally Realistic Amount of Training. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2024; 5:43-63. [PMID: 38645622 PMCID: PMC11025646 DOI: 10.1162/nol_a_00137] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Accepted: 01/09/2024] [Indexed: 04/23/2024]
Abstract
Artificial neural networks have emerged as computationally plausible models of human language processing. A major criticism of these models is that the amount of training data they receive far exceeds that of humans during language learning. Here, we use two complementary approaches to ask how the models' ability to capture human fMRI responses to sentences is affected by the amount of training data. First, we evaluate GPT-2 models trained on 1 million, 10 million, 100 million, or 1 billion words against an fMRI benchmark. We consider the 100-million-word model to be developmentally plausible in terms of the amount of training data given that this amount is similar to what children are estimated to be exposed to during the first 10 years of life. Second, we test the performance of a GPT-2 model trained on a 9-billion-token dataset to reach state-of-the-art next-word prediction performance on the human benchmark at different stages during training. Across both approaches, we find that (i) the models trained on a developmentally plausible amount of data already achieve near-maximal performance in capturing fMRI responses to sentences. Further, (ii) lower perplexity-a measure of next-word prediction performance-is associated with stronger alignment with human data, suggesting that models that have received enough training to achieve sufficiently high next-word prediction performance also acquire representations of sentences that are predictive of human fMRI responses. In tandem, these findings establish that although some training is necessary for the models' predictive ability, a developmentally realistic amount of training (∼100 million words) may suffice.
Collapse
Affiliation(s)
- Eghbal A. Hosseini
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Martin Schrimpf
- The MIT Quest for Intelligence Initiative, Cambridge, MA, USA
- Swiss Federal Institute of Technology, Lausanne, Switzerland
| | - Yian Zhang
- Computer Science Department, Stanford University, Stanford, CA, USA
| | - Samuel Bowman
- Center for Data Science, New York University, New York, NY, USA
- Department of Linguistics, New York University, New York, NY, USA
- Department of Computer Science, New York University, New York, NY, USA
| | - Noga Zaslavsky
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA
- K. Lisa Yang Integrative Computational Neuroscience (ICoN) Center, Massachusetts Institute of Technology, Cambridge, MA, USA
- Department of Language Science, University of California, Irvine, CA, USA
| | - Evelina Fedorenko
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA
- The MIT Quest for Intelligence Initiative, Cambridge, MA, USA
- Speech and Hearing Bioscience and Technology Program, Harvard University, Boston, MA, USA
| |
Collapse
|
2
|
Keogh A, Kirby S, Culbertson J. Predictability and Variation in Language Are Differentially Affected by Learning and Production. Cogn Sci 2024; 48:e13435. [PMID: 38564253 DOI: 10.1111/cogs.13435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2023] [Revised: 03/01/2024] [Accepted: 03/06/2024] [Indexed: 04/04/2024]
Abstract
General principles of human cognition can help to explain why languages are more likely to have certain characteristics than others: structures that are difficult to process or produce will tend to be lost over time. One aspect of cognition that is implicated in language use is working memory-the component of short-term memory used for temporary storage and manipulation of information. In this study, we consider the relationship between working memory and regularization of linguistic variation. Regularization is a well-documented process whereby languages become less variable (on some dimension) over time. This process has been argued to be driven by the behavior of individual language users, but the specific mechanism is not agreed upon. Here, we use an artificial language learning experiment to investigate whether limitations in working memory during either language learning or language production drive regularization behavior. We find that taxing working memory during production results in the loss of all types of variation, but the process by which random variation becomes more predictable is better explained by learning biases. A computational model offers a potential explanation for the production effect using a simple self-priming mechanism.
Collapse
Affiliation(s)
- Aislinn Keogh
- Centre for Language Evolution, University of Edinburgh
| | - Simon Kirby
- Centre for Language Evolution, University of Edinburgh
| | | |
Collapse
|
3
|
Freeman MR. Linking Language to Action: Enhancing Preschoolers' Communicative Abilities Within Language Stimulation. Lang Speech Hear Serv Sch 2023; 54:1308-1322. [PMID: 37713582 DOI: 10.1044/2023_lshss-22-00196] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/17/2023] Open
Abstract
PURPOSE Children's vocabulary and syntactic skills vary upon school entry in depth and breadth, persistently influencing academic performance, including reading. Enhancing early communicative abilities through multisensory, playful, and conversational experiences is essential and will benefit children's school readiness. This study investigated whether a language-to-action link created during language stimulation, which combines multisensory input, play, and conversation using clay, improves preschoolers' communicative abilities in terms of vocabulary, syntactic, and pragmatic language abilities more than traditional toy-based language stimulation. METHOD Language skills were examined in a pre- to posttest design in which 43 typically developing participants, ages 3-5 years, were randomly assigned to clay-based (n = 24) or traditional play-based (n = 19) language stimulation for 8 weeks. RESULTS Receptive and expressive vocabulary knowledge for items introduced in the language stimulation program, mean length of utterance (MLU), and conversational initiations improved for participants in the clay condition, whereas significant language skill growth was not observed for participants in the traditional play-based stimulation condition with toys. CONCLUSIONS A language-to-action link is created when children engage with open-ended materials, such as clay, as they craft target objects hands on and step by step, affording additional opportunities for language input and output. Results preliminarily suggest that using open-ended materials may enhance children's communicative abilities in receptive and expressive vocabulary, syntax/MLU, and pragmatics (i.e., conversational initiations) more than prefabricated toy objects during language stimulation. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.24093780.
Collapse
Affiliation(s)
- Max R Freeman
- Department of Communication Sciences and Disorders, St. John's University, Jamaica, NY
| |
Collapse
|
4
|
Sourati Z, Venkatesh VPP, Deshpande D, Rawlani H, Ilievski F, Sandlin HÂ, Mermoud A. Robust and explainable identification of logical fallacies in natural language arguments. Knowl Based Syst 2023. [DOI: 10.1016/j.knosys.2023.110418] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/27/2023]
|
5
|
Wang X, Chen Y, Zhu W. A Survey on Curriculum Learning. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:4555-4576. [PMID: 33788677 DOI: 10.1109/tpami.2021.3069908] [Citation(s) in RCA: 22] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Curriculum learning (CL) is a training strategy that trains a machine learning model from easier data to harder data, which imitates the meaningful learning order in human curricula. As an easy-to-use plug-in, the CL strategy has demonstrated its power in improving the generalization capacity and convergence rate of various models in a wide range of scenarios such as computer vision and natural language processing etc. In this survey article, we comprehensively review CL from various aspects including motivations, definitions, theories, and applications. We discuss works on curriculum learning within a general CL framework, elaborating on how to design a manually predefined curriculum or an automatic curriculum. In particular, we summarize existing CL designs based on the general framework of Difficulty Measurer + Training Scheduler and further categorize the methodologies for automatic CL into four groups, i.e., Self-paced Learning, Transfer Teacher, RL Teacher, and Other Automatic CL. We also analyze principles to select different CL designs that may benefit practical applications. Finally, we present our insights on the relationships connecting CL and other machine learning concepts including transfer learning, meta-learning, continual learning and active learning, etc., then point out challenges in CL as well as potential future research directions deserving further investigations.
Collapse
|
6
|
Schrimpf M, Blank IA, Tuckute G, Kauf C, Hosseini EA, Kanwisher N, Tenenbaum JB, Fedorenko E. The neural architecture of language: Integrative modeling converges on predictive processing. Proc Natl Acad Sci U S A 2021; 118:e2105646118. [PMID: 34737231 PMCID: PMC8694052 DOI: 10.1073/pnas.2105646118] [Citation(s) in RCA: 156] [Impact Index Per Article: 39.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/03/2021] [Indexed: 01/30/2023] Open
Abstract
The neuroscience of perception has recently been revolutionized with an integrative modeling approach in which computation, brain function, and behavior are linked across many datasets and many computational models. By revealing trends across models, this approach yields novel insights into cognitive and neural mechanisms in the target domain. We here present a systematic study taking this approach to higher-level cognition: human language processing, our species' signature cognitive skill. We find that the most powerful "transformer" models predict nearly 100% of explainable variance in neural responses to sentences and generalize across different datasets and imaging modalities (functional MRI and electrocorticography). Models' neural fits ("brain score") and fits to behavioral responses are both strongly correlated with model accuracy on the next-word prediction task (but not other language tasks). Model architecture appears to substantially contribute to neural fit. These results provide computationally explicit evidence that predictive processing fundamentally shapes the language comprehension mechanisms in the human brain.
Collapse
Affiliation(s)
- Martin Schrimpf
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139;
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
- Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Idan Asher Blank
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139
- Department of Psychology, University of California, Los Angeles, CA 90095
| | - Greta Tuckute
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Carina Kauf
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Eghbal A Hosseini
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Nancy Kanwisher
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139;
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
- Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Joshua B Tenenbaum
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139
- Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, MA 02139
| | - Evelina Fedorenko
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139;
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139
| |
Collapse
|
7
|
|
8
|
Huebner PA, Willits JA. Using lexical context to discover the noun category: Younger children have it easier. PSYCHOLOGY OF LEARNING AND MOTIVATION 2021. [DOI: 10.1016/bs.plm.2021.08.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
9
|
Machine Learning: From Expert Systems to Deep Learning. Cogn Sci 2019. [DOI: 10.1017/9781108339216.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
10
|
The Prehistory of Cognitive Science. Cogn Sci 2019. [DOI: 10.1017/9781108339216.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
11
|
Preface. Cogn Sci 2019. [DOI: 10.1017/9781108339216.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
12
|
Bibliography. Cogn Sci 2019. [DOI: 10.1017/9781108339216.021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
13
|
Bayesianism in Cognitive Science. Cogn Sci 2019. [DOI: 10.1017/9781108339216.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
14
|
Object Perception and Folk Physics. Cogn Sci 2019. [DOI: 10.1017/9781108339216.013] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
|
15
|
Glossary. Cogn Sci 2019. [DOI: 10.1017/9781108339216.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
|
16
|
Strategies for Brain Mapping. Cogn Sci 2019. [DOI: 10.1017/9781108339216.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
17
|
Modules and Architectures. Cogn Sci 2019. [DOI: 10.1017/9781108339216.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
18
|
Introduction. Cogn Sci 2019. [DOI: 10.1017/9781108339216.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
19
|
The Discipline Matures: Three Milestones. Cogn Sci 2019. [DOI: 10.1017/9781108339216.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
20
|
Models of Language Learning. Cogn Sci 2019. [DOI: 10.1017/9781108339216.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
21
|
Applying Dynamical Systems Theory to Model the Mind. Cogn Sci 2019. [DOI: 10.1017/9781108339216.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
22
|
Exploring Mindreading. Cogn Sci 2019. [DOI: 10.1017/9781108339216.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
23
|
Robotics: From GOFAI to Situated Cognition and Behavior-Based Robotics. Cogn Sci 2019. [DOI: 10.1017/9781108339216.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
24
|
The Cognitive Science of Consciousness. Cogn Sci 2019. [DOI: 10.1017/9781108339216.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
25
|
The Turn to the Brain. Cogn Sci 2019. [DOI: 10.1017/9781108339216.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
|
26
|
Index for Cognitive Science (3rd edition). Cogn Sci 2019. [DOI: 10.1017/9781108339216.022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
27
|
Mindreading: Advanced Topics. Cogn Sci 2019. [DOI: 10.1017/9781108339216.016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
28
|
Physical Symbol Systems and the Language of Thought. Cogn Sci 2019. [DOI: 10.1017/9781108339216.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
29
|
Looking Ahead: Challenges and Opportunities. Cogn Sci 2019. [DOI: 10.1017/9781108339216.019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
30
|
Neural Networks and Distributed Information Processing. Cogn Sci 2019. [DOI: 10.1017/9781108339216.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
31
|
Jost E, Brill-Schuetz K, Morgan-Short K, Christiansen MH. Input Complexity Affects Long-Term Retention of Statistically Learned Regularities in an Artificial Language Learning Task. Front Hum Neurosci 2019; 13:358. [PMID: 31680911 PMCID: PMC6803473 DOI: 10.3389/fnhum.2019.00358] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2019] [Accepted: 09/26/2019] [Indexed: 12/03/2022] Open
Abstract
Statistical learning (SL) involving sensitivity to distributional regularities in the environment has been suggested to be an important factor in many aspects of cognition, including language. However, the degree to which statistically-learned information is retained over time is not well understood. To establish whether or not learners are able to preserve such regularities over time, we examined performance on an artificial second language learning task both immediately after training and also at a follow-up session 2 weeks later. Participants were exposed to an artificial language (Brocanto2), half of them receiving simplified training items in which only 20% of sequences contained complex structures, whereas the other half were exposed to a training set in which 80% of the items were composed of complex sequences. Overall, participants showed signs of learning at the first session and retention at the second, but the degree of learning was affected by the nature of the training they received. Participants exposed to the simplified input outperformed those in the more complex training condition. A GLMM was used to model the relationship between stimulus properties and participants' endorsement strategies across both sessions. The results indicate that participants in the complex training condition relied more on an item's chunk strength than those in the simple training condition. Taken together, this set of findings shows that statistically learned regularities are retained over the course of 2 weeks. The results also demonstrate that training on input featuring simple items leads to improved learning and retention of grammatical regularities.
Collapse
Affiliation(s)
- Ethan Jost
- Department of Psychology, Cornell University, Ithaca, NY, United States
| | | | - Kara Morgan-Short
- Department of Psychology, University of Illinois at Chicago, Chicago, IL, United States
| | | |
Collapse
|
32
|
Ma T, Komarova NL. Object-Label-Order Effect When Learning From an Inconsistent Source. Cogn Sci 2019; 43:e12737. [PMID: 31446665 DOI: 10.1111/cogs.12737] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2018] [Revised: 04/12/2019] [Accepted: 04/15/2019] [Indexed: 11/30/2022]
Abstract
Learning in natural environments is often characterized by a degree of inconsistency from an input. These inconsistencies occur, for example, when learning from more than one source, or when the presence of environmental noise distorts incoming information; as a result, the task faced by the learner becomes ambiguous. In this study, we investigate how learners handle such situations. We focus on the setting where a learner receives and processes a sequence of utterances to master associations between objects and their labels, where the source is inconsistent by design: It uses both "correct" and "incorrect" object-label pairings. We hypothesize that depending on the order of presentation, the result of the learning may be different. To this end, we consider two types of symbolic learning procedures: the Object-Label (OL) and the Label-Object (LO) process. In the OL process, the learner is first exposed to the object, and then the label. In the LO process, this order is reversed. We perform experiments with human subjects, and also construct a computational model that is based on a nonlinear stochastic reinforcement learning algorithm. It is observed experimentally that OL learners are generally better at processing inconsistent input compared to LO learners. We show that the patterns observed in the learning experiments can be reproduced in the simulations if the model includes (a) an ability to regularize the input (and also to do the opposite, i.e., undermatch) and (b) an ability to take account of implicit negative evidence (i.e., interactions among different objects/labels). The model suggests that while both types of learners utilize implicit negative evidence in a similar way, there is a difference in regularization patterns: OL learners regularize the input, whereas LO learners undermatch. As a result, OL learners are able to form a more consistent system of image-utterance associations, despite the ambiguous learning task.
Collapse
Affiliation(s)
- Timmy Ma
- Department of Mathematics, Dartmouth College
| | | |
Collapse
|
33
|
Abstract
Language comprehension requires successfully navigating linguistic variability. One hypothesis for how listeners manage variability is that they rapidly update their expectations of likely linguistic events in new contexts. This process, called adaptation, allows listeners to better predict the upcoming linguistic input. In previous work, Fine, Jaeger, Farmer, and Qian (PLoS ONE, 8, e77661, 2013) found evidence for syntactic adaptation. Subjects repeatedly encountered sentences in which a verb was temporarily ambiguous between main verb (MV) and reduced relative clause (RC) interpretations. They found that subjects who had higher levels of exposure to the unexpected RC interpretation of the sentences had an easier time reading the RC sentences but a more difficult time reading the MV sentences. They concluded that syntactic adaptation occurs rapidly in unexpected structures and also results in difficulty with processing the previously expected alternative structures. This article presents two experiments. Experiment 1 was designed as a follow-up to Fine et al.'s study and failed to find evidence of adaptation. A power analysis of Fine et al.'s raw data revealed that a similar study would need double the items and four times the subjects to reach 95% power. In Experiment 2 we designed a close replication of Fine et al.'s experiment using these sample size guidelines. No evidence of rapid syntactic adaptation was found in this experiment. The failure to find evidence of adaptation in both experiments calls into question the robustness of the effect.
Collapse
|
34
|
Poletiek FH, Conway CM, Ellefson MR, Lai J, Bocanegra BR, Christiansen MH. Under What Conditions Can Recursion Be Learned? Effects of Starting Small in Artificial Grammar Learning of Center-Embedded Structure. Cogn Sci 2018; 42:2855-2889. [PMID: 30264489 PMCID: PMC6585836 DOI: 10.1111/cogs.12685] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2016] [Revised: 07/22/2018] [Accepted: 07/24/2018] [Indexed: 11/30/2022]
Abstract
It has been suggested that external and/or internal limitations paradoxically may lead to superior learning, that is, the concepts of starting small and less is more (Elman, 1993; Newport, 1990). In this paper, we explore the type of incremental ordering during training that might help learning, and what mechanism explains this facilitation. We report four artificial grammar learning experiments with human participants. In Experiments 1a and 1b we found a beneficial effect of starting small using two types of simple recursive grammars: right‐branching and center‐embedding, with recursive embedded clauses in fixed positions and fixed length. This effect was replicated in Experiment 2 (N = 100). In Experiment 3 and 4, we used a more complex center‐embedded grammar with recursive loops in variable positions, producing strings of variable length. When participants were presented an incremental ordering of training stimuli, as in natural language, they were better able to generalize their knowledge of simple units to more complex units when the training input “grew” according to structural complexity, compared to when it “grew” according to string length. Overall, the results suggest that starting small confers an advantage for learning complex center‐embedded structures when the input is organized according to structural complexity.
Collapse
Affiliation(s)
- Fenna H Poletiek
- Institute of Psychology, Leiden University.,Max Planck Institute for Psycholinguistics, Nijmegen
| | | | | | - Jun Lai
- Institute of Psychology, Leiden University
| | - Bruno R Bocanegra
- Erasmus School of Social and Behavioral Sciences, Erasmus University
| | | |
Collapse
|
35
|
Kempe V, Brooks PJ. Linking Adult Second Language Learning and Diachronic Change: A Cautionary Note. Front Psychol 2018; 9:480. [PMID: 29674993 PMCID: PMC5895708 DOI: 10.3389/fpsyg.2018.00480] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2017] [Accepted: 03/21/2018] [Indexed: 11/13/2022] Open
Affiliation(s)
- Vera Kempe
- School of Social and Health Sciences, Abertay University, Dundee, United Kingdom
| | - Patricia J Brooks
- College of Staten Island and The Graduate Center, City University of New York, Brooklyn, NY, United States
| |
Collapse
|
36
|
Fitz H, Chang F. Meaningful questions: The acquisition of auxiliary inversion in a connectionist model of sentence production. Cognition 2017; 166:225-250. [DOI: 10.1016/j.cognition.2017.05.008] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2014] [Revised: 10/31/2016] [Accepted: 05/10/2017] [Indexed: 11/16/2022]
|
37
|
Gaspers J, Cimiano P, Rohlfing K, Wrede B. Constructing a Language From Scratch: Combining Bottom–Up and Top–Down Learning Processes in a Computational Model of Language Acquisition. IEEE Trans Cogn Dev Syst 2017. [DOI: 10.1109/tcds.2016.2614958] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
38
|
Wong PCM, Vuong LC, Liu K. Personalized learning: From neurogenetics of behaviors to designing optimal language training. Neuropsychologia 2016; 98:192-200. [PMID: 27720749 DOI: 10.1016/j.neuropsychologia.2016.10.002] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2016] [Revised: 08/08/2016] [Accepted: 10/04/2016] [Indexed: 01/11/2023]
Abstract
Variability in drug responsivity has prompted the development of Personalized Medicine, which has shown great promise in utilizing genotypic information to develop safer and more effective drug regimens for patients. Similarly, individual variability in learning outcomes has puzzled researchers who seek to create optimal learning environments for students. "Personalized Learning" seeks to identify genetic, neural and behavioral predictors of individual differences in learning and aims to use predictors to help create optimal teaching paradigms. Evidence for Personalized Learning can be observed by connecting research in pharmacogenomics, cognitive genetics and behavioral experiments across domains of learning, which provides a framework for conducting empirical studies from the laboratory to the classroom and holds promise for addressing learning effectiveness in the individual learners. Evidence can also be seen in the subdomain of speech learning, thus providing initial support for the applicability of Personalized Learning to language.
Collapse
Affiliation(s)
- Patrick C M Wong
- Dept of Linguistics & Modern Languages and Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, People's Republic of China.
| | - Loan C Vuong
- Dept of Linguistics & Modern Languages and Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, People's Republic of China
| | - Kevin Liu
- Feinberg School of Medicine, Northwestern University, USA
| |
Collapse
|
39
|
Hsu A, Griffiths TL. Sampling Assumptions Affect Use of Indirect Negative Evidence in Language Learning. PLoS One 2016; 11:e0156597. [PMID: 27310576 PMCID: PMC4911062 DOI: 10.1371/journal.pone.0156597] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2015] [Accepted: 05/17/2016] [Indexed: 11/18/2022] Open
Abstract
A classic debate in cognitive science revolves around understanding how children learn complex linguistic patterns, such as restrictions on verb alternations and contractions, without negative evidence. Recently, probabilistic models of language learning have been applied to this problem, framing it as a statistical inference from a random sample of sentences. These probabilistic models predict that learners should be sensitive to the way in which sentences are sampled. There are two main types of sampling assumptions that can operate in language learning: strong and weak sampling. Strong sampling, as assumed by probabilistic models, assumes the learning input is drawn from a distribution of grammatical samples from the underlying language and aims to learn this distribution. Thus, under strong sampling, the absence of a sentence construction from the input provides evidence that it has low or zero probability of grammaticality. Weak sampling does not make assumptions about the distribution from which the input is drawn, and thus the absence of a construction from the input as not used as evidence of its ungrammaticality. We demonstrate in a series of artificial language learning experiments that adults can produce behavior consistent with both sets of sampling assumptions, depending on how the learning problem is presented. These results suggest that people use information about the way in which linguistic input is sampled to guide their learning.
Collapse
Affiliation(s)
- Anne Hsu
- School of Electronic Engineering, Queen Mary, University of London, London, United Kingdom
| | - Thomas L. Griffiths
- Department of Psychology, University of California at Berkeley, Berkeley, California, United States of America
| |
Collapse
|
40
|
Cho PW, Szkudlarek E, Tabor W. Discovery of a Recursive Principle: An Artificial Grammar Investigation of Human Learning of a Counting Recursion Language. Front Psychol 2016; 7:867. [PMID: 27375543 PMCID: PMC4897795 DOI: 10.3389/fpsyg.2016.00867] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2016] [Accepted: 05/25/2016] [Indexed: 11/13/2022] Open
Abstract
Learning is typically understood as a process in which the behavior of an organism is progressively shaped until it closely approximates a target form. It is easy to comprehend how a motor skill or a vocabulary can be progressively learned-in each case, one can conceptualize a series of intermediate steps which lead to the formation of a proficient behavior. With grammar, it is more difficult to think in these terms. For example, center embedding recursive structures seem to involve a complex interplay between multiple symbolic rules which have to be in place simultaneously for the system to work at all, so it is not obvious how the mechanism could gradually come into being. Here, we offer empirical evidence from a new artificial language (or "artificial grammar") learning paradigm, Locus Prediction, that, despite the conceptual conundrum, recursion acquisition occurs gradually, at least for a simple formal language. In particular, we focus on a variant of the simplest recursive language, a (n) b (n) , and find evidence that (i) participants trained on two levels of structure (essentially ab and aabb) generalize to the next higher level (aaabbb) more readily than participants trained on one level of structure (ab) combined with a filler sentence; nevertheless, they do not generalize immediately; (ii) participants trained up to three levels (ab, aabb, aaabbb) generalize more readily to four levels than participants trained on two levels generalize to three; (iii) when we present the levels in succession, starting with the lower levels and including more and more of the higher levels, participants show evidence of transitioning between the levels gradually, exhibiting intermediate patterns of behavior on which they were not trained; (iv) the intermediate patterns of behavior are associated with perturbations of an attractor in the sense of dynamical systems theory. We argue that all of these behaviors indicate a theory of mental representation in which recursive systems lie on a continuum of grammar systems which are organized so that grammars that produce similar behaviors are near one another, and that people learning a recursive system are navigating progressively through the space of these grammars.
Collapse
Affiliation(s)
- Pyeong Whan Cho
- Department of Psychology, University of ConnecticutStorrs, CT, USA
- Haskins LaboratoriesNew Haven, CT, USA
| | - Emily Szkudlarek
- Department of Psychology, University of ConnecticutStorrs, CT, USA
| | - Whitney Tabor
- Department of Psychology, University of ConnecticutStorrs, CT, USA
- Haskins LaboratoriesNew Haven, CT, USA
| |
Collapse
|
41
|
Thiessen ED, Girard S, Erickson LC. Statistical learning and the critical period: how a continuous learning mechanism can give rise to discontinuous learning. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2016; 7:276-88. [DOI: 10.1002/wcs.1394] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2015] [Revised: 03/31/2016] [Accepted: 04/06/2016] [Indexed: 11/08/2022]
Affiliation(s)
- Erik D. Thiessen
- Department of Psychology; Carnegie Mellon University; Pittsburgh PA USA
| | - Sandrine Girard
- Department of Psychology; Carnegie Mellon University; Pittsburgh PA USA
| | - Lucy C. Erickson
- Department of Psychology; Carnegie Mellon University; Pittsburgh PA USA
| |
Collapse
|
42
|
Regularization of languages by adults and children: A mathematical framework. Cogn Psychol 2015; 84:1-30. [PMID: 26580218 DOI: 10.1016/j.cogpsych.2015.10.001] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2014] [Revised: 10/09/2015] [Accepted: 10/20/2015] [Indexed: 11/22/2022]
Abstract
The fascinating ability of humans to modify the linguistic input and "create" a language has been widely discussed. In the work of Newport and colleagues, it has been demonstrated that both children and adults have some ability to process inconsistent linguistic input and "improve" it by making it more consistent. In Hudson Kam and Newport (2009), artificial miniature language acquisition from an inconsistent source was studied. It was shown that (i) children are better at language regularization than adults and that (ii) adults can also regularize, depending on the structure of the input. In this paper we create a learning algorithm of the reinforcement-learning type, which exhibits patterns reported in Hudson Kam and Newport (2009) and suggests a way to explain them. It turns out that in order to capture the differences between children's and adults' learning patterns, we need to introduce a certain asymmetry in the learning algorithm. Namely, we have to assume that the reaction of the learners differs depending on whether or not the source's input coincides with the learner's internal hypothesis. We interpret this result in the context of a different reaction of children and adults to implicit, expectation-based evidence, positive or negative. We propose that a possible mechanism that contributes to the children's ability to regularize an inconsistent input is related to their heightened sensitivity to positive evidence rather than the (implicit) negative evidence. In our model, regularization comes naturally as a consequence of a stronger reaction of the children to evidence supporting their preferred hypothesis. In adults, their ability to adequately process implicit negative evidence prevents them from regularizing the inconsistent input, resulting in a weaker degree of regularization.
Collapse
|
43
|
Spiliopoulos L. Transfer of conflict and cooperation from experienced games to new games: a connectionist model of learning. Front Neurosci 2015; 9:102. [PMID: 25873855 PMCID: PMC4379898 DOI: 10.3389/fnins.2015.00102] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2014] [Accepted: 03/11/2015] [Indexed: 11/26/2022] Open
Abstract
The question of whether, and if so how, learning can be transfered from previously experienced games to novel games has recently attracted the attention of the experimental game theory literature. Existing research presumes that learning operates over actions, beliefs or decision rules. This study instead uses a connectionist approach that learns a direct mapping from game payoffs to a probability distribution over own actions. Learning is operationalized as a backpropagation rule that adjusts the weights of feedforward neural networks in the direction of increasing the probability of an agent playing a myopic best response to the last game played. One advantage of this approach is that it expands the scope of the model to any possible n × n normal-form game allowing for a comprehensive model of transfer of learning. Agents are exposed to games drawn from one of seven classes of games with significantly different strategic characteristics and then forced to play games from previously unseen classes. I find significant transfer of learning, i.e., behavior that is path-dependent, or conditional on the previously seen games. Cooperation is more pronounced in new games when agents are previously exposed to games where the incentive to cooperate is stronger than the incentive to compete, i.e., when individual incentives are aligned. Prior exposure to Prisoner's dilemma, zero-sum and discoordination games led to a significant decrease in realized payoffs for all the game classes under investigation. A distinction is made between superficial and deep transfer of learning both—the former is driven by superficial payoff similarities between games, the latter by differences in the incentive structures or strategic implications of the games. I examine whether agents learn to play the Nash equilibria of games, how they select amongst multiple equilibria, and whether they transfer Nash equilibrium behavior to unseen games. Sufficient exposure to a strategically heterogeneous set of games is found to be a necessary condition for deep learning (and transfer) across game classes. Paradoxically, superficial transfer of learning is shown to lead to better outcomes than deep transfer for a wide range of game classes. The simulation results corroborate important experimental findings with human subjects, and make several novel predictions that can be tested experimentally.
Collapse
Affiliation(s)
- Leonidas Spiliopoulos
- Center for Adaptive Rationality, Max Planck Institute for Human Development Berlin, Germany
| |
Collapse
|
44
|
Fedorenko E. The role of domain-general cognitive control in language comprehension. Front Psychol 2014; 5:335. [PMID: 24803909 PMCID: PMC4009428 DOI: 10.3389/fpsyg.2014.00335] [Citation(s) in RCA: 135] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2012] [Accepted: 03/31/2014] [Indexed: 01/15/2023] Open
Abstract
What role does domain-general cognitive control play in understanding linguistic input? Although much evidence has suggested that domain-general cognitive control and working memory resources are sometimes recruited during language comprehension, many aspects of this relationship remain elusive. For example, how frequently do cognitive control mechanisms get engaged when we understand language? And is this engagement necessary for successful comprehension? I here (a) review recent brain imaging evidence for the neural separability of the brain regions that support high-level linguistic processing vs. those that support domain-general cognitive control abilities; (b) define the space of possibilities for the relationship between these sets of brain regions; and (c) review the available evidence that constrains these possibilities to some extent. I argue that we should stop asking whether domain-general cognitive control mechanisms play a role in language comprehension, and instead focus on characterizing the division of labor between the cognitive control brain regions and the more functionally specialized language regions.
Collapse
Affiliation(s)
- Evelina Fedorenko
- Psychiatry Department, Massachusetts General HospitalCharlestown, MA, USA
| |
Collapse
|
45
|
Chrysikou EG, Weber MJ, Thompson-Schill SL. A matched filter hypothesis for cognitive control. Neuropsychologia 2013; 62:341-355. [PMID: 24200920 DOI: 10.1016/j.neuropsychologia.2013.10.021] [Citation(s) in RCA: 97] [Impact Index Per Article: 8.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2013] [Revised: 10/21/2013] [Accepted: 10/28/2013] [Indexed: 11/30/2022]
Abstract
The prefrontal cortex exerts top-down influences on several aspects of higher-order cognition by functioning as a filtering mechanism that biases bottom-up sensory information toward a response that is optimal in context. However, research also indicates that not all aspects of complex cognition benefit from prefrontal regulation. Here we review and synthesize this research with an emphasis on the domains of learning and creative cognition, and outline how the appropriate level of cognitive control in a given situation can vary depending on the organism's goals and the characteristics of the given task. We offer a matched filter hypothesis for cognitive control, which proposes that the optimal level of cognitive control is task-dependent, with high levels of cognitive control best suited to tasks that are explicit, rule-based, verbal or abstract, and can be accomplished given the capacity limits of working memory and with low levels of cognitive control best suited to tasks that are implicit, reward-based, non-verbal or intuitive, and which can be accomplished irrespective of working memory limitations. Our approach promotes a view of cognitive control as a tool adapted to a subset of common challenges, rather than an all-purpose optimization system suited to every problem the organism might encounter.
Collapse
Affiliation(s)
| | - Matthew J Weber
- Department of Psychology, Center for Cognitive Neuroscience, University of Pennsylvania
| | | |
Collapse
|
46
|
Hsu AS, Chater N, Vitányi P. Language learning from positive evidence, reconsidered: a simplicity-based approach. Top Cogn Sci 2013; 5:35-55. [PMID: 23335573 DOI: 10.1111/tops.12005] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2011] [Revised: 11/25/2012] [Accepted: 11/26/2012] [Indexed: 11/27/2022]
Abstract
Children learn their native language by exposure to their linguistic and communicative environment, but apparently without requiring that their mistakes be corrected. Such learning from "positive evidence" has been viewed as raising "logical" problems for language acquisition. In particular, without correction, how is the child to recover from conjecturing an over-general grammar, which will be consistent with any sentence that the child hears? There have been many proposals concerning how this "logical problem" can be dissolved. In this study, we review recent formal results showing that the learner has sufficient data to learn successfully from positive evidence, if it favors the simplest encoding of the linguistic input. Results include the learnability of linguistic prediction, grammaticality judgments, language production, and form-meaning mappings. The simplicity approach can also be "scaled down" to analyze the learnability of specific linguistic constructions, and it is amenable to empirical testing as a framework for describing human language acquisition.
Collapse
Affiliation(s)
- Anne S Hsu
- Department of Cognitive, Perceptual and Brain Sciences, University College London, UK. anne.hsu@ ucl.ac.uk
| | | | | |
Collapse
|
47
|
Lai J, Poletiek FH. How “small” is “starting small” for learning hierarchical centre-embedded structures? JOURNAL OF COGNITIVE PSYCHOLOGY 2013. [DOI: 10.1080/20445911.2013.779247] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
48
|
Poletiek FH, Lai J. How semantic biases in simple adjacencies affect learning a complex structure with non-adjacencies in AGL: a statistical account. Philos Trans R Soc Lond B Biol Sci 2012; 367:2046-54. [PMID: 22688639 DOI: 10.1098/rstb.2012.0100] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
A major theoretical debate in language acquisition research regards the learnability of hierarchical structures. The artificial grammar learning methodology is increasingly influential in approaching this question. Studies using an artificial centre-embedded A(n)B(n) grammar without semantics draw conflicting conclusions. This study investigates the facilitating effect of distributional biases in simple AB adjacencies in the input sample--caused in natural languages, among others, by semantic biases-on learning a centre-embedded structure. A mathematical simulation of the linguistic input and the learning, comparing various distributional biases in AB pairs, suggests that strong distributional biases might help us to grasp the complex A(n)B(n) hierarchical structure in a later stage. This theoretical investigation might contribute to our understanding of how distributional features of the input--including those caused by semantic variation--help learning complex structures in natural languages.
Collapse
Affiliation(s)
- Fenna H Poletiek
- Cognitive Psychology Department, Leiden University, Pieter de la Court building, PO Box 9555, 2300 Leiden, The Netherlands.
| | | |
Collapse
|
49
|
Abstract
Learners exposed to an artificial language recognize its abstract structural regularities when instantiated in a novel vocabulary (e.g., Gómez, Gerken, & Schvaneveldt, 2000; Tunney & Altmann, 2001). We asked whether such sensitivity accelerates subsequent learning, and enables acquisition of more complex structure. In Experiment 1, pre-exposure to a category-induction language of the form aX bY sped subsequent learning when the language is instantiated in a different vocabulary. In Experiment 2, while naíve learners did not acquire an acX bcY language, in which aX and bY co-occurrence regularities were separated by a c-element, prior experience with an aX bY language provided some benefit. In Experiment 3 we replicated this finding with a 24-hour delay between learning phases, and controlled for prior experience with the aX bY language's prosodic and phonological characteristics. These findings suggest that learners, and the structure they can acquire, change as a function of experience.
Collapse
Affiliation(s)
- Jill Lany
- Department of Psychology, The University of Arizona, Tucson
| | | | | |
Collapse
|
50
|
Abstract
PURPOSE This review introduces emergentism, which is a leading theory of language development that states that language ability is the product of interactions between the child's language environment and his or her learning capabilities. The review suggests ways in which emergentism provides a theoretical rationale for interventions that are designed to address developmental language delays in young children. METHOD A review of selected literature on emergentist theory and research is presented, with a focus on the acquisition of early morphology and syntax. A significant method for developing and testing emergentist theory, connectionist modeling, is described. Key themes from both connectionist and behavioral studies are summarized and applied with specific examples to language intervention techniques. A case study is presented to integrate elements of emergentism with language intervention. CONCLUSIONS Evaluating the theoretical foundation for language interventions is an important step in evidence-based practice. This article introduces three themes in the emergentist literature that have implications for language intervention: (a) sufficiency of language input, (b) active engagement of the child with the input, and (c) factors that increase the odds for correctly mapping language form to meaning. Evidence supporting the importance of these factors in effective language intervention is presented, along with limitations in that evidence.
Collapse
Affiliation(s)
- Gerard H Poll
- Pennsylvania State University, University Park, PA, USA.
| |
Collapse
|