1
|
Mousley VL, MacSweeney M, Mercure E. Revisiting perceptual sensitivity to non-native speech in a diverse sample of bilinguals. Infant Behav Dev 2024; 76:101959. [PMID: 38781790 DOI: 10.1016/j.infbeh.2024.101959] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2024] [Revised: 04/18/2024] [Accepted: 05/12/2024] [Indexed: 05/25/2024]
Abstract
Werker and Tees (1984) prompted decades of research attempting to detail the paths infants take towards specialisation for the sounds of their native language(s). Most of this research has examined the trajectories of monolingual children. However, it has also been proposed that bilinguals, who are exposed to greater phonetic variability than monolinguals and must learn the rules of two languages, may remain perceptually open to non-native language sounds later into life than monolinguals. Using a visual habituation paradigm, the current study tests this question by comparing 15- to 18-month-old monolingual and bilingual children's developmental trajectories for non-native phonetic consonant contrast discrimination. A novel approach to the integration of stimulus presentation software with eye-tracking software was validated for objective measurement of infant looking time. The results did not support the hypothesis of a protracted period of sensitivity to non-native phonetic contrasts in bilingual compared to monolingual infants. Implications for diversification of perceptual narrowing research and implementation of increasingly sensitive measures are discussed.
Collapse
Affiliation(s)
- Victoria L Mousley
- Centre for Brain and Cognitive Development, School of Psychological Sciences, Birkbeck, University of London, London WC1E 7HX, United Kingdom.
| | - Mairéad MacSweeney
- Deafness, Cognition and Language Research Centre, University College London, London WC1H 0PD, United Kingdom; Institute of Cognitive Neuroscience, University College London, London WC1N 3AZ, United Kingdom.
| | - Evelyne Mercure
- Centre for Brain and Cognitive Development, School of Psychological Sciences, Birkbeck, University of London, London WC1E 7HX, United Kingdom; Department of Psychology, Goldsmiths, University of London, London SE14 6NW, United Kingdom.
| |
Collapse
|
2
|
Yang Y, Chen S, Chen F, Ma J. Development of perceptual similarity and discriminability: the perception of Russian phonemes by Chinese learners. PHONETICA 2023; 0:phon-2022-0023. [PMID: 37261920 DOI: 10.1515/phon-2022-0023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Accepted: 05/02/2023] [Indexed: 06/03/2023]
Abstract
This study explored the perceptual assimilation and discrimination of Russian phonemes by three groups of Chinese listeners with differing Russian learning experience. A perceptual assimilation task (PAT) and a perceptual discrimination test (PDT) were conducted to investigate if/how L1-L2 perceptual similarity would vary as a function of increased learning experience, and the development of assimilation-discrimination relations. The PAT was analyzed via assimilation rates, dispersion K' values, goodness ratings and assimilation patterns. Results revealed an intriguing phenomenon that the perceived Mandarin-Russian similarity first increased from naïve listeners to intermediate learners and then decreased slightly in relatively advanced learners. This suggests that L1-L2 perceptual similarity is subject to learning experience and could follow a potential "rise and fall" developmental pattern. The PDT results were mostly in line with the assimilation-discrimination correspondence with more experience bringing out better discriminability in general. Yet the overall sensitivity d' values from the Chinese groups were relatively low, implying acoustic/articulatory effects on L2 discriminability aside from perceptual assimilation. The results were discussed under the frameworks of L2 Perceptual Assimilation Model, Speech Learning Model and L2 Linguistic Perception Model.
Collapse
Affiliation(s)
- Yuxiao Yang
- Foreign Studies College, Hunan Normal University, Changsha, Hunan, China
| | - Sunfu Chen
- Foreign Studies College, Hunan Normal University, Changsha, Hunan, China
| | - Fei Chen
- School of Foreign Languages, Hunan University, Changsha, Hunan, China
| | - Junzhou Ma
- School of Foreign Languages, Taizhou University, Taizhou, Zhejiang, China
| |
Collapse
|
3
|
Implementation of an Early Communication Intervention for Young Children with Cerebral Palsy Using Single-Subject Research Design. J Clin Med 2022; 12:jcm12010232. [PMID: 36615031 PMCID: PMC9821676 DOI: 10.3390/jcm12010232] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2022] [Revised: 12/22/2022] [Accepted: 12/23/2022] [Indexed: 12/31/2022] Open
Abstract
The implementation of an intervention protocol aimed at increasing vocal complexity in three pre-linguistic children with cerebral palsy (two males, starting age 15 months, and one female, starting age 16 months) was evaluated utilising a repeated ABA case series design. The study progressed until the children were 36 months of age. Weekly probes with trained and untrained items were administered across each of three intervention blocks. Successive blocks targeted more advanced protophone production and speech movement patterns, individualised for each participant. Positive treatment effects were seen for all participants in terms of a greater rate of achievement of target protophone categories and speech movement patterns. Tau coefficients for trained items demonstrated overall moderate to large AB phase contrast effect sizes, with limited evidence of generalisation to untrained items. Control items featuring protophones and speech movements not targeted for intervention showed no change across phases for any participant. Our data suggest that emerging speech-production skills in prelinguistic infants with CP can be positively influenced through a multimodal intervention focused on capitalising on early periods of plasticity when language learning is most sensitive.
Collapse
|
4
|
Nittrouer S, Antonelli J, Lowenstein JH. The emergence of bifurcated structure in children's language. J Exp Psychol Gen 2022; 151:3045-3059. [PMID: 35696175 PMCID: PMC9673040 DOI: 10.1037/xge0001245] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Human language is unique among animal communication systems, in part because of its dual patterning in which meaningless phonological units combine to form meaningful words (phonological structure) and words combine to form sentences (lexicosyntactic structure). Although dual patterning is well recognized, its emergence in language development has been scarcely investigated. Chief among questions still unanswered is the extent to which development of these separate structures is independent or interdependent, and what supports acquisition of each level of structure. We explored these questions by examining growth of lexicosyntactic and phonological structure in children with normal hearing (n = 49) and children with hearing loss who use cochlear implants (n = 56). Multiple measures of each kind of structure were collected at 2-year intervals (kindergarten through eighth grade), and used to construct latent scores for each type of structure. Growth curve analysis assessed (a) the relative independence of development for each level of structure; (b) interactions between these two levels of structure in real-time language processing; and (c) contributions to growth of each level of structure made by auditory input, socioeconomic status (as proxy for linguistic experience), and speech motor control. Findings suggested that phonological and lexicosyntactic structure develop largely independently. Auditory input, socioeconomic status, and speech motor control help shape these language structures, with the last two factors exerting stronger effects for children with cochlear implants. Only for children with cochlear implants were interdependencies in real-time processing observed, reflecting compensatory mechanisms likely present to help them handle the disproportionately large phonological deficit they exhibit. (PsycInfo Database Record (c) 2022 APA, all rights reserved).
Collapse
Affiliation(s)
- Susan Nittrouer
- Department of Speech, Language, and Hearing Sciences,
University of Florida
| | | | | |
Collapse
|
5
|
Oakley M, Sande H. The Relationship between Non-Native Perception and Phonological Patterning of Implosive Consonants. LANGUAGE AND SPEECH 2022:238309221132495. [PMID: 36440824 DOI: 10.1177/00238309221132495] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
This study uses non-native perception data to examine the relationship between perceived phonetic similarity of segments and their phonological patterning. Segments that are phonetically similar to one another are anticipated to pattern together phonologically, and segments that share articulatory or acoustic properties are also expected to be perceived as similar. What is not yet clear is whether segments that pattern together phonologically are perceived as similar. This study addresses this question by examining how L1 English listeners and L1 Guébie listeners perceive non-native implosive consonants compared with plosives and sonorants. English does not have contrastive implosives, whereas Guébie has a bilabial implosive. The bilabial implosive phonologically patterns with sonorants in Guébie, to the exclusion of obstruents. Two perception experiments show English listeners make more perceptual categorization errors between implosives and voiced plosives than Guébie listeners do, but both listener groups are more likely to classify implosives as similar to voiced plosives than sonorants. The results also show that Guébie listeners are better at categorizing non-native implosive consonants (i.e., alveolar implosives) than English listeners, showing that listeners are able to extend features or gestures from their L1 to non-native implosive consonants. The results of these experiments suggest a cross-linguistic perceptual similarity hierarchy of implosives compared with other segments that are not affected by L1 phonological patterning.
Collapse
|
6
|
Anis FN, Umat C, Ahmad K, Abdul Hamid B. Arabic phoneme-grapheme correspondence by non-native, deaf children with cochlear implants and normal hearing children. Cochlear Implants Int 2022; 23:347-357. [PMID: 36005236 DOI: 10.1080/14670100.2022.2114583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/15/2022]
Abstract
OBJECTIVE This study aimed to compare the error patterns of Arabic phoneme-grapheme correspondence by a group of Malay children with cochlear implants (CIs) and normal hearing (NH) and the effects of the visual graphical features of Arabic graphemes (no-dot, single-dot, and multiple-dots) on the phoneme-grapheme correspondence. METHODS Participants were matched for hearing age (Mean, M = 7 ± 1.03 years) and duration of exposure to Arabic sounds (M = 2.7 ± 1.2 years). All 28 Arabic phonemes were presented through a loudspeaker and participants pointed to the graphemes associated with the presented phonemes. RESULTS A total of 336 and 616 tokens were collected for six children with CI and 11 NH children for each task, i.e., phonemes repetition and phoneme-grapheme correspondence. Both groups found it easier to repeat phonemes than the phoneme-grapheme correspondence. The children with CIs showed more confusion ([ظ, ز, ذ, ض, خ, ب, ه, س, ع, & ث] >10% correct scores) in phoneme-grapheme correspondence than the NH children ([ظ:14%] and [ث: 27%]). There was a significant interaction (p = 0.001) among the three visual graphical features and hearing status (CI and NH). CONCLUSION Our results infer that non-native Malay children with CIs and NH use different strategies to process the Arabic graphemes' visual features for phoneme-grapheme correspondence.
Collapse
Affiliation(s)
- Farheen Naz Anis
- Centre For Rehabilitation and Special Needs Studies, Faculty of Health Sciences, Universiti Kebangsaan Malaysia, Kuala Lumpur, Malaysia
| | - Cila Umat
- Centre For Rehabilitation and Special Needs Studies, Faculty of Health Sciences, Universiti Kebangsaan Malaysia, Kuala Lumpur, Malaysia
| | - Kartini Ahmad
- Centre For Rehabilitation and Special Needs Studies, Faculty of Health Sciences, Universiti Kebangsaan Malaysia, Kuala Lumpur, Malaysia
| | - Badrulzaman Abdul Hamid
- Centre For Rehabilitation and Special Needs Studies, Faculty of Health Sciences, Universiti Kebangsaan Malaysia, Kuala Lumpur, Malaysia
| |
Collapse
|
7
|
Zhao TC, Kuhl PK. Development of infants' neural speech processing and its relation to later language skills: A MEG study. Neuroimage 2022; 256:119242. [PMID: 35483648 DOI: 10.1016/j.neuroimage.2022.119242] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2021] [Revised: 04/14/2022] [Accepted: 04/21/2022] [Indexed: 11/28/2022] Open
Abstract
The 'sensitive period' for phonetic learning (∼6-12 months) is one of the earliest milestones in language acquisition where infants start to become specialized in processing speech sounds in their native language. In the last decade, advancements in neuroimaging technologies for infants are starting to shed light on the underlying neural mechanisms supporting this important learning period. The current study reports on a large longitudinal dataset with the aim to replicate and extend on two important questions: 1) what are the developmental changes during the 'sensitive period' for native and nonnative speech processing? 2) how does native and nonnative speech processing in infants predict later language outcomes? Fifty-four infants were recruited at 7 months of age and their neural processing of speech was measured using Magnetoencephalography (MEG). Specifically, the neural sensitivity to a native and a nonnative speech contrast was indexed by the mismatch response (MMR). They repeated the measurement again at 11 months of age and their language development was further tracked from 12 months to 30 months of age using the MacArthur-Bates Communicative Development Inventory (CDI). Using an a priori Region-of-Interest (ROI) approach, we observed significant increases for the Native MMR in the left inferior frontal region (IF) and superior temporal region (ST) from 7 to 11 months, but not for the Nonnative MMR. Complementary whole brain comparison revealed more widespread developmental changes for both contrasts. However, only individual differences in the left IF and ST for the Nonnative MMR at 11 months of age were significant predictors of individual vocabulary growth up to 30 months of age. An exploratory machine-learning based analysis further revealed that whole brain time series for both Native and Nonnative contrasts can robustly predict later outcomes, but with very different underlying spatial-temporal patterns. The current study extends our current knowledge and suggests that native and nonnative speech processing may follow different developmental trajectories and utilize different mechanisms that are relevant for later language skills.
Collapse
Affiliation(s)
- T Christina Zhao
- Institute for Learning & Brain Sciences, University of Washington, Seattle, Washington, United States; Department of Speech and Hearing Sciences, University of Washington, Seattle, Washington, United States.
| | - Patricia K Kuhl
- Institute for Learning & Brain Sciences, University of Washington, Seattle, Washington, United States; Department of Speech and Hearing Sciences, University of Washington, Seattle, Washington, United States
| |
Collapse
|
8
|
Zhao TC, Llanos F, Chandrasekaran B, Kuhl PK. Language experience during the sensitive period narrows infants' sensory encoding of lexical tones-Music intervention reverses it. Front Hum Neurosci 2022; 16:941853. [PMID: 36016666 PMCID: PMC9398460 DOI: 10.3389/fnhum.2022.941853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Accepted: 07/19/2022] [Indexed: 01/13/2023] Open
Abstract
The sensitive period for phonetic learning (6∼12 months), evidenced by improved native speech processing and declined non-native speech processing, represents an early milestone in language acquisition. We examined the extent that sensory encoding of speech is altered by experience during this period by testing two hypotheses: (1) early sensory encoding of non-native speech declines as infants gain native-language experience, and (2) music intervention reverses this decline. We longitudinally measured the frequency-following response (FFR), a robust indicator of early sensory encoding along the auditory pathway, to a Mandarin lexical tone in 7- and 11-months-old monolingual English-learning infants. Infants received either no intervention (language-experience group) or music intervention (music-intervention group) randomly between FFR recordings. The language-experience group exhibited the expected decline in FFR pitch-tracking accuracy to the Mandarin tone, while the music-intervention group did not. Our results support both hypotheses and demonstrate that both language and music experiences alter infants' speech encoding.
Collapse
Affiliation(s)
- Tian Christina Zhao
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA, United States
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, United States
| | - Fernando Llanos
- Department of Linguistics, University of Texas at Austin, Austin, TX, United States
| | - Bharath Chandrasekaran
- Department of Communication Sciences and Disorders, University of Pittsburgh, Pittsburgh, PA, United States
| | - Patricia K. Kuhl
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA, United States
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, United States
| |
Collapse
|
9
|
Matsui T, Uchida M, Fujino H, Tojo Y, Hakarino K. Perception of native and non-native phonemic contrasts in children with autistic spectrum disorder: effects of speaker variability. CLINICAL LINGUISTICS & PHONETICS 2022; 36:417-435. [PMID: 34460348 DOI: 10.1080/02699206.2021.1947385] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/30/2020] [Revised: 06/07/2021] [Accepted: 06/17/2021] [Indexed: 06/13/2023]
Abstract
The current study investigated speech perception in children with ASD by directly comparing discrimination accuracy of phonemic contrasts in the native and non-native languages. The effect of speaker variability on phoneme perception was also examined. We also explored the relation between language impairment and accuracy in phoneme discrimination in children with ASD. Significant differences in performance were found between the ASD and TD groups on discrimination of the native phonemic contrasts. By contrast, no difference was found between the two groups on discrimination of the non-native phonemic contrasts. Further subgroup analysis revealed that the ALN group (ASD without language delay or impairment) showed significantly higher discrimination accuracy for the native syllable contrasts than the non-native counterpart. No significant difference was found in the discrimination accuracy between the native and non-native phonemic contrasts in the ALD group (ASD with language delay or impairment). The effect of speaker viability on phoneme discrimination was observed in the TD group but not in the ASD subgroups. Nonverbal reasoning ability was highly related to discrimination accuracy of both the native and non-native phonemic contrasts in children with ASD. The results of the present study suggest that speech perception in children with ASD is not as attuned to their native language as in their TD peers. Our findings also indicate that language delay or impairment is related to difficulty in perception of native phonemes in children with ASD.
Collapse
Affiliation(s)
- Tomoko Matsui
- Center for Research in International Education, Tokyo Gakugei University, Tokyo, Japan
| | - Mariko Uchida
- Center for Research in International Education, Tokyo Gakugei University, Tokyo, Japan
| | - Hiroshi Fujino
- Graduate School for Teacher Education, Tokyo Gakugei University, Tokyo, Japan
| | - Yoshikuni Tojo
- College of Education, Ibaraki University, Ibaraki, Japan
| | | |
Collapse
|
10
|
Novitskiy N, Maggu AR, Lai CM, Chan PHY, Wong KHY, Lam HS, Leung TY, Leung TF, Wong PCM. Early Development of Neural Speech Encoding Depends on Age but Not Native Language Status: Evidence From Lexical Tone. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2022; 3:67-86. [PMID: 37215329 PMCID: PMC10178623 DOI: 10.1162/nol_a_00049] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/21/2020] [Accepted: 06/22/2021] [Indexed: 05/24/2023]
Abstract
We investigated the development of early-latency and long-latency brain responses to native and non-native speech to shed light on the neurophysiological underpinnings of perceptual narrowing and early language development. Specifically, we postulated a two-level process to explain the decrease in sensitivity to non-native phonemes toward the end of infancy. Neurons at the earlier stages of the ascending auditory pathway mature rapidly during infancy facilitating the encoding of both native and non-native sounds. This growth enables neurons at the later stages of the auditory pathway to assign phonological status to speech according to the infant's native language environment. To test this hypothesis, we collected early-latency and long-latency neural responses to native and non-native lexical tones from 85 Cantonese-learning children aged between 23 days and 24 months, 16 days. As expected, a broad range of presumably subcortical early-latency neural encoding measures grew rapidly and substantially during the first two years for both native and non-native tones. By contrast, long-latency cortical electrophysiological changes occurred on a much slower scale and showed sensitivity to nativeness at around six months. Our study provided a comprehensive understanding of early language development by revealing the complementary roles of earlier and later stages of speech processing in the developing brain.
Collapse
Affiliation(s)
- Nikolay Novitskiy
- Department of Linguistics and Modern Languages, Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Akshay R. Maggu
- Department of Linguistics and Modern Languages, Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
- O-lab, Duke Psychology and Neuroscience, Duke University, Durham, NC, USA
| | - Ching Man Lai
- Department of Linguistics and Modern Languages, Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Peggy H. Y. Chan
- Department of Linguistics and Modern Languages, Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
- Department of Paediatrics, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Kay H. Y. Wong
- Department of Linguistics and Modern Languages, Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Hugh Simon Lam
- Department of Paediatrics, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Tak Yeung Leung
- Department of Obsterics and Gynaecology, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Ting Fan Leung
- Department of Paediatrics, The Chinese University of Hong Kong, Hong Kong SAR, China
| | - Patrick C. M. Wong
- Department of Linguistics and Modern Languages, Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong SAR, China
| |
Collapse
|
11
|
Ward R, Hennessey N, Barty E, Elliott C, Valentine J, Cantle Moore R. Clinical utilisation of the Infant Monitor of vocal Production (IMP) for early identification of communication impairment in young infants at-risk of cerebral palsy: a prospective cohort study. Dev Neurorehabil 2022; 25:101-114. [PMID: 34241555 DOI: 10.1080/17518423.2021.1942280] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
AIM To report prospective longitudinal data of early vocaliszations of infants identified "at-risk" of cerebral palsy (CP) for early identification of communication impairment. METHOD This case-control longitudinal prospective cohort study reports on the assessment of 36 infants, 18 identified as at-risk of CP at the time of enrolment and 18 typically developing (TD) children, at three time points: 6 months, 9 months and 12 months of age, Data were obtained through criterion and norm referenced assessments of vocaliszation behaviors. RESULTS Early vocal behaviors of infants identified as at-risk of CP did not differ from their age matched peers at 6 months of age, however, significant group differences emerged at 9 and 12 months when pre-canonical and canonical babble typically emerge. Generalized linear mixed models analysis showed that the rate of development of early language ability and more complex speech-related vocal behaviors was slower for infants at risk of CP when compared to TD infants, with over 75% of infants with CP showing below normal vocal production and impaired language by 12 months of age. INTERPRETATION Our data suggest characteristics of infant vocalizations associated with pre-canonical and canonical babbling provide a strong evidence base for predicting communication outcomes in infants at risk of CP.
Collapse
Affiliation(s)
- R Ward
- Kids Rehab, Perth Children's, Hospital, Perth, Australia.,School of Allied Health, Curtin University, Perth, Australia.,Institute of Health Research, University of Notre Dame Australia, Fremantle, Australia
| | - N Hennessey
- School of Allied Health, Curtin University, Perth, Australia
| | - E Barty
- Kids Rehab, Perth Children's, Hospital, Perth, Australia
| | - C Elliott
- Kids Rehab, Perth Children's, Hospital, Perth, Australia.,School of Allied Health, Curtin University, Perth, Australia.,Telethon Kids Institute, Perth, Australia
| | - J Valentine
- Kids Rehab, Perth Children's, Hospital, Perth, Australia
| | - R Cantle Moore
- NextSense Institute/Macquarie University, Sydney, New South Wales
| |
Collapse
|
12
|
Polka L, Masapollo M, Ménard L. Setting the Stage for Speech Production: Infants Prefer Listening to Speech Sounds With Infant Vocal Resonances. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:109-120. [PMID: 34889651 DOI: 10.1044/2021_jslhr-21-00412] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
PURPOSE Current models of speech development argue for an early link between speech production and perception in infants. Recent data show that young infants (at 4-6 months) preferentially attend to speech sounds (vowels) with infant vocal properties compared to those with adult vocal properties, suggesting the presence of special "memory banks" for one's own nascent speech-like productions. This study investigated whether the vocal resonances (formants) of the infant vocal tract are sufficient to elicit this preference and whether this perceptual bias changes with age and emerging vocal production skills. METHOD We selectively manipulated the fundamental frequency (f0 ) of vowels synthesized with formants specifying either an infant or adult vocal tract, and then tested the effects of those manipulations on the listening preferences of infants who were slightly older than those previously tested (at 6-8 months). RESULTS Unlike findings with younger infants (at 4-6 months), slightly older infants in Experiment 1 displayed a robust preference for vowels with infant formants over adult formants when f0 was matched. The strength of this preference was also positively correlated with age among infants between 4 and 8 months. In Experiment 2, this preference favoring infant over adult formants was maintained when f0 values were modulated. CONCLUSIONS Infants between 6 and 8 months of age displayed a robust and distinct preference for speech with resonances specifying a vocal tract that is similar in size and length to their own. This finding, together with data indicating that this preference is not present in younger infants and appears to increase with age, suggests that nascent knowledge of the motor schema of the vocal tract may play a role in shaping this perceptual bias, lending support to current models of speech development. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.17131805.
Collapse
Affiliation(s)
- Linda Polka
- School of Communication Sciences and Disorders, McGill University, Montréal, Québec, Canada
- Center for Research on Brain, Language and Music, McGill University, Montréal, Québec, Canada
| | - Matthew Masapollo
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville
| | - Lucie Ménard
- Center for Research on Brain, Language and Music, McGill University, Montréal, Québec, Canada
- Department of Linguistics, Université du Québec à Montréal, Canada
| |
Collapse
|
13
|
Falk S, Audibert N. Acoustic signatures of communicative dimensions in codified mother-infant interactions. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:4429. [PMID: 34972287 DOI: 10.1121/10.0008977] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/18/2021] [Accepted: 11/20/2021] [Indexed: 06/14/2023]
Abstract
Nursery rhymes, lullabies, or traditional stories are pieces of oral tradition that constitute an integral part of communication between caregivers and preverbal infants. Caregivers use a distinct acoustic style when singing or narrating to their infants. Unlike spontaneous infant-directed (ID) interactions, codified interactions benefit from highly stable acoustics due to their repetitive character. The aim of the study was to determine whether specific combinations of acoustic traits (i.e., vowel pitch, duration, spectral structure, and their variability) form characteristic "signatures" of different communicative dimensions during codified interactions, such as vocalization type, interactive stimulation, and infant-directedness. Bayesian analysis, applied to over 14 000 vowels from codified live interactions between mothers and their 6-months-old infants, showed that a few acoustic traits prominently characterize arousing vs calm interactions and sung vs spoken interactions. While pitch and duration and their variation played a prominent role in constituting these signatures, more linguistic aspects such as vowel clarity showed small or no effects. Infant-directedness was identifiable in a larger set of acoustic cues than the other dimensions. These findings provide insights into the functions of acoustic variation of ID communication and into the potential role of codified interactions for infants' learning about communicative intent and expressive forms typical of language and music.
Collapse
Affiliation(s)
- Simone Falk
- Département de Linguistique et de traduction, Université de Montréal, 3150, rue Jean-Brillant, Montreal, Quebec, H3T 1N8, Canada
| | - Nicolas Audibert
- Laboratoire de Phonétique et Phonologie, UMR7018, CNRS/Université Sorbonne-Nouvelle, 19 Rue des Bernardins, Paris, 75005, France
| |
Collapse
|
14
|
White KS, Daub O. When it's not appropriate to adapt: Toddlers' learning of novel speech patterns is affected by visual information. BRAIN AND LANGUAGE 2021; 222:105022. [PMID: 34536771 DOI: 10.1016/j.bandl.2021.105022] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2021] [Revised: 08/22/2021] [Accepted: 08/30/2021] [Indexed: 06/13/2023]
Abstract
In adults, perceptual learning for speech is constrained, such that learning of novel pronunciations is less likely to occur if the (e.g., visual) context indicates that they are transient. However, adults have had a lifetime of experience with the types of cues that signal stable vs. transient speech variation. We ask whether visual context affects toddlers' learning of a novel speech pattern. Across conditions, 19-month-olds (N = 117) were exposed to familiar words either pronounced typically or in a novel, consonant-shifting accent. During exposure, some toddlers heard the accented pronunciations without a face present; others saw a video of the speaker producing the words with a lollipop against her cheek or in her mouth. Toddlers showed the weakest learning of the accent when the speaker had the lollipop in her mouth, suggesting that they treated the lollipop as the cause of the atypical pronunciations. These results demonstrate that toddlers' adaptation to a novel speech pattern is influenced by extra-linguistic context.
Collapse
|
15
|
Ma J, Zhu J, Yang Y, Chen F. The Development of Categorical Perception of Segments and Suprasegments in Mandarin-Speaking Preschoolers. Front Psychol 2021; 12:693366. [PMID: 34354636 PMCID: PMC8329735 DOI: 10.3389/fpsyg.2021.693366] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2021] [Accepted: 05/27/2021] [Indexed: 11/13/2022] Open
Abstract
This study investigated the developmental trajectories of categorical perception (CP) of segments (i.e., stops) and suprasegments (i.e., lexical tones) in an attempt to examine the perceptual development of phonological categories and whether CP of suprasegments develops in parallel with that of segments. Forty-seven Mandarin-speaking monolingual preschoolers aged four to six years old, and fourteen adults completed both identification and discrimination tasks of the Tone 1-2 continuum and the /pa/-/pha/ continuum. Results revealed that children could perceive both lexical tones and aspiration of stops in a categorical manner by age four. The boundary position did not depend on age, with children having similar positions to adults regardless of speech continuum types. The boundary width, on the other hand, reached the adult-like level at age six for lexical tones, but not for stops. In addition, the within-category discrimination score did not differ significantly between children and adults for both continua. The between-category discrimination score improved with age and achieved the adult-like level at age five for lexical tones, but still not for stops even at age six. It suggests that the fine-grained perception of phonological categories is a protracted process, and the improvement and varying timeline of the development of segments and suprasegments are discussed in relation to statistical learning of the regularities of speech sounds in ambient language, ongoing maturation of perceptual systems, the memory mechanism underlying perceptual learning, and the intrinsic nature of speech elements.
Collapse
Affiliation(s)
- Junzhou Ma
- School of Foreign Languages, Taizhou University, Taizhou, China
| | - Jiaqiang Zhu
- School of Foreign Languages, Hunan University, Changsha, China
| | - Yuxiao Yang
- Foreign Studies College, Hunan Normal University, Changsha, China
| | - Fei Chen
- School of Foreign Languages, Hunan University, Changsha, China
| |
Collapse
|
16
|
Polka L, Molnar M, Zhao TC, Masapollo M. Neurophysiological Correlates of Asymmetries in Vowel Perception: An English-French Cross-Linguistic Event-Related Potential Study. Front Hum Neurosci 2021; 15:607148. [PMID: 34149375 PMCID: PMC8209302 DOI: 10.3389/fnhum.2021.607148] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2020] [Accepted: 05/03/2021] [Indexed: 11/14/2022] Open
Abstract
Behavioral studies examining vowel perception in infancy indicate that, for many vowel contrasts, the ease of discrimination changes depending on the order of stimulus presentation, regardless of the language from which the contrast is drawn and the ambient language that infants have experienced. By adulthood, linguistic experience has altered vowel perception; analogous asymmetries are observed for non-native contrasts but are mitigated for native contrasts. Although these directional effects are well documented behaviorally, the brain mechanisms underlying them are poorly understood. In the present study we begin to address this gap. We first review recent behavioral work which shows that vowel perception asymmetries derive from phonetic encoding strategies, rather than general auditory processes. Two existing theoretical models-the Natural Referent Vowel framework and the Native Language Magnet model-are invoked as a means of interpreting these findings. Then we present the results of a neurophysiological study which builds on this prior work. Using event-related brain potentials, we first measured and assessed the mismatch negativity response (MMN, a passive neurophysiological index of auditory change detection) in English and French native-speaking adults to synthetic vowels that either spanned two different phonetic categories (/y/vs./u/) or fell within the same category (/u/). Stimulus presentation was organized such that each vowel was presented as standard and as deviant in different blocks. The vowels were presented with a long (1,600-ms) inter-stimulus interval to restrict access to short-term memory traces and tap into a "phonetic mode" of processing. MMN analyses revealed weak asymmetry effects regardless of the (i) vowel contrast, (ii) language group, and (iii) MMN time window. Then, we conducted time-frequency analyses of the standard epochs for each vowel. In contrast to the MMN analysis, time-frequency analysis revealed significant differences in brain oscillations in the theta band (4-8 Hz), which have been linked to attention and processing efficiency. Collectively, these findings suggest that early-latency (pre-attentive) mismatch responses may not be a strong neurophysiological correlate of asymmetric behavioral vowel discrimination. Rather, asymmetries may reflect differences in neural processing efficiency for vowels with certain inherent acoustic-phonetic properties, as revealed by theta oscillatory activity.
Collapse
Affiliation(s)
- Linda Polka
- School of Communication Sciences and Disorders, McGill University, Montreal, QC, Canada
- Center for Research on Brain, Language, and Music, McGill University, Montreal, QC, Canada
| | - Monika Molnar
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
| | - T. Christina Zhao
- Institute for Learning and Brain Sciences, University of Washington, Seattle, WA, United States
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, United States
| | - Matthew Masapollo
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville, FL, United States
| |
Collapse
|
17
|
Zhao TC, Boorom O, Kuhl PK, Gordon R. Infants' neural speech discrimination predicts individual differences in grammar ability at 6 years of age and their risk of developing speech-language disorders. Dev Cogn Neurosci 2021; 48:100949. [PMID: 33823366 PMCID: PMC8047161 DOI: 10.1016/j.dcn.2021.100949] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2020] [Revised: 03/23/2021] [Accepted: 03/27/2021] [Indexed: 11/17/2022] Open
Abstract
The 'sensitive period' for phonetic learning posits that between 6 and 12 months of age, infants' discrimination of native and nonnative speech sounds diverge. Individual differences in this dynamic processing of speech have been shown to predict later language acquisition up to 30 months of age, using parental surveys. Yet, it is unclear whether infant speech discrimination could predict longer-term language outcome and risk for developmental speech-language disorders, which affect up to 16 % of the population. The current study reports a prospective prediction of speech-language skills at a much later age-6 years-old-from the same children's nonnative speech discrimination at 11 months-old, indexed by MEG mismatch responses. Children's speech-language skills at 6 were comprehensively evaluated by a speech-language pathologist in two ways: individual differences in spoken grammar, and the presence versus absence of speech-language disorders. Results showed that the prefrontal MEG mismatch response at 11 months not only significantly predicted individual differences in spoken grammar skills at 6 years, but also accurately identified the presence versus absence of speech-language disorders, using a machine-learning classification. These results represent new evidence that advance our theoretical understanding of the neurodevelopmental trajectory of language acquisition and early risk factors for developmental speech-language disorders.
Collapse
Affiliation(s)
- T Christina Zhao
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA, USA; Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA.
| | - Olivia Boorom
- Department of Hearing & Speech Sciences, Vanderbilt University Medical Center, Nashville, TN, USA
| | - Patricia K Kuhl
- Institute for Learning & Brain Sciences, University of Washington, Seattle, WA, USA; Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Reyna Gordon
- Department of Otolaryngology and Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA; Vanderbilt Brain Institute, Vanderbilt University Medical Center, Nashville, TN, USA; Vanderbilt Kennedy Center, Vanderbilt University Medical Center, Nashville, TN, USA
| |
Collapse
|
18
|
Zhu J, Chen X, Yang Y. Effects of Amateur Musical Experience on Categorical Perception of Lexical Tones by Native Chinese Adults: An ERP Study. Front Psychol 2021; 12:611189. [PMID: 33790832 PMCID: PMC8005537 DOI: 10.3389/fpsyg.2021.611189] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2020] [Accepted: 02/17/2021] [Indexed: 01/29/2023] Open
Abstract
Music impacting on speech processing is vividly evidenced in most reports involving professional musicians, while the question of whether the facilitative effects of music are limited to experts or may extend to amateurs remains to be resolved. Previous research has suggested that analogous to language experience, musicianship also modulates lexical tone perception but the influence of amateur musical experience in adulthood is poorly understood. Furthermore, little is known about how acoustic information and phonological information of lexical tones are processed by amateur musicians. This study aimed to provide neural evidence of cortical plasticity by examining categorical perception of lexical tones in Chinese adults with amateur musical experience relative to the non-musician counterparts. Fifteen adult Chinese amateur musicians and an equal number of non-musicians participated in an event-related potential (ERP) experiment. Their mismatch negativities (MMNs) to lexical tones from Mandarin Tone 2–Tone 4 continuum and non-speech tone analogs were measured. It was hypothesized that amateur musicians would exhibit different MMNs to their non-musician counterparts in processing two aspects of information in lexical tones. Results showed that the MMN mean amplitude evoked by within-category deviants was significantly larger for amateur musicians than non-musicians regardless of speech or non-speech condition. This implies the strengthened processing of acoustic information by adult amateur musicians without the need of focused attention, as the detection of subtle acoustic nuances of pitch was measurably improved. In addition, the MMN peak latency elicited by across-category deviants was significantly shorter than that by within-category deviants for both groups, indicative of the earlier processing of phonological information than acoustic information of lexical tones at the pre-attentive stage. The results mentioned above suggest that cortical plasticity can still be induced in adulthood, hence non-musicians should be defined more strictly than before. Besides, the current study enlarges the population demonstrating the beneficial effects of musical experience on perceptual and cognitive functions, namely, the effects of enhanced speech processing from music are not confined to a small group of experts but extend to a large population of amateurs.
Collapse
Affiliation(s)
- Jiaqiang Zhu
- School of Foreign Languages, Hunan University, Changsha, China
| | - Xiaoxiang Chen
- School of Foreign Languages, Hunan University, Changsha, China
| | - Yuxiao Yang
- Foreign Studies College, Hunan Normal University, Changsha, China
| |
Collapse
|
19
|
Falk S, Fasolo M, Genovese G, Romero‐Lauro L, Franco F. Sing for me, Mama! Infants' discrimination of novel vowels in song. INFANCY 2021; 26:248-270. [DOI: 10.1111/infa.12387] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2020] [Revised: 12/08/2020] [Accepted: 12/14/2020] [Indexed: 01/22/2023]
Affiliation(s)
- Simone Falk
- Department of Linguistics and Translation University of Montreal, Montreal Quebec Canada
- Laboratory for Brain, Music and Sound Research (BRAMS) University of Montreal, Montreal Quebec Canada
| | - Mirco Fasolo
- Department of Neuroscience, Imaging, and Clinical Sciences University "G. d'Annunzio" of Chieti‐Pescara Chieti Italy
| | | | - Leonor Romero‐Lauro
- Department of Psychology University of Milan‐Bicocca Milano Italy
- Neuromi Milan Center for Neuroscience Milano Italy
| | - Fabia Franco
- Department of Psychology Faculty of Science and Technology Middlesex University London London UK
| |
Collapse
|
20
|
Masapollo M, Guenther FH. Engaging the Articulators Enhances Perception of Concordant Visible Speech Movements. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:3679-3688. [PMID: 31577522 PMCID: PMC7201334 DOI: 10.1044/2019_jslhr-s-19-0167] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/17/2023]
Abstract
Purpose This study aimed to test whether (and how) somatosensory feedback signals from the vocal tract affect concurrent unimodal visual speech perception. Method Participants discriminated pairs of silent visual utterances of vowels under 3 experimental conditions: (a) normal (baseline) and while holding either (b) a bite block or (c) a lip tube in their mouths. To test the specificity of somatosensory-visual interactions during perception, we assessed discrimination of vowel contrasts optically distinguished based on their mandibular (English /ɛ/-/æ/) or labial (English /u/-French /u/) postures. In addition, we assessed perception of each contrast using dynamically articulating videos and static (single-frame) images of each gesture (at vowel midpoint). Results Engaging the jaw selectively facilitated perception of the dynamic gestures optically distinct in terms of jaw height, whereas engaging the lips selectively facilitated perception of the dynamic gestures optically distinct in terms of their degree of lip compression and protrusion. Thus, participants perceived visible speech movements in relation to the configuration and shape of their own vocal tract (and possibly their ability to produce covert vowel production-like movements). In contrast, engaging the articulators had no effect when the speaking faces did not move, suggesting that the somatosensory inputs affected perception of time-varying kinematic information rather than changes in target (movement end point) mouth shapes. Conclusions These findings suggest that orofacial somatosensory inputs associated with speech production prime premotor and somatosensory brain regions involved in the sensorimotor control of speech, thereby facilitating perception of concordant visible speech movements. Supplemental Material https://doi.org/10.23641/asha.9911846.
Collapse
Affiliation(s)
- Matthew Masapollo
- Department of Speech, Language and Hearing Sciences, Boston University, MA
| | - Frank H. Guenther
- Department of Speech, Language and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
| |
Collapse
|
21
|
Redford MA. Speech Production From a Developmental Perspective. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:2946-2962. [PMID: 31465709 PMCID: PMC6813032 DOI: 10.1044/2019_jslhr-s-csmc7-18-0130] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
Purpose Current approaches to speech production aim to explain adult behavior and so make assumptions that, when taken to their logical conclusion, fail to adequately account for development. This failure is problematic if adult behavior can be understood to emerge from the developmental process. This problem motivates the proposal of a developmentally sensitive theory of speech production. The working hypothesis, which structures the theory, is that feedforward representations and processes mature earlier than central feedback control processes in speech production. Method Theoretical assumptions that underpin the 2 major approaches to adult speech production are reviewed. Strengths and weaknesses are evaluated with respect to developmental patterns. A developmental approach is then pursued. The strengths of existing theories are borrowed, and the ideas are resynthesized under the working hypothesis. The speech production process is then reimagined in developmental stages, with each stage building on the previous one. Conclusion The resulting theory proposes that speech production relies on conceptually linked representations that are information-reduced holistic perceptual and motoric forms, constituting the phonological aspect of a system that is acquired with the lexicon. These forms are referred to as exemplars and schemas, respectively. When a particular exemplar and schema are activated with the selection of a particular lexical concept, their forms are used to define unique trajectories through an endogenous perceptual-motor space that guides implementation. This space is not linguistic, reflecting its origin in the prespeech period. Central feedback control over production emerges with failures in communication and the development of a self-concept.
Collapse
|
22
|
Yang Y, Chen X. Within-organ contrast in second language perception: The perception of Russian initial /r-l/ contrast by Chinese learners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:EL117. [PMID: 31472589 DOI: 10.1121/1.5120549] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/14/2019] [Accepted: 07/16/2019] [Indexed: 06/10/2023]
Abstract
This study examined the perception of the Russian initial /r-l/ contrast by Chinese learners in an attempt to extend the Perceptual Assimilation Model with the Articulatory Organ Hypothesis (PAM-AOH) to second language (L2) perception. Results of a perceptual assimilation task and a discrimination test showed that the assimilation pattern of Russian /r-l/ relative to Mandarin was uncategorized categorized, the discriminability of this within-organ contrast, however, was only moderate even for the more advanced Chinese learners of Russian. These findings suggest that L2 perception could be susceptible to both the L1/L2 assimilation patterns and the contrast types as suggested by PAM-AOH.
Collapse
Affiliation(s)
- Yuxiao Yang
- School of Foreign Languages, Hunan University, South Lushan Road, Yuelu District, Changsha 410082, ,
| | - Xiaoxiang Chen
- School of Foreign Languages, Hunan University, South Lushan Road, Yuelu District, Changsha 410082, ,
| |
Collapse
|
23
|
Kröger BJ, Bafna T, Cao M. Emergence of an Action Repository as Part of a Biologically Inspired Model of Speech Processing: The Role of Somatosensory Information in Learning Phonetic-Phonological Sound Features. Front Psychol 2019; 10:1462. [PMID: 31354560 PMCID: PMC6635888 DOI: 10.3389/fpsyg.2019.01462] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2019] [Accepted: 06/07/2019] [Indexed: 11/13/2022] Open
Abstract
A comprehensive model of speech processing and speech learning has been established. The model comprises a mental lexicon, an action repository and an articulatory-acoustic module for executing motor plans and generating auditory and somatosensory feedback information (Kröger and Cao, 2015). In this study a “model language” based on three auditory and motor realizations of 70 monosyllabic words has been trained in order to simulate early phases of speech acquisition (babbling and imitation). We were able to show that (i) the emergence of phonetic-phonological features results from an increasing degree of ordering of syllable representations within the action repository and that (ii) this ordering or arrangement of syllables is mainly shaped by auditory information. Somatosensory information helps to increase the speed of learning. Especially consonantal features like place of articulation are learned earlier if auditory information is accompanied by somatosensory information. It can be concluded that somatosensory information as it is generated already during the babbling and the imitation phase of speech acquisition is very helpful especially for learning features like place of articulation. After learning is completed acoustic information together with semantic information is sufficient for determining the phonetic-phonological information from the speech signal. Moreover it is possible to learn phonetic-phonological features like place of articulation from auditory and semantic information only but not as fast as when somatosensory information is also available during the early stages of learning.
Collapse
Affiliation(s)
- Bernd J Kröger
- Neurophonetics Group, Department of Phoniatrics, Pedaudiology, and Communication Disorders, Medical School, RWTH Aachen University, Aachen, Germany
| | - Tanya Bafna
- Medical School, RWTH Aachen University, Aachen, Germany
| | - Mengxue Cao
- School of Chinese Language and Literature, Beijing Normal University, Beijing, China
| |
Collapse
|
24
|
Imafuku M, Kanakogi Y, Butler D, Myowa M. Demystifying infant vocal imitation: The roles of mouth looking and speaker's gaze. Dev Sci 2019; 22:e12825. [PMID: 30980494 DOI: 10.1111/desc.12825] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2017] [Revised: 01/08/2019] [Accepted: 03/01/2019] [Indexed: 12/20/2022]
Abstract
Vocal imitation plays a fundamental role in human language acquisition from infancy. Little is known, however, about how infants imitate other's sounds. We focused on three factors: (a) whether infants receive information from upright faces, (b) the infant's observation of the speaker's mouth and (c) the speaker directing their gaze towards the infant. We recorded the eye movements of 6-month-olds who participated in experiments watching videos of a speaker producing vowel sounds. We found that an infants' tendency to vocally imitate such videos increased as a function of (a) seeing upright rather than inverted faces, (b) their increased looking towards the speaker's mouth and (c) whether the speaker directed their gaze towards, rather than away from infants. These latter findings are consistent with theories of motor resonance and natural pedagogy respectively. New light has been shed on the cues and underlying mechanisms linking infant speech perception and production.
Collapse
Affiliation(s)
- Masahiro Imafuku
- Graduate School of Education, Kyoto University, Kyoto, Japan.,Faculty of Education, Musashino University, Tokyo, Japan
| | | | - David Butler
- Graduate School of Education, Kyoto University, Kyoto, Japan.,The Institute for Social Neuroscience Psychology, Heidelberg, Victoria, Australia
| | - Masako Myowa
- Graduate School of Education, Kyoto University, Kyoto, Japan
| |
Collapse
|
25
|
Anis FN, Umat C, Ahmad K, Hamid BA. Patterns of recognition of Arabic consonants by non-native children with cochlear implants and normal hearing. Cochlear Implants Int 2018; 20:12-22. [PMID: 30293522 DOI: 10.1080/14670100.2018.1530420] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Abstract
OBJECTIVE This study examined the patterns of recognition of Arabic consonants, via information transmission analysis for phonological features, in a group of Malay children with normal hearing (NH) and cochlear implants (CI). METHOD A total of 336 and 616 acoustic tokens were collected from six CI and 11 NH Malay children, respectively. The groups were matched for hearing age and duration of exposure to Arabic sounds. All the 28 Arabic consonants in the form of consonant-vowel /a/ were presented randomly twice via a loudspeaker at approximately 65 dB SPL. The participants were asked to repeat verbally the stimulus heard in each presentation. RESULTS Within the native Malay perceptual space, the two groups responded differently to the Arabic consonants. The dispersed uncategorized assimilation in the CI group was distinct in the confusion matrix (CM), as compared to the NH children. Consonants /ħ/, /tˁ/, /sˁ/ and /ʁ/ were difficult for the CI children, while the most accurate item was /k/ (84%). The CI group transmitted significantly reduced information, especially for place feature transmission, then the NH group (p < 0.001). Significant interactions between place-hearing status and manner-hearing status were also obtained, suggesting there were information transmission differences in the pattern of consonants recognition between the study groups. CONCLUSION CI and NH Malay children may be using different acoustic cues to recognize Arabic sounds, which contribute to the different assimilation categories' patterns within the Malay perceptual space.
Collapse
Affiliation(s)
- Farheen Naz Anis
- a Centre For Rehabilitation and Special Needs, Faculty of Health Sciences , Universiti Kebangsaan Malaysia , Jalan Raja Muda Abdul Aziz 50300 , Kuala Lumpur , Malaysia
| | - Cila Umat
- a Centre For Rehabilitation and Special Needs, Faculty of Health Sciences , Universiti Kebangsaan Malaysia , Jalan Raja Muda Abdul Aziz 50300 , Kuala Lumpur , Malaysia.,b Institute of Ear, Hearing & Speech, Universiti Kebangsaan Malaysia , Kuala Lumpur , Malaysia
| | - Kartini Ahmad
- a Centre For Rehabilitation and Special Needs, Faculty of Health Sciences , Universiti Kebangsaan Malaysia , Jalan Raja Muda Abdul Aziz 50300 , Kuala Lumpur , Malaysia
| | - Badrulzaman Abdul Hamid
- a Centre For Rehabilitation and Special Needs, Faculty of Health Sciences , Universiti Kebangsaan Malaysia , Jalan Raja Muda Abdul Aziz 50300 , Kuala Lumpur , Malaysia
| |
Collapse
|
26
|
Masapollo M, Polka L, Ménard L, Franklin L, Tiede M, Morgan J. Asymmetries in unimodal visual vowel perception: The roles of oral-facial kinematics, orientation, and configuration. J Exp Psychol Hum Percept Perform 2018; 44:1103-1118. [PMID: 29517257 PMCID: PMC6037555 DOI: 10.1037/xhp0000518] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Masapollo, Polka, and Ménard (2017) recently reported a robust directional asymmetry in unimodal visual vowel perception: Adult perceivers discriminate a change from an English /u/ viseme to a French /u/ viseme significantly better than a change in the reverse direction. This asymmetry replicates a frequent pattern found in unimodal auditory vowel perception that points to a universal bias favoring more extreme vocalic articulations, which lead to acoustic signals with increased formant convergence. In the present article, the authors report 5 experiments designed to investigate whether this asymmetry in the visual realm reflects a speech-specific or general processing bias. They successfully replicated the directional effect using Masapollo et al.'s dynamically articulating faces but failed to replicate the effect when the faces were shown under static conditions. Asymmetries also emerged during discrimination of canonically oriented point-light stimuli that retained the kinematics and configuration of the articulating mouth. In contrast, no asymmetries emerged during discrimination of rotated point-light stimuli or Lissajou patterns that retained the kinematics, but not the canonical orientation or spatial configuration, of the labial gestures. These findings suggest that the perceptual processes underlying asymmetries in unimodal visual vowel discrimination are sensitive to speech-specific motion and configural properties and raise foundational questions concerning the role of specialized and general processes in vowel perception. (PsycINFO Database Record
Collapse
Affiliation(s)
- Matthew Masapollo
- Brown University
- McGill University
- Centre for Research on Brain, Language, and Music
| | - Linda Polka
- McGill University
- Centre for Research on Brain, Language, and Music
| | - Lucie Ménard
- Centre for Research on Brain, Language, and Music
- University of Quebec at Montreal
| | | | | | | |
Collapse
|
27
|
Early phonology revealed by international adoptees' birth language retention. Proc Natl Acad Sci U S A 2017; 114:7307-7312. [PMID: 28652342 DOI: 10.1073/pnas.1706405114] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Until at least 6 mo of age, infants show good discrimination for familiar phonetic contrasts (i.e., those heard in the environmental language) and contrasts that are unfamiliar. Adult-like discrimination (significantly worse for nonnative than for native contrasts) appears only later, by 9-10 mo. This has been interpreted as indicating that infants have no knowledge of phonology until vocabulary development begins, after 6 mo of age. Recently, however, word recognition has been observed before age 6 mo, apparently decoupling the vocabulary and phonology acquisition processes. Here we show that phonological acquisition is also in progress before 6 mo of age. The evidence comes from retention of birth-language knowledge in international adoptees. In the largest ever such study, we recruited 29 adult Dutch speakers who had been adopted from Korea when young and had no conscious knowledge of Korean language at all. Half were adopted at age 3-5 mo (before native-specific discrimination develops) and half at 17 mo or older (after word learning has begun). In a short intensive training program, we observe that adoptees (compared with 29 matched controls) more rapidly learn tripartite Korean consonant distinctions without counterparts in their later-acquired Dutch, suggesting that the adoptees retained phonological knowledge about the Korean distinction. The advantage is equivalent for the younger-adopted and the older-adopted groups, and both groups not only acquire the tripartite distinction for the trained consonants but also generalize it to untrained consonants. Although infants younger than 6 mo can still discriminate unfamiliar phonetic distinctions, this finding indicates that native-language phonological knowledge is nonetheless being acquired at that age.
Collapse
|
28
|
Masapollo M, Polka L, Ménard L. A universal bias in adult vowel perception - By ear or by eye. Cognition 2017; 166:358-370. [PMID: 28601721 DOI: 10.1016/j.cognition.2017.06.001] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2016] [Revised: 05/03/2017] [Accepted: 06/01/2017] [Indexed: 12/29/2022]
Abstract
Speech perceivers are universally biased toward "focal" vowels (i.e., vowels whose adjacent formants are close in frequency, which concentrates acoustic energy into a narrower spectral region). This bias is demonstrated in phonetic discrimination tasks as a directional asymmetry: a change from a relatively less to a relatively more focal vowel results in significantly better performance than a change in the reverse direction. We investigated whether the critical information for this directional effect is limited to the auditory modality, or whether visible articulatory information provided by the speaker's face also plays a role. Unimodal auditory and visual as well as bimodal (auditory-visual) vowel stimuli were created from video recordings of a speaker producing variants of /u/, differing in both their degree of focalization and visible lip rounding (i.e., lip compression and protrusion). In Experiment 1, we confirmed that subjects showed an asymmetry while discriminating the auditory vowel stimuli. We then found, in Experiment 2, a similar asymmetry when subjects lip-read those same vowels. In Experiment 3, we found asymmetries, comparable to those found for unimodal vowels, for bimodal vowels when the audio and visual channels were phonetically-congruent. In contrast, when the audio and visual channels were phonetically-incongruent (as in the "McGurk effect"), this asymmetry was disrupted. These findings collectively suggest that the perceptual processes underlying the "focal" vowel bias are sensitive to articulatory information available across sensory modalities, and raise foundational issues concerning the extent to which vowel perception derives from general-auditory or speech-gesture-specific processes.
Collapse
Affiliation(s)
- Matthew Masapollo
- School of Communication Sciences and Disorders, McGill University, 2001 McGill College, 8th Floor, Montreal, QC H3A 1G1, Canada; Centre for Research on Brain, Language, and Music, McGill University, 3640 de la Montagne, Montreal, Quebec H3G 2A8, Canada.
| | - Linda Polka
- School of Communication Sciences and Disorders, McGill University, 2001 McGill College, 8th Floor, Montreal, QC H3A 1G1, Canada; Centre for Research on Brain, Language, and Music, McGill University, 3640 de la Montagne, Montreal, Quebec H3G 2A8, Canada
| | - Lucie Ménard
- Département de Linguistique, Université du Québec à Montréal, Pavillon J.-A. De sève, DS-4425, 320, Sainte-Catherine Est, Montréal, QC H2X 1L7, Canada; Centre for Research on Brain, Language, and Music, McGill University, 3640 de la Montagne, Montreal, Quebec H3G 2A8, Canada
| |
Collapse
|
29
|
Choi J, Cutler A, Broersma M. Early development of abstract language knowledge: evidence from perception-production transfer of birth-language memory. ROYAL SOCIETY OPEN SCIENCE 2017; 4:160660. [PMID: 28280567 PMCID: PMC5319333 DOI: 10.1098/rsos.160660] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/01/2016] [Accepted: 12/13/2016] [Indexed: 05/21/2023]
Abstract
Children adopted early in life into another linguistic community typically forget their birth language but retain, unaware, relevant linguistic knowledge that may facilitate (re)learning of birth-language patterns. Understanding the nature of this knowledge can shed light on how language is acquired. Here, international adoptees from Korea with Dutch as their current language, and matched Dutch-native controls, provided speech production data on a Korean consonantal distinction unlike any Dutch distinctions, at the outset and end of an intensive perceptual training. The productions, elicited in a repetition task, were identified and rated by Korean listeners. Adoptees' production scores improved significantly more across the training period than control participants' scores, and, for adoptees only, relative production success correlated significantly with the rate of learning in perception (which had, as predicted, also surpassed that of the controls). Of the adoptee group, half had been adopted at 17 months or older (when talking would have begun), while half had been prelinguistic (under six months). The former group, with production experience, showed no advantage over the group without. Thus the adoptees' retained knowledge of Korean transferred from perception to production and appears to be abstract in nature rather than dependent on the amount of experience.
Collapse
Affiliation(s)
- Jiyoun Choi
- Hanyang Phonetics and Psycholinguistics Lab, Hanyang University, Seoul, South Korea
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- ARC Centre of Excellence for the Dynamics of Language, Australia
| | - Anne Cutler
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- ARC Centre of Excellence for the Dynamics of Language, Australia
- The MARCS Institute, Western Sydney University, New South Wales, Australia
| | - Mirjam Broersma
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Centre for Language Studies, Radboud University, Nijmegen, The Netherlands
| |
Collapse
|