1
|
Lester NA, Katsika A. The Syntactic Pasts of Nouns Shape Their Prosodic Future: Lexico-Syntactic Effects on Position and Duration. LANGUAGE AND SPEECH 2024; 67:639-675. [PMID: 37403367 DOI: 10.1177/00238309231177884] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/06/2023]
Abstract
Phrasal prosody is often viewed as a level of linguistic representation at which the phonetic profile of an utterance varies independently of the lexical items it contains. For example, the same word, when produced at the edges of prosodic phrases, will take longer to produce than when it is produced within the edges of a phrase. Lengthening effects have also been found for words when placed in different syntactic or lexical contexts. Recent evidence suggests that lexico-syntactic information-for example, the global syntactic distributions of words-affects phonetic duration in production, irrespective of other factors. The present study asks whether these lexico-syntactic effects on duration interact with prosodic position within the phrase. Specifically, we ask whether (a) the lexico-syntactic information of a word determines its prosodic position, and (b) whether, beyond any categorical effects on positioning, lexico-syntactic factors affect duration within prosodic positions. We address these questions using the Santa Barbara Corpus of Spoken American English. We operationalize syntactic information as the diversity and the typicality of the syntactic distributions of nouns based on a dependency parse of the British National Corpus. We find that earlier positions in the prosodic phrase generally prefer words with higher syntactic diversity. In addition, diversity and typicality modulate duration more reliably in nonfinal positions. Together, our results point to an early influence of lexico-syntactic considerations on prosodic planning.
Collapse
Affiliation(s)
| | - Argyro Katsika
- Department of Linguistics, University of California, Santa Barbara, USA
| |
Collapse
|
2
|
Türk O, Calhoun S. Phrasal Synchronization of Gesture With Prosody and Information Structure. LANGUAGE AND SPEECH 2024; 67:702-743. [PMID: 37522627 DOI: 10.1177/00238309231185308] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/01/2023]
Abstract
This study investigates the synchronization of manual gestures with prosody and information structure using Turkish natural speech data. Prosody has long been linked to gesture as a key driver of gesture-speech synchronization. Gesture has a hierarchical phrasal structure similar to prosody. At the lowest level, gesture has been shown to be synchronized with prosody (e.g., apexes and pitch accents). However, less is known about higher levels. Even less is known about timing relationships with information structure, though this is signaled by prosody and linked to gesture. The present study analyzed phrase synchronization in 3 hr of narrations in Turkish annotated for gesture, prosody, and information structure-topics and foci. The analysis of 515 gesture phrases showed that there was no one-to-one synchronization with intermediate phrases, but their onsets and offsets were synchronized. Moreover, information structural units, topics, and foci were closely synchronized with gesture phrase medial stroke + post-hold combinations (i.e., apical areas). In addition, iconic and metaphoric gestures were more likely to be paired with foci, and deictics with topics. Overall, the results confirm synchronization of gesture and prosody at the phrasal level and provide evidence that gesture shows a direct sensitivity to information structure. These show that speech and gesture production are more connected than assumed in existing production models.
Collapse
Affiliation(s)
| | - Sasha Calhoun
- Te Herenga Waka - Victoria University of Wellington, New Zealand
| |
Collapse
|
3
|
Degano G, Donhauser PW, Gwilliams L, Merlo P, Golestani N. Speech prosody enhances the neural processing of syntax. Commun Biol 2024; 7:748. [PMID: 38902370 PMCID: PMC11190187 DOI: 10.1038/s42003-024-06444-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Accepted: 06/12/2024] [Indexed: 06/22/2024] Open
Abstract
Human language relies on the correct processing of syntactic information, as it is essential for successful communication between speakers. As an abstract level of language, syntax has often been studied separately from the physical form of the speech signal, thus often masking the interactions that can promote better syntactic processing in the human brain. However, behavioral and neural evidence from adults suggests the idea that prosody and syntax interact, and studies in infants support the notion that prosody assists language learning. Here we analyze a MEG dataset to investigate how acoustic cues, specifically prosody, interact with syntactic representations in the brains of native English speakers. More specifically, to examine whether prosody enhances the cortical encoding of syntactic representations, we decode syntactic phrase boundaries directly from brain activity, and evaluate possible modulations of this decoding by the prosodic boundaries. Our findings demonstrate that the presence of prosodic boundaries improves the neural representation of phrase boundaries, indicating the facilitative role of prosodic cues in processing abstract linguistic features. This work has implications for interactive models of how the brain processes different linguistic features. Future research is needed to establish the neural underpinnings of prosody-syntax interactions in languages with different typological characteristics.
Collapse
Affiliation(s)
- Giulio Degano
- Department of Psychology, Faculty of Psychology and Educational Sciences, University of Geneva, Geneva, Switzerland.
| | - Peter W Donhauser
- Ernst Strüngmann Institute for Neuroscience in Cooperation with Max Planck Society, Frankfurt am Main, Germany
| | - Laura Gwilliams
- Department of Psychology, Stanford University, Stanford, CA, USA
| | - Paola Merlo
- Department of Linguistics, University of Geneva, Geneva, Switzerland
- University Centre for Informatics, University of Geneva, Geneva, Switzerland
| | - Narly Golestani
- Department of Psychology, Faculty of Psychology and Educational Sciences, University of Geneva, Geneva, Switzerland
- Brain and Language Lab, Cognitive Science Hub, University of Vienna, Vienna, Austria
- Department of Behavioral and Cognitive Biology, Faculty of Life Sciences, University of Vienna, Vienna, Austria
| |
Collapse
|
4
|
Lo CW, Meyer L. Chunk boundaries disrupt dependency processing in an AG: Reconciling incremental processing and discrete sampling. PLoS One 2024; 19:e0305333. [PMID: 38889141 PMCID: PMC11185458 DOI: 10.1371/journal.pone.0305333] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2024] [Accepted: 05/29/2024] [Indexed: 06/20/2024] Open
Abstract
Language is rooted in our ability to compose: We link words together, fusing their meanings. Links are not limited to neighboring words but often span intervening words. The ability to process these non-adjacent dependencies (NADs) conflicts with the brain's sampling of speech: We consume speech in chunks that are limited in time, containing only a limited number of words. It is unknown how we link words together that belong to separate chunks. Here, we report that we cannot-at least not so well. In our electroencephalography (EEG) study, 37 human listeners learned chunks and dependencies from an artificial grammar (AG) composed of syllables. Multi-syllable chunks to be learned were equal-sized, allowing us to employ a frequency-tagging approach. On top of chunks, syllable streams contained NADs that were either confined to a single chunk or crossed a chunk boundary. Frequency analyses of the EEG revealed a spectral peak at the chunk rate, showing that participants learned the chunks. NADs that cross boundaries were associated with smaller electrophysiological responses than within-chunk NADs. This shows that NADs are processed readily when they are confined to the same chunk, but not as well when crossing a chunk boundary. Our findings help to reconcile the classical notion that language is processed incrementally with recent evidence for discrete perceptual sampling of speech. This has implications for language acquisition and processing as well as for the general view of syntax in human language.
Collapse
Affiliation(s)
- Chia-Wen Lo
- Research Group Language Cycles, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Lars Meyer
- Research Group Language Cycles, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
- University Clinic Münster, Münster, Germany
| |
Collapse
|
5
|
Buss E, Richter ME, Sweeney VN, Davis AG, Dillon MT, Park LR. Effect of Age and Unaided Acoustic Hearing on Pediatric Cochlear Implant Users' Ability to Distinguish Yes/No Statements and Questions. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:1932-1944. [PMID: 38748909 DOI: 10.1044/2024_jslhr-23-00631] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2024]
Abstract
PURPOSE The purpose of this study was to evaluate the ability to discriminate yes/no questions from statements in three groups of children: bilateral cochlear implant (CI) users, nontraditional CI users with aidable hearing preoperatively in the ear to be implanted, and controls with normal hearing. Half of the nontraditional CI users had sufficient postoperative acoustic hearing in the implanted ear to use electric-acoustic stimulation, and half used a CI alone. METHOD Participants heard recorded sentences that were produced either as yes/no questions or as statements by three male and three female talkers. Three raters scored each participant response as either a question or a statement. Bilateral CI users (n = 40, 4-12 years old) and normal-hearing controls (n = 10, 4-12 years old) were tested binaurally in the free field. Nontraditional CI recipients (n = 22, 6-17 years old) were tested with direct audio input to the study ear. RESULTS For the bilateral CI users, performance was predicted by age but not by 125-Hz acoustic thresholds; just under half (n = 17) of the participants in this group had measurable 125-Hz thresholds in their better ear. For nontraditional CI recipients, better performance was predicted by lower 125-Hz acoustic thresholds in the test ear, and there was no association with participant age. Performance approached that of the normal-hearing controls for some participants in each group. CONCLUSIONS Results suggest that a 125-Hz acoustic hearing supports discrimination of yes/no questions and statements in pediatric CI users. Bilateral CI users with little or no acoustic hearing at 125 Hz develop the ability to perform this task, but that ability emerges later than for children with better acoustic hearing. These results underscore the importance of preserving acoustic hearing for pediatric CI users when possible.
Collapse
Affiliation(s)
- Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| | - Margaret E Richter
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| | - Victoria N Sweeney
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
- Center for Hearing Research, Boys Town National Research Hospitals, Omaha, NE
| | - Amanda G Davis
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| | - Margaret T Dillon
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| | - Lisa R Park
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill
| |
Collapse
|
6
|
Filipe MG, Severino C, Vigário M, Frota S. Development and validation of a parental report of toddlers' prosodic skills. CLINICAL LINGUISTICS & PHONETICS 2024; 38:509-528. [PMID: 37348063 DOI: 10.1080/02699206.2023.2226302] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/24/2022] [Accepted: 06/12/2023] [Indexed: 06/24/2023]
Abstract
This study describes the development and validation of Proso-Quest, a parental report of toddlers' prosodic skills that aims to assess early prosodic development in European Portuguese. The development and validation of Proso-Quest proceeded in three phases. Phase 1 was undertaken (a) to establish the structure of the parental report and select the items considering previous work, (b) to collect input from experts on prosodic development, and (c) to revise the report after a pilot study. Phase 2 examined internal consistency, reliability, test-retest reliability, and correlations between Proso-Quest and a valid measure of vocabulary development. Finally, Phase 3 evaluated the discriminant validity of this report in a clinical sample that frequently presents prosodic impairments. The psychometric properties of Proso-Quest indicated an excellent internal consistency, high test-retest reliability, significant correlations with a valid measure of vocabulary development, and sensitivity to identify prosodic delays. This parental report showed evidence of reliability and validity in describing early prosodic development and impairment, and it may be a useful tool in research and educational assessments, as well as in clinical-based assessments.
Collapse
Affiliation(s)
- Marisa G Filipe
- Center of Linguistics, University of Lisbon, Lisbon, Portugal
| | - Cátia Severino
- Center of Linguistics, University of Lisbon, Lisbon, Portugal
| | - Marina Vigário
- Center of Linguistics, University of Lisbon, Lisbon, Portugal
| | - Sónia Frota
- Center of Linguistics, University of Lisbon, Lisbon, Portugal
| |
Collapse
|
7
|
Ertürk A, Gürses E, Kulak Kayıkcı ME. Sex related differences in the perception and production of emotional prosody in adults. PSYCHOLOGICAL RESEARCH 2024; 88:449-457. [PMID: 37542581 DOI: 10.1007/s00426-023-01865-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2022] [Accepted: 07/30/2023] [Indexed: 08/07/2023]
Abstract
This study aimed to investigate the features of sex-related emotional prosody production patterns and perception abilities in adult speakers. The study involved 42 native Turkish speakers (27 females and 15 males). Sex-related perception and production of the emotions "anger," "joy," "sadness," and "neutral" were examined. Participants were first asked to indicate the actor's emotional state by selecting one of the given emotion alternatives provided. They were then instructed to produce the same stimuli with varying emotions. We analyzed the change in voice characteristics employed in different emotions in terms of F0 (Hz), speaking rate (seconds), and intensity (dB) using pairwise emotion comparison. The findings showed no sex differences in emotional prosody perceptions (p = 0.725). However, differences in the production of emotional prosody between sex have been documented in pitch variation of speech. Within-group analyses revealed that women tended to use a higher pitch when expressing joy versus sadness and a neutral state of feeling. Both men and women exhibited varying loudness levels for different emotional states in the speech loudness analysis. When expressing sadness, both men and women speak slower than when expressing as contrasted to anger, joy, or neutral states of feeling. Although Turkish speakers' ability to perceive emotional prosody is similar to that of other languages, they favor speech loudness fluctuation in the production of emotional prosody.
Collapse
Affiliation(s)
- Ayşe Ertürk
- Department of Audiology, Hacettepe University, 06100, Sıhhiye, Ankara, Turkey.
| | - Emre Gürses
- Department of Audiology, Hacettepe University, 06100, Sıhhiye, Ankara, Turkey
| | | |
Collapse
|
8
|
Petitpain A, Brabant-Thery J, Patin C, Ott L, Basirat A. Status of prosody in the practices of speech-language pathologists in France: A survey on theoretical and clinical dimensions. JOURNAL OF COMMUNICATION DISORDERS 2024; 107:106404. [PMID: 38266444 DOI: 10.1016/j.jcomdis.2023.106404] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/03/2023] [Revised: 11/28/2023] [Accepted: 12/22/2023] [Indexed: 01/26/2024]
Abstract
PURPOSE Many populations in the scope of speech and language pathologists (SLPs) are at risk of prosodic disorders that can impact communicative ability. While a few studies have examined the importance of prosody in the clinical practice of English-speaking SLPs, there is a lack of similar research in non-English-speaking countries. The current study aimed to establish an overview of the situation in France and to determine factors that might be associated with assessing and treating prosody. METHOD We collected data about the theoretical and clinical dimensions related to prosody from SLPs in France using an online survey. RESULTS One hundred and twenty-eight valid responses were analyzed. Results suggest that both emotional and linguistic prosodies are examined by SLPs. Some aspects of prosody seem to be mastered better than others. While they are aware of the major prosodic difficulties in various pathologies and their negative impact, most SLPs in France rarely or never assess prosody. They feel more uncomfortable with it than with other speech-language domains and often do not consider it as a therapeutic target. Importantly, the perception of prosody as a therapeutic tool is associated with the frequency of assessing and addressing prosodic impairment. CONCLUSIONS Further studies should focus on identifying the factors that could improve practices and developing comprehensive assessment tools and evidence-based intervention methods in French. Providing better training in prosody and setting up more collaborations between SLPs and researchers in linguistics and psycholinguistics would be keys in developing this field.
Collapse
Affiliation(s)
- Aline Petitpain
- Department of Speech and Language Therapy, Faculty of Medicine, UFR3S, F-59000 Lille, France
| | - Juliette Brabant-Thery
- Department of Speech and Language Therapy, Faculty of Medicine, UFR3S, F-59000 Lille, France; Department of Neurology, Roger Salengro Hospital, CHU Lille, F-59000 Lille, France
| | - Cédric Patin
- Univ. Lille, CNRS, UMR 8163 - STL -Savoirs, Textes, Langage, F-59000 Lille, France
| | - Laurent Ott
- Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France
| | - Anahita Basirat
- Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France.
| |
Collapse
|
9
|
Zora H, Wester J, Csépe V. Predictions about prosody facilitate lexical access: Evidence from P50/N100 and MMN components. Int J Psychophysiol 2023; 194:112262. [PMID: 37924955 DOI: 10.1016/j.ijpsycho.2023.112262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2023] [Revised: 08/23/2023] [Accepted: 10/28/2023] [Indexed: 11/06/2023]
Abstract
Research into the neural foundation of perception asserts a model where top-down predictions modulate the bottom-up processing of sensory input. Despite becoming increasingly influential in cognitive neuroscience, the precise account of this predictive coding framework remains debated. In this study, we aim to contribute to this debate by investigating how predictions about prosody facilitate speech perception, and to shed light especially on lexical access influenced by simultaneous predictions in different domains, inter alia, prosodic and semantic. Using a passive auditory oddball paradigm, we examined neural responses to prosodic changes, leading to a semantic change as in Dutch nouns canon ['kaːnɔn] 'canon' vs kanon [kaː'nɔn] 'cannon', and used acoustically identical pseudowords as controls. Results from twenty-eight native speakers of Dutch (age range 18-32 years) indicated an enhanced P50/N100 complex to prosodic change in pseudowords as well as an MMN response to both words and pseudowords. The enhanced P50/N100 response to pseudowords is claimed to indicate that all relevant auditory information is still processed by the brain, whereas the reduced response to words might reflect the suppression of information that has already been encoded. The MMN response to pseudowords and words, on the other hand, is best justified by the unification of previously established prosodic representations with sensory and semantic input respectively. This pattern of results is in line with the predictive coding framework acting on multiple levels and is of crucial importance to indicate that predictions about linguistic prosodic information are utilized by the brain as early as 50 ms.
Collapse
Affiliation(s)
- Hatice Zora
- Max Planck Institute for Psycholinguistics, P.O. Box 310 6500, AH, Nijmegen, the Netherlands.
| | - Janniek Wester
- Max Planck Institute for Psycholinguistics, P.O. Box 310 6500, AH, Nijmegen, the Netherlands
| | - Valéria Csépe
- HUN-REN Research Centre of Natural Sciences, Brain Imaging Centre, P.O. Box 286 1519, Budapest, Hungary
| |
Collapse
|
10
|
Zhang Y, Ding R, Frassinelli D, Tuomainen J, Klavinskis-Whiting S, Vigliocco G. The role of multimodal cues in second language comprehension. Sci Rep 2023; 13:20824. [PMID: 38012193 PMCID: PMC10682458 DOI: 10.1038/s41598-023-47643-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Accepted: 11/16/2023] [Indexed: 11/29/2023] Open
Abstract
In face-to-face communication, multimodal cues such as prosody, gestures, and mouth movements can play a crucial role in language processing. While several studies have addressed how these cues contribute to native (L1) language processing, their impact on non-native (L2) comprehension is largely unknown. Comprehension of naturalistic language by L2 comprehenders may be supported by the presence of (at least some) multimodal cues, as these provide correlated and convergent information that may aid linguistic processing. However, it is also the case that multimodal cues may be less used by L2 comprehenders because linguistic processing is more demanding than for L1 comprehenders, leaving more limited resources for the processing of multimodal cues. In this study, we investigated how L2 comprehenders use multimodal cues in naturalistic stimuli (while participants watched videos of a speaker), as measured by electrophysiological responses (N400) to words, and whether there are differences between L1 and L2 comprehenders. We found that prosody, gestures, and informative mouth movements each reduced the N400 in L2, indexing easier comprehension. Nevertheless, L2 participants showed weaker effects for each cue compared to L1 comprehenders, with the exception of meaningful gestures and informative mouth movements. These results show that L2 comprehenders focus on specific multimodal cues - meaningful gestures that support meaningful interpretation and mouth movements that enhance the acoustic signal - while using multimodal cues to a lesser extent than L1 comprehenders overall.
Collapse
Affiliation(s)
- Ye Zhang
- Experimental Psychology, University College London, London, UK
| | - Rong Ding
- Language and Computation in Neural Systems, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Diego Frassinelli
- Department of Linguistics, University of Konstanz, Konstanz, Germany
| | - Jyrki Tuomainen
- Speech, Hearing and Phonetic Sciences, University College London, London, UK
| | | | | |
Collapse
|
11
|
Anikin A, Canessa-Pollard V, Pisanski K, Massenet M, Reby D. Beyond speech: Exploring diversity in the human voice. iScience 2023; 26:108204. [PMID: 37908309 PMCID: PMC10613903 DOI: 10.1016/j.isci.2023.108204] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Revised: 07/20/2023] [Accepted: 10/11/2023] [Indexed: 11/02/2023] Open
Abstract
Humans have evolved voluntary control over vocal production for speaking and singing, while preserving the phylogenetically older system of spontaneous nonverbal vocalizations such as laughs and screams. To test for systematic acoustic differences between these vocal domains, we analyzed a broad, cross-cultural corpus representing over 2 h of speech, singing, and nonverbal vocalizations. We show that, while speech is relatively low-pitched and tonal with mostly regular phonation, singing and especially nonverbal vocalizations vary enormously in pitch and often display harsh-sounding, irregular phonation owing to nonlinear phenomena. The evolution of complex supralaryngeal articulatory spectro-temporal modulation has been critical for speech, yet has not significantly constrained laryngeal source modulation. In contrast, articulation is very limited in nonverbal vocalizations, which predominantly contain minimally articulated open vowels and rapid temporal modulation in the roughness range. We infer that vocal source modulation works best for conveying affect, while vocal filter modulation mainly facilitates semantic communication.
Collapse
Affiliation(s)
- Andrey Anikin
- Division of Cognitive Science, Lund University, Lund, Sweden
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
| | - Valentina Canessa-Pollard
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
- Psychology, Institute of Psychology, Business and Human Sciences, University of Chichester, Chichester, West Sussex PO19 6PE, UK
| | - Katarzyna Pisanski
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
- CNRS French National Centre for Scientific Research, DDL Dynamics of Language Lab, University of Lyon 2, 69007 Lyon, France
- Institute of Psychology, University of Wrocław, Dawida 1, 50-527 Wrocław, Poland
| | - Mathilde Massenet
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
| | - David Reby
- ENES Bioacoustics Research Lab, CRNL, University of Saint-Etienne, CNRS, Inserm, 23 rue Michelon, 42023 Saint-Etienne, France
| |
Collapse
|
12
|
Yu L, Huang D, Wang S, Zhang Y. Reduced Neural Specialization for Word-level Linguistic Prosody in Children with Autism. J Autism Dev Disord 2023; 53:4351-4367. [PMID: 36038793 DOI: 10.1007/s10803-022-05720-x] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 08/10/2022] [Indexed: 10/15/2022]
Abstract
Children with autism often show atypical brain lateralization for speech and language processing, however, it is unclear what linguistic component contributes to this phenomenon. Here we measured event-related potential (ERP) responses in 21 school-age autistic children and 25 age-matched neurotypical (NT) peers during listening to word-level prosodic stimuli. We found that both groups displayed larger late negative response (LNR) amplitude to native prosody than to nonnative prosody; however, unlike the NT group exhibiting left-lateralized LNR distinction of prosodic phonology, the autism group showed no evidence of LNR lateralization. Moreover, in both groups, the LNR effects were only present for prosodic phonology but not for phoneme-free prosodic acoustics. These results extended the findings of inadequate neural specialization for language in autism to sub-lexical prosodic structures.
Collapse
Affiliation(s)
- Luodi Yu
- Center for Autism Research, School of Education, Guangzhou University, Wenyi Bldg, Guangzhou, China.
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University) , Ministry of Education, Guangzhou, China.
| | - Dan Huang
- Guangzhou Rehabilitation & Research Center for Children with ASD, Guangzhou Cana School, Guangzhou, China
| | - Suiping Wang
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University) , Ministry of Education, Guangzhou, China.
| | - Yang Zhang
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, MN, USA
| |
Collapse
|
13
|
Lo CW, Anderson M, Henke L, Meyer L. Periodic fluctuations in reading times reflect multi-word-chunking. Sci Rep 2023; 13:18522. [PMID: 37898645 PMCID: PMC10613263 DOI: 10.1038/s41598-023-45536-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 10/20/2023] [Indexed: 10/30/2023] Open
Abstract
Memory is fleeting. To avoid information loss, humans need to recode verbal stimuli into chunks of limited duration, each containing multiple words. Chunk duration may also be limited neurally by the wavelength of periodic brain activity, so-called neural oscillations. While both cognitive and neural constraints predict some degree of behavioral regularity in processing, this remains to be shown. Our analysis of self-paced reading data from 181 participants reveals periodic patterns at a frequency of [Formula: see text] 2 Hz. We defined multi-word chunks by using a computational formalization based on dependency annotations and part-of-speech tags. Potential chunk outputs were first generated from the computational formalization and the final chunk outputs were selected based on normalized pointwise mutual information. We show that behavioral periodicity is time-aligned to multi-word chunks, suggesting that the multi-word chunks generated from local dependency clusters may minimize memory demands. This is the first evidence that sentence processing behavior is periodic, consistent with a role of both memory constraints and endogenous electrophysiological rhythms in the formation of chunks during language comprehension.
Collapse
Affiliation(s)
- Chia-Wen Lo
- Research Group Language Cycles, Max Planck Institute for Human Cognitive and Brain Sciences, 04013, Leipzig, Germany.
| | | | - Lena Henke
- Research Group Language Cycles, Max Planck Institute for Human Cognitive and Brain Sciences, 04013, Leipzig, Germany
| | - Lars Meyer
- Research Group Language Cycles, Max Planck Institute for Human Cognitive and Brain Sciences, 04013, Leipzig, Germany
- Clinic for Phoniatrics and Pedaudiology, University Clinic Münster, 48149, Münster, Germany
| |
Collapse
|
14
|
Gilbert AC, Lee JG, Wolpert M, Baum SR. Phrase parsing in a second language as indexed by the closure positive shift: The impact of language experience and acoustic cue salience. Eur J Neurosci 2023; 58:3838-3858. [PMID: 37667595 DOI: 10.1111/ejn.16134] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2023] [Revised: 07/22/2023] [Accepted: 08/12/2023] [Indexed: 09/06/2023]
Abstract
Despite the importance of prosodic processing in utterance parsing, a majority of studies investigating boundary localization in a second language focus on word segmentation. The goal of the present study was to investigate the parsing of phrase boundaries in first and second languages from different prosodic typologies (stress-timed vs. syllable-timed). Fifty English-French bilingual adults who varied in native language (French or English) and second language proficiency listened to English and French utterances with different prosodic structures while event-related brain potentials were recorded. The utterances were built around target words presented either in phrase-final position (bearing phrase-final lengthening) or in penultimate position. Each participant listened to both English and French stimuli, providing data in their native language (used as reference) and their second language. Target words in phrase-final position elicited closure positive shifts across listeners in both languages, regardless of the language-specific acoustic cues associated with phrase-final lengthening (shorter phrase-final lengthening in English compared to French). Interestingly, directional effects were observed, where learning to parse English as a second language in a native-like manner seemed to require a higher proficiency level than learning to parse French as a second language. This pattern of results supports the idea that L2 listeners need to learn to recognize L2-specific phrase-final lengthening regardless of the apparent similarity across languages and that some language combinations might present greater challenges than others.
Collapse
Affiliation(s)
- Annie C Gilbert
- School of Communication Sciences and Disorders, McGill University, Montréal, Canada
- Centre for Research on Brain, Language, and Music, Montréal, Canada
| | - Jasmine G Lee
- Centre for Research on Brain, Language, and Music, Montréal, Canada
- Integrated Program in Neuroscience, McGill University, Montréal, Canada
| | - Max Wolpert
- Centre for Research on Brain, Language, and Music, Montréal, Canada
- Integrated Program in Neuroscience, McGill University, Montréal, Canada
| | - Shari R Baum
- School of Communication Sciences and Disorders, McGill University, Montréal, Canada
- Centre for Research on Brain, Language, and Music, Montréal, Canada
| |
Collapse
|
15
|
Heller D, Brown-Schmidt S. The Multiple Perspectives Theory of Mental States in Communication. Cogn Sci 2023; 47:e13322. [PMID: 37483115 DOI: 10.1111/cogs.13322] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2021] [Revised: 06/11/2023] [Accepted: 07/02/2023] [Indexed: 07/25/2023]
Abstract
Inspired by early proposals in philosophy, dominant accounts of language posit a central role for mutual knowledge, either encoded directly in common ground, or approximated through other cognitive mechanisms. Using existing empirical evidence from language and memory, we challenge this tradition, arguing that mutual knowledge captures only a subset of the mental states needed to support communication. In a novel theoretical proposal, we argue for a cognitive architecture that includes separate, distinct representations of the self and other, and a cognitive process that compares these representations continuously during conversation, outputting both similarities and differences in perspective. Our theory accounts for existing data, interfaces with findings from other cognitive domains, and makes novel predictions about the role of perspective in language use. We term this new account the Multiple Perspectives Theory of mental states in communication.
Collapse
|
16
|
Martínez-Nicolás I, Martínez-Sánchez F, Ivanova O, Meilán JJG. Reading and lexical-semantic retrieval tasks outperforms single task speech analysis in the screening of mild cognitive impairment and Alzheimer's disease. Sci Rep 2023; 13:9728. [PMID: 37322073 PMCID: PMC10272227 DOI: 10.1038/s41598-023-36804-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Accepted: 06/12/2023] [Indexed: 06/17/2023] Open
Abstract
Age-related cognitive impairment have increased dramatically in recent years, which has risen the interes in developing screening tools for mild cognitive impairment and Alzheimer's disease. Speech analysis allows to exploit the behavioral consequences of cognitive deficits on the patient's vocal performance so that it is possible to identify pathologies affecting speech production such as dementia. Previous studies have further shown that the speech task used determines how the speech parameters are altered. We aim to combine the impairments in several speech production tasks in order to improve the accuracy of screening through speech analysis. The sample consists of 72 participants divided into three equal groups of healthy older adults, people with mild cognitive impairment, or Alzheimer's disease, matched by age and education. A complete neuropsychological assessment and two voice recordings were performed. The tasks required the participants to read a text, and complete a sentence with semantic information. A stepwise linear discriminant analysis was performed to select speech parameters with discriminative power. The discriminative functions obtained an accuracy of 83.3% in simultaneous classifications of several levels of cognitive impairment. It would therefore be a promising screening tool for dementia.
Collapse
Affiliation(s)
| | | | - Olga Ivanova
- Faculty of Philology, University of Salamanca, 37008, Salamanca, Spain
| | - Juan J G Meilán
- Faculty of Psychology, University of Salamanca, 37008, Salamanca, Spain
- Institute of Neuroscience of Castilla y León, 37007, Salamanca, Spain
| |
Collapse
|
17
|
Hu N, Chen A, Quené H, Sanders TJM. The role of prosody in interpreting causality in English discourse. PLoS One 2023; 18:e0286003. [PMID: 37267347 DOI: 10.1371/journal.pone.0286003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2023] [Accepted: 05/06/2023] [Indexed: 06/04/2023] Open
Abstract
Previous studies have well established that certain causal connectives encode information about the semantic-pragmatic distinction between different types of causal relations such as CAUSE-CONSEQUENCE versus CLAIM-ARGUMENT relations. These "specialized" causal connectives assist listeners in discerning different types of causality. Additionally, research has demonstrated that utterances expressing CLAIM-ARGUMENT relations exhibit distinct prosodic characteristics compared to utterances expressing CAUSE-CONSEQUENCE relations. However, it remains unknown whether the prosodic characteristics of utterances expressing causality can aid listeners in determining the specific type of causality being conveyed. To address this knowledge gap, this study investigates the impact of the prosody, specifically the prosody of the causal connective so in English, on listeners' interpretation of the type of causality expressed. We conducted a perception experiment employing a forced-choice discourse completion task, where the participants were required to select a continuation for each sound clip they heard. The sound clip consisted of factual events followed by the causal connective so. We found that the odds of listeners choosing subjective continuations over objective continuations increased when the connective so at the end of the sound clip was pronounced with subjective causality prosodic features, such as prolonged duration and a concave f0 contour. This finding suggests that the prosody of the connective so plays a role in conveying subjectivity in causality, guiding listeners in interpreting causal relations. In addition, it is important to note that our data revealed individual variation among listeners in their interpretations of prosodic information related to subjective-objective causality contrast.
Collapse
Affiliation(s)
- Na Hu
- Department of Language, Literature and Communication, Institute for Language Sciences, Utrecht University, Utrecht, the Netherlands
| | - Aoju Chen
- Department of Language, Literature and Communication, Institute for Language Sciences, Utrecht University, Utrecht, the Netherlands
| | - Hugo Quené
- Department of Language, Literature and Communication, Institute for Language Sciences, Utrecht University, Utrecht, the Netherlands
| | - Ted J M Sanders
- Department of Language, Literature and Communication, Institute for Language Sciences, Utrecht University, Utrecht, the Netherlands
| |
Collapse
|
18
|
Belz M, Rasskazova O, Krivokapić J, Mooshammer C. Interaction between Phrasal Structure and Vowel Tenseness in German: An Acoustic and Articulatory Study. LANGUAGE AND SPEECH 2023; 66:3-34. [PMID: 35021902 PMCID: PMC9975821 DOI: 10.1177/00238309211064857] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Phrase-final lengthening affects the segments preceding a prosodic boundary. This prosodic variation is generally assumed to be independent of the phonemic identity. We refer to this as the 'uniform lengthening hypothesis' (ULH). However, in German, lax vowels do not undergo lengthening for word stress or shortening for increased speech rate, indicating that temporal properties might interact with phonemic identity. We test the ULH by comparing the effect of the boundary on acoustic and kinematic measures for tense and lax vowels and several coda consonants. We further examine if the boundary effect decreases with distance from the boundary. Ten native speakers of German were recorded by means of electromagnetic articulography (EMA) while reading sentences that contained six minimal pairs varying in vowel tenseness and boundary type. In line with the ULH, the results show that the acoustic durations of lax vowels are lengthened phrase-finally, similarly to tense vowels. We find that acoustic lengthening is stronger the closer the segments are to the boundary. Articulatory parameters of the closing movements toward the post-vocalic consonants are affected by both phrasal position and identity of the preceding vowel. The results are discussed with regard to the interaction between prosodic structure and vowel tenseness.
Collapse
Affiliation(s)
- Malte Belz
- Malte Belz, Institut für deutsche Sprache und Linguistik, Humboldt-Universität zu Berlin, Unter den Linden 6, Berlin 10099, Germany.
| | | | | | | |
Collapse
|
19
|
de Beer C, Wartenburger I, Huttenlauch C, Hanne S. A systematic review on production and comprehension of linguistic prosody in people with acquired language and communication disorders resulting from unilateral brain lesions. JOURNAL OF COMMUNICATION DISORDERS 2023; 101:106298. [PMID: 36623377 DOI: 10.1016/j.jcomdis.2022.106298] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/16/2021] [Revised: 11/14/2022] [Accepted: 12/25/2022] [Indexed: 06/17/2023]
Abstract
BACKGROUND Prosody serves central functions in language processing including linguistic functions (linguistic prosody), like structuring the speech signal. Impairments in production and comprehension of linguistic prosody have been described for persons with unilateral right (RHDP) or left hemisphere damage (LHDP). However, reported results differ with respect to the characteristics and severities of these impairments AIMS: We conducted a systematic literature review focusing on production and comprehension of linguistic prosody at the prosody-syntax interface (i.e., phrase or sentence level) in LHDP and RHDP. METHODS & PROCEDURES In a systematic literature search we included: (i) empirical studies with (ii) adult RHDP and/or LHDP (iii) investigating production and/or comprehension of linguistic prosody at the (iv) phrase or sentence level (v) reporting quantitative data on prosodic measures. We excluded overview papers; studies involving participants with dysarthria, apraxia of speech, foreign accent syndrome, psychiatric diseases, and/or neurodegenerative diseases; studies focusing primarily on emotional prosody; and on lexical stress / word level; studies of which no full text was available and/or that were published in a language other than English. We searched the databases BIOSIS, MEDLINE, EMBASE, PubMed, Web of Science, CINAHL, Cochrane Library, PSYNDEX, PsycINFO and speechBITE, last searched on January 13th 2022.We found 2,631 studies without duplicates. We identified 43 studies which were included into our systematic review. For data extraction and synthesis of results, we grouped studies by (i) modality (production vs. comprehension), (ii) function (syntactic structure vs. information structure), and (iii) by experiment task. For production studies, outcome measures were defined as the productive use of the different prosodic cues (lengthening, pause, f0, amplitude). For comprehension studies, performance measures (accuracy and reaction times) were defined as outcome measures. In accordance with the PRISMA 2020 statement (Page et al., 2021), we conducted a quality check to assess study risk of bias. Our review was pre-registered with PROSPERO (CRD42019120308). OUTCOMES & RESULTS Of the 43 studies reviewed, 30 studies involved RHDP (n = 309), assessing production in 15 studies and focusing on comprehension of prosody in 16 studies (one study investigated production and comprehension). LHDP (n = 438) were included in 35 studies of which 15 studied production and 21 evaluated comprehension of prosody (one study investigated production and comprehension). Despite the heterogeneity of results in the studies reviewed, our synthesis of results suggests that both LHDP and RHDP show limitations, but no complete impairment, in their production and/or comprehension of linguistic prosody. Prosodic limitations are evident in different areas of processing linguistic prosody, like syntactic disambiguation or the distinction between sentence types. There is a tendency towards more severe limitations in LHDP as compared to RHDP. CONCLUSIONS We only included published studies into our review and did not perform an assessment of risk of reporting bias as well as systematic certainty assessments of the outcomes. Despite these limitations, we conclude that both groups show deficits in production and comprehension of linguistic prosody, but neither LHDP nor RHDP are completely impaired in their prosodic processing. This suggests that prosody is a relevant communicative resource for LHDP and RHDP worth being addressed in speech-language-therapy.
Collapse
Affiliation(s)
- Carola de Beer
- SFB1287, Cognitive Sciences, Department of Linguistic, University of Potsdam, Germany; Faculty of Linguistics and Literary Studies & Medical School OWL, University of Bielefeld, Germany.
| | - Isabell Wartenburger
- SFB1287, Cognitive Sciences, Department of Linguistic, University of Potsdam, Germany
| | - Clara Huttenlauch
- SFB1287, Cognitive Sciences, Department of Linguistic, University of Potsdam, Germany
| | - Sandra Hanne
- SFB1287, Cognitive Sciences, Department of Linguistic, University of Potsdam, Germany
| |
Collapse
|
20
|
Armstrong M, Breen M, Gooden S, Levon E, Yu KM. Sociolectal and Dialectal Variation in Prosody. LANGUAGE AND SPEECH 2022; 65:783-790. [PMID: 36127821 DOI: 10.1177/00238309221122105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
As in many linguistics subfields, studies of prosody have mainly focused on majority languages and dialects and on speakers who hold power in social structures. The goal of this Special Issue is to diversify prosody research in terms of the languages and dialects being investigated, as well as the social structures that influence prosodic variation. The Special Issue brings together prosody researchers and researchers exploring sociological variation in prosody, with a focus on the prosody of marginalized dialects and on prosodic differences based on gender, sexuality, race, and ethnicity. The papers in this volume don't just advance our understanding of critical issues in sociolinguistics, but they also challenge some of the received wisdom in the exploration of sociolinguistic influences on prosody. Not only does this collection highlight the value of this work to informing theories of prosodic variation and change, but the collected papers also provide examples of methodological innovations in the field that will be valuable for all prosody researchers.
Collapse
Affiliation(s)
- Meghan Armstrong
- Spanish and Portuguese Studies, University of Massachusetts Amherst, USA
| | - Mara Breen
- Department of Psychology and Education, Mount Holyoke College, USA
| | - Shelome Gooden
- Department of Linguistics, University of Pittsburgh, USA
| | - Erez Levon
- Center for the Study of Language and Society, University of Bern, Switzerland
| | - Kristine M Yu
- Department of Linguistics, University of Massachusetts Amherst, USA
| |
Collapse
|
21
|
Young NJ. The Sociolectal and Stylistic Variability of Rhythm in Stockholm. LANGUAGE AND SPEECH 2022; 65:1034-1070. [PMID: 33509037 DOI: 10.1177/0023830920969727] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The question of "staccato" rhythm in Stockholm's multiethnolect is investigated by comparing nPVIV measurements of the speech of 36 adult male speakers. The men, ages 24-43, come from a stratified sample of social classes and racial groups. Three contextual styles were recorded and analyzed: informal, formal, and very formal. The distribution of nPVIV values in informal speech across class and racial group indicates that speech rhythm splits three ways: low-alternation "staccato" rhythm among the racialized lower-class men, high-alternation rhythm among the white lower-class men, and an intermediate level of rhythm among higher-class men, regardless of racialized category. The "staccato" low-alternation feature is also less stylistically sensitive than the high-alternation feature, implying that the latter is a more established feature than the former. Further, the "staccato" feature is more stylistically sensitive among younger speakers than older speakers, implying an ongoing change from indicator to marker status. For all speakers, age has a stable main effect, which means that younger speakers, independent of racial group and class, have lower alternation than older speakers. Implied here is that low-alternation is a change from below that originates within the racialized working class. While it may be incrementally transmitting into the wider speech community, the white working class is the most resistant to its incursion.
Collapse
Affiliation(s)
- Nathan J Young
- Centre for Research on Bilingualism, Stockholm University, Sweden
| |
Collapse
|
22
|
Yuen I, Xu Rattanasone N, Schmidt E, Holt R, Demuth K. The acoustic realization of contrastive focus by 6-year-old Australian English-speaking children. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:3313. [PMID: 36586851 DOI: 10.1121/10.0016387] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/14/2022] [Accepted: 11/17/2022] [Indexed: 06/17/2023]
Abstract
Children as young as five have some ability to produce contrastive focus [Wells et al. (2004) J. Child Lang. 31, 749-778]. However, adult listeners' ability to identify the location of contrastive focus depends on whether the speech came from a 4-, 7-, or 11-year-old [Patel and Brayton (2009) J. Speech. Lang. Hear. Res. 52, 790-801]. Recent acoustic studies have also reported the use of F0 vs duration in contrastive focus productions by American English-speaking 2-year-olds [Thorson and Morgan (2021) J. Child Lang. 48, 541-568] and 4-year-olds [Wonnacott and Watson (2008) Cognition 107, 1093-1101], respectively. This study, therefore, evaluated the extent to which older 6-year-olds, with more language experience, used F0 and/or duration when producing contrastive focus, and compared this to adult speech. Monosyllabic and disyllabic adjective + noun targets (e.g., green ball) in utterance medial and final position were elicited from 20 Australian English-speaking 6-year-olds and 14 adults in adjective focus and noun focus conditions. Although both adults and children used high F0, only adults consistently used word and stressed syllable duration as well. This suggests that children may master the different acoustic cues to contrastive focus at different stages of development, with late cue integration.
Collapse
Affiliation(s)
- Ivan Yuen
- Department of Linguistics, Macquarie University, C5A, Balaclava Road, North Ryde, New South Wales, 2109, Australia
| | - Nan Xu Rattanasone
- Department of Linguistics, Macquarie University, C5A, Balaclava Road, North Ryde, New South Wales, 2109, Australia
| | - Elaine Schmidt
- Cambridge Assessment, University of Cambridge, Cambridge, United Kingdom
| | - Rebecca Holt
- Department of Linguistics, Macquarie University, C5A, Balaclava Road, North Ryde, New South Wales, 2109, Australia
| | - Katherine Demuth
- Department of Linguistics, Macquarie University, C5A, Balaclava Road, North Ryde, New South Wales, 2109, Australia
| |
Collapse
|
23
|
van der Burght CL, Numssen O, Schlaak B, Goucha T, Hartwigsen G. Differential contributions of inferior frontal gyrus subregions to sentence processing guided by intonation. Hum Brain Mapp 2022; 44:585-598. [PMID: 36189774 PMCID: PMC9842926 DOI: 10.1002/hbm.26086] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2022] [Revised: 08/30/2022] [Accepted: 08/31/2022] [Indexed: 01/25/2023] Open
Abstract
Auditory sentence comprehension involves processing content (semantics), grammar (syntax), and intonation (prosody). The left inferior frontal gyrus (IFG) is involved in sentence comprehension guided by these different cues, with neuroimaging studies preferentially locating syntactic and semantic processing in separate IFG subregions. However, this regional specialisation has not been confirmed with a neurostimulation method. Consequently, the causal role of such a specialisation remains unclear. This study probed the role of the posterior IFG (pIFG) for syntactic processing and the anterior IFG (aIFG) for semantic processing with repetitive transcranial magnetic stimulation (rTMS) in a task that required the interpretation of the sentence's prosodic realisation. Healthy participants performed a sentence completion task with syntactic and semantic decisions, while receiving 10 Hz rTMS over either left aIFG, pIFG, or vertex (control). Initial behavioural analyses showed an inhibitory effect on accuracy without task-specificity. However, electric field simulations revealed differential effects for both subregions. In the aIFG, stronger stimulation led to slower semantic processing, with no effect of pIFG stimulation. In contrast, we found a facilitatory effect on syntactic processing in both aIFG and pIFG, where higher stimulation strength was related to faster responses. Our results provide first evidence for the functional relevance of left aIFG in semantic processing guided by intonation. The stimulation effect on syntactic responses emphasises the importance of the IFG for syntax processing, without supporting the hypothesis of a pIFG-specific involvement. Together, the results support the notion of functionally specialised IFG subregions for diverse but fundamental cues for language processing.
Collapse
Affiliation(s)
- Constantijn L. van der Burght
- Department of NeuropsychologyMax Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany,Lise Meitner Research Group Cognition and PlasticityMax Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany,Psychology of Language DepartmentMax Planck Institute for PsycholinguisticsNijmegen
| | - Ole Numssen
- Lise Meitner Research Group Cognition and PlasticityMax Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany
| | - Benito Schlaak
- Lise Meitner Research Group Cognition and PlasticityMax Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany
| | - Tomás Goucha
- Department of NeuropsychologyMax Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany
| | - Gesa Hartwigsen
- Lise Meitner Research Group Cognition and PlasticityMax Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany
| |
Collapse
|
24
|
Wang LC, Xu Z, Liu D, Kwan-Chen LLY, Chung KKH, Cho HY, Chen JK. Age differences in the relation between Chinese students’ prosodic sensitivity and reading comprehension: From nine to fifteen years. COGNITIVE DEVELOPMENT 2022. [DOI: 10.1016/j.cogdev.2022.101234] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
25
|
Baek H. Prosodic Disambiguation in First and Second Language Production: English and Korean. LANGUAGE AND SPEECH 2022; 65:598-624. [PMID: 34605716 DOI: 10.1177/00238309211042041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
This study investigates the use of prosodic cues for syntactic ambiguity resolution by first language (L1) and second language (L2) speakers. In a production experiment, sentences with relative clause attachment ambiguity were elicited in three language conditions: native English speakers' L1 productions as well as Korean-English bilingual speakers' L1 Korean and L2 English productions. The results show that English uses both boundary marking (pause) and relative word prominence (elevated pitch and intensity) for disambiguation, while Korean mainly relies on boundary marking (pre-boundary lengthening and pause). The bilingual speakers have learned to use the English phonological categories such as pitch accents for disambiguation, but their use of phonetic cues to realize these categories still differed from that of native English speakers. In addition, they did not show a significant use of boundary cues. These results are discussed in relation to the typological differences between the prosody of English and of Korean.
Collapse
|
26
|
Glushko A, Poeppel D, Steinhauer K. Overt and implicit prosody contribute to neurophysiological responses previously attributed to grammatical processing. Sci Rep 2022; 12:14759. [PMID: 36042220 PMCID: PMC9427746 DOI: 10.1038/s41598-022-18162-3] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Accepted: 08/05/2022] [Indexed: 11/09/2022] Open
Abstract
Recent neurophysiological research suggests that slow cortical activity tracks hierarchical syntactic structure during online sentence processing. Here we tested an alternative hypothesis: electrophysiological activity peaks at constituent phrase as well as sentence frequencies reflect cortical tracking of overt or covert (implicit) prosodic grouping. Participants listened to series of sentences presented in three conditions while electroencephalography (EEG) was recorded. First, prosodic cues in the sentence materials were neutralized. We found an EEG spectral power peak elicited at a frequency that only 'tagged' covert, implicit prosodic change, but not any major syntactic constituents. In the second condition, participants listened to a series of sentences with overt prosodic grouping cues that either aligned or misaligned with the syntactic phrasing in the sentences (initial overt prosody trials). Following each overt prosody trial, participants were presented with a second series of sentences lacking overt prosodic cues (instructed prosody trial) and were instructed to imagine the prosodic contour present in the previous, overt prosody trial. The EEG responses reflected an interactive relationship between syntactic processing and prosodic tracking at the frequencies of syntactic constituents (sentences and phrases): alignment of syntax and prosody boosted EEG responses, whereas their misalignment had an opposite effect. This was true for both overt and imagined prosody conditions. We conclude that processing of both overt and covert prosody is reflected in the frequency-tagged neural responses at sentence constituent frequencies. These findings need to be incorporated in any account that aims to identify neural markers reflecting syntactic processing.
Collapse
Affiliation(s)
| | - David Poeppel
- Department of Psychology, New York University, New York City, NY, USA
- Ernst Struengmann Institute for Neuroscience, Frankfurt, Germany
- Center for Language, Music, and Emotion (CLaME), New York, USA
| | - Karsten Steinhauer
- Centre for Research on Brain, Language and Music, Montreal, Canada
- School of Communication Sciences and Disorders, McGill University, 2001 McGill College Avenue, Unit 800, Montreal, QC, H3A 1G1, Canada
| |
Collapse
|
27
|
Lau JCY, Patel S, Kang X, Nayar K, Martin GE, Choy J, Wong PCM, Losh M. Cross-linguistic patterns of speech prosodic differences in autism: A machine learning study. PLoS One 2022; 17:e0269637. [PMID: 35675372 PMCID: PMC9176813 DOI: 10.1371/journal.pone.0269637] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Accepted: 05/24/2022] [Indexed: 11/19/2022] Open
Abstract
Differences in speech prosody are a widely observed feature of Autism Spectrum Disorder (ASD). However, it is unclear how prosodic differences in ASD manifest across different languages that demonstrate cross-linguistic variability in prosody. Using a supervised machine-learning analytic approach, we examined acoustic features relevant to rhythmic and intonational aspects of prosody derived from narrative samples elicited in English and Cantonese, two typologically and prosodically distinct languages. Our models revealed successful classification of ASD diagnosis using rhythm-relative features within and across both languages. Classification with intonation-relevant features was significant for English but not Cantonese. Results highlight differences in rhythm as a key prosodic feature impacted in ASD, and also demonstrate important variability in other prosodic properties that appear to be modulated by language-specific differences, such as intonation.
Collapse
Affiliation(s)
- Joseph C. Y. Lau
- Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, Evanston, Illinois, United States of America
| | - Shivani Patel
- Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, Evanston, Illinois, United States of America
| | - Xin Kang
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong S.A.R., China
- Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong S.A.R., China
- Research Centre for Language, Cognition and Language Application, Chongqing University, Chongqing, China
- School of Foreign Languages and Cultures, Chongqing University, Chongqing, China
| | - Kritika Nayar
- Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, Evanston, Illinois, United States of America
| | - Gary E. Martin
- Department of Communication Sciences and Disorders, St. John’s University, Staten Island, New York, United States of America
| | - Jason Choy
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong S.A.R., China
| | - Patrick C. M. Wong
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong S.A.R., China
- Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong S.A.R., China
| | - Molly Losh
- Roxelyn and Richard Pepper Department of Communication Sciences and Disorders, Northwestern University, Evanston, Illinois, United States of America
| |
Collapse
|
28
|
Carlson K, Potter D. Focus Attracts Attachment. LANGUAGE AND SPEECH 2022; 65:491-512. [PMID: 34289730 PMCID: PMC8776885 DOI: 10.1177/00238309211033321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
There is growing evidence that pitch accents as well as prosodic boundaries can affect syntactic attachment. But is this an effect of their perceptual salience (the Salience Hypothesis), or is it because accents mark the position of focus (the Focus Attraction Hypothesis)? A pair of auditory comprehension experiments shows that focus position, as indicated by preceding wh-questions instead of by pitch accents, affects attachment by drawing the ambiguous phrase to the focus. This supports the Focus Attraction Hypothesis (or a pragmatic version of salience) for both these results and previous results of accents on attachment. These experiments show that information structure, as indicated with prosody or other means, influences sentence interpretation, and suggests a view on which modifiers are drawn to the most important information in a sentence.
Collapse
|
29
|
Guan CQ, Meng W, Morett LM, Fraundorf SH. Mapping Pitch Accents to Memory Representations in Spoken Discourse Among Chinese Learners of English: Effects of L2 Proficiency and Working Memory. Front Psychol 2022; 13:870152. [PMID: 35664143 PMCID: PMC9161639 DOI: 10.3389/fpsyg.2022.870152] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2022] [Accepted: 03/09/2022] [Indexed: 11/14/2022] Open
Abstract
We examined L2 learners' interpretation of pitch accent cues in discourse memory and how these effects vary with proficiency and working memory (WM). One hundred sixty-eight L1-Chinese participants learning L2-English listened to recorded discourses containing pairs of contrastive alternatives and then took a later recognition memory test. Their language proficiency and WM were measured through standard tests and the participants were categorized into low, medium, advanced, and high advanced language proficiency groups. We analyzed recognition memory task performance using signal detection theory to tease apart response bias (an overall tendency to affirm memory probes) from sensitivity (the ability to discern whether a specific probe statement is true). The results showed a benefit of contrastive L + H* pitch accents in rejecting probes referring to items unmentioned in a discourse, but not contrastive alternatives themselves. More proficient participants also showed more accurate memory for the discourses overall, as well as a reduced overall bias to affirm the presented statements as true. Meanwhile, that the benefit of L + H* accents in rejecting either contrast probes or unmentioned probes was modulated for people with greater working memory. Participants with higher WM were quite sure that it did not exist in the memory trace as this part of discourse wasn't mentioned. The results support a contrast-uncertainty hypothesis, in which comprehenders recall the contrast set but fail to distinguish which is the correct item. Further, these effects were influenced by proficiency and by working memory, suggesting they reflect incomplete mapping between pitch accent and discourse representation.
Collapse
Affiliation(s)
- Connie Qun Guan
- School of Foreign Studies, Beijing Language and Culture University, Beijing, China
- Department of Psychology, Carnegie Mellon University, Pittsburgh, PA, United States
| | - Wanjin Meng
- China National Institute of Education Sciences, Beijing, China
| | - Laura M. Morett
- Department of Educational Studies in Psychology, Research Methodology, and Counseling, University of Alabama, Tuscaloosa, AL, United States
| | - Scott H. Fraundorf
- Department of Psychology and Learning Research and Development Center, University of Pittsburgh, Pittsburgh, PA, United States
| |
Collapse
|
30
|
Kawar K, Kishon-Rabin L, Segal O. Identification and Comprehension of Narrow Focus by Arabic-Speaking Adolescents With Moderate-to-Profound Hearing Loss. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:2029-2046. [PMID: 35472256 DOI: 10.1044/2022_jslhr-21-00296] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
PURPOSE Processing narrow focus (NF), the stressed word in the sentence, includes both the perceptual ability to identify the stressed word in the sentence and the pragmatic-semantic ability to comprehend the nonexplicit linguistic message. NF and its underlying meaning can be conveyed only via the auditory modality. Therefore, NF can be considered as a measure for assessing the efficacy of the hearing aid (HA) and cochlear implants (CIs) for acquiring nonexplicit language skills. The purpose of this study was to assess identification and comprehension of NF by HA and CI users who are native speakers of Arabic and to associate NF outcomes with speech perception and cognitive and linguistic abilities. METHOD A total of 46 adolescents (age range: 11;2-18;8) participated: 18 with moderate-to-severe hearing loss who used HAs, 10 with severe-to-profound hearing loss who used CIs, and 18 with typical hearing (TH). Test materials included the Arabic Narrow Focus Test (ANFT), which includes three subtests assessing identification (ANFT1), comprehension of NF in simple four-word sentences (ANFT2), and longer sentences with a construction list at the clause or noun phrase level (ANFT3). In addition, speech perception, vocabulary, and working memory were assessed. RESULTS All the participants successfully identified the word carrying NF, with no significant difference between the groups. Comprehension of NF in ANFT2 and ANFT3 was reduced for HA and CI users compared with TH peers, and speech perception, hearing status, and memory for digits predicted the variability in the overall results of ANFT1, ANFT2, and ANFT3, respectively. CONCLUSIONS Arabic speakers who used HAs or CIs were able to identify NF successfully, suggesting that the acoustic cues were perceptually available to them. However, HA and CI users had considerable difficulty in understanding NF. Different factors may contribute to this difficulty, including the memory load during the task as well as pragmatic-linguistic knowledge on the possible meanings of NF.
Collapse
Affiliation(s)
- Khaloob Kawar
- Department of Special Education, Beit Berl College, Kfar Saba, Israel
- Department of Communication Disorders, Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Israel
| | - Liat Kishon-Rabin
- Department of Communication Disorders, Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Israel
| | - Osnat Segal
- Department of Communication Disorders, Steyer School of Health Professions, Sackler Faculty of Medicine, Tel Aviv University, Israel
| |
Collapse
|
31
|
Motoki K, Pathak A, Spence C. Tasting prosody: Crossmodal correspondences between voice quality and basic tastes. Food Qual Prefer 2022. [DOI: 10.1016/j.foodqual.2022.104621] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022]
|
32
|
Chauvette L, Fournier P, Sharp A. The frequency-following response to assess the neural representation of spectral speech cues in older adults. Hear Res 2022; 418:108486. [DOI: 10.1016/j.heares.2022.108486] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/01/2021] [Revised: 03/12/2022] [Accepted: 03/15/2022] [Indexed: 11/04/2022]
|
33
|
Many Changes in Speech through Aging Are Actually a Consequence of Cognitive Changes. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:ijerph19042137. [PMID: 35206323 PMCID: PMC8871602 DOI: 10.3390/ijerph19042137] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/20/2021] [Revised: 01/21/2022] [Accepted: 02/08/2022] [Indexed: 11/16/2022]
Abstract
Background: During aging, changes in human speech may arise because of the neurophysiological deterioration associated with age, or as the result of an impairment in the cognitive processes underlying speech production. Some speech parameters show specific alterations under the presence of dementia. The objective of our study is to identify which of these parameters change because of age, cognitive state, or the interaction of both. Methods: The sample includes 400 people over 55 years old, who were divided into four groups, according to their age. The cognitive state of the participants was assessed through the MMSE test and three ranks were stablished. Gender was also considered in the analysis. Results: Certain temporal, fluency, rhythm, amplitude and voice quality parameters were found to be related to the cognitive state, while disturbance parameters changed due to age. Frequency parameters were exclusively influenced by gender. Conclusions: Understanding how speech parameters are specifically affected by age, cognitive state, or the interaction of both, is determinant to advance in the use of speech as a clinical marker for the detection of cognitive impairments.
Collapse
|
34
|
What is Functional Communication? A Theoretical Framework for Real-World Communication Applied to Aphasia Rehabilitation. Neuropsychol Rev 2022; 32:937-973. [PMID: 35076868 PMCID: PMC9630202 DOI: 10.1007/s11065-021-09531-2] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
Aphasia is an impairment of language caused by acquired brain damage such as stroke or traumatic brain injury, that affects a person’s ability to communicate effectively. The aim of rehabilitation in aphasia is to improve everyday communication, improving an individual’s ability to function in their day-to-day life. For that reason, a thorough understanding of naturalistic communication and its underlying mechanisms is imperative. The field of aphasiology currently lacks an agreed, comprehensive, theoretically founded definition of communication. Instead, multiple disparate interpretations of functional communication are used. We argue that this makes it nearly impossible to validly and reliably assess a person’s communicative performance, to target this behaviour through therapy, and to measure improvements post-therapy. In this article we propose a structured, theoretical approach to defining the concept of functional communication. We argue for a view of communication as “situated language use”, borrowed from empirical psycholinguistic studies with non-brain damaged adults. This framework defines language use as: (1) interactive, (2) multimodal, and (3) contextual. Existing research on each component of the framework from non-brain damaged adults and people with aphasia is reviewed. The consequences of adopting this approach to assessment and therapy for aphasia rehabilitation are discussed. The aim of this article is to encourage a more systematic, comprehensive approach to the study and treatment of situated language use in aphasia.
Collapse
|
35
|
Amichetti NM, Neukam J, Kinney AJ, Capach N, March SU, Svirsky MA, Wingfield A. Adults with cochlear implants can use prosody to determine the clausal structure of spoken sentences. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:4315. [PMID: 34972310 PMCID: PMC8674009 DOI: 10.1121/10.0008899] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/24/2021] [Revised: 11/04/2021] [Accepted: 11/08/2021] [Indexed: 06/14/2023]
Abstract
Speech prosody, including pitch contour, word stress, pauses, and vowel lengthening, can aid the detection of the clausal structure of a multi-clause sentence and this, in turn, can help listeners determine the meaning. However, for cochlear implant (CI) users, the reduced acoustic richness of the signal raises the question of whether CI users may have difficulty using sentence prosody to detect syntactic clause boundaries within sentences or whether this ability is rescued by the redundancy of the prosodic features that normally co-occur at clause boundaries. Twenty-two CI users, ranging in age from 19 to 77 years old, recalled three types of sentences: sentences in which the prosodic pattern was appropriate to the location of a clause boundary within the sentence (congruent prosody), sentences with reduced prosodic information, or sentences in which the location of the clause boundary and the prosodic marking of a clause boundary were placed in conflict. The results showed the presence of congruent prosody to be associated with superior sentence recall and a reduced processing effort as indexed by the pupil dilation. The individual differences in a standard test of word recognition (consonant-nucleus-consonant score) were related to the recall accuracy as well as the processing effort. The outcomes are discussed in terms of the redundancy of the prosodic features, which normally accompany a clause boundary and processing effort.
Collapse
Affiliation(s)
- Nicole M Amichetti
- Department of Psychology, Brandeis University, Waltham, Massachusetts 02453, USA
| | - Jonathan Neukam
- Department of Otolaryngology, New York University (NYU) Langone Medical Center, New York, New York 10016, USA
| | - Alexander J Kinney
- Department of Psychology, Brandeis University, Waltham, Massachusetts 02453, USA
| | - Nicole Capach
- Department of Otolaryngology, New York University (NYU) Langone Medical Center, New York, New York 10016, USA
| | - Samantha U March
- Department of Psychology, Brandeis University, Waltham, Massachusetts 02453, USA
| | - Mario A Svirsky
- Department of Otolaryngology, New York University (NYU) Langone Medical Center, New York, New York 10016, USA
| | - Arthur Wingfield
- Department of Psychology, Brandeis University, Waltham, Massachusetts 02453, USA
| |
Collapse
|
36
|
Ots N. Cognitive constraints on advance planning of sentence intonation. PLoS One 2021; 16:e0259343. [PMID: 34784351 PMCID: PMC8594795 DOI: 10.1371/journal.pone.0259343] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2020] [Accepted: 10/18/2021] [Indexed: 12/04/2022] Open
Abstract
Pitch peaks tend to be higher at the beginning of longer than shorter sentences (e.g., 'A farmer is pulling donkeys' vs 'A farmer is pulling a donkey and goat'), whereas pitch valleys at the ends of sentences are rather constant for a given speaker. These data seem to imply that speakers avoid dropping their voice pitch too low by planning the height of sentence-initial pitch peaks prior to speaking. However, the length effect on sentence-initial pitch peaks appears to vary across different types of sentences, speakers and languages. Therefore, the notion that speakers plan sentence intonation in advance due to the limitations in low voice pitch leaves part of the data unexplained. Consequently, this study suggests a complementary cognitive account of length-dependent pitch scaling. In particular, it proposes that the sentence-initial pitch raise in long sentences is related to high demands on mental resources during the early stages of sentence planning. To tap into the cognitive underpinnings of planning sentence intonation, this study adopts the methodology of recording eye movements during a picture description task, as the eye movements are the established approximation of the real-time planning processes. Measures of voice pitch (Fundamental Frequency) and incrementality (eye movements) are used to examine the relationship between (verbal) working memory (WM), incrementality of sentence planning and the height of sentence-initial pitch peaks.
Collapse
Affiliation(s)
- Nele Ots
- Institute of Linguistics, Wolfgang Goethe University, Frankfurt am Main, Hessen, Germany
| |
Collapse
|
37
|
Fiveash A, Bedoin N, Gordon RL, Tillmann B. Processing rhythm in speech and music: Shared mechanisms and implications for developmental speech and language disorders. Neuropsychology 2021; 35:771-791. [PMID: 34435803 PMCID: PMC8595576 DOI: 10.1037/neu0000766] [Citation(s) in RCA: 42] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022] Open
Abstract
OBJECTIVE Music and speech are complex signals containing regularities in how they unfold in time. Similarities between music and speech/language in terms of their auditory features, rhythmic structure, and hierarchical structure have led to a large body of literature suggesting connections between the two domains. However, the precise underlying mechanisms behind this connection remain to be elucidated. METHOD In this theoretical review article, we synthesize previous research and present a framework of potentially shared neural mechanisms for music and speech rhythm processing. We outline structural similarities of rhythmic signals in music and speech, synthesize prominent music and speech rhythm theories, discuss impaired timing in developmental speech and language disorders, and discuss music rhythm training as an additional, potentially effective therapeutic tool to enhance speech/language processing in these disorders. RESULTS We propose the processing rhythm in speech and music (PRISM) framework, which outlines three underlying mechanisms that appear to be shared across music and speech/language processing: Precise auditory processing, synchronization/entrainment of neural oscillations to external stimuli, and sensorimotor coupling. The goal of this framework is to inform directions for future research that integrate cognitive and biological evidence for relationships between rhythm processing in music and speech. CONCLUSION The current framework can be used as a basis to investigate potential links between observed timing deficits in developmental disorders, impairments in the proposed mechanisms, and pathology-specific deficits which can be targeted in treatment and training supporting speech therapy outcomes. On these grounds, we propose future research directions and discuss implications of our framework. (PsycInfo Database Record (c) 2021 APA, all rights reserved).
Collapse
Affiliation(s)
- Anna Fiveash
- Lyon Neuroscience Research Center, CRNL, CNRS, UMR5292, INSERM, U1028, F-69000, Lyon, France
- University Lyon 1, Lyon, France
| | - Nathalie Bedoin
- Lyon Neuroscience Research Center, CRNL, CNRS, UMR5292, INSERM, U1028, F-69000, Lyon, France
- University Lyon 1, Lyon, France
- University of Lyon 2, CNRS, UMR5596, Lyon, F-69000, France
| | - Reyna L. Gordon
- Department of Otolaryngology – Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, Tennessee, USA
- Vanderbilt Brain Institute, Vanderbilt University, Nashville, Tennessee
- Vanderbilt Genetics Institute, Vanderbilt University, Nashville, Tennessee
- Vanderbilt Kennedy Center, Vanderbilt University Medical Center, Nashville, Tennessee
| | - Barbara Tillmann
- Lyon Neuroscience Research Center, CRNL, CNRS, UMR5292, INSERM, U1028, F-69000, Lyon, France
- University Lyon 1, Lyon, France
| |
Collapse
|
38
|
Kurumada C, Roettger TB. Thinking probabilistically in the study of intonational speech prosody. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2021; 13:e1579. [PMID: 34599647 DOI: 10.1002/wcs.1579] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/31/2021] [Revised: 08/09/2021] [Accepted: 08/26/2021] [Indexed: 11/07/2022]
Abstract
Speech prosody, the melodic and rhythmic properties of a language, plays a critical role in our everyday communication. Researchers have identified unique patterns of prosody that segment words and phrases, highlight focal elements in a sentence, and convey holistic meanings and speech acts that interact with the information shared in context. The mapping between the sound and meaning represented in prosody is suggested to be probabilistic-the same physical instance of sounds can support multiple meanings across talkers and contexts while the same meaning can be encoded in physically distinct sound patterns (e.g., pitch movements). The current overview presents an analysis framework for probing the nature of this probabilistic relationship. Illustrated by examples from the literature and a dataset of German focus marking, we discuss the production variability within and across talkers and consider challenges that this variability imposes on the comprehension system. A better understanding of these challenges, we argue, will illuminate how the human perceptual, cognitive, and computational mechanisms may navigate the variability to arrive at a coherent understanding of speech prosody. The current paper is intended to be an introduction for those who are interested in thinking probabilistically about the sound-meaning mapping in prosody. Open questions for future research are discussed with proposals for examining prosodic production and comprehension within a comprehensive, mathematically-motivated framework of probabilistic inference under uncertainty. This article is categorized under: Linguistics > Language in Mind and Brain Psychology > Language.
Collapse
Affiliation(s)
- Chigusa Kurumada
- Department of Brain and Cognitive Sciences, University of Rochester, Rochester, New York, USA
| | - Timo B Roettger
- Department of Linguistics & Scandinavian Studies, Universitetet i Oslo, Oslo, Norway
| |
Collapse
|
39
|
An X, Soong FK, Yang S, Xie L. Effective and direct control of neural TTS prosody by removing interactions between different attributes. Neural Netw 2021; 143:250-260. [PMID: 34157649 DOI: 10.1016/j.neunet.2021.06.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2020] [Revised: 05/24/2021] [Accepted: 06/03/2021] [Indexed: 10/21/2022]
Abstract
End-to-end TTS advancement has shown that synthesized speech prosody can be controlled by conditioning the decoder with speech prosody attribute labels. However, to annotate quantitatively the prosody patterns of a large set of training data is both time consuming and expensive. To use unannotated data, variational autoencoder (VAE) has been proposed to model individual prosody attribute as a random variable in the latent space. The VAE is an unsupervised approach and the corresponding latent variables are in general correlated with each other. For more effective and direct control of speech prosody along each attribute dimension, it is highly desirable to disentangle the correlated latent variables. Additionally, being able to interpret the disentangled attributes as speech perceptual cues is useful for designing more efficient prosody control of TTS. In this paper, we propose two attribute separation schemes: (1) using 3 separate VAEs to model the real-valued, different prosodic features, i.e., F0, energy and duration; (2) minimizing mutual information between different prosody attributes to remove their mutual correlations, for facilitating more direct prosody control. Experimental results confirm that the two proposed schemes can indeed make individual prosody attributes more interpretable and direct TTS prosody control more effective. The improvements are measured objectively by F0 Frame Error (FFE) and subjectively with MOS and A/B comparison listening tests, respectively. The scatter diagrams of t-SNE also demonstrate the correlations between prosody attributes, which are well disentangled by minimizing their mutual information. Synthesized TTS samples can be found at https://xiaochunan.github.io/prosody/index.html.
Collapse
Affiliation(s)
- Xiaochun An
- Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science, Northwestern Polytechnical University, Xi'an, China.
| | | | - Shan Yang
- Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science, Northwestern Polytechnical University, Xi'an, China.
| | - Lei Xie
- Audio, Speech and Language Processing Group (ASLP@NPU), School of Computer Science, Northwestern Polytechnical University, Xi'an, China.
| |
Collapse
|
40
|
Llanos F, German JS, Gnanateja GN, Chandrasekaran B. The neural processing of pitch accents in continuous speech. Neuropsychologia 2021; 158:107883. [PMID: 33989647 DOI: 10.1016/j.neuropsychologia.2021.107883] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Revised: 04/29/2021] [Accepted: 05/03/2021] [Indexed: 12/21/2022]
Abstract
Pitch accents are local pitch patterns that convey differences in word prominence and modulate the information structure of the discourse. Despite the importance to discourse in languages like English, neural processing of pitch accents remains understudied. The current study investigates the neural processing of pitch accents by native and non-native English speakers while they are listening to or ignoring 45 min of continuous, natural speech. Leveraging an approach used to study phonemes in natural speech, we analyzed thousands of electroencephalography (EEG) segments time-locked to pitch accents in a prosodic transcription. The optimal neural discrimination between pitch accent categories emerged at latencies between 100 and 200 ms. During these latencies, we found a strong structural alignment between neural and phonetic representations of pitch accent categories. In the same latencies, native listeners exhibited more robust processing of pitch accent contrasts than non-native listeners. However, these group differences attenuated when the speech signal was ignored. We can reliably capture the neural processing of discrete and contrastive pitch accent categories in continuous speech. Our analytic approach also captures how language-specific knowledge and selective attention influences the neural processing of pitch accent categories.
Collapse
Affiliation(s)
- Fernando Llanos
- Department of Communication Science and Disorders, University of Pittsburgh, Pittsburgh, PA, USA; Department of Linguistics, The University of Texas at Austin, Austin, TX, USA
| | - James S German
- Aix-Marseille University, CNRS, LPL, Aix-en-Provence, France
| | - G Nike Gnanateja
- Department of Communication Science and Disorders, University of Pittsburgh, Pittsburgh, PA, USA
| | - Bharath Chandrasekaran
- Department of Communication Science and Disorders, University of Pittsburgh, Pittsburgh, PA, USA.
| |
Collapse
|
41
|
Biron T, Baum D, Freche D, Matalon N, Ehrmann N, Weinreb E, Biron D, Moses E. Automatic detection of prosodic boundaries in spontaneous speech. PLoS One 2021; 16:e0250969. [PMID: 33939754 PMCID: PMC8092678 DOI: 10.1371/journal.pone.0250969] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2020] [Accepted: 04/16/2021] [Indexed: 11/19/2022] Open
Abstract
Automatic speech recognition (ASR) and natural language processing (NLP) are expected to benefit from an effective, simple, and reliable method to automatically parse conversational speech. The ability to parse conversational speech depends crucially on the ability to identify boundaries between prosodic phrases. This is done naturally by the human ear, yet has proved surprisingly difficult to achieve reliably and simply in an automatic manner. Efforts to date have focused on detecting phrase boundaries using a variety of linguistic and acoustic cues. We propose a method which does not require model training and utilizes two prosodic cues that are based on ASR output. Boundaries are identified using discontinuities in speech rate (pre-boundary lengthening and phrase-initial acceleration) and silent pauses. The resulting phrases preserve syntactic validity, exhibit pitch reset, and compare well with manual tagging of prosodic boundaries. Collectively, our findings support the notion of prosodic phrases that represent coherent patterns across textual and acoustic parameters.
Collapse
Affiliation(s)
- Tirza Biron
- Department of Physics of Complex Systems, Weizmann Institute of Science, Rehovot, Israel
| | - Daniel Baum
- Department of Physics of Complex Systems, Weizmann Institute of Science, Rehovot, Israel
| | - Dominik Freche
- Sagol Center for Brain and Mind, Interdisciplinary Center, Herzliya, Israel
| | - Nadav Matalon
- Department of Linguistics, The Hebrew University, Jerusalem, Israel
| | - Netanel Ehrmann
- Department of Physics of Complex Systems, Weizmann Institute of Science, Rehovot, Israel
| | - Eyal Weinreb
- Department of Physics of Complex Systems, Weizmann Institute of Science, Rehovot, Israel
| | - David Biron
- Department of Physics of Complex Systems, Weizmann Institute of Science, Rehovot, Israel
| | - Elisha Moses
- Department of Physics of Complex Systems, Weizmann Institute of Science, Rehovot, Israel
| |
Collapse
|
42
|
THORSON JC, MORGAN JL. Prosodic realizations of new, given, and corrective referents in the spontaneous speech of toddlers. JOURNAL OF CHILD LANGUAGE 2021; 48:541-568. [PMID: 34726145 PMCID: PMC8567208 DOI: 10.1017/s0305000920000434] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Our motivation was to examine how toddler (2;6) and adult speakers of American English prosodically realize information status categories. The aims were three-fold: 1) to analyze how adults phonologically make information status distinctions; 2) to examine how these same categories are signaled in toddlers' spontaneous speech; and 3) to analyze the three primary acoustic correlates of prosody (F0, intensity, and duration). During a spontaneous speech task designed as an interactive game, a set of target nouns was elicited as one of three types (new, given, corrective). Results show that toddlers primarily used H* across information status categories, with secondary preferences for deaccenting given information and for using L+H* for corrective information. Only duration distinguished information status, and duration, average pitch, and intensity differentiated pitch accent types for both adults and children. Discussion includes how pitch accent selection and input play a role in guiding prosodic realizations of information status.
Collapse
Affiliation(s)
- Jill C. THORSON
- University of New Hampshire, Department of Communication Sciences and Disorders, Durham, New Hampshire, USA
| | - James L. MORGAN
- Brown University, Department of Cognitive, Linguistic & Psychological Sciences, Providence, Rhode Island, USA
| |
Collapse
|
43
|
Martínez-Nicolás I, Llorente TE, Martínez-Sánchez F, Meilán JJG. Ten Years of Research on Automatic Voice and Speech Analysis of People With Alzheimer's Disease and Mild Cognitive Impairment: A Systematic Review Article. Front Psychol 2021; 12:620251. [PMID: 33833713 PMCID: PMC8021952 DOI: 10.3389/fpsyg.2021.620251] [Citation(s) in RCA: 54] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2020] [Accepted: 02/15/2021] [Indexed: 11/25/2022] Open
Abstract
Background: The field of voice and speech analysis has become increasingly popular over the last 10 years, and articles on its use in detecting neurodegenerative diseases have proliferated. Many studies have identified characteristic speech features that can be used to draw an accurate distinction between healthy aging among older people and those with mild cognitive impairment and Alzheimer's disease. Speech analysis has been singled out as a cost-effective and reliable method for detecting the presence of both conditions. In this research, a systematic review was conducted to determine these features and their diagnostic accuracy. Methods: Peer-reviewed literature was located across multiple databases, involving studies that apply new procedures of automatic speech analysis to collect behavioral evidence of linguistic impairments along with their diagnostic accuracy on Alzheimer's disease and mild cognitive impairment. The risk of bias was assessed by using JBI and QUADAS-2 checklists. Results: Thirty-five papers met the inclusion criteria; of these, 11 were descriptive studies that either identified voice features or explored their cognitive correlates, and the rest were diagnostic studies. Overall, the studies were of good quality and presented solid evidence of the usefulness of this technique. The distinctive acoustic and rhythmic features found are gathered. Most studies record a diagnostic accuracy over 88% for Alzheimer's and 80% for mild cognitive impairment. Conclusion: Automatic speech analysis is a promising tool for diagnosing mild cognitive impairment and Alzheimer's disease. The reported features seem to be indicators of the cognitive changes in older people. The specific features and the cognitive changes involved could be the subject of further research.
Collapse
Affiliation(s)
- Israel Martínez-Nicolás
- Faculty of Psychology, University of Salamanca, Salamanca, Spain.,Institute of Neuroscience of Castilla y León, University of Salamanca, Salamanca, Spain
| | - Thide E Llorente
- Faculty of Psychology, University of Salamanca, Salamanca, Spain.,Institute of Neuroscience of Castilla y León, University of Salamanca, Salamanca, Spain
| | | | - Juan José G Meilán
- Faculty of Psychology, University of Salamanca, Salamanca, Spain.,Institute of Neuroscience of Castilla y León, University of Salamanca, Salamanca, Spain
| |
Collapse
|
44
|
de Simone J, Cevasco J. The Role of the Establishment of Causal Connections and the Modality of Presentation of Discourse in the Generation of Emotion Inferences by Argentine College Students. READING PSYCHOLOGY 2020. [DOI: 10.1080/02702711.2020.1837314] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Affiliation(s)
| | - Jazmín Cevasco
- Department of Psychology, University of Buenos Aires, Buenos Aires, Argentina
- National Research and Technical Research Council, Buenos Aires, Argentina
| |
Collapse
|
45
|
Chien PJ, Friederici AD, Hartwigsen G, Sammler D. Intonation processing increases task-specific fronto-temporal connectivity in tonal language speakers. Hum Brain Mapp 2020; 42:161-174. [PMID: 32996647 PMCID: PMC7721241 DOI: 10.1002/hbm.25214] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2020] [Revised: 09/08/2020] [Accepted: 09/13/2020] [Indexed: 01/08/2023] Open
Abstract
Language comprehension depends on tight functional interactions between distributed brain regions. While these interactions are established for semantic and syntactic processes, the functional network of speech intonation – the linguistic variation of pitch – has been scarcely defined. Particularly little is known about intonation in tonal languages, in which pitch not only serves intonation but also expresses meaning via lexical tones. The present study used psychophysiological interaction analyses of functional magnetic resonance imaging data to characterise the neural networks underlying intonation and tone processing in native Mandarin Chinese speakers. Participants categorised either intonation or tone of monosyllabic Mandarin words that gradually varied between statement and question and between Tone 2 and Tone 4. Intonation processing induced bilateral fronto‐temporal activity and increased functional connectivity between left inferior frontal gyrus and bilateral temporal regions, likely linking auditory perception and labelling of intonation categories in a phonological network. Tone processing induced bilateral temporal activity, associated with the auditory representation of tonal (phonemic) categories. Together, the present data demonstrate the breadth of the functional intonation network in a tonal language including higher‐level phonological processes in addition to auditory representations common to both intonation and tone.
Collapse
Affiliation(s)
- Pei-Ju Chien
- International Max Planck Research School NeuroCom, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.,Otto Hahn Group 'Neural Bases of Intonation in Speech and Music', Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.,Lise Meitner Research Group 'Cognition and Plasticity', Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.,Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Angela D Friederici
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Gesa Hartwigsen
- Lise Meitner Research Group 'Cognition and Plasticity', Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Daniela Sammler
- Otto Hahn Group 'Neural Bases of Intonation in Speech and Music', Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| |
Collapse
|
46
|
Hirose Y. Sequential Interpretation of Pitch Prominence as Contrastive and Syntactic Information: Contrast Comes First, but Syntax Takes Over. LANGUAGE AND SPEECH 2020; 63:455-478. [PMID: 31286829 DOI: 10.1177/0023830919854476] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Pitch accent serves multiple duties (encoding lexical accent, syntactic structure, and focus) in spoken Japanese. This study investigates how listeners interpret a role-ambiguous pitch prominence surfacing as F0 rise, which could be a cue to the resolution of a syntactic ambiguity between two possible branching structures, or a signal of contrastive focus on the constituent accompanied by the rise. Two visual world paradigm experiments tested the same Japanese linguistic stimuli with and without pitch emphasis on the second word of structures of the following form: modifier + N1 + N2. In Experiment 1, the visual context suppressed the availability of the contrastive interpretation; in Experiment 2, the visual context made the contrastive interpretation available. We found that the same pitch event can be interpreted as both syntax-encoding and contrast-encoding information within the course of processing the same sentence, as long as contextual information is made visually available. When contrastive focus is pragmatically felicitous, it is computed immediately, as soon as the incoming input is accompanied by a notable pitch prominence (Experiment 2). The same prosodic cue can then be re-interpreted as a signal to syntax after the branching ambiguity is recognized due to subsequent input (Experiments 1 and 2). This is most consistent with the view that an initially assigned prosodic boundary is exploited for re-interpretation.
Collapse
Affiliation(s)
- Yuki Hirose
- Graduate School of Arts and Sciences, University of Tokyo, Japan
| |
Collapse
|
47
|
Hahn LE, Benders T, Snijders TM, Fikkert P. Six-month-old infants recognize phrases in song and speech. INFANCY 2020; 25:699-718. [PMID: 32794372 DOI: 10.1111/infa.12357] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2019] [Revised: 06/09/2020] [Accepted: 07/02/2020] [Indexed: 11/29/2022]
Abstract
Infants exploit acoustic boundaries to perceptually organize phrases in speech. This prosodic parsing ability is well-attested and is a cornerstone to the development of speech perception and grammar. However, infants also receive linguistic input in child songs. This study provides evidence that infants parse songs into meaningful phrasal units and replicates previous research for speech. Six-month-old Dutch infants (n = 80) were tested in the song or speech modality in the head-turn preference procedure. First, infants were familiarized to two versions of the same word sequence: One version represented a well-formed unit, and the other contained a phrase boundary halfway through. At test, infants were presented two passages, each containing one version of the familiarized sequence. The results for speech replicated the previously observed preference for the passage containing the well-formed sequence, but only in a more fine-grained analysis. The preference for well-formed phrases was also observed in the song modality, indicating that infants recognize phrase structure in song. There were acoustic differences between stimuli of the current and previous studies, suggesting that infants are flexible in their processing of boundary cues while also providing a possible explanation for differences in effect sizes.
Collapse
Affiliation(s)
- Laura E Hahn
- Centre for Language Studies, Radboud University, Nijmegen, The Netherlands.,International Max Planck Research School for Language Sciences, Nijmegen, The Netherlands
| | - Titia Benders
- Department of Linguistics, Macquarie University, Sydney, NSW, Australia
| | - Tineke M Snijders
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.,Donders Institute for Brain, Cognition and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, Nijmegen, The Netherlands
| | - Paula Fikkert
- Centre for Language Studies, Radboud University, Nijmegen, The Netherlands
| |
Collapse
|
48
|
Fitzroy AB, Breen M. Metric Structure and Rhyme Predictability Modulate Speech Intensity During Child-Directed and Read-Alone Productions of Children's Literature. LANGUAGE AND SPEECH 2020; 63:292-305. [PMID: 31074328 DOI: 10.1177/0023830919843158] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Temporal and phonological predictability in children's literature may support early literacy acquisition. Realization of predictive structure in caregiver prosody could guide children's attention during shared reading, thereby supporting reading subskill development. However, little is known about how predictive structure is realized prosodically during child-directed reading. We investigated whether speakers use word intensity to signal predictive metric and rhyme structure in child-directed and read-alone productions of The Cat in the Hat (Dr. Seuss, 1957), by modeling maximum intensity (dB) of monosyllabic words as a function of metric strength, rhyme predictability, and a set of control parameters. In the control model, intensity increased with lower lexical frequency, capitalization, first mention, and likelihood of a syntactic boundary. Metric structure predicted word intensity beyond these control factors in a hierarchical manner: words aligned with beat one in a 6/8 metric structure were produced with highest intensity, words aligned with beat four were produced with intermediate intensity, and words aligned with all other beats were produced with the lowest intensity. Additionally, phonologically predictable rhyme targets were reduced in intensity. The effects of meter and rhyme were not moderated by the presence of a child audience. These results demonstrate that predictability along multiple dimensions is encoded during reading of poetic children's literature, and that metric structure is realized hierarchically in word intensity. Further, the manner by which predictability is encoded in word intensity differs from that previously reported for word duration in this corpus (Breen, 2018), demonstrating that intensity and duration present nonidentical prosodic information channels.
Collapse
Affiliation(s)
- Ahren B Fitzroy
- Department of Psychology and Education, Mount Holyoke College; Department of Psychological and Brain Sciences, University of Massachusetts, USA
| | - Mara Breen
- Department of Psychology and Education, Mount Holyoke College, USA
| |
Collapse
|
49
|
Meilán JJG, Martínez-Sánchez F, Martínez-Nicolás I, Llorente TE, Carro J. Changes in the Rhythm of Speech Difference between People with Nondegenerative Mild Cognitive Impairment and with Preclinical Dementia. Behav Neurol 2020; 2020:4683573. [PMID: 32351632 PMCID: PMC7178534 DOI: 10.1155/2020/4683573] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2019] [Revised: 03/25/2020] [Accepted: 03/26/2020] [Indexed: 11/17/2022] Open
Abstract
This study explores several speech parameters related to mild cognitive impairment, as well as those that might be flagging the presence of an underlying neurodegenerative process. Speech is an excellent biomarker because it is not invasive and, what is more, its analysis is rapid and economical. Our aim has been to ascertain whether the typical speech patterns of people with Alzheimer's disease are also present during the disorder's preclinical stages. To do so, we shall be using a task that involves reading out aloud. This is followed by an analysis of the recordings, looking for the possible parameters differentiating between those older people with MCI and a high probability of developing dementia and those with MCI that will not do so. We found that the disease's most differentiating parameters prior to its onset involve changes in speech duration and an alteration in rhythm rate and intensity. These parameters seem to be related to the first difficulties in lexical access among older people with AD.
Collapse
Affiliation(s)
- Juan J. G. Meilán
- Faculty of Psychology, University of Salamanca, Salamanca, Spain
- Institute of Neurosciences of Castile and Leon, Salamanca., Spain
| | | | - Israel Martínez-Nicolás
- Faculty of Psychology, University of Salamanca, Salamanca, Spain
- Institute of Neurosciences of Castile and Leon, Salamanca., Spain
| | - Thide E. Llorente
- Faculty of Psychology, University of Salamanca, Salamanca, Spain
- Institute of Neurosciences of Castile and Leon, Salamanca., Spain
| | - Juan Carro
- Faculty of Psychology, University of Salamanca, Salamanca, Spain
- Institute of Neurosciences of Castile and Leon, Salamanca., Spain
| |
Collapse
|
50
|
Abstract
The processing of emotional nonlinguistic information in speech is defined as emotional prosody. This auditory nonlinguistic information is essential in the decoding of social interactions and in our capacity to adapt and react adequately by taking into account contextual information. An integrated model is proposed at the functional and brain levels, encompassing 5 main systems that involve cortical and subcortical neural networks relevant for the processing of emotional prosody in its major dimensions, including perception and sound organization; related action tendencies; and associated values that integrate complex social contexts and ambiguous situations.
Collapse
Affiliation(s)
- Didier Grandjean
- Department of Psychology and Educational Sciences and Swiss Center for Affective Sciences, University of Geneva, Switzerland
| |
Collapse
|