1
|
Magnuson JS, You H, Hannagan T. Lexical Feedback in the Time-Invariant String Kernel (TISK) Model of Spoken Word Recognition. J Cogn 2024; 7:38. [PMID: 38681820 PMCID: PMC11049678 DOI: 10.5334/joc.362] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Accepted: 04/03/2024] [Indexed: 05/01/2024] Open
Abstract
The Time-Invariant String Kernel (TISK) model of spoken word recognition (Hannagan, Magnuson & Grainger, 2013; You & Magnuson, 2018) is an interactive activation model with many similarities to TRACE (McClelland & Elman, 1986). However, by replacing most time-specific nodes in TRACE with time-invariant open-diphone nodes, TISK uses orders of magnitude fewer nodes and connections than TRACE. Although TISK performed remarkably similarly to TRACE in simulations reported by Hannagan et al., the original TISK implementation did not include lexical feedback, precluding simulation of top-down effects, and leaving open the possibility that adding feedback to TISK might fundamentally alter its performance. Here, we demonstrate that when lexical feedback is added to TISK, it gains the ability to simulate top-down effects without losing the ability to simulate the fundamental phenomena tested by Hannagan et al. Furthermore, with feedback, TISK demonstrates graceful degradation when noise is added to input, although parameters can be found that also promote (less) graceful degradation without feedback. We review arguments for and against feedback in cognitive architectures, and conclude that feedback provides a computationally efficient basis for robust constraint-based processing.
Collapse
Affiliation(s)
- James S. Magnuson
- BCBL: Basque Center on Cognition, Brain & Language, Donostia-San Sebastián, Spain
- Ikerbasque: Basque Foundation for Science, Bilbao, Spain
| | - Heejo You
- Department of Psychological Sciences and CT Institute for the Brain and Cognitive Sciences, University of Connecticut, Storrs, CT, USA
| | - Thomas Hannagan
- Hyundai Motor Group Robotics LAB, Uiwang, South Korea
- Stellantis Group, The Netherlands
| |
Collapse
|
2
|
Jeppsen C, Baxelbaum K, Tomblin B, Klein K, McMurray B. The development of lexical processing: Real-time phonological competition and semantic activation in school age children. Q J Exp Psychol (Hove) 2024:17470218241244799. [PMID: 38508999 DOI: 10.1177/17470218241244799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/22/2024]
Abstract
Prior research suggests that the development of speech perception and word recognition stabilises in early childhood. However, recent work suggests that development of these processes continues throughout adolescence. This study aimed to investigate whether these developmental changes are based solely within the lexical system or are due to domain general changes, and to extend this investigation to lexical-semantic processing. We used two Visual World Paradigm tasks: one to examine phonological and semantic processing, one to capture non-linguistic domain-general skills. We tested 43 seven- to nine-year-olds, 42 ten- to thirteen-year-olds, and 30 sixteen- to seventeen-year-olds. Older children were quicker to fixate the target word and exhibited earlier onset and offset of fixations to both semantic and phonological competitors. Visual/cognitive skills explained significant, but not all, variance in the development of these effects. Developmental changes in semantic activation were largely attributable to changes in upstream phonological processing. These results suggest that the concurrent development of linguistic processes and broader visual/cognitive skills lead to developmental changes in real-time phonological competition, while semantic activation is more stable across these ages.
Collapse
Affiliation(s)
- Charlotte Jeppsen
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City, IA, USA
| | - Keith Baxelbaum
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City, IA, USA
| | - Bruce Tomblin
- Department of Communication Sciences and Disorders, The University of Iowa, Iowa City, IA, USA
| | - Kelsey Klein
- Department of Audiology and Speech Pathology, The University of Tennessee Health Science Center, Memphis, TN, USA
| | - Bob McMurray
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City, IA, USA
- Department of Communication Sciences and Disorders, The University of Iowa, Iowa City, IA, USA
| |
Collapse
|
3
|
Hunter CR, Abrahamyan H. Sensitivity, reliability and convergent validity of sequential dual-task measures of listening effort. Int J Audiol 2024; 63:30-39. [PMID: 36427054 DOI: 10.1080/14992027.2022.2145513] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Accepted: 11/04/2022] [Indexed: 11/27/2022]
Abstract
OBJECTIVE The aim of the current study was to assess the sensitivity, reliability and convergent validity of objective measures of listening effort collected in a sequential dual-task. DESIGN On each trial, participants viewed a set of digits and listened to a spoken sentence presented at one of a range of signal-to-noise ratios (SNR) and then typed the sentence-final word and recalled the digits. Listening effort measures included word response time, digit recall accuracy and digit response time. In Experiment 1, SNR on each trial was randomised. In Experiment 2, SNR varied in a blocked design, and in each block self-reported listening effort was also collected. STUDY SAMPLES Separate groups of 40 young adults participated in each experiment. RESULTS Effects of SNR were observed for all measures. Linear effects of SNR were generally observed even with word recognition accuracy factored out of the models. Among the objective measures, reliability was excellent, and repeated-measures correlations, though not between-subjects correlations, were nearly all significant. CONCLUSION The objective measures assessed appear to be sensitive and reliable indices of listening effort that are non-redundant with speech intelligibility and have strong within-participants convergent validity. Results support use of these measures in future studies of listening effort.
Collapse
Affiliation(s)
- Cynthia R Hunter
- Speech Perception, Cognition, and Hearing Laboratory, Department of Speech-Language-Hearing: Sciences and Disorders, University of Kansas, Lawrence, KS, USA
| | - Hayk Abrahamyan
- Language Perception Laboratory, Department of Psychology, State University of New York at Buffalo, Buffalo, NY, USA
| |
Collapse
|
4
|
Dufour S, Mirault J, Grainger J. Rime Priming Effects in Spoken Word Recognition. Exp Psychol 2023; 70:336-343. [PMID: 38288915 PMCID: PMC11060140 DOI: 10.1027/1618-3169/a000598] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Revised: 09/19/2023] [Accepted: 11/22/2023] [Indexed: 04/12/2024]
Abstract
In this study, we re-examined the facilitation that occurs when auditorily presented monosyllabic primes and targets share their final phonemes, and in particular the rime (e.g., /vɔʀd/-/kɔʀd/). More specifically, we asked whether this rime facilitation effect is also observed when the two last consonants of the rime are transposed (e.g., /vɔʀd/-/kɔʀd/). In comparison to a control condition in which the primes and the targets were unrelated (e.g., /pylt/-/kɔʀd/), we found significant priming effects in both the rime (/vɔdʀ/-/kɔʀd/) and the transposed-phoneme "rime" /vɔdʀ/-/kɔʀd/ conditions. We also observed a significantly greater priming effect in the former condition than in the latter condition. We use the theoretical framework of the TISK model (Hannagan et al., 2013) to propose a novel account of final overlap phonological priming in terms of activation of both position-independent phoneme representations and bi-phone representations.
Collapse
Affiliation(s)
- Sophie Dufour
- CNRS, LPL, UMR 7309, Aix-Marseille Université, Aix-en-Provence, France
- Institute for Language, Communication, and the Brain, Aix-Marseille Université, Aix-en-Provence, France
| | - Jonathan Mirault
- Laboratoire de Psychologie Cognitive, Aix-Marseille Université & CNRS, Aix-en-Provence, France
- Pôle pilote AMPIRIC, Institut National Supérieur du Professorat et de l'Éducation (INSPÉ), Aix-Marseille Université, Aix-en-Provence, France
| | - Jonathan Grainger
- Institute for Language, Communication, and the Brain, Aix-Marseille Université, Aix-en-Provence, France
- Laboratoire de Psychologie Cognitive, Aix-Marseille Université & CNRS, Aix-en-Provence, France
| |
Collapse
|
5
|
Dufour S, Mirault J, Grainger J. When phonological neighbours cooperate during spoken sentence processing. Q J Exp Psychol (Hove) 2023:17470218231196823. [PMID: 37578078 DOI: 10.1177/17470218231196823] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/15/2023]
Abstract
This study examined for the first time the impact of the presence of a phonological neighbour on word recognition when the target word and its neighbour co-occur in a spoken sentence. To do so, we developed a new task, the verb detection task, in which participants were instructed to respond as soon as they detected a verb in a sequence of words, thus allowing us to probe spoken word recognition processes in real time. We found that participants were faster at detecting a verb when it was phonologically related to the preceding noun than when it was phonologically unrelated. This effect was found with both correct sentences (Experiment 1) and with ungrammatical sequences of words (Experiment 2). The effect was also found in Experiment 3 where adjacent phonologically related words were included in the non-verb condition (i.e., word sequences not containing a verb), thus ruling out any strategic influences. These results suggest that activation persists across different words during spoken sentence processing such that processing of a word at position n + 1 benefits from the sublexical phonology activated during processing of the word at position n. We discuss how different models of spoken word recognition might be able (or not) to account for these findings.
Collapse
Affiliation(s)
- Sophie Dufour
- CNRS, LPL, UMR 7309, Aix-Marseille Université, Aix-en-Provence, France
- Institute for Language, Communication, and the Brain, Aix-Marseille Université, Aix-en-Provence, France
| | - Jonathan Mirault
- Laboratoire de Psychologie Cognitive, CNRS, Aix-Marseille Université, Marseille, France
- Pôle pilote AMPIRIC, Institut National Supérieur du Professorat et de l'Éducation (INSPÉ), Aix-Marseille Université, Marseille, France
| | - Jonathan Grainger
- Institute for Language, Communication, and the Brain, Aix-Marseille Université, Aix-en-Provence, France
- Laboratoire de Psychologie Cognitive, CNRS, Aix-Marseille Université, Marseille, France
| |
Collapse
|
6
|
Dossey E, Jones Z, Clopper CG. Relative Contributions of Social, Contextual, and Lexical Factors in Speech Processing. Lang Speech 2023; 66:322-353. [PMID: 35787020 DOI: 10.1177/00238309221107870] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
This exploratory study examined the simultaneous interactions and relative contributions of bottom-up social information (regional dialect, speaking style), top-down contextual information (semantic predictability), and the internal dynamics of the lexicon (neighborhood density, lexical frequency) to lexical access and word recognition. Cross-modal matching and intelligibility in noise tasks were conducted with a community sample of adults at a local science museum. Each task featured one condition in which keywords were presented in isolation and one condition in which they were presented within a multiword phrase. Lexical processing was slower and more accurate when keywords were presented in their phrasal context, and was both faster and more accurate for auditory stimuli produced in the local Midland dialect. In both tasks, interactions were observed among stimulus dialect, speaking style, semantic predictability, phonological neighborhood density, and lexical frequency. These interactions revealed that bottom-up social information and top-down contextual information contribute more to speech processing than the internal dynamics of the lexicon. Moreover, the relatively stronger bottom-up social effects were observed in both the isolated word and multiword phrase conditions, suggesting that social variation is central to speech processing, even in non-interactive laboratory tasks. At the same time, the specific interactions observed differed between the two experiments, reflecting task-specific demands related to processing time constraints and signal degradation.
Collapse
Affiliation(s)
- Ellen Dossey
- Department of Linguistics, The Ohio State University, USA
| | - Zack Jones
- Department of Linguistics, The Ohio State University, USA
| | | |
Collapse
|
7
|
Abstract
Spoken word recognition depends on variations in fine-grained phonetics as listeners decode speech. However, many models of second language (L2) speech perception focus on units such as isolated syllables, and not on words. In two eye-tracking experiments, we investigated how fine-grained phonetic details (i.e. duration of nasalization on contrastive and coarticulatory nasalized vowels in Canadian French) influenced spoken word recognition in an L2, as compared to a group of native (L1) listeners. Results from L2 listeners (English-native speakers) indicated that fine-grained phonetics impacted the recognition of words, i.e. they were able to use nasalization duration variability in a way similar to L1-French listeners, providing evidence that lexical representations can be highly specified in an L2. Specifically, L2 listeners were able to distinguish minimal word pairs (differentiated by the presence of phonological vowel nasalization in French) and were able to use variability in a way approximating L1-French listeners. Furthermore, the robustness of the French "nasal vowel" category in L2 listeners depended on age of exposure. Early bilinguals displayed greater sensitivity to some ambiguity in the stimuli than late bilinguals, suggesting that early bilinguals had greater sensitivity to small variations in the signal and thus better knowledge of the phonetic cue associated with phonological vowel nasalization in French, similarly to L1 listeners.
Collapse
Affiliation(s)
- Félix Desmeules-Trudel
- Félix Desmeules-Trudel, Department of Psychology, University of Toronto Mississauga, 3359 Mississauga Road, CCT Building, Room 4019, Mississauga, ON L5L 1C6, Canada.
| | | |
Collapse
|
8
|
Zhang J, Meng Y, Wu C, Yuan Z. Spoken Word Recognition across Language Boundary: ERP Evidence of Prosodic Transfer Driven by Pitch. Brain Sci 2023; 13. [PMID: 36831746 DOI: 10.3390/brainsci13020202] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2023] [Revised: 01/20/2023] [Accepted: 01/21/2023] [Indexed: 01/27/2023] Open
Abstract
Extensive research has explored the perception of English lexical stress by Chinese EFL learners and tried to unveil the underlying mechanism of the prosodic transfer from a native tonal language to a non-native stress language. However, the role of the pitch as the shared cue by lexical stress and lexical tone during the transfer remains controversial when the segmental cue (i.e., reduced vowel) is absent. By employing event-related potential (ERP) measurements, the current study aimed to further investigate the role of the pitch during the prosodic transfer from L1 lexical tone to L2 lexical stress and the underlying neural responses. Two groups of adult Chinese EFL learners were compared, as both Mandarin and Cantonese are tonal languages with different levels of complexity. The results showed that Cantonese speakers relied more than Mandarin speakers on pitch cues, not only in their processing of English lexical stress but also in word recognition. Our findings are consistent with the arguments of Cue Weighting and attest to the influence of native tonal language experience on second language acquisition. The results may have implications on pedagogical methods that pitch could be an important clue in second language teaching.
Collapse
|
9
|
Tuft SE, Incera S, MᶜLennan CT. Examining long-term repetition priming effects in spoken word recognition using computer mouse tracking. Front Psychol 2023; 13:1074784. [PMID: 36687990 PMCID: PMC9850077 DOI: 10.3389/fpsyg.2022.1074784] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 12/07/2022] [Indexed: 01/07/2023] Open
Abstract
Language researchers in a variety of disciplines have used priming as a tool to investigate theoretical questions. In spoken word recognition, long-term repetition priming effects have been obtained across a number of behavioral tasks (e.g., lexical decision, shadowing). Repeated - primed - words are responded to more efficiently than new - unprimed - words. However, to our knowledge, long-term repetition priming effects have not been examined using computer mouse tracking, which would provide data regarding the time course of long-term repetition priming effects. Consequently, we compared participants' lexical decision responses using a computer mouse to primed and unprimed words. We predicted that participants would respond more efficiently to primed words compared to unprimed words. Indeed, across all of the dependent variables investigated (accuracy, reaction time, mouse trajectories) and across environments (in person, online), participants responded more efficiently to primed words than to unprimed words. We also performed additional exploratory analyses examining long-term repetition priming effects for nonwords. Across environments (in person, online), participants had more errors to primed nonwords than to unprimed nonwords, but there were no differences in reaction times and mouse trajectories. The current data demonstrating long-term repetition priming effects in mouse tracking are expected to motivate future investigations examining the time course of various long-term repetition priming effects for both words and nonwords.
Collapse
Affiliation(s)
- Samantha E. Tuft
- Language Research Laboratory, Department of Psychology, Cleveland State University, Cleveland, OH, United States
| | - Sara Incera
- Multilingual Laboratory, Department of Psychology, Eastern Kentucky University, Richmond, KY, United States
| | - Conor T. MᶜLennan
- Language Research Laboratory, Department of Psychology, Cleveland State University, Cleveland, OH, United States,*Correspondence: Conor T. MᶜLennan,
| |
Collapse
|
10
|
Abstract
Efficient word recognition depends on the ability to overcome competition from overlapping words. The nature of the overlap depends on the input modality: spoken words have temporal overlap from other words that share phonemes in the same positions, whereas written words have spatial overlap from other words with letters in the same places. It is unclear how these differences in input format affect the ability to recognise a word and the types of competitors that become active while doing so. This study investigates word recognition in both modalities in children between 7 and 15. Children complete a visual-world paradigm eye-tracking task that measures competition from words with several types of overlap, using identical word lists between modalities. Results showed correlated developmental changes in the speed of target recognition in both modalities. In addition, developmental changes were seen in the efficiency of competitor suppression for some competitor types in the spoken modality. These data reveal some developmental continuity in the process of word recognition independent of modality but also some instances of independence in how competitors are activated. Stimuli, data, and analyses from this project are available at: https://osf.io/eav72.
Collapse
Affiliation(s)
- Keith S Apfelbaum
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA
| | - Claire Goodwin
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA
| | - Christina Blomquist
- Department of Communication Sciences and Disorders, University of Maryland, College Park, MD, USA
| | - Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA
- Department of Communication Sciences and Disorders, Department of Linguistics, Department of Otolaryngology, University of Iowa, Iowa City, IA, USA
| |
Collapse
|
11
|
Abstract
As a spoken word unfolds over time, similar sounding words (cap and cat) compete until one word "wins". Lexical competition becomes more efficient from infancy through adolescence. We examined one potential mechanism underlying this development: lexical inhibition, by which activated candidates suppress competitors. In Experiment 1, younger (7-8 years) and older (12-13 years) children heard words (cap) in which the onset was manipulated to briefly boost competition from a cohort competitor (cat). This was compared to a condition with a nonword (cack) onset that would not inhibit the target. Words were presented in a visual world task during which eye movements were recorded. Both groups showed less looking to the target when perceiving the competitor-splice relative to the nonword-splice, showing engagement of lexical inhibition. Exploratory analyses of linguistic adaptation across the experiment revealed that older children demonstrated consistent lexical inhibition across the experiment and younger children did not, initially showing no effect in the first half of trials and then a robust effect in the latter half. In Experiment 2, adults also displayed consistent lexical inhibition in the same task. These findings suggest that younger children do not consistently engage lexical inhibition in typical listening but can quickly bring it online in response to certain linguistic experiences. Computational modeling showed that age-related differences are best explained by increased engagement of inhibition rather than growth in activation. These findings suggest that continued development of lexical inhibition in later childhood may underlie increases in efficiency of spoken word recognition. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Collapse
|
12
|
Abstract
Dominant theoretical accounts of developmental language disorder (DLD) commonly invoke working memory capacity limitations. In the current report, we present an alternative view: That working memory in DLD is not under-resourced but overloaded due to operating on speech representations with low discriminability. This account is developed through computational simulations involving deep convolutional neural networks trained on spoken word spectrograms in which information is either retained to mimic typical development or degraded to mimic the auditory processing deficits identified among some children with DLD. We assess not only spoken word recognition accuracy and predictive probability and entropy (i.e., predictive distribution spread), but also use mean-field-theory based manifold analysis to assess; (a) internal speech representation dimensionality and (b) classification capacity, a measure of the networks' ability to isolate any given internal speech representation that is used as a proxy for attentional control. We show that instantiating a low-level auditory processing deficit results in the formation of internal speech representations with atypically high dimensionality, and that classification capacity is exhausted due to low representation separability. These representation and control deficits underpin not only lower performance accuracy but also greater uncertainty even when making accurate predictions in a simulated spoken word recognition task (i.e., predictive distributions with low maximum probability and high entropy), which replicates the response delays and word finding difficulties often seen in DLD. Overall, these simulations demonstrate a theoretical account of speech representation and processing deficits in DLD in which working memory capacity limitations play no causal role. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Collapse
|
13
|
Poulton VR, Nieuwland MS. Can You Hear What's Coming? Failure to Replicate ERP Evidence for Phonological Prediction. Neurobiol Lang (Camb) 2022; 3:556-574. [PMID: 37215344 PMCID: PMC10158594 DOI: 10.1162/nol_a_00078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/05/2021] [Accepted: 07/18/2022] [Indexed: 05/24/2023]
Abstract
Prediction-based theories of language comprehension assume that listeners predict both the meaning and phonological form of likely upcoming words. In alleged event-related potential (ERP) demonstrations of phonological prediction, prediction-mismatching words elicit a phonological mismatch negativity (PMN), a frontocentral negativity that precedes the centroparietal N400 component. However, classification and replicability of the PMN has proven controversial, with ongoing debate on whether the PMN is a distinct component or merely an early part of the N400. In this electroencephalography (EEG) study, we therefore attempted to replicate the PMN effect and its separability from the N400, using a participant sample size (N = 48) that was more than double that of previous studies. Participants listened to sentences containing either a predictable word or an unpredictable word with/without phonological overlap with the predictable word. Preregistered analyses revealed a widely distributed negative-going ERP in response to unpredictable words in both the early (150-250 ms) and the N400 (300-500 ms) time windows. Bayes factor analysis yielded moderate evidence against a different scalp distribution of the effects in the two time windows. Although our findings do not speak against phonological prediction during sentence comprehension, they do speak against the PMN effect specifically as a marker of phonological prediction mismatch. Instead of an PMN effect, our results demonstrate the early onset of the auditory N400 effect associated with unpredictable words. Our failure to replicate further highlights the risk associated with commonly employed data-contingent analyses (e.g., analyses involving time windows or electrodes that were selected based on visual inspection) and small sample sizes in the cognitive neuroscience of language.
Collapse
Affiliation(s)
- Victoria R. Poulton
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Mante S. Nieuwland
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Nijmegen, Radboud University, The Netherlands
| |
Collapse
|
14
|
Oganyan M, Wright RA. The Role of the Root in Spoken Word Recognition in Hebrew: An Auditory Gating Paradigm. Brain Sci 2022; 12:750. [PMID: 35741635 DOI: 10.3390/brainsci12060750] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2022] [Revised: 05/24/2022] [Accepted: 06/01/2022] [Indexed: 11/17/2022] Open
Abstract
Very few studies have investigated online spoken word recognition in templatic languages. In this study, we investigated both lexical (neighborhood density and frequency) and morphological (role of root morpheme) aspects of spoken word recognition of Hebrew, a templatic language, using the traditional gating paradigm. Additionally, we compared the traditional gating paradigm with a novel, phoneme-based gating paradigm. The phoneme-based approach allows for better control of information available at each gate. We found lexical effects with high-frequency words and low neighborhood density words being recognized at earlier gates. We also found that earlier access to root-morpheme information enabled word recognition at earlier gates. Finally, we showed that both the traditional gating paradigm and gating by phoneme paradigm yielded equivalent results.
Collapse
|
15
|
Brown VA, Fox NP, Strand JF. "Where are the . . . Fixations?": Grammatical number cues guide anticipatory fixations to upcoming referents and reduce lexical competition. J Exp Psychol Learn Mem Cogn 2022; 48:643-657. [PMID: 34014757 PMCID: PMC8605024 DOI: 10.1037/xlm0001019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Listeners make use of contextual cues during continuous speech processing that help overcome the limitations of the acoustic input. These semantic, grammatical, and pragmatic cues facilitate prediction of upcoming words and/or reduce the lexical search space by inhibiting activation of contextually inappropriate words that share phonological information with the target. The current study used the visual world paradigm to assess whether and how listeners use contextual cues about grammatical number during sentence processing by presenting target words in carrier phrases that were grammatically unconstraining ("Click on the . . .") or grammatically constraining ("Where is/are the . . ."). Prior to the onset of the target word, listeners were already more likely to fixate on plural objects in the "Where are the . . ." context than the "Where is the . . ." context, indicating that they used the construction of the verb to anticipate the referent. Further, participants showed less interference from cohort competitors when the sentence frame made them contextually inappropriate, but still fixated on those words more than on phonologically unrelated distractor words. These results suggest that listeners rapidly and flexibly make use of contextual cues about grammatical number while maintaining sensitivity to the bottom-up input. (PsycInfo Database Record (c) 2022 APA, all rights reserved).
Collapse
|
16
|
Zou T, Liu Y, Zhong H. The Roles of Consonant, Rime, and Tone in Mandarin Spoken Word Recognition: An Eye-Tracking Study. Front Psychol 2022; 12:740444. [PMID: 35069318 PMCID: PMC8766742 DOI: 10.3389/fpsyg.2021.740444] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2021] [Accepted: 12/06/2021] [Indexed: 11/13/2022] Open
Abstract
This study investigated the relative role of sub-syllabic components (initial consonant, rime, and tone) in spoken word recognition of Mandarin Chinese using an eye-tracking experiment with a visual world paradigm. Native Mandarin speakers (all born and grew up in Beijing) were presented with four pictures and an auditory stimulus. They were required to click the picture according to the sound stimulus they heard, and their eye movements were tracked during this process. For a target word (e.g., tang2 "candy"), nine conditions of competitors were constructed in terms of the amount of their phonological overlap with the target: consonant competitor (e.g., ti1 "ladder"), rime competitor (e.g., lang4 "wave"), tone competitor (e.g., niu2 "cow"), consonant plus rime competitor (e.g., tang1"soup"), consonant plus tone competitor (e.g., tou2 "head"), rime plus tone competitor (e.g., yang2 "sheep"), cohort competitor (e.g., ta3 "tower"), cohort plus tone competitor (e.g., tao2 "peach"), and baseline competitor (e.g., xue3 "snow"). A growth curve analysis was conducted with the fixation to competitors, targets, and distractors, and the results showed that (1) competitors with consonant or rime overlap can be adequately activated, while tone overlap plays a weaker role since additional tonal information can strengthen the competitive effect only when it was added to a candidate that already bears much phonological similarity with the target. (2) Mandarin words are processed in an incremental way in the time course of word recognition since different partially overlapping competitors could be activated immediately; (3) like the pattern found in English, both cohort and rime competitors were activated to compete for lexical activation, but these two competitors were not temporally distinctive and mainly differed in the size of their competitive effects. Generally, the gradation of activation based on the phonological similarity between target and candidates found in this study was in line with the continuous mapping models and may reflect a strategy of native speakers shaped by the informative characteristics of the interaction among different sub-syllabic components.
Collapse
Affiliation(s)
- Ting Zou
- School of English and International Studies, Beijing Foreign Studies University, Beijing, China
| | | | | |
Collapse
|
17
|
Vitevitch MS, Mullin GJD. What Do Cognitive Networks Do? Simulations of Spoken Word Recognition Using the Cognitive Network Science Approach. Brain Sci 2021; 11:brainsci11121628. [PMID: 34942930 PMCID: PMC8699506 DOI: 10.3390/brainsci11121628] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Revised: 12/05/2021] [Accepted: 12/09/2021] [Indexed: 11/16/2022] Open
Abstract
Cognitive network science is an emerging approach that uses the mathematical tools of network science to map the relationships among representations stored in memory to examine how that structure might influence processing. In the present study, we used computer simulations to compare the ability of a well-known model of spoken word recognition, TRACE, to the ability of a cognitive network model with a spreading activation-like process to account for the findings from several previously published behavioral studies of language processing. In all four simulations, the TRACE model failed to retrieve a sufficient number of words to assess if it could replicate the behavioral findings. The cognitive network model successfully replicated the behavioral findings in Simulations 1 and 2. However, in Simulation 3a, the cognitive network did not replicate the behavioral findings, perhaps because an additional mechanism was not implemented in the model. However, in Simulation 3b, when the decay parameter in spreadr was manipulated to model this mechanism the cognitive network model successfully replicated the behavioral findings. The results suggest that models of cognition need to take into account the multi-scale structure that exists among representations in memory, and how that structure can influence processing.
Collapse
|
18
|
Wang Y, Zang X, Zhang H, Shen W. The Processing of the Second Syllable in Recognizing Chinese Disyllabic Spoken Words: Evidence From Eye Tracking. Front Psychol 2021; 12:681337. [PMID: 34777085 PMCID: PMC8580174 DOI: 10.3389/fpsyg.2021.681337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2021] [Accepted: 10/07/2021] [Indexed: 11/13/2022] Open
Abstract
In the current study, two experiments were conducted to investigate the processing of the second syllable (which was considered as the rhyme at the word level) during Chinese disyllabic spoken word recognition using a printed-word paradigm. In Experiment 1, participants heard a spoken target word and were simultaneously presented with a visual display of four printed words: a target word, a phonological competitor, and two unrelated distractors. The phonological competitors were manipulated to share either full phonemic overlap of the second syllable with targets (the syllabic overlap condition; e.g., , xiao3zhuan4, "calligraphy" vs. , gong1zhuan4, "revolution") or the initial phonemic overlap of the second syllable (the sub-syllabic overlap condition; e.g., , yuan2zhu4, "cylinder" vs. , gong1zhuan4, "revolution") with targets. Participants were asked to select the target words and their eye movements were simultaneously recorded. The results did not show any phonological competition effect in either the syllabic overlap condition or the sub-syllabic overlap condition. In Experiment 2, to maximize the likelihood of observing the phonological competition effect, a target-absent version of the printed-word paradigm was adopted, in which target words were removed from the visual display. The results of Experiment 2 showed significant phonological competition effects in both conditions, i.e., more fixations were made to the phonological competitors than to the distractors. Moreover, the phonological competition effect was found to be larger in the syllabic overlap condition than in the sub-syllabic overlap condition. These findings shed light on the effect of the second syllable competition at the word level during spoken word recognition and, more importantly, showed that the initial phonemes of the second syllable at the syllabic level are also accessed during Chinese disyllabic spoken word recognition.
Collapse
Affiliation(s)
- Youxi Wang
- Center for Cognition and Brain Disorders, The Affiliated Hospital of Hangzhou Normal University, Hangzhou, China.,Institute of Psychological Science, Hangzhou Normal University, Hangzhou, China.,Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
| | - Xuelian Zang
- Center for Cognition and Brain Disorders, The Affiliated Hospital of Hangzhou Normal University, Hangzhou, China.,Institute of Psychological Science, Hangzhou Normal University, Hangzhou, China.,Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
| | - Hua Zhang
- Institute of Psychological Science, Hangzhou Normal University, Hangzhou, China
| | - Wei Shen
- Center for Cognition and Brain Disorders, The Affiliated Hospital of Hangzhou Normal University, Hangzhou, China.,Institute of Psychological Science, Hangzhou Normal University, Hangzhou, China.,Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
| |
Collapse
|
19
|
Hendrickson K, Apfelbaum K, Goodwin C, Blomquist C, Klein K, McMurray B. The profile of real-time competition in spoken and written word recognition: More similar than different. Q J Exp Psychol (Hove) 2021; 75:1653-1673. [PMID: 34666573 DOI: 10.1177/17470218211056842] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Word recognition occurs across two sensory modalities: auditory (spoken words) and visual (written words). While each faces different challenges, they are often described in similar terms as a competition process by which multiple lexical candidates are activated and compete for recognition. While there is a general consensus regarding the types of words that compete during spoken word recognition, there is less consensus for written word recognition. The present study develops a novel version of the Visual World Paradigm (VWP) to examine written word recognition and uses this to assess the nature of the competitor set during word recognition in both modalities using the same experimental design. For both spoken and written words, we found evidence for activation of onset competitors (cohorts, e.g., cat, cap) and words that contain the same phonemes or letters in reverse order (anadromes, e.g., cat, tack). We found no evidence of activation for rhymes (e.g., cat, hat). The results across modalities were quite similar, with the exception that for spoken words, cohorts were more active than anadromes, whereas for written words activation was similar. These results suggest a common characterisation of lexical similarity across spoken and written words: temporal or spatial order is coarsely coded, and onsets may receive more weight in both systems. However, for spoken words, temporary ambiguity during the moment of processing gives cohorts an additional boost during real-time recognition.
Collapse
Affiliation(s)
- Kristi Hendrickson
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA, USA
| | - Keith Apfelbaum
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA
| | - Claire Goodwin
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA.,University of Iowa Health Network Rehabilitation Hospital, Coralville, IA, USA
| | - Christina Blomquist
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Kelsey Klein
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA, USA
| | - Bob McMurray
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA, USA.,Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA.,Department of Otolaryngology, University of Iowa, Iowa City, IA, USA
| |
Collapse
|
20
|
Zeng Y, Fiorentino R, Zhang J. Electrophysiological Signatures of Perceiving Alternated Tone in Mandarin Chinese: Mismatch Negativity to Underlying Tone Conflict. Front Psychol 2021; 12:735593. [PMID: 34646215 PMCID: PMC8504678 DOI: 10.3389/fpsyg.2021.735593] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2021] [Accepted: 08/23/2021] [Indexed: 11/16/2022] Open
Abstract
Although phonological alternation is prevalent in languages, the process of perceiving phonologically alternated sounds is poorly understood, especially at the neurolinguistic level. We examined the process of perceiving Mandarin 3rd tone sandhi (T3 + T3 → T2 + T3) with a mismatch negativity (MMN) experiment. Our design has two independent variables (whether the deviant undergoes tone sandhi; whether the standard and the deviant have matched underlying tone). These two independent variables modulated ERP responses in both the first and the second syllables. Notably, despite the apparent segmental conflict between the standard and the deviant in all conditions, MMN is only observed when neither the standard nor the deviant undergoes tone sandhi, suggesting that discovering the underlying representation of an alternated sound could interfere with the generation of MMN. A tentative model with three hypothesized underlying processing mechanisms is proposed to explain the observed latency and amplitude differences across conditions. The results are also discussed in light of the potential electrophysiological signatures involved in the process of perceiving alternated sounds.
Collapse
Affiliation(s)
- Yuyu Zeng
- Phonetics and Psycholinguistics Laboratory, Department of Linguistics, University of Kansas, Lawrence, KS, United States.,Neurolinguistics and Language Processing Laboratory, Department of Linguistics, University of Kansas, Lawrence, KS, United States
| | - Robert Fiorentino
- Neurolinguistics and Language Processing Laboratory, Department of Linguistics, University of Kansas, Lawrence, KS, United States
| | - Jie Zhang
- Phonetics and Psycholinguistics Laboratory, Department of Linguistics, University of Kansas, Lawrence, KS, United States
| |
Collapse
|
21
|
Duta M, Plunkett K. A Neural Network Model of Lexical-Semantic Competition During Spoken Word Recognition. Front Hum Neurosci 2021; 15:700281. [PMID: 34602993 PMCID: PMC8484523 DOI: 10.3389/fnhum.2021.700281] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2021] [Accepted: 08/19/2021] [Indexed: 11/25/2022] Open
Abstract
Visual world studies show that upon hearing a word in a target-absent visual context containing related and unrelated items, toddlers and adults briefly direct their gaze toward phonologically related items, before shifting toward semantically and visually related ones. We present a neural network model that processes dynamic unfolding phonological representations of words and maps them to static internal lexical, semantic, and visual representations. The model, trained on representations derived from real corpora, simulates this early phonological over semantic/visual preference. Our results support the hypothesis that incremental unfolding of a spoken word is in itself sufficient to account for the transient preference for phonological competitors over both unrelated and semantically and visually related ones. Phonological representations mapped dynamically in a bottom-up fashion to semantic-visual representations capture the early phonological preference effects reported in visual world tasks. The semantic visual preference typically observed later in such a task does not require top-down feedback from a semantic or visual system.
Collapse
Affiliation(s)
- Mihaela Duta
- Oxford Research Software Engineering Group, Department of Computer Science, University of Oxford, Oxford, United Kingdom
- Oxford University Babylab, Department of Experimental Psychology, University of Oxford, Oxford, United Kingdom
| | - Kim Plunkett
- Oxford University Babylab, Department of Experimental Psychology, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
22
|
Nabé M, Schwartz JL, Diard J. COSMO-Onset: A Neurally-Inspired Computational Model of Spoken Word Recognition, Combining Top-Down Prediction and Bottom-Up Detection of Syllabic Onsets. Front Syst Neurosci 2021; 15:653975. [PMID: 34421549 PMCID: PMC8371689 DOI: 10.3389/fnsys.2021.653975] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2021] [Accepted: 07/02/2021] [Indexed: 11/13/2022] Open
Abstract
Recent neurocognitive models commonly consider speech perception as a hierarchy of processes, each corresponding to specific temporal scales of collective oscillatory processes in the cortex: 30-80 Hz gamma oscillations in charge of phonetic analysis, 4-9 Hz theta oscillations in charge of syllabic segmentation, 1-2 Hz delta oscillations processing prosodic/syntactic units and the 15-20 Hz beta channel possibly involved in top-down predictions. Several recent neuro-computational models thus feature theta oscillations, driven by the speech acoustic envelope, to achieve syllabic parsing before lexical access. However, it is unlikely that such syllabic parsing, performed in a purely bottom-up manner from envelope variations, would be totally efficient in all situations, especially in adverse sensory conditions. We present a new probabilistic model of spoken word recognition, called COSMO-Onset, in which syllabic parsing relies on fusion between top-down, lexical prediction of onset events and bottom-up onset detection from the acoustic envelope. We report preliminary simulations, analyzing how the model performs syllabic parsing and phone, syllable and word recognition. We show that, while purely bottom-up onset detection is sufficient for word recognition in nominal conditions, top-down prediction of syllabic onset events allows overcoming challenging adverse conditions, such as when the acoustic envelope is degraded, leading either to spurious or missing onset events in the sensory signal. This provides a proposal for a possible computational functional role of top-down, predictive processes during speech recognition, consistent with recent models of neuronal oscillatory processes.
Collapse
Affiliation(s)
- Mamady Nabé
- Université Grenoble Alpes, CNRS, GIPSA-Lab, Grenoble, France.,Université Grenoble Alpes, CNRS, Laboratoire de Psychologie et NeuroCognition, Grenoble, France
| | | | - Julien Diard
- Université Grenoble Alpes, CNRS, Laboratoire de Psychologie et NeuroCognition, Grenoble, France
| |
Collapse
|
23
|
Bakhtiar M, Mokhlesin M, Pattamadilok C, Politzer-Ahles S, Zhang C. The Effect of Orthographic Transparency on Auditory Word Recognition Across the Development of Reading Proficiency. Front Psychol 2021; 12:691989. [PMID: 34385960 PMCID: PMC8353368 DOI: 10.3389/fpsyg.2021.691989] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2021] [Accepted: 07/02/2021] [Indexed: 11/24/2022] Open
Abstract
A question under debate in psycholinguistics is the nature of the relationship between spoken and written languages. Although it has been extensively shown that orthographic transparency, which varies across writing systems, strongly affects reading performance, its role in speech processing is much less investigated. The present study addressed this issue in Persian, whose writing system provides a possibility to assess the impact of orthographic transparency on spoken word recognition in young children at different stages of reading acquisition. In Persian, the long vowels are systematically present in the script, whereas the spelling correspondence of short vowels is progressively omitted from the script in the course of reading acquisition, thus, turning transparent into opaque spelling. Based on this unique characteristic, we tested 144 monolingual Persian-speaking nonreaders (i.e., preschoolers) and readers (second graders to fifth graders and young adults) in an auditory lexical decision task using transparent and opaque words. Overall, the results showed that, in accordance with the fact that the diacritics of short vowels are progressively omitted during the second year of schooling, the stimuli containing short vowels (opaque words) were recognized more slowly than transparent ones in third graders. Interestingly, there is a hint that the emergence of the transparency effect in the third graders was associated with an overall slower recognition speed in this group compared to their younger peers. These findings indicate that learning opaque spelling-sound correspondence might not only generate interference between the two language codes but also induce a general processing cost in the entire spoken language system.
Collapse
Affiliation(s)
- Mehdi Bakhtiar
- Unit of Human Communication, Development, and Information Sciences, The University of Hong Kong, Hong Kong, China
| | - Maryam Mokhlesin
- Neuromuscular Rehabilitation Research Center, Semnan University of Medical Sciences, Semnan, Iran
- Department of Speech Therapy, University of Social Welfare and Rehabilitation Sciences, Tehran, Iran
| | | | - Stephen Politzer-Ahles
- Department of Chinese and Bilingual Studies, The Hong Kong Polytechnic University, Hong Kong, China
| | - Caicai Zhang
- Department of Chinese and Bilingual Studies, The Hong Kong Polytechnic University, Hong Kong, China
| |
Collapse
|
24
|
Sikora J, Stein C, Ubellacker D, Walker A, Tippett DC. Longitudinal decline in spoken word recognition and object knowledge in primary progressive aphasia. Medicine (Baltimore) 2021; 100:e26163. [PMID: 34087875 PMCID: PMC8183769 DOI: 10.1097/md.0000000000026163] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/08/2021] [Accepted: 05/11/2021] [Indexed: 11/26/2022] Open
Abstract
The premise of this study is that spoken word recognition and object knowledge are impaired in semantic variant primary progressive aphasia (PPA) (svPPA) and are spared in logopenic variant (lvPPA) and nonfluent agrammatic primary progressive aphasia (nfaPPA) at disease onset. Over time, however, there may be heterogeneity in these abilities in lvPPA and nfaPPA. We hypothesized that individuals with svPPA would demonstrate poorer performance on baseline spoken word recognition and object knowledge than those with lvPPA and nfaPPA) as documented in the literature, but that rates of decline over time on spoken word recognition and object knowledge would be similar in all 3 PPA variants because these become less distinguishable with disease progression.The aim of this study was to investigate longitudinal patterns of decline in spoken word recognition and object knowledge across PPA variants.Ninety-five individuals with PPA completed the Semantic Word Picture Matching and Semantic Associates tests at baseline to establish expected performance in these areas. Thirty-five individuals completed follow-up testing.The distributions of trichotomized mean rates of decline in object knowledge were similar for lvPPA and svPPA (P = .05). There were weak negative correlations between symptom duration and baseline scores on Semantic Word Picture Matching (r[37] = -0.399, P = .01), and baseline scores on Semantic Associates (r[37] = -0.394, P = .01) in lvPPA.Degradation of spoken word recognition and object knowledge occurs over time in lvPPA. Further investigation of the receptive language deficits in PPA is warranted to characterize language changes that lessen the distinctions between PPA variants with disease progression.
Collapse
Affiliation(s)
| | | | | | | | - Donna C. Tippett
- Department of Neurology
- Department of Otolaryngology–Head and Neck Surgery
- Department of Physical Medicine and Rehabilitation, Johns Hopkins University School of Medicine, Baltimore, MD
| |
Collapse
|
25
|
Luthra S, Li MYC, You H, Brodbeck C, Magnuson JS. Does signal reduction imply predictive coding in models of spoken word recognition? Psychon Bull Rev 2021; 28:1381-9. [PMID: 33852158 DOI: 10.3758/s13423-021-01924-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/24/2021] [Indexed: 12/29/2022]
Abstract
Pervasive behavioral and neural evidence for predictive processing has led to claims that language processing depends upon predictive coding. Formally, predictive coding is a computational mechanism where only deviations from top-down expectations are passed between levels of representation. In many cognitive neuroscience studies, a reduction of signal for expected inputs is taken as being diagnostic of predictive coding. In the present work, we show that despite not explicitly implementing prediction, the TRACE model of speech perception exhibits this putative hallmark of predictive coding, with reductions in total lexical activation, total lexical feedback, and total phoneme activation when the input conforms to expectations. These findings may indicate that interactive activation is functionally equivalent or approximant to predictive coding or that caution is warranted in interpreting neural signal reduction as diagnostic of predictive coding.
Collapse
|
26
|
Abstract
This study uses a response mouse-tracking paradigm to examine the role of sub-phonemic information in online lexical ambiguity resolution of continuous speech. We examine listeners' sensitivity to the sub-phonemic information that is specific to the ambiguous internal open juncture /s/-stop sequences in American English (e.g., "place kin" vs. "play skin"), that is, voice onset time (VOT) indicating different degrees of aspiration (e.g., long VOT for "kin" vs. short VOT for "skin") in connected speech contexts. A cross-splicing method was used to create two-word sequences (e.g., "place kin" or "play skin") with matching VOTs (long for "kin"; short for "skin") or mismatching VOTs (short for "kin"; long for "skin"). Participants (n = 20) heard the two-word sequences, while looking at computer displays with the second word in the left/right corner ("KIN" and "SKIN"). Then, listeners' click responses and mouse movement trajectories were recorded. Click responses show significant effects of VOT manipulation, while mouse trajectories do not. Our results show that stop-release information, whether temporal or spectral, can (mis)guide listeners' interpretation of the possible location of a word boundary between /s/ and a following stop, even when other aspects in the acoustic signal (e.g., duration of /s/) point to the alternative segmentation. Taken together, our results suggest that segmentation and lexical access are highly attuned to bottom-up phonetic information; our results have implications for a model of spoken language recognition with position-specific representations available at the prelexical level and also allude to the possibility that detailed phonetic information may be stored in the listeners' lexicons.
Collapse
Affiliation(s)
- Yoonjeong Lee
- Department of Head and Neck Surgery, University of California, USA
| | | | - Louis Goldstein
- Department of Linguistics, University of Southern California, USA
| |
Collapse
|
27
|
Abstract
The paper aims to examine how the acoustic input (the surface form) and the abstract linguistic representation (the underlying representation) interact during spoken word recognition by investigating left-dominant tone sandhi, a tonal alternation in which the underlying tone of the first syllable spreads to the sandhi domain. We conducted two auditory-auditory priming lexical decision experiments on Shanghai left-dominant sandhi words with less-frequent and frequent Shanghai users, in which each disyllabic target was preceded by monosyllabic primes either sharing the same underlying tone, surface tone, or being unrelated to the tone of the first syllable of the sandhi targets. Results showed a surface priming effect but not an underlying priming effect in younger speakers who used Shanghai less frequently, but no surface or underlying priming effect in older speakers who used Shanghai more often. Moreover, the surface priming did not interact with speakers' familiarity ratings to the sandhi targets. These patterns suggest that left-dominant Shanghai sandhi words may be represented in the sandhi form in the mental lexicon. The results are discussed in the context of how phonological opacity, productivity, the non-structure-preserving nature of tone spreading, and speakers' semantic knowledge influence the representation and processing of tone sandhi words.
Collapse
Affiliation(s)
- Hanbo Yan
- School of Chinese Studies and Exchange, Shanghai International Studies University, China
| | - Yu-Fu Chien
- Department of Chinese Language and Literature, Fudan University, China
| | - Jie Zhang
- Department of Linguistics, University of Kansas, USA
| |
Collapse
|
28
|
Chien YF, Yang X, Fiorentino R, Sereno JA. The Role of Surface and Underlying Forms When Processing Tonal Alternations in Mandarin Chinese: A Mismatch Negativity Study. Front Psychol 2020; 11:646. [PMID: 32322230 PMCID: PMC7156642 DOI: 10.3389/fpsyg.2020.00646] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2019] [Accepted: 03/18/2020] [Indexed: 11/13/2022] Open
Abstract
Phonological alternation (sound change depending on the phonological environment) poses challenges to spoken word recognition models. Mandarin Chinese T3 sandhi is such a phenomenon in which a tone 3 (T3) changes into a tone 2 (T2) when followed by another T3. In a mismatch negativity (MMN) study examining Mandarin Chinese T3 sandhi, participants passively listened to either a T2 word [tʂu2 je4] /tʂu2 je4/, a T3 word [tʂu3 je4] /tʂu3 je4/, a sandhi word [tʂu2 jen3] /tʂu3 jen3/, or a mix of T3 and sandhi word standards. The deviant in each condition was a T2 word [tʂu2]. Results showed an MMN only in the T2 and T3 conditions but not in the Sandhi or Mix conditions. All conditions also yielded omission MMNs. This pattern cannot be explained based on the surface forms of standards and deviants; rather these data suggest an underspecified or underlying T3 stored linguistic representation used in spoken word processing.
Collapse
Affiliation(s)
- Yu-Fu Chien
- Department of Chinese Language and Literature, Fudan University, Shanghai, China.,Department of Modern Languages, DePaul University, Chicago, IL, United States
| | - Xiao Yang
- Department of Linguistics, University of Kansas, Lawrence, KS, United States
| | - Robert Fiorentino
- Department of Linguistics, University of Kansas, Lawrence, KS, United States
| | - Joan A Sereno
- Department of Linguistics, University of Kansas, Lawrence, KS, United States
| |
Collapse
|
29
|
Strand JF, Brown VA, Barbour DL. Talking Points: A Modulating Circle Increases Listening Effort Without Improving Speech Recognition in Young Adults. Psychon Bull Rev 2020; 27:536-43. [PMID: 32128719 DOI: 10.3758/s13423-020-01713-y] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Speech recognition is improved when the acoustic input is accompanied by visual cues provided by a talking face (Erber in Journal of Speech and Hearing Research, 12(2), 423-425, 1969; Sumby & Pollack in The Journal of the Acoustical Society of America, 26(2), 212-215, 1954). One way that the visual signal facilitates speech recognition is by providing the listener with information about fine phonetic detail that complements information from the auditory signal. However, given that degraded face stimuli can still improve speech recognition accuracy (Munhall, Kroos, Jozan, & Vatikiotis-Bateson in Perception & Psychophysics, 66(4), 574-583, 2004), and static or moving shapes can improve speech detection accuracy (Bernstein, Auer, & Takayanagi in Speech Communication, 44(1-4), 5-18, 2004), aspects of the visual signal other than fine phonetic detail may also contribute to the perception of speech. In two experiments, we show that a modulating circle providing information about the onset, offset, and acoustic amplitude envelope of the speech does not improve recognition of spoken sentences (Experiment 1) or words (Experiment 2). Further, contrary to our hypothesis, the modulating circle increased listening effort despite subjective reports that it made the word recognition task seem easier to complete (Experiment 2). These results suggest that audiovisual speech processing, even when the visual stimulus only conveys temporal information about the acoustic signal, may be a cognitively demanding process.
Collapse
|
30
|
Wiener S, Lee CY. Multi-Talker Speech Promotes Greater Knowledge-Based Spoken Mandarin Word Recognition in First and Second Language Listeners. Front Psychol 2020; 11:214. [PMID: 32161560 PMCID: PMC7052525 DOI: 10.3389/fpsyg.2020.00214] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2019] [Accepted: 01/30/2020] [Indexed: 11/15/2022] Open
Abstract
Spoken word recognition involves a perceptual tradeoff between the reliance on the incoming acoustic signal and knowledge about likely sound categories and their co-occurrences as words. This study examined how adult second language (L2) learners navigate between acoustic-based and knowledge-based spoken word recognition when listening to highly variable, multi-talker truncated speech, and whether this perceptual tradeoff changes as L2 listeners gradually become more proficient in their L2 after multiple months of structured classroom learning. First language (L1) Mandarin Chinese listeners and L1 English-L2 Mandarin adult listeners took part in a gating experiment. The L2 listeners were tested twice - once at the start of their intermediate/advanced L2 language class and again 2 months later. L1 listeners were only tested once. Participants were asked to identify syllable-tone words that varied in syllable token frequency (high/low according to a spoken word corpus) and syllable-conditioned tonal probability (most probable/least probable in speech given the syllable). The stimuli were recorded by 16 different talkers and presented at eight gates ranging from onset-only (gate 1) through onset +40 ms increments (gates 2 through 7) to the full word (gate 8). Mixed-effects regression modeling was used to compare performance to our previous study which used single-talker stimuli (Wiener et al., 2019). The results indicated that multi-talker speech caused both L1 and L2 listeners to rely greater on knowledge-based processing of tone. L1 listeners were able to draw on distributional knowledge of syllable-tone probabilities in early gates and switch to predominantly acoustic-based processing when more of the signal was available. In contrast, L2 listeners, with their limited experience with talker range normalization, were less able to effectively transition from probability-based to acoustic-based processing. Moreover, for the L2 listeners, the reliance on such distributional information for spoken word recognition appeared to be conditioned by the nature of the acoustic signal. Single-talker speech did not result in the same pattern of probability-based tone processing, suggesting that knowledge-based processing of L2 speech may only occur under certain acoustic conditions, such as multi-talker speech.
Collapse
Affiliation(s)
- Seth Wiener
- Language Acquisition, Processing and Pedagogy Lab, Department of Modern Languages, Carnegie Mellon University, Pittsburgh, PA, United States
| | - Chao-Yang Lee
- Speech Processing Lab, Communication Sciences and Disorders, Ohio University, Athens, OH, United States
| |
Collapse
|
31
|
Guang C, Lefkowitz E, Dillman-Hasso N, Brown VA, Strand JF. Recall of Speech is Impaired by Subsequent Masking Noise: A Replication of Experiment 2. ACTA ACUST UNITED AC 2020; 3:158-167. [PMID: 34240010 DOI: 10.1080/25742442.2021.1896908] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
Abstract
Introduction The presence of masking noise can impair speech intelligibility and increase the attentional and cognitive resources necessary to understand speech. The first study to demonstrate the negative cognitive effects of noisy speech found that participants had poorer recall for aurally-presented digits early in a list when later digits were presented in noise relative to quiet (Rabbitt, 1968). However, despite being cited nearly 500 times and providing the foundation for a wealth of subsequent research on the topic, the original study has never been directly replicated. Methods This study replicated Rabbitt (1968) with a large online sample and tested its robustness to a variety of analytical and scoring techniques. Results We replicated Rabbitt's key finding that listening to speech in noise impairs recall for items that came earlier in the list. The results were consistent when we used the original analytical technique (an ANOVA) and a more powerful analytical technique (generalized linear mixed effects models) that was not available when the original paper was published. Discussion These findings support the claim that effortful listening can interfere with encoding or rehearsal of previously presented information.
Collapse
Affiliation(s)
| | | | | | - Violet A Brown
- Washington University in St. Louis, Department of Psychological and Brain Sciences
| | | |
Collapse
|
32
|
Abstract
A long-term priming experiment examined the way stress information is processed and represented in French speakers' mind. Repeated prime and target words either matched (/bã'do/ - /bã'do/ "headband") or mismatched their stress pattern (/bãdo/ - /bã'do/). In comparison to a control condition (/maʁ[Formula: see text]/ - /bã'do/), the results showed that matching and mismatching primes were equally effective in facilitating the processing of the target words. Thus, despite the fact that French speakers routinely produce and hear words in their stressed and unstressed versions, this study suggests that stress in French is not integrated into lexical representations.
Collapse
Affiliation(s)
- Amandine Michelas
- Laboratoire Parole et Langage, CNRS & Aix-Marseille University, Aix-en-Provence, France
| | - Sophie Dufour
- Laboratoire Parole et Langage, CNRS & Aix-Marseille University, Aix-en-Provence, France
| |
Collapse
|
33
|
Zou L, Packard JL, Xia Z, Liu Y, Shu H. Morphological and Whole-Word Semantic Processing Are Distinct: Event Related Potentials Evidence From Spoken Word Recognition in Chinese. Front Hum Neurosci 2019; 13:133. [PMID: 31057382 PMCID: PMC6478770 DOI: 10.3389/fnhum.2019.00133] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2018] [Accepted: 04/03/2019] [Indexed: 01/23/2023] Open
Abstract
Behavioral and imaging studies in alphabetic languages have shown that morphological processing is a discrete and independent element of lexical access. However, there is no explicit marker of morphological structure in Chinese complex words, such that the extent to which morpheme meaning is related to word meaning is unknown. Event-related potentials (ERPs) were used in the present study to investigate the dissociation of morphemic and whole-word meaning in an auditory-auditory primed lexical decision task. All the prime and target words are compounds consisting of two Chinese morphemes. The relationship between morpheme and whole-word meaning was manipulated while controlling the phonology and orthography of the first syllable in each prime-target pair. A clear dissociation was found between morphemic and whole-word meaning on N400 amplitude and topography. Specifically, sharing a morpheme produced a larger N400 in the anterior-central electrode sites, while sharing whole-word meaning produced a smaller N400 in central-posterior electrode sites. In addition, the morphological N400 effect was negatively correlated with the participants' reading ability, with better readers needing less orthographic information to distinguish different morphemes in compound words. These findings indicate that morphological and whole-word meaning are dissociated in spoken Chinese compound word recognition and that even in the spoken language modality, good readers are better able to access the meaning of individual morphemes in Chinese compound word processing.
Collapse
Affiliation(s)
- Lijuan Zou
- School of Psychology and Education, Zaozhuang University, Zaozhuang, China
| | - Jerome L. Packard
- Beckman Institute, University of Illinois, Urbana, IL, United States
| | - Zhichao Xia
- State Key Laboratory of Cognitive Neuroscience and Learning & IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China
- Center for Collaboration and Innovation in Brain and Learning Sciences, Beijing Normal University, Beijing, China
| | - Youyi Liu
- State Key Laboratory of Cognitive Neuroscience and Learning & IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China
- Center for Collaboration and Innovation in Brain and Learning Sciences, Beijing Normal University, Beijing, China
| | - Hua Shu
- State Key Laboratory of Cognitive Neuroscience and Learning & IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China
| |
Collapse
|
34
|
Warner N. Reduced speech: All is variability. Wiley Interdiscip Rev Cogn Sci 2019; 10:e1496. [PMID: 30811095 DOI: 10.1002/wcs.1496] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/02/2018] [Revised: 01/25/2019] [Accepted: 01/26/2019] [Indexed: 11/12/2022]
Abstract
Humans often communicate through reduced speech, where many sounds are altered or deleted. Even whole syllables or occasionally whole words can be deleted, as in the production [th jyth ɛ̃m] for do you have time shown in the visual abstract. Such reduction is more common in spontaneous, casual, conversational speech, but it occurs in connected speech in formal settings, and rarely even in careful word-list reading. Such speech poses challenges for phonetic transcription, since it contains many sounds that do not seem to be any particular sound of the language and that are difficult to classify into categories. I argue this is because reduction operates largely outside the phonological system of the language. Reduced speech also poses questions about how listeners recognize the many possible forms of words, since speech seems to be made up entirely of variability. This article is categorized under: Linguistics > Linguistic Theory Linguistics > Language in Mind and Brain Psychology > Perception and Psychophysics.
Collapse
Affiliation(s)
- Natasha Warner
- Department of Linguistics, University of Arizona, Tucson, Arizona
| |
Collapse
|
35
|
Abstract
BACKGROUND Deficits in fluent language production are a hallmark of aphasia and may arise from impairments at different levels in the language system. It has been proposed that difficulty resolving lexical competition contributes to fluency deficits. AIMS The present study tested this hypothesis in a novel way: by examining whether narrative speech production fluency is associated with difficulty resolving lexical competition in spoken word recognition as measured by sensitivity to phonological neighborhood density. METHODS & PROCEDURES Nineteen participants with aphasia and 15 neurologically intact older adults identified spoken words that varied in phonological neighborhood density and were presented in moderate noise. OUTCOMES & RESULTS Neurologically intact participants exhibited the standard inhibitory effect of phonological neighborhood density on response times: slower recognition of spoken words from denser neighborhoods. Among participants with aphasia, the inhibitory effect of phonological neighborhood density (less accurate recognition of spoken words from denser neighborhoods) was smaller for participants with greater fluency. The neighborhood effect was larger for participants with greater receptive vocabulary knowledge, indicating that the fluency effect was not a result of general lexical deficits. CONCLUSIONS These results are consistent with the hypothesis that impaired lexical selection is a contributing factor in fluency deficits in post-stroke aphasia.
Collapse
Affiliation(s)
- Mona Roxana Botezatu
- Department of Communication Science and Disorders, University of Missouri, Columbia, MO, 65211, USA, ,
| | - Daniel Mirman
- Department of Psychology, University of Alabama, Birmingham, AL, 35294, USA, ;
- Moss Rehabilitation Research Institute, Elkins Park, PA, 19027, USA
| |
Collapse
|
36
|
Abstract
The current study investigated the phonetic adjustment mechanisms that underlie perceptual adaptation in first and second language (Dutch-English) listeners by exposing them to a novel English accent containing controlled deviations from the standard accent (e.g. /i/-to-/ɪ/ yielding /krɪm/ instead of /krim/ for 'cream'). These deviations involved contrasts that either were contrastive or were not contrastive in Dutch. Following accent exposure with disambiguating feedback, listeners completed lexical decision and word identification tasks. Both native and second language listeners demonstrated adaptation, evidenced by higher lexical endorsement rates and word identification accuracy than untrained control listeners for items containing trained accent patterns. However, for L2 listeners, adaptation was modulated by the phonemic contrast, that is, whether or not it was contrastive in the listeners' native language. Specifically, the training-induced criterion loosening for the L2 listeners was limited to contrasts that exist in both their L1, Dutch, and L2, English. For contrasts that are either absent or neutralized in Dutch, the L2 listeners demonstrated relatively loose pre-training criteria compared to L1 listeners. The results indicate that accent exposure induces both a general increase in tolerance for atypical speech input as well as targeted adjustments to specific categories for both L1 and L2 listeners.
Collapse
Affiliation(s)
- Angela Cooper
- Correspondence to: Angela Cooper, Department of Psychology, University of Toronto, 3359 Mississauga Rd., Mississauga, ON, L5L 1C6. Tel: 647-774-8967.,
| | | |
Collapse
|
37
|
Magnuson JS, Mirman D, Luthra S, Strauss T, Harris HD. Interaction in Spoken Word Recognition Models: Feedback Helps. Front Psychol 2018; 9:369. [PMID: 29666593 PMCID: PMC5891609 DOI: 10.3389/fpsyg.2018.00369] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2017] [Accepted: 03/06/2018] [Indexed: 11/13/2022] Open
Abstract
Human perception, cognition, and action requires fast integration of bottom-up signals with top-down knowledge and context. A key theoretical perspective in cognitive science is the interactive activation hypothesis: forward and backward flow in bidirectionally connected neural networks allows humans and other biological systems to approximate optimal integration of bottom-up and top-down information under real-world constraints. An alternative view is that online feedback is neither necessary nor helpful; purely feed forward alternatives can be constructed for any feedback system, and online feedback could not improve processing and would preclude veridical perception. In the domain of spoken word recognition, the latter view was apparently supported by simulations using the interactive activation model, TRACE, with and without feedback: as many words were recognized more quickly without feedback as were recognized faster with feedback, However, these simulations used only a small set of words and did not address a primary motivation for interaction: making a model robust in noise. We conducted simulations using hundreds of words, and found that the majority were recognized more quickly with feedback than without. More importantly, as we added noise to inputs, accuracy and recognition times were better with feedback than without. We follow these simulations with a critical review of recent arguments that online feedback in interactive activation models like TRACE is distinct from other potentially helpful forms of feedback. We conclude that in addition to providing the benefits demonstrated in our simulations, online feedback provides a plausible means of implementing putatively distinct forms of feedback, supporting the interactive activation hypothesis.
Collapse
Affiliation(s)
- James S. Magnuson
- Department of Psychological Sciences, University of Connecticut, Storrs, CT, United States
- Connecticut Institute for the Brain and Cognitive Sciences, University of Connecticut, Storrs, CT, United States
| | - Daniel Mirman
- Department of Psychology, University of Alabama at Birmingham, Birmingham, AL, United States
| | - Sahil Luthra
- Department of Psychological Sciences, University of Connecticut, Storrs, CT, United States
- Connecticut Institute for the Brain and Cognitive Sciences, University of Connecticut, Storrs, CT, United States
| | - Ted Strauss
- McConnell Brain Imaging Centre, McGill University, Montreal, QC, Canada
| | - Harlan D. Harris
- Department of Psychological Sciences, University of Connecticut, Storrs, CT, United States
- Connecticut Institute for the Brain and Cognitive Sciences, University of Connecticut, Storrs, CT, United States
| |
Collapse
|
38
|
Xie X, Myers EB. Learning a Talker or Learning an Accent: Acoustic Similarity Constrains Generalization of Foreign Accent Adaptation to New Talkers. J Mem Lang 2017; 97:30-46. [PMID: 28890602 PMCID: PMC5589144 DOI: 10.1016/j.jml.2017.07.005] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Past research has revealed that native listeners use top-down information to adjust the mapping from speech sounds to phonetic categories. Such phonetic adjustments help listeners adapt to foreign-accented speech. However, the mechanism by which talker-specific adaptation generalizes to other talkers is poorly understood. Here we asked what conditions induce crosstalker generalization in talker accent adaptation. Native-English listeners were exposed to Mandarin-accented words, produced by a single talker or multiple talkers. Following exposure, adaptation to the accent was tested by recognition of novel words in a task that assesses online lexical access. Crucially, test words were novel words and were produced by a novel Mandarin-accented talker. Results indicated that regardless of exposure condition (single or multiple talker exposure), generalization was greatest when the talkers were acoustically similar to one another, suggesting that listeners were not developing an accent-wide schema for Mandarin talkers, but rather attuning to the specific acoustic-phonetic properties of the talkers. Implications for general mechanisms of talker generalization in speech adaptation are discussed.
Collapse
|
39
|
Abstract
The present study examined how the network science measure known as closeness centrality (which measures the average distance between a node and all other nodes in the network) influences lexical processing. In the mental lexicon, a word such as CAN has high closeness centrality, because it is close to many other words in the lexicon. Whereas, a word such as CURE has low closeness centrality because it is far from other words in the lexicon. In an auditory lexical decision task (Experiment 1) participants responded more quickly to words with high closeness centrality. In Experiment 2 an auditory lexical decision task was again used, but with a wider range of stimulus characteristics. Although, there was no main effect of closeness centrality in Experiment 2, an interaction between closeness centrality and frequency of occurrence was observed on reaction times. The results are explained in terms of partial activation gradually strengthening over time word-forms that are centrally located in the phonological network.
Collapse
|
40
|
Veivo O, Järvikivi J, Porretta V, Hyönä J. Orthographic Activation in L2 Spoken Word Recognition Depends on Proficiency: Evidence from Eye-Tracking. Front Psychol 2016; 7:1120. [PMID: 27512381 PMCID: PMC4961715 DOI: 10.3389/fpsyg.2016.01120] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2016] [Accepted: 07/12/2016] [Indexed: 12/02/2022] Open
Abstract
The use of orthographic and phonological information in spoken word recognition was studied in a visual world task where L1 Finnish learners of L2 French (n = 64) and L1 French native speakers (n = 24) were asked to match spoken word forms with printed words while their eye movements were recorded. In Experiment 1, French target words were contrasted with competitors having a longer ( vs. ) or a shorter word initial phonological overlap ( vs. ) and an identical orthographic overlap. In Experiment 2, target words were contrasted with competitors of either longer ( vs. ) or shorter word initial orthographic overlap ( vs. ) and of an identical phonological overlap. A general phonological effect was observed in the L2 listener group but not in the L1 control group. No general orthographic effects were observed in the L2 or L1 groups, but a significant effect of proficiency was observed for orthographic overlap over time: higher proficiency L2 listeners used also orthographic information in the matching task in a time-window from 400 to 700 ms, whereas no such effect was observed for lower proficiency listeners. These results suggest that the activation of orthographic information in L2 spoken word recognition depends on proficiency in L2.
Collapse
Affiliation(s)
- Outi Veivo
- Department of French, University of TurkuTurku, Finland
| | - Juhani Järvikivi
- Department of Linguistics, University of AlbertaEdmonton, Alberta, Canada
| | - Vincent Porretta
- Department of Linguistics, University of AlbertaEdmonton, Alberta, Canada
| | - Jukka Hyönä
- Department of Psychology, University of TurkuTurku, Finland
| |
Collapse
|
41
|
Goh WD, Yap MJ, Lau MC, Ng MMR, Tan LC. Semantic Richness Effects in Spoken Word Recognition: A Lexical Decision and Semantic Categorization Megastudy. Front Psychol 2016; 7:976. [PMID: 27445936 PMCID: PMC4923159 DOI: 10.3389/fpsyg.2016.00976] [Citation(s) in RCA: 32] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2016] [Accepted: 06/13/2016] [Indexed: 11/30/2022] Open
Abstract
A large number of studies have demonstrated that semantic richness dimensions [e.g., number of features, semantic neighborhood density, semantic diversity , concreteness, emotional valence] influence word recognition processes. Some of these richness effects appear to be task-general, while others have been found to vary across tasks. Importantly, almost all of these findings have been found in the visual word recognition literature. To address this gap, we examined the extent to which these semantic richness effects are also found in spoken word recognition, using a megastudy approach that allows for an examination of the relative contribution of the various semantic properties to performance in two tasks: lexical decision, and semantic categorization. The results show that concreteness, valence, and number of features accounted for unique variance in latencies across both tasks in a similar direction—faster responses for spoken words that were concrete, emotionally valenced, and with a high number of features—while arousal, semantic neighborhood density, and semantic diversity did not influence latencies. Implications for spoken word recognition processes are discussed.
Collapse
Affiliation(s)
- Winston D Goh
- Department of Psychology, National University of Singapore Singapore, Singapore
| | - Melvin J Yap
- Department of Psychology, National University of Singapore Singapore, Singapore
| | - Mabel C Lau
- Department of Psychology, National University of Singapore Singapore, Singapore
| | - Melvin M R Ng
- Department of Psychology, National University of Singapore Singapore, Singapore
| | - Luuan-Chin Tan
- Department of Psychology, National University of Singapore Singapore, Singapore
| |
Collapse
|
42
|
Abstract
Deficits in visual disengagement are one of the earliest emerging differences in infants who are later diagnosed with autism spectrum disorder. Although researchers have speculated that deficits in visual disengagement could have negative effects on the development of children with autism spectrum disorder, we do not know which skills are disrupted or how this disruption takes place. As a first step in understanding this issue, this study investigated the relationship between visual disengagement and a critical skill in early language development: spoken word recognition. Participants were 18 children with autism spectrum disorder (aged 4-7 years). Consistent with our predictions, children with poorer visual disengagement were slower and less accurate to process familiar words; disengagement explained over half of the variance in spoken word recognition. Visual disengagement remained uniquely associated with spoken word recognition after accounting for children's vocabulary size and age. These findings align with a recently proposed developmental model in which poor visual disengagement decreases the speed and accuracy of real-time spoken word recognition in children with autism spectrum disorder-which, in turn, may negatively affect their language development.
Collapse
|
43
|
Abstract
Research has shown that syllables play a relevant role in lexical access in Spanish, a shallow language with a transparent syllabic structure. Syllable frequency has been shown to have an inhibitory effect on visual word recognition in Spanish. However, no study has examined the syllable frequency effect on spoken word recognition. The present study tested the effect of the frequency of the first syllable on recognition of spoken Spanish words. A sample of 45 young adults (33 women, 12 men; M = 20.4, SD = 2.8; college students) performed an auditory lexical decision on 128 Spanish disyllabic words and 128 disyllabic nonwords. Words were selected so that lexical and first syllable frequency were manipulated in a within-subject 2 × 2 design, and six additional independent variables were controlled: token positional frequency of the second syllable, number of phonemes, position of lexical stress, number of phonological neighbors, number of phonological neighbors that have higher frequencies than the word, and acoustical durations measured in milliseconds. Decision latencies and error rates were submitted to linear mixed models analysis. Results showed a typical facilitatory effect of the lexical frequency and, importantly, an inhibitory effect of the first syllable frequency on reaction times and error rates.
Collapse
Affiliation(s)
- Julio González-Alvarez
- Department of Psicología Básica, Clínica y Psicobiología, University Jaume I, Castellón, Spain
| | | |
Collapse
|
44
|
Schmidtke J. The Bilingual Disadvantage in Speech Understanding in Noise Is Likely a Frequency Effect Related to Reduced Language Exposure. Front Psychol 2016; 7:678. [PMID: 27242592 PMCID: PMC4865492 DOI: 10.3389/fpsyg.2016.00678] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2016] [Accepted: 04/22/2016] [Indexed: 12/03/2022] Open
Abstract
The present study sought to explain why bilingual speakers are disadvantaged relative to monolingual speakers when it comes to speech understanding in noise. Exemplar models of the mental lexicon hold that each encounter with a word leaves a memory trace in long-term memory. Words that we encounter frequently will be associated with richer phonetic representations in memory and therefore recognized faster and more accurately than less frequently encountered words. Because bilinguals are exposed to each of their languages less often than monolinguals by virtue of speaking two languages, they encounter all words less frequently and may therefore have poorer phonetic representations of all words compared to monolinguals. In the present study, vocabulary size was taken as an estimate for language exposure and the prediction was made that both vocabulary size and word frequency would be associated with recognition accuracy for words presented in noise. Forty-eight early Spanish–English bilingual and 53 monolingual English young adults were tested on speech understanding in noise (SUN) ability, English oral verbal ability, verbal working memory (WM), and auditory attention. Results showed that, as a group, monolinguals recognized significantly more words than bilinguals. However, this effect was attenuated by language proficiency; higher proficiency was associated with higher accuracy on the SUN test in both groups. This suggests that greater language exposure is associated with better SUN. Word frequency modulated recognition accuracy and the difference between groups was largest for low frequency words, suggesting that the bilinguals’ insufficient exposure to these words hampered recognition. The effect of WM was not significant, likely because of its large shared variance with language proficiency. The effect of auditory attention was small but significant. These results are discussed within the Ease of Language Understanding model (Rönnberg et al., 2013), which provides a framework for explaining individual differences in SUN.
Collapse
|
45
|
Schneider BA, Avivi-Reich M, Leung C, Heinrich A. How Age and Linguistic Competence Affect Memory for Heard Information. Front Psychol 2016; 7:618. [PMID: 27242569 PMCID: PMC4860395 DOI: 10.3389/fpsyg.2016.00618] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2015] [Accepted: 04/13/2016] [Indexed: 11/13/2022] Open
Abstract
The short-term memory performance of a group of younger adults, for whom English was a second language (young EL2 listeners), was compared to that of younger and older adults for whom English was their first language (EL1 listeners). To-be-remembered words were presented in noise and in quiet. When presented in noise, the listening situation was adjusted to ensure that the likelihood of recognizing the individual words was comparable for all groups. Previous studies which used the same paradigm found memory performance of older EL1 adults on this paired-associate task to be poorer than that of their younger EL1 counterparts both in quiet and in a background of babble. The purpose of the present study was to investigate whether the less well-established semantic and linguistic skills of EL2 listeners would also lead to memory deficits even after equating for word recognition as was done for the younger and older EL1 listeners. No significant differences in memory performance were found between young EL1 and EL2 listeners after equating for word recognition, indicating that the EL2 listeners' poorer semantic and linguistic skills had little effect on their ability to memorize and recall paired associates. This result is consistent with the hypothesis that age-related declines in memory are primarily due to age-related declines in higher-order processes supporting stream segregation and episodic memory. Such declines are likely to increase the load on higher-order (possibly limited) cognitive processes supporting memory. The problems that these results pose for the comprehension of spoken language in these three groups are discussed.
Collapse
Affiliation(s)
- Bruce A Schneider
- Human Communication Laboratory, Psychology, University of Toronto Mississauga Mississauga, ON, Canada
| | | | - Caterina Leung
- Human Communication Laboratory, Psychology, University of Toronto Mississauga Mississauga, ON, Canada
| | - Antje Heinrich
- Medical Research Council Institute of Hearing Research Nottingham, UK
| |
Collapse
|
46
|
Abstract
Language learning is generally described as a problem of acquiring new information (e.g., new words). However, equally important are changes in how the system processes known information. For example, a wealth of studies has suggested dramatic changes over development in how efficiently children recognize familiar words, but it is unknown what kind of experience-dependent mechanisms of plasticity give rise to such changes in real-time processing. We examined the plasticity of the language processing system by testing whether a fundamental aspect of spoken word recognition, lexical interference, can be altered by experience. Adult participants were trained on a set of familiar words over a series of 4 tasks. In the high-competition (HC) condition, tasks were designed to encourage coactivation of similar words (e.g., net and neck) and to require listeners to resolve this competition. Tasks were similar in the low-competition (LC) condition, but did not enhance this competition. Immediately after training, interlexical interference was tested using a visual world paradigm task. Participants in the HC group resolved interference to a fuller degree than those in the LC group, demonstrating that experience can shape the way competition between words is resolved. TRACE simulations showed that the observed late differences in the pattern of interference resolution can be attributed to differences in the strength of lexical inhibition. These findings inform cognitive models in many domains that involve competition/interference processes, and suggest an experience-dependent mechanism of plasticity that may underlie longer term changes in processing efficiency associated with both typical and atypical development.
Collapse
Affiliation(s)
- Efthymia C Kapnoula
- Department of Psychological and Brain Sciences, Delta Center, University of Iowa
| | - Bob McMurray
- Department of Psychological and Brain Sciences, Delta Center, University of Iowa
| |
Collapse
|
47
|
Abstract
Top-down contextual influences play a major part in speech understanding, especially in hearing-impaired patients with deteriorated auditory input. Those influences are most obvious in difficult listening situations, such as listening to sentences in noise but can also be observed at the word level under more favorable conditions, as in one of the most commonly used tasks in audiology, i.e., repeating isolated words in silence. This study aimed to explore the role of top-down contextual influences and their dependence on lexical factors and patient-specific factors using standard clinical linguistic material. Spondaic word perception was tested in 160 hearing-impaired patients aged 23-88 years with a four-frequency average pure-tone threshold ranging from 21 to 88 dB HL. Sixty spondaic words were randomly presented at a level adjusted to correspond to a speech perception score ranging between 40 and 70% of the performance intensity function obtained using monosyllabic words. Phoneme and whole-word recognition scores were used to calculate two context-influence indices (the j factor and the ratio of word scores to phonemic scores) and were correlated with linguistic factors, such as the phonological neighborhood density and several indices of word occurrence frequencies. Contextual influence was greater for spondaic words than in similar studies using monosyllabic words, with an overall j factor of 2.07 (SD = 0.5). For both indices, context use decreased with increasing hearing loss once the average hearing loss exceeded 55 dB HL. In right-handed patients, significantly greater context influence was observed for words presented in the right ears than for words presented in the left, especially in patients with many years of education. The correlations between raw word scores (and context influence indices) and word occurrence frequencies showed a significant age-dependent effect, with a stronger correlation between perception scores and word occurrence frequencies when the occurrence frequencies were based on the years corresponding to the patients' youth, showing a "historic" word frequency effect. This effect was still observed for patients with few years of formal education, but recent occurrence frequencies based on current word exposure had a stronger influence for those patients, especially for younger ones.
Collapse
Affiliation(s)
- Annie Moulin
- INSERM, U1028, Lyon Neuroscience Research Center, Brain Dynamics and Cognition TeamLyon, France
- CNRS, UMR5292, Lyon Neuroscience Research Center, Brain Dynamics and Cognition TeamLyon, France
- University of LyonLyon, France
| | - Céline Richard
- Otorhinolaryngology Department, Vaudois University Hospital Center and University of LausanneLausanne, Switzerland
- The Laboratory for Investigative Neurophysiology, Department of Radiology and Department of Clinical Neurosciences, Vaudois University Hospital Center and University of LausanneLausanne, Switzerland
| |
Collapse
|
48
|
Abstract
This study investigated the developmental time course of spoken word recognition in older children using eye tracking to assess how the real-time processing dynamics of word recognition change over development. We found that 9-year-olds were slower to activate the target words and showed more early competition from competitor words than 16-year-olds; however, both age groups ultimately fixated targets to the same degree. This contrasts with a prior study of adolescents with language impairment (McMurray, Samelson, Lee, & Tomblin, 2010) that showed a different pattern of real-time processes. These findings suggest that the dynamics of word recognition are still developing even at these late ages, and developmental changes may derive from different sources than individual differences in relative language ability.
Collapse
Affiliation(s)
- Hannah Rigler
- Department of Psychological and Brain Sciences, University of Iowa
| | | | - Lea Greiner
- Department of Psychological and Brain Sciences, University of Iowa
| | - Jessica Walker
- Department of Psychological and Brain Sciences, University of Iowa
| | - J Bruce Tomblin
- Department of Psychological and Brain Sciences, University of Iowa
| | - Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa
| |
Collapse
|
49
|
Riordan B, Dye M, Jones MN. Grammatical number processing and anticipatory eye movements are not tightly coordinated in English spoken language comprehension. Front Psychol 2015; 6:590. [PMID: 25999900 PMCID: PMC4423439 DOI: 10.3389/fpsyg.2015.00590] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2015] [Accepted: 04/21/2015] [Indexed: 11/30/2022] Open
Abstract
Recent studies of eye movements in world-situated language comprehension have demonstrated that rapid processing of morphosyntactic information – e.g., grammatical gender and number marking – can produce anticipatory eye movements to referents in the visual scene. We investigated how type of morphosyntactic information and the goals of language users in comprehension affected eye movements, focusing on the processing of grammatical number morphology in English-speaking adults. Participants’ eye movements were recorded as they listened to simple English declarative (There are the lions.) and interrogative (Where are the lions?) sentences. In Experiment 1, no differences were observed in speed to fixate target referents when grammatical number information was informative relative to when it was not. The same result was obtained in a speeded task (Experiment 2) and in a task using mixed sentence types (Experiment 3). We conclude that grammatical number processing in English and eye movements to potential referents are not tightly coordinated. These results suggest limits on the role of predictive eye movements in concurrent linguistic and scene processing. We discuss how these results can inform and constrain predictive approaches to language processing.
Collapse
Affiliation(s)
| | - Melody Dye
- Department of Psychological and Brain Sciences, Indiana University Bloomington, IN, USA
| | - Michael N Jones
- Department of Psychological and Brain Sciences, Indiana University Bloomington, IN, USA
| |
Collapse
|
50
|
Toscano JC, McMurray B. The time-course of speaking rate compensation: Effects of sentential rate and vowel length on voicing judgments. Lang Cogn Neurosci 2015; 30:529-543. [PMID: 25780801 PMCID: PMC4358767 DOI: 10.1080/23273798.2014.946427] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
Many sources of context information in speech (such as speaking rate) occur either before or after the phonetic cues they influence, yet there is little work examining the time-course of these effects. Here, we investigate how listeners compensate for preceding sentence rate and subsequent vowel length (a secondary cue that has been used as a proxy for speaking rate) when categorizing words varying in voice-onset time (VOT). Participants selected visual objects in a display while their eye-movements were recorded, allowing us to examine when each source of information had an effect on lexical processing. We found that the effect of VOT preceded that of vowel length, suggesting that each cue is used as it becomes available. In a second experiment, we found that, in contrast, the effect of preceding sentence rate occurred simultaneously with VOT, suggesting that listeners interpret VOT relative to preceding rate.
Collapse
Affiliation(s)
- Joseph C Toscano
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, 405 N Mathews Ave, Urbana, IL 61801
| | - Bob McMurray
- Dept. of Psychology and Dept. of Communication Sciences & Disorders, University of Iowa, E11 Seashore Hall, Iowa City, IA 52242
| |
Collapse
|