1
|
Moberly AC, Du L, Tamati TN. Individual Differences in the Recognition of Spectrally Degraded Speech: Associations With Neurocognitive Functions in Adult Cochlear Implant Users and With Noise-Vocoded Simulations. Trends Hear 2025; 29:23312165241312449. [PMID: 39819389 PMCID: PMC11742172 DOI: 10.1177/23312165241312449] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2024] [Revised: 12/02/2024] [Accepted: 12/12/2024] [Indexed: 01/19/2025] Open
Abstract
When listening to speech under adverse conditions, listeners compensate using neurocognitive resources. A clinically relevant form of adverse listening is listening through a cochlear implant (CI), which provides a spectrally degraded signal. CI listening is often simulated through noise-vocoding. This study investigated the neurocognitive mechanisms supporting recognition of spectrally degraded speech in adult CI users and normal-hearing (NH) peers listening to noise-vocoded speech, with the hypothesis that an overlapping set of neurocognitive functions would contribute to speech recognition in both groups. Ninety-seven adults with either a CI (54 CI individuals, mean age 66.6 years, range 45-87 years) or age-normal hearing (43 NH individuals, mean age 66.8 years, range 50-81 years) participated. Listeners heard materials varying in linguistic complexity consisting of isolated words, meaningful sentences, anomalous sentences, high-variability sentences, and audiovisually (AV) presented sentences. Participants were also tested for vocabulary knowledge, nonverbal reasoning, working memory capacity, inhibition-concentration, and speed of lexical and phonological access. Linear regression analyses with robust standard errors were performed for speech recognition tasks on neurocognitive functions. Nonverbal reasoning contributed to meaningful sentence recognition in NH peers and anomalous sentence recognition in CI users. Speed of lexical access contributed to performance on most speech tasks for CI users but not for NH peers. Finally, inhibition-concentration and vocabulary knowledge contributed to AV sentence recognition in NH listeners alone. Findings suggest that the complexity of speech materials may determine the particular contributions of neurocognitive skills, and that NH processing of noise-vocoded speech may not represent how CI listeners process speech.
Collapse
Affiliation(s)
- Aaron C. Moberly
- Department of Otolaryngology – Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
| | - Liping Du
- Department of Biostatistics, Vanderbilt University Medical Center, Nashville, TN, USA
| | - Terrin N. Tamati
- Department of Otolaryngology – Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
| |
Collapse
|
2
|
Arras T, Rachman L, van Wieringen A, Başkent D. Perception of voice cues and speech-in-speech by children with prelingual single-sided deafness and a cochlear implant. Hear Res 2024; 454:109133. [PMID: 39546877 DOI: 10.1016/j.heares.2024.109133] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/16/2024] [Revised: 10/10/2024] [Accepted: 10/14/2024] [Indexed: 11/17/2024]
Abstract
Voice cues, such as fundamental frequency (F0) and vocal tract length (VTL), help listeners identify the speaker's gender, perceive the linguistic and emotional prosody, and segregate competing talkers. Postlingually implanted adult cochlear implant (CI) users seem to have difficulty in perceiving and making use of voice cues, especially of VTL. Early implanted child CI users, in contrast, perceive and make use of both voice cues better than CI adults, and in patterns similar to their peers with normal hearing (NH). In our study, we investigated the perception and use of voice cues in children with single-sided deafness (SSD) who received their CI at an early age (SSD+CI), in an attempt to bridge the gap between these two groups. The SSD+CI children have access to bilateral auditory information and often receive their CI at an early age, similar to CI children. They may also have dominant acoustic representations, similar to CI adults who acquired hearing loss at a later age. As such, the current study aimed to investigate the perception and use of voice cues by a group of nine early-implanted children with prelingual SSD. The study consisted of three experiments: F0 and VTL discrimination, voice gender categorization, and speech-in-speech perception. In each experiment, the results of the SSD group are compared to children and adults with CIs (for their CI ear) and with typical hearing (for their NH ear). Overall, the SSD+CI children had poorer VTL detection thresholds with their CI compared to their NH ear, while their F0 perception was similar across ears. Detection thresholds for both F0 and VTL with their CI ear was comparable to those of bilaterally implanted CI children, suggesting that SSD+CI children do not only rely on their NH ear, but actually make use of their CI. SSD+CI children relied more heavily on F0 cues than on VTL cues for voice gender categorization, with cue weighting patterns comparable to those of CI adults. In contrast to CI children, the SSD+CI children showed limited speech perception benefit based on F0 and VTL differences between the target and masker speaker, which again corresponded to the results of CI adults. Altogether, the SSD+CI children make good use of their CI, despite a good-hearing ear, however, the perceptual patterns seem to fall in-between those of CI children and CI adults. Perhaps a combination of childhood neuroplasticity, limited experience with relying only on the CI, and a dominant acoustic representation of voice gender explain these results.
Collapse
Affiliation(s)
- Tine Arras
- ExpORL, Dept. Neurosciences, KU Leuven, Belgium; Cochlear Technology Centre, Belgium
| | - Laura Rachman
- Dept. of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, The Netherlands; Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, The Netherlands; W.J. Kolff Institute for Biomedical Engineering and Materials Science, Graduate School of Medical Sciences, University of Groningen, The Netherlands
| | - Astrid van Wieringen
- ExpORL, Dept. Neurosciences, KU Leuven, Belgium; Dept. of Special Needs Education, University of Oslo, Norway
| | - Deniz Başkent
- Dept. of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, The Netherlands; Research School of Behavioral and Cognitive Neuroscience, Graduate School of Medical Sciences, University of Groningen, The Netherlands; W.J. Kolff Institute for Biomedical Engineering and Materials Science, Graduate School of Medical Sciences, University of Groningen, The Netherlands.
| |
Collapse
|
3
|
Nagels L, Gaudrain E, Vickers D, Hendriks P, Başkent D. Prelingually Deaf Children With Cochlear Implants Show Better Perception of Voice Cues and Speech in Competing Speech Than Postlingually Deaf Adults With Cochlear Implants. Ear Hear 2024; 45:952-968. [PMID: 38616318 PMCID: PMC11175806 DOI: 10.1097/aud.0000000000001489] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2022] [Accepted: 01/10/2024] [Indexed: 04/16/2024]
Abstract
OBJECTIVES Postlingually deaf adults with cochlear implants (CIs) have difficulties with perceiving differences in speakers' voice characteristics and benefit little from voice differences for the perception of speech in competing speech. However, not much is known yet about the perception and use of voice characteristics in prelingually deaf implanted children with CIs. Unlike CI adults, most CI children became deaf during the acquisition of language. Extensive neuroplastic changes during childhood could make CI children better at using the available acoustic cues than CI adults, or the lack of exposure to a normal acoustic speech signal could make it more difficult for them to learn which acoustic cues they should attend to. This study aimed to examine to what degree CI children can perceive voice cues and benefit from voice differences for perceiving speech in competing speech, comparing their abilities to those of normal-hearing (NH) children and CI adults. DESIGN CI children's voice cue discrimination (experiment 1), voice gender categorization (experiment 2), and benefit from target-masker voice differences for perceiving speech in competing speech (experiment 3) were examined in three experiments. The main focus was on the perception of mean fundamental frequency (F0) and vocal-tract length (VTL), the primary acoustic cues related to speakers' anatomy and perceived voice characteristics, such as voice gender. RESULTS CI children's F0 and VTL discrimination thresholds indicated lower sensitivity to differences compared with their NH-age-equivalent peers, but their mean discrimination thresholds of 5.92 semitones (st) for F0 and 4.10 st for VTL indicated higher sensitivity than postlingually deaf CI adults with mean thresholds of 9.19 st for F0 and 7.19 st for VTL. Furthermore, CI children's perceptual weighting of F0 and VTL cues for voice gender categorization closely resembled that of their NH-age-equivalent peers, in contrast with CI adults. Finally, CI children had more difficulties in perceiving speech in competing speech than their NH-age-equivalent peers, but they performed better than CI adults. Unlike CI adults, CI children showed a benefit from target-masker voice differences in F0 and VTL, similar to NH children. CONCLUSION Although CI children's F0 and VTL voice discrimination scores were overall lower than those of NH children, their weighting of F0 and VTL cues for voice gender categorization and their benefit from target-masker differences in F0 and VTL resembled that of NH children. Together, these results suggest that prelingually deaf implanted CI children can effectively utilize spectrotemporally degraded F0 and VTL cues for voice and speech perception, generally outperforming postlingually deaf CI adults in comparable tasks. These findings underscore the presence of F0 and VTL cues in the CI signal to a certain degree and suggest other factors contributing to the perception challenges faced by CI adults.
Collapse
Affiliation(s)
- Leanne Nagels
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
| | - Etienne Gaudrain
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
- CNRS UMR 5292, Lyon Neuroscience Research Center, Auditory Cognition and Psychoacoustics, Inserm UMRS 1028, Université Claude Bernard Lyon 1, Université de Lyon, Lyon, France
| | - Deborah Vickers
- Cambridge Hearing Group, Sound Lab, Clinical Neurosciences Department, University of Cambridge, Cambridge, United Kingdom
| | - Petra Hendriks
- Center for Language and Cognition Groningen (CLCG), University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, University of Groningen, Groningen, The Netherlands
- W.J. Kolff Institute for Biomedical Engineering and Materials Science, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
| |
Collapse
|
4
|
Saba JN, Ali H, Hansen JHL. The effects of estimation accuracy, estimation approach, and number of selected channels using formant-priority channel selection for an "n-of-m" sound processing strategy for cochlear implants. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:3100. [PMID: 37227411 PMCID: PMC10219683 DOI: 10.1121/10.0019416] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Revised: 04/16/2023] [Accepted: 04/28/2023] [Indexed: 05/26/2023]
Abstract
Previously, selection of l channels was prioritized according to formant frequency locations in an l-of-n-of-m-based signal processing strategy to provide important voicing information independent of listening environments for cochlear implant (CI) users. In this study, ideal, or ground truth, formants were incorporated into the selection stage to determine the effect of accuracy on (1) subjective speech intelligibility, (2) objective channel selection patterns, and (3) objective stimulation patterns (current). An average +11% improvement (p < 0.05) was observed across six CI users in quiet, but not for noise or reverberation conditions. Analogous increases in channel selection and current for the upper range of F1 and a decrease across mid-frequencies with higher corresponding current, were both observed at the expense of noise-dominant channels. Objective channel selection patterns were analyzed a second time to determine the effects of estimation approach and number of selected channels (n). A significant effect of estimation approach was only observed in the noise and reverberation condition with minor differences in channel selection and significantly decreased stimulated current. Results suggest that estimation method, accuracy, and number of channels in the proposed strategy using ideal formants may improve intelligibility when corresponding stimulated current of formant channels are not masked by noise-dominant channels.
Collapse
Affiliation(s)
- Juliana N Saba
- University of Texas at Dallas, Center for Robust Speech Systems, Cochlear Implant Laboratory, 800 W. Campbell Rd, EC 33, Richardson, Texas 75080, USA
| | - Hussnain Ali
- University of Texas at Dallas, Center for Robust Speech Systems, Cochlear Implant Laboratory, 800 W. Campbell Rd, EC 33, Richardson, Texas 75080, USA
| | - John H L Hansen
- University of Texas at Dallas, Center for Robust Speech Systems, Cochlear Implant Laboratory, 800 W. Campbell Rd, EC 33, Richardson, Texas 75080, USA
| |
Collapse
|
5
|
Haumann NT, Petersen B, Friis Andersen AS, Faulkner KF, Brattico E, Vuust P. Mismatch negativity as a marker of music perception in individual cochlear implant users: A spike density component analysis study. Clin Neurophysiol 2023; 148:76-92. [PMID: 36822119 DOI: 10.1016/j.clinph.2023.01.015] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2022] [Revised: 01/19/2023] [Accepted: 01/24/2023] [Indexed: 02/10/2023]
Abstract
OBJECTIVE Ninety percent of cochlear implant (CI) users are interested in improving their music perception. However, only few objective behavioral and neurophysiological tests have been developed for tracing the development of music discrimination skills in CI users. In this study, we aimed to obtain an accurate individual mismatch negativity (MMN) marker that could predict behavioral auditory discrimination thresholds. METHODS We measured the individual MMN response to four magnitudes of deviations in four different musical features (intensity, pitch, timbre, and rhythm) in a rare sample of experienced CI users and a control sample of normally hearing participants. We applied a recently developed spike density component analysis (SCA), which can suppress confounding alpha waves, and contrasted it with previously proposed methods. RESULTS Statistically detected individual MMN predicted attentive sound discrimination ability with high accuracy: for CI users 89.2% (278/312 cases) and for controls 90.5% (384/424 cases). As expected, MMN was detected for fewer CI users when the sound deviants were of smaller magnitude. CONCLUSIONS The findings support the use of MMN responses in individual CI users as a diagnostic tool for testing music perception. SIGNIFICANCE For CI users, the new SCA method provided more accurate and replicable diagnostic detections than preceding state-of-the-art.
Collapse
Affiliation(s)
- Niels Trusbak Haumann
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus/Aalborg, Universitetsbyen 3, 8000 Aarhus C, Denmark.
| | - Bjørn Petersen
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus/Aalborg, Universitetsbyen 3, 8000 Aarhus C, Denmark
| | - Anne Sofie Friis Andersen
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus/Aalborg, Universitetsbyen 3, 8000 Aarhus C, Denmark
| | | | - Elvira Brattico
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus/Aalborg, Universitetsbyen 3, 8000 Aarhus C, Denmark
| | - Peter Vuust
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University and The Royal Academy of Music, Aarhus/Aalborg, Universitetsbyen 3, 8000 Aarhus C, Denmark
| |
Collapse
|
6
|
The Use of Static and Dynamic Cues for Vowel Identification by Children Wearing Hearing Aids or Cochlear Implants. Ear Hear 2019; 41:72-81. [PMID: 30998549 DOI: 10.1097/aud.0000000000000735] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE To examine vowel perception based on dynamic formant transition and/or static formant pattern cues in children with hearing loss while using their hearing aids or cochlear implants. We predicted that the sensorineural hearing loss would degrade formant transitions more than static formant patterns, and that shortening the duration of cues would cause more difficulty for vowel identification for these children than for their normal-hearing peers. DESIGN A repeated-measures, between-group design was used. Children 4 to 9 years of age from a university hearing services clinic who were fit for hearing aids (13 children) or who wore cochlear implants (10 children) participated. Chronologically age-matched children with normal hearing served as controls (23 children). Stimuli included three naturally produced syllables (/ba/, /bi/, and /bu/), which were presented either in their entirety or segmented to isolate the formant transition or the vowel static formant center. The stimuli were presented to listeners via loudspeaker in the sound field. Aided participants wore their own devices and listened with their everyday settings. Participants chose the vowel presented by selecting from corresponding pictures on a computer screen. RESULTS Children with hearing loss were less able to use shortened transition or shortened vowel centers to identify vowels as compared to their normal-hearing peers. Whole syllable and initial transition yielded better identification performance than the vowel center for /ɑ/, but not for /i/ or /u/. CONCLUSIONS The children with hearing loss may require a longer time window than children with normal hearing to integrate vowel cues over time because of altered peripheral encoding in spectrotemporal domains. Clinical implications include cognizance of the importance of vowel perception when developing habilitative programs for children with hearing loss.
Collapse
|
7
|
VAN DE Velde DJ, Schiller NO, Levelt CC, VAN Heuven VJ, Beers M, Briaire JJ, Frijns JHM. Prosody perception and production by children with cochlear implants. JOURNAL OF CHILD LANGUAGE 2019; 46:111-141. [PMID: 30334510 DOI: 10.1017/s0305000918000387] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
The perception and production of emotional and linguistic (focus) prosody were compared in children with cochlear implants (CI) and normally hearing (NH) peers. Thirteen CI and thirteen hearing-age-matched school-aged NH children were tested, as baseline, on non-verbal emotion understanding, non-word repetition, and stimulus identification and naming. Main tests were verbal emotion discrimination, verbal focus position discrimination, acted emotion production, and focus production. Productions were evaluated by NH adult Dutch listeners. All scores between groups were comparable, except a lower score for the CI group for non-word repetition. Emotional prosody perception and production scores correlated weakly for CI children but were uncorrelated for NH children. In general, hearing age weakly predicted emotion production but not perception. Non-verbal emotional (but not linguistic) understanding predicted CI children's (but not controls') emotion perception and production. In conclusion, increasing time in sound might facilitate vocal emotional expression, possibly requiring independently maturing emotion perception skills.
Collapse
Affiliation(s)
- Daan J VAN DE Velde
- Leiden University Centre for Linguistics, Leiden University,Van Wijkplaats 3,2311 BX,Leiden
| | - Niels O Schiller
- Leiden University Centre for Linguistics, Leiden University,Van Wijkplaats 3,2311 BX,Leiden
| | - Claartje C Levelt
- Leiden University Centre for Linguistics, Leiden University,Van Wijkplaats 3,2311 BX,Leiden
| | - Vincent J VAN Heuven
- Department of Hungarian and Applied Linguistics,Pannon Egyetem,10 Egyetem Ut.,8200 Veszprém,Hungary
| | - Mieke Beers
- Leiden University Medical Center,ENT Department,Postbus 9600,2300 RC,Leiden
| | - Jeroen J Briaire
- Leiden University Medical Center,ENT Department,Postbus 9600,2300 RC,Leiden
| | - Johan H M Frijns
- Leiden Institute for Brain and Cognition,Postbus 9600, 2300 RC,Leiden
| |
Collapse
|
8
|
Sagi E, Svirsky MA. Contribution of formant frequency information to vowel perception in steady-state noise by cochlear implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:1027. [PMID: 28253672 PMCID: PMC5392095 DOI: 10.1121/1.4976059] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/16/2016] [Revised: 01/12/2017] [Accepted: 01/18/2017] [Indexed: 06/06/2023]
Abstract
Cochlear implant (CI) recipients have difficulty understanding speech in noise even at moderate signal-to-noise ratios. Knowing the mechanisms they use to understand speech in noise may facilitate the search for better speech processing algorithms. In the present study, a computational model is used to assess whether CI users' vowel identification in noise can be explained by formant frequency cues (F1 and F2). Vowel identification was tested with 12 unilateral CI users in quiet and in noise. Formant cues were measured from vowels in each condition, specific to each subject's speech processor. Noise distorted the location of vowels in the F2 vs F1 plane in comparison to quiet. The best fit model to subjects' data in quiet produced model predictions in noise that were within 8% of actual scores on average. Predictions in noise were much better when assuming that subjects used a priori knowledge regarding how formant information is degraded in noise (experiment 1). However, the model's best fit to subjects' confusion matrices in noise was worse than in quiet, suggesting that CI users utilize formant cues to identify vowels in noise, but to a different extent than how they identify vowels in quiet (experiment 2).
Collapse
Affiliation(s)
- Elad Sagi
- Department of Otolaryngology, New York University School of Medicine, New York, New York 10016, USA
| | - Mario A Svirsky
- Department of Otolaryngology, New York University School of Medicine, New York, New York 10016, USA
| |
Collapse
|
9
|
Kong YY, Winn MB, Poellmann K, Donaldson GS. Discriminability and Perceptual Saliency of Temporal and Spectral Cues for Final Fricative Consonant Voicing in Simulated Cochlear-Implant and Bimodal Hearing. Trends Hear 2016; 20:20/0/2331216516652145. [PMID: 27317666 PMCID: PMC5562340 DOI: 10.1177/2331216516652145] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
Multiple redundant acoustic cues can contribute to the perception of a single phonemic contrast. This study investigated the effect of spectral degradation on the discriminability and perceptual saliency of acoustic cues for identification of word-final fricative voicing in "loss" versus "laws", and possible changes that occurred when low-frequency acoustic cues were restored. Three acoustic cues that contribute to the word-final /s/-/z/ contrast (first formant frequency [F1] offset, vowel-consonant duration ratio, and consonant voicing duration) were systematically varied in synthesized words. A discrimination task measured listeners' ability to discriminate differences among stimuli within a single cue dimension. A categorization task examined the extent to which listeners make use of a given cue to label a syllable as "loss" versus "laws" when multiple cues are available. Normal-hearing listeners were presented with stimuli that were either unprocessed, processed with an eight-channel noise-band vocoder to approximate spectral degradation in cochlear implants, or low-pass filtered. Listeners were tested in four listening conditions: unprocessed, vocoder, low-pass, and a combined vocoder + low-pass condition that simulated bimodal hearing. Results showed a negative impact of spectral degradation on F1 cue discrimination and a trading relation between spectral and temporal cues in which listeners relied more heavily on the temporal cues for "loss-laws" identification when spectral cues were degraded. Furthermore, the addition of low-frequency fine-structure cues in simulated bimodal hearing increased the perceptual saliency of the F1 cue for "loss-laws" identification compared with vocoded speech. Findings suggest an interplay between the quality of sensory input and cue importance.
Collapse
Affiliation(s)
- Ying-Yee Kong
- Department of Communication Sciences and Disorders, Northeastern University, Boston, MA, USA
| | - Matthew B Winn
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Katja Poellmann
- Department of Communication Sciences and Disorders, Northeastern University, Boston, MA, USA
| | - Gail S Donaldson
- Department of Communication Sciences & Disorders, University of South Florida, Tampa, FL, USA
| |
Collapse
|
10
|
Donaldson GS, Rogers CL, Johnson LB, Oh SH. Vowel identification by cochlear implant users: Contributions of duration cues and dynamic spectral cues. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:65-73. [PMID: 26233007 PMCID: PMC4491094 DOI: 10.1121/1.4922173] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/15/2014] [Revised: 05/14/2015] [Accepted: 05/15/2015] [Indexed: 06/01/2023]
Abstract
A recent study from our laboratory assessed vowel identification in cochlear implant (CI) users, using full /dVd/ syllables and partial (center- and edges-only) syllables with duration cues neutralized [Donaldson, Rogers, Cardenas, Russell, and Hanna (2013). J. Acoust. Soc. Am. 134, 3021-3028]. CI users' poorer performance for partial syllables as compared to full syllables, and for edges-only syllables as compared to center-only syllables, led to the hypotheses (1) that CI users may rely strongly on vowel duration cues; and (2) that CI users have more limited access to dynamic spectral cues than steady-state spectral cues. The present study tested those hypotheses. Ten CI users and ten young normal hearing (YNH) listeners heard full /dVd/ syllables and modified (center- and edges-only) syllables in which vowel duration cues were either preserved or eliminated. The presence of duration cues significantly improved vowel identification scores in four CI users, suggesting a strong reliance on duration cues. Duration effects were absent for the other CI users and the YNH listeners. On average, CI users and YNH listeners demonstrated similar performance for center-only stimuli and edges-only stimuli having the same total duration of vowel information. However, three CI users demonstrated significantly poorer performance for the edges-only stimuli, indicating apparent deficits of dynamic spectral processing.
Collapse
Affiliation(s)
- Gail S Donaldson
- Department of Communication Sciences and Disorders, University of South Florida, PCD 1017, 4202 East Fowler Avenue, Tampa, Florida 33620, USA
| | - Catherine L Rogers
- Department of Communication Sciences and Disorders, University of South Florida, PCD 1017, 4202 East Fowler Avenue, Tampa, Florida 33620, USA
| | - Lindsay B Johnson
- Department of Communication Sciences and Disorders, University of South Florida, PCD 1017, 4202 East Fowler Avenue, Tampa, Florida 33620, USA
| | - Soo Hee Oh
- Department of Communication Sciences and Disorders, University of South Florida, PCD 1017, 4202 East Fowler Avenue, Tampa, Florida 33620, USA
| |
Collapse
|
11
|
Joseph S, Iverson P, Manohar S, Fox Z, Scott SK, Husain M. Precision of working memory for speech sounds. Q J Exp Psychol (Hove) 2015; 68:2022-40. [PMID: 25607721 DOI: 10.1080/17470218.2014.1002799] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
Memory for speech sounds is a key component of models of verbal working memory (WM). But how good is verbal WM? Most investigations assess this using binary report measures to derive a fixed number of items that can be stored. However, recent findings in visual WM have challenged such "quantized" views by employing measures of recall precision with an analogue response scale. WM for speech sounds might rely on both continuous and categorical storage mechanisms. Using a novel speech matching paradigm, we measured WM recall precision for phonemes. Vowel qualities were sampled from a formant space continuum. A probe vowel had to be adjusted to match the vowel quality of a target on a continuous, analogue response scale. Crucially, this provided an index of the variability of a memory representation around its true value and thus allowed us to estimate how memories were distorted from the original sounds. Memory load affected the quality of speech sound recall in two ways. First, there was a gradual decline in recall precision with increasing number of items, consistent with the view that WM representations of speech sounds become noisier with an increase in the number of items held in memory, just as for vision. Based on multidimensional scaling (MDS), the level of noise appeared to be reflected in distortions of the formant space. Second, as memory load increased, there was evidence of greater clustering of participants' responses around particular vowels. A mixture model captured both continuous and categorical responses, demonstrating a shift from continuous to categorical memory with increasing WM load. This suggests that direct acoustic storage can be used for single items, but when more items must be stored, categorical representations must be used.
Collapse
Affiliation(s)
- Sabine Joseph
- a Institute of Cognitive Neuroscience, University College London , London , UK
| | | | | | | | | | | |
Collapse
|
12
|
Moberly AC, Lowenstein JH, Tarr E, Caldwell-Tarr A, Welling DB, Shahin AJ, Nittrouer S. Do adults with cochlear implants rely on different acoustic cues for phoneme perception than adults with normal hearing? JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2014; 57:566-82. [PMID: 24686722 PMCID: PMC4008700 DOI: 10.1044/2014_jslhr-h-12-0323] [Citation(s) in RCA: 44] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
PURPOSE Several acoustic cues specify any single phonemic contrast. Nonetheless, adult, native speakers of a language share weighting strategies, showing preferential attention to some properties over others. Cochlear implant (CI) signal processing disrupts the salience of some cues: In general, amplitude structure remains readily available, but spectral structure less so. This study asked how well speech recognition is supported if CI users shift attention to salient cues not weighted strongly by native speakers. METHOD Twenty adults with CIs participated. The /bɑ/-/wɑ/ contrast was used because spectral and amplitude structure varies in correlated fashion for this contrast. Adults with normal hearing weight the spectral cue strongly but the amplitude cue negligibly. Three measurements were made: labeling decisions, spectral and amplitude discrimination, and word recognition. RESULTS Outcomes varied across listeners: Some weighted the spectral cue strongly, some weighted the amplitude cue, and some weighted neither. Spectral discrimination predicted spectral weighting. Spectral weighting explained the most variance in word recognition. Age of onset of hearing loss predicted spectral weighting but not unique variance in word recognition. CONCLUSION The weighting strategies of listeners with normal hearing likely support speech recognition best, so efforts in implant design, fitting, and training should focus on developing those strategies.
Collapse
|
13
|
Donaldson GS, Rogers CL, Cardenas ES, Russell BA, Hanna NH. Vowel identification by cochlear implant users: contributions of static and dynamic spectral cues. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 134:3021-3028. [PMID: 24116437 DOI: 10.1121/1.4820894] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Previous research has shown that normal hearing listeners can identify vowels in syllables on the basis of either quasi-static or dynamic spectral cues; however, it is not known how well cochlear implant (CI) users with current-generation devices can make use of these cues. The present study assessed vowel identification in adult CI users and a comparison group of young normal hearing (YNH) listeners. Stimuli were naturally spoken /dVd/ syllables and modified syllables that retained only quasi-static spectral cues from an 80-ms segment of the vowel center ("C80" stimuli) or dynamic spectral cues from two 20-ms segments of the vowel edges ("E20" stimuli). YNH listeners exhibited near-perfect performance for the unmodified (99.8%) and C80 (92.9%) stimuli and maintained good performance for the E20 stimuli (70.2%). CI users exhibited poorer average performance than YNH listeners for the Full stimuli (72.3%) and proportionally larger reductions in performance for the C80 stimuli (41.8%) and E20 stimuli (29.0%). Findings suggest that CI users have difficulty identifying vowels on the basis of spectral cues in the absence of duration cues, and have limited access to brief dynamic spectral cues. Error analyses suggest that CI users may rely strongly on vowel duration cues when those cues are available.
Collapse
Affiliation(s)
- Gail S Donaldson
- Department of Communication Sciences and Disorders, University of South Florida, PCD 1017, 4202 East Fowler Avenue, Tampa, Florida 33620
| | | | | | | | | |
Collapse
|
14
|
Van Zyl M, Hanekom JJ. Perception of vowels and prosody by cochlear implant recipients in noise. JOURNAL OF COMMUNICATION DISORDERS 2013; 46:449-464. [PMID: 24157128 DOI: 10.1016/j.jcomdis.2013.09.002] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/02/2012] [Revised: 09/13/2013] [Accepted: 09/16/2013] [Indexed: 06/02/2023]
Abstract
UNLABELLED The aim of the present study was to compare the ability of cochlear implant (CI) recipients to recognise speech prosody in the presence of speech-weighted noise to their ability to recognise vowels in the same test paradigm and listening condition. All test materials were recorded from four different speakers (two male, two female). Two prosody recognition tasks were developed, both using single words as stimuli. The first task involved a question/statement distinction, while the second task required listeners to make a judgement about the speaker's attitude. Vowel recognition tests were conducted using vowel pairs selected on the basis of specific acoustic cues (frequencies of the first two formants and duration). Ten CI users and ten normal-hearing controls were tested in both quiet and an adaptive noise condition, using a two-alternative forced-choice test paradigm for all the tests. Results indicated that vowel recognition was significantly better than prosody recognition in both listener groups in both quiet and noise, and that question/statement discrimination was the most difficult task for CI listeners in noise. Data from acoustic analyses were used to interpret differences in performance on different tasks and with different speakers. LEARNING OUTCOMES As a result of this activity, readers will be able to (1) describe suitable methods for comparing vowel and prosody perception in noise, (2) compare performance on vowel and prosody perception tasks in quiet in normal-hearing listeners and cochlear implant recipients, (3) compare performance on vowel and prosody perception tasks in noise in normal-hearing listeners and cochlear implant recipients and (4) relate performance on prosody tasks in quiet to performance on these tasks in noise.
Collapse
Affiliation(s)
- Marianne Van Zyl
- Department of Electrical, Electronic and Computer Engineering, University of Pretoria, Lynnwood Road, Pretoria 0002, South Africa
| | | |
Collapse
|
15
|
Peng SC, Chatterjee M, Lu N. Acoustic cue integration in speech intonation recognition with cochlear implants. Trends Amplif 2012; 16:67-82. [PMID: 22790392 PMCID: PMC3560417 DOI: 10.1177/1084713812451159] [Citation(s) in RCA: 44] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The present article reports on the perceptual weighting of prosodic cues in question-statement identification by adult cochlear implant (CI) listeners. Acoustic analyses of normal-hearing (NH) listeners' production of sentences spoken as questions or statements confirmed that in English the last bisyllabic word in a sentence carries the dominant cues (F0, duration, and intensity patterns) for the contrast. Furthermore, these analyses showed that the F0 contour is the primary cue for the question-statement contrast, with intensity and duration changes conveying important but less reliable information. On the basis of these acoustic findings, the authors examined adult CI listeners' performance in two question-statement identification tasks. In Task 1, 13 CI listeners' question-statement identification accuracy was measured using naturally uttered sentences matched for their syntactic structures. In Task 2, the same listeners' perceptual cue weighting in question-statement identification was assessed using resynthesized single-word stimuli, within which fundamental frequency (F0), intensity, and duration properties were systematically manipulated. Both tasks were also conducted with four NH listeners with full-spectrum and noise-band-vocoded stimuli. Perceptual cue weighting was assessed by comparing the estimated coefficients in logistic models fitted to the data. Of the 13 CI listeners, 7 achieved high performance levels in Task 1. The results of Task 2 indicated that multiple sources of acoustic cues for question-statement identification were utilized to different extents depending on the listening conditions (e.g., full spectrum vs. spectrally degraded) or the listeners' hearing and amplification status (e.g., CI vs. NH).
Collapse
Affiliation(s)
- Shu-Chen Peng
- Division of Ophthalmic, Neurological, and Ear, Nose and Throat Devices, Office of Device Evaluation, U.S. Food and Drug Administration, 10903 New Hampshire Ave, Silver Spring, MD 20993, USA.
| | | | | |
Collapse
|
16
|
Winn MB, Chatterjee M, Idsardi WJ. The use of acoustic cues for phonetic identification: effects of spectral degradation and electric hearing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 131:1465-1479. [PMID: 22352517 PMCID: PMC3292615 DOI: 10.1121/1.3672705] [Citation(s) in RCA: 45] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/07/2010] [Revised: 10/10/2011] [Accepted: 12/05/2011] [Indexed: 05/30/2023]
Abstract
Although some cochlear implant (CI) listeners can show good word recognition accuracy, it is not clear how they perceive and use the various acoustic cues that contribute to phonetic perceptions. In this study, the use of acoustic cues was assessed for normal-hearing (NH) listeners in optimal and spectrally degraded conditions, and also for CI listeners. Two experiments tested the tense/lax vowel contrast (varying in formant structure, vowel-inherent spectral change, and vowel duration) and the word-final fricative voicing contrast (varying in F1 transition, vowel duration, consonant duration, and consonant voicing). Identification results were modeled using mixed-effects logistic regression. These experiments suggested that under spectrally-degraded conditions, NH listeners decrease their use of formant cues and increase their use of durational cues. Compared to NH listeners, CI listeners showed decreased use of spectral cues like formant structure and formant change and consonant voicing, and showed greater use of durational cues (especially for the fricative contrast). The results suggest that although NH and CI listeners may show similar accuracy on basic tests of word, phoneme or feature recognition, they may be using different perceptual strategies in the process.
Collapse
Affiliation(s)
- Matthew B Winn
- Department of Hearing and Speech Sciences, University of Maryland, College Park, 0100 Lefrak Hall, College Park, Maryland 20742, USA.
| | | | | |
Collapse
|
17
|
Välimaa TT, Sorri MJ, Laitakari J, Sivonen V, Muhli A. Vowel confusion patterns in adults during initial 4 years of implant use. CLINICAL LINGUISTICS & PHONETICS 2011; 25:121-144. [PMID: 21070135 DOI: 10.3109/02699206.2010.514692] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
This study investigated adult cochlear implant users' (n = 39) vowel recognition and confusions by an open-set syllable test during 4 years of implant use, in a prospective repeated-measures design. Subjects' responses were coded for phoneme errors and estimated by the generalized mixed model. Improvement in overall vowel recognition was highest during the first 6 months, showing statistically significant change until 4 years, especially for the mediocre performers. The best performers improved statistically significantly until 18 months. The poorest performers improved until 12 months and exhibited more vowel confusions. No differences were found in overall vowel recognition between Nucleus24M/24R and Med-ElC40+ device users (matched comparison), but certain vowels showed statistically significant differences. Vowel confusions between adjacent vowels were evident, probably due to the implant users' inability to discriminate formant frequencies. Vowel confusions were also dominated by vowels whose average F1 and/or F2 frequencies were higher than the target vowel, indicating a basalward shift in the confusions.
Collapse
Affiliation(s)
- Taina T Välimaa
- Faculty of Humanities, Logopedics, and Department of Otorhinolaryngology. Oulu University Hospital, University of Oulu, Finland.
| | | | | | | | | |
Collapse
|
18
|
Giezen MR, Escudero P, Baker A. Use of acoustic cues by children with cochlear implants. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2010; 53:1440-1457. [PMID: 20689031 DOI: 10.1044/1092-4388(2010/09-0252)] [Citation(s) in RCA: 26] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
PURPOSE This study examined the use of different acoustic cues in auditory perception of consonant and vowel contrasts by profoundly deaf children with a cochlear implant (CI) in comparison to age-matched children and young adults with normal hearing. METHOD A speech sound categorization task in an XAB format was administered to 15 children ages 5-6 with a CI (mean age at implant: 1;8 [years;months]), 20 normal-hearing age-matched children, and 21 normal-hearing adults. Four contrasts were examined: //-/a/, /i/-/i/, /bu/-/pu/, and /fu/-/su/. Measures included phoneme endpoint identification, individual cue reliance, cue weighting, and classification slope. RESULTS The children with a CI used the spectral cues in the /fu/-/su/ contrast less effectively than the children with normal hearing, resulting in poorer phoneme endpoint identification and a shallower classification slope. Performance on the other 3 contrasts did not differ significantly. Adults consistently showed steeper classification slopes than the children, but similar cue-weighting patterns were observed in all 3 groups. CONCLUSIONS Despite their different auditory input, children with a CI appear to be able to use many acoustic cues effectively in speech perception. Most importantly, children with a CI and normal-hearing children were observed to use similar cue-weighting patterns.
Collapse
Affiliation(s)
- Marcel R Giezen
- Amsterdam Center for Language and Communication, University of Amsterdam, the Netherlands.
| | | | | |
Collapse
|
19
|
Sagi E, Meyer TA, Kaiser AR, Teoh SW, Svirsky MA. A mathematical model of vowel identification by users of cochlear implants. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 127:1069-83. [PMID: 20136228 PMCID: PMC2830268 DOI: 10.1121/1.3277215] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/01/2009] [Revised: 11/25/2009] [Accepted: 11/30/2009] [Indexed: 05/28/2023]
Abstract
A simple mathematical model is presented that predicts vowel identification by cochlear implant users based on these listeners' resolving power for the mean locations of first, second, and/or third formant energies along the implanted electrode array. This psychophysically based model provides hypotheses about the mechanism cochlear implant users employ to encode and process the input auditory signal to extract information relevant for identifying steady-state vowels. Using one free parameter, the model predicts most of the patterns of vowel confusions made by users of different cochlear implant devices and stimulation strategies, and who show widely different levels of speech perception (from near chance to near perfect). Furthermore, the model can predict results from the literature, such as Skinner, et al. [(1995). Ann. Otol. Rhinol. Laryngol. 104, 307-311] frequency mapping study, and the general trend in the vowel results of Zeng and Galvin's [(1999). Ear Hear. 20, 60-74] studies of output electrical dynamic range reduction. The implementation of the model presented here is specific to vowel identification by cochlear implant users, but the framework of the model is more general. Computational models such as the one presented here can be useful for advancing knowledge about speech perception in hearing impaired populations, and for providing a guide for clinical research and clinical practice.
Collapse
Affiliation(s)
- Elad Sagi
- Department of Otolaryngology, New York University School of Medicine, New York, New York 10016, USA.
| | | | | | | | | |
Collapse
|
20
|
Iverson P, Evans BG. Learning English vowels with different first-language vowel systems II: Auditory training for native Spanish and German speakers. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2009; 126:866-877. [PMID: 19640051 DOI: 10.1121/1.3148196] [Citation(s) in RCA: 38] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
This study investigated whether individuals with small and large native-language (L1) vowel inventories learn second-language (L2) vowel systems differently, in order to better understand how L1 categories interfere with new vowel learning. Listener groups whose L1 was Spanish (5 vowels) or German (18 vowels) were given five sessions of high-variability auditory training for English vowels, after having been matched to assess their pre-test English vowel identification accuracy. Listeners were tested before and after training in terms of their identification accuracy for English vowels, the assimilation of these vowels into their L1 vowel categories, and their best exemplars for English (i.e., perceptual vowel space map). The results demonstrated that Germans improved more than Spanish speakers, despite the Germans' more crowded L1 vowel space. A subsequent experiment demonstrated that Spanish listeners were able to improve as much as the German group after an additional ten sessions of training, and that both groups were able to retain this learning. The findings suggest that a larger vowel category inventory may facilitate new learning, and support a hypothesis that auditory training improves identification by making the application of existing categories to L2 phonemes more automatic and efficient.
Collapse
Affiliation(s)
- Paul Iverson
- Division of Psychology and Language Sciences, University College London, Chandler House, 2 Wakefield Street, London, United Kingdom
| | | |
Collapse
|
21
|
Hattori K, Iverson P. English /r/-/l/ category assimilation by Japanese adults: individual differences and the link to identification accuracy. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2009; 125:469-479. [PMID: 19173432 DOI: 10.1121/1.3021295] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/27/2023]
Abstract
Native speakers of Japanese often have difficulty identifying English /r/ and /l/, and it has been thought that second-language (L2) learning difficulties like this are caused by how L2 phonemes are assimilated into ones native phonological system. This study took an individual difference approach to examining this relationship by testing the category assimilation of Japanese speakers with a wide range of English /r/-/l/ identification abilities. All Japanese subjects were assessed in terms of (1) their accuracy in identifying English /r/ and /l/, (2) their assimilation of /r/ and /l/ into their Japanese flap category, (3) their production of /r/ and /l/, and (4) their best-exemplar locations for /r/, /l/, and Japanese flap in a five-dimensional set of synthetic stimuli (F1, F2, F3, closure duration, and transition duration). The results demonstrated that Japanese speakers assimilate /l/ into their flap category more strongly than they assimilate /r/. However, there was little evidence that category assimilation was predictive of English /r/-/l/ perception and production. Japanese speakers had three distinct best exemplars for /r/, /l/, and flap, and only their representation of F3 in /r/ and /l/ was predictive of identification ability.
Collapse
Affiliation(s)
- Kota Hattori
- Division of Psychology and Language Sciences, University College London, London, United Kingdom
| | | |
Collapse
|
22
|
Iverson P, Evans BG. Learning English vowels with different first-language vowel systems: perception of formant targets, formant movement, and duration. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 122:2842-2854. [PMID: 18189574 DOI: 10.1121/1.2783198] [Citation(s) in RCA: 36] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
This study examined whether individuals with a wide range of first-language vowel systems (Spanish, French, German, and Norwegian) differ fundamentally in the cues that they use when they learn the English vowel system (e.g., formant movement and duration). All subjects: (1) identified natural English vowels in quiet; (2) identified English vowels in noise that had been signal processed to flatten formant movement or equate duration; (3) perceptually mapped best exemplars for first- and second-language synthetic vowels in a five-dimensional vowel space that included formant movement and duration; and (4) rated how natural English vowels assimilated into their L1 vowel categories. The results demonstrated that individuals with larger and more complex first-language vowel systems (German and Norwegian) were more accurate at recognizing English vowels than were individuals with smaller first-language systems (Spanish and French). However, there were no fundamental differences in what these individuals learned. That is, all groups used formant movement and duration to recognize English vowels, and learned new aspects of the English vowel system rather than simply assimilating vowels into existing first-language categories. The results suggest that there is a surprising degree of uniformity in the ways that individuals with different language backgrounds perceive second language vowels.
Collapse
Affiliation(s)
- Paul Iverson
- Department of Phonetics and Linguistics, University College London, 4 Stephenson Way, London NW1 2HE, United Kingdom
| | | |
Collapse
|
23
|
Evans BG, Iverson P. Plasticity in vowel perception and production: a study of accent change in young adults. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2007; 121:3814-26. [PMID: 17552729 DOI: 10.1121/1.2722209] [Citation(s) in RCA: 48] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
This study investigated changes in vowel production and perception among university students from the north of England, as individuals adapt their accent from regional to educated norms. Subjects were tested in their production and perception at regular intervals over a period of 2 years: before beginning university, 3 months later, and at the end of their first and second years at university. At each testing session, subjects were recorded reading a set of experimental words and a short passage. Subjects also completed two perceptual tasks; they chose best exemplar locations for vowels embedded in either northern or southern English accented carrier sentences and identified words in noise spoken with either a northern or southern English accent. The results demonstrated that subjects at a late stage in their language development, early adulthood, changed their spoken accent after attending university. There were no reliable changes in perception over time, but there was evidence for a between-subjects link between production and perception; subjects chose similar vowels to the ones they produced, and subjects who had a more southern English accent were better at identifying southern English speech in noise.
Collapse
Affiliation(s)
- Bronwen G Evans
- Department of Phonetics and Linguistics, University College London, 4 Stephenson Way, London NW1 2HE, United Kingdom
| | | |
Collapse
|