1
|
MacLellan LE, Stepp CE, Fager SK, Mentis M, Boucher AR, Abur D, Cler GJ. Evaluating Camera Mouse as a computer access system for augmentative and alternative communication in cerebral palsy: A case study. Assist Technol 2024; 36:217-223. [PMID: 37699111 PMCID: PMC10927611 DOI: 10.1080/10400435.2023.2242893] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/26/2023] [Indexed: 09/14/2023] Open
Abstract
Camera Mouse is a freely available software program that visually tracks the movement of facial features to allow individuals with motor impairments to control a computer mouse. The goal of this case study was to provide an evaluation of Camera Mouse as a computer access method as part of a multiple modality communication system for an individual with cerebral palsy. The participant was asked to reproduce sentences and respond to ethical dilemmas for language sampling. Tasks were completed using natural speech and an AAC solution consisting of Camera Mouse paired with an orthographic selection interface and speech synthesis. The participant completed a questionnaire for satisfaction with the introduced assistive technology. Camera Mouse resulted in higher intelligibility than natural speech, while natural speech had a higher rate. She used more complex language with her natural speech. The participant rated Camera Mouse as at least 3/5 on all measures, including 5/5 on weight and safety. The results of this case study suggest Camera Mouse is a promising computer access system for communication supported by the participant's satisfaction rating, expressive language, and synthesized speech production capabilities.
Collapse
Affiliation(s)
- Lauren E. MacLellan
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA
| | - Cara E. Stepp
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA
- Graduate Program for Neuroscience, Boston University, Boston, MA
- Department of Biomedical Engineering, Boston University, Boston, MA
- Department of Otolaryngology – Head and Neck Surgery, Boston University School of Medicine, Boston, MA
| | - Susan K. Fager
- Institute for Rehabilitation Science and Engineering, Madonna Rehabilitation Hospitals, Lincoln, NE
| | - Michelle Mentis
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA
| | - Alyssa R. Boucher
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA
| | - Defne Abur
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA
| | - Gabriel J. Cler
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA
- Graduate Program for Neuroscience, Boston University, Boston, MA
- Department of Speech & Hearing Sciences, University of Washington, Seattle, WA
| |
Collapse
|
2
|
Asaridou SS, Cler GJ, Wiedemann A, Krishnan S, Smith HJ, Willis HE, Healy MP, Watkins KE. Microstructural Properties of the Cerebellar Peduncles in Children with Developmental Language Disorder. bioRxiv 2023:2023.07.13.548858. [PMID: 37503009 PMCID: PMC10370025 DOI: 10.1101/2023.07.13.548858] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/29/2023]
Abstract
Children with developmental language disorder (DLD) struggle to learn their native language for no apparent reason. While research on the neurobiological underpinnings of the disorder has focused on the role of cortico-striatal systems, little is known about the role of the cerebellum in DLD. Cortico-cerebellar circuits might be involved in the disorder as they contribute to complex sensorimotor skill learning, including the acquisition of spoken language. Here, we used diffusion-weighted imaging data from 77 typically developing and 54 children with DLD and performed probabilistic tractography to identify the cerebellum's white matter tracts: the inferior, middle, and superior cerebellar peduncles. Children with DLD showed lower fractional anisotropy (FA) in the inferior cerebellar peduncles (ICP), fiber tracts that carry motor and sensory input via the inferior olive to the cerebellum. Lower FA in DLD was driven by lower axial diffusivity. Probing this further with more sophisticated modeling of diffusion data, we found higher orientation dispersion but no difference in neurite density in the ICP of DLD. Reduced FA is therefore unlikely to be reflecting microstructural differences in myelination in this tract, rather the organization of axons in these pathways is disrupted. ICP microstructure was not associated with language or motor coordination performance in our sample. We also found no differences in the middle and superior peduncles, the main pathways connecting the cerebellum with the cortex. To conclude, it is not cortico-cerebellar but atypical olivocerebellar white matter connections that characterize DLD and suggest the involvement of the olivocerebellar system in speech acquisition and development.
Collapse
Affiliation(s)
- Salomi S. Asaridou
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
| | - Gabriel J. Cler
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
- Department of Speech & Hearing Sciences, University of Washington, Seattle, USA
| | - Anna Wiedemann
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
- Department of Psychiatry, University of Cambridge, Cambridge, UK
| | - Saloni Krishnan
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
- Department of Psychology, Royal Holloway, University of London, Egham Hill, Surrey, UK
| | - Harriet J. Smith
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Hanna E. Willis
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
- Nuffield Department of Clinical Neuroscience, University of Oxford, Oxford, UK
| | - Máiréad P. Healy
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
- Department of Psychology, University of Cambridge, Cambridge, UK
| | - Kate E. Watkins
- Department of Experimental Psychology, Wellcome Centre for Integrative Neuroimaging, University of Oxford, Oxford, UK
| |
Collapse
|
3
|
Krishnan S, Cler GJ, Smith HJ, Willis HE, Asaridou SS, Healy MP, Papp D, Watkins KE. Quantitative MRI reveals differences in striatal myelin in children with DLD. eLife 2022; 11:e74242. [PMID: 36164824 PMCID: PMC9514847 DOI: 10.7554/elife.74242] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2021] [Accepted: 07/21/2022] [Indexed: 12/25/2022] Open
Abstract
Developmental language disorder (DLD) is a common neurodevelopmental disorder characterised by receptive or expressive language difficulties or both. While theoretical frameworks and empirical studies support the idea that there may be neural correlates of DLD in frontostriatal loops, findings are inconsistent across studies. Here, we use a novel semiquantitative imaging protocol - multi-parameter mapping (MPM) - to investigate microstructural neural differences in children with DLD. The MPM protocol allows us to reproducibly map specific indices of tissue microstructure. In 56 typically developing children and 33 children with DLD, we derived maps of (1) longitudinal relaxation rate R1 (1/T1), (2) transverse relaxation rate R2* (1/T2*), and (3) Magnetization Transfer saturation (MTsat). R1 and MTsat predominantly index myelin, while R2* is sensitive to iron content. Children with DLD showed reductions in MTsat values in the caudate nucleus bilaterally, as well as in the left ventral sensorimotor cortex and Heschl's gyrus. They also had globally lower R1 values. No group differences were noted in R2* maps. Differences in MTsat and R1 were coincident in the caudate nucleus bilaterally. These findings support our hypothesis of corticostriatal abnormalities in DLD and indicate abnormal levels of myelin in the dorsal striatum in children with DLD.
Collapse
Affiliation(s)
- Saloni Krishnan
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
- Department of Psychology, Royal Holloway, University of London, Egham HillLondonUnited Kingdom
| | - Gabriel J Cler
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
- Department of Speech and Hearing Sciences, University of WashingtonSeattleUnited States
| | - Harriet J Smith
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
- MRC Cognition and Brain Sciences Unit, University of CambridgeCambridgeUnited Kingdom
| | - Hanna E Willis
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
- Nuffield Department of Clinical Neurosciences, John Radcliffe HospitalOxfordUnited Kingdom
| | - Salomi S Asaridou
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
| | - Máiréad P Healy
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
- Department of Psychology, University of CambridgeCambridgeUnited Kingdom
| | - Daniel Papp
- NeuroPoly Lab, Biomedical Engineering Department, Polytechnique MontrealMontrealCanada
- Wellcome Centre for Integrative Neuroimaging, FMRIB Centre, Nuffield Department of Clinical Neuroscience, University of OxfordOxfordUnited Kingdom
| | - Kate E Watkins
- Wellcome Centre for Integrative Neuroimaging, Dept of Experimental Psychology, University of OxfordOxfordUnited Kingdom
| |
Collapse
|
4
|
Abstract
This study introduces an ability-based method for personalized keyboard generation, wherein an individual’s own movement and human–computer interaction data are used to automatically compute a personalized virtual keyboard layout. Our approach integrates a multidirectional point-select task to characterize cursor control over time, distance, and direction. The characterization is automatically employed to develop a computationally efficient keyboard layout that prioritizes each user’s movement abilities through capturing directional constraints and preferences. We evaluated our approach in a study involving 16 participants using inertial sensing and facial electromyography as an access method, resulting in significantly increased communication rates using the personalized keyboard (52.0 bits/min) when compared to a generically optimized keyboard (47.9 bits/min). Our results demonstrate the ability to effectively characterize an individual’s movement abilities to design a personalized keyboard for improved communication. This work underscores the importance of integrating a user’s motor abilities when designing virtual interfaces.
Collapse
Affiliation(s)
| | - Gabriel J. Cler
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA 98105, USA
| | - Susan K. Fager
- Institute of Rehabilitation Science and Engineering, Madonna Rehabilitation Hospital, Lincoln, NE 68506, USA
| | - Paola Contessa
- Delsys, Inc., Natick, MA 01760, USA
- Altec, Inc., Natick, MA 01760, USA
| | - Serge H. Roy
- Delsys, Inc., Natick, MA 01760, USA
- Altec, Inc., Natick, MA 01760, USA
| | - Gianluca De Luca
- Delsys, Inc., Natick, MA 01760, USA
- Altec, Inc., Natick, MA 01760, USA
| | - Joshua C. Kline
- Delsys, Inc., Natick, MA 01760, USA
- Altec, Inc., Natick, MA 01760, USA
| | - Jennifer M. Vojtech
- Delsys, Inc., Natick, MA 01760, USA
- Altec, Inc., Natick, MA 01760, USA
- Correspondence:
| |
Collapse
|
5
|
Weerathunge HR, Alzamendi GA, Cler GJ, Guenther FH, Stepp CE, Zañartu M. LaDIVA: A neurocomputational model providing laryngeal motor control for speech acquisition and production. PLoS Comput Biol 2022; 18:e1010159. [PMID: 35737706 PMCID: PMC9258861 DOI: 10.1371/journal.pcbi.1010159] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Revised: 07/06/2022] [Accepted: 05/02/2022] [Indexed: 11/18/2022] Open
Abstract
Many voice disorders are the result of intricate neural and/or biomechanical impairments that are poorly understood. The limited knowledge of their etiological and pathophysiological mechanisms hampers effective clinical management. Behavioral studies have been used concurrently with computational models to better understand typical and pathological laryngeal motor control. Thus far, however, a unified computational framework that quantitatively integrates physiologically relevant models of phonation with the neural control of speech has not been developed. Here, we introduce LaDIVA, a novel neurocomputational model with physiologically based laryngeal motor control. We combined the DIVA model (an established neural network model of speech motor control) with the extended body-cover model (a physics-based vocal fold model). The resulting integrated model, LaDIVA, was validated by comparing its model simulations with behavioral responses to perturbations of auditory vocal fundamental frequency (fo) feedback in adults with typical speech. LaDIVA demonstrated capability to simulate different modes of laryngeal motor control, ranging from short-term (i.e., reflexive) and long-term (i.e., adaptive) auditory feedback paradigms, to generating prosodic contours in speech. Simulations showed that LaDIVA’s laryngeal motor control displays properties of motor equivalence, i.e., LaDIVA could robustly generate compensatory responses to reflexive vocal fo perturbations with varying initial laryngeal muscle activation levels leading to the same output. The model can also generate prosodic contours for studying laryngeal motor control in running speech. LaDIVA can expand the understanding of the physiology of human phonation to enable, for the first time, the investigation of causal effects of neural motor control in the fine structure of the vocal signal.
Collapse
Affiliation(s)
- Hasini R. Weerathunge
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts, United States of America
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts, United States of America
- * E-mail:
| | - Gabriel A. Alzamendi
- Department of Electronic Engineering, Universidad Técnica Federico Santa María, Valparaíso, Chile
- Institute for Research and Development on Bioengineering and Bioinformatics (IBB), CONICET-UNER, Oro Verde, Argentina
| | - Gabriel J. Cler
- Department of Speech & Hearing Sciences, University of Washington, Seattle, Washington, United States of America
| | - Frank H. Guenther
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts, United States of America
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts, United States of America
| | - Cara E. Stepp
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts, United States of America
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head and Neck Surgery, Boston University School of Medicine, Boston, Massachusetts, United States of America
| | - Matías Zañartu
- Department of Electronic Engineering, Universidad Técnica Federico Santa María, Valparaíso, Chile
| |
Collapse
|
6
|
Cler GJ, Krishnan S, Papp D, Wiltshire CEE, Chesters J, Watkins KE. Elevated iron concentration in putamen and cortical speech motor network in developmental stuttering. Brain 2021; 144:2979-2984. [PMID: 34750604 PMCID: PMC8634076 DOI: 10.1093/brain/awab283] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2021] [Revised: 06/17/2021] [Accepted: 07/16/2021] [Indexed: 11/12/2022] Open
Abstract
Theoretical accounts of developmental stuttering implicate dysfunctional cortico-striatal-thalamo-cortical motor loops through the putamen. However, the analysis of conventional MRI brain scans in individuals who stutter has failed to yield strong support for this theory in terms of reliable differences in the structure or function of the basal ganglia. Here, we performed quantitative mapping of brain tissue, which can be used to measure iron content alongside markers sensitive to myelin and thereby offers particular sensitivity to the measurement of iron-rich structures such as the basal ganglia. Analysis of these quantitative maps in 41 men and women who stutter and 32 individuals who are typically fluent revealed significant group differences in maps of R2*, indicative of higher iron content in individuals who stutter in the left putamen and in left hemisphere cortical regions important for speech motor control. Higher iron levels in brain tissue in individuals who stutter could reflect elevated dopamine levels or lysosomal dysfunction, both of which are implicated in stuttering. This study represents the first use of these quantitative measures in developmental stuttering and provides new evidence of microstructural differences in the basal ganglia and connected frontal cortical regions.
Collapse
Affiliation(s)
- Gabriel J Cler
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, UK
| | - Saloni Krishnan
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, UK
- Department of Psychology, Royal Holloway, University of London, Egham Hill, Surrey TW20 0EX, UK
| | - Daniel Papp
- Wellcome Centre for Integrative Neuroimaging, FMRIB Centre, Nuffield Department of Clinical Neuroscience, University of Oxford, Oxford OX3 9DU, UK
| | - Charlotte E E Wiltshire
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, UK
| | - Jennifer Chesters
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, UK
- Bristol Speech and Language Therapy Research Unit, North Bristol NHS Trust, Bristol BS10 5NB, UK
| | - Kate E Watkins
- Wellcome Centre for Integrative Neuroimaging, Department of Experimental Psychology, University of Oxford, Oxford OX2 6GG, UK
| |
Collapse
|
7
|
Brown KM, Dahl KL, Cler GJ, Stepp CE. Listener Age and Gender Diversity: Effects on Voice-based Perception of Gender. J Voice 2021; 35:739-745. [PMID: 32165021 PMCID: PMC7483284 DOI: 10.1016/j.jvoice.2020.02.004] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2019] [Revised: 02/05/2020] [Accepted: 02/06/2020] [Indexed: 10/24/2022]
Abstract
OBJECTIVE An important clinical outcome of voice masculinization treatments in transmasculine speakers is voice-based perception of gender. Rigorous assessments of voice treatment that utilize ratings of perception of gender typically do not control for demographic characteristics of the listeners. The objective of the present study was to determine the effect of listeners' age and gender diversity on voice-based judgments of speaker gender. METHODS Speech stimuli were produced by a single transmasculine individual over approximately one year of hormone replacement therapy, during which he experienced significant changes in his voice. Three groups of listeners rated speech stimuli on a visual analog scale with anchors ranging from "definitely male" to "guessing male" to "guessing female" to "definitely female." Listener groups were N = 10 cisgender young adults, N = 10 cisgender older adults, and N = 10 gender diverse individuals. RESULTS All groups rated the speaker as consistently female through week 14 of hormone replacement therapy and consistently male after week 28. Mean responses of the three groups of listeners were highly correlated (Pearson's correlations all r > 0.97). CONCLUSION Given reasonable group sizes, average ratings of gender perception of a transmasculine speaker are not highly influenced by varying listener age and gender minority status.
Collapse
Affiliation(s)
- Katherine M Brown
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts
| | - Kimberly L Dahl
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts
| | - Gabriel J Cler
- Department of Experimental Psychology, University of Oxford, Oxford, UK
| | - Cara E Stepp
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts; Department of Otolaryngology - Head and Neck Surgery, Boston University School of Medicine, Boston, Massachusetts; Department of Biomedical Engineering, Boston University, Boston, Massachusetts.
| |
Collapse
|
8
|
Cler GJ, Perkell JS, Stepp CE. Oral configurations during vowel nasalization in English. Speech Commun 2021; 129:17-24. [PMID: 34621100 PMCID: PMC8492006 DOI: 10.1016/j.specom.2021.02.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Speech nasalization is achieved primarily through the opening and closing of the velopharyngeal port. However, the resultant acoustic features can also be influenced by tongue configuration. Although vowel nasalization is not contrastive in English, two previous studies have found possible differences in the oral articulation of nasal and oral vowel productions, albeit with inconsistent results. In an attempt to further understand the conflicting findings, we evaluated the oral kinematics of nasalized and non-nasalized vowels in a cohort of both male and female American English speakers via electromagnetic articulography. Tongue body and lip positions were captured during vowels produced in nasal and oral contexts (e.g., /mɑm/, /bɑb/). Large contrasts were seen in all participants between tongue position of /æ/ in oral and nasal contexts, in which tongue positions were higher and more forward during /mæm/ than /bæb/. Lip aperture was smaller in a nasal context for /æ/. Lip protrusion was not different between vowels in oral and nasal contexts. Smaller contrasts in tongue and lip position were seen for vowels /ɑ, i, u/; this is consistent with biomechanical accounts of vowel production that suggest that /i, u/ are particularly constrained, whereas /æ/ has fewer biomechanical constraints, allowing for more flexibility for articulatory differences in different contexts. Thus we conclude that speakers of American English do indeed use different oral configurations for vowels that are in nasal and oral contexts, despite vowel nasalization being non-contrastive. This effect was consistent across speakers for only one vowel, perhaps accounting for previously-conflicting results.
Collapse
Affiliation(s)
- Gabriel J Cler
- Graduate Program for Neuroscience - Computational Neuroscience, Boston University
- Department of Speech, Language, and Hearing Sciences, Boston University
- Department of Speech and Hearing Sciences, University of Washington
| | - Joseph S Perkell
- Department of Speech, Language, and Hearing Sciences, Boston University
- Research Laboratory of Electronics, Massachusetts Institute of Technology
| | - Cara E Stepp
- Graduate Program for Neuroscience - Computational Neuroscience, Boston University
- Department of Speech, Language, and Hearing Sciences, Boston University
- Department of Biomedical Engineering, Boston University
- Department of Otolaryngology - Head and Neck Surgery, Boston University School of Medicine
| |
Collapse
|
9
|
Krishnan S, Asaridou SS, Cler GJ, Smith HJ, Willis HE, Healy MP, Thompson PA, Bishop DVM, Watkins KE. Functional organisation for verb generation in children with developmental language disorder. Neuroimage 2020; 226:117599. [PMID: 33285329 PMCID: PMC7836232 DOI: 10.1016/j.neuroimage.2020.117599] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2020] [Revised: 11/13/2020] [Accepted: 11/17/2020] [Indexed: 11/30/2022] Open
Abstract
Developmental language disorder (DLD) is characterised by difficulties in learning one's native language for no apparent reason. These language difficulties occur in 7% of children and are known to limit future academic and social achievement. Our understanding of the brain abnormalities associated with DLD is limited. Here, we used a simple four-minute verb generation task (children saw a picture of an object and were instructed to say an action that goes with that object) to test children between the ages of 10-15 years (DLD N = 50, typically developing N = 67). We also tested 26 children with poor language ability who did not meet our criteria for DLD. Contrary to our registered predictions, we found that children with DLD did not have (i) reduced activity in language relevant regions such as the left inferior frontal cortex; (ii) dysfunctional striatal activity during overt production; or (iii) a reduction in left-lateralised activity in frontal cortex. Indeed, performance of this simple language task evoked activity in children with DLD in the same regions and to a similar level as in typically developing children. Consistent with previous reports, we found sub-threshold group differences in the left inferior frontal gyrus and caudate nuclei, but only when analysis was limited to a subsample of the DLD group (N = 14) who had the poorest performance on the task. Additionally, we used a two-factor model to capture variation in all children studied (N = 143) on a range of neuropsychological tests and found that these language and verbal memory factors correlated with activity in different brain regions. Our findings indicate a lack of support for some neurological models of atypical language learning, such as the procedural deficit hypothesis or the atypical lateralization hypothesis, at least when using simple language tasks that children can perform. These results also emphasise the importance of controlling for and monitoring task performance.
Collapse
Affiliation(s)
- Saloni Krishnan
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK; Department of Psychology, Royal Holloway, University of London, Egham Hill, Surrey TW20 0EX, UK.
| | - Salomi S Asaridou
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK
| | - Gabriel J Cler
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK
| | - Harriet J Smith
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK; MRC Cognition and Brain Sciences Unit, University of Cambridge, 15 Chaucer Road, Cambridge CB2 7EF, UK
| | - Hannah E Willis
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK; Nuffield Department of Clinical Neurosciences, John Radcliffe Hospital, Headley Way, Headington, Oxford OX3 9DU, UK
| | - Máiréad P Healy
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK; Department of Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, UK
| | - Paul A Thompson
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK
| | - Dorothy V M Bishop
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK
| | - Kate E Watkins
- Department of Experimental Psychology & Wellcome Trust Centre for Integrative Neuroimaging, University of Oxford, Anna Watts Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford, OX2 6GG, UK
| |
Collapse
|
10
|
Abstract
This study evaluated the performance of two alternate computer access methods that could be used for two-dimensional cursor control. The first method, ACC/sEMG, integrates head acceleration and facial surface electromyography. The second method, Camera Mouse, is a free-to-use, computer vision-based access method. Twenty-four healthy adults performed a target acquisition task using each computer access method across two lighting conditions and three computer orientations. Performance in the task was evaluated using metrics of target selection accuracy, movement time, and path efficiency. Using ACC/sEMG resulted in better mean path efficiency and target selection accuracy, whereas using Camera Mouse resulted in faster target selection. Moreover, performance in the task when using Camera Mouse depended on lighting conditions in the room. The findings of this study show that the ACC/sEMG system is an effective computer access method across different lighting conditions and computer orientations. However, there is a tradeoff between speed and accuracy: ACC/sEMG system provided higher target selection accuracy compared to Camera Mouse, while the latter provided faster target selection. Future development should focus on evaluating performance of each method in populations with limited motor abilities.
Collapse
|
11
|
Vojtech JM, Noordzij JP, Cler GJ, Stepp CE. The Effects of Modulating Fundamental Frequency and Speech Rate on the Intelligibility, Communication Efficiency, and Perceived Naturalness of Synthetic Speech. Am J Speech Lang Pathol 2019; 28:875-886. [PMID: 31306599 PMCID: PMC6802860 DOI: 10.1044/2019_ajslp-msc18-18-0052] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/16/2018] [Revised: 09/28/2018] [Accepted: 02/04/2019] [Indexed: 06/10/2023]
Abstract
Purpose This study investigated how modulating fundamental frequency (f0) and speech rate differentially impact the naturalness, intelligibility, and communication efficiency of synthetic speech. Method Sixteen sentences of varying prosodic content were developed via a speech synthesizer. The f0 contour and speech rate of these sentences were altered to produce 4 stimulus sets: (a) normal rate with a fixed f0 level, (b) slow rate with a fixed f0 level, (c) normal rate with prosodically natural f0 variation, and (d) normal rate with prosodically unnatural f0 variation. Sixteen listeners provided orthographic transcriptions and judgments of naturalness for these stimuli. Results Sentences with f0 variation were rated as more natural than those with a fixed f0 level. Conversely, sentences with a fixed f0 level demonstrated higher intelligibility than those with f0 variation. Speech rate did not affect the intelligibility of stimuli with a fixed f0 level. Communication efficiency was highest for sentences produced at a normal rate and a fixed f0 level. Conclusions Sentence-level f0 variation increased naturalness ratings of synthesized speech, whether the variation was prosodically natural or not. However, these f0 variations reduced intelligibility. There is evidence of a trade-off in naturalness and intelligibility of synthesized speech, which may impact future speech synthesis designs. Supplemental Material https://doi.org/10.23641/asha.8847833.
Collapse
Affiliation(s)
- Jennifer M. Vojtech
- Department of Biomedical Engineering, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
| | - Jacob P. Noordzij
- Department of Biomedical Engineering, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
| | - Gabriel J. Cler
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
| | - Cara E. Stepp
- Department of Biomedical Engineering, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Otolaryngology–Head and Neck Surgery, Boston University School of Medicine, MA
| |
Collapse
|
12
|
Cler GJ, Kolin KR, Noordzij JP, Vojtech JM, Fager SK, Stepp CE. Optimized and Predictive Phonemic Interfaces for Augmentative and Alternative Communication. J Speech Lang Hear Res 2019; 62:2065-2081. [PMID: 31306607 PMCID: PMC6808364 DOI: 10.1044/2019_jslhr-s-msc18-18-0187] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/17/2018] [Revised: 12/06/2018] [Accepted: 03/15/2019] [Indexed: 06/10/2023]
Abstract
Purpose We empirically assessed the results of computational optimization and prediction in communication interfaces that were designed to allow individuals with severe motor speech disorders to select phonemes and generate speech output. Method Interface layouts were either random or optimized, in which phoneme targets that were likely to be selected together were located in proximity. Target sizes were either static or predictive, such that likely targets were dynamically enlarged following each selection. Communication interfaces were evaluated by 36 users without motor impairments using an alternate access method. Each user was assigned to 1 of 4 interfaces varying in layout and whether prediction was implemented (random/static, random/predictive, optimized/static, optimized/predictive) and participated in 12 sessions over a 3-week period. Six participants with severe motor impairments used both the optimized/static and optimized/predictive interfaces in 1-2 sessions. Results In individuals without motor impairments, prediction provided significantly faster communication rates during training (Sessions 1-9), as users were learning the interface target locations and the novel access method. After training, optimization acted to significantly increase communication rates. The optimization likely became relevant only after training when participants knew the target locations and moved directly to the targets. Participants with motor impairments could use the interfaces with alternate access methods and generally rated the interface with prediction as preferred. Conclusions Optimization and prediction led to increases in communication rates in users without motor impairments. Predictive interfaces were preferred by users with motor impairments. Future research is needed to translate these results into clinical practice. Supplemental Material https://doi.org/10.23641/asha.8636948.
Collapse
Affiliation(s)
- Gabriel J. Cler
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
| | - Katharine R. Kolin
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
| | - Jacob P. Noordzij
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
| | - Jennifer M. Vojtech
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
| | - Susan K. Fager
- Institute for Rehabilitation Science and Engineering, Madonna Rehabilitation Hospital, Lincoln, NE
| | - Cara E. Stepp
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
- Department of Otolaryngology—Head and Neck Surgery, Boston University School of Medicine, MA
| |
Collapse
|
13
|
Cler GJ, McKenna VS, Dahl KL, Stepp CE. Longitudinal Case Study of Transgender Voice Changes Under Testosterone Hormone Therapy. J Voice 2019; 34:748-762. [PMID: 30987859 DOI: 10.1016/j.jvoice.2019.03.006] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2018] [Revised: 03/14/2019] [Accepted: 03/14/2019] [Indexed: 10/27/2022]
Abstract
The purpose of this study was to comprehensively evaluate voice and speech changes in one healthy 30-year-old transgender male undergoing testosterone therapy for transition. Testing occurred at three timepoints before cross-sex hormone therapy and every 2 weeks thereafter for 1 year. Data collected included measures of acoustics, aerodynamics, and laryngeal structure and function via flexible laryngoscopy. Analysis included acoustic correlates of pitch, loudness, voice quality, and vocal tract length, as well as perceptual measures of voice quality and gender. Speaking fundamental frequency (fo) lowered from 183 Hz to 134 Hz. Phonatory frequency range (ie, minimum and maximum singing range) shifted from a range of D#3-E6 to a range of A2-A5. Perceptual measures of voice quality indicated no negative changes. Naïve listeners reliably rated the participant's speech samples as male after 37 weeks on testosterone. Few studies document in detail the variety of voice changes that occur during cross-sex hormone therapy, focusing instead on fo alone. This study adds to the literature a comprehensive case study of speech and voice changes experienced by one transmasculine participant undergoing testosterone therapy.
Collapse
Affiliation(s)
- Gabriel J Cler
- Graduate Program for Neuroscience, Boston University, Boston, Massachusetts; Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts.
| | - Victoria S McKenna
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts
| | - Kimberly L Dahl
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts
| | - Cara E Stepp
- Graduate Program for Neuroscience, Boston University, Boston, Massachusetts; Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts; Department of Biomedical Engineering, Boston University, Boston, Massachusetts; Department of Otolaryngology - Head and Neck Surgery, Boston University School of Medicine, Boston, Massachusetts
| |
Collapse
|
14
|
Vojtech JM, Cler GJ, Stepp CE. Prediction of Optimal Facial Electromyographic Sensor Configurations for Human-Machine Interface Control. IEEE Trans Neural Syst Rehabil Eng 2018; 26:1566-1576. [PMID: 29994124 DOI: 10.1109/tnsre.2018.2849202] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/09/2023]
Abstract
Surface electromyography (sEMG) is a promising computer access method for individuals with motor impairments. However, optimal sensor placement is a tedious task requiring trial-and-error by an expert, particularly when recording from facial musculature likely to be spared in individuals with neurological impairments. We sought to reduce the sEMG sensor configuration complexity by using quantitative signal features extracted from a short calibration task to predict human-machine interface (HMI) performance. A cursor control system allowed individuals to activate specific sEMG-targeted muscles to control an onscreen cursor and navigate a target selection task. The task was repeated for a range of sensor configurations to elicit a range of signal qualities. Signal features were extracted from the calibration of each configuration and examined via a principle component factor analysis in order to predict the HMI performance during subsequent tasks. Feature components most influenced by the energy and the complexity of the EMG signal and muscle activity between the sensors were significantly predictive of the HMI performance. However, configuration order had a greater effect on performance than the configurations, suggesting that non-experts can place sEMG sensors in the vicinity of usable muscle sites for computer access and healthy individuals will learn to efficiently control the HMI system.
Collapse
|
15
|
Heller Murray ES, Michener CM, Enflo L, Cler GJ, Stepp CE. The Impact of Glottal Configuration on Speech Breathing. J Voice 2017; 32:420-427. [PMID: 28838793 DOI: 10.1016/j.jvoice.2017.07.001] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2017] [Revised: 06/28/2017] [Accepted: 07/05/2017] [Indexed: 11/18/2022]
Abstract
OBJECTIVE The purpose of this study was to examine whether changes in respiratory patterns occurred in response to volitional changes in glottal configuration. METHODS Twelve vocally healthy participants read a passage while wearing the Inductotrace respiratory inductive plethysmograph, which measures the excursions of the rib cage and abdomen. Participants read the passage 5 times in a typical speaking voice (baseline phase), 10 times in an experimental voice, which was similar to a breathy vocal quality (experimental phase), and 5 times again in a typical speaking voice (return phase). Kinematic estimates of lung volume (LV) initiation, LV termination, and LV excursion were collected for each speech breath. RESULTS Participants spoke with larger LV excursions during the experimental phase, characterized by increased LV initiation and decreased LV termination compared with the baseline phase. CONCLUSION In response to volitional changes in glottal configuration, healthy individuals spoke with increased LV excursion. They both responded to changes (decreasing LV termination) and planned for more efficient future utterances (increasing LV initiation) during the experimental phase. This study demonstrated that respiratory patterns change in response to changes in glottal configuration; future work will examine these patterns in individuals with voice disorders.
Collapse
Affiliation(s)
| | - Carolyn M Michener
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts
| | - Laura Enflo
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts
| | - Gabriel J Cler
- Graduate Program for Neuroscience-Computational, Boston University, Boston, Massachusetts
| | - Cara E Stepp
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts; Department of Biomedical Engineering, Boston University, Boston, Massachusetts; Department of Otolaryngology-Head and Neck Surgery, Boston University School of Medicine, Boston, Massachusetts
| |
Collapse
|
16
|
Cler GJ, Mittelman T, Braden MN, Woodnorth GH, Stepp CE. Video Game Rehabilitation of Velopharyngeal Dysfunction: A Case Series. J Speech Lang Hear Res 2017; 60:1800-1809. [PMID: 28655049 PMCID: PMC5544406 DOI: 10.1044/2017_jslhr-s-16-0231] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/14/2016] [Revised: 10/12/2016] [Accepted: 11/22/2016] [Indexed: 06/07/2023]
Abstract
PURPOSE Video games provide a promising platform for rehabilitation of speech disorders. Although video games have been used to train speech perception in foreign language learners and have been proposed for aural rehabilitation, their use in speech therapy has been limited thus far. We present feasibility results from at-home use in a case series of children with velopharyngeal dysfunction (VPD) using an interactive video game that provided real-time biofeedback to facilitate appropriate nasalization. METHOD Five participants were recruited across a range of ages, VPD severities, and VPD etiologies. Participants completed multiple weeks of individual game play with a video game that provides feedback on nasalization measured via nasal accelerometry. Nasalization was assessed before and after training by using nasometry, aerodynamic measures, and expert perceptual judgments. RESULTS Four participants used the game at home or school, with the remaining participant unwilling to have the nasal accelerometer secured to his nasal skin, perhaps due to his young age. The remaining participants showed a tendency toward decreased nasalization after training, particularly for the words explicitly trained in the video game. CONCLUSION Results suggest that video game-based systems may provide a useful rehabilitation platform for providing real-time feedback of speech nasalization in VPD. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.5116828.
Collapse
Affiliation(s)
- Gabriel J. Cler
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
| | - Talia Mittelman
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
| | - Maia N. Braden
- Department of Surgery, Division of Otolaryngology, Voice and Swallow Clinics, University of Wisconsin, Madison
| | - Geralyn Harvey Woodnorth
- Speech-Language Pathology Program, Otolaryngology and Communication Enhancement, Boston Children's Hospital, MA
| | - Cara E. Stepp
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
- Department of Otolaryngology–Head and Neck Surgery, Boston University School of Medicine, MA
| |
Collapse
|
17
|
Cler GJ, Lee JC, Mittelman T, Stepp CE, Bohland JW. Kinematic Analysis of Speech Sound Sequencing Errors Induced by Delayed Auditory Feedback. J Speech Lang Hear Res 2017; 60:1695-1711. [PMID: 28655038 PMCID: PMC5544401 DOI: 10.1044/2017_jslhr-s-16-0234] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/14/2016] [Revised: 10/07/2016] [Accepted: 11/16/2016] [Indexed: 06/07/2023]
Abstract
PURPOSE Delayed auditory feedback (DAF) causes speakers to become disfluent and make phonological errors. Methods for assessing the kinematics of speech errors are lacking, with most DAF studies relying on auditory perceptual analyses, which may be problematic, as errors judged to be categorical may actually represent blends of sounds or articulatory errors. METHOD Eight typical speakers produced nonsense syllable sequences under normal and DAF (200 ms). Lip and tongue kinematics were captured with electromagnetic articulography. Time-locked acoustic recordings were transcribed, and the kinematics of utterances with and without perceived errors were analyzed with existing and novel quantitative methods. RESULTS New multivariate measures showed that for 5 participants, kinematic variability for productions perceived to be error free was significantly increased under delay; these results were validated by using the spatiotemporal index measure. Analysis of error trials revealed both typical productions of a nontarget syllable and productions with articulatory kinematics that incorporated aspects of both the target and the perceived utterance. CONCLUSIONS This study is among the first to characterize articulatory changes under DAF and provides evidence for different classes of speech errors, which may not be perceptually salient. New methods were developed that may aid visualization and analysis of large kinematic data sets. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.5103067.
Collapse
Affiliation(s)
- Gabriel J. Cler
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
| | | | | | - Cara E. Stepp
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Biomedical Engineering, Boston University, MA
- School of Medicine, Department of Otolaryngology–Head and Neck Surgery, Boston University, MA
| | - Jason W. Bohland
- Graduate Program for Neuroscience–Computational Neuroscience, Boston University, MA
- Department of Speech, Language, and Hearing Sciences, Boston University, MA
- Department of Health Sciences, Boston University, MA
| |
Collapse
|
18
|
Cler GJ, Nieto-Castañón A, Guenther FH, Fager SK, Stepp CE. Surface electromyographic control of a novel phonemic interface for speech synthesis. Augment Altern Commun 2016; 32:120-30. [PMID: 27141992 DOI: 10.3109/07434618.2016.1170205] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022] Open
Abstract
Many individuals with minimal movement capabilities use AAC to communicate. These individuals require both an interface with which to construct a message (e.g., a grid of letters) and an input modality with which to select targets. This study evaluated the interaction of two such systems: (a) an input modality using surface electromyography (sEMG) of spared facial musculature, and (b) an onscreen interface from which users select phonemic targets. These systems were evaluated in two experiments: (a) participants without motor impairments used the systems during a series of eight training sessions, and (b) one individual who uses AAC used the systems for two sessions. Both the phonemic interface and the electromyographic cursor show promise for future AAC applications.
Collapse
Affiliation(s)
- Gabriel J Cler
- a Graduate Program for Neuroscience - Computational Neuroscience , Boston University , MA , USA
| | - Alfonso Nieto-Castañón
- b Department of Speech, Language, and Hearing Sciences , Boston University , Boston , MA , USA
| | - Frank H Guenther
- b Department of Speech, Language, and Hearing Sciences , Boston University , Boston , MA , USA ;,c Department of Biomedical Engineering , Boston University , Boston , MA , USA
| | - Susan K Fager
- d Institute for Rehabilitation Science and Engineering , Madonna Rehabilitation Hospital , Lincoln , NE , USA
| | - Cara E Stepp
- b Department of Speech, Language, and Hearing Sciences , Boston University , Boston , MA , USA ;,c Department of Biomedical Engineering , Boston University , Boston , MA , USA
| |
Collapse
|
19
|
Cler GJ, Stepp CE. Discrete Versus Continuous Mapping of Facial Electromyography for Human-Machine Interface Control: Performance and Training Effects. IEEE Trans Neural Syst Rehabil Eng 2015; 23:572-80. [PMID: 25616053 DOI: 10.1109/tnsre.2015.2391054] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Individuals with high spinal cord injuries are unable to operate a keyboard and mouse with their hands. In this experiment, we compared two systems using surface electromyography (sEMG) recorded from facial muscles to control an onscreen keyboard to type five-letter words. Both systems used five sEMG sensors to capture muscle activity during five distinct facial gestures that were mapped to five cursor commands: move left, move right, move up, move down, and "click". One system used a discrete movement and feedback algorithm in which the user produced one quick facial gesture, causing a corresponding discrete movement to an adjacent letter. The other system was continuously updated and allowed the user to control the cursor's velocity by relative activation between different sEMG channels. Participants were trained on one system for four sessions on consecutive days, followed by one crossover session on the untrained system. Information transfer rates (ITRs) were high for both systems compared to other potential input modalities, both initially and with training (Session 1: 62.1 bits/min, Session 4: 105.1 bits/min). Users of the continuous system showed significantly higher ITRs than the discrete users. Future development will focus on improvements to both systems, which may offer differential advantages for users with various motor impairments.
Collapse
|