1
|
Gordienko Y, Gordienko N, Taran V, Rojbi A, Telenyk S, Stirenko S. Effect of natural and synthetic noise data augmentation on physical action classification by brain-computer interface and deep learning. Front Neuroinform 2025; 19:1521805. [PMID: 40083893 PMCID: PMC11903462 DOI: 10.3389/fninf.2025.1521805] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2024] [Accepted: 02/13/2025] [Indexed: 03/16/2025] Open
Abstract
Analysis of electroencephalography (EEG) signals gathered by brain-computer interface (BCI) recently demonstrated that deep neural networks (DNNs) can be effectively used for investigation of time sequences for physical actions (PA) classification. In this study, the relatively simple DNN with fully connected network (FCN) components and convolutional neural network (CNN) components was considered to classify finger-palm-hand manipulations each from the grasp-and-lift (GAL) dataset. The main aim of this study was to imitate and investigate environmental influence by the proposed noise data augmentation (NDA) of two kinds: (i) natural NDA by inclusion of noise EEG data from neighboring regions by increasing the sampling size N and the different offset values for sample labeling and (ii) synthetic NDA by adding the generated Gaussian noise. The natural NDA by increasing N leads to the higher micro and macro area under the curve (AUC) for receiver operating curve values for the bigger N values than usage of synthetic NDA. The detrended fluctuation analysis (DFA) was applied to investigate the fluctuation properties and calculate the correspondent Hurst exponents H for the quantitative characterization of the fluctuation variability. H values for the low time window scales (< 2 s) are higher in comparison with ones for the bigger time window scales. For example, H more than 2-3 times higher for some PAs, i.e., it means that the shorter EEG fragments (< 2 s) demonstrate the scaling behavior of the higher complexity than the longer fragments. As far as these results were obtained by the relatively small DNN with the low resource requirements, this approach can be promising for porting such models to Edge Computing infrastructures on devices with the very limited computational resources.
Collapse
Affiliation(s)
- Yuri Gordienko
- Computer Engineering Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute,”Kyiv, Ukraine
| | - Nikita Gordienko
- Computer Engineering Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute,”Kyiv, Ukraine
| | - Vladyslav Taran
- Computer Engineering Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute,”Kyiv, Ukraine
| | - Anis Rojbi
- Laboratoire Cognitions Humaine et Artificielle, Université Paris 8, Paris, France
| | - Sergii Telenyk
- Computer Engineering Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute,”Kyiv, Ukraine
- Department of Automation and Computer Science, Faculty of Electrical and Computer Engineering, Cracow University of Technology, Cracow, Poland
| | - Sergii Stirenko
- Computer Engineering Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute,”Kyiv, Ukraine
| |
Collapse
|
2
|
Hjortkjær J, Wong DDE, Catania A, Märcher-Rørsted J, Ceolini E, Fuglsang SA, Kiselev I, Di Liberto G, Liu SC, Dau T, Slaney M, de Cheveigné A. Real-time control of a hearing instrument with EEG-based attention decoding. J Neural Eng 2025; 22:016027. [PMID: 39996608 DOI: 10.1088/1741-2552/ad867c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Accepted: 10/14/2024] [Indexed: 02/26/2025]
Abstract
Enhancing speech perception in everyday noisy acoustic environments remains an outstanding challenge for hearing aids. Speech separation technology is improving rapidly, but hearing devices cannot fully exploit this advance without knowing which sound sources the user wants to hear. Even with high-quality source separation, the hearing aid must know which speech streams to enhance and which to suppress. Advances in EEG-based decoding of auditory attention raise the potential of neurosteering, in which a hearing instrument selectively enhances the sound sources that a hearing-impaired listener is focusing their attention on. Here, we present and discuss a real-time brain-computer interface system that combines a stimulus-response model based on canonical correlation analysis for real-time EEG attention decoding, coupled with a multi-microphone hardware platform enabling low-latency real-time speech separation through spatial beamforming. We provide an overview of the system and its various components, discuss prospects and limitations of the technology, and illustrate its application with case studies of listeners steering acoustic feedback of competing speech streams via real-time attention decoding. A software implementation code of the system is publicly available for further research and explorations.
Collapse
Affiliation(s)
- Jens Hjortkjær
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
- Danish Research Centre for Magnetic Resonance, Centre for Functional and Diagnostic Imaging and Research, Copenhagen University Hospital-Amager and Hvidovre, Copenhagen, Denmark
| | - Daniel D E Wong
- Laboratoire des Systèmes Perceptifs, CNRS UMR, Paris 8248, France
- Département d'Etudes Cognitives, Ecole Normale Supérieure, PSL, Paris, France
| | - Alessandro Catania
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Jonatan Märcher-Rørsted
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Enea Ceolini
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Søren A Fuglsang
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
- Danish Research Centre for Magnetic Resonance, Centre for Functional and Diagnostic Imaging and Research, Copenhagen University Hospital-Amager and Hvidovre, Copenhagen, Denmark
| | - Ilya Kiselev
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Giovanni Di Liberto
- School of Computer Science and Statistics, Institute of Neuroscience, Trinity College, The University of Dublin, Dublin, Ireland
| | - Shih-Chii Liu
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Kgs. Lyngby, Denmark
| | - Malcolm Slaney
- Center for Computer Research in Music and Acoustics (CCRMA), Stanford University, Stanford, CA, United States of America
| | - Alain de Cheveigné
- Laboratoire des Systèmes Perceptifs, CNRS UMR, Paris 8248, France
- Département d'Etudes Cognitives, Ecole Normale Supérieure, PSL, Paris, France
| |
Collapse
|
3
|
Geirnaert S, Yao Y, Francart T, Bertrand A. Stimulus-Informed Generalized Canonical Correlation Analysis for Group Analysis of Neural Responses to Natural Stimuli. IEEE J Biomed Health Inform 2025; 29:970-983. [PMID: 39292590 DOI: 10.1109/jbhi.2024.3462991] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/20/2024]
Abstract
Various new brain-computer interface technologies or neuroscience applications require decoding stimulus-following neural responses to natural stimuli such as speech and video from, e.g., electroencephalography (EEG) signals. In this context, generalized canonical correlation analysis (GCCA) is often used as a group analysis technique, which allows the extraction of correlated signal components from the neural activity of multiple subjects attending to the same stimulus. GCCA can be used to improve the signal-to-noise ratio of the stimulus-following neural responses relative to all other irrelevant (non-)neural activity, or to quantify the correlated neural activity across multiple subjects in a group-wise coherence metric. However, the traditional GCCA technique is stimulus-unaware: no information about the stimulus is used to estimate the correlated components from the neural data of several subjects. Therefore, the GCCA technique might fail to extract relevant correlated signal components in practical situations where the amount of information is limited, for example, because of a limited amount of training data or group size. This motivates a new stimulus-informed GCCA (SI-GCCA) framework that allows taking the stimulus into account to extract the correlated components. We show that SI-GCCA outperforms GCCA in various practical settings, for both auditory and visual stimuli. Moreover, we showcase how SI-GCCA can be used to steer the estimation of the components towards the stimulus. As such, SI-GCCA substantially improves upon GCCA for various purposes, ranging from preprocessing to quantifying attention.
Collapse
|
4
|
Ronca V, Capotorto R, Di Flumeri G, Giorgi A, Vozzi A, Germano D, Virgilio VD, Borghini G, Cartocci G, Rossi D, Inguscio BMS, Babiloni F, Aricò P. Optimizing EEG Signal Integrity: A Comprehensive Guide to Ocular Artifact Correction. Bioengineering (Basel) 2024; 11:1018. [PMID: 39451394 PMCID: PMC11505294 DOI: 10.3390/bioengineering11101018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Revised: 09/30/2024] [Accepted: 10/08/2024] [Indexed: 10/26/2024] Open
Abstract
Ocular artifacts, including blinks and saccades, pose significant challenges in the analysis of electroencephalographic (EEG) data, often obscuring crucial neural signals. This tutorial provides a comprehensive guide to the most effective methods for correcting these artifacts, with a focus on algorithms designed for both laboratory and real-world settings. We review traditional approaches, such as regression-based techniques and Independent Component Analysis (ICA), alongside more advanced methods like Artifact Subspace Reconstruction (ASR) and deep learning-based algorithms. Through detailed step-by-step instructions and comparative analysis, this tutorial equips researchers with the tools necessary to maintain the integrity of EEG data, ensuring accurate and reliable results in neurophysiological studies. The strategies discussed are particularly relevant for wearable EEG systems and real-time applications, reflecting the growing demand for robust and adaptable solutions in applied neuroscience.
Collapse
Affiliation(s)
- Vincenzo Ronca
- Department of Computer, Control, and Management Engineering, Sapienza University of Rome, 00185 Roma, Italy; (D.G.); (V.D.V.); (B.M.S.I.); or (P.A.)
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
| | - Rossella Capotorto
- Department of Anatomical, Histological, Forensic and Orthopaedic Sciences, Sapienza University of Rome, 00185 Roma, Italy;
| | - Gianluca Di Flumeri
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
- Department of Molecular Medicine, Sapienza University of Rome, 00185 Roma, Italy;
| | - Andrea Giorgi
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
- Department of Anatomical, Histological, Forensic and Orthopaedic Sciences, Sapienza University of Rome, 00185 Roma, Italy;
| | - Alessia Vozzi
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
| | - Daniele Germano
- Department of Computer, Control, and Management Engineering, Sapienza University of Rome, 00185 Roma, Italy; (D.G.); (V.D.V.); (B.M.S.I.); or (P.A.)
| | - Valerio Di Virgilio
- Department of Computer, Control, and Management Engineering, Sapienza University of Rome, 00185 Roma, Italy; (D.G.); (V.D.V.); (B.M.S.I.); or (P.A.)
| | - Gianluca Borghini
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
- Department of Molecular Medicine, Sapienza University of Rome, 00185 Roma, Italy;
| | - Giulia Cartocci
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
- Department of Molecular Medicine, Sapienza University of Rome, 00185 Roma, Italy;
| | - Dario Rossi
- Department of Molecular Medicine, Sapienza University of Rome, 00185 Roma, Italy;
| | - Bianca M. S. Inguscio
- Department of Computer, Control, and Management Engineering, Sapienza University of Rome, 00185 Roma, Italy; (D.G.); (V.D.V.); (B.M.S.I.); or (P.A.)
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
| | - Fabio Babiloni
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
- Department of Physiology and Pharmacology “Vittorio Erspamer”, Sapienza University of Rome, Piazzale Aldo Moro 5, 00185 Rome, Italy
- College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310005, China
| | - Pietro Aricò
- Department of Computer, Control, and Management Engineering, Sapienza University of Rome, 00185 Roma, Italy; (D.G.); (V.D.V.); (B.M.S.I.); or (P.A.)
- BrainSigns S.r.l., Industrial Neurosciences Lab, 00198 Rome, Italy; (G.D.F.); (A.G.); (A.V.); (G.B.); (G.C.); (F.B.)
| |
Collapse
|
5
|
Herbert C, Northoff G. Editorial: Analyzing and computing humans - the role of language, culture, brain and health. Front Hum Neurosci 2024; 18:1439729. [PMID: 39015823 PMCID: PMC11250248 DOI: 10.3389/fnhum.2024.1439729] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Accepted: 06/06/2024] [Indexed: 07/18/2024] Open
Affiliation(s)
- Cornelia Herbert
- Applied Emotion and Motivation Psychology, Institute of Psychology and Education, Faculty of Engineering, Computer Science and Psychology, Ulm University, Ulm, Germany
| | - Georg Northoff
- Institute of Mental Health Research, University of Ottawa, Ottawa, ON, Canada
| |
Collapse
|
6
|
Simon A, Bech S, Loquet G, Østergaard J. Cortical linear encoding and decoding of sounds: Similarities and differences between naturalistic speech and music listening. Eur J Neurosci 2024; 59:2059-2074. [PMID: 38303522 DOI: 10.1111/ejn.16265] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2023] [Revised: 11/02/2023] [Accepted: 01/12/2024] [Indexed: 02/03/2024]
Abstract
Linear models are becoming increasingly popular to investigate brain activity in response to continuous and naturalistic stimuli. In the context of auditory perception, these predictive models can be 'encoding', when stimulus features are used to reconstruct brain activity, or 'decoding' when neural features are used to reconstruct the audio stimuli. These linear models are a central component of some brain-computer interfaces that can be integrated into hearing assistive devices (e.g., hearing aids). Such advanced neurotechnologies have been widely investigated when listening to speech stimuli but rarely when listening to music. Recent attempts at neural tracking of music show that the reconstruction performances are reduced compared with speech decoding. The present study investigates the performance of stimuli reconstruction and electroencephalogram prediction (decoding and encoding models) based on the cortical entrainment of temporal variations of the audio stimuli for both music and speech listening. Three hypotheses that may explain differences between speech and music stimuli reconstruction were tested to assess the importance of the speech-specific acoustic and linguistic factors. While the results obtained with encoding models suggest different underlying cortical processing between speech and music listening, no differences were found in terms of reconstruction of the stimuli or the cortical data. The results suggest that envelope-based linear modelling can be used to study both speech and music listening, despite the differences in the underlying cortical mechanisms.
Collapse
Affiliation(s)
- Adèle Simon
- Artificial Intelligence and Sound, Department of Electronic Systems, Aalborg University, Aalborg, Denmark
- Research Department, Bang & Olufsen A/S, Struer, Denmark
| | - Søren Bech
- Artificial Intelligence and Sound, Department of Electronic Systems, Aalborg University, Aalborg, Denmark
- Research Department, Bang & Olufsen A/S, Struer, Denmark
| | - Gérard Loquet
- Department of Audiology and Speech Pathology, University of Melbourne, Melbourne, Victoria, Australia
| | - Jan Østergaard
- Artificial Intelligence and Sound, Department of Electronic Systems, Aalborg University, Aalborg, Denmark
| |
Collapse
|
7
|
Álvarez-Meza AM, Torres-Cardona HF, Orozco-Alzate M, Pérez-Nastar HD, Castellanos-Dominguez G. Affective Neural Responses Sonified through Labeled Correlation Alignment. SENSORS (BASEL, SWITZERLAND) 2023; 23:5574. [PMID: 37420740 DOI: 10.3390/s23125574] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Revised: 06/10/2023] [Accepted: 06/11/2023] [Indexed: 07/09/2023]
Abstract
Sound synthesis refers to the creation of original acoustic signals with broad applications in artistic innovation, such as music creation for games and videos. Nonetheless, machine learning architectures face numerous challenges when learning musical structures from arbitrary corpora. This issue involves adapting patterns borrowed from other contexts to a concrete composition objective. Using Labeled Correlation Alignment (LCA), we propose an approach to sonify neural responses to affective music-listening data, identifying the brain features that are most congruent with the simultaneously extracted auditory features. For dealing with inter/intra-subject variability, a combination of Phase Locking Value and Gaussian Functional Connectivity is employed. The proposed two-step LCA approach embraces a separate coupling stage of input features to a set of emotion label sets using Centered Kernel Alignment. This step is followed by canonical correlation analysis to select multimodal representations with higher relationships. LCA enables physiological explanation by adding a backward transformation to estimate the matching contribution of each extracted brain neural feature set. Correlation estimates and partition quality represent performance measures. The evaluation uses a Vector Quantized Variational AutoEncoder to create an acoustic envelope from the tested Affective Music-Listening database. Validation results demonstrate the ability of the developed LCA approach to generate low-level music based on neural activity elicited by emotions while maintaining the ability to distinguish between the acoustic outputs.
Collapse
Affiliation(s)
| | | | - Mauricio Orozco-Alzate
- Signal Processing and Recognition Group, Universidad Nacional de Colombia, Manizales 170003, Colombia
| | - Hernán Darío Pérez-Nastar
- Signal Processing and Recognition Group, Universidad Nacional de Colombia, Manizales 170003, Colombia
| | | |
Collapse
|
8
|
Brain–computer interface in an inter-individual approach using spatial coherence: Identification of better channels and tests repetition using auditory selective attention. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104573] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/30/2023]
|
9
|
Arif S, Munawar S, Ali H. Driving drowsiness detection using spectral signatures of EEG-based neurophysiology. Front Physiol 2023; 14:1153268. [PMID: 37064914 PMCID: PMC10097971 DOI: 10.3389/fphys.2023.1153268] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2023] [Accepted: 03/09/2023] [Indexed: 03/31/2023] Open
Abstract
Introduction: Drowsy driving is a significant factor causing dire road crashes and casualties around the world. Detecting it earlier and more effectively can significantly reduce the lethal aftereffects and increase road safety. As physiological conditions originate from the human brain, so neurophysiological signatures in drowsy and alert states may be investigated for this purpose. In this preface, A passive brain-computer interface (pBCI) scheme using multichannel electroencephalography (EEG) brain signals is developed for spatially localized and accurate detection of human drowsiness during driving tasks.Methods: This pBCI modality acquired electrophysiological patterns of 12 healthy subjects from the prefrontal (PFC), frontal (FC), and occipital cortices (OC) of the brain. Neurological states are recorded using six EEG channels spread over the right and left hemispheres in the PFC, FC, and OC of the sleep-deprived subjects during simulated driving tasks. In post-hoc analysis, spectral signatures of the δ, θ, α, and β rhythms are extracted in terms of spectral band powers and their ratios with a temporal correlation over the complete span of the experiment. Minimum redundancy maximum relevance, Chi-square, and ReliefF feature selection methods are used and aggregated with a Z-score based approach for global feature ranking. The extracted drowsiness attributes are classified using decision trees, discriminant analysis, logistic regression, naïve Bayes, support vector machines, k-nearest neighbors, and ensemble classifiers. The binary classification results are reported with confusion matrix-based performance assessment metrics.Results: In inter-classifier comparison, the optimized ensemble model achieved the best results of drowsiness classification with 85.6% accuracy and precision, 89.7% recall, 87.6% F1-score, 80% specificity, 70.3% Matthews correlation coefficient, 70.2% Cohen’s kappa score, and 91% area under the receiver operating characteristic curve with 76-ms execution time. In inter-channel comparison, the best results were obtained at the F8 electrode position in the right FC of the brain. The significance of all the results was validated with a p-value of less than 0.05 using statistical hypothesis testing methods.Conclusions: The proposed scheme has achieved better results for driving drowsiness detection with the accomplishment of multiple objectives. The predictor importance approach has reduced the feature extraction cost and computational complexity is minimized with the use of conventional machine learning classifiers resulting in low-cost hardware and software requirements. The channel selection approach has spatially localized the most promising brain region for drowsiness detection with only a single EEG channel (F8) which reduces the physical intrusiveness in normal driving operation. This pBCI scheme has a good potential for practical applications requiring earlier, more accurate, and less disruptive drowsiness detection using the spectral information of EEG biosignals.
Collapse
Affiliation(s)
- Saad Arif
- Department of Mechanical Engineering, HITEC University Taxila, Taxila Cantt, Pakistan
| | - Saba Munawar
- Department of Electrical and Computer Engineering, COMSATS University Islamabad, Wah Campus, Wah Cantt, Pakistan
| | - Hashim Ali
- Department of Computer Science, School of Engineering and Digital Sciences, Nazarbayev University, Astana, Kazakhstan
- *Correspondence: Hashim Ali,
| |
Collapse
|
10
|
A neuroscience-inspired spiking neural network for EEG-based auditory spatial attention detection. Neural Netw 2022; 152:555-565. [DOI: 10.1016/j.neunet.2022.05.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2021] [Revised: 03/02/2022] [Accepted: 05/02/2022] [Indexed: 11/18/2022]
|
11
|
Straetmans L, Holtze B, Debener S, Jaeger M, Mirkovic B. Neural tracking to go: auditory attention decoding and saliency detection with mobile EEG. J Neural Eng 2021; 18. [PMID: 34902846 DOI: 10.1088/1741-2552/ac42b5] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2021] [Accepted: 12/13/2021] [Indexed: 11/11/2022]
Abstract
OBJECTIVE Neuro-steered assistive technologies have been suggested to offer a major advancement in future devices like neuro-steered hearing aids. Auditory attention decoding methods would in that case allow for identification of an attended speaker within complex auditory environments, exclusively from neural data. Decoding the attended speaker using neural information has so far only been done in controlled laboratory settings. Yet, it is known that ever-present factors like distraction and movement are reflected in the neural signal parameters related to attention. APPROACH Thus, in the current study we applied a two-competing speaker paradigm to investigate performance of a commonly applied EEG-based auditory attention decoding (AAD) model outside of the laboratory during leisure walking and distraction. Unique environmental sounds were added to the auditory scene and served as distractor events. MAIN RESULTS The current study shows, for the first time, that the attended speaker can be accurately decoded during natural movement. At a temporal resolution of as short as 5-seconds and without artifact attenuation, decoding was found to be significantly above chance level. Further, as hypothesized, we found a decrease in attention to the to-be-attended and the to-be-ignored speech stream after the occurrence of a salient event. Additionally, we demonstrate that it is possible to predict neural correlates of distraction with a computational model of auditory saliency based on acoustic features. CONCLUSION Taken together, our study shows that auditory attention tracking outside of the laboratory in ecologically valid conditions is feasible and a step towards the development of future neural-steered hearing aids.
Collapse
Affiliation(s)
- Lisa Straetmans
- Department of Psychology, Carl von Ossietzky Universität Oldenburg Fakultät für Medizin und Gesundheitswissenschaften, Ammerländer Heerstraße 114-118, Oldenburg, Niedersachsen, 26129, GERMANY
| | - B Holtze
- Department of Psychology, Carl von Ossietzky Universität Oldenburg Fakultät für Medizin und Gesundheitswissenschaften, Ammerländer Heerstr. 114-118, Oldenburg, Niedersachsen, 26129, GERMANY
| | - Stefan Debener
- Department of Psychology, Carl von Ossietzky Universität Oldenburg Fakultät für Medizin und Gesundheitswissenschaften, Ammerländer Heerstr. 114-118, Oldenburg, Niedersachsen, 26129, GERMANY
| | - Manuela Jaeger
- Department of Psychology, Carl von Ossietzky Universität Oldenburg Fakultät für Medizin und Gesundheitswissenschaften, Ammerländer Heerstr. 114-118, Oldenburg, Niedersachsen, 26129, GERMANY
| | - Bojana Mirkovic
- Department of Psychology , Carl von Ossietzky Universität Oldenburg Fakultät für Medizin und Gesundheitswissenschaften, Ammerländer Heerstr. 114-118, Oldenburg, Niedersachsen, 26129, GERMANY
| |
Collapse
|