1
|
McKearney RM, Simpson DM, Bell SL. Automated wave labelling of the auditory brainstem response using machine learning. Int J Audiol 2024:1-6. [PMID: 39363648 DOI: 10.1080/14992027.2024.2404537] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2024] [Revised: 07/26/2024] [Accepted: 09/05/2024] [Indexed: 10/05/2024]
Abstract
OBJECTIVE To compare the performance of a selection of machine learning algorithms, trained to label peaks I, III, and V of the auditory brainstem response (ABR) waveform. An additional algorithm was trained to provide a confidence measure related to the ABR wave latency estimates. DESIGN Secondary data analysis of a previously published ABR dataset. Five types of machine learning algorithm were compared within a nested k-fold cross-validation procedure. STUDY SAMPLE A set of 482 suprathreshold ABR waveforms were used. These were recorded from 81 participants with audiometric thresholds within normal limits. RESULTS A convolutional recurrent neural network (CRNN) outperformed the other algorithms evaluated. The algorithm labelled 95.9% of ABR waves within ±0.1 ms of the target. The mean absolute error was 0.025 ms, averaged across the outer validation folds of the nested cross-validation procedure. High confidence levels were generally associated with greater wave-labelling accuracy. CONCLUSIONS Machine learning algorithms have the potential to assist clinicians with ABR interpretation. The present work identifies a promising machine learning approach, but any algorithm to be used in clinical practice would need to be trained on a large, accurately labelled, heterogeneous dataset and evaluated in clinical settings in follow-on work.
Collapse
Affiliation(s)
- Richard M McKearney
- Institute of Sound and Vibration Research, Faculty of Engineering and Physical Sciences, University of Southampton, Southampton, UK
| | - David M Simpson
- Institute of Sound and Vibration Research, Faculty of Engineering and Physical Sciences, University of Southampton, Southampton, UK
| | - Steven L Bell
- Institute of Sound and Vibration Research, Faculty of Engineering and Physical Sciences, University of Southampton, Southampton, UK
| |
Collapse
|
2
|
Wimalarathna H, Ankmnal-Veeranna S, Allan C, Agrawal SK, Samarabandu J, Ladak HM, Allen P. Machine learning approaches used to analyze auditory evoked responses from the human auditory brainstem: A systematic review. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 226:107118. [PMID: 36122495 DOI: 10.1016/j.cmpb.2022.107118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/08/2022] [Revised: 08/01/2022] [Accepted: 09/06/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND The application of machine learning algorithms for assessing the auditory brainstem response has gained interest over recent years with a considerable number of publications in the literature. In this systematic review, we explore how machine learning has been used to develop algorithms to assess auditory brainstem responses. A clear and comprehensive overview is provided to allow clinicians and researchers to explore the domain and the potential translation to clinical care. METHODS The systematic review was performed based on PRISMA guidelines. A search was conducted of PubMed, IEEE-Xplore, and Scopus databases focusing on human studies that have used machine learning to assess auditory brainstem responses. The duration of the search was from January 1, 1990, to April 3, 2021. The Covidence systematic review platform (www.covidence.org) was used throughout the process. RESULTS A total of 5812 studies were found through the database search and 451 duplicates were removed. The title and abstract screening process further reduced the article count to 89 and in the proceeding full-text screening, 34 articles met our full inclusion criteria. CONCLUSION Three categories of applications were found, namely neurologic diagnosis, hearing threshold estimation, and other (does not relate to neurologic or hearing threshold estimation). Neural networks and support vector machines were the most commonly used machine learning algorithms in all three categories. Only one study had conducted a clinical trial to evaluate the algorithm after development. Challenges remain in the amount of data required to train machine learning models. Suggestions for future research avenues are mentioned with recommended reporting methods for researchers.
Collapse
Affiliation(s)
- Hasitha Wimalarathna
- Department of Electrical & Computer Engineering, Western University, London, Ontario, Canada; National Centre for Audiology, Western University, London, Ontario, Canada.
| | - Sangamanatha Ankmnal-Veeranna
- National Centre for Audiology, Western University, London, Ontario, Canada; College of Nursing and Health Professions, School of Speech and Hearing Sciences, The University of Southern Mississippi, J.B. George Building, Hattiesburg, MS, USA
| | - Chris Allan
- National Centre for Audiology, Western University, London, Ontario, Canada; School of Communication Sciences & Disorders, Western University, London, Ontario, Canada
| | - Sumit K Agrawal
- Department of Electrical & Computer Engineering, Western University, London, Ontario, Canada; National Centre for Audiology, Western University, London, Ontario, Canada; School of Biomedical Engineering, Western University, London, Ontario, Canada; Department of Medical Biophysics, Western University, London, Ontario, Canada; Department of Otolaryngology - Head and Neck Surgery, Western University, London, Ontario, Canada
| | - Jagath Samarabandu
- Department of Electrical & Computer Engineering, Western University, London, Ontario, Canada
| | - Hanif M Ladak
- Department of Electrical & Computer Engineering, Western University, London, Ontario, Canada; National Centre for Audiology, Western University, London, Ontario, Canada; School of Biomedical Engineering, Western University, London, Ontario, Canada; Department of Medical Biophysics, Western University, London, Ontario, Canada; Department of Otolaryngology - Head and Neck Surgery, Western University, London, Ontario, Canada
| | - Prudence Allen
- National Centre for Audiology, Western University, London, Ontario, Canada; School of Communication Sciences & Disorders, Western University, London, Ontario, Canada
| |
Collapse
|
3
|
Trzaskowski B, Jedrzejczak WW, Pilka E, Kochanek K, Skarzynski H. Automatic removal of sonomotor waves from auditory brainstem responses. Comput Biol Med 2013; 43:524-32. [PMID: 23566398 DOI: 10.1016/j.compbiomed.2013.02.016] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2011] [Revised: 02/13/2013] [Accepted: 02/15/2013] [Indexed: 10/27/2022]
Abstract
We have developed a computerized technique for automatic detection and removal of sonomotor waves (SMWs) from auditory brainstem responses (ABRs). Our approach is based on adaptive decomposition using a redundant set of Gaussian and 1-cycle-limited Gabor functions. In order to find optimal parameters and evaluate the efficiency of the methods, simulated data were first used before applying it to clinical data. Results were good and confirmed by an expert with years of clinical experience in ABR evaluation.
Collapse
Affiliation(s)
- B Trzaskowski
- Institute of Physiology and Pathology of Hearing, Warszawa, Poland.
| | | | | | | | | |
Collapse
|
4
|
Acır N, Erkan Y, Bahtiyar YA. Auditory brainstem response classification for threshold detection using estimated evoked potential data: comparison with ensemble averaged data. Neural Comput Appl 2011. [DOI: 10.1007/s00521-011-0776-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
5
|
|
6
|
Vedel-Larsen E, Fuglø J, Channir F, Thomsen CE, Sørensen HBD. A comparative study between a simplified Kalman filter and Sliding Window Averaging for single trial dynamical estimation of event-related potentials. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2010; 99:252-260. [PMID: 20227130 DOI: 10.1016/j.cmpb.2009.12.006] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/18/2008] [Revised: 10/01/2009] [Accepted: 12/18/2009] [Indexed: 05/28/2023]
Abstract
The classical approach for extracting event-related potentials (ERPs) from the brain is ensemble averaging. For long latency ERPs this is not optimal, partly due to the time-delay in obtaining a response and partly because the latency and amplitude for the ERP components, like the P300, are variable and depend on cognitive function. This study compares the performance of a simplified Kalman filter with Sliding Window Averaging in tracking dynamical changes in single trial P300. The comparison is performed on simulated P300 data with added background noise consisting of both simulated and real background EEG in various input signal to noise ratios. While both methods can be applied to track dynamical changes, the simplified Kalman filter has an advantage over the Sliding Window Averaging, most notable in a better noise suppression when both are optimized for faster changing latency and amplitude in the P300 component and in a considerably higher robustness towards suboptimal settings. The latter is of great importance in a clinical setting where the optimal setting cannot be determined.
Collapse
Affiliation(s)
- Esben Vedel-Larsen
- Department of Electrical Engineering, Technical University of Denmark, 2800, Lyngby, Denmark
| | | | | | | | | |
Collapse
|
7
|
Davey R, McCullagh P, Lightbody G, McAllister G. Auditory brainstem response classification: A hybrid model using time and frequency features. Artif Intell Med 2007; 40:1-14. [PMID: 16930965 DOI: 10.1016/j.artmed.2006.07.001] [Citation(s) in RCA: 15] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2006] [Revised: 06/23/2006] [Accepted: 07/03/2006] [Indexed: 11/28/2022]
Abstract
OBJECTIVE The auditory brainstem response (ABR) is an evoked response obtained from brain electrical activity when an auditory stimulus is applied to the ear. An audiologist can determine the threshold level of hearing by applying stimuli at reducing levels of intensity, and can also diagnose various otological, audiological, and neurological abnormalities by examining the morphology of the waveform and the latencies of the individual waves. This is a subjective process requiring considerable expertise. The aim of this research was to develop software classification models to assist the audiologist with an automated detection of the ABR waveform and also to provide objectivity and consistency in this detection. MATERIALS AND METHODS The dataset used in this study consisted of 550 waveforms derived from tests using a range of stimulus levels applied to 85 subjects ranging in hearing ability. Each waveform had been classified by a human expert as 'response=Yes' or 'response=No'. Individual software classification models were generated using time, frequency and cross-correlation measures. Classification employed both artificial neural networks (NNs) and the C5.0 decision tree algorithm. Accuracies were validated using six-fold cross-validation, and by randomising training, validation and test datasets. RESULTS The result was a two stage classification process whereby strong responses were classified to an accuracy of 95.6% in the first stage. This used a ratio of post-stimulus to pre-stimulus power in the time domain, with power measures at 200, 500 and 900Hz in the frequency domain. In the second stage, outputs from time, frequency and cross-correlation classifiers were combined using the Dempster-Shafer method to produce a hybrid model with an accuracy of 85% (126 repeat waveforms). CONCLUSION By combining the different approaches a hybrid system has been created that emulates the approach used by an audiologist in analysing an ABR waveform. Interpretation did not rely on one particular feature but brought together power and frequency analysis as well as consistency of subaverages. This provided a system that enhanced robustness to artefacts while maintaining classification accuracy.
Collapse
Affiliation(s)
- Robert Davey
- Department of Language and Communication Science, City University, Northampton Square, London EC1V 0HB, UK
| | | | | | | |
Collapse
|
8
|
Zhang R, McAllister G, Scotney B, McClean S, Houston G. Combining Wavelet Analysis and Bayesian Networks for the Classification of Auditory Brainstem Response. ACTA ACUST UNITED AC 2006; 10:458-67. [PMID: 16871712 DOI: 10.1109/titb.2005.863865] [Citation(s) in RCA: 35] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
The auditory brainstem response (ABR) has become a routine clinical tool for hearing and neurological assessment. In order to pick out the ABR from the background EEG activity that obscures it, stimulus-synchronized averaging of many repeated trials is necessary, typically requiring up to 2000 repetitions. This number of repetitions can be very difficult, time consuming and uncomfortable for some subjects. In this study, a method combining wavelet analysis and Bayesian networks is introduced to reduce the required number of repetitions, which could offer a great advantage in the clinical situation. 314 ABRs with 64 repetitions and 155 ABRs with 128 repetitions recorded from eight subjects are used here. A wavelet transform is applied to each of the ABRs, and the important features of the ABRs are extracted by thresholding and matching the wavelet coefficients. The significant wavelet coefficients that represent the extracted features of the ABRs are then used as the variables to build the Bayesian network for classification of the ABRs. In order to estimate the performance of this approach, stratified ten-fold cross-validation is used.
Collapse
Affiliation(s)
- Rui Zhang
- School of Computing and Mathematics, Faculty of Engineering, University of Ulster, Jordanstown, UK.
| | | | | | | | | |
Collapse
|
9
|
Nayak A, Roy RJ. Anesthesia control using midlatency auditory evoked potentials. IEEE Trans Biomed Eng 1998; 45:409-21. [PMID: 9556958 DOI: 10.1109/10.664197] [Citation(s) in RCA: 27] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
This paper shows the development of a system to control inhalation anesthetic concentration delivered to a patient based upon that patient's midlatency auditory evoked potentials (MLAEP's). It was developed and tested in dogs by determining response to the supramaximal stimulus of tail clamping. Prior to tail clamp, the MLAEP was recorded along with inhalational anesthetic concentration and classified as responders or nonresponders as determined by tail clamping. This was performed at a number of different anesthetic levels to obtain a data training set. The MLAEP's were compacted by means of discrete time wavelet transform (DTWT), and together with anesthetic concentration value, a stepwise discriminant analysis (SDA) was performed to determine those features which could separate responders from nonresponders. It was determined that only three features were necessary for this recognition. These features were then used to train a four-layer artificial neural network (ANN) to separate the responders from nonresponders. The network was tested using a separate set of data, resulting in a 93% recognition rate in the anesthetic transition zone between responders and nonresponders, and 100% recognition rate outside this zone. The anesthetic controller used this ANN combined with fuzzy logic and rule-based control. A set of ten animal experiments were performed to test the robustness of this controller. Acceptable clinical performance was obtained, showing the feasibility of this approach.
Collapse
Affiliation(s)
- A Nayak
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, USA
| | | |
Collapse
|
10
|
Tian J, Juhola M, Grönfors T. Latency estimation of auditory brainstem response by neural networks. Artif Intell Med 1997; 10:115-28. [PMID: 9201382 DOI: 10.1016/s0933-3657(97)00389-8] [Citation(s) in RCA: 20] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
In the clinical application of auditory brainstem responses (ABRs), the latencies of five to seven main peaks are extremely important parameters for diagnosis. In practice, the latencies have mainly been done by manual measurement so far. In recent years, some new techniques have been developed involving automatic computer recognition. Computer recognition is difficult, however, since some peaks are complicated and vary a lot individually. In this paper, we introduce an artificial neural network method for ABR research. The detection of ABR is performed by using artificial neural networks. A proper bandpass filter is designed for peak extraction. Moreover, a new approach to estimate the latencies of the peaks by artificial neural networks is presented. The neural networks are studied in relation to the selection of model, number of layers and number of neurons in each hidden layer. Experimental results are described showing that artificial neural networks are a promising method in the study of ABR.
Collapse
Affiliation(s)
- J Tian
- Department of Computer Science and Applied Mathematics, University of Kuopio, Finland.
| | | | | |
Collapse
|
11
|
Tian J, Juhola M, Grönfors T. Segmentation of auditory brainstem response signals. INTERNATIONAL JOURNAL OF BIO-MEDICAL COMPUTING 1996; 43:215-26. [PMID: 9032010 DOI: 10.1016/s0020-7101(96)01212-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2023]
Abstract
Auditory brainstem responses are used to detect hearing defects in audiology and otoneurology. The use of computer programs for the analysis of such recordings is increasing. To identify their detailed properties a pattern recognition algorithm implemented in an analysis program must be highly reliable. For the recognition process, some preprocessing phases after recording the necessary, such as filtering and often also segmentation. In the following, we will explore segmentation, which can be used in preprocessing of biomedical signals after filtering. We studied linear segmentation, where slopes of short signal segments are computed and divided into different classes according to their values. A segment length of 8 samples for a sampling frequency of 50 kHz employed was best according to our tests and error criteria. Using clustering, we found that less than 10 segment classes is suitable for pattern recognition.
Collapse
Affiliation(s)
- J Tian
- Department of Computer Science and Applied Mathematics, University of Kuopio, Finland
| | | | | |
Collapse
|
12
|
Gade J, Rosenfalck A, van Gils M, Cluitmans P. Modelling techniques and their application for monitoring in high dependency environments--learning models. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 1996; 51:75-84. [PMID: 8894392 DOI: 10.1016/0169-2607(96)01763-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
This paper reviews the use of learning models including Bayesian classifiers and artificial neural networks in monitoring and interpreting biosignals. Generally learning models applied for analysis of biosignals are "black-box' types trained on the basis of measured signals. It is illustrated that the training and application of learning models more or less follow the same sequences. The main focus is the interpretation of electrical signals from the brain (electroencephalogram (EEG) and evoked potentials (EP)). Current analysis of these signals often reveals sudden changes in the EEG or evoked potentials to be the earliest discernible signs of inadequate perfusion of the brain. They may reflect problems such as systemic arterial oxygen desaturation or hypotension arising from other body system failures during critical illness. It is suggested that these brain signals should be recorded in the critical care unit, and that they should form part of the annotated database of biosignals established during the IMPROVE project. This would allow for the development of new methods for on-line warning of impending damage to the central nervous system, such that corrective actions could be taken before permanent damage occurred.
Collapse
Affiliation(s)
- J Gade
- Department of Medical Informatics and Image Analysis, Aalborg University, Denmark.
| | | | | | | |
Collapse
|
13
|
|