1
|
Lee H, Choi YR, Lee HK, Jeong J, Hong J, Shin HW, Kim HS. Explainable vision transformer for automatic visual sleep staging on multimodal PSG signals. NPJ Digit Med 2025; 8:55. [PMID: 39863774 PMCID: PMC11762271 DOI: 10.1038/s41746-024-01378-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Accepted: 12/10/2024] [Indexed: 01/27/2025] Open
Abstract
Polysomnography (PSG) is crucial for diagnosing sleep disorders, but manual scoring of PSG is time-consuming and subjective, leading to high variability. While machine-learning models have improved PSG scoring, their clinical use is hindered by the 'black-box' nature. In this study, we present SleepXViT, an automatic sleep staging system using Vision Transformer (ViT) that provides intuitive, consistent explanations by mimicking human 'visual scoring'. Tested on KISS-a PSG image dataset from 7745 patients across four hospitals-SleepXViT achieved a Macro F1 score of 81.94%, outperforming baseline models and showing robust performances on public datasets SHHS1 and SHHS2. Furthermore, SleepXViT offers well-calibrated confidence scores, enabling expert review for low-confidence predictions, alongside high-resolution heatmaps highlighting essential features and relevance scores for adjacent epochs' influence on sleep stage predictions. Together, these explanations reinforce the scoring consistency of SleepXViT, making it both reliable and interpretable, thereby facilitating the synergy between the AI model and human scorers in clinical settings.
Collapse
Affiliation(s)
- Hyojin Lee
- Graduate School of Data Science, Seoul National University, Seoul, Republic of Korea
| | - You Rim Choi
- Graduate School of Data Science, Seoul National University, Seoul, Republic of Korea
| | - Hyun Kyung Lee
- Obstructive Upper Airway Research (OUaR) Laboratory, Department of Pharmacology, Seoul National University College of Medicine, Seoul, Republic of Korea
- Department of Biomedical Sciences, Seoul National University Graduate School, Seoul, Republic of Korea
| | - Jaemin Jeong
- Department of Computer Engineering, School of Software, Hallym University, Chuncheon, Republic of Korea
| | - Joopyo Hong
- Graduate School of Data Science, Seoul National University, Seoul, Republic of Korea
| | - Hyun-Woo Shin
- Obstructive Upper Airway Research (OUaR) Laboratory, Department of Pharmacology, Seoul National University College of Medicine, Seoul, Republic of Korea.
- OUaR LaB, Inc, Seoul, Republic of Korea.
- Department of Otorhinolaryngology-Head and Neck Surgery, Seoul National University Hospital, Seoul, Republic of Korea.
- Sensory Organ Research Institute, Seoul National University Medical Research Center, Seoul, Republic of Korea.
- Department of Biomedical Sciences, Seoul National University Graduate School, Seoul, Republic of Korea.
| | - Hyung-Sin Kim
- Graduate School of Data Science, Seoul National University, Seoul, Republic of Korea.
| |
Collapse
|
2
|
Zhou W, Zhu H, Chen W, Chen C, Xu J. Outlier Handling Strategy of Ensembled-Based Sequential Convolutional Neural Networks for Sleep Stage Classification. Bioengineering (Basel) 2024; 11:1226. [PMID: 39768044 PMCID: PMC11673830 DOI: 10.3390/bioengineering11121226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2024] [Revised: 11/22/2024] [Accepted: 11/30/2024] [Indexed: 01/11/2025] Open
Abstract
The pivotal role of sleep has led to extensive research endeavors aimed at automatic sleep stage classification. However, existing methods perform poorly when classifying small groups or individuals, and these results are often considered outliers in terms of overall performance. These outliers may introduce bias during model training, adversely affecting feature selection and diminishing model performance. To address the above issues, this paper proposes an ensemble-based sequential convolutional neural network (E-SCNN) that incorporates a clustering module and neural networks. E-SCNN effectively ensembles machine learning and deep learning techniques to minimize outliers, thereby enhancing model robustness at the individual level. Specifically, the clustering module categorizes individuals based on similarities in feature distribution and assigns personalized weights accordingly. Subsequently, by combining these tailored weights with the robust feature extraction capabilities of convolutional neural networks, the model generates more accurate sleep stage classifications. The proposed model was verified on two public datasets, and experimental results demonstrate that the proposed method obtains overall accuracies of 84.8% on the Sleep-EDF Expanded dataset and 85.5% on the MASS dataset. E-SCNN can alleviate the outlier problem, which is important for improving sleep quality monitoring for individuals.
Collapse
Affiliation(s)
- Wei Zhou
- Jiangsu Key Laboratory of Intelligent Medical Image Computing, Nanjing 210044, China;
- School of Future Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
| | - Hangyu Zhu
- Center for Intelligent Medical Electronics (CIME), School of Information Science and Engineering, Fudan University, Shanghai 200433, China;
| | - Wei Chen
- School of Biomedical Engineering, The University of Sydney, Sydney, NSW 2006, Australia;
| | - Chen Chen
- Center for Medical Research and Innovation, Shanghai Pudong Hosptial, Fudan University Pudong Medical Center, Shanghai 201203, China
- Human Phenome Institute, Fudan University, Shanghai 200438, China
| | - Jun Xu
- Jiangsu Key Laboratory of Intelligent Medical Image Computing, Nanjing 210044, China;
- School of Future Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
| |
Collapse
|
3
|
Chen Y, Lv Y, Sun X, Poluektov M, Zhang Y, Penzel T. ESSN: An Efficient Sleep Sequence Network for Automatic Sleep Staging. IEEE J Biomed Health Inform 2024; 28:7447-7456. [PMID: 39141450 DOI: 10.1109/jbhi.2024.3443340] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/16/2024]
Abstract
By modeling the temporal dependencies of sleep sequence, advanced automatic sleep staging algorithms have achieved satisfactory performance, approaching the level of medical technicians and laying the foundation for clinical assistance. However, existing algorithms cannot adapt well to computing scenarios with limited computing power, such as portable sleep detection and consumer-level sleep disorder screening. In addition, existing algorithms still have the problem of N1 confusion. To address these issues, we propose an efficient sleep sequence network (ESSN) with an ingenious structure to achieve efficient automatic sleep staging at a low computational cost. A novel N1 structure loss is introduced based on the prior knowledge of N1 transition probability to alleviate the N1 stage confusion problem. On the SHHS dataset containing 5,793 subjects, the overall accuracy, macro F1, and Cohen's kappa of ESSN are 88.0%, 81.2%, and 0.831, respectively. When the input length is 200, the parameters and floating-point operations of ESSN are 0.27M and 0.35G, respectively. With a lead in accuracy, ESSN inference is twice as fast as L-SeqSleepNet on the same device. Therefore, our proposed model exhibits solid competitive advantages comparing to other state-of-the-art automatic sleep staging methods.
Collapse
|
4
|
Bao J, Wang G, Wang T, Wu N, Hu S, Lee WH, Lo SL, Yan X, Zheng Y, Wang G. A Feature Fusion Model Based on Temporal Convolutional Network for Automatic Sleep Staging Using Single-Channel EEG. IEEE J Biomed Health Inform 2024; 28:6641-6652. [PMID: 39504300 DOI: 10.1109/jbhi.2024.3457969] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2024]
Abstract
Sleep staging is a crucial task in sleep monitoring and diagnosis, but clinical sleep staging is both time-consuming and subjective. In this study, we proposed a novel deep learning algorithm named feature fusion temporal convolutional network (FFTCN) for automatic sleep staging using single-channel EEG data. This algorithm employed a one-dimensional convolutional neural network (1D-CNN) to extract temporal features from raw EEG, and a two-dimensional CNN (2D-CNN) to extract time-frequency features from spectrograms generated through continuous wavelet transform (CWT) at the epoch level. These features were subsequently fused and further fed into a temporal convolutional network (TCN) to classify sleep stages at the sequence level. Moreover, a two-step training strategy was used to enhance the model's performance on an imbalanced dataset. Our proposed method exhibits superior performance in the 5-class classification task for healthy subjects, as evaluated on the SHHS-1, Sleep-EDF-153, and ISRUC-S1 datasets. This work provided a straightforward and promising method for improving the accuracy of automatic sleep staging using only single-channel EEG, and the proposed method exhibited great potential for future applications in professional sleep monitoring, which could effectively alleviate the workload of sleep technicians.
Collapse
|
5
|
Jirakittayakorn N, Wongsawat Y, Mitrirattanakul S. An enzyme-inspired specificity in deep learning model for sleep stage classification using multi-channel PSG signals input: Separating training approach and its performance on cross-dataset validation for generalizability. Comput Biol Med 2024; 182:109138. [PMID: 39305732 DOI: 10.1016/j.compbiomed.2024.109138] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2024] [Revised: 08/15/2024] [Accepted: 09/08/2024] [Indexed: 11/14/2024]
Abstract
Numerous automatic sleep stage classification systems have been developed, but none have become effective assistive tools for sleep technicians due to issues with generalization. Four key factors hinder the generalization of these models are instruments, montage of recording, subject type, and scoring manual factors. This study aimed to develop a deep learning model that addresses generalization problems by integrating enzyme-inspired specificity and employing separating training approaches. Subject type and scoring manual factors were controlled, while the focus was on instruments and montage of recording factors. The proposed model consists of three sets of signal-specific models including EEG-, EOG-, and EMG-specific model. The EEG-specific models further include three sets of channel-specific models. All signal-specific and channel-specific models were established with data manipulation and weighted loss strategies, resulting in three sets of data manipulation models and class-specific models, respectively. These models were CNNs. Additionally, BiLSTM models were applied to EEG- and EOG-specific models to obtain temporal information. Finally, classification task for sleep stage was handled by 'the-last-dense' layer. The optimal sampling frequency for each physiological signal was identified and used during the training process. The proposed model was trained on MGH dataset and evaluated using both within dataset and cross-dataset. For MGH dataset, overall accuracy of 81.05 %, MF1 of 79.05 %, Kappa of 0.7408, and per-class F1-scores: W (84.98 %), N1 (58.06 %), N2 (84.82 %), N3 (79.20 %), and REM (88.17 %) can be achieved. Performances on cross-datasets are as follows: SHHS1 200 records reached 79.54 %, 70.56 %, and 0.7078; SHHS2 200 records achieved 76.77 %, 66.30 %, and 0.6632; Sleep-EDF 153 records gained 78.52 %, 72.13 %, and 0.7031; and BCI-MU (local dataset) 94 records achieved 83.57 %, 82.17 %, and 0.7769 for overall accuracy, MF1, and Kappa respectively. Additionally, the proposed model has approximately 9.3 M trainable parameters and takes around 26 s to process one PSG record. The results indicate that the proposed model demonstrates generalizability in sleep stage classification and shows potential as a feasibility tool for real-world applications. Additionally, enzyme-inspired specificity effectively addresses the challenges posed by varying montage of recording, while the identified optimal frequencies mitigate instrument-related issues.
Collapse
Affiliation(s)
| | - Yodchanan Wongsawat
- Department of Biomedical Engineering, Faculty of Engineering, Mahidol University, Thailand.
| | | |
Collapse
|
6
|
Tveter M, Tveitstøl T, Hatlestad-Hall C, Pérez T AS, Taubøll E, Yazidi A, Hammer HL, Haraldsen IRJH. Advancing EEG prediction with deep learning and uncertainty estimation. Brain Inform 2024; 11:27. [PMID: 39461914 PMCID: PMC11512943 DOI: 10.1186/s40708-024-00239-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 09/18/2024] [Indexed: 10/29/2024] Open
Abstract
Deep Learning (DL) has the potential to enhance patient outcomes in healthcare by implementing proficient systems for disease detection and diagnosis. However, the complexity and lack of interpretability impede their widespread adoption in critical high-stakes predictions in healthcare. Incorporating uncertainty estimations in DL systems can increase trustworthiness, providing valuable insights into the model's confidence and improving the explanation of predictions. Additionally, introducing explainability measures, recognized and embraced by healthcare experts, can help address this challenge. In this study, we investigate DL models' ability to predict sex directly from electroencephalography (EEG) data. While sex prediction have limited direct clinical application, its binary nature makes it a valuable benchmark for optimizing deep learning techniques in EEG data analysis. Furthermore, we explore the use of DL ensembles to improve performance over single models and as an approach to increase interpretability and performance through uncertainty estimation. Lastly, we use a data-driven approach to evaluate the relationship between frequency bands and sex prediction, offering insights into their relative importance. InceptionNetwork, a single DL model, achieved 90.7% accuracy and an AUC of 0.947, and the best-performing ensemble, combining variations of InceptionNetwork and EEGNet, achieved 91.1% accuracy in predicting sex from EEG data using five-fold cross-validation. Uncertainty estimation through deep ensembles led to increased prediction performance, and the models were able to classify sex in all frequency bands, indicating sex-specific features across all bands.
Collapse
Affiliation(s)
- Mats Tveter
- Department of Neurology, Oslo University Hospital, Oslo, Norway.
- Institute of Clinical Medicine, Faculty of Medicine, University of Oslo, Oslo, Norway.
| | - Thomas Tveitstøl
- Department of Neurology, Oslo University Hospital, Oslo, Norway
- Institute of Clinical Medicine, Faculty of Medicine, University of Oslo, Oslo, Norway
| | | | - Ana S Pérez T
- Department of Neurology, Oslo University Hospital, Oslo, Norway
- Institute of Clinical Medicine, Faculty of Medicine, University of Oslo, Oslo, Norway
| | - Erik Taubøll
- Institute of Clinical Medicine, Faculty of Medicine, University of Oslo, Oslo, Norway
| | - Anis Yazidi
- Department of Computer Science, Oslo Metropolitan University, Oslo, Norway
| | - Hugo L Hammer
- Department of Computer Science, Oslo Metropolitan University, Oslo, Norway
- Department of Holistic Systems, SimulaMet, Oslo, Norway
| | | |
Collapse
|
7
|
Rusanen M, Jouan G, Huttunen R, Nikkonen S, Sigurðardóttir S, Töyräs J, Duce B, Myllymaa S, Arnardottir ES, Leppänen T, Islind AS, Kainulainen S, Korkalainen H. Retrospective validation of automatic sleep analysis with grey areas model for human-in-the-loop scoring approach. J Sleep Res 2024:e14362. [PMID: 39443165 DOI: 10.1111/jsr.14362] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2024] [Revised: 09/02/2024] [Accepted: 09/09/2024] [Indexed: 10/25/2024]
Abstract
State-of-the-art automatic sleep staging methods have demonstrated comparable reliability and superior time efficiency to manual sleep staging. However, fully automatic black-box solutions are difficult to adapt into clinical workflow due to the lack of transparency in decision-making processes. Transparency would be crucial for interaction between automatic methods and the work of sleep experts, i.e., in human-in-the-loop applications. To address these challenges, we propose an automatic sleep staging model (aSAGA) that effectively utilises both electroencephalography and electro-oculography channels while incorporating transparency of uncertainty in the decision-making process. We validated the model through extensive retrospective testing using a range of datasets, including open-access, clinical, and research-driven sources. Our channel-wise ensemble model, trained on both electroencephalography and electro-oculography signals, demonstrated robustness and the ability to generalise across various types of sleep recordings, including novel self-applied home polysomnography. Additionally, we compared model uncertainty with human uncertainty in sleep staging and studied various uncertainty mapping metrics to identify ambiguous regions, or "grey areas", that may require manual re-evaluation. The validation of this grey area concept revealed its potential to enhance sleep staging accuracy and to highlight regions in the recordings where sleep experts may struggle to reach a consensus. In conclusion, this study provides a technical basis and understanding of automatic sleep staging uncertainty. Our approach has the potential to improve the integration of automatic sleep staging into clinical practice; however, further studies are needed to test the model prospectively in real-world clinical settings and human-in-the-loop scoring applications.
Collapse
Grants
- 965417 European Union's Horizon 2020 research and innovation programme
- 5041789 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- 5041794 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- 5041797 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- 504180 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- 5041803 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- 5041807 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- 5041809 the State Research Funding for University-Level Health Research, Kuopio University Hospital, Wellbeing Service County of North Savo
- Finnish Cultural Foundation
- Eemil Aaltonen Foundation
- The Research Foundation of the Pulmonary Diseases
- Foundation of the Finnish Anti-Tuberculosis Association
- Tampereen Tuberkuloosisäätiö
- 90458 NordForsk
- 5133/31/2018 Business Finland
- ANR-15-IDEX-02 Agence Nationale de la Recherche
- ANR-19-P3IA-0003 Agence Nationale de la Recherche
- Sigrid Jusélius Foundation
Collapse
Affiliation(s)
- Matias Rusanen
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
- HP2 Laboratory, INSERM U1300, Grenoble Alpes University, Grenoble Alpes University Hospital, Grenoble, France
| | - Gabriel Jouan
- Reykjavik University Sleep Institute, School of Technology, Reykjavik University, Reykjavik, Iceland
| | - Riku Huttunen
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
| | - Sami Nikkonen
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
| | - Sigríður Sigurðardóttir
- Reykjavik University Sleep Institute, School of Technology, Reykjavik University, Reykjavik, Iceland
| | - Juha Töyräs
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Science Service Centre, Kuopio University Hospital, Kuopio, Finland
- School of Electrical Engineering and Computer Science, The University of Queensland, Brisbane, Queensland, Australia
| | - Brett Duce
- Princess Alexandra Hospital, Sleep Disorders Centre, Princess Alexandra Hospital, Brisbane, Queensland, Australia
| | - Sami Myllymaa
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
| | - Erna Sif Arnardottir
- Reykjavik University Sleep Institute, School of Technology, Reykjavik University, Reykjavik, Iceland
- Landspitali-The National University Hospital of Iceland, Reykjavik, Iceland
| | - Timo Leppänen
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
- School of Electrical Engineering and Computer Science, The University of Queensland, Brisbane, Queensland, Australia
| | | | - Samu Kainulainen
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
| | - Henri Korkalainen
- Department of Technical Physics, University of Eastern Finland, Kuopio, Finland
- Diagnostic Imaging Centre, Kuopio University Hospital, Kuopio, Finland
| |
Collapse
|
8
|
Zhou W, Shen N, Zhou L, Liu M, Zhang Y, Fu C, Yu H, Shu F, Chen W, Chen C. PSEENet: A Pseudo-Siamese Neural Network Incorporating Electroencephalography and Electrooculography Characteristics for Heterogeneous Sleep Staging. IEEE J Biomed Health Inform 2024; 28:5189-5200. [PMID: 38771683 DOI: 10.1109/jbhi.2024.3403878] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/23/2024]
Abstract
Sleep staging plays a critical role in evaluating the quality of sleep. Currently, most studies are either suffering from dramatic performance drops when coping with varying input modalities or unable to handle heterogeneous signals. To handle heterogeneous signals and guarantee favorable sleep staging performance when a single modality is available, a pseudo-siamese neural network (PSN) to incorporate electroencephalography (EEG), electrooculography (EOG) characteristics is proposed (PSEENet). PSEENet consists of two parts, spatial mapping modules (SMMs) and a weight-shared classifier. SMMs are used to extract high-dimensional features. Meanwhile, joint linkages among multi-modalities are provided by quantifying the similarity of features. Finally, with the cooperation of heterogeneous characteristics, associations within various sleep stages can be established by the classifier. The evaluation of the model is validated on two public datasets, namely, Montreal Archive of Sleep Studies (MASS) and SleepEDFX, and one clinical dataset from Huashan Hospital of Fudan University (HSFU). Experimental results show that the model can handle heterogeneous signals, provide superior results under multimodal signals and show good performance with single modality. PSEENet obtains accuracy of 79.1%, 82.1% with EEG, EEG and EOG on Sleep-EDFX, and significantly improves the accuracy with EOG from 73.7% to 76% by introducing similarity information.
Collapse
|
9
|
Pradeepkumar J, Anandakumar M, Kugathasan V, Suntharalingham D, Kappel SL, De Silva AC, Edussooriya CUS. Toward Interpretable Sleep Stage Classification Using Cross-Modal Transformers. IEEE Trans Neural Syst Rehabil Eng 2024; 32:2893-2904. [PMID: 39102323 DOI: 10.1109/tnsre.2024.3438610] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/07/2024]
Abstract
Accurate sleep stage classification is significant for sleep health assessment. In recent years, several machine-learning based sleep staging algorithms have been developed, and in particular, deep-learning based algorithms have achieved performance on par with human annotation. Despite improved performance, a limitation of most deep-learning based algorithms is their black-box behavior, which have limited their use in clinical settings. Here, we propose a cross-modal transformer, which is a transformer-based method for sleep stage classification. The proposed cross-modal transformer consists of a cross-modal transformer encoder architecture along with a multi-scale one-dimensional convolutional neural network for automatic representation learning. The performance of our method is on-par with the state-of-the-art methods and eliminates the black-box behavior of deep-learning models by utilizing the interpretability aspect of the attention modules. Furthermore, our method provides considerable reductions in the number of parameters and training time compared to the state-of-the-art methods. Our code is available at https://github.com/Jathurshan0330/Cross-Modal-Transformer. A demo of our work can be found at https://bit.ly/Cross_modal_transformer_demo.
Collapse
|
10
|
Moctezuma LA, Suzuki Y, Furuki J, Molinas M, Abe T. GRU-powered sleep stage classification with permutation-based EEG channel selection. Sci Rep 2024; 14:17952. [PMID: 39095608 PMCID: PMC11297028 DOI: 10.1038/s41598-024-68978-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Accepted: 07/30/2024] [Indexed: 08/04/2024] Open
Abstract
We present a new approach to classifying the sleep stage that incorporates a computationally inexpensive method based on permutations for channel selection and takes advantage of deep learning power, specifically the gated recurrent unit (GRU) model, along with other deep learning methods. By systematically permuting the electroencephalographic (EEG) channels, different combinations of EEG channels are evaluated to identify the most informative subset for the classification of the 5-class sleep stage. For analysis, we used an EEG dataset that was collected at the International Institute for Integrative Sleep Medicine (WPI-IIIS) at the University of Tsukuba in Japan. The results of these explorations provide many new insights such as the (1) drastic decrease in performance when channels are fewer than 3, (2) 3-random channels selected by permutation provide the same or better prediction than the 3 channels recommended by the American Academy of Sleep Medicine (AASM), (3) N1 class suffers the most in prediction accuracy as the channels drop from 128 to 3 random or 3 AASM, and (4) no single channel provides acceptable levels of accuracy in the prediction of 5 classes. The results obtained show the GRU's ability to retain essential temporal information from EEG data, which allows capturing the underlying patterns associated with each sleep stage effectively. Using permutation-based channel selection, we enhance or at least maintain as high model efficiency as when using high-density EEG, incorporating only the most informative EEG channels.
Collapse
Affiliation(s)
- Luis Alfredo Moctezuma
- International Institute for Integrative Sleep Medicine (WPI-IIIS), University of Tsukuba, Tsukuba, Ibaraki, Japan.
| | - Yoko Suzuki
- International Institute for Integrative Sleep Medicine (WPI-IIIS), University of Tsukuba, Tsukuba, Ibaraki, Japan
| | - Junya Furuki
- International Institute for Integrative Sleep Medicine (WPI-IIIS), University of Tsukuba, Tsukuba, Ibaraki, Japan
| | - Marta Molinas
- Department of Engineering Cybernetics, Norwegian University of Science and Technology, Trondheim, Norway
| | - Takashi Abe
- International Institute for Integrative Sleep Medicine (WPI-IIIS), University of Tsukuba, Tsukuba, Ibaraki, Japan
| |
Collapse
|
11
|
Lee M, Kang H, Yu SH, Cho H, Oh J, van der Lande G, Gosseries O, Jeong JH. Automatic Sleep Stage Classification Using Nasal Pressure Decoding Based on a Multi-Kernel Convolutional BiLSTM Network. IEEE Trans Neural Syst Rehabil Eng 2024; 32:2533-2544. [PMID: 38941194 DOI: 10.1109/tnsre.2024.3420715] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/30/2024]
Abstract
Sleep quality is an essential parameter of a healthy human life, while sleep disorders such as sleep apnea are abundant. In the investigation of sleep and its malfunction, the gold-standard is polysomnography, which utilizes an extensive range of variables for sleep stage classification. However, undergoing full polysomnography, which requires many sensors that are directly connected to the heaviness of the setup and the discomfort of sleep, brings a significant burden. In this study, sleep stage classification was performed using the single dimension of nasal pressure, dramatically decreasing the complexity of the process. In turn, such improvements could increase the much needed clinical applicability. Specifically, we propose a deep learning structure consisting of multi-kernel convolutional neural networks and bidirectional long short-term memory for sleep stage classification. Sleep stages of 25 healthy subjects were classified into 3-class (wake, rapid eye movement (REM), and non-REM) and 4-class (wake, REM, light, and deep sleep) based on nasal pressure. Following a leave-one-subject-out cross-validation, in the 3-class the accuracy was 0.704, the F1-score was 0.490, and the kappa value was 0.283 for the overall metrics. In the 4-class, the accuracy was 0.604, the F1-score was 0.349, and the kappa value was 0.217 for the overall metrics. This was higher than the four comparative models, including the class-wise F1-score. This result demonstrates the possibility of a sleep stage classification model only using easily applicable and highly practical nasal pressure recordings. This is also likely to be used with interventions that could help treat sleep-related diseases.
Collapse
|
12
|
Moctezuma LA, Suzuki Y, Furuki J, Molinas M, Abe T. Enhancing sleep stage classification with 2-class stratification and permutation-based channel selection. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-4. [PMID: 40039811 DOI: 10.1109/embc53108.2024.10781882] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
We present a method that uses a convolutional neural network (CNN) called EEGNeX to extract and classify the characteristics of sleep-related waveforms from electroencephalographic (EEG) signals in different stages of sleep. Our results showed that the CNN model with 128 channels achieved high performance, distributing the sleep stages into 2-class models. We used a permutation-based channel selection process and using the top 3 channels, we achieved a performance greater than 80% in accuracy, Fscore, precision, recall, area under the receiver operating characteristic (AUROC) and kappa value, except when classifying N1 versus N2, where the average kappa value was 0.52. Performance is shown to decrease when using the 3 channels recommended by the American Academy of Sleep Medicine (AASM) or 3 random channels. Overall, the results showed that 2-class CNN models with 3 channels selected with a permutation-based approach achieve good performance in the classification of sleep stages from EEG signals, with a computational cost much lower than using 128 EEG channels.
Collapse
|
13
|
Bechny M, Monachino G, Fiorillo L, van der Meer J, Schmidt MH, Bassetti CLA, Tzovara A, Faraci FD. Bridging AI and Clinical Practice: Integrating Automated Sleep Scoring Algorithm with Uncertainty-Guided Physician Review. Nat Sci Sleep 2024; 16:555-572. [PMID: 38827394 PMCID: PMC11143488 DOI: 10.2147/nss.s455649] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Accepted: 04/18/2024] [Indexed: 06/04/2024] Open
Abstract
Purpose This study aims to enhance the clinical use of automated sleep-scoring algorithms by incorporating an uncertainty estimation approach to efficiently assist clinicians in the manual review of predicted hypnograms, a necessity due to the notable inter-scorer variability inherent in polysomnography (PSG) databases. Our efforts target the extent of review required to achieve predefined agreement levels, examining both in-domain (ID) and out-of-domain (OOD) data, and considering subjects' diagnoses. Patients and Methods A total of 19,578 PSGs from 13 open-access databases were used to train U-Sleep, a state-of-the-art sleep-scoring algorithm. We leveraged a comprehensive clinical database of an additional 8832 PSGs, covering a full spectrum of ages (0-91 years) and sleep-disorders, to refine the U-Sleep, and to evaluate different uncertainty-quantification approaches, including our novel confidence network. The ID data consisted of PSGs scored by over 50 physicians, and the two OOD sets comprised recordings each scored by a unique senior physician. Results U-Sleep demonstrated robust performance, with Cohen's kappa (K) at 76.2% on ID and 73.8-78.8% on OOD data. The confidence network excelled at identifying uncertain predictions, achieving AUROC scores of 85.7% on ID and 82.5-85.6% on OOD data. Independently of sleep-disorder status, statistical evaluations revealed significant differences in confidence scores between aligning vs discording predictions, and significant correlations of confidence scores with classification performance metrics. To achieve κ ≥ 90% with physician intervention, examining less than 29.0% of uncertain epochs was required, substantially reducing physicians' workload, and facilitating near-perfect agreement. Conclusion Inter-scorer variability limits the accuracy of the scoring algorithms to ~80%. By integrating an uncertainty estimation with U-Sleep, we enhance the review of predicted hypnograms, to align with the scoring taste of a responsible physician. Validated across ID and OOD data and various sleep-disorders, our approach offers a strategy to boost automated scoring tools' usability in clinical settings.
Collapse
Affiliation(s)
- Michal Bechny
- Institute of Computer Science, University of Bern, Bern, Switzerland
- Institute of Digital Technologies for Personalized Healthcare (Meditech), University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland
| | - Giuliana Monachino
- Institute of Computer Science, University of Bern, Bern, Switzerland
- Institute of Digital Technologies for Personalized Healthcare (Meditech), University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland
| | - Luigi Fiorillo
- Institute of Digital Technologies for Personalized Healthcare (Meditech), University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland
| | | | - Markus H Schmidt
- Department of Neurology, University Hospital of Bern, Bern, Switzerland
- Ohio Sleep Medicine Institute, Dublin, OH, USA
| | | | - Athina Tzovara
- Institute of Computer Science, University of Bern, Bern, Switzerland
- Department of Neurology, University Hospital of Bern, Bern, Switzerland
| | - Francesca D Faraci
- Institute of Digital Technologies for Personalized Healthcare (Meditech), University of Applied Sciences and Arts of Southern Switzerland, Lugano, Switzerland
| |
Collapse
|
14
|
Shao Y, Huang B, Du L, Wang P, Li Z, Liu Z, Zhou L, Song Y, Chen X, Fang Z. Reliable automatic sleep stage classification based on hybrid intelligence. Comput Biol Med 2024; 173:108314. [PMID: 38513392 DOI: 10.1016/j.compbiomed.2024.108314] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 02/10/2024] [Accepted: 03/12/2024] [Indexed: 03/23/2024]
Abstract
Sleep staging is a vital aspect of sleep assessment, serving as a critical tool for evaluating the quality of sleep and identifying sleep disorders. Manual sleep staging is a laborious process, while automatic sleep staging is seldom utilized in clinical practice due to issues related to the inadequate accuracy and interpretability of classification results in automatic sleep staging models. In this work, a hybrid intelligent model is presented for automatic sleep staging, which integrates data intelligence and knowledge intelligence, to attain a balance between accuracy, interpretability, and generalizability in the sleep stage classification. Specifically, it is built on any combination of typical electroencephalography (EEG) and electrooculography (EOG) channels, including a temporal fully convolutional network based on the U-Net architecture and a multi-task feature mapping structure. The experimental results show that, compared to current interpretable automatic sleep staging models, our model achieves a Macro-F1 score of 0.804 on the ISRUC dataset and 0.780 on the Sleep-EDFx dataset. Moreover, we use knowledge intelligence to address issues of excessive jumps and unreasonable sleep stage transitions in the coarse sleep graphs obtained by the model. We also explore the different ways knowledge intelligence affects coarse sleep graphs by combining different sleep graph correction methods. Our research can offer convenient support for sleep physicians, indicating its significant potential in improving the efficiency of clinical sleep staging.
Collapse
Affiliation(s)
- Yizi Shao
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China.
| | - Bokai Huang
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China.
| | - Lidong Du
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China; Personalized Management of Chronic Respiratory Disease, Chinese Academy of Medical Sciences, Beijing, China.
| | - Peng Wang
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China; Personalized Management of Chronic Respiratory Disease, Chinese Academy of Medical Sciences, Beijing, China.
| | - Zhenfeng Li
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China; Personalized Management of Chronic Respiratory Disease, Chinese Academy of Medical Sciences, Beijing, China.
| | - Zhe Liu
- Hunan VentMed Medical Technology Co., Ltd, Shaoyang, China.
| | - Lei Zhou
- Qingpu Branch of Zhongshan Hospital, Fudan University, Shanghai, China.
| | - Yuanlin Song
- Zhongshan Hospital Fudan University, Shanghai, China.
| | - Xianxiang Chen
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China; Personalized Management of Chronic Respiratory Disease, Chinese Academy of Medical Sciences, Beijing, China.
| | - Zhen Fang
- Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China; University of Chinese Academy of Sciences, Beijing, China; Personalized Management of Chronic Respiratory Disease, Chinese Academy of Medical Sciences, Beijing, China.
| |
Collapse
|
15
|
Jirakittayakorn N, Wongsawat Y, Mitrirattanakul S. ZleepAnlystNet: a novel deep learning model for automatic sleep stage scoring based on single-channel raw EEG data using separating training. Sci Rep 2024; 14:9859. [PMID: 38684765 PMCID: PMC11058251 DOI: 10.1038/s41598-024-60796-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 04/26/2024] [Indexed: 05/02/2024] Open
Abstract
Numerous models for sleep stage scoring utilizing single-channel raw EEG signal have typically employed CNN and BiLSTM architectures. While these models, incorporating temporal information for sequence classification, demonstrate superior overall performance, they often exhibit low per-class performance for N1-stage, necessitating an adjustment of loss function. However, the efficacy of such adjustment is constrained by the training process. In this study, a pioneering training approach called separating training is introduced, alongside a novel model, to enhance performance. The developed model comprises 15 CNN models with varying loss function weights for feature extraction and 1 BiLSTM for sequence classification. Due to its architecture, this model cannot be trained using an end-to-end approach, necessitating separate training for each component using the Sleep-EDF dataset. Achieving an overall accuracy of 87.02%, MF1 of 82.09%, Kappa of 0.8221, and per-class F1-socres (W 90.34%, N1 54.23%, N2 89.53%, N3 88.96%, and REM 87.40%), our model demonstrates promising performance. Comparison with sleep technicians reveals a Kappa of 0.7015, indicating alignment with reference sleep stags. Additionally, cross-dataset validation and adaptation through training with the SHHS dataset yield an overall accuracy of 84.40%, MF1 of 74.96% and Kappa of 0.7785 when tested with the Sleep-EDF-13 dataset. These findings underscore the generalization potential in model architecture design facilitated by our novel training approach.
Collapse
Affiliation(s)
- Nantawachara Jirakittayakorn
- Institute for Innovative Learning, Mahidol University, Nakhon Pathom, Thailand
- Faculty of Dentistry, Mahidol University, Bangkok, Thailand
| | - Yodchanan Wongsawat
- Department of Biomedical Engineering, Faculty of Engineering, Mahidol University, Nakhon Pathom, Thailand
| | - Somsak Mitrirattanakul
- Department of Masticatory Science, Faculty of Dentistry, Mahidol University, Bangkok, Thailand.
| |
Collapse
|
16
|
Rashidi S, Asl BM. Strength of ensemble learning in automatic sleep stages classification using single-channel EEG and ECG signals. Med Biol Eng Comput 2024; 62:997-1015. [PMID: 38114690 DOI: 10.1007/s11517-023-02980-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2023] [Accepted: 11/26/2023] [Indexed: 12/21/2023]
Abstract
Healthy sleep plays an essential role in human daily life. Classification of sleep stages is a crucial tool for assisting physicians in diagnosing and treating sleep disorders. In this study, a strong ensemble learning model is proposed to enhance the ability of classification models in accurate sleep staging, particularly in multi-class classification. We asserted that high-accuracy sleep classification is achievable using only single-channel electroencephalogram (EEG) and electrocardiogram (ECG) by combining their best-extractable features in the time and frequency domains we recommended. More importantly, the superiority of the recommended method, which is the simultaneous use of stacking and bagging, over conventional machine learning classifiers in sleep staging was demonstrated, using the MIT-BIH Polysomnographic and Sleep-EDF expanded databases. Finally, K-fold cross-validation was used to fairly estimate these models. The best mean test accuracy rates for distinguishing between two classes of "sleep vs. wake," "rapid vs. non-rapid eye movement," and "deep vs. light sleep," were obtained 99.93%, 99.64%, and 99.69%, respectively. Furthermore, our proposed method achieved accuracies of 97.14%, 95.18%, 92.7%, and 85.64% for separating three, four, five, and six sleep classes, respectively. Compared to recent studies, our method outperforms other sleep stage classification schemes, especially in multi-class staging.
Collapse
Affiliation(s)
- Samandokht Rashidi
- Department of Electrical and Computer Engineering, Tarbiat Modares University, Tehran, Iran
| | - Babak Mohammadzadeh Asl
- Department of Electrical and Computer Engineering, Tarbiat Modares University, Tehran, Iran.
| |
Collapse
|
17
|
Heremans ERM, Seedat N, Buyse B, Testelmans D, van der Schaar M, De Vos M. U-PASS: An uncertainty-guided deep learning pipeline for automated sleep staging. Comput Biol Med 2024; 171:108205. [PMID: 38401452 DOI: 10.1016/j.compbiomed.2024.108205] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Revised: 02/16/2024] [Accepted: 02/20/2024] [Indexed: 02/26/2024]
Abstract
With the increasing prevalence of machine learning in critical fields like healthcare, ensuring the safety and reliability of these systems is crucial. Estimating uncertainty plays a vital role in enhancing reliability by identifying areas of high and low confidence and reducing the risk of errors. This study introduces U-PASS, a specialized human-centered machine learning pipeline tailored for clinical applications, which effectively communicates uncertainty to clinical experts and collaborates with them to improve predictions. U-PASS incorporates uncertainty estimation at every stage of the process, including data acquisition, training, and model deployment. Training is divided into a supervised pre-training step and a semi-supervised recording-wise finetuning step. We apply U-PASS to the challenging task of sleep staging and demonstrate that it systematically improves performance at every stage. By optimizing the training dataset, actively seeking feedback from domain experts for informative samples, and deferring the most uncertain samples to experts, U-PASS achieves an impressive expert-level accuracy of 85% on a challenging clinical dataset of elderly sleep apnea patients. This represents a significant improvement over the starting point at 75% accuracy. The largest improvement gain is due to the deferral of uncertain epochs to a sleep expert. U-PASS presents a promising AI approach to incorporating uncertainty estimation in machine learning pipelines, improving their reliability and unlocking their potential in clinical settings.
Collapse
Affiliation(s)
- Elisabeth R M Heremans
- KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Kasteelpark Arenberg 10, B-3001 Leuven, Belgium.
| | | | - Bertien Buyse
- UZ Leuven, Department of Pneumology, Herestraat 49, B-3000 Leuven, Belgium
| | - Dries Testelmans
- UZ Leuven, Department of Pneumology, Herestraat 49, B-3000 Leuven, Belgium
| | | | - Maarten De Vos
- KU Leuven, Department of Electrical Engineering (ESAT), STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Kasteelpark Arenberg 10, B-3001 Leuven, Belgium.
| |
Collapse
|
18
|
Nam B, Bark B, Lee J, Kim IY. InsightSleepNet: the interpretable and uncertainty-aware deep learning network for sleep staging using continuous Photoplethysmography. BMC Med Inform Decis Mak 2024; 24:50. [PMID: 38355559 PMCID: PMC10865603 DOI: 10.1186/s12911-024-02437-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Accepted: 01/23/2024] [Indexed: 02/16/2024] Open
Abstract
BACKGROUND This study was conducted to address the existing drawbacks of inconvenience and high costs associated with sleep monitoring. In this research, we performed sleep staging using continuous photoplethysmography (PPG) signals for sleep monitoring with wearable devices. Furthermore, our aim was to develop a more efficient sleep monitoring method by considering both the interpretability and uncertainty of the model's prediction results, with the goal of providing support to medical professionals in their decision-making process. METHOD The developed 4-class sleep staging model based on continuous PPG data incorporates several key components: a local attention module, an InceptionTime module, a time-distributed dense layer, a temporal convolutional network (TCN), and a 1D convolutional network (CNN). This model prioritizes both interpretability and uncertainty estimation in its prediction results. The local attention module is introduced to provide insights into the impact of each epoch within the continuous PPG data. It achieves this by leveraging the TCN structure. To quantify the uncertainty of prediction results and facilitate selective predictions, an energy score estimation is employed. By enhancing both the performance and interpretability of the model and taking into consideration the reliability of its predictions, we developed the InsightSleepNet for accurate sleep staging. RESULT InsightSleepNet was evaluated using three distinct datasets: MESA, CFS, and CAP. Initially, we assessed the model's classification performance both before and after applying an energy score threshold. We observed a significant improvement in the model's performance with the implementation of the energy score threshold. On the MESA dataset, prior to applying the energy score threshold, the accuracy was 84.2% with a Cohen's kappa of 0.742 and weighted F1 score of 0.842. After implementing the energy score threshold, the accuracy increased to a range of 84.8-86.1%, Cohen's kappa values ranged from 0.75 to 0.78 and weighted F1 scores ranged from 0.848 to 0.861. In the case of the CFS dataset, we also noted enhanced performance. Before the application of the energy score threshold, the accuracy stood at 80.6% with a Cohen's kappa of 0.72 and weighted F1 score of 0.808. After thresholding, the accuracy improved to a range of 81.9-85.6%, Cohen's kappa values ranged from 0.74 to 0.79 and weighted F1 scores ranged from 0.821 to 0.857. Similarly, on the CAP dataset, the initial accuracy was 80.6%, accompanied by a Cohen's kappa of 0.73 and weighted F1 score was 0.805. Following the application of the threshold, the accuracy increased to a range of 81.4-84.3%, Cohen's kappa values ranged from 0.74 to 0.79 and weighted F1 scores ranged from 0.813 to 0.842. Additionally, by interpreting the model's predictions, we obtained results indicating a correlation between the peak of the PPG signal and sleep stage classification. CONCLUSION InsightSleepNet is a 4-class sleep staging model that utilizes continuous PPG data, serves the purpose of continuous sleep monitoring with wearable devices. Beyond its primary function, it might facilitate in-depth sleep analysis by medical professionals and empower them with interpretability for intervention-based predictions. This capability can also support well-informed clinical decision-making, providing valuable insights and serving as a reliable second opinion in medical settings.
Collapse
Affiliation(s)
- Borum Nam
- Department of Electronic Engineering, Hanyang University, Seoul, Republic of Korea
| | - Beomjun Bark
- Department of Biomedical Engineering, Hanyang University, 222, Wangsimni-ro, Seoul, 04763, Republic of Korea
| | - Jeyeon Lee
- Department of Biomedical Engineering, Hanyang University, 222, Wangsimni-ro, Seoul, 04763, Republic of Korea
| | - In Young Kim
- Department of Biomedical Engineering, Hanyang University, 222, Wangsimni-ro, Seoul, 04763, Republic of Korea.
| |
Collapse
|
19
|
Jain R, G RA. Modality-Specific Feature Selection, Data Augmentation and Temporal Context for Improved Performance in Sleep Staging. IEEE J Biomed Health Inform 2024; 28:1031-1042. [PMID: 38051608 DOI: 10.1109/jbhi.2023.3339713] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
This work attempts to design an effective sleep staging system, making the best use of the available signals, strategies, and features in the literature. It must not only perform well on different datasets comprising healthy and clinical populations but also achieve good accuracy in cross-dataset experiments. Toward this end, we propose a model comprising multiple binary classifiers in a hierarchical fashion, where, at each level, one or more of EEG, EOG, and EMG are selected to best differentiate between two sleep stages. The best set of 100 features is chosen out of all the features derived from selected signals. The class imbalance in data is addressed by random undersampling and boosting techniques with decision trees as weak learners. Temporal context and data augmentation are used to improve the performance. We also evaluate the performance of our model by training and testing on different datasets. We compare the results of five approaches: using only EEG, EEG+EOG, EEG+EMG+EOG, EEG+EMG, and selective modality with a specific combination of EEG, EMG, and/or EOG at each level. The best results are obtained by considering features from EEG+EMG+EOG at each hierarchical level. The proposed model achieves average accuracies of 83.1%, 90.0%, 84.4%, 82.1%, 81.5%, 79.9%, and 73.7% on Sleep-EDF, Exp Sleep-EDF, ISRUC-S1, S2 and S3, DRMS-SUB, and DRMS-PAT datasets, respectively. For all the datasets except DRMS-SUB, the proposed method outperforms all the state-of-the-art approaches. Cross-dataset performance exceeds 80% for all datasets except DRMS-PAT; independent of whether the test data is from normal subjects or patients.
Collapse
|
20
|
Brodersen PJN, Alfonsa H, Krone LB, Blanco-Duque C, Fisk AS, Flaherty SJ, Guillaumin MCC, Huang YG, Kahn MC, McKillop LE, Milinski L, Taylor L, Thomas CW, Yamagata T, Foster RG, Vyazovskiy VV, Akerman CJ. Somnotate: A probabilistic sleep stage classifier for studying vigilance state transitions. PLoS Comput Biol 2024; 20:e1011793. [PMID: 38232122 PMCID: PMC10824458 DOI: 10.1371/journal.pcbi.1011793] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2023] [Revised: 01/29/2024] [Accepted: 01/02/2024] [Indexed: 01/19/2024] Open
Abstract
Electrophysiological recordings from freely behaving animals are a widespread and powerful mode of investigation in sleep research. These recordings generate large amounts of data that require sleep stage annotation (polysomnography), in which the data is parcellated according to three vigilance states: awake, rapid eye movement (REM) sleep, and non-REM (NREM) sleep. Manual and current computational annotation methods ignore intermediate states because the classification features become ambiguous, even though intermediate states contain important information regarding vigilance state dynamics. To address this problem, we have developed "Somnotate"-a probabilistic classifier based on a combination of linear discriminant analysis (LDA) with a hidden Markov model (HMM). First we demonstrate that Somnotate sets new standards in polysomnography, exhibiting annotation accuracies that exceed human experts on mouse electrophysiological data, remarkable robustness to errors in the training data, compatibility with different recording configurations, and an ability to maintain high accuracy during experimental interventions. However, the key feature of Somnotate is that it quantifies and reports the certainty of its annotations. We leverage this feature to reveal that many intermediate vigilance states cluster around state transitions, whereas others correspond to failed attempts to transition. This enables us to show for the first time that the success rates of different types of transition are differentially affected by experimental manipulations and can explain previously observed sleep patterns. Somnotate is open-source and has the potential to both facilitate the study of sleep stage transitions and offer new insights into the mechanisms underlying sleep-wake dynamics.
Collapse
Affiliation(s)
- Paul J. N. Brodersen
- Department of Pharmacology, University of Oxford; Mansfield Road, Oxford, United Kingdom
| | - Hannah Alfonsa
- Department of Pharmacology, University of Oxford; Mansfield Road, Oxford, United Kingdom
| | - Lukas B. Krone
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Cristina Blanco-Duque
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Angus S. Fisk
- Nuffield Department of Clinical Neurosciences, University of Oxford; John Radcliffe Hospital, Oxford, United Kingdom
| | - Sarah J. Flaherty
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Mathilde C. C. Guillaumin
- Nuffield Department of Clinical Neurosciences, University of Oxford; John Radcliffe Hospital, Oxford, United Kingdom
- Sleep and Circadian Neuroscience Institute, University of Oxford; Oxford, United Kingdom
- Institute for Neuroscience, Department of Health Sciences and Technology, ETH Zurich; Schwerzenbach, Switzerland
| | - Yi-Ge Huang
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Martin C. Kahn
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Laura E. McKillop
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Linus Milinski
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Lewis Taylor
- Nuffield Department of Clinical Neurosciences, University of Oxford; John Radcliffe Hospital, Oxford, United Kingdom
| | - Christopher W. Thomas
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Tomoko Yamagata
- Nuffield Department of Clinical Neurosciences, University of Oxford; John Radcliffe Hospital, Oxford, United Kingdom
| | - Russell G. Foster
- Sleep and Circadian Neuroscience Institute, University of Oxford; Oxford, United Kingdom
| | - Vladyslav V. Vyazovskiy
- Department of Physiology, Anatomy and Genetics, University of Oxford; Parks Road, United Kingdom
| | - Colin J. Akerman
- Department of Pharmacology, University of Oxford; Mansfield Road, Oxford, United Kingdom
| |
Collapse
|
21
|
Li J, Wu C, Pan J, Wang F. Few-shot EEG sleep staging based on transductive prototype optimization network. Front Neuroinform 2023; 17:1297874. [PMID: 38125309 PMCID: PMC10730933 DOI: 10.3389/fninf.2023.1297874] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2023] [Accepted: 11/13/2023] [Indexed: 12/23/2023] Open
Abstract
Electroencephalography (EEG) is a commonly used technology for monitoring brain activities and diagnosing sleep disorders. Clinically, doctors need to manually stage sleep based on EEG signals, which is a time-consuming and laborious task. In this study, we propose a few-shot EEG sleep staging termed transductive prototype optimization network (TPON) method, which aims to improve the performance of EEG sleep staging. Compared with traditional deep learning methods, TPON uses a meta-learning algorithm, which generalizes the classifier to new classes that are not visible in the training set, and only have a few examples for each new class. We learn the prototypes of existing objects through meta-training, and capture the sleep features of new objects through the "learn to learn" method of meta-learning. The prototype distribution of the class is optimized and captured by using support set and unlabeled high confidence samples to increase the authenticity of the prototype. Compared with traditional prototype networks, TPON can effectively solve too few samples in few-shot learning and improve the matching degree of prototypes in prototype network. The experimental results on the public SleepEDF-2013 dataset show that the proposed algorithm outperform than most advanced algorithms in the overall performance. In addition, we experimentally demonstrate the feasibility of cross-channel recognition, which indicates that there are many similar sleep EEG features between different channels. In future research, we can further explore the common features among different channels and investigate the combination of universal features in sleep EEG. Overall, our method achieves high accuracy in sleep stage classification, demonstrating the effectiveness of this approach and its potential applications in other medical fields.
Collapse
Affiliation(s)
| | | | | | - Fei Wang
- School of Software, South China Normal University, Guangzhou, China
| |
Collapse
|
22
|
Yeckle J, Manian V. Automated Sleep Stage Classification in Home Environments: An Evaluation of Seven Deep Neural Network Architectures. SENSORS (BASEL, SWITZERLAND) 2023; 23:8942. [PMID: 37960641 PMCID: PMC10649735 DOI: 10.3390/s23218942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Revised: 10/25/2023] [Accepted: 10/31/2023] [Indexed: 11/15/2023]
Abstract
Sleep is an essential human physiological need that has garnered increasing scientific attention due to the burgeoning prevalence of sleep-related disorders and their impact on public health. Among contemporary challenges, the demand for authentic sleep monitoring outside the confines of specialized laboratories, ideally within the home environment, has arisen. Addressing this, we explore the development of pragmatic approaches that facilitate implementation within domestic settings. Such approaches necessitate the deployment of streamlined, computationally efficient automated classifiers. In pursuit of a sleep stage classifier tailored for home use, this study rigorously assessed seven conventional neural network architectures prominent in deep learning (LeNet, ResNet, VGG, MLP, LSTM-CNN, LSTM, BLSTM). Leveraging sleep recordings from a cohort of 20 subjects, we elucidate that LeNet, VGG, and ResNet exhibit superior performance compared to recent advancements reported in the literature. Furthermore, a comprehensive architectural analysis was conducted, illuminating the strengths and limitations of each in the context of home-based sleep monitoring. Our findings distinctly identify LeNet as the most-amenable architecture for this purpose, with LSTM and BLSTM demonstrating relatively lesser compatibility. Ultimately, this research substantiates the feasibility of automating sleep stage classification employing lightweight neural networks, thereby accommodating scenarios with constrained computational resources. This advancement aims at revolutionizing the field of sleep monitoring, making it more accessible and reliable for individuals in their homes.
Collapse
Affiliation(s)
- Jaime Yeckle
- Department of Electrical and Computer Engineering, University of Puerto Rico, Mayaguez, PR 00681, USA;
| | | |
Collapse
|
23
|
van Gorp H, van Gilst MM, Fonseca P, Overeem S, van Sloun RJG. Modeling the Impact of Inter-Rater Disagreement on Sleep Statistics Using Deep Generative Learning. IEEE J Biomed Health Inform 2023; 27:5599-5609. [PMID: 37561616 DOI: 10.1109/jbhi.2023.3304010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/12/2023]
Abstract
Sleep staging is the process by which an overnight polysomnographic measurement is segmented into epochs of 30 seconds, each of which is annotated as belonging to one of five discrete sleep stages. The resulting scoring is graphically depicted as a hypnogram, and several overnight sleep statistics are derived, such as total sleep time and sleep onset latency. Gold standard sleep staging as performed by human technicians is time-consuming, costly, and comes with imperfect inter-scorer agreement, which also results in inter-scorer disagreement about the overnight statistics. Deep learning algorithms have shown promise in automating sleep scoring, but struggle to model inter-scorer disagreement in sleep statistics. To that end, we introduce a novel technique using conditional generative models based on Normalizing Flows that permits the modeling of the inter-rater disagreement of overnight sleep statistics, termed U-Flow. We compare U-Flow to other automatic scoring methods on a hold-out test set of 70 subjects, each scored by six independent scorers. The proposed method achieves similar sleep staging performance in terms of accuracy and Cohen's kappa on the majority-voted hypnograms. At the same time, U-Flow outperforms the other methods in terms of modeling the inter-rater disagreement of overnight sleep statistics. The consequences of inter-rater disagreement about overnight sleep statistics may be great, and the disagreement potentially carries diagnostic and scientifically relevant information about sleep structure. U-Flow is able to model this disagreement efficiently and can support further investigations into the impact inter-rater disagreement has on sleep medicine and basic sleep research.
Collapse
|
24
|
Liu Z, Qin M, Lu Y, Luo S, Zhang Q. DenSleepNet: DenseNet based model for sleep staging with two-frequency feature fusion and coordinate attention. Biomed Eng Lett 2023; 13:751-761. [PMID: 37872995 PMCID: PMC10590351 DOI: 10.1007/s13534-023-00301-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2023] [Revised: 06/15/2023] [Accepted: 06/23/2023] [Indexed: 10/25/2023] Open
Abstract
Sleep staging is often applied to assess the quality of sleep and also be used to prevent and monitor psychiatric disorders caused by sleep. However, it remains a challenge to extract the discriminative features of salient waveforms in sleep EEG and enable the network to effectively classify sleep stages by emphasizing these crucial features, thus achieving higher accuracy. In this study, an end-to-end deep learning model based on DenseNet for automatic sleep staging is designed and constructed. In the framework, two convolutional branches are devised to extract the underlying features (Two-Frequency Feature) at various frequencies, which are then fused and input into the DenseNet module to extract salient waveform features. After that, the Coordinate Attention mechanism is employed to enhance the localization of salient waveform features by emphasizing the position of salient waveforms and the spatial relationship across the entire frequency spectrum. Finally, the obtained features are accessed to the fully connected for sleep staging. The model was validated with a 20-fold cross-validation procedure on two public available datasets, and the overall accuracy, kappa coefficient, and MF1 score reached 92.9%, 78.7, 0.86 and 90.0%, 75.8, 0.80 on Sleep-EDF-20 and Sleep-EDFx, respectively. Experimental results show that the proposed model achieves competitive performance for sleep staging compared with the reported approaches under the identical conditions.
Collapse
Affiliation(s)
- Zhi Liu
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Meiqiao Qin
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Yunhua Lu
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Sixin Luo
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| | - Qinhan Zhang
- School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China
| |
Collapse
|
25
|
Wong S, Simmons A, Villicana JR, Barnett S. Estimating Patient-Level Uncertainty in Seizure Detection Using Group-Specific Out-of-Distribution Detection Technique. SENSORS (BASEL, SWITZERLAND) 2023; 23:8375. [PMID: 37896469 PMCID: PMC10611125 DOI: 10.3390/s23208375] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/07/2023] [Revised: 09/29/2023] [Accepted: 10/05/2023] [Indexed: 10/29/2023]
Abstract
Epilepsy is a chronic neurological disorder affecting around 1% of the global population, characterized by recurrent epileptic seizures. Accurate diagnosis and treatment are crucial for reducing mortality rates. Recent advancements in machine learning (ML) algorithms have shown potential in aiding clinicians with seizure detection in electroencephalography (EEG) data. However, these algorithms face significant challenges due to the patient-specific variability in seizure patterns and the limited availability of high-quality EEG data for training, causing erratic predictions. These erratic predictions are harmful, especially for high-stake domains in healthcare, negatively affecting patients. Therefore, ensuring safety in AI is of the utmost importance. In this study, we propose a novel ensemble method for uncertainty quantification to identify patients with low-confidence predictions in ML-based seizure detection algorithms. Our approach aims to mitigate high-risk predictions in previously unseen seizure patients, thereby enhancing the robustness of existing seizure detection algorithms. Additionally, our method can be implemented with most of the deep learning (DL) models. We evaluated the proposed method against established uncertainty detection techniques, demonstrating its effectiveness in identifying patients for whom the model's predictions are less certain. Our proposed method managed to achieve 87%, 89% and 75% in accuracy, specificity and sensitivity, respectively. This study represents a novel attempt to improve the reliability and robustness of DL algorithms in the domain of seizure detection. This study underscores the value of integrating uncertainty quantification into ML algorithms for seizure detection, offering clinicians a practical tool to gauge the applicability of ML models for individual patients.
Collapse
Affiliation(s)
- Sheng Wong
- Applied Artificial Intelligence Institute, Deakin University, Burwood, VIC 3125, Australia
| | - Anj Simmons
- Applied Artificial Intelligence Institute, Deakin University, Burwood, VIC 3125, Australia
| | | | - Scott Barnett
- Applied Artificial Intelligence Institute, Deakin University, Burwood, VIC 3125, Australia
| |
Collapse
|
26
|
Phan H, Lorenzen KP, Heremans E, Chen OY, Tran MC, Koch P, Mertins A, Baumert M, Mikkelsen KB, De Vos M. L-SeqSleepNet: Whole-cycle Long Sequence Modeling for Automatic Sleep Staging. IEEE J Biomed Health Inform 2023; 27:4748-4757. [PMID: 37552591 DOI: 10.1109/jbhi.2023.3303197] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/10/2023]
Abstract
Human sleep is cyclical with a period of approximately 90 minutes, implying long temporal dependency in the sleep data. Yet, exploring this long-term dependency when developing sleep staging models has remained untouched. In this work, we show that while encoding the logic of a whole sleep cycle is crucial to improve sleep staging performance, the sequential modelling approach in existing state-of-the-art deep learning models are inefficient for that purpose. We thus introduce a method for efficient long sequence modelling and propose a new deep learning model, L-SeqSleepNet, which takes into account whole-cycle sleep information for sleep staging. Evaluating L-SeqSleepNet on four distinct databases of various sizes, we demonstrate state-of-the-art performance obtained by the model over three different EEG setups, including scalp EEG in conventional Polysomnography (PSG), in-ear EEG, and around-the-ear EEG (cEEGrid), even with a single EEG channel input. Our analyses also show that L-SeqSleepNet is able to alleviate the predominance of N2 sleep (the major class in terms of classification) to bring down errors in other sleep stages. Moreover the network becomes much more robust, meaning that for all subjects where the baseline method had exceptionally poor performance, their performance are improved significantly. Finally, the computation time only grows at a sub-linear rate when the sequence length increases.
Collapse
|
27
|
Lee M, Kwak HG, Kim HJ, Won DO, Lee SW. SeriesSleepNet: an EEG time series model with partial data augmentation for automatic sleep stage scoring. Front Physiol 2023; 14:1188678. [PMID: 37700762 PMCID: PMC10494443 DOI: 10.3389/fphys.2023.1188678] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Accepted: 08/10/2023] [Indexed: 09/14/2023] Open
Abstract
Introduction: We propose an automatic sleep stage scoring model, referred to as SeriesSleepNet, based on convolutional neural network (CNN) and bidirectional long short-term memory (bi-LSTM) with partial data augmentation. We used single-channel raw electroencephalography signals for automatic sleep stage scoring. Methods: Our framework was focused on time series information, so we applied partial data augmentation to learn the connected time information in small series. In specific, the CNN module learns the time information of one epoch (intra-epoch) whereas the bi-LSTM trains the sequential information between the adjacent epochs (inter-epoch). Note that the input of the bi-LSTM is the augmented CNN output. Moreover, the proposed loss function was used to fine-tune the model by providing additional weights. To validate the proposed framework, we conducted two experiments using the Sleep-EDF and SHHS datasets. Results and Discussion: The results achieved an overall accuracy of 0.87 and 0.84 and overall F1-score of 0.80 and 0.78 and kappa value of 0.81 and 0.78 for five-class classification, respectively. We showed that the SeriesSleepNet was superior to the baselines based on each component in the proposed framework. Our architecture also outperformed the state-of-the-art methods with overall F1-score, accuracy, and kappa value. Our framework could provide information on sleep disorders or quality of sleep to automatically classify sleep stages with high performance.
Collapse
Affiliation(s)
- Minji Lee
- Department of Biomedical Software Engineering, The Catholic University of Korea, Bucheon, Republic of Korea
| | - Heon-Gyu Kwak
- Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea
| | - Hyeong-Jin Kim
- Department of Brain and Cognitive Engineering, Korea University, Seoul, Republic of Korea
| | - Dong-Ok Won
- Department of Artificial Intelligence Convergence, Hallym University, Chuncheon, Republic of Korea
| | - Seong-Whan Lee
- Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea
| |
Collapse
|
28
|
Wenjian W, Qian X, Jun X, Zhikun H. DynamicSleepNet: a multi-exit neural network with adaptive inference time for sleep stage classification. Front Physiol 2023; 14:1171467. [PMID: 37250117 PMCID: PMC10213983 DOI: 10.3389/fphys.2023.1171467] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2023] [Accepted: 04/26/2023] [Indexed: 05/31/2023] Open
Abstract
Sleep is an essential human physiological behavior, and the quality of sleep directly affects a person's physical and mental state. In clinical medicine, sleep stage is an important basis for doctors to diagnose and treat sleep disorders. The traditional method of classifying sleep stages requires sleep experts to classify them manually, and the whole process is time-consuming and laborious. In recent years, with the help of deep learning, automatic sleep stage classification has made great progress, especially networks using multi-modal electrophysiological signals, which have greatly improved in terms of accuracy. However, we found that the existing multimodal networks have a large number of redundant calculations in the process of using multiple electrophysiological signals, and the networks become heavier due to the use of multiple signals, and difficult to be used in small devices. To solve these two problems, this paper proposes DynamicSleepNet, a network that can maximize the use of multiple electrophysiological signals and can dynamically adjust between accuracy and efficiency. DynamicSleepNet consists of three effective feature extraction modules (EFEMs) and three classifier modules, each EFEM is connected to a classifier. Each EFEM is able to extract signal features while making the effective features more prominent and the invalid features are suppressed. The samples processed by the EFEM are given to the corresponding classifier for classification, and if the classifier considers the uncertainty of the sample to be below the threshold we set, the sample can be output early without going through the whole network. We validated our model on four datasets. The results show that the highest accuracy of our model outperforms all baselines. With accuracy close to baselines, our model is faster than the baselines by a factor of several to several tens, and the number of parameters of the model is lower or close. The implementation code is available at: https://github.com/Quinella7291/A-Multi-exit-Neural-Network-with-Adaptive-Inference-Time-for-Sleep-Stage-Classification/.
Collapse
Affiliation(s)
- Wang Wenjian
- School of Information Science, Yunnan University, Kunming, China
| | | | | | | |
Collapse
|
29
|
Nasiri S, Ganglberger W, Sun H, Thomas RJ, Westover MB. Exploiting labels from multiple experts in automated sleep scoring. Sleep 2023; 46:zsad034. [PMID: 36795078 PMCID: PMC10171620 DOI: 10.1093/sleep/zsad034] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2023] [Indexed: 02/17/2023] Open
Affiliation(s)
- Samaneh Nasiri
- Department of Neurology, Massachusetts General Hospital (MGH), Boston, MA, USA
- Harvard Medical School, Boston, MA, USA
- Clinical Data Animation Center (CDAC), Boston, MA, USA
- McCance Center for Brain Health, MGH, Boston, MA, USA
| | - Wolfgang Ganglberger
- Harvard Medical School, Boston, MA, USA
- Clinical Data Animation Center (CDAC), Boston, MA, USA
- McCance Center for Brain Health, MGH, Boston, MA, USA
| | - Haoqi Sun
- Harvard Medical School, Boston, MA, USA
- Clinical Data Animation Center (CDAC), Boston, MA, USA
- McCance Center for Brain Health, MGH, Boston, MA, USA
- Department of Medicine, Division of Pulmonary, Critical Care & Sleep, Beth Israel Deaconess Medical Center, Boston, MA, USA
| | - Robert J Thomas
- Harvard Medical School, Boston, MA, USA
- McCance Center for Brain Health, MGH, Boston, MA, USA
| | - M Brandon Westover
- Department of Neurology, Massachusetts General Hospital (MGH), Boston, MA, USA
- Harvard Medical School, Boston, MA, USA
- Clinical Data Animation Center (CDAC), Boston, MA, USA
- McCance Center for Brain Health, MGH, Boston, MA, USA
- Department of Medicine, Division of Pulmonary, Critical Care & Sleep, Beth Israel Deaconess Medical Center, Boston, MA, USA
| |
Collapse
|
30
|
Do not sleep on traditional machine learning. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104429] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
31
|
Supratak A, Haddawy P. Quantifying the impact of data characteristics on the transferability of sleep stage scoring models. Artif Intell Med 2023; 139:102540. [PMID: 37100508 DOI: 10.1016/j.artmed.2023.102540] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Revised: 03/18/2023] [Accepted: 03/27/2023] [Indexed: 04/03/2023]
Abstract
Deep learning models for scoring sleep stages based on single-channel EEG have been proposed as a promising method for remote sleep monitoring. However, applying these models to new datasets, particularly from wearable devices, raises two questions. First, when annotations on a target dataset are unavailable, which different data characteristics affect the sleep stage scoring performance the most and by how much? Second, when annotations are available, which dataset should be used as the source of transfer learning to optimize performance? In this paper, we propose a novel method for computationally quantifying the impact of different data characteristics on the transferability of deep learning models. Quantification is accomplished by training and evaluating two models with significant architectural differences, TinySleepNet and U-Time, under various transfer configurations in which the source and target datasets have different recording channels, recording environments, and subject conditions. For the first question, the environment had the highest impact on sleep stage scoring performance, with performance degrading by over 14% when sleep annotations were unavailable. For the second question, the most useful transfer sources for TinySleepNet and the U-Time models were MASS-SS1 and ISRUC-SG1, containing a high percentage of N1 (the rarest sleep stage) relative to the others. The frontal and central EEGs were preferred for TinySleepNet. The proposed approach enables full utilization of existing sleep datasets for training and planning model transfer to maximize the sleep stage scoring performance on a target problem when sleep annotations are limited or unavailable, supporting the realization of remote sleep monitoring.
Collapse
|
32
|
Fu G, Zhou Y, Gong P, Wang P, Shao W, Zhang D. A Temporal-Spectral Fused and Attention-Based Deep Model for Automatic Sleep Staging. IEEE Trans Neural Syst Rehabil Eng 2023; 31:1008-1018. [PMID: 37022069 DOI: 10.1109/tnsre.2023.3238852] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
Sleep staging is a vital process for evaluating sleep quality and diagnosing sleep-related diseases. Most of the existing automatic sleep staging methods focus on time-domain information and often ignore the transformation relationship between sleep stages. To deal with the above problems, we propose a Temporal-Spectral fused and Attention-based deep neural Network model (TSA-Net) for automatic sleep staging, using a single-channel electroencephalogram (EEG) signal. The TSA-Net is composed of a two-stream feature extractor, feature context learning, and conditional random field (CRF). Specifically, the two-stream feature extractor module can automatically extract and fuse EEG features from time and frequency domains, considering that both temporal and spectral features can provide abundant distinguishing information for sleep staging. Subsequently, the feature context learning module learns the dependencies between features using the multi-head self-attention mechanism and outputs a preliminary sleep stage. Finally, the CRF module further applies transition rules to improve classification performance. We evaluate our model on two public datasets, Sleep-EDF-20 and Sleep-EDF-78. In terms of accuracy, the TSA-Net achieves 86.64% and 82.21% on the Fpz-Cz channel, respectively. The experimental results illustrate that our TSA-Net can optimize the performance of sleep staging and achieve better staging performance than state-of-the-art methods.
Collapse
|
33
|
Lu C, Pathak S, Englebienne G, Seifert C. Channel Contribution in Deep Learning Based Automatic Sleep Scoring-How Many Channels Do We Need? IEEE Trans Neural Syst Rehabil Eng 2023; 31:494-505. [PMID: 37015588 DOI: 10.1109/tnsre.2022.3227040] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
Abstract
Machine learning based sleep scoring methods aim to automate the process of annotating polysomnograms with sleep stages. Although sleep signals of multiple modalities and channels should contain more information according to sleep guidelines, most multi-channel multi-modal models in the literature showed only a little performance improvement compared to single-channel EEG models and sometimes even failed to outperform them. In this paper, we investigate whether the high performance of single-channel EEG models can be attributed to specific model features in their deep learning architectures and to which extent multi-channel multi-modal models take the information from different channels of modalities into account. First, we transfer the model features from single-channel EEG models, such as combinations of small and large filters in CNNs, to multi-channel multi-modal models and measure their impacts. Second, we employ two explainability methods, the layer-wise relevance propagation as post-hoc and the embedded channel attention network as intrinsic interpretability methods, to measure the contribution of different channels on predictive performance. We find that i) single-channel model features can improve the performance of multi-channel multi-modal models and ii) multi-channel multi-modal models focus on one important channel per modality and use the remaining channels to complement the information of the focused channels. Our results suggest that more advanced methods for aggregating channel information using complementary information from other channels may improve sleep scoring performance for multi-channel multi-modal models.
Collapse
|
34
|
Fang Y, Xia Y, Chen P, Zhang J, Zhang Y. A dual-stream deep neural network integrated with adaptive boosting for sleep staging. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
35
|
Chen Z, Yang Z, Wang D, Zhu X, Ono N, Altaf-Ul-Amin MD, Kanaya S, Huang M. Sleep Staging Framework with Physiologically Harmonized Sub-Networks. Methods 2023; 209:18-28. [PMID: 36436760 DOI: 10.1016/j.ymeth.2022.11.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Revised: 11/15/2022] [Accepted: 11/21/2022] [Indexed: 11/26/2022] Open
Abstract
Sleep screening is an important tool for both healthcare and neuroscientific research. Automatic sleep scoring is an alternative to the time-consuming gold-standard manual scoring procedure. Recently there have seen promising results on automatic stage scoring by extracting spatio-temporal features via deep neural networks from electroencephalogram (EEG). However, such methods fail to consistently yield good performance due to a missing piece in data representation: the medical criterion of the sleep scoring task on top of EEG features. We argue that capturing stage-specific features that satisfy the criterion of sleep medicine is non-trivial for automatic sleep scoring. This paper considers two criteria: Transient stage marker and Overall profile of EEG features, then we propose a physiologically meaningful framework for sleep stage scoring via mixed deep neural networks. The framework consists of two sub-networks: feature extraction networks, constructed in consideration of the physiological characteristics of sleep, and an attention-based scoring decision network. Moreover, we quantize the framework for potential use under an IoT setting. For proof-of-concept, the performance of the proposed framework is demonstrated by introducing multiple sleep datasets with the largest comprising 42,560 h recorded from 5,793 subjects. From the experiment results, the proposed method achieves a competitive stage scoring performance, especially for Wake, N2, and N3, with higher F1 scores of 0.92, 0.86, and 0.88, respectively. Moreover, the feasibility analysis of framework quantization provides a potential for future implementation in the edge computing field and clinical settings.
Collapse
Affiliation(s)
- Zheng Chen
- Graduate School of Engineering Science, Osaka University, Japan.
| | - Ziwei Yang
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Japan
| | - Dong Wang
- Graduate School and Faculty of Information Science and Electrical Engineering, Kyushu University, Japan
| | - Xin Zhu
- Biomedical Information Engineering Lab, The University of Aizu, Japan
| | - Naoaki Ono
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Japan; Data Science Center, Nara Insitute of Science and Technology, Japan
| | - M D Altaf-Ul-Amin
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Japan
| | - Shigehiko Kanaya
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Japan; Data Science Center, Nara Insitute of Science and Technology, Japan
| | - Ming Huang
- Graduate School of Science and Technology, Nara Institute of Science and Technology, Japan; Data Science Center, Nara Insitute of Science and Technology, Japan.
| |
Collapse
|
36
|
Sholeyan AE, Rahatabad FN, Setarehdan SK. Designing an Automatic Sleep Staging System Using Deep Convolutional Neural Network Fed by Nonlinear Dynamic Transformation. J Med Biol Eng 2022. [DOI: 10.1007/s40846-022-00771-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2022]
|
37
|
Yu R, Zhou Z, Wu S, Gao X, Bin G. MRASleepNet: a multi-resolution attention network for sleep stage classification using single-channel EEG. J Neural Eng 2022; 19. [PMID: 36379059 DOI: 10.1088/1741-2552/aca2de] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2022] [Accepted: 11/15/2022] [Indexed: 11/16/2022]
Abstract
Objective. Computerized classification of sleep stages based on single-lead electroencephalography (EEG) signals is important, but still challenging. In this paper, we proposed a deep neural network called MRASleepNet for automatic sleep stage classification using single-channel EEG signals.Approach. The proposed MRASleepNet model consisted of a feature extraction (FE) module, a multi-resolution attention (MRA) module, and a gated multilayer perceptron (gMLP) module, as well as a direct pathway for computing statistical features. The FE, MRA, and gMLP modules were used to extract features, establish feature attention, and obtain temporal relationships between features, respectively. EEG signals were normalized and cut into 30 s segments, and enhanced by incorporating contextual information from adjacent data segments. After data enhancement, the 40 s data segments were input to the MRASleepNet model. The model was evaluated on the SleepEDF and the cyclic alternating pattern (CAP) databases, using such metrics as the accuracy, Kappa, and macro-F1 (MF1).Main results.For the SleepEDF-20 database, the proposed model had an accuracy of 84.5%, an MF1 of 0.789, and a Kappa of 0.786. For the SleepEDF-78 database, the model had an accuracy of 81.4%, an MF1 of 0.754, and a Kappa of 0.743. For the CAP database, the model had an accuracy of 74.3%, an MF1 of 0.656, and a Kappa of 0.652. The proposed model achieved satisfactory performance in automatic sleep stage classification tasks.Significance. The time- and frequency-domain features extracted by the FE module and filtered by the MRA module, together with the temporal features extracted by the gMLP module and the statistical features extracted by the statistical highway, enabled the proposed model to obtain a satisfying performance in sleep staging. The proposed MRASleepNet model may be used as a new deep learning method for automatic sleep stage classification. The code of MRASleepNet will be made available publicly onhttps://github.com/YuRui8879/.
Collapse
Affiliation(s)
- Rui Yu
- Department of Biomedical Engineering, Faculty of Environment and Life, Beijing University of Technology, Beijing 100124, People's Republic of China
| | - Zhuhuang Zhou
- Department of Biomedical Engineering, Faculty of Environment and Life, Beijing University of Technology, Beijing 100124, People's Republic of China
| | - Shuicai Wu
- Department of Biomedical Engineering, Faculty of Environment and Life, Beijing University of Technology, Beijing 100124, People's Republic of China
| | - Xiaorong Gao
- Department of Biomedical Engineering, Tsinghua University, 100084 Beijing, People's Republic of China
| | - Guangyu Bin
- Department of Biomedical Engineering, Faculty of Environment and Life, Beijing University of Technology, Beijing 100124, People's Republic of China
| |
Collapse
|
38
|
Lemkhenter A, Favaro P. Towards Sleep Scoring Generalization Through Self-Supervised Meta-Learning. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:2961-2966. [PMID: 36085742 DOI: 10.1109/embc48229.2022.9871056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
In this work we introduce a novel meta-learning method for sleep scoring based on self-supervised learning. Our approach aims at building models for sleep scoring that can generalize across different patients and recording facilities, but do not require a further adaptation step to the target data. Towards this goal, we build our method on top of the Model Agnostic Meta-Learning (MAML) framework by incorporating a self-supervised learning (SSL) stage, and call it S2MAML. We show that S2MAML can significantly outperform MAML. The gain in performance comes from the SSL stage, which we base on a general purpose pseudo-task that limits the overfitting to the subject-specific patterns present in the training dataset. We show that S2MAML outperforms standard supervised learning and MAML on the SC, ST, ISRUC, UCD and CAP datasets. Clinical relevance- Our work tackles the generalization problem of automatic sleep scoring models. This is one of the main hurdles that limits the adoption of such models for clinical and research sleep studies.
Collapse
|
39
|
Zhao C, Li J, Guo Y. SleepContextNet: A temporal context network for automatic sleep staging based single-channel EEG. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 220:106806. [PMID: 35461126 DOI: 10.1016/j.cmpb.2022.106806] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/30/2021] [Revised: 04/07/2022] [Accepted: 04/07/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND AND OBJECTIVE Single-channel EEG is the most popular choice of sensing modality in sleep staging studies, because it widely conforms to the sleep staging guidelines. The current deep learning method using single-channel EEG signals for sleep staging mainly extracts the features of its surrounding epochs to obtain the short-term temporal context information of EEG epochs, and ignore the influence of the long-term temporal context information on sleep staging. However, the long-term context information includes sleep stage transition rules in a sleep cycle, which can further improve the performance of sleep staging. The aim of this research is to develop a temporal context network to capture the long-term context between EEG sleep stages. METHODS In this paper, we design a sleep staging network named SleepContextNet for sleep stage sequence. SleepContextNet can extract and utilize the long-term temporal context between consecutive EEG epochs, and combine it with the short-term context. we utilize Convolutional Neural Network(CNN) layers for learning representative features from each sleep stage and the representation features sequence learned are fed into a Recurrent Neural Network(RNN) layer for learning long-term and short-term context information among sleep stage in chronological order. In addition, we design a data augmentation algorithm for EEG to retain the long-term context information without changing the number of samples. RESULTS We evaluate the performance of our proposed network using four public datasets, the 2013 version of Sleep-EDF (SEDF), the 2018 version of Sleep-EDF Expanded (SEDFX), Sleep Heart Health Study (SHHS) and the CAP Sleep Database. The experimental results demonstrate that SleepContextNet outperforms state-of-the-art techniques in terms of different evaluation metrics by capturing long-term and short-term temporal context information. On average, accuracy of 84.8% in SEDF, 82.7% in SEDFX, 86.4% in SHHS and 78.8% in CAP are obtained under subject-independent cross validation. CONCLUSIONS The network extracts the long-term and short-term temporal context information of sleep stages from the sequence features, which utilizes the temporal dependencies among the EEG epochs effectively and improves the accuracy of sleep stages. The sleep staging method based on forward temporal context information is suitable for real-time family sleep monitoring system.
Collapse
Affiliation(s)
- Caihong Zhao
- School of Electronic and Engineer, Heilongjiang University, Harbin, 150080, China; School of Computer Science and Technology, Heilongjiang University, Harbin, 150080, China
| | - Jinbao Li
- Shandong Artificial Intelligence Institute, Qilu University of Technology (Shandong Academy of Sciences), Jinan, 250353, China.
| | - Yahong Guo
- School of Computer Science and Technology, Qilu University of Technology (Shandong Academy of Sciences), Jinan, 250353, China.
| |
Collapse
|
40
|
Zhu H, Wu Y, Shen N, Fan J, Tao L, Fu C, Yu H, Wan F, Pun SH, Chen C, Chen W. The Masking Impact of Intra-artifacts in EEG on Deep Learning-based Sleep Staging Systems: A Comparative Study. IEEE Trans Neural Syst Rehabil Eng 2022; 30:1452-1463. [PMID: 35536800 DOI: 10.1109/tnsre.2022.3173994] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Elimination of intra-artifacts in EEG has been overlooked in most of the existing sleep staging systems, especially in deep learning-based approaches. Whether intra-artifacts, originated from the eye movement, chin muscle firing, or heart beating, etc., in EEG signals would lead to a positive or a negative masking effect on deep learning-based sleep staging systems was investigated in this paper. We systematically analyzed several traditional pre-processing methods involving fast Independent Component Analysis (FastICA), Information Maximization (Infomax), and Second-order Blind Source Separation (SOBI). On top of these methods, a SOBI-WT method based on the joint use of the SOBI and Wavelet Transform (WT) is proposed. It offered an effective solution for suppressing artifact components while retaining residual informative data. To provide a comprehensive comparative analysis, these pre-processing methods were applied to eliminate the intra-artifacts and the processed signals were fed to two ready-to-use deep learning models, namely two-step hierarchical neural network (THNN) and SimpleSleepNet for automatic sleep staging. The evaluation was performed on two widely used public datasets, Montreal Archive of Sleep Studies (MASS) and Sleep-EDF Expanded, and a clinical dataset that was collected in Huashan Hospital of Fudan University, Shanghai, China (HSFU). The proposed SOBI-WT method increased the accuracy from 79.0% to 81.3% on MASS, 83.3% to 85.7% on Sleep-EDF Expanded, and 75.5% to 77.1% on HSFU compared with the raw EEG signal, respectively. Experimental results demonstrate that the intra-artifacts bring out a masking negative impact on the deep learning-based sleep staging systems and the proposed SOBI-WT method has the best performance in diminishing this negative impact compared with other artifact elimination methods.
Collapse
|
41
|
Phan H, Mikkelsen K. Automatic sleep staging of EEG signals: recent development, challenges, and future directions. Physiol Meas 2022; 43. [PMID: 35320788 DOI: 10.1088/1361-6579/ac6049] [Citation(s) in RCA: 47] [Impact Index Per Article: 15.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2021] [Accepted: 03/23/2022] [Indexed: 11/11/2022]
Abstract
Modern deep learning holds a great potential to transform clinical practice on human sleep. Teaching a machine to carry out routine tasks would be a tremendous reduction in workload for clinicians. Sleep staging, a fundamental step in sleep practice, is a suitable task for this and will be the focus in this article. Recently, automatic sleep staging systems have been trained to mimic manual scoring, leading to similar performance to human sleep experts, at least on scoring of healthy subjects. Despite tremendous progress, we have not seen automatic sleep scoring adopted widely in clinical environments. This review aims to give a shared view of the authors on the most recent state-of-the-art development in automatic sleep staging, the challenges that still need to be addressed, and the future directions for automatic sleep scoring to achieve clinical value.
Collapse
Affiliation(s)
- Huy Phan
- School of Electronic Engineering and Computer Science, Queen Mary University of London, Mile End Rd, London, E1 4NS, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
| | - Kaare Mikkelsen
- Department of Electrical and Computer Engineering, Aarhus Universitet, Finlandsgade 22, Aarhus, 8000, DENMARK
| |
Collapse
|