1
|
Jin Y, Liu J, Zhou Y, Chen R, Chen H, Duan W, Chen Y, Zhang XL. CRDet: A circle representation detector for lung granulomas based on multi-scale attention features with center point calibration. Comput Med Imaging Graph 2024; 113:102354. [PMID: 38341946 DOI: 10.1016/j.compmedimag.2024.102354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Revised: 02/04/2024] [Accepted: 02/04/2024] [Indexed: 02/13/2024]
Abstract
Lung granuloma is a very common lung disease, and its specific diagnosis is important for determining the exact cause of the disease as well as the prognosis of the patient. And, an effective lung granuloma detection model based on computer-aided diagnostics (CAD) can help pathologists to localize granulomas, thereby improving the efficiency of the specific diagnosis. However, for lung granuloma detection models based on CAD, the significant size differences between granulomas and how to better utilize the morphological features of granulomas are both critical challenges to be addressed. In this paper, we propose an automatic method CRDet to localize granulomas in histopathological images and deal with these challenges. We first introduce the multi-scale feature extraction network with self-attention to extract features at different scales at the same time. Then, the features will be converted to circle representations of granulomas by circle representation detection heads to achieve the alignment of features and ground truth. In this way, we can also more effectively use the circular morphological features of granulomas. Finally, we propose a center point calibration method at the inference stage to further optimize the circle representation. For model evaluation, we built a lung granuloma circle representation dataset named LGCR, including 288 images from 50 subjects. Our method yielded 0.316 mAP and 0.571 mAR, outperforming the state-of-the-art object detection methods on our proposed LGCR.
Collapse
Affiliation(s)
- Yu Jin
- Institute of Artificial Intelligence, School of Computer Science, Wuhan University, Wuhan, China
| | - Juan Liu
- Institute of Artificial Intelligence, School of Computer Science, Wuhan University, Wuhan, China.
| | - Yuanyuan Zhou
- Department of Immunology, TaiKang Medical School (School of Basic Medical Sciences), Wuhan University, Wuhan, China; Hubei Province Key Laboratory of Allergy and Immunology, Wuhan University, Wuhan, China
| | - Rong Chen
- Wuhan Jinyintan Hospital, Tongji Medical College of Huazhong University of Science and Technology, Wuhan, China
| | - Hua Chen
- Institute of Artificial Intelligence, School of Computer Science, Wuhan University, Wuhan, China
| | - Wensi Duan
- Institute of Artificial Intelligence, School of Computer Science, Wuhan University, Wuhan, China
| | - Yuqi Chen
- Institute of Artificial Intelligence, School of Computer Science, Wuhan University, Wuhan, China
| | - Xiao-Lian Zhang
- Department of Immunology, TaiKang Medical School (School of Basic Medical Sciences), Wuhan University, Wuhan, China; Hubei Province Key Laboratory of Allergy and Immunology, Wuhan University, Wuhan, China
| |
Collapse
|
2
|
Wei W, Zhang L, Yang K, Li J, Cui N, Han Y, Zhang N, Yang X, Tan H, Wang K. A lightweight network for traffic sign recognition based on multi-scale feature and attention mechanism. Heliyon 2024; 10:e26182. [PMID: 38420439 PMCID: PMC10900943 DOI: 10.1016/j.heliyon.2024.e26182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/24/2023] [Revised: 01/29/2024] [Accepted: 02/08/2024] [Indexed: 03/02/2024] Open
Abstract
Traffic sign recognition is an important part of intelligent transportation system. It uses computer vision and traffic sign recognition technology to detect and recognize traffic signs on the road automatically. In this paper, we propose a lightweight model for traffic sign recognition based on convolutional neural networks called ConvNeSe. Firstly, the feature extraction module of the model is constructed using the Depthwise Separable Convolution and Inverted Residuals structures. The model extracts multi-scale features with strong representation ability by optimizing the structure of convolutional neural networks and fusing of features. Then, the model introduces Squeeze and Excitation Block (SE Block) to improve the attention to important features, which can capture key information of traffic sign images. Finally, the accuracy of the model in the German Traffic Sign Recognition Benchmark Database (GTSRB) is 99.85%. At the same time, the model has good robustness according to the results of ablation experiments.
Collapse
Affiliation(s)
- Wei Wei
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Lili Zhang
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Kang Yang
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Jing Li
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Ning Cui
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Yucheng Han
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Ning Zhang
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Xudong Yang
- Beijing Institute of Petrochemical Technology, Beijing, 102617, China
| | - Hongxin Tan
- Science and Technology on Complex Aviation Systems Simulation Laboratory, Beijing, 100076, China
| | - Kai Wang
- Institute of National Defense Science and Technology Innovation, Academy of Military Sciences, Beijing, 100036, China
| |
Collapse
|
3
|
Zhang J, Zhang Y, Jin Y, Xu J, Xu X. MDU-Net: multi-scale densely connected U-Net for biomedical image segmentation. Health Inf Sci Syst 2023; 11:13. [PMID: 36925619 PMCID: PMC10011258 DOI: 10.1007/s13755-022-00204-9] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2022] [Accepted: 11/02/2022] [Indexed: 03/18/2023] Open
Abstract
Biomedical image segmentation plays a central role in quantitative analysis, clinical diagnosis, and medical intervention. In the light of the fully convolutional networks (FCN) and U-Net, deep convolutional networks (DNNs) have made significant contributions to biomedical image segmentation applications. In this paper, we propose three different multi-scale dense connections (MDC) for the encoder, the decoder of U-shaped architectures, and across them. Based on three dense connections, we propose a multi-scale densely connected U-Net (MDU-Net) for biomedical image segmentation. MDU-Net directly fuses the neighboring feature maps with different scales from both higher layers and lower layers to strengthen feature propagation in the current layer. Multi-scale dense connections, which contain shorter connections between layers close to the input and output, also make a much deeper U-Net possible. Besides, we introduce quantization to alleviate the potential overfitting in dense connections, and further improve the segmentation performance. We evaluate our proposed model on the MICCAI 2015 Gland Segmentation (GlaS) dataset. The three MDC improve U-Net performance by up to 1.8% on test A and 3.5% on test B in the MICCAI Gland dataset. Meanwhile, the MDU-Net with quantization obviously improves the segmentation performance of original U-Net.
Collapse
Affiliation(s)
- Jiawei Zhang
- The Department of New Networks, Peng Cheng Laboratory, Shenzhen, Guangdong China
- Department of Cardiovascular Surgery, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences),Southern Medical University, Guangzhou, Guangdong China
- Cyberspace Institute of Advanced Technology, Guangzhou University, Guangzhou, Guangdong China
- Institute for Sustainable Industries & Livable Cities, Victoria University, Melbourne, VIC Australia
| | - Yanchun Zhang
- The Department of New Networks, Peng Cheng Laboratory, Shenzhen, Guangdong China
- Cyberspace Institute of Advanced Technology, Guangzhou University, Guangzhou, Guangdong China
- Institute for Sustainable Industries & Livable Cities, Victoria University, Melbourne, VIC Australia
| | - Yuzhen Jin
- Shanghai Key Laboratory of Data Science, School of Computer Science, Fudan University, Shanghai, China
| | - Jilan Xu
- Shanghai Key Laboratory of Data Science, School of Computer Science, Fudan University, Shanghai, China
| | - Xiaowei Xu
- Department of Cardiovascular Surgery, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences),Southern Medical University, Guangzhou, Guangdong China
| |
Collapse
|
4
|
彭 昆, 张 桂, 王 杰, 储 珺. [Non-rigid registration for medical images based on deformable convolution and multi-scale feature focusing modules]. Sheng Wu Yi Xue Gong Cheng Xue Za Zhi 2023; 40:492-498. [PMID: 37380388 PMCID: PMC10307602 DOI: 10.7507/1001-5515.202301012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Revised: 04/21/2023] [Indexed: 06/30/2023]
Abstract
Non-rigid registration plays an important role in medical image analysis. U-Net has been proven to be a hot research topic in medical image analysis and is widely used in medical image registration. However, existing registration models based on U-Net and its variants lack sufficient learning ability when dealing with complex deformations, and do not fully utilize multi-scale contextual information, resulting insufficient registration accuracy. To address this issue, a non-rigid registration algorithm for X-ray images based on deformable convolution and multi-scale feature focusing module was proposed. First, it used residual deformable convolution to replace the standard convolution of the original U-Net to enhance the expression ability of registration network for image geometric deformations. Then, stride convolution was used to replace the pooling operation of the downsampling operation to alleviate feature loss caused by continuous pooling. In addition, a multi-scale feature focusing module was introduced to the bridging layer in the encoding and decoding structure to improve the network model's ability of integrating global contextual information. Theoretical analysis and experimental results both showed that the proposed registration algorithm could focus on multi-scale contextual information, handle medical images with complex deformations, and improve the registration accuracy. It is suitable for non-rigid registration of chest X-ray images.
Collapse
Affiliation(s)
- 昆 彭
- 南昌航空大学 计算机视觉研究所(南昌 330063)Institute of Computer Vision, Nanchang Hangkong University, Nanchang 330063, P. R. China
| | - 桂梅 张
- 南昌航空大学 计算机视觉研究所(南昌 330063)Institute of Computer Vision, Nanchang Hangkong University, Nanchang 330063, P. R. China
| | - 杰 王
- 南昌航空大学 计算机视觉研究所(南昌 330063)Institute of Computer Vision, Nanchang Hangkong University, Nanchang 330063, P. R. China
| | - 珺 储
- 南昌航空大学 计算机视觉研究所(南昌 330063)Institute of Computer Vision, Nanchang Hangkong University, Nanchang 330063, P. R. China
| |
Collapse
|
5
|
Tong N, Xu Y, Zhang J, Gou S, Li M. Robust and efficient abdominal CT segmentation using shape constrained multi-scale attention network. Phys Med 2023; 110:102595. [PMID: 37178624 DOI: 10.1016/j.ejmp.2023.102595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/22/2022] [Revised: 03/02/2023] [Accepted: 04/17/2023] [Indexed: 05/15/2023] Open
Abstract
PURPOSE Although many deep learning-based abdominal multi-organ segmentation networks have been proposed, the various intensity distributions and organ shapes of the CT images from multi-center, multi-phase with various diseases introduce new challenges for robust abdominal CT segmentation. To achieve robust and efficient abdominal multi-organ segmentation, a new two-stage method is presented in this study. METHODS A binary segmentation network is used for coarse localization, followed by a multi-scale attention network for the fine segmentation of liver, kidney, spleen, and pancreas. To constrain the organ shapes produced by the fine segmentation network, an additional network is pre-trained to learn the shape features of the organs with serious diseases and then employed to constrain the training of the fine segmentation network. RESULTS The performance of the presented segmentation method was extensively evaluated on the multi-center data set from the Fast and Low GPU Memory Abdominal oRgan sEgmentation (FLARE) challenge, which was held in conjunction with International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) 2021. Dice Similarity Coefficient (DSC) and Normalized Surface Dice (NSD) were calculated to quantitatively evaluate the segmentation accuracy and efficiency. An average DSC and NSD of 83.7% and 64.4% were achieved, and our method finally won the second place among more than 90 participating teams. CONCLUSIONS The evaluation results on the public challenge demonstrate that our method shows promising performance in robustness and efficiency, which may promote the clinical application of the automatic abdominal multi-organ segmentation.
Collapse
Affiliation(s)
- Nuo Tong
- AI-based Big Medical Imaging Data Frontier Research Center, Academy of Advanced Interdisciplinary Research, Xidian University, Xi'an, Shaanxi 710071, China
| | - Yinan Xu
- Key Lab of Intelligent Perception and Image Understanding of Ministry of Education, Xidian University, Xi'an, Shaanxi, 710071, China
| | - Jinsong Zhang
- Xijing Hospital of Air Force Military Medical University, Xian, Shaanxi 710032, China
| | - Shuiping Gou
- AI-based Big Medical Imaging Data Frontier Research Center, Academy of Advanced Interdisciplinary Research, Xidian University, Xi'an, Shaanxi 710071, China; Key Lab of Intelligent Perception and Image Understanding of Ministry of Education, Xidian University, Xi'an, Shaanxi, 710071, China.
| | - Mengbin Li
- Xijing Hospital of Air Force Military Medical University, Xian, Shaanxi 710032, China.
| |
Collapse
|
6
|
Liu P, Du J, Vong CM. A novel sequential structure for lightweight multi-scale feature learning under limited available images. Neural Netw 2023; 164:124-134. [PMID: 37148608 DOI: 10.1016/j.neunet.2023.04.023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2022] [Revised: 04/10/2023] [Accepted: 04/16/2023] [Indexed: 05/08/2023]
Abstract
Although multi-scale feature learning can improve the performances of deep models, its parallel structure quadratically increases the model parameters and causes deep models to become larger and larger when enlarging the receptive fields (RFs). This leads to deep models easily suffering from over-fitting issue in many practical applications where the available training samples are always insufficient or limited. In addition, under this limited situation, although lightweight models (with fewer model parameters) can effectively reduce over-fitting, they may suffer from under-fitting because of insufficient training data for effective feature learning. In this work, a lightweight model called Sequential Multi-scale Feature Learning Network (SMF-Net) is proposed to alleviate these two issues simultaneously using a novel sequential structure of multi-scale feature learning. Compared to both deep and lightweight models, the proposed sequential structure in SMF-Net can easily extract features with larger RFs for multi-scale feature learning only with a few and linearly increased model parameters. The experimental results on both classification and segmentation tasks demonstrate that our SMF-Net only has 1.25M model parameters (5.3% of Res2Net50) with 0.7G FLOPS (14.6% of Res2Net50) for classification and 1.54M parameters (8.9% of UNet) with 3.35G FLOPs (10.9% of UNet) for segmentation but achieves higher accuracy than SOTA deep models and lightweight models, even when the training data is very limited available.
Collapse
Affiliation(s)
- Peng Liu
- Department of Computer and Information Science, University of Macau, 999078, Macao Special Administrative Region of China.
| | - Jie Du
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, 518060, China.
| | - Chi-Man Vong
- Department of Computer and Information Science, University of Macau, 999078, Macao Special Administrative Region of China.
| |
Collapse
|
7
|
Pan X, Gao X, Wang H, Zhang W, Mu Y, He X. Temporal-based Swin Transformer network for workflow recognition of surgical video. Int J Comput Assist Radiol Surg 2023; 18:139-47. [PMID: 36331795 DOI: 10.1007/s11548-022-02785-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2022] [Accepted: 10/21/2022] [Indexed: 11/06/2022]
Abstract
PURPOSE Surgical workflow recognition has emerged as an important part of computer-assisted intervention systems for the modern operating room, which also is a very challenging problem. Although the CNN-based approach achieves excellent performance, it does not learn global and long-range semantic information interactions well due to the inductive bias inherent in convolution. METHODS In this paper, we propose a temporal-based Swin Transformer network (TSTNet) for the surgical video workflow recognition task. TSTNet contains two main parts: the Swin Transformer and the LSTM. The Swin Transformer incorporates the attention mechanism to encode remote dependencies and learn highly expressive representations. The LSTM is capable of learning long-range dependencies and is used to extract temporal information. The TSTNet organically combines the two components to extract spatiotemporal features that contain more contextual information. In particular, based on a full understanding of the natural features of the surgical video, we propose a priori revision algorithm (PRA) using a priori information about the sequence of the surgical phase. This strategy optimizes the output of TSTNet and further improves the recognition performance. RESULTS We conduct extensive experiments using the Cholec80 dataset to validate the effectiveness of the TSTNet-PRA method. Our method achieves excellent performance on the Cholec80 dataset, which accuracy is up to 92.8% and greatly exceeds the state-of-the-art methods. CONCLUSION By modelling remote temporal information and multi-scale visual information, we propose the TSTNet-PRA method. It was evaluated on a large public dataset, showing a high recognition capability superior to other spatiotemporal networks.
Collapse
|
8
|
吴 玉, 林 岚, 吴 水. [Multimodal high-grade glioma semantic segmentation network with multi-scale and multi-attention fusion mechanism]. Sheng Wu Yi Xue Gong Cheng Xue Za Zhi 2022; 39:433-440. [PMID: 35788512 PMCID: PMC10950780 DOI: 10.7507/1001-5515.202103021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 03/05/2021] [Revised: 04/25/2022] [Indexed: 06/15/2023]
Abstract
Glioma is a primary brain tumor with high incidence rate. High-grade gliomas (HGG) are those with the highest degree of malignancy and the lowest degree of survival. Surgical resection and postoperative adjuvant chemoradiotherapy are often used in clinical treatment, so accurate segmentation of tumor-related areas is of great significance for the treatment of patients. In order to improve the segmentation accuracy of HGG, this paper proposes a multi-modal glioma semantic segmentation network with multi-scale feature extraction and multi-attention fusion mechanism. The main contributions are, (1) Multi-scale residual structures were used to extract features from multi-modal gliomas magnetic resonance imaging (MRI); (2) Two types of attention modules were used for features aggregating in channel and spatial; (3) In order to improve the segmentation performance of the whole network, the branch classifier was constructed using ensemble learning strategy to adjust and correct the classification results of the backbone classifier. The experimental results showed that the Dice coefficient values of the proposed segmentation method in this article were 0.909 7, 0.877 3 and 0.839 6 for whole tumor, tumor core and enhanced tumor respectively, and the segmentation results had good boundary continuity in the three-dimensional direction. Therefore, the proposed semantic segmentation network has good segmentation performance for high-grade gliomas lesions.
Collapse
Affiliation(s)
- 玉超 吴
- 北京工业大学 环境与生命科学学院 生物医学工程系 智能化生理测量与临床转化北京市国际科研合作基地(北京 100124)Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Intelligent Physiological Measurement and Clinical Translation, Beijing International Base for Scientific and Technological Cooperation, Beijing 100124, P. R. China
| | - 岚 林
- 北京工业大学 环境与生命科学学院 生物医学工程系 智能化生理测量与临床转化北京市国际科研合作基地(北京 100124)Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Intelligent Physiological Measurement and Clinical Translation, Beijing International Base for Scientific and Technological Cooperation, Beijing 100124, P. R. China
| | - 水才 吴
- 北京工业大学 环境与生命科学学院 生物医学工程系 智能化生理测量与临床转化北京市国际科研合作基地(北京 100124)Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Intelligent Physiological Measurement and Clinical Translation, Beijing International Base for Scientific and Technological Cooperation, Beijing 100124, P. R. China
| |
Collapse
|
9
|
Yan B, Cao M, Gong W, Wei B. Multi-scale brain tumor segmentation combined with deep supervision. Int J Comput Assist Radiol Surg 2021. [PMID: 34894336 DOI: 10.1007/s11548-021-02515-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2021] [Accepted: 09/29/2021] [Indexed: 10/19/2022]
Abstract
PURPOSE Fully convolutional neural networks (FCNNs) have achieved good performance in the field of medical image segmentation. FCNNs that use multimodal images and multi-scale feature extraction have higher accuracy for brain tumor segmentation. Therefore, we have made some improvements to U-Net for fully automated segmentation of gliomas using multimodal images. And we named it multi-scale dilate network with deep supervision (MSD-Net). METHODS MSD-Net is a symmetrical structure composed of a down-sampling process and an up-sampling process. In the down-sampling process, we use the multi-scale feature extraction block (ME) to extract multi-scale features and focus on primary features. Unlike other methods, ME consists of dilate convolution and standard convolution. Dilate convolution extracts multi-scale informations and standard convolution merges features of different scales. Hence, the output of the ME contains local information and global information. During the up-sampling process, we add a deep supervision block (DSB), which can shorten the length of back-propagation. In this paper, we pay more attention to the importance of shallow features for feature restoration. RESULTS Our network validated in the BraTS17's validation dataset. The DSC scores of MSD-Net for complete tumor, tumor core and enhancing tumor were 0.88, 0.81 and 0.78, respectively, which outperforms most networks. CONCLUSION This study shows that ME enhances the feature extraction ability of the network and improves the accuracy of segmentation results. DSB speeds up the convergence of the network. In addition, we should also pay attention to the contribution of shallow features to feature restoration.
Collapse
|
10
|
Chen K, Zhang C, Ma J, Wang G, Zhang J. Sleep staging from single-channel EEG with multi-scale feature and contextual information. Sleep Breath 2019; 23:1159-67. [PMID: 30863994 DOI: 10.1007/s11325-019-01789-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2018] [Revised: 01/16/2019] [Accepted: 01/26/2019] [Indexed: 01/16/2023]
Abstract
PURPOSE Portable sleep monitoring devices with less-attached sensors and high-accuracy sleep staging methods can expedite sleep disorder diagnosis. The aim of this study was to propose a single-channel EEG sleep staging model, SleepStageNet, which extracts sleep EEG features by multi-scale convolutional neural networks (CNN) and then infers the type of sleep stages by capturing the contextual information between adjacent epochs using recurrent neural networks (RNN) and conditional random field (CRF). METHODS To verify the feasibility of our model, two datasets, one composed by two different single-channel EEGs (Fpz-Cz and Pz-Oz) on 20 healthy people and one composed by a single-channel EEG (F4-M1) on 104 obstructive sleep apnea (OSA) patients with different severities, were examined. The corresponding sleep stages were scored as four states (wake, REM, light sleep, and deep sleep). The accuracy measures were obtained from epoch-by-epoch comparison between the model and PSG scorer, and the agreement between them was quantified with Cohen's kappa (ҡ). RESULTS Our model achieved superior performance with average accuracy (Fpz-Cz, 0.88; Pz-Oz, 0.85) and ҡ (Fpz-Cz, 0.82; Pz-Oz, 0.77) on the healthy people. Furthermore, we validated this model on the OSA patients with average accuracy (F4-M1, 0.80) and ҡ (F4-M1, 0.67). Our model significantly improved the accuracy and ҡ compared to previous methods. CONCLUSIONS The proposed SleepStageNet has proved feasible for assessment of sleep architecture among OSA patients using single-channel EEG. We suggest that this technological advancement could augment the current use of home sleep apnea testing.
Collapse
|