1
|
Jannat M, Birahim SA, Hasan MA, Roy T, Sultana L, Sarker H, Fairuz S, Abdallah HA. Lung Segmentation with Lightweight Convolutional Attention Residual U-Net. Diagnostics (Basel) 2025; 15:854. [PMID: 40218203 PMCID: PMC11988706 DOI: 10.3390/diagnostics15070854] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2025] [Revised: 03/22/2025] [Accepted: 03/25/2025] [Indexed: 04/14/2025] Open
Abstract
Background: Examining chest radiograph images (CXR) is an intricate and time-consuming process, sometimes requiring the identification of many anomalies at the same time. Lung segmentation is key to overcoming this challenge through different deep learning (DL) techniques. Many researchers are working to improve the performance and efficiency of lung segmentation models. This article presents a DL-based approach to accurately identify the lung mask region in CXR images to assist radiologists in recognizing early signs of high-risk lung diseases. Methods: This paper proposes a novel technique, Lightweight Residual U-Net, combining the strengths of the convolutional block attention module (CBAM), the Atrous Spatial Pyramid Pooling (ASPP) block, and the attention module, which consists of only 3.24 million trainable parameters. Furthermore, the proposed model has been trained using both the RELU and LeakyReLU activation functions, with LeakyReLU yielding superior performance. The study indicates that the Dice loss function is more effective in achieving better results. Results: The proposed model is evaluated on three benchmark datasets: JSRT, SZ, and MC, achieving a Dice score of 98.72%, 97.49%, and 99.08%, respectively, outperforming the state-of-the-art models. Conclusions: Using the capabilities of DL and cutting-edge attention processes, the proposed model improves current efforts to enhance lung segmentation for the early identification of many serious lung diseases.
Collapse
Affiliation(s)
- Meftahul Jannat
- Department of Electronics and Telecommunication Engineering, Rajshahi University of Engineering and Technology, Rajshahi 6204, Bangladesh; (M.J.); (H.S.); (S.F.)
| | - Shaikh Afnan Birahim
- Department of Computer Science and Engineering, Royal University of Dhaka, Dhaka 1208, Bangladesh;
| | - Mohammad Asif Hasan
- Department of Electronics and Telecommunication Engineering, Rajshahi University of Engineering and Technology, Rajshahi 6204, Bangladesh; (M.J.); (H.S.); (S.F.)
| | - Tonmoy Roy
- Department of Data Analytics and Information Systems, Utah State University, Old Main Hill, Logan, UT 84322, USA;
| | - Lubna Sultana
- Department of Computer Science and Engineering, International Islamic University Chittagong, Chittagong 4318, Bangladesh;
| | - Hasan Sarker
- Department of Electronics and Telecommunication Engineering, Rajshahi University of Engineering and Technology, Rajshahi 6204, Bangladesh; (M.J.); (H.S.); (S.F.)
| | - Samia Fairuz
- Department of Electronics and Telecommunication Engineering, Rajshahi University of Engineering and Technology, Rajshahi 6204, Bangladesh; (M.J.); (H.S.); (S.F.)
| | - Hanaa A. Abdallah
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| |
Collapse
|
2
|
Chen Y, Chen C, Li X, Xiao R. SIAM: Spatial and Intensity Awareness Module for cerebrovascular segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 260:108511. [PMID: 39626410 DOI: 10.1016/j.cmpb.2024.108511] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Revised: 10/20/2024] [Accepted: 11/07/2024] [Indexed: 02/09/2025]
Abstract
BACKGROUND AND OBJECTIVES Cerebrovascular segmentation plays a crucial role in guiding the diagnosis and treatment of cerebrovascular diseases. With the rapid advancements in deep learning models, significant progress has been made in 3D cerebrovascular segmentation. However, they often rely on massive images and annotations, which is still challenging in cerebrovascular segmentation. METHODS Considering the unique pixel and spatial features inherent to vascular structures, such as vessel shape, location, and high pixel intensity characteristics, we propose a novel Spatial and Intensity Awareness Module (SIAM) for limited cerebrovascular segmentation. This module introduces spatial and pixel intensity perturbations to construct new matching data for model learning. Using collaborative training and shared features, SIAM gains the awareness of spatial and pixel intensity, thereby endowing the model with cerebrovascular semantics. Owing to the awareness learning belonging to an independent training module, SIAM satisfies the attribute of plug-and-play. RESULTS To validate SIAM, we carried out experiments on three cerebrovascular datasets with different modalities. The results demonstrate that SIAM enables the models to perform remarkably in normal and limited cerebrovascular segmentation. It can be seamlessly integrated into existing segmentation models without disrupting structural integrity. CONCLUSION SIAM effectively learns and adapts to the unique spatial and pixel intensity features of vascular structures through collaborative training and shared features. Our experiments on three different cerebrovascular datasets confirm its robustness and efficacy even with limited data. Furthermore, its plug-and-play nature allows for seamless integration into existing models, preserving their structural integrity. Our code is available at https://github.com/QingYunA/SIAM-Spatial-and-Intensity-Awareness-Module-for-3D-Cerebrovascular-Segmentation.
Collapse
Affiliation(s)
- Yunqing Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China.
| | - Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China.
| | - Xiaoheng Li
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China; Shunde Innovation School, University of Science and Technology Beijing, Foshan, 100024, China.
| |
Collapse
|
3
|
Chen C, Song Y, Chen D, Zhu J, Ning H, Xiao R. Design and application of pneumatic rehabilitation glove system based on brain-computer interface. THE REVIEW OF SCIENTIFIC INSTRUMENTS 2024; 95:095108. [PMID: 39248624 DOI: 10.1063/5.0225972] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/28/2024] [Accepted: 08/22/2024] [Indexed: 09/10/2024]
Abstract
Stroke has been the second leading cause of death and disability worldwide. With the innovation of therapeutic schedules, its death rate has decreased significantly but still guides chronic movement disorders. Due to the lack of independent activities and minimum exercise standards, the traditional rehabilitation means of occupational therapy and constraint-induced movement therapy pose challenges in stroke patients with severe impairments. Therefore, specific and effective rehabilitation methods seek innovation. To address the overlooked limitation, we design a pneumatic rehabilitation glove system. Specially, we developed a pneumatic glove, which utilizes ElectroEncephaloGram (EEG) acquisition to gain the EEG signals. A proposed EEGTran model is inserted into the system to distinguish the specific motor imagination behavior, thus, the glove can perform specific activities according to the patient's imagination, facilitating the patients with severe movement disorders and promoting the rehabilitation technology. The experimental results show that the proposed EEGTrans reached an accuracy of 87.3% and outperformed that of competitors. It demonstrates that our pneumatic rehabilitation glove system contributes to the rehabilitation training of stroke patients.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Yize Song
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Duoyou Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Jiahua Zhu
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Huansheng Ning
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
- Shunde Graduate School of University of Science and Technology Beijing, Foshan 100024, China
| |
Collapse
|
4
|
Shafi SM, Chinnappan SK. Segmenting and classifying lung diseases with M-Segnet and Hybrid Squeezenet-CNN architecture on CT images. PLoS One 2024; 19:e0302507. [PMID: 38753712 PMCID: PMC11098347 DOI: 10.1371/journal.pone.0302507] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2024] [Accepted: 04/07/2024] [Indexed: 05/18/2024] Open
Abstract
Diagnosing lung diseases accurately and promptly is essential for effectively managing this significant public health challenge on a global scale. This paper introduces a new framework called Modified Segnet-based Lung Disease Segmentation and Severity Classification (MSLDSSC). The MSLDSSC model comprises four phases: "preprocessing, segmentation, feature extraction, and classification." Initially, the input image undergoes preprocessing using an improved Wiener filter technique. This technique estimates the power spectral density of the noisy and original images and computes the SNR assisted by PSNR to evaluate image quality. Next, the preprocessed image undergoes Segmentation to identify and separate the RoI from the background objects in the lung image. We employ a Modified Segnet mechanism that utilizes a proposed hard tanh-Softplus activation function for effective Segmentation. Following Segmentation, features such as MLDN, entropy with MRELBP, shape features, and deep features are extracted. Following the feature extraction phase, the retrieved feature set is input into a hybrid severity classification model. This hybrid model comprises two classifiers: SDPA-Squeezenet and DCNN. These classifiers train on the retrieved feature set and effectively classify the severity level of lung diseases.
Collapse
Affiliation(s)
- Syed Mohammed Shafi
- School of Computer Science and Engineering Vellore Institute of Technology, Vellore, India
| | | |
Collapse
|
5
|
Chen C, Chen Y, Li X, Ning H, Xiao R. Linear semantic transformation for semi-supervised medical image segmentation. Comput Biol Med 2024; 173:108331. [PMID: 38522252 DOI: 10.1016/j.compbiomed.2024.108331] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2024] [Revised: 02/29/2024] [Accepted: 03/17/2024] [Indexed: 03/26/2024]
Abstract
Medical image segmentation is a focus research and foundation in developing intelligent medical systems. Recently, deep learning for medical image segmentation has become a standard process and succeeded significantly, promoting the development of reconstruction, and surgical planning of disease diagnosis. However, semantic learning is often inefficient owing to the lack of supervision of feature maps, resulting in that high-quality segmentation models always rely on numerous and accurate data annotations. Learning robust semantic representation in latent spaces remains a challenge. In this paper, we propose a novel semi-supervised learning framework to learn vital attributes in medical images, which constructs generalized representation from diverse semantics to realize medical image segmentation. We first build a self-supervised learning part that achieves context recovery by reconstructing space and intensity of medical images, which conduct semantic representation for feature maps. Subsequently, we combine semantic-rich feature maps and utilize simple linear semantic transformation to convert them into image segmentation. The proposed framework was tested using five medical segmentation datasets. Quantitative assessments indicate the highest scores of our method on IXI (73.78%), ScaF (47.50%), COVID-19-Seg (50.72%), PC-Seg (65.06%), and Brain-MR (72.63%) datasets. Finally, we compared our method with the latest semi-supervised learning methods and obtained 77.15% and 75.22% DSC values, respectively, ranking first on two representative datasets. The experimental results not only proved that the proposed linear semantic transformation was effectively applied to medical image segmentation, but also presented its simplicity and ease-of-use to pursue robust segmentation in semi-supervised learning. Our code is now open at: https://github.com/QingYunA/Linear-Semantic-Transformation-for-Semi-Supervised-Medical-Image-Segmentation.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Yunqing Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Xiaoheng Li
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Huansheng Ning
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China; Shunde Innovation School, University of Science and Technology Beijing, Foshan, 100024, China.
| |
Collapse
|
6
|
Zhang J, Wang S, Jiang Z, Chen Z, Bai X. CD-Net: Cascaded 3D Dilated convolutional neural network for pneumonia lesion segmentation. Comput Biol Med 2024; 173:108311. [PMID: 38513395 DOI: 10.1016/j.compbiomed.2024.108311] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 02/22/2024] [Accepted: 03/12/2024] [Indexed: 03/23/2024]
Abstract
COVID-19 is a global pandemic that has caused significant global, social, and economic disruption. To effectively assist in screening and monitoring diagnosed cases, it is crucial to accurately segment lesions from Computer Tomography (CT) scans. Due to the lack of labeled data and the presence of redundant parameters in 3D CT, there are still significant challenges in diagnosing COVID-19 in related fields. To address the problem, we have developed a new model called the Cascaded 3D Dilated convolutional neural network (CD-Net) for directly processing CT volume data. To reduce memory consumption when cutting volume data into small patches, we initially design a cascade architecture in CD-Net to preserve global information. Then, we construct a Multi-scale Parallel Dilated Convolution (MPDC) block to aggregate features of different sizes and simultaneously reduce the parameters. Moreover, to alleviate the shortage of labeled data, we employ classical transfer learning, which requires only a small amount of data while achieving better performance. Experimental results conducted on the different public-available datasets verify that the proposed CD-Net has reduced the negative-positive ratio and outperformed other existing segmentation methods while requiring less data.
Collapse
Affiliation(s)
- Jinli Zhang
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.
| | - Shaomeng Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.
| | - Zongli Jiang
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.
| | - Zhijie Chen
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.
| | - Xiaolu Bai
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.
| |
Collapse
|
7
|
Hossain MSA, Gul S, Chowdhury MEH, Khan MS, Sumon MSI, Bhuiyan EH, Khandakar A, Hossain M, Sadique A, Al-Hashimi I, Ayari MA, Mahmud S, Alqahtani A. Deep Learning Framework for Liver Segmentation from T1-Weighted MRI Images. SENSORS (BASEL, SWITZERLAND) 2023; 23:8890. [PMID: 37960589 PMCID: PMC10650219 DOI: 10.3390/s23218890] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/23/2023] [Revised: 08/08/2023] [Accepted: 08/15/2023] [Indexed: 11/15/2023]
Abstract
The human liver exhibits variable characteristics and anatomical information, which is often ambiguous in radiological images. Machine learning can be of great assistance in automatically segmenting the liver in radiological images, which can be further processed for computer-aided diagnosis. Magnetic resonance imaging (MRI) is preferred by clinicians for liver pathology diagnosis over volumetric abdominal computerized tomography (CT) scans, due to their superior representation of soft tissues. The convenience of Hounsfield unit (HoU) based preprocessing in CT scans is not available in MRI, making automatic segmentation challenging for MR images. This study investigates multiple state-of-the-art segmentation networks for liver segmentation from volumetric MRI images. Here, T1-weighted (in-phase) scans are investigated using expert-labeled liver masks from a public dataset of 20 patients (647 MR slices) from the Combined Healthy Abdominal Organ Segmentation grant challenge (CHAOS). The reason for using T1-weighted images is that it demonstrates brighter fat content, thus providing enhanced images for the segmentation task. Twenty-four different state-of-the-art segmentation networks with varying depths of dense, residual, and inception encoder and decoder backbones were investigated for the task. A novel cascaded network is proposed to segment axial liver slices. The proposed framework outperforms existing approaches reported in the literature for the liver segmentation task (on the same test set) with a dice similarity coefficient (DSC) score and intersect over union (IoU) of 95.15% and 92.10%, respectively.
Collapse
Affiliation(s)
- Md. Sakib Abrar Hossain
- NSU Genome Research Institute (NGRI), North South University, Dhaka 1229, Bangladesh
- Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
| | - Sidra Gul
- Department of Computer Systems Engineering, University of Engineering and Technology Peshawar, Peshawar 25000, Pakistan
- Artificial Intelligence in Healthcare, IIPL, National Center of Artificial Intelligence, Peshawar 25000, Pakistan
| | | | | | | | - Enamul Haque Bhuiyan
- Center for Magnetic Resonance Research, University of Illinois Chicago, Chicago, IL 60607, USA
| | - Amith Khandakar
- Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
| | - Maqsud Hossain
- NSU Genome Research Institute (NGRI), North South University, Dhaka 1229, Bangladesh
| | - Abdus Sadique
- NSU Genome Research Institute (NGRI), North South University, Dhaka 1229, Bangladesh
| | | | | | - Sakib Mahmud
- Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
| | - Abdulrahman Alqahtani
- Department of Medical Equipment Technology, College of Applied, Medical Science, Majmaah University, Majmaah City 11952, Saudi Arabia
- Department of Biomedical Technology, College of Applied Medical Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
| |
Collapse
|
8
|
Wu Y, Zhao S, Qi S, Feng J, Pang H, Chang R, Bai L, Li M, Xia S, Qian W, Ren H. Two-stage contextual transformer-based convolutional neural network for airway extraction from CT images. Artif Intell Med 2023; 143:102637. [PMID: 37673569 DOI: 10.1016/j.artmed.2023.102637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2023] [Revised: 06/14/2023] [Accepted: 08/11/2023] [Indexed: 09/08/2023]
Abstract
Accurate airway segmentation from computed tomography (CT) images is critical for planning navigation bronchoscopy and realizing a quantitative assessment of airway-related chronic obstructive pulmonary disease (COPD). Existing methods face difficulty in airway segmentation, particularly for the small branches of the airway. These difficulties arise due to the constraints of limited labeling and failure to meet clinical use requirements in COPD. We propose a two-stage framework with a novel 3D contextual transformer for segmenting the overall airway and small airway branches using CT images. The method consists of two training stages sharing the same modified 3D U-Net network. The novel 3D contextual transformer block is integrated into both the encoder and decoder path of the network to effectively capture contextual and long-range information. In the first training stage, the proposed network segments the overall airway with the overall airway mask. To improve the performance of the segmentation result, we generate the intrapulmonary airway branch label, and train the network to focus on producing small airway branches in the second training stage. Extensive experiments were performed on in-house and multiple public datasets. Quantitative and qualitative analyses demonstrate that our proposed method extracts significantly more branches and longer lengths of the airway tree while accomplishing state-of-the-art airway segmentation performance. The code is available at https://github.com/zhaozsq/airway_segmentation.
Collapse
Affiliation(s)
- Yanan Wu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China; Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China.
| | - Shuiqing Zhao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
| | - Shouliang Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| | - Jie Feng
- School of Chemical Equipment, Shenyang University of Technology, Liaoyang, China.
| | - Haowen Pang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
| | - Runsheng Chang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
| | - Long Bai
- Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China.
| | - Mengqi Li
- Department of Respiratory, the Second Affiliated Hospital of Dalian Medical University, Dalian, China.
| | - Shuyue Xia
- Respiratory Department, Central Hospital Affiliated to Shenyang Medical College, Shenyang, China.
| | - Wei Qian
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
| | - Hongliang Ren
- Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China.
| |
Collapse
|
9
|
Zhou GQ, Wei H, Wang X, Wang KN, Chen Y, Xiong F, Ren G, Liu C, Li L, Huang Q. BSMNet: Boundary-salience multi-branch network for intima-media identification in carotid ultrasound images. Comput Biol Med 2023; 162:107092. [PMID: 37263149 DOI: 10.1016/j.compbiomed.2023.107092] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Revised: 05/05/2023] [Accepted: 05/27/2023] [Indexed: 06/03/2023]
Abstract
Carotid artery intima-media thickness (CIMT) is an essential factor in signaling the risk of cardiovascular diseases, which is commonly evaluated using ultrasound imaging. However, automatic intima-media segmentation and thickness measurement are still challenging due to the boundary ambiguity of intima-media and inherent speckle noises in ultrasound images. In this work, we propose an end-to-end boundary-salience multi-branch network, BSMNet, to tackle the carotid intima-media identification from ultrasound images, where the prior shape knowledge and anatomical dependence are exploited using a parallel linear structure learning modules followed by a boundary refinement module. Moreover, we design a strip attention model to boost the thin strip region segmentation with shape priors, in which an anisotropic kernel shape captures long-range global relations and scrutinizes meaningful local salient contexts simultaneously. Extensive experimental results on an in-house carotid ultrasound (US) dataset demonstrate the promising performance of our method, which achieves about 0.02 improvement in Dice and HD95 than other state-of-the-art methods. Our method is promising in advancing the analysis of systemic arterial disease with ultrasound imaging.
Collapse
Affiliation(s)
- Guang-Quan Zhou
- The School of Biological Science and Medical Engineering, Southeast University, Nanjing, China; Jiangsu Key Laboratory of Biomaterials and Devices, Southeast University, Nanjing, China; State Key Laboratory of Digital Medical Engineering, Southeast University, Nanjing, China.
| | - Hao Wei
- The School of Biological Science and Medical Engineering, Southeast University, Nanjing, China
| | - Xiaoyi Wang
- Shenzhen Delica Medical Equipment Co., Ltd, Shenzhen, 518132, China.
| | - Kai-Ni Wang
- The School of Biological Science and Medical Engineering, Southeast University, Nanjing, China; Jiangsu Key Laboratory of Biomaterials and Devices, Southeast University, Nanjing, China; State Key Laboratory of Digital Medical Engineering, Southeast University, Nanjing, China
| | - Yuzhao Chen
- The School of Biological Science and Medical Engineering, Southeast University, Nanjing, China
| | - Fei Xiong
- Ethics Committee of Medical and Experimental Animals, Northwestern Polytechnical University, Xi'an, China
| | - Guanqing Ren
- Shenzhen Delica Medical Equipment Co., Ltd, Shenzhen, 518132, China
| | - Chunying Liu
- Ethics Committee of Medical and Experimental Animals, Northwestern Polytechnical University, Xi'an, China
| | - Le Li
- Institute of Medical Research, Northwestern Polytechnical University, Xi'an, China.
| | - Qinghua Huang
- School of Artificial Intelligence, OPtics and ElectroNics (iOPEN), Northwestern Polytechnical University, Xi'an, China.
| |
Collapse
|
10
|
Yang Y, Zhang L, Ren L, Zhou L, Wang X. SuperMini-seg: An ultra lightweight network for COVID-19 lung infection segmentation from CT images. Biomed Signal Process Control 2023; 85:104896. [PMID: 36998783 PMCID: PMC10028361 DOI: 10.1016/j.bspc.2023.104896] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2022] [Revised: 01/31/2023] [Accepted: 03/18/2023] [Indexed: 03/24/2023]
Abstract
The automatic segmentation of lung lesions from COVID-19 computed tomography (CT) images is helpful in establishing a quantitative model to diagnose and treat COVID-19. To this end, this study proposes a lightweight segmentation network called the SuperMini-Seg. We propose a new module called the transformer parallel convolution module (TPCB), which introduces both transformer and convolution operations in one module. SuperMini-seg adopts the structure of a double-branch parallel to downsample the image and designs a gated attention mechanism in the middle of the two parallel branches. At the same time, the attentive hierarchical spatial pyramid (AHSP) module and criss-cross attention module are adopted, and more than 100K parameters are present in the model. At the same time, the model is scalable, and the parameter quantity of SuperMini-seg-V2 reaches more than 70K. Compared with other advanced methods, the segmentation accuracy was almost reached the state-of-art method. The calculation efficiency was high, which is convenient for practical deployment.
Collapse
Affiliation(s)
- Yuan Yang
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, School of Medicine and Engineering, No. 37 Xueyuan Road, Haidian District, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Ministry of Industry and Information Technology, No. 37 Xueyuan Road, Haidian District, Beijing, China
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| | - Lin Zhang
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, School of Medicine and Engineering, No. 37 Xueyuan Road, Haidian District, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Ministry of Industry and Information Technology, No. 37 Xueyuan Road, Haidian District, Beijing, China
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| | - Lei Ren
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, School of Medicine and Engineering, No. 37 Xueyuan Road, Haidian District, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Ministry of Industry and Information Technology, No. 37 Xueyuan Road, Haidian District, Beijing, China
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| | - Longfei Zhou
- Duke University School of Medicine, 2424 Erwin Rd, Suite 302, Durham, USA
| | - Xiaohan Wang
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| |
Collapse
|
11
|
Park J, Mah AJ, Nguyen T, Park S, Ghazi Zadeh L, Shadgan B, Gandjbakhche AH. Modification of a Conventional Deep Learning Model to Classify Simulated Breathing Patterns: A Step toward Real-Time Monitoring of Patients with Respiratory Infectious Diseases. SENSORS (BASEL, SWITZERLAND) 2023; 23:5592. [PMID: 37420758 PMCID: PMC10300752 DOI: 10.3390/s23125592] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/19/2023] [Revised: 06/08/2023] [Accepted: 06/13/2023] [Indexed: 07/09/2023]
Abstract
The emergence of the global coronavirus pandemic in 2019 (COVID-19 disease) created a need for remote methods to detect and continuously monitor patients with infectious respiratory diseases. Many different devices, including thermometers, pulse oximeters, smartwatches, and rings, were proposed to monitor the symptoms of infected individuals at home. However, these consumer-grade devices are typically not capable of automated monitoring during both day and night. This study aims to develop a method to classify and monitor breathing patterns in real-time using tissue hemodynamic responses and a deep convolutional neural network (CNN)-based classification algorithm. Tissue hemodynamic responses at the sternal manubrium were collected in 21 healthy volunteers using a wearable near-infrared spectroscopy (NIRS) device during three different breathing conditions. We developed a deep CNN-based classification algorithm to classify and monitor breathing patterns in real time. The classification method was designed by improving and modifying the pre-activation residual network (Pre-ResNet) previously developed to classify two-dimensional (2D) images. Three different one-dimensional CNN (1D-CNN) classification models based on Pre-ResNet were developed. By using these models, we were able to obtain an average classification accuracy of 88.79% (without Stage 1 (data size reducing convolutional layer)), 90.58% (with 1 × 3 Stage 1), and 91.77% (with 1 × 5 Stage 1).
Collapse
Affiliation(s)
- Jinho Park
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.)
| | - Aaron James Mah
- Implantable Biosensing Laboratory, International Collaboration on Repair Discoveries, Vancouver, BC V5Z 1M9, Canada; (A.J.M.); (L.G.Z.); (B.S.)
- Department of Pathology & Laboratory Medicine, University of British Columbia, Vancouver, BC V6T 1Z7, Canada
| | - Thien Nguyen
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.)
| | - Soongho Park
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.)
| | - Leili Ghazi Zadeh
- Implantable Biosensing Laboratory, International Collaboration on Repair Discoveries, Vancouver, BC V5Z 1M9, Canada; (A.J.M.); (L.G.Z.); (B.S.)
- Department of Pathology & Laboratory Medicine, University of British Columbia, Vancouver, BC V6T 1Z7, Canada
| | - Babak Shadgan
- Implantable Biosensing Laboratory, International Collaboration on Repair Discoveries, Vancouver, BC V5Z 1M9, Canada; (A.J.M.); (L.G.Z.); (B.S.)
- Department of Pathology & Laboratory Medicine, University of British Columbia, Vancouver, BC V6T 1Z7, Canada
| | - Amir H. Gandjbakhche
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.)
| |
Collapse
|
12
|
He X, Wang Y, Poiesi F, Song W, Xu Q, Feng Z, Wan Y. Exploiting multi-granularity visual features for retinal layer segmentation in human eyes. Front Bioeng Biotechnol 2023; 11:1191803. [PMID: 37324431 PMCID: PMC10267414 DOI: 10.3389/fbioe.2023.1191803] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Accepted: 05/22/2023] [Indexed: 06/17/2023] Open
Abstract
Accurate segmentation of retinal layer boundaries can facilitate the detection of patients with early ophthalmic disease. Typical segmentation algorithms operate at low resolutions without fully exploiting multi-granularity visual features. Moreover, several related studies do not release their datasets that are key for the research on deep learning-based solutions. We propose a novel end-to-end retinal layer segmentation network based on ConvNeXt, which can retain more feature map details by using a new depth-efficient attention module and multi-scale structures. In addition, we provide a semantic segmentation dataset containing 206 retinal images of healthy human eyes (named NR206 dataset), which is easy to use as it does not require any additional transcoding processing. We experimentally show that our segmentation approach outperforms state-of-the-art approaches on this new dataset, achieving, on average, a Dice score of 91.3% and mIoU of 84.4%. Moreover, our approach achieves state-of-the-art performance on a glaucoma dataset and a diabetic macular edema (DME) dataset, showing that our model is also suitable for other applications. We will make our source code and the NR206 dataset publicly available at (https://github.com/Medical-Image-Analysis/Retinal-layer-segmentation).
Collapse
Affiliation(s)
- Xiang He
- School of Mechanical Engineering, Shandong University, Jinan, China
- Joint SDU-NTU Centre for Artificial Intelligence Research (C-FAIR), Shandong University, Jinan, China
| | | | | | - Weiye Song
- School of Mechanical Engineering, Shandong University, Jinan, China
| | - Quanqing Xu
- School of Mechanical Engineering, Shandong University, Jinan, China
| | - Zixuan Feng
- School of Mechanical Engineering, Shandong University, Jinan, China
| | - Yi Wan
- School of Mechanical Engineering, Shandong University, Jinan, China
| |
Collapse
|
13
|
Chen C, Zhou K, Wang Z, Zhang Q, Xiao R. All answers are in the images: A review of deep learning for cerebrovascular segmentation. Comput Med Imaging Graph 2023; 107:102229. [PMID: 37043879 DOI: 10.1016/j.compmedimag.2023.102229] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2022] [Revised: 03/03/2023] [Accepted: 04/03/2023] [Indexed: 04/14/2023]
Abstract
Cerebrovascular imaging is a common examination. Its accurate cerebrovascular segmentation become an important auxiliary method for the diagnosis and treatment of cerebrovascular diseases, which has received extensive attention from researchers. Deep learning is a heuristic method that encourages researchers to derive answers from the images by driving datasets. With the continuous development of datasets and deep learning theory, it has achieved important success for cerebrovascular segmentation. Detailed survey is an important reference for researchers. To comprehensively analyze the newest cerebrovascular segmentation, we have organized and discussed researches centered on deep learning. This survey comprehensively reviews deep learning for cerebrovascular segmentation since 2015, it mainly includes sliding window based models, U-Net based models, other CNNs based models, small-sample based models, semi-supervised or unsupervised models, fusion based models, Transformer based models, and graphics based models. We organize the structures, improvement, and important parameters of these models, as well as analyze development trends and quantitative assessment. Finally, we have discussed the challenges and opportunities of possible research directions, hoping that our survey can provide researchers with convenient reference.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Zhiliang Wang
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Qian Zhang
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, Beijing 100070, China; China National Clinical Research Center for Neurological Diseases, Beijing 100070, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; Shunde Innovation School, University of Science and Technology Beijing, Foshan 100024, China.
| |
Collapse
|
14
|
Chen C, Zhou K, Qi S, Lu T, Xiao R. A learnable Gabor Convolution kernel for vessel segmentation. Comput Biol Med 2023; 158:106892. [PMID: 37028143 DOI: 10.1016/j.compbiomed.2023.106892] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Revised: 03/26/2023] [Accepted: 04/01/2023] [Indexed: 04/05/2023]
Abstract
Vessel segmentation is significant for characterizing vascular diseases, receiving wide attention of researchers. The common vessel segmentation methods are mainly based on convolutional neural networks (CNNs), which have excellent feature learning capabilities. Owing to inability to predict learning direction, CNNs generate large channels or sufficient depth to obtain sufficient features. It may engender redundant parameters. Drawing on performance ability of Gabor filters in vessel enhancement, we built Gabor convolution kernel and designed its optimization. Unlike traditional filter using and common modulation, its parameters are automatically updated using gradients in the back propagation. Since the structural shape of Gabor convolution kernels is the same as that of regular convolution kernels, it can be integrated into any CNNs architecture. We built Gabor ConvNet using Gabor convolution kernels and tested it using three vessel datasets. It scored 85.06%, 70.52% and 67.11%, respectively, ranking first on three datasets. Results shows that our method outperforms advanced models in vessel segmentation. Ablations also proved that Gabor kernel has better vessel extraction ability than the regular convolution kernel.
Collapse
|
15
|
Ding W, Abdel-Basset M, Hawash H, Pedrycz W. MIC-Net: A deep network for cross-site segmentation of COVID-19 infection in the fog-assisted IoMT. Inf Sci (N Y) 2023; 623:20-39. [PMID: 36532157 PMCID: PMC9745980 DOI: 10.1016/j.ins.2022.12.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2022] [Revised: 12/02/2022] [Accepted: 12/07/2022] [Indexed: 12/15/2022]
Abstract
The automatic segmentation of COVID-19 pneumonia from a computerized tomography (CT) scan has become a major interest for scholars in developing a powerful diagnostic framework in the Internet of Medical Things (IoMT). Federated deep learning (FDL) is considered a promising approach for efficient and cooperative training from multi-institutional image data. However, the nonindependent and identically distributed (Non-IID) data from health care remain a remarkable challenge, limiting the applicability of FDL in the real world. The variability in features incurred by different scanning protocols, scanners, or acquisition parameters produces the learning drift phenomena during the training, which impairs both the training speed and segmentation performance of the model. This paper proposes a novel FDL approach for reliable and efficient multi-institutional COVID-19 segmentation, called MIC-Net. MIC-Net consists of three main building modules: the down-sampler, context enrichment (CE) module, and up-sampler. The down-sampler was designed to effectively learn both local and global representations from input CT scans by combining the advantages of lightweight convolutional and attention modules. The contextual enrichment (CE) module is introduced to enable the network to capture the contextual representation that can be later exploited to enrich the semantic knowledge of the up-sampler through skip connections. To further tackle the inter-site heterogeneity within the model, the approach uses an adaptive and switchable normalization (ASN) to adaptively choose the best normalization strategy according to the underlying data. A novel federated periodic selection protocol (FED-PCS) is proposed to fairly select the training participants according to their resource state, data quality, and loss of a local model. The results of an experimental evaluation of MIC-Net on three publicly available data sets show its robust performance, with an average dice score of 88.90% and an average surface dice of 87.53%.
Collapse
Affiliation(s)
- Weiping Ding
- School of Information Science and Technology, Nantong University, Nantong, China
- Faculty of Data Science, City University of Macau, Macau, China
| | | | | | - Witold Pedrycz
- Electrical and Computer Engineering, University of Alberta, Edmonton, AB T6R 2V4, Canada
| |
Collapse
|
16
|
Chen H, Yu MA, Chen C, Zhou K, Qi S, Chen Y, Xiao R. FDE-net: Frequency-domain enhancement network using dynamic-scale dilated convolution for thyroid nodule segmentation. Comput Biol Med 2023; 153:106514. [PMID: 36628913 DOI: 10.1016/j.compbiomed.2022.106514] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2022] [Revised: 12/22/2022] [Accepted: 12/31/2022] [Indexed: 01/05/2023]
Abstract
Thyroid nodules, a common disease of endocrine system, have a probability of nearly 10% to turn into malignant nodules and thus pose a serious threat to health. Automatic segmentation of thyroid nodules is of great importance for clinicopathological diagnosis. This work proposes FDE-Net, a combined segmental frequency domain enhancement and dynamic scale cavity convolutional network for thyroid nodule segmentation. In FDE-Net, traditional image omics method is introduced to enhance the feature image in the segmented frequency domain. Such an approach reduces the influence of noise and strengthens the detail and contour information of the image. The proposed method introduces a cascade cross-scale attention module, which addresses the insensitivity of the network to the change in target scale by fusing the features of different receptive fields and improves the ability of the network to identify multiscale target regions. It repeatedly uses the high-dimensional feature image to improve segmentation accuracy in accordance with the simple structure of thyroid nodules. In this study, 1355 ultrasound images are used for training and testing. Quantitative evaluation results showed that the Dice coefficient of FDE-Net in thyroid nodule segmentation was 83.54%, which is better than other methods. Therefore, FDE-Net can enable the accurate and rapid segmentation of thyroid nodules.
Collapse
Affiliation(s)
- Hongyu Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ming-An Yu
- Department of Interventional Medicine, China-Japan Friendship Hospital, Beijing, 100029, China
| | - Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Siyu Qi
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Yunqing Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China; Shunde Innovation School, University of Science and Technology Beijing, Foshan, 100024, China.
| |
Collapse
|
17
|
Chen C, Zhou K, Wang Z, Xiao R. Generative Consistency for Semi-Supervised Cerebrovascular Segmentation From TOF-MRA. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:346-353. [PMID: 35727774 DOI: 10.1109/tmi.2022.3184675] [Citation(s) in RCA: 22] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Cerebrovascular segmentation from Time-of-flight magnetic resonance angiography (TOF-MRA) is a critical step in computer-aided diagnosis. In recent years, deep learning models have proved its powerful feature extraction for cerebrovascular segmentation. However, they require many labeled datasets to implement effective driving, which are expensive and professional. In this paper, we propose a generative consistency for semi-supervised (GCS) model. Considering the rich information contained in the feature map, the GCS model utilizes the generation results to constrain the segmentation model. The generated data comes from labeled data, unlabeled data, and unlabeled data after perturbation, respectively. The GCS model also calculates the consistency of the perturbed data to improve the feature mining ability. Subsequently, we propose a new model as the backbone of the GSC model. It transfers TOF-MRA into graph space and establishes correlation using Transformer. We demonstrated the effectiveness of the proposed model on TOF-MRA representations, and tested the GCS model with state-of-the-art semi-supervised methods using the proposed model as backbone. The experiments prove the important role of the GCS model in cerebrovascular segmentation. Code is available at https://github.com/MontaEllis/SSL-For-Medical-Segmentation.
Collapse
|
18
|
Chen C, Qi S, Zhou K, Lu T, Ning H, Xiao R. Pairwise attention-enhanced adversarial model for automatic bone segmentation in CT images. Phys Med Biol 2023; 68. [PMID: 36634367 DOI: 10.1088/1361-6560/acb2ab] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2022] [Accepted: 01/12/2023] [Indexed: 01/14/2023]
Abstract
Objective. Bone segmentation is a critical step in screw placement navigation. Although the deep learning methods have promoted the rapid development for bone segmentation, the local bone separation is still challenging due to irregular shapes and similar representational features.Approach. In this paper, we proposed the pairwise attention-enhanced adversarial model (Pair-SegAM) for automatic bone segmentation in computed tomography images, which includes the two parts of the segmentation model and discriminator. Considering that the distributions of the predictions from the segmentation model contains complicated semantics, we improve the discriminator to strengthen the awareness ability of the target region, improving the parsing of semantic information features. The Pair-SegAM has a pairwise structure, which uses two calculation mechanics to set up pairwise attention maps, then we utilize the semantic fusion to filter unstable regions. Therefore, the improved discriminator provides more refinement information to capture the bone outline, thus effectively enhancing the segmentation models for bone segmentation.Main results. To test the Pair-SegAM, we selected the two bone datasets for assessment. We evaluated our method against several bone segmentation models and latest adversarial models on the both datasets. The experimental results prove that our method not only exhibits superior bone segmentation performance, but also states effective generalization.Significance. Our method provides a more efficient segmentation of specific bones and has the potential to be extended to other semantic segmentation domains.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Siyu Qi
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Tong Lu
- Visual 3D Medical Science and Technology Development Co. Ltd, Beijing 100082, People's Republic of China
| | - Huansheng Ning
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China.,Shunde Innovation School, University of Science and Technology Beijing, Foshan 100024, People's Republic of China
| |
Collapse
|
19
|
Wang X, Yang B, Pan X, Liu F, Zhang S. BPCN: bilateral progressive compensation network for lung infection image segmentation. Phys Med Biol 2023; 68. [PMID: 36580682 DOI: 10.1088/1361-6560/acaf21] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2022] [Accepted: 12/29/2022] [Indexed: 12/31/2022]
Abstract
Lung infection image segmentation is a key technology for autonomous understanding of the potential illness. However, current approaches usually lose the low-level details, which leads to a considerable accuracy decrease for lung infection areas with varied shapes and sizes. In this paper, we propose bilateral progressive compensation network (BPCN), a bilateral progressive compensation network to improve the accuracy of lung lesion segmentation through complementary learning of spatial and semantic features. The proposed BPCN are mainly composed of two deep branches. One branch is the multi-scale progressive fusion for main region features. The other branch is a flow-field based adaptive body-edge aggregation operations to explicitly learn detail features of lung infection areas which is supplement to region features. In addition, we propose a bilateral spatial-channel down-sampling to generate a hierarchical complementary feature which avoids losing discriminative features caused by pooling operations. Experimental results show that our proposed network outperforms state-of-the-art segmentation methods in lung infection segmentation on two public image datasets with or without a pseudo-label training strategy.
Collapse
Affiliation(s)
- Xiaoyan Wang
- Zhejiang University of Technology, Zhejiang Province, People's Republic of China
| | - Baoqi Yang
- Zhejiang University of Technology, Zhejiang Province, People's Republic of China
| | - Xiang Pan
- Zhejiang University of Technology, Zhejiang Province, People's Republic of China
| | - Fuchang Liu
- Hangzhou Normal University, Zhejiang Province, People's Republic of China
| | - Sanyuan Zhang
- Zhejiang University, Zhejiang Province, People's Republic of China
| |
Collapse
|
20
|
Hasan MM, Islam MU, Sadeq MJ, Fung WK, Uddin J. Review on the Evaluation and Development of Artificial Intelligence for COVID-19 Containment. SENSORS (BASEL, SWITZERLAND) 2023; 23:527. [PMID: 36617124 PMCID: PMC9824505 DOI: 10.3390/s23010527] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/28/2022] [Revised: 12/23/2022] [Accepted: 12/29/2022] [Indexed: 06/17/2023]
Abstract
Artificial intelligence has significantly enhanced the research paradigm and spectrum with a substantiated promise of continuous applicability in the real world domain. Artificial intelligence, the driving force of the current technological revolution, has been used in many frontiers, including education, security, gaming, finance, robotics, autonomous systems, entertainment, and most importantly the healthcare sector. With the rise of the COVID-19 pandemic, several prediction and detection methods using artificial intelligence have been employed to understand, forecast, handle, and curtail the ensuing threats. In this study, the most recent related publications, methodologies and medical reports were investigated with the purpose of studying artificial intelligence's role in the pandemic. This study presents a comprehensive review of artificial intelligence with specific attention to machine learning, deep learning, image processing, object detection, image segmentation, and few-shot learning studies that were utilized in several tasks related to COVID-19. In particular, genetic analysis, medical image analysis, clinical data analysis, sound analysis, biomedical data classification, socio-demographic data analysis, anomaly detection, health monitoring, personal protective equipment (PPE) observation, social control, and COVID-19 patients' mortality risk approaches were used in this study to forecast the threatening factors of COVID-19. This study demonstrates that artificial-intelligence-based algorithms integrated into Internet of Things wearable devices were quite effective and efficient in COVID-19 detection and forecasting insights which were actionable through wide usage. The results produced by the study prove that artificial intelligence is a promising arena of research that can be applied for disease prognosis, disease forecasting, drug discovery, and to the development of the healthcare sector on a global scale. We prove that artificial intelligence indeed played a significantly important role in helping to fight against COVID-19, and the insightful knowledge provided here could be extremely beneficial for practitioners and research experts in the healthcare domain to implement the artificial-intelligence-based systems in curbing the next pandemic or healthcare disaster.
Collapse
Affiliation(s)
- Md. Mahadi Hasan
- Department of Computer Science and Engineering, Asian University of Bangladesh, Ashulia 1349, Bangladesh
| | - Muhammad Usama Islam
- School of Computing and Informatics, University of Louisiana at Lafayette, Lafayette, LA 70504, USA
| | - Muhammad Jafar Sadeq
- Department of Computer Science and Engineering, Asian University of Bangladesh, Ashulia 1349, Bangladesh
| | - Wai-Keung Fung
- Department of Applied Computing and Engineering, Cardiff School of Technologies, Cardiff Metropolitan University, Cardiff CF5 2YB, UK
| | - Jasim Uddin
- Department of Applied Computing and Engineering, Cardiff School of Technologies, Cardiff Metropolitan University, Cardiff CF5 2YB, UK
| |
Collapse
|
21
|
Wang L. Deep Learning Techniques to Diagnose Lung Cancer. Cancers (Basel) 2022; 14:5569. [PMID: 36428662 PMCID: PMC9688236 DOI: 10.3390/cancers14225569] [Citation(s) in RCA: 34] [Impact Index Per Article: 11.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Revised: 11/11/2022] [Accepted: 11/11/2022] [Indexed: 11/15/2022] Open
Abstract
Medical imaging tools are essential in early-stage lung cancer diagnostics and the monitoring of lung cancer during treatment. Various medical imaging modalities, such as chest X-ray, magnetic resonance imaging, positron emission tomography, computed tomography, and molecular imaging techniques, have been extensively studied for lung cancer detection. These techniques have some limitations, including not classifying cancer images automatically, which is unsuitable for patients with other pathologies. It is urgently necessary to develop a sensitive and accurate approach to the early diagnosis of lung cancer. Deep learning is one of the fastest-growing topics in medical imaging, with rapidly emerging applications spanning medical image-based and textural data modalities. With the help of deep learning-based medical imaging tools, clinicians can detect and classify lung nodules more accurately and quickly. This paper presents the recent development of deep learning-based imaging techniques for early lung cancer detection.
Collapse
Affiliation(s)
- Lulu Wang
- Biomedical Device Innovation Center, Shenzhen Technology University, Shenzhen 518118, China
| |
Collapse
|
22
|
Sadik F, Dastider AG, Subah MR, Mahmud T, Fattah SA. A dual-stage deep convolutional neural network for automatic diagnosis of COVID-19 and pneumonia from chest CT images. Comput Biol Med 2022; 149:105806. [PMID: 35994932 PMCID: PMC9295386 DOI: 10.1016/j.compbiomed.2022.105806] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2022] [Revised: 06/05/2022] [Accepted: 06/26/2022] [Indexed: 11/15/2022]
Abstract
In the Coronavirus disease-2019 (COVID-19) pandemic, for fast and accurate diagnosis of a large number of patients, besides traditional methods, automated diagnostic tools are now extremely required. In this paper, a deep convolutional neural network (CNN) based scheme is proposed for automated accurate diagnosis of COVID-19 from lung computed tomography (CT) scan images. First, for the automated segmentation of lung regions in a chest CT scan, a modified CNN architecture, namely SKICU-Net is proposed by incorporating additional skip interconnections in the U-Net model that overcome the loss of information in dimension scaling. Next, an agglomerative hierarchical clustering is deployed to eliminate the CT slices without significant information. Finally, for effective feature extraction and diagnosis of COVID-19 and pneumonia from the segmented lung slices, a modified DenseNet architecture, namely P-DenseCOVNet is designed where parallel convolutional paths are introduced on top of the conventional DenseNet model for getting better performance through overcoming the loss of positional arguments. Outstanding performances have been achieved with an F1 score of 0.97 in the segmentation task along with an accuracy of 87.5% in diagnosing COVID-19, common pneumonia, and normal cases. Significant experimental results and comparison with other studies show that the proposed scheme provides very satisfactory performances and can serve as an effective diagnostic tool in the current pandemic.
Collapse
Affiliation(s)
- Farhan Sadik
- Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka 1205, Bangladesh
| | - Ankan Ghosh Dastider
- Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka 1205, Bangladesh
| | - Mohseu Rashid Subah
- Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka 1205, Bangladesh
| | - Tanvir Mahmud
- Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka 1205, Bangladesh
| | - Shaikh Anowarul Fattah
- Department of Electrical and Electronic Engineering, Bangladesh University of Engineering and Technology, Dhaka 1205, Bangladesh.
| |
Collapse
|
23
|
Abstract
Background: Ultrasonography is the main examination method for breast diseases. Ultrasound imaging is currently relied upon by doctors to form statements of characteristics and locations of lesions, which severely limits the completeness and effectiveness of ultrasound image information. Moreover, analyzing ultrasonography requires experienced ultrasound doctors, which are not common in hospitals. Thus, this work proposes a 3D-based breast ultrasound system, which can automatically diagnose ultrasound images of the breasts and generate a representative 3D breast lesion model through typical ultrasonography. Methods: In this system, we use a weighted ensemble method to combine three different neural networks and explore different combinations of the neural networks. On this basis, a breast locator was designed to measure and transform the spatial position of lesions. The breast ultrasound software generates a 3D visualization report through the selection and geometric transformation of the nodular model. Results: The ensemble neural network improved in all metrics compared with the classical neural network (DenseNet, AlexNet, GoogLeNet, etc.). It proved that the ensemble neural network proposed in this work can be used for intelligent diagnosis of breast ultrasound images. For 3D visualization, magnetic resonance imaging (MRI) scans were performed to achieve their 3D reconstructions. By comparing two types of visualized results (MRI and our 3D model), we determined that models generated by the 3D-based breast ultrasound system have similar nodule characteristics and spatial relationships with the MRI. Conclusions: In summary, this system implements automatic diagnosis of ultrasound images and presents lesions through 3D models, which can obtain complete and accurate ultrasound image information. Thus, it has clinical potential.
Collapse
|
24
|
Chen C, Zhou K, Guo X, Wang Z, Xiao R, Wang G. Cerebrovascular segmentation in phase-contrast magnetic resonance angiography by multi-feature fusion and vessel completion. Comput Med Imaging Graph 2022; 98:102070. [DOI: 10.1016/j.compmedimag.2022.102070] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2021] [Revised: 04/12/2022] [Accepted: 04/18/2022] [Indexed: 10/18/2022]
|
25
|
Liu W, Luo J, Yang Y, Wang W, Deng J, Yu L. Automatic lung segmentation in chest X-ray images using improved U-Net. Sci Rep 2022; 12:8649. [PMID: 35606509 PMCID: PMC9127108 DOI: 10.1038/s41598-022-12743-y] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2021] [Accepted: 05/16/2022] [Indexed: 11/08/2022] Open
Abstract
The automatic segmentation of the lung region for chest X-ray (CXR) can help doctors diagnose many lung diseases. However, extreme lung shape changes and fuzzy lung regions caused by serious lung diseases may incorrectly make the automatic lung segmentation model. We improved the U-Net network by using the pre-training Efficientnet-b4 as the encoder and the Residual block and the LeakyReLU activation function in the decoder. The network can extract Lung field features efficiently and avoid the gradient instability caused by the multiplication effect in gradient backpropagation. Compared with the traditional U-Net model, our method improves about 2.5% dice coefficient and 6% Jaccard Index for the two benchmark lung segmentation datasets. Our model improves about 5% dice coefficient and 9% Jaccard Index for the private lung segmentation datasets compared with the traditional U-Net model. Comparative experiments show that our method can improve the accuracy of lung segmentation of CXR images and it has a lower standard deviation and good robustness.
Collapse
Affiliation(s)
- Wufeng Liu
- Henan University of Technology, Zhengzhou, 450001, China.
| | - Jiaxin Luo
- Henan University of Technology, Zhengzhou, 450001, China
| | - Yan Yang
- Henan University of Technology, Zhengzhou, 450001, China
| | - Wenlian Wang
- Nanyang Central Hospital, Nanyang, 473009, China
| | - Junkui Deng
- Nanyang Central Hospital, Nanyang, 473009, China
| | - Liang Yu
- Henan University of Technology, Zhengzhou, 450001, China
| |
Collapse
|
26
|
Rahman AU, Alqahtani A, Aldhafferi N, Nasir MU, Khan MF, Khan MA, Mosavi A. Histopathologic Oral Cancer Prediction Using Oral Squamous Cell Carcinoma Biopsy Empowered with Transfer Learning. SENSORS (BASEL, SWITZERLAND) 2022; 22:3833. [PMID: 35632242 PMCID: PMC9146317 DOI: 10.3390/s22103833] [Citation(s) in RCA: 23] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/03/2022] [Revised: 05/06/2022] [Accepted: 05/17/2022] [Indexed: 02/06/2023]
Abstract
Oral cancer is a dangerous and extensive cancer with a high death ratio. Oral cancer is the most usual cancer in the world, with more than 300,335 deaths every year. The cancerous tumor appears in the neck, oral glands, face, and mouth. To overcome this dangerous cancer, there are many ways to detect like a biopsy, in which small chunks of tissues are taken from the mouth and tested under a secure and hygienic microscope. However, microscope results of tissues to detect oral cancer are not up to the mark, a microscope cannot easily identify the cancerous cells and normal cells. Detection of cancerous cells using microscopic biopsy images helps in allaying and predicting the issues and gives better results if biologically approaches apply accurately for the prediction of cancerous cells, but during the physical examinations microscopic biopsy images for cancer detection there are major chances for human error and mistake. So, with the development of technology deep learning algorithms plays a major role in medical image diagnosing. Deep learning algorithms are efficiently developed to predict breast cancer, oral cancer, lung cancer, or any other type of medical image. In this study, the proposed model of transfer learning model using AlexNet in the convolutional neural network to extract rank features from oral squamous cell carcinoma (OSCC) biopsy images to train the model. Simulation results have shown that the proposed model achieved higher classification accuracy 97.66% and 90.06% of training and testing, respectively.
Collapse
Affiliation(s)
- Atta-ur Rahman
- Department of Computer Science (CS), College of Computer Science and Information Technology (CCSIT), Imam Abdulrahman Bin Faisal University (IAU), P.O. Box 1982, Dammam 31441, Saudi Arabia;
| | - Abdullah Alqahtani
- Department of Computer Information Systems (CIS), College of Computer Science and Information Technology (CCSIT), Imam Abdulrahman Bin Faisal University (IAU), P.O. Box 1982, Dammam 31441, Saudi Arabia; (A.A.); (N.A.)
| | - Nahier Aldhafferi
- Department of Computer Information Systems (CIS), College of Computer Science and Information Technology (CCSIT), Imam Abdulrahman Bin Faisal University (IAU), P.O. Box 1982, Dammam 31441, Saudi Arabia; (A.A.); (N.A.)
| | - Muhammad Umar Nasir
- Faculty of Computing, Riphah School of Computing and Innovation, Riphah International University, Lahore Campus, Lahore 54000, Pakistan;
| | - Muhammad Farhan Khan
- Department of Forensic Sciences, University of Health Sciences, Lahore 54000, Pakistan;
| | | | - Amir Mosavi
- John von Neumann Faculty of Informatics, Obuda University, 1034 Budapest, Hungary;
- Institute of Information Engineering, Automation and Mathematics, Slovak University of Technology in Bratislava, 81107 Bratislava, Slovakia
- Faculty of Civil Engineering, TU-Dresden, 01062 Dresden, Germany
| |
Collapse
|
27
|
Kumar S, Chandra Sekhar Redd L, George Joseph S, Kumar Sharma V, H S. Deep learning based model for classification of COVID -19 images for healthcare research progress. MATERIALS TODAY. PROCEEDINGS 2022; 62:5008-5012. [PMID: 35602305 PMCID: PMC9113957 DOI: 10.1016/j.matpr.2022.04.884] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
As imaging technology plays an important role in the diagnosis and evaluation of the new coronavirus pneumonia (COVID-19), COVID-19 related data sets have been published one after another, but there are relatively few data sets and research progress in related literature. To this end, through COVID-19-related journal papers, reports, and related open-source data set websites, organize and analyze the new coronary pneumonia data set and the deep learning models involved, including computed tomography (CT) image data sets and X-ray (CXR) Image dataset. Analyze the characteristics of the medical images presented in these data sets; focus on open-source data sets, as well as classification and segmentation models that perform well on related data sets. Finally, the future development trend of lung imaging technology is discussed.
Collapse
Affiliation(s)
- Saroj Kumar
- Sr. Data Scientist, DCG Data-Core Systems India Pvt Ltd, Kolkata, India
| | | | - Susheel George Joseph
- Department of Computer Application,Kristu Jyoti College of Management and Technology, Changanasery, India
| | - Vinay Kumar Sharma
- CSE, AIT, Chandigarh University, NH-95 Chandigarh-Ludhiana Highway, Mohali, Punjab, India
| | - Sabireen H
- Vellore Institute of Technology, Chennai, India
| |
Collapse
|
28
|
A Saliency Prediction Model Based on Re-Parameterization and Channel Attention Mechanism. ELECTRONICS 2022. [DOI: 10.3390/electronics11081180] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Deep saliency models can effectively imitate the attention mechanism of human vision, and they perform considerably better than classical models that rely on handcrafted features. However, deep models also require higher-level information, such as context or emotional content, to further approach human performance. Therefore, this study proposes a multilevel saliency prediction network that aims to use a combination of spatial and channel information to find possible high-level features, further improving the performance of a saliency model. Firstly, we use a VGG style network with an identity block as the primary network architecture. With the help of re-parameterization, we can obtain rich features similar to multiscale networks and effectively reduce computational cost. Secondly, a subnetwork with a channel attention mechanism is designed to find potential saliency regions and possible high-level semantic information in an image. Finally, image spatial features and a channel enhancement vector are combined after quantization to improve the overall performance of the model. Compared with classical models and other deep models, our model exhibits superior overall performance.
Collapse
|
29
|
Chen C, Zhou K, Wang H, Lu Y, Wang Z, Xiao R, Lu T. TMSF-Net: Multi-series fusion network with treeconnect for colorectal tumor segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 215:106613. [PMID: 34998166 DOI: 10.1016/j.cmpb.2021.106613] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/08/2021] [Revised: 11/29/2021] [Accepted: 12/29/2021] [Indexed: 06/14/2023]
Abstract
PURPOSE Colorectal tumors are common clinical diseases. Automatic segmentation of colorectal tumors captured in computed tomography (CT) images can provide numerous possibilities for computer-assisted treatment. Obtaining large datasets is expensive, and completing labeling is time- and manpower-consuming. To solve the challenge using a limited pathological dataset, this paper proposes a multi-series fusion network with treeconnect (TMSF-Net), which can automatically achieve colorectal tumor segmentation using CT images. METHODS To drive the TMSF-Net, three-series enhanced CT images were collected from all patients to improve the data characteristics. In the TMSF-Net, the coding path was designed as a three-branch structure to realize the feature extraction of the different series. Subsequently, the three branches were merged to start the feature analysis in the decoding path. To achieve the objective of feature fusion, different layers in the decoding path fused feature maps from the upper layer in the encoding path to achieve a cross-scale fusion. In addition, to reduce the problem of parameter redundancy, this study adopted a three-dimensional treeconnect to complete data connection on three branches. RESULTS A total of 22 cases were conducted by ablation and comparative experiments to test the TMSF-Net. The results showed that the TMSF-Net can improve the network performance by multiseries fusion, and its expressiveness is better than many classic networks. CONCLUSION The TMSF-Net is a many-to-one structure network, which can enhance the network learning ability and improve the analysis of potential features. Therefore, it yields good results in colorectal tumor segmentation. It can provide a new direction for neural network models based on feature fusion.
Collapse
Affiliation(s)
- Cheng Chen
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Kangneng Zhou
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Huilin Wang
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - YuanYuan Lu
- The Department of Ultrasound, Chinese PLA General Hospital, Beijing, 100853, China
| | - Zhiliang Wang
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Ruoxiu Xiao
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; Beijing Engineering and Technology Center for Convergence Networks and Ubiquitous Services, School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China.
| | - Tingting Lu
- The Department of Anorectal Surgery, The First Hospital of China Medical University, Shenyang 110122, China.
| |
Collapse
|
30
|
Hassan H, Ren Z, Zhao H, Huang S, Li D, Xiang S, Kang Y, Chen S, Huang B. Review and classification of AI-enabled COVID-19 CT imaging models based on computer vision tasks. Comput Biol Med 2022; 141:105123. [PMID: 34953356 PMCID: PMC8684223 DOI: 10.1016/j.compbiomed.2021.105123] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Revised: 12/03/2021] [Accepted: 12/03/2021] [Indexed: 01/12/2023]
Abstract
This article presents a systematic overview of artificial intelligence (AI) and computer vision strategies for diagnosing the coronavirus disease of 2019 (COVID-19) using computerized tomography (CT) medical images. We analyzed the previous review works and found that all of them ignored classifying and categorizing COVID-19 literature based on computer vision tasks, such as classification, segmentation, and detection. Most of the COVID-19 CT diagnosis methods comprehensively use segmentation and classification tasks. Moreover, most of the review articles are diverse and cover CT as well as X-ray images. Therefore, we focused on the COVID-19 diagnostic methods based on CT images. Well-known search engines and databases such as Google, Google Scholar, Kaggle, Baidu, IEEE Xplore, Web of Science, PubMed, ScienceDirect, and Scopus were utilized to collect relevant studies. After deep analysis, we collected 114 studies and reported highly enriched information for each selected research. According to our analysis, AI and computer vision have substantial potential for rapid COVID-19 diagnosis as they could significantly assist in automating the diagnosis process. Accurate and efficient models will have real-time clinical implications, though further research is still required. Categorization of literature based on computer vision tasks could be helpful for future research; therefore, this review article will provide a good foundation for conducting such research.
Collapse
Affiliation(s)
- Haseeb Hassan
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China; Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Shenzhen University Health Science Center, Shenzhen, China
| | - Zhaoyu Ren
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Huishi Zhao
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Shoujin Huang
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Dan Li
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Shaohua Xiang
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Yan Kang
- Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Shenzhen University Health Science Center, Shenzhen, China; Medical Device Innovation Research Center, Shenzhen Technology University, Shenzhen, China
| | - Sifan Chen
- Guangdong Provincial Key Laboratory of Malignant Tumor Epigenetics and Gene Regulation, Guangdong-Hong Kong Joint Laboratory for RNA Medicine, Sun Yat-Sen Memorial Hospital, Sun Yat-Sen University, Guangzhou, China; Medical Research Center, Sun Yat-Sen Memorial Hospital, Sun Yat-Sen University, Guangzhou, China
| | - Bingding Huang
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China.
| |
Collapse
|
31
|
Efficient tumor volume measurement and segmentation approach for CT image based on twin support vector machines. Neural Comput Appl 2022. [DOI: 10.1007/s00521-021-06769-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
32
|
Jadhav S, Deng G, Zawin M, Kaufman AE. COVID-view: Diagnosis of COVID-19 using Chest CT. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2022; 28:227-237. [PMID: 34587075 PMCID: PMC8981756 DOI: 10.1109/tvcg.2021.3114851] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2021] [Revised: 06/13/2021] [Accepted: 08/08/2021] [Indexed: 05/02/2023]
Abstract
Significant work has been done towards deep learning (DL) models for automatic lung and lesion segmentation and classification of COVID-19 on chest CT data. However, comprehensive visualization systems focused on supporting the dual visual+DL diagnosis of COVID-19 are non-existent. We present COVID-view, a visualization application specially tailored for radiologists to diagnose COVID-19 from chest CT data. The system incorporates a complete pipeline of automatic lungs segmentation, localization/isolation of lung abnormalities, followed by visualization, visual and DL analysis, and measurement/quantification tools. Our system combines the traditional 2D workflow of radiologists with newer 2D and 3D visualization techniques with DL support for a more comprehensive diagnosis. COVID-view incorporates a novel DL model for classifying the patients into positive/negative COVID-19 cases, which acts as a reading aid for the radiologist using COVID-view and provides the attention heatmap as an explainable DL for the model output. We designed and evaluated COVID-view through suggestions, close feedback and conducting case studies of real-world patient data by expert radiologists who have substantial experience diagnosing chest CT scans for COVID-19, pulmonary embolism, and other forms of lung infections. We present requirements and task analysis for the diagnosis of COVID-19 that motivate our design choices and results in a practical system which is capable of handling real-world patient cases.
Collapse
Affiliation(s)
| | - Gaofeng Deng
- Department of Computer ScienceStony Brook UniversityUSA
| | - Marlene Zawin
- Department of RadiologyStony Brook University HospitalUSA
| | | |
Collapse
|
33
|
Chen C, Zhou J, Zhou K, Wang Z, Xiao R. DW-UNet: Loss Balance under Local-Patch for 3D Infection Segmentation from COVID-19 CT Images. Diagnostics (Basel) 2021; 11:1942. [PMID: 34829289 PMCID: PMC8623821 DOI: 10.3390/diagnostics11111942] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2021] [Revised: 10/17/2021] [Accepted: 10/18/2021] [Indexed: 12/23/2022] Open
Abstract
(1) Background: COVID-19 has been global epidemic. This work aims to extract 3D infection from COVID-19 CT images; (2) Methods: Firstly, COVID-19 CT images are processed with lung region extraction and data enhancement. In this strategy, gradient changes of voxels in different directions respond to geometric characteristics. Due to the complexity of tubular tissues in lung region, they are clustered to the lung parenchyma center based on their filtered possibility. Thus, infection is improved after data enhancement. Then, deep weighted UNet is established to refining 3D infection texture, and weighted loss function is introduced. It changes cost calculation of different samples, causing target samples to dominate convergence direction. Finally, the trained network effectively extracts 3D infection from CT images by adjusting driving strategy of different samples. (3) Results: Using Accuracy, Precision, Recall and Coincidence rate, 20 subjects from a private dataset and eight subjects from Kaggle Competition COVID-19 CT dataset tested this method in hold-out validation framework. This work achieved good performance both in the private dataset (99.94-00.02%, 60.42-11.25%, 70.79-09.35% and 63.15-08.35%) and public dataset (99.73-00.12%, 77.02-06.06%, 41.23-08.61% and 52.50-08.18%). We also applied some extra indicators to test data augmentation and different models. The statistical tests have verified the significant difference of different models. (4) Conclusions: This study provides a COVID-19 infection segmentation technology, which provides an important prerequisite for the quantitative analysis of COVID-19 CT images.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (C.C.); (K.Z.); (Z.W.)
| | - Jiancang Zhou
- Department of Critical Care Medicine, Sir Run Run Shaw Hospital, Zhejiang University School of Medicine, Hangzhou 310016, China;
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (C.C.); (K.Z.); (Z.W.)
| | - Zhiliang Wang
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (C.C.); (K.Z.); (Z.W.)
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (C.C.); (K.Z.); (Z.W.)
| |
Collapse
|
34
|
Chen C, Liu B, Zhou K, He W, Yan F, Wang Z, Xiao R. CSR-Net: Cross-Scale Residual Network for multi-objective scaphoid fracture segmentation. Comput Biol Med 2021; 137:104776. [PMID: 34461504 DOI: 10.1016/j.compbiomed.2021.104776] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2021] [Revised: 08/12/2021] [Accepted: 08/17/2021] [Indexed: 11/16/2022]
Abstract
The scaphoid is located in the carpals. Owing to the body structure and location of the scaphoid, scaphoid fractures are common and it is difficult to heal. Three-dimensional reconstruction of scaphoid fracture can accurately display the fracture surface and provide important support for the surgical plan involving screw placement. To achieve this goal, in this study, the cross-scale residual network (CSR-Net) is proposed for scaphoid fracture segmentation. In the CSR-Net, the features of different layers are used to achieve fusion through cross-scale residual connection, which realizes scale and channel conversions between the features of different layers. It can establish close connections between different scale features. The structures of the output layer and channel are designed to establish the CSR-Net as a multi-objective architecture, which can realize scaphoid fracture and hand bone segmentations synchronously. In this study, 65 computed tomography images of scaphoid fracture are tested. Quantitative metrics are used for assessment, and the results obtained show that the CSR-Net achieves higher performance in hand bone and scaphoid fracture segmentations. In the visually detailed display, the fracture surface is clearer and more intuitive than those obtained from other methods. Therefore, the CSR-Net can achieve accurate and rapid scaphoid fracture segmentation. Its multi-objective design provides not only an accurate digital model, but also a prerequisite for navigation in the hand bone.
Collapse
Affiliation(s)
- Cheng Chen
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Bo Liu
- The Department of Hand Surgery, Beijing Jishuitan Hospital, Beijing, 100035, China
| | - Kangneng Zhou
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Wanzhang He
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Fei Yan
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Zhiliang Wang
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China
| | - Ruoxiu Xiao
- The School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, China; The Institute of Artificial Intelligence, University of Science and Technology Beijing, Beijing, 100083, China.
| |
Collapse
|
35
|
Dulf EH, Bledea M, Mocan T, Mocan L. Automatic Detection of Colorectal Polyps Using Transfer Learning. SENSORS (BASEL, SWITZERLAND) 2021; 21:5704. [PMID: 34502594 PMCID: PMC8433882 DOI: 10.3390/s21175704] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/26/2021] [Revised: 08/18/2021] [Accepted: 08/19/2021] [Indexed: 12/25/2022]
Abstract
Colorectal cancer is the second leading cause of cancer death and ranks third worldwide in diagnosed malignant pathologies (1.36 million new cases annually). An increase in the diversity of treatment options as well as a rising population require novel diagnostic tools. Current diagnostics involve critical human thinking, but the decisional process loses accuracy due to the increased number of modulatory factors involved. The proposed computer-aided diagnosis system analyses each colonoscopy and provides predictions that will help the clinician to make the right decisions. Artificial intelligence is included in the system both offline and online image processing tools. Aiming to improve the diagnostic process of colon cancer patients, an application was built that allows the easiest and most intuitive interaction between medical staff and the proposed diagnosis system. The developed tool uses two networks. The first, a convolutional neural network, is capable of classifying eight classes of tissue with a sensitivity of 98.13% and an F1 score of 98.14%, while the second network, based on semantic segmentation, can identify the malignant areas with a Jaccard index of 75.18%. The results could have a direct impact on personalised medicine combining clinical knowledge with the computing power of intelligent algorithms.
Collapse
Affiliation(s)
- Eva-H. Dulf
- Department of Automation, Faculty of Automation and Computer Science, Technical University of Cluj-Napoca, Memorandumului Str. 28, 400014 Cluj-Napoca, Romania;
| | - Marius Bledea
- Department of Automation, Faculty of Automation and Computer Science, Technical University of Cluj-Napoca, Memorandumului Str. 28, 400014 Cluj-Napoca, Romania;
| | - Teodora Mocan
- Department of Physiology, Iuliu Hatieganu University of Medicine and Pharmacy, 400349 Cluj-Napoca, Romania;
- Nanomedicine Department, Regional Institute of Gatroenterology and Hepatology, 400000 Cluj-Napoca, Romania
| | - Lucian Mocan
- Department of Surgery, 3-rd Surgery Clinic, Iuliu Hatieganu University of Medicine and Pharmacy, 400349 Cluj-Napoca, Romania;
| |
Collapse
|