51
|
Bougourzi F, Distante C, Dornaika F, Taleb-Ahmed A, Hadid A, Chaudhary S, Yang W, Qiang Y, Anwar T, Breaban ME, Hsu CC, Tai SC, Chen SN, Tricarico D, Chaudhry HAH, Fiandrotti A, Grangetto M, Spatafora MAN, Ortis A, Battiato S. COVID-19 Infection Percentage Estimation from Computed Tomography Scans: Results and Insights from the International Per-COVID-19 Challenge. SENSORS (BASEL, SWITZERLAND) 2024; 24:1557. [PMID: 38475092 PMCID: PMC10934842 DOI: 10.3390/s24051557] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/10/2023] [Revised: 11/29/2023] [Accepted: 02/26/2024] [Indexed: 03/14/2024]
Abstract
COVID-19 analysis from medical imaging is an important task that has been intensively studied in the last years due to the spread of the COVID-19 pandemic. In fact, medical imaging has often been used as a complementary or main tool to recognize the infected persons. On the other hand, medical imaging has the ability to provide more details about COVID-19 infection, including its severity and spread, which makes it possible to evaluate the infection and follow-up the patient's state. CT scans are the most informative tool for COVID-19 infection, where the evaluation of COVID-19 infection is usually performed through infection segmentation. However, segmentation is a tedious task that requires much effort and time from expert radiologists. To deal with this limitation, an efficient framework for estimating COVID-19 infection as a regression task is proposed. The goal of the Per-COVID-19 challenge is to test the efficiency of modern deep learning methods on COVID-19 infection percentage estimation (CIPE) from CT scans. Participants had to develop an efficient deep learning approach that can learn from noisy data. In addition, participants had to cope with many challenges, including those related to COVID-19 infection complexity and crossdataset scenarios. This paper provides an overview of the COVID-19 infection percentage estimation challenge (Per-COVID-19) held at MIA-COVID-2022. Details of the competition data, challenges, and evaluation metrics are presented. The best performing approaches and their results are described and discussed.
Collapse
Affiliation(s)
- Fares Bougourzi
- Institute of Applied Sciences and Intelligent Systems, National Research Council of Italy, 73100 Lecce, Italy;
- Laboratoire LISSI, University Paris-Est Creteil, Vitry sur Seine, 94400 Paris, France
| | - Cosimo Distante
- Institute of Applied Sciences and Intelligent Systems, National Research Council of Italy, 73100 Lecce, Italy;
| | - Fadi Dornaika
- Department of Computer Science and Artificial Intelligence, University of the Basque Country UPV/EHU, Manuel Lardizabal, 1, 20018 San Sebastian, Spain;
- IKERBASQUE, Basque Foundation for Science, 48011 Bilbao, Spain
| | - Abdelmalik Taleb-Ahmed
- Institut d’Electronique de Microélectronique et de Nanotechnologie (IEMN), UMR 8520, Universite Polytechnique Hauts-de-France, Université de Lille, CNRS, 59313 Valenciennes, France;
| | - Abdenour Hadid
- Sorbonne Center for Artificial Intelligence, Sorbonne University of Abu Dhabi, Abu Dhabi P.O. Box 38044, United Arab Emirates
| | - Suman Chaudhary
- College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan 030024, China; (S.C.)
| | - Wanting Yang
- College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan 030024, China; (S.C.)
| | - Yan Qiang
- College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan 030024, China; (S.C.)
| | - Talha Anwar
- School of Computing, National University of Computer and Emerging Sciences, Islamabad 44000, Pakistan
| | | | - Chih-Chung Hsu
- Institute of Data Science, National Cheng Kung University, No. 1, University Rd., East Dist., Tainan City 701, Taiwan
| | - Shen-Chieh Tai
- Institute of Data Science, National Cheng Kung University, No. 1, University Rd., East Dist., Tainan City 701, Taiwan
| | - Shao-Ning Chen
- Institute of Data Science, National Cheng Kung University, No. 1, University Rd., East Dist., Tainan City 701, Taiwan
| | - Davide Tricarico
- Dipartimento di Informatica, Universita degli Studi di Torino, Corso Svizzera 185, 10149 Torino, Italy; (D.T.); (H.A.H.C.)
| | - Hafiza Ayesha Hoor Chaudhry
- Dipartimento di Informatica, Universita degli Studi di Torino, Corso Svizzera 185, 10149 Torino, Italy; (D.T.); (H.A.H.C.)
| | - Attilio Fiandrotti
- Dipartimento di Informatica, Universita degli Studi di Torino, Corso Svizzera 185, 10149 Torino, Italy; (D.T.); (H.A.H.C.)
| | - Marco Grangetto
- Dipartimento di Informatica, Universita degli Studi di Torino, Corso Svizzera 185, 10149 Torino, Italy; (D.T.); (H.A.H.C.)
| | | | - Alessandro Ortis
- Department of Mathematics and Computer Science, University of Catania, 95125 Catania, Italy (S.B.)
| | - Sebastiano Battiato
- Department of Mathematics and Computer Science, University of Catania, 95125 Catania, Italy (S.B.)
| |
Collapse
|
52
|
Murmu A, Kumar P. GIFNet: an effective global infection feature network for automatic COVID-19 lung lesions segmentation. Med Biol Eng Comput 2024:10.1007/s11517-024-03024-z. [PMID: 38308670 DOI: 10.1007/s11517-024-03024-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Accepted: 01/11/2024] [Indexed: 02/05/2024]
Abstract
The ongoing COronaVIrus Disease 2019 (COVID-19) pandemic carried by the SARS-CoV-2 virus spread worldwide in early 2019, bringing about an existential health catastrophe. Automatic segmentation of infected lungs from COVID-19 X-ray and computer tomography (CT) images helps to generate a quantitative approach for treatment and diagnosis. The multi-class information about the infected lung is often obtained from the patient's CT dataset. However, the main challenge is the extensive range of infected features and lack of contrast between infected and normal areas. To resolve these issues, a novel Global Infection Feature Network (GIFNet)-based Unet with ResNet50 model is proposed for segmenting the locations of COVID-19 lung infections. The Unet layers have been used to extract the features from input images and select the region of interest (ROI) by using the ResNet50 technique for training it faster. Moreover, integrating the pooling layer into the atrous spatial pyramid pooling (ASPP) mechanism in the bottleneck helps for better feature selection and handles scale variation during training. Furthermore, the partial differential equation (PDE) approach is used to enhance the image quality and intensity value for particular ROI boundary edges in the COVID-19 images. The proposed scheme has been validated on two datasets, namely the SARS-CoV-2 CT scan and COVIDx-19, for detecting infected lung segmentation (ILS). The experimental findings have been subjected to a comprehensive analysis using various evaluation metrics, including accuracy (ACC), area under curve (AUC), recall (REC), specificity (SPE), dice similarity coefficient (DSC), mean absolute error (MAE), precision (PRE), and mean squared error (MSE) to ensure rigorous validation. The results demonstrate the superior performance of the proposed system compared to the state-of-the-art (SOTA) segmentation models on both X-ray and CT datasets.
Collapse
Affiliation(s)
- Anita Murmu
- Computer Science and Engineering Department, National Institute of Technology Patna, Ashok Rajpath, Patna, Bihar, 800005, India.
| | - Piyush Kumar
- Computer Science and Engineering Department, National Institute of Technology Patna, Ashok Rajpath, Patna, Bihar, 800005, India
| |
Collapse
|
53
|
Zhao J, Sun L, Sun Z, Zhou X, Si H, Zhang D. MSEF-Net: Multi-scale edge fusion network for lumbosacral plexus segmentation with MR image. Artif Intell Med 2024; 148:102771. [PMID: 38325928 DOI: 10.1016/j.artmed.2024.102771] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Revised: 12/08/2023] [Accepted: 01/14/2024] [Indexed: 02/09/2024]
Abstract
Nerve damage of spine areas is a common cause of disability and paralysis. The lumbosacral plexus segmentation from magnetic resonance imaging (MRI) scans plays an important role in many computer-aided diagnoses and surgery of spinal nerve lesions. Due to the complex structure and low contrast of the lumbosacral plexus, it is difficult to delineate the regions of edges accurately. To address this issue, we propose a Multi-Scale Edge Fusion Network (MSEF-Net) to fully enhance the edge feature in the encoder and adaptively fuse multi-scale features in the decoder. Specifically, to highlight the edge structure feature, we propose an edge feature fusion module (EFFM) by combining the Sobel operator edge detection and the edge-guided attention module (EAM), respectively. To adaptively fuse the multi-scale feature map in the decoder, we introduce an adaptive multi-scale fusion module (AMSF). Our proposed MSEF-Net method was evaluated on the collected spinal MRI dataset with 89 patients (a total of 2848 MR images). Experimental results demonstrate that our MSEF-Net is effective for lumbosacral plexus segmentation with MR images, when compared with several state-of-the-art segmentation methods.
Collapse
Affiliation(s)
- Junyong Zhao
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, the Key Laboratory of Brain-Machine Intelligence Technology, Ministry of Education, Nanjing 211106, China
| | - Liang Sun
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, the Key Laboratory of Brain-Machine Intelligence Technology, Ministry of Education, Nanjing 211106, China; Nanjing University of Aeronautics and Astronautics Shenzhen Research Institute, Shenzhen 518063, China.
| | - Zhi Sun
- Department of Medical Imaging, Shandong Provincial Hospital, Jinan 250021, China
| | - Xin Zhou
- Department of Orthopedics, Qilu Hospital, Cheeloo College of Medicine, Shandong University, Jinan 250012, China
| | - Haipeng Si
- Department of Orthopedics, Qilu Hospital, Cheeloo College of Medicine, Shandong University, Jinan 250012, China.
| | - Daoqiang Zhang
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, the Key Laboratory of Brain-Machine Intelligence Technology, Ministry of Education, Nanjing 211106, China; Nanjing University of Aeronautics and Astronautics Shenzhen Research Institute, Shenzhen 518063, China.
| |
Collapse
|
54
|
Moosavi AS, Mahboobi A, Arabzadeh F, Ramezani N, Moosavi HS, Mehrpoor G. Segmentation and classification of lungs CT-scan for detecting COVID-19 abnormalities by deep learning technique: U-Net model. J Family Med Prim Care 2024; 13:691-698. [PMID: 38605799 PMCID: PMC11006039 DOI: 10.4103/jfmpc.jfmpc_695_23] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Revised: 07/12/2023] [Accepted: 09/22/2023] [Indexed: 04/13/2024] Open
Abstract
Background Artificial intelligence (AI) techniques have been ascertained useful in the analysis and description of infectious areas in radiological images promptly. Our aim in this study was to design a web-based application for detecting and labeling infected tissues on CT (computed tomography) lung images of patients based on the deep learning (DL) method as a type of AI. Materials and Methods The U-Net architecture, one of the DL networks, is used as a hybrid model with pre-trained densely connected convolutional network 121 (DenseNet121) architecture for the segmentation process. The proposed model was constructed on 1031 persons' CT-scan images from Ibn Sina Hospital of Iran in 2021 and some publicly available datasets. The network was trained using 6000 slices, validated on 1000 slices images, and tested against the 150 slices. Accuracy, sensitivity, specificity, and area under the receiver operating characteristics (ROC) curve (AUC) were calculated to evaluate model performance. Results The results indicate the acceptable ability of the U-Net-DenseNet121 model in detecting COVID-19 abnormality (accuracy = 0.88 and AUC = 0.96 for thresholds of 0.13 and accuracy = 0.88 and AUC = 0.90 for thresholds of 0.2). Based on this model, we developed the "Imaging-Tech" web-based application for use at hospitals and clinics to make our project's output more practical and attractive in the market. Conclusion We designed a DL-based model for the segmentation of COVID-19 CT scan images and, based on this model, constructed a web-based application that, according to the results, is a reliable detector for infected tissue in lung CT-scans. The availability of such tools would aid in automating, prioritizing, fastening, and broadening the treatment of COVID-19 patients globally.
Collapse
Affiliation(s)
| | - Ashraf Mahboobi
- Department of Radiologist, Babol University of Medical Sciences, Babol, Iran
| | - Farzin Arabzadeh
- Department of Radiologist, Dr. Arabzadeh Radiology and Sonography Clinic, Behbahan, Iran
| | - Nazanin Ramezani
- School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
| | - Helia S. Moosavi
- Computer Science Bachelor Degree, University of Toronto, On, Canada
| | - Golbarg Mehrpoor
- Department of Rheumatologist, Alborz University of Medical Sciences, Karaj, Iran
| |
Collapse
|
55
|
Reska D, Kretowski M. GPU-accelerated lung CT segmentation based on level sets and texture analysis. Sci Rep 2024; 14:1444. [PMID: 38228773 PMCID: PMC10792028 DOI: 10.1038/s41598-024-51452-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2023] [Accepted: 01/05/2024] [Indexed: 01/18/2024] Open
Abstract
This paper presents a novel semi-automatic method for lung segmentation in thoracic CT datasets. The fully three-dimensional algorithm is based on a level set representation of an active surface and integrates texture features to improve its robustness. The method's performance is enhanced by the graphics processing unit (GPU) acceleration. The segmentation process starts with a manual initialisation of 2D contours on a few representative slices of the analysed volume. Next, the starting regions for the active surface are generated according to the probability maps of texture features. The active surface is then evolved to give the final segmentation result. The recent implementation employs features based on grey-level co-occurrence matrices and Gabor filters. The algorithm was evaluated on real medical imaging data from the LCTCS 2017 challenge. The results were also compared with the outcomes of other segmentation methods. The proposed approach provided high segmentation accuracy while offering very competitive performance.
Collapse
Affiliation(s)
- Daniel Reska
- Faculty of Computer Science, Bialystok University of Technology, Białystok, Poland.
| | - Marek Kretowski
- Faculty of Computer Science, Bialystok University of Technology, Białystok, Poland
| |
Collapse
|
56
|
Xiao H, Li L, Liu Q, Zhang Q, Liu J, Liu Z. Context-aware and local-aware fusion with transformer for medical image segmentation. Phys Med Biol 2024; 69:025011. [PMID: 38086076 DOI: 10.1088/1361-6560/ad14c6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Accepted: 12/12/2023] [Indexed: 01/11/2024]
Abstract
Objective. Convolutional neural networks (CNNs) have made significant progress in medical image segmentation tasks. However, for complex segmentation tasks, CNNs lack the ability to establish long-distance relationships, resulting in poor segmentation performance. The characteristics of intra-class diversity and inter-class similarity in images increase the difficulty of segmentation. Additionally, some focus areas exhibit a scattered distribution, making segmentation even more challenging.Approach. Therefore, this work proposed a new Transformer model, FTransConv, to address the issues of inter-class similarity, intra-class diversity, and scattered distribution in medical image segmentation tasks. To achieve this, three Transformer-CNN modules were designed to extract global and local information, and a full-scale squeeze-excitation module was proposed in the decoder using the idea of full-scale connections.Main results. Without any pre-training, this work verified the effectiveness of FTransConv on three public COVID-19 CT datasets and MoNuSeg. Experiments have shown that FTransConv, which has only 26.98M parameters, outperformed other state-of-the-art models, such as Swin-Unet, TransAttUnet, UCTransNet, LeViT-UNet, TransUNet, UTNet, and SAUNet++. This model achieved the best segmentation performance with a DSC of 83.22% in COVID-19 datasets and 79.47% in MoNuSeg.Significance. This work demonstrated that our method provides a promising solution for regions with high inter-class similarity, intra-class diversity and scatter distribution in image segmentation.
Collapse
Affiliation(s)
- Hanguang Xiao
- College of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, People's Republic of China
| | - Li Li
- College of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, People's Republic of China
| | - Qiyuan Liu
- College of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, People's Republic of China
| | - Qihang Zhang
- College of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, People's Republic of China
| | - Junqi Liu
- College of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, People's Republic of China
| | - Zhi Liu
- College of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, People's Republic of China
| |
Collapse
|
57
|
Li Y, Yan B, Hou J, Bai B, Huang X, Xu C, Fang L. UNet based on dynamic convolution decomposition and triplet attention. Sci Rep 2024; 14:271. [PMID: 38168684 PMCID: PMC10761743 DOI: 10.1038/s41598-023-50989-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2023] [Accepted: 12/28/2023] [Indexed: 01/05/2024] Open
Abstract
The robustness and generalization of medical image segmentation models are being challenged by the differences between different disease types, different image types, and different cases.Deep learning based semantic segmentation methods have been providing state-of-the-art performance in the last few years. One deep learning technique, U-Net, has become the most popular architecture in the medical imaging segmentation. Despite outstanding overall performance in segmenting medical images, it still has the problems of limited feature expression ability and inaccurate segmentation. To this end, we propose a DTA-UNet based on Dynamic Convolution Decomposition (DCD) and Triple Attention (TA). Firstly, the model with Attention U-Net as the baseline network uses DCD to replace all the conventional convolution in the encoding-decoding process to enhance its feature extraction capability. Secondly, we combine TA with Attention Gate (AG) to be used for skip connection in order to highlight lesion regions by removing redundant information in both spatial and channel dimensions. The proposed model are tested on the two public datasets and actual clinical dataset such as the public COVID-SemiSeg dataset, the ISIC 2018 dataset, and the cooperative hospital stroke segmentation dataset. Ablation experiments on the clinical stroke segmentation dataset show the effectiveness of DCD and TA with only a 0.7628 M increase in the number of parameters compared to the baseline model. The proposed DTA-UNet is further evaluated on the three datasets of different types of images to verify its universality. Extensive experimental results show superior performance on different segmentation metrics compared to eight state-of-art methods.The GitHub URL of our code is https://github.com/shuaihou1234/DTA-UNet .
Collapse
Affiliation(s)
- Yang Li
- Academy for Advanced Interdisciplinary Studies, Northeast Normal University, Changchun, 130024, Jilin, China
- Shanghai Zhangjiang Institute of Mathematics, Shanghai, 201203, China
| | - Bobo Yan
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
- Pazhou Lab, Guangzhou, China
| | - Jianxin Hou
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Bingyang Bai
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Xiaoyu Huang
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Canfei Xu
- The Third Hospital of Jilin University, Changchun, 130117, Jilin, China
| | - Limei Fang
- Encephalopathy Center, The Third Affiliated Hospital of Changchun University of Chinese Medicine, Changchun, 130117, Jilin, China.
| |
Collapse
|
58
|
Vaikunta Pai T, Maithili K, Arun Kumar R, Nagaraju D, Anuradha D, Kumar S, Ravuri A, Sunilkumar Reddy T, Sivaram M, Vidhya RG. DKCNN: Improving deep kernel convolutional neural network-based COVID-19 identification from CT images of the chest. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:913-930. [PMID: 38820059 DOI: 10.3233/xst-230424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2024]
Abstract
BACKGROUND An efficient deep convolutional neural network (DeepCNN) is proposed in this article for the classification of Covid-19 disease. OBJECTIVE A novel structure known as the Pointwise-Temporal-pointwise convolution unit is developed incorporated with the varying kernel-based depth wise temporal convolution before and after the pointwise convolution operations. METHODS The outcome is optimized by the Slap Swarm algorithm (SSA). The proposed Deep CNN is composed of depth wise temporal convolution and end-to-end automatic detection of disease. First, the datasets SARS-COV-2 Ct-Scan Dataset and CT scan COVID Prediction dataset are preprocessed using the min-max approach and the features are extracted for further processing. RESULTS The experimental analysis is conducted between the proposed and some state-of-art works and stated that the proposed work effectively classifies the disease than the other approaches. CONCLUSION The proposed structural unit is used to design the deep CNN with the increasing kernel sizes. The classification process is improved by the inclusion of depth wise temporal convolutions along with the kernel variation. The computational complexity is reduced by the introduction of stride convolutions are used in the residual linkage among the adjacent structural units.
Collapse
Affiliation(s)
- T Vaikunta Pai
- Department of Information Science and Engineering, NMAM Institute of Technology-Affiliated to NITTE (Deemed to be University), Bangalore, Karnataka, India
| | - K Maithili
- Department of Computer Science and Engineering (Ai & ML), KG Reddy College of Engineering and Technology, Hyderabad, Telangana, India
| | - Ravula Arun Kumar
- Department of Computer Science and Engineering, Vardhaman College of Engineering, Hyderabad, Telangana, India
| | - D Nagaraju
- Department of Computer Science and Engineering, Sri Venkatesa Perumal College of Engineering and Technology, Puttur, Andhra Pradesh, India
| | - D Anuradha
- Department of Computer Science and Business Systems, Panimalar Engineering College, Chennai, India
| | - Shailendra Kumar
- Department of Electronics and Communication Engineering, Integral University Lucknow, Uttar Pradesh, India
| | | | - T Sunilkumar Reddy
- Department of Computer Science and Engineering, Sri Venkatesa Perumal College of Engineering and Technology, Puttur, Andhra Pradesh, India
| | - M Sivaram
- Saveetha School of Engineering, Saveetha Institute of Medical and Technical Sciences, Saveetha Nagar, Thandalam, Tamil Nadu, India
| | - R G Vidhya
- Department of ECE, HKBKCE, Bangalore, India
| |
Collapse
|
59
|
Shui Y, Wang Z, Liu B, Wang W, Fu S, Li Y. A three-path network with multi-scale selective feature fusion, edge-inspiring and edge-guiding for liver tumor segmentation. Comput Biol Med 2024; 168:107841. [PMID: 38081117 DOI: 10.1016/j.compbiomed.2023.107841] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2023] [Revised: 11/04/2023] [Accepted: 12/07/2023] [Indexed: 01/10/2024]
Abstract
Automatic liver tumor segmentation is one of the most important tasks in computer-aided diagnosis and treatment. Deep learning techniques have gained increasing popularity for medical image segmentation in recent years. However, due to the various shapes, sizes, and obscure boundaries of tumors, it is still difficult to automatically extract tumor regions from CT images. Based on the complementarity of edge detection and region segmentation, a three-path structure with multi-scale selective feature fusion (MSFF) module, multi-channel feature fusion (MFF) module, edge-inspiring (EI) module, and edge-guiding (EG) module is proposed in this paper. The MSFF module includes the process of generation, fusion, and selection of multi-scale features, which can adaptively correct the response weights in multiple branches to filter redundant information. The MFF module integrates richer hierarchical features to capture targets at different scales. The EI module aggregates high-level semantic information at different levels to obtain fine edge semantics, which is injected into the EG module for representation learning of segmentation features. Experiments on the LiTs2017 dataset show that our proposed method achieves a Dice index of 85.55% and a Jaccard index of 81.11%, which are higher than what can be obtained by the current state-of-the-art methods. Cross-dataset validation experiments conducted on 3Dircadb and Clinical datasets show the generalization and robustness of the proposed method by achieving dice indices of 80.14% and 81.68%, respectively.
Collapse
Affiliation(s)
- Yuanyuan Shui
- School of Mathematics, Shandong University, Jinan, 250100, China
| | - Zhendong Wang
- School of Mathematics, Shandong University, Jinan, 250100, China
| | - Bin Liu
- Department of Intervention Medicine, The Second Hospital of Shandong University, Jinan, 250033, China
| | - Wei Wang
- Department of Intervention Medicine, The Second Hospital of Shandong University, Jinan, 250033, China
| | - Shujun Fu
- School of Mathematics, Shandong University, Jinan, 250100, China; Department of Intervention Medicine, The Second Hospital of Shandong University, Jinan, 250033, China.
| | - Yuliang Li
- Department of Intervention Medicine, The Second Hospital of Shandong University, Jinan, 250033, China.
| |
Collapse
|
60
|
Gopatoti A, Jayakumar R, Billa P, Patteeswaran V. DDA-SSNets: Dual decoder attention-based semantic segmentation networks for COVID-19 infection segmentation and classification using chest X-Ray images. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:623-649. [PMID: 38607728 DOI: 10.3233/xst-230421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/14/2024]
Abstract
BACKGROUND COVID-19 needs to be diagnosed and staged to be treated accurately. However, prior studies' diagnostic and staging abilities for COVID-19 infection needed to be improved. Therefore, new deep learning-based approaches are required to aid radiologists in detecting and quantifying COVID-19-related lung infections. OBJECTIVE To develop deep learning-based models to classify and quantify COVID-19-related lung infections. METHODS Initially, Dual Decoder Attention-based Semantic Segmentation Networks (DDA-SSNets) such as Dual Decoder Attention-UNet (DDA-UNet) and Dual Decoder Attention-SegNet (DDA-SegNet) are proposed to facilitate the dual segmentation tasks such as lung lobes and infection segmentation in chest X-ray (CXR) images. The lung lobe and infection segmentations are mapped to grade the severity of COVID-19 infection in both the lungs of CXRs. Later, a Genetic algorithm-based Deep Convolutional Neural Network classifier with the optimum number of layers, namely GADCNet, is proposed to classify the extracted regions of interest (ROI) from the CXR lung lobes into COVID-19 and non-COVID-19. RESULTS The DDA-SegNet shows better segmentation with an average BCSSDC of 99.53% and 99.97% for lung lobes and infection segmentations, respectively, compared with DDA-UNet with an average BCSSDC of 99.14% and 99.92%. The proposed DDA-SegNet with GADCNet classifier offered excellent classification results with an average BCCAC of 99.98%, followed by the GADCNet with DDA-UNet with an average BCCAC of 99.92% after extensive testing and analysis. CONCLUSIONS The results show that the proposed DDA-SegNet has superior performance in the segmentation of lung lobes and COVID-19-infected regions in CXRs, along with improved severity grading compared to the DDA-UNet and improved accuracy of the GADCNet classifier in classifying the CXRs into COVID-19, and non-COVID-19.
Collapse
Affiliation(s)
- Anandbabu Gopatoti
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Ramya Jayakumar
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Poornaiah Billa
- Department of Electronics and Communication Engineering, Lakireddy Bali Reddy College of Engineering, Mylavaram, Andhra Pradesh, India
| | - Vijayalakshmi Patteeswaran
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| |
Collapse
|
61
|
Amgothu S, Koppu S. COVID-19 prediction using Caviar Squirrel Jellyfish Search Optimization technique in fog-cloud based architecture. PLoS One 2023; 18:e0295599. [PMID: 38127990 PMCID: PMC10735048 DOI: 10.1371/journal.pone.0295599] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Accepted: 11/24/2023] [Indexed: 12/23/2023] Open
Abstract
In the pandemic of COVID-19 patients approach to the hospital for prescription, yet due to extreme line up the patient gets treatment after waiting for more than one hour. Generally, wearable devices directly measure the preliminary data of the patient stored in capturing mode. In order to store the data, the hospitals require large storage devices that make the progression of data more complex. To bridge this gap, a potent scheme is established for COVID-19 prediction based fog-cloud named Caviar Squirrel Jellyfish Search Optimization (CSJSO). Here, CSJSO is the amalgamation of CAViar Squirrel Search Algorithm (CSSA) and Jellyfish Search Optimization (JSO), where CSSA is blended by the Conditional Autoregressive Value-at-Risk (CAViar) and Squirrel Search Algorithm (SSA). This architecture comprises the healthcare IoT sensor layer, fog layer and cloud layer. In the healthcare IoT sensor layer, the routing process with the collection of patient health condition data is carried out. On the other hand, in the fog layer COVID-19 detection is performed by employing a Deep Neuro Fuzzy Network (DNFN) trained by the proposed Remora Namib Beetle JSO (RNBJSO). Here, RNBJSO is the combination of Namib Beetle Optimization (NBO), Remora Optimization Algorithm (ROA) and Jellyfish Search optimization (JSO). Finally, in the cloud layer, the detection of COVID-19 employing Deep Long Short Term Memory (Deep LSTM) trained utilizing proposed CSJSO is performed. The evaluation measures utilized for CSJSO_Deep LSTM in database-1, such as Mean Squared Error (MSE) and Root Mean Squared Error (RMSE) observed 0.062 and 0.252 in confirmed cases. The measures employed in database-2 are accuracy, sensitivity and specificity achieved 0.925, 0.928 and 0.925 in K-set.
Collapse
Affiliation(s)
- Shanthi Amgothu
- School of Computer Science Engineering and Information Systems, Vellore, India
| | - Srinivas Koppu
- School of Computer Science Engineering and Information Systems, Vellore, India
| |
Collapse
|
62
|
Antar S, Abd El-Sattar HKH, Abdel-Rahman MH, F M Ghaleb F. COVID-19 infection segmentation using hybrid deep learning and image processing techniques. Sci Rep 2023; 13:22737. [PMID: 38123587 PMCID: PMC10733411 DOI: 10.1038/s41598-023-49337-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2023] [Accepted: 12/07/2023] [Indexed: 12/23/2023] Open
Abstract
The coronavirus disease 2019 (COVID-19) epidemic has become a worldwide problem that continues to affect people's lives daily, and the early diagnosis of COVID-19 has a critical importance on the treatment of infected patients for medical and healthcare organizations. To detect COVID-19 infections, medical imaging techniques, including computed tomography (CT) scan images and X-ray images, are considered some of the helpful medical tests that healthcare providers carry out. However, in addition to the difficulty of segmenting contaminated areas from CT scan images, these approaches also offer limited accuracy for identifying the virus. Accordingly, this paper addresses the effectiveness of using deep learning (DL) and image processing techniques, which serve to expand the dataset without the need for any augmentation strategies, and it also presents a novel approach for detecting COVID-19 virus infections in lung images, particularly the infection prediction issue. In our proposed method, to reveal the infection, the input images are first preprocessed using a threshold then resized to 128 × 128. After that, a density heat map tool is used for coloring the resized lung images. The three channels (red, green, and blue) are then separated from the colored image and are further preprocessed through image inverse and histogram equalization, and are subsequently fed, in independent directions, into three separate U-Nets with the same architecture for segmentation. Finally, the segmentation results are combined and run through a convolution layer one by one to get the detection. Several evaluation metrics using the CT scan dataset were used to measure the performance of the proposed approach in comparison with other state-of-the-art techniques in terms of accuracy, sensitivity, precision, and the dice coefficient. The experimental results of the proposed approach reached 99.71%, 0.83, 0.87, and 0.85, respectively. These results show that coloring the CT scan images dataset and then dividing each image into its RGB image channels can enhance the COVID-19 detection, and it also increases the U-Net power in the segmentation when merging the channel segmentation results. In comparison to other existing segmentation techniques employing bigger 512 × 512 images, this study is one of the few that can rapidly and correctly detect the COVID-19 virus with high accuracy on smaller 128 × 128 images using the metrics of accuracy, sensitivity, precision, and dice coefficient.
Collapse
Affiliation(s)
- Samar Antar
- Computer Science Division, Department of Mathematics, Faculty of Science, Ain Shams University, Abbassia, Cairo, 11566, Egypt
| | | | - Mohammad H Abdel-Rahman
- Computer Science Division, Department of Mathematics, Faculty of Science, Ain Shams University, Abbassia, Cairo, 11566, Egypt
| | - Fayed F M Ghaleb
- Computer Science Division, Department of Mathematics, Faculty of Science, Ain Shams University, Abbassia, Cairo, 11566, Egypt
| |
Collapse
|
63
|
Yang D, Xu T, Zhang Y, An D, Wang Q, Pan Z, Liu G, Yue Y. Image-fusion-based object detection using a time-of-flight camera. OPTICS EXPRESS 2023; 31:43100-43114. [PMID: 38178412 DOI: 10.1364/oe.510101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Accepted: 11/23/2023] [Indexed: 01/06/2024]
Abstract
In this work, we demonstrate an innovative object detection framework based on depth and active infrared intensity images fusion with a time-of-flight (ToF) camera. A slide window weight fusion (SWWF) method provides fuse image with two modalities to localize targets. Then, the depth and intensity information is extracted to construct a joint feature space. Next, we utilize four machine learning methods to achieve object recognition. To verify this method, experiments are performed on an in-house dataset containing 1066 images, which are categorized into six different surface materials. Consequently, the approach performs well on localization with a 0.778 intersection over union (IoU). The best classification results are obtained with K-Nearest Neighbor (KNN) with a 98.01% total accuracy. Furthermore, our demonstrated method is less affected by various illumination conditions.
Collapse
|
64
|
Buongiorno R, Del Corso G, Germanese D, Colligiani L, Python L, Romei C, Colantonio S. Enhancing COVID-19 CT Image Segmentation: A Comparative Study of Attention and Recurrence in UNet Models. J Imaging 2023; 9:283. [PMID: 38132701 PMCID: PMC10744014 DOI: 10.3390/jimaging9120283] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Revised: 12/09/2023] [Accepted: 12/13/2023] [Indexed: 12/23/2023] Open
Abstract
Imaging plays a key role in the clinical management of Coronavirus disease 2019 (COVID-19) as the imaging findings reflect the pathological process in the lungs. The visual analysis of High-Resolution Computed Tomography of the chest allows for the differentiation of parenchymal abnormalities of COVID-19, which are crucial to be detected and quantified in order to obtain an accurate disease stratification and prognosis. However, visual assessment and quantification represent a time-consuming task for radiologists. In this regard, tools for semi-automatic segmentation, such as those based on Convolutional Neural Networks, can facilitate the detection of pathological lesions by delineating their contour. In this work, we compared four state-of-the-art Convolutional Neural Networks based on the encoder-decoder paradigm for the binary segmentation of COVID-19 infections after training and testing them on 90 HRCT volumetric scans of patients diagnosed with COVID-19 collected from the database of the Pisa University Hospital. More precisely, we started from a basic model, the well-known UNet, then we added an attention mechanism to obtain an Attention-UNet, and finally we employed a recurrence paradigm to create a Recurrent-Residual UNet (R2-UNet). In the latter case, we also added attention gates to the decoding path of an R2-UNet, thus designing an R2-Attention UNet so as to make the feature representation and accumulation more effective. We compared them to gain understanding of both the cognitive mechanism that can lead a neural model to the best performance for this task and the good compromise between the amount of data, time, and computational resources required. We set up a five-fold cross-validation and assessed the strengths and limitations of these models by evaluating the performances in terms of Dice score, Precision, and Recall defined both on 2D images and on the entire 3D volume. From the results of the analysis, it can be concluded that Attention-UNet outperforms the other models by achieving the best performance of 81.93%, in terms of 2D Dice score, on the test set. Additionally, we conducted statistical analysis to assess the performance differences among the models. Our findings suggest that integrating the recurrence mechanism within the UNet architecture leads to a decline in the model's effectiveness for our particular application.
Collapse
Affiliation(s)
- Rossana Buongiorno
- Institute of Information Science and Technologies, National Research Council of Italy (ISTI-CNR), 56124 Pisa, PI, Italy; (G.D.C.); (S.C.)
| | - Giulio Del Corso
- Institute of Information Science and Technologies, National Research Council of Italy (ISTI-CNR), 56124 Pisa, PI, Italy; (G.D.C.); (S.C.)
| | - Danila Germanese
- Institute of Information Science and Technologies, National Research Council of Italy (ISTI-CNR), 56124 Pisa, PI, Italy; (G.D.C.); (S.C.)
| | - Leonardo Colligiani
- Department of Translational Research, Academic Radiology, University of Pisa, 56124 Pisa, PI, Italy;
| | - Lorenzo Python
- 2nd Radiology Unit, Pisa University Hospital, 56124 Pisa, PI, Italy; (L.P.)
| | - Chiara Romei
- 2nd Radiology Unit, Pisa University Hospital, 56124 Pisa, PI, Italy; (L.P.)
| | - Sara Colantonio
- Institute of Information Science and Technologies, National Research Council of Italy (ISTI-CNR), 56124 Pisa, PI, Italy; (G.D.C.); (S.C.)
| |
Collapse
|
65
|
Lu F, Zhang Z, Liu T, Tang C, Bai H, Zhai G, Chen J, Wu X. A weakly supervised inpainting-based learning method for lung CT image segmentation. PATTERN RECOGNITION 2023; 144:109861. [DOI: 10.1016/j.patcog.2023.109861] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/05/2024]
|
66
|
Ozaltin O, Yeniay O, Subasi A. OzNet: A New Deep Learning Approach for Automated Classification of COVID-19 Computed Tomography Scans. BIG DATA 2023; 11:420-436. [PMID: 36927081 DOI: 10.1089/big.2022.0042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Coronavirus disease 2019 (COVID-19) is spreading rapidly around the world. Therefore, the classification of computed tomography (CT) scans alleviates the workload of experts, whose workload increased considerably during the pandemic. Convolutional neural network (CNN) architectures are successful for the classification of medical images. In this study, we have developed a new deep CNN architecture called OzNet. Moreover, we have compared it with pretrained architectures namely AlexNet, DenseNet201, GoogleNet, NASNetMobile, ResNet-50, SqueezeNet, and VGG-16. In addition, we have compared the classification success of three preprocessing methods with raw CT scans. We have not only classified the raw CT scans, but also have performed the classification with three different preprocessing methods, which are discrete wavelet transform (DWT), intensity adjustment, and gray to color red, green, blue image conversion on the data sets. Furthermore, it is known that the architecture's performance increases with the use of DWT preprocessing method rather than using the raw data set. The results are extremely promising with the CNN algorithms using the COVID-19 CT scans processed with the DWT. The proposed DWT-OzNet has achieved a high classification performance of more than 98.8% for each calculated metric.
Collapse
Affiliation(s)
- Oznur Ozaltin
- Department of Statistics, Institute of Science, Hacettepe University, Ankara, Turkey
| | - Ozgur Yeniay
- Department of Statistics, Institute of Science, Hacettepe University, Ankara, Turkey
| | - Abdulhamit Subasi
- Institute of Biomedicine, Faculty of Medicine, University of Turku, Turku, Finland
- Department of Computer Science, College of Engineering, Effat University, Jeddah, Saudi Arabia
| |
Collapse
|
67
|
Song P, Li J, Fan H, Fan L. TGDAUNet: Transformer and GCNN based dual-branch attention UNet for medical image segmentation. Comput Biol Med 2023; 167:107583. [PMID: 37890420 DOI: 10.1016/j.compbiomed.2023.107583] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Revised: 09/28/2023] [Accepted: 10/15/2023] [Indexed: 10/29/2023]
Abstract
Accurate and automatic segmentation of medical images is a key step in clinical diagnosis and analysis. Currently, the successful application of Transformers' model in the field of computer vision, researchers have begun to gradually explore the application of Transformers in medical segmentation of images, especially in combination with convolutional neural networks with coding-decoding structure, which have achieved remarkable results in the field of medical segmentation. However, most studies have combined Transformers with CNNs at a single scale or processed only the highest-level semantic feature information, ignoring the rich location information in the lower-level semantic feature information. At the same time, for problems such as blurred structural boundaries and heterogeneous textures in images, most existing methods usually simply connect contour information to capture the boundaries of the target. However, these methods cannot capture the precise outline of the target and ignore the potential relationship between the boundary and the region. In this paper, we propose the TGDAUNet, which consists of a dual-branch backbone network of CNNs and Transformers and a parallel attention mechanism, to achieve accurate segmentation of lesions in medical images. Firstly, high-level semantic feature information of the CNN backbone branches is fused at multiple scales, and the high-level and low-level feature information complement each other's location and spatial information. We further use the polarised self-attentive (PSA) module to reduce the impact of redundant information caused by multiple scales, to better couple with the feature information extracted from the Transformers backbone branch, and to establish global contextual long-range dependencies at multiple scales. In addition, we have designed the Reverse Graph-reasoned Fusion (RGF) module and the Feature Aggregation (FA) module to jointly guide the global context. The FA module aggregates high-level semantic feature information to generate an original global predictive segmentation map. The RGF module captures non-significant features of the boundaries in the original or secondary global prediction segmentation graph through a reverse attention mechanism, establishing a graph reasoning module to explore the potential semantic relationships between boundaries and regions, further refining the target boundaries. Finally, to validate the effectiveness of our proposed method, we compare our proposed method with the current popular methods in the CVC-ClinicDB, Kvasir-SEG, ETIS, CVC-ColonDB, CVC-300,datasets as well as the skin cancer segmentation datasets ISIC-2016 and ISIC-2017. The large number of experimental results show that our method outperforms the currently popular methods. Source code is released at https://github.com/sd-spf/TGDAUNet.
Collapse
Affiliation(s)
- Pengfei Song
- Co-Innovation Center of Shandong Colleges and Universities: Future Intelligent Computing, School of Computer Science and Technology, Shandong Technology and Business University, Laishan District, Yantai, 264005, China
| | - Jinjiang Li
- Co-Innovation Center of Shandong Colleges and Universities: Future Intelligent Computing, School of Computer Science and Technology, Shandong Technology and Business University, Laishan District, Yantai, 264005, China
| | - Hui Fan
- Co-Innovation Center of Shandong Colleges and Universities: Future Intelligent Computing, School of Computer Science and Technology, Shandong Technology and Business University, Laishan District, Yantai, 264005, China.
| | - Linwei Fan
- School of Computer Science and Technology, Shandong University of Finance and Economics, Jinan, Shandong, 250014, China
| |
Collapse
|
68
|
Yue G, Yang C, Zhao Z, An Z, Yang Y. ERGPNet: lesion segmentation network for COVID-19 chest X-ray images based on embedded residual convolution and global perception. Front Physiol 2023; 14:1296185. [PMID: 38028767 PMCID: PMC10679680 DOI: 10.3389/fphys.2023.1296185] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 11/02/2023] [Indexed: 12/01/2023] Open
Abstract
The Segmentation of infected areas from COVID-19 chest X-ray (CXR) images is of great significance for the diagnosis and treatment of patients. However, accurately and effectively segmenting infected areas of CXR images is still challenging due to the inherent ambiguity of CXR images and the cross-scale variations in infected regions. To address these issues, this article proposes a ERGPNet based on embedded residuals and global perception, to segment lesion regions in COVID-19 CXR images. First, aiming at the inherent fuzziness of CXR images, an embedded residual convolution structure is proposed to enhance the ability of internal feature extraction. Second, a global information perception module is constructed to guide the network in generating long-distance information flow, alleviating the interferences of cross-scale variations on the algorithm's discrimination ability. Finally, the network's sensitivity to target regions is improved, and the interference of noise information is suppressed through the utilization of parallel spatial and serial channel attention modules. The interactions between each module fully establish the mapping relationship between feature representation and information decision-making and improve the accuracy of lesion segmentation. Extensive experiments on three datasets of COVID-19 CXR images, and the results demonstrate that the proposed method outperforms other state-of-the-art segmentation methods of CXR images.
Collapse
Affiliation(s)
- Gongtao Yue
- School of Computer Science, Xijing University, Xi’an, China
| | - Chen Yang
- School of Computer Science, Xijing University, Xi’an, China
| | - Zhengyang Zhao
- School of Information and Navigation, Air Force Engineering University, Xi’an, China
| | - Ziheng An
- School of Integrated Circuits, Anhui University, Hefei, China
| | - Yongsheng Yang
- School of Computer Science, Xijing University, Xi’an, China
| |
Collapse
|
69
|
Mu N, Guo J, Wang R. Automated polyp segmentation based on a multi-distance feature dissimilarity-guided fully convolutional network. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:20116-20134. [PMID: 38052639 DOI: 10.3934/mbe.2023891] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
Colorectal malignancies often arise from adenomatous polyps, which typically begin as solitary, asymptomatic growths before progressing to malignancy. Colonoscopy is widely recognized as a highly efficacious clinical polyp detection method, offering valuable visual data that facilitates precise identification and subsequent removal of these tumors. Nevertheless, accurately segmenting individual polyps poses a considerable difficulty because polyps exhibit intricate and changeable characteristics, including shape, size, color, quantity and growth context during different stages. The presence of similar contextual structures around polyps significantly hampers the performance of commonly used convolutional neural network (CNN)-based automatic detection models to accurately capture valid polyp features, and these large receptive field CNN models often overlook the details of small polyps, which leads to the occurrence of false detections and missed detections. To tackle these challenges, we introduce a novel approach for automatic polyp segmentation, known as the multi-distance feature dissimilarity-guided fully convolutional network. This approach comprises three essential components, i.e., an encoder-decoder, a multi-distance difference (MDD) module and a hybrid loss (HL) module. Specifically, the MDD module primarily employs a multi-layer feature subtraction (MLFS) strategy to propagate features from the encoder to the decoder, which focuses on extracting information differences between neighboring layers' features at short distances, and both short and long-distance feature differences across layers. Drawing inspiration from pyramids, the MDD module effectively acquires discriminative features from neighboring layers or across layers in a continuous manner, which helps to strengthen feature complementary across different layers. The HL module is responsible for supervising the feature maps extracted at each layer of the network to improve prediction accuracy. Our experimental results on four challenge datasets demonstrate that the proposed approach exhibits superior automatic polyp performance in terms of the six evaluation criteria compared to five current state-of-the-art approaches.
Collapse
Affiliation(s)
- Nan Mu
- College of Computer Science, Sichuan Normal University, Chengdu 610101, China
- Visual Computing and Virtual Reality Key Laboratory of Sichuan, Sichuan Normal University, Chengdu 610068, China
- Education Big Data Collaborative Innovation Center of Sichuan 2011, Chengdu 610101, China
| | - Jinjia Guo
- Chongqing University-University of Cincinnati Joint Co-op Institution, Chongqing University, Chongqing 400044, China
| | - Rong Wang
- College of Computer Science, Sichuan Normal University, Chengdu 610101, China
- Visual Computing and Virtual Reality Key Laboratory of Sichuan, Sichuan Normal University, Chengdu 610068, China
- Education Big Data Collaborative Innovation Center of Sichuan 2011, Chengdu 610101, China
| |
Collapse
|
70
|
Xiang Z, Mao Q, Wang J, Tian Y, Zhang Y, Wang W. Dmbg-Net: Dilated multiresidual boundary guidance network for COVID-19 infection segmentation. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:20135-20154. [PMID: 38052640 DOI: 10.3934/mbe.2023892] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
Accurate segmentation of infected regions in lung computed tomography (CT) images is essential for the detection and diagnosis of coronavirus disease 2019 (COVID-19). However, lung lesion segmentation has some challenges, such as obscure boundaries, low contrast and scattered infection areas. In this paper, the dilated multiresidual boundary guidance network (Dmbg-Net) is proposed for COVID-19 infection segmentation in CT images of the lungs. This method focuses on semantic relationship modelling and boundary detail guidance. First, to effectively minimize the loss of significant features, a dilated residual block is substituted for a convolutional operation, and dilated convolutions are employed to expand the receptive field of the convolution kernel. Second, an edge-attention guidance preservation block is designed to incorporate boundary guidance of low-level features into feature integration, which is conducive to extracting the boundaries of the region of interest. Third, the various depths of features are used to generate the final prediction, and the utilization of a progressive multi-scale supervision strategy facilitates enhanced representations and highly accurate saliency maps. The proposed method is used to analyze COVID-19 datasets, and the experimental results reveal that the proposed method has a Dice similarity coefficient of 85.6% and a sensitivity of 84.2%. Extensive experimental results and ablation studies have shown the effectiveness of Dmbg-Net. Therefore, the proposed method has a potential application in the detection, labeling and segmentation of other lesion areas.
Collapse
Affiliation(s)
- Zhenwu Xiang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Qi Mao
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Jintao Wang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Yi Tian
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Yan Zhang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Wenfeng Wang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| |
Collapse
|
71
|
Khan SH, Iqbal J, Hassnain SA, Owais M, Mostafa SM, Hadjouni M, Mahmoud A. COVID-19 detection and analysis from lung CT images using novel channel boosted CNNs. EXPERT SYSTEMS WITH APPLICATIONS 2023; 229:120477. [PMID: 37220492 PMCID: PMC10186852 DOI: 10.1016/j.eswa.2023.120477] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Revised: 05/10/2023] [Accepted: 05/10/2023] [Indexed: 05/25/2023]
Abstract
In December 2019, the global pandemic COVID-19 in Wuhan, China, affected human life and the worldwide economy. Therefore, an efficient diagnostic system is required to control its spread. However, the automatic diagnostic system poses challenges with a limited amount of labeled data, minor contrast variation, and high structural similarity between infection and background. In this regard, a new two-phase deep convolutional neural network (CNN) based diagnostic system is proposed to detect minute irregularities and analyze COVID-19 infection. In the first phase, a novel SB-STM-BRNet CNN is developed, incorporating a new channel Squeezed and Boosted (SB) and dilated convolutional-based Split-Transform-Merge (STM) block to detect COVID-19 infected lung CT images. The new STM blocks performed multi-path region-smoothing and boundary operations, which helped to learn minor contrast variation and global COVID-19 specific patterns. Furthermore, the diverse boosted channels are achieved using the SB and Transfer Learning concepts in STM blocks to learn texture variation between COVID-19-specific and healthy images. In the second phase, COVID-19 infected images are provided to the novel COVID-CB-RESeg segmentation CNN to identify and analyze COVID-19 infectious regions. The proposed COVID-CB-RESeg methodically employed region-homogeneity and heterogeneity operations in each encoder-decoder block and boosted-decoder using auxiliary channels to simultaneously learn the low illumination and boundaries of the COVID-19 infected region. The proposed diagnostic system yields good performance in terms of accuracy: 98.21 %, F-score: 98.24%, Dice Similarity: 96.40 %, and IOU: 98.85 % for the COVID-19 infected region. The proposed diagnostic system would reduce the burden and strengthen the radiologist's decision for a fast and accurate COVID-19 diagnosis.
Collapse
Affiliation(s)
- Saddam Hussain Khan
- Department of Computer Systems Engineering, University of Engineering and Applied Science, Swat 19060, Pakistan
| | - Javed Iqbal
- Department of Computer Systems Engineering, University of Engineering and Applied Science, Swat 19060, Pakistan
| | - Syed Agha Hassnain
- Ocean College, Zhejiang University, Zheda Road 1, Zhoushan, Zhejiang 316021, China
| | - Muhammad Owais
- KUCARS and C2PS, Department of Electrical Engineering and Computer Science, Khalifa University, UAE
| | - Samih M Mostafa
- Computer Science Department, Faculty of Computers and Information, South Valley University, Qena 83523, Egypt
- Faculty of Industry and Energy Technology, New Assiut Technological University (N.A.T.U.), New Assiut City, Egypt
| | - Myriam Hadjouni
- Department of Computer Sciences, College of Computer and Information Science, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Amena Mahmoud
- Faculty of Computers and Information, Department of Computer Science, KafrElSkeikh University, Egypt
| |
Collapse
|
72
|
Li Y, Zhou T, He K, Zhou Y, Shen D. Multi-Scale Transformer Network With Edge-Aware Pre-Training for Cross-Modality MR Image Synthesis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3395-3407. [PMID: 37339020 DOI: 10.1109/tmi.2023.3288001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/22/2023]
Abstract
Cross-modality magnetic resonance (MR) image synthesis can be used to generate missing modalities from given ones. Existing (supervised learning) methods often require a large number of paired multi-modal data to train an effective synthesis model. However, it is often challenging to obtain sufficient paired data for supervised training. In reality, we often have a small number of paired data while a large number of unpaired data. To take advantage of both paired and unpaired data, in this paper, we propose a Multi-scale Transformer Network (MT-Net) with edge-aware pre-training for cross-modality MR image synthesis. Specifically, an Edge-preserving Masked AutoEncoder (Edge-MAE) is first pre-trained in a self-supervised manner to simultaneously perform 1) image imputation for randomly masked patches in each image and 2) whole edge map estimation, which effectively learns both contextual and structural information. Besides, a novel patch-wise loss is proposed to enhance the performance of Edge-MAE by treating different masked patches differently according to the difficulties of their respective imputations. Based on this proposed pre-training, in the subsequent fine-tuning stage, a Dual-scale Selective Fusion (DSF) module is designed (in our MT-Net) to synthesize missing-modality images by integrating multi-scale features extracted from the encoder of the pre-trained Edge-MAE. Furthermore, this pre-trained encoder is also employed to extract high-level features from the synthesized image and corresponding ground-truth image, which are required to be similar (consistent) in the training. Experimental results show that our MT-Net achieves comparable performance to the competing methods even using 70% of all available paired data. Our code will be released at https://github.com/lyhkevin/MT-Net.
Collapse
|
73
|
Lyu F, Ye M, Yip TCF, Wong GLH, Yuen PC. Local Style Transfer via Latent Space Manipulation for Cross-Disease Lesion Segmentation. IEEE J Biomed Health Inform 2023; PP:273-284. [PMID: 37883256 DOI: 10.1109/jbhi.2023.3327726] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2023]
Abstract
Automatic lesion segmentation is important for assisting doctors in the diagnostic process. Recent deep learning approaches heavily rely on large-scale datasets, which are difficult to obtain in many clinical applications. Leveraging external labelled datasets is an effective solution to tackle the problem of insufficient training data. In this paper, we propose a new framework, namely LatenTrans, to utilize existing datasets for boosting the performance of lesion segmentation in extremely low data regimes. LatenTrans translates non-target lesions into target-like lesions and expands the training dataset with target-like data for better performance. Images are first projected to the latent space via aligned style-based generative models, and rich lesion semantics are encoded using the latent codes. A novel consistency-aware latent code manipulation module is proposed to enable high-quality local style transfer from non-target lesions to target-like lesions while preserving other parts. Moreover, we propose a new metric, Normalized Latent Distance, to solve the question of how to select an adequate one from various existing datasets for knowledge transfer. Extensive experiments are conducted on segmenting lung and brain lesions, and the experimental results demonstrate that our proposed LatenTrans is superior to existing methods for cross-disease lesion segmentation.
Collapse
|
74
|
Tayebi Arasteh S, Romanowicz J, Pace DF, Golland P, Powell AJ, Maier AK, Truhn D, Brosch T, Weese J, Lotfinia M, van der Geest RJ, Moghari MH. Automated segmentation of 3D cine cardiovascular magnetic resonance imaging. Front Cardiovasc Med 2023; 10:1167500. [PMID: 37904806 PMCID: PMC10613522 DOI: 10.3389/fcvm.2023.1167500] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2023] [Accepted: 09/18/2023] [Indexed: 11/01/2023] Open
Abstract
Introduction As the life expectancy of children with congenital heart disease (CHD) is rapidly increasing and the adult population with CHD is growing, there is an unmet need to improve clinical workflow and efficiency of analysis. Cardiovascular magnetic resonance (CMR) is a noninvasive imaging modality for monitoring patients with CHD. CMR exam is based on multiple breath-hold 2-dimensional (2D) cine acquisitions that should be precisely prescribed and is expert and institution dependent. Moreover, 2D cine images have relatively thick slices, which does not allow for isotropic delineation of ventricular structures. Thus, development of an isotropic 3D cine acquisition and automatic segmentation method is worthwhile to make CMR workflow straightforward and efficient, as the present work aims to establish. Methods Ninety-nine patients with many types of CHD were imaged using a non-angulated 3D cine CMR sequence covering the whole-heart and great vessels. Automatic supervised and semi-supervised deep-learning-based methods were developed for whole-heart segmentation of 3D cine images to separately delineate the cardiac structures, including both atria, both ventricles, aorta, pulmonary arteries, and superior and inferior vena cavae. The segmentation results derived from the two methods were compared with the manual segmentation in terms of Dice score, a degree of overlap agreement, and atrial and ventricular volume measurements. Results The semi-supervised method resulted in a better overlap agreement with the manual segmentation than the supervised method for all 8 structures (Dice score 83.23 ± 16.76% vs. 77.98 ± 19.64%; P-value ≤0.001). The mean difference error in atrial and ventricular volumetric measurements between manual segmentation and semi-supervised method was lower (bias ≤ 5.2 ml) than the supervised method (bias ≤ 10.1 ml). Discussion The proposed semi-supervised method is capable of cardiac segmentation and chamber volume quantification in a CHD population with wide anatomical variability. It accurately delineates the heart chambers and great vessels and can be used to accurately calculate ventricular and atrial volumes throughout the cardiac cycle. Such a segmentation method can reduce inter- and intra- observer variability and make CMR exams more standardized and efficient.
Collapse
Affiliation(s)
- Soroosh Tayebi Arasteh
- Department of Cardiology, Boston Children’s Hospital, and Department of Pediatrics, Harvard Medical School, Boston, MA, United States
- Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
- Department of Diagnostic and Interventional Radiology, University Hospital RWTH Aachen, Aachen, Germany
| | - Jennifer Romanowicz
- Department of Cardiology, Boston Children’s Hospital, and Department of Pediatrics, Harvard Medical School, Boston, MA, United States
- Department of Cardiology, Children’s Hospital Colorado, and School of Medicine, University of Colorado, Aurora, CO, United States
| | - Danielle F. Pace
- Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA, United States
- Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Polina Golland
- Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Andrew J. Powell
- Department of Cardiology, Boston Children’s Hospital, and Department of Pediatrics, Harvard Medical School, Boston, MA, United States
| | - Andreas K. Maier
- Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
| | - Daniel Truhn
- Department of Diagnostic and Interventional Radiology, University Hospital RWTH Aachen, Aachen, Germany
| | - Tom Brosch
- Philips Research Laboratories, Hamburg, Germany
| | | | - Mahshad Lotfinia
- Institute of Heat and Mass Transfer, RWTH Aachen University, Aachen, Germany
| | | | - Mehdi H. Moghari
- Department of Radiology, Children’s Hospital Colorado, and School of Medicine, University of Colorado, Aurora, CO, United States
| |
Collapse
|
75
|
Chu WT, Reza SMS, Anibal JT, Landa A, Crozier I, Bağci U, Wood BJ, Solomon J. Artificial Intelligence and Infectious Disease Imaging. J Infect Dis 2023; 228:S322-S336. [PMID: 37788501 PMCID: PMC10547369 DOI: 10.1093/infdis/jiad158] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2022] [Accepted: 05/06/2023] [Indexed: 10/05/2023] Open
Abstract
The mass production of the graphics processing unit and the coronavirus disease 2019 (COVID-19) pandemic have provided the means and the motivation, respectively, for rapid developments in artificial intelligence (AI) and medical imaging techniques. This has led to new opportunities to improve patient care but also new challenges that must be overcome before these techniques are put into practice. In particular, early AI models reported high performances but failed to perform as well on new data. However, these mistakes motivated further innovation focused on developing models that were not only accurate but also stable and generalizable to new data. The recent developments in AI in response to the COVID-19 pandemic will reap future dividends by facilitating, expediting, and informing other medical AI applications and educating the broad academic audience on the topic. Furthermore, AI research on imaging animal models of infectious diseases offers a unique problem space that can fill in evidence gaps that exist in clinical infectious disease research. Here, we aim to provide a focused assessment of the AI techniques leveraged in the infectious disease imaging research space, highlight the unique challenges, and discuss burgeoning solutions.
Collapse
Affiliation(s)
- Winston T Chu
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
- Integrated Research Facility at Fort Detrick, Division of Clinical Research, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Frederick, Maryland, USA
| | - Syed M S Reza
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
| | - James T Anibal
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
| | - Adam Landa
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
| | - Ian Crozier
- Clinical Monitoring Research Program Directorate, Frederick National Laboratory for Cancer Research sponsored by the National Cancer Institute, Frederick, Maryland, USA
| | - Ulaş Bağci
- Department of Radiology, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, USA
| | - Bradford J Wood
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
- Center for Interventional Oncology, National Cancer Institute, National Institutes of Health, Bethesda, Maryland, USA
| | - Jeffrey Solomon
- Clinical Monitoring Research Program Directorate, Frederick National Laboratory for Cancer Research sponsored by the National Cancer Institute, Frederick, Maryland, USA
| |
Collapse
|
76
|
Du P, Niu X, Li X, Ying C, Zhou Y, He C, Lv S, Liu X, Du W, Wu W. Automatically transferring supervised targets method for segmenting lung lesion regions with CT imaging. BMC Bioinformatics 2023; 24:332. [PMID: 37667214 PMCID: PMC10478337 DOI: 10.1186/s12859-023-05435-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Accepted: 08/02/2023] [Indexed: 09/06/2023] Open
Abstract
BACKGROUND To present an approach that autonomously identifies and selects a self-selective optimal target for the purpose of enhancing learning efficiency to segment infected regions of the lung from chest computed tomography images. We designed a semi-supervised dual-branch framework for training, where the training set consisted of limited expert-annotated data and a large amount of coarsely annotated data that was automatically segmented based on Hu values, which were used to train both strong and weak branches. In addition, we employed the Lovasz scoring method to automatically switch the supervision target in the weak branch and select the optimal target as the supervision object for training. This method can use noisy labels for rapid localization during the early stages of training, and gradually use more accurate targets for supervised training as the training progresses. This approach can utilize a large number of samples that do not require manual annotation, and with the iterations of training, the supervised targets containing noise become closer and closer to the fine-annotated data, which significantly improves the accuracy of the final model. RESULTS The proposed dual-branch deep learning network based on semi-supervision together with cost-effective samples achieved 83.56 ± 12.10 and 82.67 ± 8.04 on our internal and external test benchmarks measured by the mean Dice similarity coefficient (DSC). Through experimental comparison, the DSC value of the proposed algorithm was improved by 13.54% and 2.02% on the internal benchmark and 13.37% and 2.13% on the external benchmark compared with U-Net without extra sample assistance and the mean-teacher frontier algorithm, respectively. CONCLUSION The cost-effective pseudolabeled samples assisted the training of DL models and achieved much better results compared with traditional DL models with manually labeled samples only. Furthermore, our method also achieved the best performance compared with other up-to-date dual branch structures.
Collapse
Affiliation(s)
- Peng Du
- Hangzhou AiSmartIoT Co., Ltd., Hangzhou, Zhejiang, China
| | - Xiaofeng Niu
- Artificial Intelligence Lab, Hangzhou AiSmartVision Co., Ltd., Hangzhou, Zhejiang, China
| | - Xukun Li
- Artificial Intelligence Lab, Hangzhou AiSmartVision Co., Ltd., Hangzhou, Zhejiang, China
| | - Chiqing Ying
- State Key Laboratory for Diagnosis and Treatment of Infectious Diseases, National Clinical Research Center for Infectious Diseases, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, The First Affiliated Hospital, School of Medicine, Zhejiang University, 79 QingChun Road, Hangzhou, 310003, Zhejiang, China
| | - Yukun Zhou
- Artificial Intelligence Lab, Hangzhou AiSmartVision Co., Ltd., Hangzhou, Zhejiang, China
| | - Chang He
- State Key Laboratory for Diagnosis and Treatment of Infectious Diseases, National Clinical Research Center for Infectious Diseases, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, The First Affiliated Hospital, School of Medicine, Zhejiang University, 79 QingChun Road, Hangzhou, 310003, Zhejiang, China
| | - Shuangzhi Lv
- Department of Radiology The First Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, Zhejiang, China
| | - Xiaoli Liu
- State Key Laboratory for Diagnosis and Treatment of Infectious Diseases, National Clinical Research Center for Infectious Diseases, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, The First Affiliated Hospital, School of Medicine, Zhejiang University, 79 QingChun Road, Hangzhou, 310003, Zhejiang, China
| | - Weibo Du
- State Key Laboratory for Diagnosis and Treatment of Infectious Diseases, National Clinical Research Center for Infectious Diseases, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, The First Affiliated Hospital, School of Medicine, Zhejiang University, 79 QingChun Road, Hangzhou, 310003, Zhejiang, China.
| | - Wei Wu
- State Key Laboratory for Diagnosis and Treatment of Infectious Diseases, National Clinical Research Center for Infectious Diseases, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, The First Affiliated Hospital, School of Medicine, Zhejiang University, 79 QingChun Road, Hangzhou, 310003, Zhejiang, China.
| |
Collapse
|
77
|
He A, Wang K, Li T, Du C, Xia S, Fu H. H2Former: An Efficient Hierarchical Hybrid Transformer for Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2763-2775. [PMID: 37018111 DOI: 10.1109/tmi.2023.3264513] [Citation(s) in RCA: 44] [Impact Index Per Article: 22.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Accurate medical image segmentation is of great significance for computer aided diagnosis. Although methods based on convolutional neural networks (CNNs) have achieved good results, it is weak to model the long-range dependencies, which is very important for segmentation task to build global context dependencies. The Transformers can establish long-range dependencies among pixels by self-attention, providing a supplement to the local convolution. In addition, multi-scale feature fusion and feature selection are crucial for medical image segmentation tasks, which is ignored by Transformers. However, it is challenging to directly apply self-attention to CNNs due to the quadratic computational complexity for high-resolution feature maps. Therefore, to integrate the merits of CNNs, multi-scale channel attention and Transformers, we propose an efficient hierarchical hybrid vision Transformer (H2Former) for medical image segmentation. With these merits, the model can be data-efficient for limited medical data regime. The experimental results show that our approach exceeds previous Transformer, CNNs and hybrid methods on three 2D and two 3D medical image segmentation tasks. Moreover, it keeps computational efficiency in model parameters, FLOPs and inference time. For example, H2Former outperforms TransUNet by 2.29% in IoU score on KVASIR-SEG dataset with 30.77% parameters and 59.23% FLOPs.
Collapse
|
78
|
Xu Y, Zheng X, Li Y, Ye X, Cheng H, Wang H, Lyu J. Exploring patient medication adherence and data mining methods in clinical big data: A contemporary review. J Evid Based Med 2023; 16:342-375. [PMID: 37718729 DOI: 10.1111/jebm.12548] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Accepted: 08/30/2023] [Indexed: 09/19/2023]
Abstract
BACKGROUND Increasingly, patient medication adherence data are being consolidated from claims databases and electronic health records (EHRs). Such databases offer an indirect avenue to gauge medication adherence in our data-rich healthcare milieu. The surge in data accessibility, coupled with the pressing need for its conversion to actionable insights, has spotlighted data mining, with machine learning (ML) emerging as a pivotal technique. Nonadherence poses heightened health risks and escalates medical costs. This paper elucidates the synergistic interaction between medical database mining for medication adherence and the role of ML in fostering knowledge discovery. METHODS We conducted a comprehensive review of EHR applications in the realm of medication adherence, leveraging ML techniques. We expounded on the evolution and structure of medical databases pertinent to medication adherence and harnessed both supervised and unsupervised ML paradigms to delve into adherence and its ramifications. RESULTS Our study underscores the applications of medical databases and ML, encompassing both supervised and unsupervised learning, for medication adherence in clinical big data. Databases like SEER and NHANES, often underutilized due to their intricacies, have gained prominence. Employing ML to excavate patient medication logs from these databases facilitates adherence analysis. Such findings are pivotal for clinical decision-making, risk stratification, and scholarly pursuits, aiming to elevate healthcare quality. CONCLUSION Advanced data mining in the era of big data has revolutionized medication adherence research, thereby enhancing patient care. Emphasizing bespoke interventions and research could herald transformative shifts in therapeutic modalities.
Collapse
Affiliation(s)
- Yixian Xu
- Department of Anesthesiology, The First Affiliated Hospital of Jinan University, Guangzhou, China
| | - Xinkai Zheng
- Department of Dermatology, The First Affiliated Hospital of Jinan University, Guangzhou, China
| | - Yuanjie Li
- Planning & Discipline Construction Office, The First Affiliated Hospital of Jinan University, Guangzhou, China
| | - Xinmiao Ye
- Department of Anesthesiology, The First Affiliated Hospital of Jinan University, Guangzhou, China
| | - Hongtao Cheng
- School of Nursing, Jinan University, Guangzhou, China
| | - Hao Wang
- Department of Anesthesiology, The First Affiliated Hospital of Jinan University, Guangzhou, China
| | - Jun Lyu
- Department of Clinical Research, The First Affiliated Hospital of Jinan University, Guangzhou, China
- Guangdong Provincial Key Laboratory of Traditional Chinese Medicine Informatization, Guangzhou, China
| |
Collapse
|
79
|
Xu X, Gao L, Yu L. GOLF-Net: Global and local association fusion network for COVID-19 lung infection segmentation. Comput Biol Med 2023; 164:107361. [PMID: 37595522 DOI: 10.1016/j.compbiomed.2023.107361] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2023] [Revised: 07/27/2023] [Accepted: 08/12/2023] [Indexed: 08/20/2023]
Abstract
The global spread of the Corona Virus Disease 2019 (COVID-19) has caused significant health hazards, leading researchers to explore new methods for detecting lung infections that can supplement molecular diagnosis. Computer tomography (CT) has emerged as a promising tool, although accurately segmenting infected areas in COVID-19 CT scans, especially given the limited available data, remains a challenge for deep learning models. To address this issue, we propose a novel segmentation network, the GlObal and Local association Fusion Network (GOLF-Net), that combines global and local features from Convolutional Neural Networks and Transformers, respectively. Our network leverages attention mechanisms to enhance the correlation and representation of local features, improving the accuracy of infected area segmentation. Additionally, we implement transfer learning to pretrain our network parameters, providing a robust solution to the issue of limited COVID-19 CT data. Our experimental results demonstrate that the segmentation performance of our network exceeds that of most existing models, with a Dice coefficient of 95.09% and an IoU of 92.58%. © 2014 Hosting by Elsevier B.V. All rights reserved.
Collapse
Affiliation(s)
- Xinyu Xu
- School of Computer Science and Technology, Xidian University, China
| | - Lin Gao
- School of Computer Science and Technology, Xidian University, China
| | - Liang Yu
- School of Computer Science and Technology, Xidian University, China.
| |
Collapse
|
80
|
Hu K, Zhang X, Lee D, Xiong D, Zhang Y, Gao X. Boundary-Guided and Region-Aware Network With Global Scale-Adaptive for Accurate Segmentation of Breast Tumors in Ultrasound Images. IEEE J Biomed Health Inform 2023; 27:4421-4432. [PMID: 37310830 DOI: 10.1109/jbhi.2023.3285789] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Breast ultrasound (BUS) image segmentation is a critical procedure in the diagnosis and quantitative analysis of breast cancer. Most existing methods for BUS image segmentation do not effectively utilize the prior information extracted from the images. In addition, breast tumors have very blurred boundaries, various sizes and irregular shapes, and the images have a lot of noise. Thus, tumor segmentation remains a challenge. In this article, we propose a BUS image segmentation method using a boundary-guided and region-aware network with global scale-adaptive (BGRA-GSA). Specifically, we first design a global scale-adaptive module (GSAM) to extract features of tumors of different sizes from multiple perspectives. GSAM encodes the features at the top of the network in both channel and spatial dimensions, which can effectively extract multi-scale context and provide global prior information. Moreover, we develop a boundary-guided module (BGM) for fully mining boundary information. BGM guides the decoder to learn the boundary context by explicitly enhancing the extracted boundary features. Simultaneously, we design a region-aware module (RAM) for realizing the cross-fusion of diverse layers of breast tumor diversity features, which can facilitate the network to improve the learning ability of contextual features of tumor regions. These modules enable our BGRA-GSA to capture and integrate rich global multi-scale context, multi-level fine-grained details, and semantic information to facilitate accurate breast tumor segmentation. Finally, the experimental results on three publicly available datasets show that our model achieves highly effective segmentation of breast tumors even with blurred boundaries, various sizes and shapes, and low contrast.
Collapse
|
81
|
Li W, Cao Y, Wang S, Wan B. Fully feature fusion based neural network for COVID-19 lesion segmentation in CT images. Biomed Signal Process Control 2023; 86:104939. [PMID: 37082352 PMCID: PMC10083211 DOI: 10.1016/j.bspc.2023.104939] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2022] [Revised: 03/07/2023] [Accepted: 04/05/2023] [Indexed: 04/22/2023]
Abstract
Coronavirus Disease 2019 (COVID-19) spreads around the world, seriously affecting people's health. Computed tomography (CT) images contain rich semantic information as an auxiliary diagnosis method. However, the automatic segmentation of COVID-19 lesions in CT images faces several challenges, including inconsistency in size and shape of the lesion, the high variability of the lesion, and the low contrast of pixel values between the lesion and normal tissue surrounding the lesion. Therefore, this paper proposes a Fully Feature Fusion Based Neural Network for COVID-19 Lesion Segmentation in CT Images (F3-Net). F3-Net uses an encoder-decoder architecture. In F3-Net, the Multiple Scale Module (MSM) can sense features of different scales, and Dense Path Module (DPM) is used to eliminate the semantic gap between features. The Attention Fusion Module (AFM) is the attention module, which can better fuse the multiple features. Furthermore, we proposed an improved loss function L o s s C o v i d - B C E that pays more attention to the lesions based on the prior knowledge of the distribution of COVID-19 lesions in the lungs. Finally, we verified the superior performance of F3-Net on a COVID-19 segmentation dataset, experiments demonstrate that the proposed model can segment COVID-19 lesions more accurately in CT images than benchmarks of state of the art.
Collapse
Affiliation(s)
- Wei Li
- Key Laboratory of Intelligent Computing in Medical Image (MIIC), Northeastern University, Ministry of Education, Shenyang, China
| | - Yangyong Cao
- School of Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Shanshan Wang
- School of Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Bolun Wan
- School of Computer Science and Engineering, Northeastern University, Shenyang, China
| |
Collapse
|
82
|
Liu S, Cai T, Tang X, Wang C. MRL-Net: Multi-Scale Representation Learning Network for COVID-19 Lung CT Image Segmentation. IEEE J Biomed Health Inform 2023; 27:4317-4328. [PMID: 37314916 DOI: 10.1109/jbhi.2023.3285936] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Accuracy segmentation of COVID-19 lesions in lung CT images can aid patient screening and diagnosis. However, the blurred, inconsistent shape and location of the lesion area poses a great challenge to this vision task. To tackle this issue, we propose a multi-scale representation learning network (MRL-Net) that integrates CNN with Transformer via two bridge unit: Dual Multi-interaction Attention (DMA) and Dual Boundary Attention (DBA). First, to obtain multi-scale local detailed feature and global contextual information, we combine low-level geometric information and high-level semantic features extracted by CNN and Transformer, respectively. Secondly, for enhanced feature representation, DMA is proposed to fuse the local detailed feature of CNN and the global context information of Transformer. Finally, DBA makes our network focus on the boundary features of the lesion, further enhancing the representational learning. Amounts of experimental results show that MRL-Net is superior to current state-of-the-art methods and achieves better COVID-19 image segmentation performance.
Collapse
|
83
|
Liu Z, Lv Q, Yang Z, Li Y, Lee CH, Shen L. Recent progress in transformer-based medical image analysis. Comput Biol Med 2023; 164:107268. [PMID: 37494821 DOI: 10.1016/j.compbiomed.2023.107268] [Citation(s) in RCA: 23] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Revised: 05/30/2023] [Accepted: 07/16/2023] [Indexed: 07/28/2023]
Abstract
The transformer is primarily used in the field of natural language processing. Recently, it has been adopted and shows promise in the computer vision (CV) field. Medical image analysis (MIA), as a critical branch of CV, also greatly benefits from this state-of-the-art technique. In this review, we first recap the core component of the transformer, the attention mechanism, and the detailed structures of the transformer. After that, we depict the recent progress of the transformer in the field of MIA. We organize the applications in a sequence of different tasks, including classification, segmentation, captioning, registration, detection, enhancement, localization, and synthesis. The mainstream classification and segmentation tasks are further divided into eleven medical image modalities. A large number of experiments studied in this review illustrate that the transformer-based method outperforms existing methods through comparisons with multiple evaluation metrics. Finally, we discuss the open challenges and future opportunities in this field. This task-modality review with the latest contents, detailed information, and comprehensive comparison may greatly benefit the broad MIA community.
Collapse
Affiliation(s)
- Zhaoshan Liu
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Qiujie Lv
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore; School of Intelligent Systems Engineering, Sun Yat-sen University, No. 66, Gongchang Road, Guangming District, 518107, China.
| | - Ziduo Yang
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore; School of Intelligent Systems Engineering, Sun Yat-sen University, No. 66, Gongchang Road, Guangming District, 518107, China.
| | - Yifan Li
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Chau Hung Lee
- Department of Radiology, Tan Tock Seng Hospital, 11 Jalan Tan Tock Seng, Singapore, 308433, Singapore.
| | - Lei Shen
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| |
Collapse
|
84
|
Shen N, Xu T, Bian Z, Huang S, Mu F, Huang B, Xiao Y, Li J. SCANet: A Unified Semi-Supervised Learning Framework for Vessel Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2476-2489. [PMID: 35862338 DOI: 10.1109/tmi.2022.3193150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Automatic subcutaneous vessel imaging with near-infrared (NIR) optical apparatus can promote the accuracy of locating blood vessels, thus significantly contributing to clinical venipuncture research. Though deep learning models have achieved remarkable success in medical image segmentation, they still struggle in the subfield of subcutaneous vessel segmentation due to the scarcity and low-quality of annotated data. To relieve it, this work presents a novel semi-supervised learning framework, SCANet, that achieves accurate vessel segmentation through an alternate training strategy. The SCANet is composed of a multi-scale recurrent neural network that embeds coarse-to-fine features and two auxiliary branches, a consistency decoder and an adversarial learning branch, responsible for strengthening fine-grained details and eliminating differences between ground-truths and predictions, respectively. Equipped with a novel semi-supervised alternate training strategy, the three components work collaboratively, enabling SCANet to accurately segment vessel regions with only a handful of labeled data and abounding unlabeled data. Moreover, to mitigate the shortage of annotated data in this field, we provide a new subcutaneous vessel dataset, VESSEL-NIR. Extensive experiments on a wide variety of tasks, including the segmentation of subcutaneous vessels, retinal vessels, and skin lesions, well demonstrate the superiority and generality of our approach.
Collapse
|
85
|
Zhang M, Wen G, Zhong J, Chen D, Wang C, Huang X, Zhang S. MLP-Like Model With Convolution Complex Transformation for Auxiliary Diagnosis Through Medical Images. IEEE J Biomed Health Inform 2023; 27:4385-4396. [PMID: 37467088 DOI: 10.1109/jbhi.2023.3292312] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/21/2023]
Abstract
Medical images such as facial and tongue images have been widely used for intelligence-assisted diagnosis, which can be regarded as the multi-label classification task for disease location (DL) and disease nature (DN) of biomedical images. Compared with complicated convolutional neural networks and Transformers for this task, recent MLP-like architectures are not only simple and less computationally expensive, but also have stronger generalization capabilities. However, MLP-like models require better input features from the image. Thus, this study proposes a novel convolution complex transformation MLP-like (CCT-MLP) model for the multi-label DL and DN recognition task for facial and tongue images. Notably, the convolutional Tokenizer and multiple convolutional layers are first used to extract the better shallow features from input biomedical images to make up for the loss of spatial information obtained by the simple MLP structure. Subsequently, the Channel-MLP architecture with complex transformations is used to extract deep-level contextual features. In this way, multi-channel features are extracted and mixed to perform the multi-label classification of the input biomedical images. Experimental results on our constructed multi-label facial and tongue image datasets demonstrate that our method outperforms existing methods in terms of both accuracy (Acc) and mean average precision (mAP).
Collapse
|
86
|
Reza SMS, Chu WT, Homayounieh F, Blain M, Firouzabadi FD, Anari PY, Lee JH, Worwa G, Finch CL, Kuhn JH, Malayeri A, Crozier I, Wood BJ, Feuerstein IM, Solomon J. Deep-Learning-Based Whole-Lung and Lung-Lesion Quantification Despite Inconsistent Ground Truth: Application to Computerized Tomography in SARS-CoV-2 Nonhuman Primate Models. Acad Radiol 2023; 30:2037-2045. [PMID: 36966070 PMCID: PMC9968618 DOI: 10.1016/j.acra.2023.02.027] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2023] [Revised: 02/21/2023] [Accepted: 02/22/2023] [Indexed: 03/01/2023]
Abstract
RATIONALE AND OBJECTIVES Animal modeling of infectious diseases such as coronavirus disease 2019 (COVID-19) is important for exploration of natural history, understanding of pathogenesis, and evaluation of countermeasures. Preclinical studies enable rigorous control of experimental conditions as well as pre-exposure baseline and longitudinal measurements, including medical imaging, that are often unavailable in the clinical research setting. Computerized tomography (CT) imaging provides important diagnostic, prognostic, and disease characterization to clinicians and clinical researchers. In that context, automated deep-learning systems for the analysis of CT imaging have been broadly proposed, but their practical utility has been limited. Manual outlining of the ground truth (i.e., lung-lesions) requires accurate distinctions between abnormal and normal tissues that often have vague boundaries and is subject to reader heterogeneity in interpretation. Indeed, this subjectivity is demonstrated as wide inconsistency in manual outlines among experts and from the same expert. The application of deep-learning data-science tools has been less well-evaluated in the preclinical setting, including in nonhuman primate (NHP) models of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) infection/COVID-19, in which the translation of human-derived deep-learning tools is challenging. The automated segmentation of the whole lung and lung lesions provides a potentially standardized and automated method to detect and quantify disease. MATERIALS AND METHODS We used deep-learning-based quantification of the whole lung and lung lesions on CT scans of NHPs exposed to SARS-CoV-2. We proposed a novel multi-model ensemble technique to address the inconsistency in the ground truths for deep-learning-based automated segmentation of the whole lung and lung lesions. Multiple models were obtained by training the convolutional neural network (CNN) on different subsets of the training data instead of having a single model using the entire training dataset. Moreover, we employed a feature pyramid network (FPN), a CNN that provides predictions at different resolution levels, enabling the network to predict objects with wide size variations. RESULTS We achieved an average of 99.4 and 60.2% Dice coefficients for whole-lung and lung-lesion segmentation, respectively. The proposed multi-model FPN outperformed well-accepted methods U-Net (50.5%), V-Net (54.5%), and Inception (53.4%) for the challenging lesion-segmentation task. We show the application of segmentation outputs for longitudinal quantification of lung disease in SARS-CoV-2-exposed and mock-exposed NHPs. CONCLUSION Deep-learning methods should be optimally characterized for and targeted specifically to preclinical research needs in terms of impact, automation, and dynamic quantification independently from purely clinical applications.
Collapse
Affiliation(s)
- Syed M S Reza
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Winston T Chu
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Fatemeh Homayounieh
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Maxim Blain
- Center for Interventional Oncology, Radiology and Imaging Sciences, NIH Clinical Center and National Cancer Institute, Center for Cancer Research, National Institutes of Health, Bethesda, Maryland
| | - Fatemeh D Firouzabadi
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Pouria Y Anari
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Ji Hyun Lee
- Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Gabriella Worwa
- Integrated Research Facility at Fort Detrick, Division of Clinical Research, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Frederick, Maryland
| | - Courtney L Finch
- Integrated Research Facility at Fort Detrick, Division of Clinical Research, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Frederick, Maryland
| | - Jens H Kuhn
- Integrated Research Facility at Fort Detrick, Division of Clinical Research, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Frederick, Maryland
| | - Ashkan Malayeri
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland
| | - Ian Crozier
- Clinical Monitoring Research Program Directorate, Frederick National Laboratory for Cancer Research, Frederick, Maryland
| | - Bradford J Wood
- Center for Interventional Oncology, Radiology and Imaging Sciences, NIH Clinical Center and National Cancer Institute, Center for Cancer Research, National Institutes of Health, Bethesda, Maryland
| | - Irwin M Feuerstein
- Integrated Research Facility at Fort Detrick, Division of Clinical Research, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Frederick, Maryland
| | - Jeffrey Solomon
- Clinical Monitoring Research Program Directorate, Frederick National Laboratory for Cancer Research, Frederick, Maryland.
| |
Collapse
|
87
|
Santosh KC, GhoshRoy D, Nakarmi S. A Systematic Review on Deep Structured Learning for COVID-19 Screening Using Chest CT from 2020 to 2022. Healthcare (Basel) 2023; 11:2388. [PMID: 37685422 PMCID: PMC10486542 DOI: 10.3390/healthcare11172388] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 08/16/2023] [Accepted: 08/22/2023] [Indexed: 09/10/2023] Open
Abstract
The emergence of the COVID-19 pandemic in Wuhan in 2019 led to the discovery of a novel coronavirus. The World Health Organization (WHO) designated it as a global pandemic on 11 March 2020 due to its rapid and widespread transmission. Its impact has had profound implications, particularly in the realm of public health. Extensive scientific endeavors have been directed towards devising effective treatment strategies and vaccines. Within the healthcare and medical imaging domain, the application of artificial intelligence (AI) has brought significant advantages. This study delves into peer-reviewed research articles spanning the years 2020 to 2022, focusing on AI-driven methodologies for the analysis and screening of COVID-19 through chest CT scan data. We assess the efficacy of deep learning algorithms in facilitating decision making processes. Our exploration encompasses various facets, including data collection, systematic contributions, emerging techniques, and encountered challenges. However, the comparison of outcomes between 2020 and 2022 proves intricate due to shifts in dataset magnitudes over time. The initiatives aimed at developing AI-powered tools for the detection, localization, and segmentation of COVID-19 cases are primarily centered on educational and training contexts. We deliberate on their merits and constraints, particularly in the context of necessitating cross-population train/test models. Our analysis encompassed a review of 231 research publications, bolstered by a meta-analysis employing search keywords (COVID-19 OR Coronavirus) AND chest CT AND (deep learning OR artificial intelligence OR medical imaging) on both the PubMed Central Repository and Web of Science platforms.
Collapse
Affiliation(s)
- KC Santosh
- 2AI: Applied Artificial Intelligence Research Lab, Vermillion, SD 57069, USA
| | - Debasmita GhoshRoy
- School of Automation, Banasthali Vidyapith, Tonk 304022, Rajasthan, India;
| | - Suprim Nakarmi
- Department of Computer Science, University of South Dakota, Vermillion, SD 57069, USA;
| |
Collapse
|
88
|
Chen M, Yi S, Yang M, Yang Z, Zhang X. UNet segmentation network of COVID-19 CT images with multi-scale attention. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:16762-16785. [PMID: 37920033 DOI: 10.3934/mbe.2023747] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/04/2023]
Abstract
In recent years, the global outbreak of COVID-19 has posed an extremely serious life-safety risk to humans, and in order to maximize the diagnostic efficiency of physicians, it is extremely valuable to investigate the methods of lesion segmentation in images of COVID-19. Aiming at the problems of existing deep learning models, such as low segmentation accuracy, poor model generalization performance, large model parameters and difficult deployment, we propose an UNet segmentation network integrating multi-scale attention for COVID-19 CT images. Specifically, the UNet network model is utilized as the base network, and the structure of multi-scale convolutional attention is proposed in the encoder stage to enhance the network's ability to capture multi-scale information. Second, a local channel attention module is proposed to extract spatial information by modeling local relationships to generate channel domain weights, to supplement detailed information about the target region to reduce information redundancy and to enhance important information. Moreover, the network model encoder segment uses the Meta-ACON activation function to avoid the overfitting phenomenon of the model and to improve the model's representational ability. A large number of experimental results on publicly available mixed data sets show that compared with the current mainstream image segmentation algorithms, the pro-posed method can more effectively improve the accuracy and generalization performance of COVID-19 lesions segmentation and provide help for medical diagnosis and analysis.
Collapse
Affiliation(s)
- Mingju Chen
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| | - Sihang Yi
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| | - Mei Yang
- Zigong Third People's Hospital, Zigong 643000, China
| | - Zhiwen Yang
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| | - Xingyue Zhang
- School of Automation and Information Engineering, Sichuan University of Science & Engineering, Yibin 644002, China
- Artificial Intelligence Key Laboratory of Sichuan Province, Sichuan University of Science & Engineering, Yibin 644002, China
| |
Collapse
|
89
|
He Z, Wong ANN, Yoo JS. Co-ERA-Net: Co-Supervision and Enhanced Region Attention for Accurate Segmentation in COVID-19 Chest Infection Images. Bioengineering (Basel) 2023; 10:928. [PMID: 37627813 PMCID: PMC10451793 DOI: 10.3390/bioengineering10080928] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Revised: 07/31/2023] [Accepted: 08/03/2023] [Indexed: 08/27/2023] Open
Abstract
Accurate segmentation of infected lesions in chest images remains a challenging task due to the lack of utilization of lung region information, which could serve as a strong location hint for infection. In this paper, we propose a novel segmentation network Co-ERA-Net for infections in chest images that leverages lung region information by enhancing supervised information and fusing multi-scale lung region and infection information at different levels. To achieve this, we introduce a Co-supervision scheme incorporating lung region information to guide the network to accurately locate infections within the lung region. Furthermore, we design an Enhanced Region Attention Module (ERAM) to highlight regions with a high probability of infection by incorporating infection information into the lung region information. The effectiveness of the proposed scheme is demonstrated using COVID-19 CT and X-ray datasets, with the results showing that the proposed schemes and modules are promising. Based on the baseline, the Co-supervision scheme, when integrated with lung region information, improves the Dice coefficient by 7.41% and 2.22%, and the IoU by 8.20% and 3.00% in CT and X-ray datasets respectively. Moreover, when this scheme is combined with the Enhanced Region Attention Module, the Dice coefficient sees further improvement of 14.24% and 2.97%, with the IoU increasing by 28.64% and 4.49% for the same datasets. In comparison with existing approaches across various datasets, our proposed method achieves better segmentation performance in all main metrics and exhibits the best generalization and comprehensive performance.
Collapse
Affiliation(s)
| | | | - Jung Sun Yoo
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China; (Z.H.); (A.N.N.W.)
| |
Collapse
|
90
|
Saha S, Dutta S, Goswami B, Nandi D. ADU-Net: An Attention Dense U-Net based deep supervised DNN for automated lesion segmentation of COVID-19 from chest CT images. Biomed Signal Process Control 2023; 85:104974. [PMID: 37122956 PMCID: PMC10121143 DOI: 10.1016/j.bspc.2023.104974] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Revised: 04/01/2023] [Accepted: 04/15/2023] [Indexed: 05/02/2023]
Abstract
An automatic method for qualitative and quantitative evaluation of chest Computed Tomography (CT) images is essential for diagnosing COVID-19 patients. We aim to develop an automated COVID-19 prediction framework using deep learning. We put forth a novel Deep Neural Network (DNN) composed of an attention-based dense U-Net with deep supervision for COVID-19 lung lesion segmentation from chest CT images. We incorporate dense U-Net where convolution kernel size 5×5 is used instead of 3×3. The dense and transition blocks are introduced to implement a densely connected network on each encoder level. Also, the attention mechanism is applied between the encoder, skip connection, and decoder. These are used to keep both the high and low-level features efficiently. The deep supervision mechanism creates secondary segmentation maps from the features. Deep supervision combines secondary supervision maps from various resolution levels and produces a better final segmentation map. The trained artificial DNN model takes the test data at its input and generates a prediction output for COVID-19 lesion segmentation. The proposed model has been applied to the MedSeg COVID-19 chest CT segmentation dataset. Data pre-processing methods help the training process and improve performance. We compare the performance of the proposed DNN model with state-of-the-art models by computing the well-known metrics: dice coefficient, Jaccard coefficient, accuracy, specificity, sensitivity, and precision. As a result, the proposed model outperforms the state-of-the-art models. This new model may be considered an efficient automated screening system for COVID-19 diagnosis and can potentially improve patient health care and management system.
Collapse
Affiliation(s)
- Sanjib Saha
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, 713209, West Bengal, India
- Department of Computer Science and Engineering, Dr. B. C. Roy Engineering College, Durgapur, 713206, West Bengal, India
| | - Subhadeep Dutta
- Department of Computer Science and Engineering, Dr. B. C. Roy Engineering College, Durgapur, 713206, West Bengal, India
| | - Biswarup Goswami
- Department of Respiratory Medicine, Health and Family Welfare, Government of West Bengal, Kolkata, 700091, West Bengal, India
| | - Debashis Nandi
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, 713209, West Bengal, India
| |
Collapse
|
91
|
Zeng LL, Gao K, Hu D, Feng Z, Hou C, Rong P, Wang W. SS-TBN: A Semi-Supervised Tri-Branch Network for COVID-19 Screening and Lesion Segmentation. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:10427-10442. [PMID: 37022260 DOI: 10.1109/tpami.2023.3240886] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Insufficient annotated data and minor lung lesions pose big challenges for computed tomography (CT)-aided automatic COVID-19 diagnosis at an early outbreak stage. To address this issue, we propose a Semi-Supervised Tri-Branch Network (SS-TBN). First, we develop a joint TBN model for dual-task application scenarios of image segmentation and classification such as CT-based COVID-19 diagnosis, in which pixel-level lesion segmentation and slice-level infection classification branches are simultaneously trained via lesion attention, and individual-level diagnosis branch aggregates slice-level outputs for COVID-19 screening. Second, we propose a novel hybrid semi-supervised learning method to make full use of unlabeled data, combining a new double-threshold pseudo labeling method specifically designed to the joint model and a new inter-slice consistency regularization method specifically tailored to CT images. Besides two publicly available external datasets, we collect internal and our own external datasets including 210,395 images (1,420 cases versus 498 controls) from ten hospitals. Experimental results show that the proposed method achieves state-of-the-art performance in COVID-19 classification with limited annotated data even if lesions are subtle, and that segmentation results promote interpretability for diagnosis, suggesting the potential of the SS-TBN in early screening in insufficient labeled data situations at the early stage of a pandemic outbreak like COVID-19.
Collapse
|
92
|
Zaeri N. Artificial intelligence and machine learning responses to COVID-19 related inquiries. J Med Eng Technol 2023; 47:301-320. [PMID: 38625639 DOI: 10.1080/03091902.2024.2321846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2021] [Accepted: 02/18/2024] [Indexed: 04/17/2024]
Abstract
Researchers and scientists can use computational-based models to turn linked data into useful information, aiding in disease diagnosis, examination, and viral containment due to recent artificial intelligence and machine learning breakthroughs. In this paper, we extensively study the role of artificial intelligence and machine learning in delivering efficient responses to the COVID-19 pandemic almost four years after its start. In this regard, we examine a large number of critical studies conducted by various academic and research communities from multiple disciplines, as well as practical implementations of artificial intelligence algorithms that suggest potential solutions in investigating different COVID-19 decision-making scenarios. We identify numerous areas where artificial intelligence and machine learning can impact this context, including diagnosis (using chest X-ray imaging and CT imaging), severity, tracking, treatment, and the drug industry. Furthermore, we analyse the dilemma's limits, restrictions, and hazards.
Collapse
Affiliation(s)
- Naser Zaeri
- Faculty of Computer Studies, Arab Open University, Kuwait
| |
Collapse
|
93
|
Yang Y, Zhang L, Ren L, Zhou L, Wang X. SuperMini-seg: An ultra lightweight network for COVID-19 lung infection segmentation from CT images. Biomed Signal Process Control 2023; 85:104896. [PMID: 36998783 PMCID: PMC10028361 DOI: 10.1016/j.bspc.2023.104896] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2022] [Revised: 01/31/2023] [Accepted: 03/18/2023] [Indexed: 03/24/2023]
Abstract
The automatic segmentation of lung lesions from COVID-19 computed tomography (CT) images is helpful in establishing a quantitative model to diagnose and treat COVID-19. To this end, this study proposes a lightweight segmentation network called the SuperMini-Seg. We propose a new module called the transformer parallel convolution module (TPCB), which introduces both transformer and convolution operations in one module. SuperMini-seg adopts the structure of a double-branch parallel to downsample the image and designs a gated attention mechanism in the middle of the two parallel branches. At the same time, the attentive hierarchical spatial pyramid (AHSP) module and criss-cross attention module are adopted, and more than 100K parameters are present in the model. At the same time, the model is scalable, and the parameter quantity of SuperMini-seg-V2 reaches more than 70K. Compared with other advanced methods, the segmentation accuracy was almost reached the state-of-art method. The calculation efficiency was high, which is convenient for practical deployment.
Collapse
Affiliation(s)
- Yuan Yang
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, School of Medicine and Engineering, No. 37 Xueyuan Road, Haidian District, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Ministry of Industry and Information Technology, No. 37 Xueyuan Road, Haidian District, Beijing, China
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| | - Lin Zhang
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, School of Medicine and Engineering, No. 37 Xueyuan Road, Haidian District, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Ministry of Industry and Information Technology, No. 37 Xueyuan Road, Haidian District, Beijing, China
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| | - Lei Ren
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, School of Medicine and Engineering, No. 37 Xueyuan Road, Haidian District, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Ministry of Industry and Information Technology, No. 37 Xueyuan Road, Haidian District, Beijing, China
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| | - Longfei Zhou
- Duke University School of Medicine, 2424 Erwin Rd, Suite 302, Durham, USA
| | - Xiaohan Wang
- School of Automation Science and Electrical Engineering, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing, China
| |
Collapse
|
94
|
Yuan C, Song S, Yang J, Sun Y, Yang B, Xu L. Pulmonary arteries segmentation from CT images using PA-Net with attention module and contour loss. Med Phys 2023; 50:4887-4898. [PMID: 36752170 DOI: 10.1002/mp.16265] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2022] [Revised: 01/03/2023] [Accepted: 01/18/2023] [Indexed: 02/09/2023] Open
Abstract
BACKGROUND Pulmonary embolism is a kind of cardiovascular disease that threatens human life and health. Since pulmonary embolism exists in the pulmonary artery, improving the segmentation accuracy of pulmonary artery is the key to the diagnosis of pulmonary embolism. Traditional medical image segmentation methods have limited effectiveness in pulmonary artery segmentation. In recent years, deep learning methods have been gradually adopted to solve complex problems in the field of medical image segmentation. PURPOSE Due to the irregular shape of the pulmonary artery and the adjacent-complex tissues, the accuracy of the existing pulmonary artery segmentation methods based on deep learning needs to be improved. Therefore, the purpose of this paper is to develop a segmentation network, which can obtain higher segmentation accuracy and further improve the diagnosis effect. METHODS In this study, the pulmonary artery segmentation performance from the network model and loss function is improved, proposing a pulmonary artery segmentation network (PA-Net) to segment the pulmonary artery region from 2D CT images. Reverse Attention and edge attention are used to enhance the expression ability of the boundary. In addition, to better use feature information, the channel attention module is introduced in the decoder to highlight the important channel features and suppress the unimportant channels. Due to blurred boundaries, pixels near the boundaries of the pulmonary artery may be difficult to segment. Therefore, a new contour loss function based on the active contour model is proposed in this study to segment the target region by assigning dynamic weights to false positive and false negative regions and accurately predict the boundary structure. RESULTS The experimental results show that the segmentation accuracy of this proposed method is significantly improved in comparison with state-of-the-art segmentation methods, and the Dice coefficient is 0.938 ± 0.035, which is also confirmed from the 3D reconstruction results. CONCLUSIONS Our proposed method can accurately segment pulmonary artery structure. This new development will provide the possibility for further rapid diagnosis of pulmonary artery diseases such as pulmonary embolism. Code is available at https://github.com/Yuanyan19/PA-Net.
Collapse
Affiliation(s)
- Chengyan Yuan
- School of Science, Northeastern University, Shenyang, China
| | - Shuni Song
- School of Data and Computer Science, Guangdong Peizheng College, Guangzhou, China
| | - Jinzhong Yang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yu Sun
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of Northern Theater Command, Shenyang, China
- Key Laboratory of Cardiovascular Imaging and Research of Liaoning Province, Shenyang, China
| | - Benqiang Yang
- Department of Radiology, General Hospital of Northern Theater Command, Shenyang, China
- Key Laboratory of Cardiovascular Imaging and Research of Liaoning Province, Shenyang, China
| | - Lisheng Xu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Medical Image Computing, Ministry of Education, Shenyang, Liaoning, China
- Neusoft Research of Intelligent Healthcare Technology, Co. Ltd., Shenyang, Liaoning, China
| |
Collapse
|
95
|
Pan X, Zhu H, Du J, Hu G, Han B, Jia Y. MS-DCANet: A Novel Segmentation Network For Multi-Modality COVID-19 Medical Images. J Multidiscip Healthc 2023; 16:2023-2043. [PMID: 37489133 PMCID: PMC10363353 DOI: 10.2147/jmdh.s417068] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Accepted: 07/10/2023] [Indexed: 07/26/2023] Open
Abstract
Aim The Coronavirus Disease 2019 (COVID-19) pandemic has increased the public health burden and brought profound disaster to humans. For the particularity of the COVID-19 medical images with blurred boundaries, low contrast and different infection sites, some researchers have improved the accuracy by adding more complexity. Also, they overlook the complexity of lesions, which hinder their ability to capture the relationship between segmentation sites and the background, as well as the edge contours and global context. However, increasing the computational complexity, parameters and inference speed is unfavorable for model transfer from laboratory to clinic. A perfect segmentation network needs to balance the above three factors completely. To solve the above issues, this paper propose a symmetric automatic segmentation framework named MS-DCANet. We introduce Tokenized MLP block, a novel attention scheme that uses a shift-window mechanism to conditionally fuse local and global features to get more continuous boundaries and spatial positioning capabilities. It has greater understanding of irregular lesion contours. MS-DCANet also uses several Dual Channel blocks and a Res-ASPP block to improve the ability to recognize small targets. On multi-modality COVID-19 tasks, MS-DCANet achieved state-of-the-art performance compared with other baselines. It can well trade off the accuracy and complexity. To prove the strong generalization ability of our proposed model, we apply it to other tasks (ISIC 2018 and BAA) and achieve satisfactory results. Patients The X-ray dataset from Qatar University which contains 3379 cases for light, normal and heavy COVID-19 lung infection. The CT dataset contains the scans of 10 patient cases with COVID-19, a total of 1562 CT axial slices. The BAA dataset is obtained from the hospital and includes 387 original images. The ISIC 2018 dataset is from the International Skin Imaging Collaborative (ISIC) containing 2594 original images. Results The proposed MS-DCANet achieved evaluation metrics (MIOU) of 73.86, 97.26, 89.54, and 79.54 on the four datasets, respectively, far exceeding other current state-of-the art baselines. Conclusion The proposed MS-DCANet can help clinicians to automate the diagnosis of COVID-19 patients with different symptoms.
Collapse
Affiliation(s)
- Xiaoyu Pan
- College of Medical Informatics, Chongqing Medical University, Chongqing, People’s Republic of China
| | - Huazheng Zhu
- College of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing, People’s Republic of China
| | - Jinglong Du
- College of Medical Informatics, Chongqing Medical University, Chongqing, People’s Republic of China
| | - Guangtao Hu
- College of Medical Informatics, Chongqing Medical University, Chongqing, People’s Republic of China
| | - Baoru Han
- College of Medical Informatics, Chongqing Medical University, Chongqing, People’s Republic of China
| | - Yuanyuan Jia
- College of Medical Informatics, Chongqing Medical University, Chongqing, People’s Republic of China
| |
Collapse
|
96
|
Jin Q, Hou H, Zhang G, Li Z. FEGNet: A Feedback Enhancement Gate Network for Automatic Polyp Segmentation. IEEE J Biomed Health Inform 2023; 27:3420-3430. [PMID: 37126617 DOI: 10.1109/jbhi.2023.3272168] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
Regular colonoscopy is an effective way to prevent colorectal cancer by detecting colorectal polyps. Automatic polyp segmentation significantly aids clinicians in precisely locating polyp areas for further diagnosis. However, polyp segmentation is a challenge problem, since polyps appear in a variety of shapes, sizes and textures, and they tend to have ambiguous boundaries. In this paper, we propose a U-shaped model named Feedback Enhancement Gate Network (FEGNet) for accurate polyp segmentation to overcome these difficulties. Specifically, for the high-level features, we design a novel Recurrent Gate Module (RGM) based on the feedback mechanism, which can refine attention maps without any additional parameters. RGM consists of Feature Aggregation Attention Gate (FAAG) and Multi-Scale Module (MSM). FAAG can aggregate context and feedback information, and MSM is applied for capturing multi-scale information, which is critical for the segmentation task. In addition, we propose a straightforward but effective edge extraction module to detect boundaries of polyps for low-level features, which is used to guide the training of early features. In our experiments, quantitative and qualitative evaluations show that the proposed FEGNet has achieved the best results in polyp segmentation compared to other state-of-the-art models on five colonoscopy datasets.
Collapse
|
97
|
Nikulin P, Zschaeck S, Maus J, Cegla P, Lombardo E, Furth C, Kaźmierska J, Rogasch JMM, Holzgreve A, Albert NL, Ferentinos K, Strouthos I, Hajiyianni M, Marschner SN, Belka C, Landry G, Cholewinski W, Kotzerke J, Hofheinz F, van den Hoff J. A convolutional neural network with self-attention for fully automated metabolic tumor volume delineation of head and neck cancer in
[
18
F]FDG PET/CT. Eur J Nucl Med Mol Imaging 2023; 50:2751-2766. [PMID: 37079128 PMCID: PMC10317885 DOI: 10.1007/s00259-023-06197-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Accepted: 03/14/2023] [Indexed: 04/21/2023]
Abstract
PURPOSE PET-derived metabolic tumor volume (MTV) and total lesion glycolysis of the primary tumor are known to be prognostic of clinical outcome in head and neck cancer (HNC). Including evaluation of lymph node metastases can further increase the prognostic value of PET but accurate manual delineation and classification of all lesions is time-consuming and prone to interobserver variability. Our goal, therefore, was development and evaluation of an automated tool for MTV delineation/classification of primary tumor and lymph node metastases in PET/CT investigations of HNC patients. METHODS Automated lesion delineation was performed with a residual 3D U-Net convolutional neural network (CNN) incorporating a multi-head self-attention block. 698[ 18 F]FDG PET/CT scans from 3 different sites and 5 public databases were used for network training and testing. An external dataset of 181[ 18 F]FDG PET/CT scans from 2 additional sites was employed to assess the generalizability of the network. In these data, primary tumor and lymph node (LN) metastases were interactively delineated and labeled by two experienced physicians. Performance of the trained network models was assessed by 5-fold cross-validation in the main dataset and by pooling results from the 5 developed models in the external dataset. The Dice similarity coefficient (DSC) for individual delineation tasks and the primary tumor/metastasis classification accuracy were used as evaluation metrics. Additionally, a survival analysis using univariate Cox regression was performed comparing achieved group separation for manual and automated delineation, respectively. RESULTS In the cross-validation experiment, delineation of all malignant lesions with the trained U-Net models achieves DSC of 0.885, 0.805, and 0.870 for primary tumor, LN metastases, and the union of both, respectively. In external testing, the DSC reaches 0.850, 0.724, and 0.823 for primary tumor, LN metastases, and the union of both, respectively. The voxel classification accuracy was 98.0% and 97.9% in cross-validation and external data, respectively. Univariate Cox analysis in the cross-validation and the external testing reveals that manually and automatically derived total MTVs are both highly prognostic with respect to overall survival, yielding essentially identical hazard ratios (HR) (HR man = 1.9 ;p < 0.001 vs.HR cnn = 1.8 ;p < 0.001 in cross-validation andHR man = 1.8 ;p = 0.011 vs.HR cnn = 1.9 ;p = 0.004 in external testing). CONCLUSION To the best of our knowledge, this work presents the first CNN model for successful MTV delineation and lesion classification in HNC. In the vast majority of patients, the network performs satisfactory delineation and classification of primary tumor and lymph node metastases and only rarely requires more than minimal manual correction. It is thus able to massively facilitate study data evaluation in large patient groups and also does have clear potential for supervised clinical application.
Collapse
Affiliation(s)
- Pavel Nikulin
- Helmholtz-Zentrum Dresden-Rossendorf, PET Center, Institute of Radiopharmaceutical Cancer Research, Bautzner Landstrasse 400, 01328, Dresden, Germany.
| | - Sebastian Zschaeck
- Department of Radiation Oncology, Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Berlin, Germany
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Jens Maus
- Helmholtz-Zentrum Dresden-Rossendorf, PET Center, Institute of Radiopharmaceutical Cancer Research, Bautzner Landstrasse 400, 01328, Dresden, Germany
| | - Paulina Cegla
- Department of Nuclear Medicine, Greater Poland Cancer Centre, Poznan, Poland
| | - Elia Lombardo
- Department of Radiation Oncology, University Hospital, Ludwig-Maximilians-University (LMU) Munich, Munich, Germany
| | - Christian Furth
- Department of Nuclear Medicine, Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Berlin, Germany
| | - Joanna Kaźmierska
- Electroradiology Department, University of Medical Sciences, Poznan, Poland
- Radiotherapy Department II, Greater Poland Cancer Centre, Poznan, Poland
| | - Julian M M Rogasch
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Berlin, Germany
- Department of Nuclear Medicine, Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Berlin, Germany
| | - Adrien Holzgreve
- Department of Nuclear Medicine, University Hospital, Ludwig-Maximilians-University (LMU) Munich, Munich, Germany
| | - Nathalie L Albert
- Department of Nuclear Medicine, University Hospital, Ludwig-Maximilians-University (LMU) Munich, Munich, Germany
| | - Konstantinos Ferentinos
- Department of Radiation Oncology, German Oncology Center, European University Cyprus, Limassol, Cyprus
| | - Iosif Strouthos
- Department of Radiation Oncology, German Oncology Center, European University Cyprus, Limassol, Cyprus
| | - Marina Hajiyianni
- Department of Radiation Oncology, Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt-Universität zu Berlin, Berlin, Germany
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Sebastian N Marschner
- Department of Radiation Oncology, University Hospital, Ludwig-Maximilians-University (LMU) Munich, Munich, Germany
| | - Claus Belka
- Department of Radiation Oncology, University Hospital, Ludwig-Maximilians-University (LMU) Munich, Munich, Germany
- German Cancer Consortium (DKTK), Partner Site Munich, Munich, Germany
| | - Guillaume Landry
- Department of Radiation Oncology, University Hospital, Ludwig-Maximilians-University (LMU) Munich, Munich, Germany
| | - Witold Cholewinski
- Department of Nuclear Medicine, Greater Poland Cancer Centre, Poznan, Poland
- Electroradiology Department, University of Medical Sciences, Poznan, Poland
| | - Jörg Kotzerke
- Department of Nuclear Medicine, University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany
| | - Frank Hofheinz
- Helmholtz-Zentrum Dresden-Rossendorf, PET Center, Institute of Radiopharmaceutical Cancer Research, Bautzner Landstrasse 400, 01328, Dresden, Germany
| | - Jörg van den Hoff
- Helmholtz-Zentrum Dresden-Rossendorf, PET Center, Institute of Radiopharmaceutical Cancer Research, Bautzner Landstrasse 400, 01328, Dresden, Germany
- Department of Nuclear Medicine, University Hospital Carl Gustav Carus, Technische Universität Dresden, Dresden, Germany
| |
Collapse
|
98
|
Zeng Y, Zeng P, Shen S, Liang W, Li J, Zhao Z, Zhang K, Shen C. DCTR U-Net: automatic segmentation algorithm for medical images of nasopharyngeal cancer in the context of deep learning. Front Oncol 2023; 13:1190075. [PMID: 37546396 PMCID: PMC10402756 DOI: 10.3389/fonc.2023.1190075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2023] [Accepted: 05/30/2023] [Indexed: 08/08/2023] Open
Abstract
Nasopharyngeal carcinoma (NPC) is a malignant tumor that occurs in the wall of the nasopharyngeal cavity and is prevalent in Southern China, Southeast Asia, North Africa, and the Middle East. According to studies, NPC is one of the most common malignant tumors in Hainan, China, and it has the highest incidence rate among otorhinolaryngological malignancies. We proposed a new deep learning network model to improve the segmentation accuracy of the target region of nasopharyngeal cancer. Our model is based on the U-Net-based network, to which we add Dilated Convolution Module, Transformer Module, and Residual Module. The new deep learning network model can effectively solve the problem of restricted convolutional fields of perception and achieve global and local multi-scale feature fusion. In our experiments, the proposed network was trained and validated using 10-fold cross-validation based on the records of 300 clinical patients. The results of our network were evaluated using the dice similarity coefficient (DSC) and the average symmetric surface distance (ASSD). The DSC and ASSD values are 0.852 and 0.544 mm, respectively. With the effective combination of the Dilated Convolution Module, Transformer Module, and Residual Module, we significantly improved the segmentation performance of the target region of the NPC.
Collapse
Affiliation(s)
- Yan Zeng
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
- ChinaPersonnel Department, Hainan Medical University, Haikou, China
| | - PengHui Zeng
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
| | - ShaoDong Shen
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
| | - Wei Liang
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
| | - Jun Li
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
| | - Zhe Zhao
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
| | - Kun Zhang
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
- School of Information Science and Technology, Hainan Normal University, Haikou, China
| | - Chong Shen
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of Information and Communication Engineering, Hainan University, Haikou, China
| |
Collapse
|
99
|
Li Q, Chen M, Geng J, Adamu MJ, Guan X. High-Resolution Network with Dynamic Convolution and Coordinate Attention for Classification of Chest X-ray Images. Diagnostics (Basel) 2023; 13:2165. [PMID: 37443559 DOI: 10.3390/diagnostics13132165] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2023] [Revised: 06/12/2023] [Accepted: 06/21/2023] [Indexed: 07/15/2023] Open
Abstract
The development of automatic chest X-ray (CXR) disease classification algorithms is significant for diagnosing thoracic diseases. Owing to the characteristics of lesions in CXR images, including high similarity in appearance of the disease, varied sizes, and different occurrence locations, most existing convolutional neural network-based methods have insufficient feature extraction for thoracic lesions and struggle to adapt to changes in lesion size and location. To address these issues, this study proposes a high-resolution classification network with dynamic convolution and coordinate attention (HRCC-Net). In the method, this study suggests a parallel multi-resolution network in which a high-resolution branch acquires essential detailed features of the lesion and multi-resolution feature swapping and fusion to obtain multiple receptive fields to extract complicated disease features adequately. Furthermore, this study proposes dynamic convolution to enhance the network's ability to represent multi-scale information to accommodate lesions of diverse scales. In addition, this study introduces a coordinate attention mechanism, which enables automatic focus on pathologically relevant regions and capturing the variations in lesion location. The proposed method is evaluated on ChestX-ray14 and CheXpert datasets. The average AUC (area under ROC curve) values reach 0.845 and 0.913, respectively, indicating this method's advantages compared with the currently available methods. Meanwhile, with its specificity and sensitivity to measure the performance of medical diagnostic systems, the network can improve diagnostic efficiency while reducing the rate of misdiagnosis. The proposed algorithm has great potential for thoracic disease diagnosis and treatment.
Collapse
Affiliation(s)
- Qiang Li
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| | - Mingyu Chen
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| | - Jingjing Geng
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| | | | - Xin Guan
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| |
Collapse
|
100
|
Butt MJ, Malik AK, Qamar N, Yar S, Malik AJ, Rauf U. A Survey on COVID-19 Data Analysis Using AI, IoT, and Social Media. SENSORS (BASEL, SWITZERLAND) 2023; 23:5543. [PMID: 37420714 DOI: 10.3390/s23125543] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/30/2023] [Revised: 06/04/2023] [Accepted: 06/07/2023] [Indexed: 07/09/2023]
Abstract
Coronaviruses are a well-established and deadly group of viruses that cause illness in both humans and animals. The novel type of this virus group, named COVID-19, was firstly reported in December 2019, and, with the passage of time, coronavirus has spread to almost all parts of the world. Coronavirus has been the cause of millions of deaths around the world. Furthermore, many countries are struggling with COVID-19 and have experimented with various kinds of vaccines to eliminate the deadly virus and its variants. This survey deals with COVID-19 data analysis and its impact on human social life. Data analysis and information related to coronavirus can greatly help scientists and governments in controlling the spread and symptoms of the deadly coronavirus. In this survey, we cover many areas of discussion related to COVID-19 data analysis, such as how artificial intelligence, along with machine learning, deep learning, and IoT, have worked together to fight against COVID-19. We also discuss artificial intelligence and IoT techniques used to forecast, detect, and diagnose patients of the novel coronavirus. Moreover, this survey also describes how fake news, doctored results, and conspiracy theories were spread over social media sites, such as Twitter, by applying various social network analysis and sentimental analysis techniques. A comprehensive comparative analysis of existing techniques has also been conducted. In the end, the Discussion section presents different data analysis techniques, provides future directions for research, and suggests general guidelines for handling coronavirus, as well as changing work and life conditions.
Collapse
Affiliation(s)
- Muhammad Junaid Butt
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Ahmad Kamran Malik
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Nafees Qamar
- School of Health and Behavioral Sciences, Bryant University, Smithfield, RI 02917, USA
| | - Samad Yar
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Arif Jamal Malik
- Department of Software Engineering, Foundation University, Islamabad 44000, Pakistan
| | - Usman Rauf
- Department of Mathematics and Computer Science, Mercy College, Dobbs Ferry, NY 10522, USA
| |
Collapse
|