1
|
Chen K, Liu J, Wan R, Ho-Fun Lee V, Vardhanabhuti V, Yan H, Li H. Unsupervised Domain Adaptation for Low-Dose CT Reconstruction via Bayesian Uncertainty Alignment. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:8525-8539. [PMID: 38985555 DOI: 10.1109/tnnls.2024.3409573] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/12/2024]
Abstract
Low-dose computed tomography (LDCT) image reconstruction techniques can reduce patient radiation exposure while maintaining acceptable imaging quality. Deep learning (DL) is widely used in this problem, but the performance of testing data (also known as target domain) is often degraded in clinical scenarios due to the variations that were not encountered in training data (also known as source domain). Unsupervised domain adaptation (UDA) of LDCT reconstruction has been proposed to solve this problem through distribution alignment. However, existing UDA methods fail to explore the usage of uncertainty quantification, which is crucial for reliable intelligent medical systems in clinical scenarios with unexpected variations. Moreover, existing direct alignment for different patients would lead to content mismatch issues. To address these issues, we propose to leverage a probabilistic reconstruction framework to conduct a joint discrepancy minimization between source and target domains in both the latent and image spaces. In the latent space, we devise a Bayesian uncertainty alignment to reduce the epistemic gap between the two domains. This approach reduces the uncertainty level of target domain data, making it more likely to render well-reconstructed results on target domains. In the image space, we propose a sharpness-aware distribution alignment (SDA) to achieve a match of second-order information, which can ensure that the reconstructed images from the target domain have similar sharpness to normal-dose CT (NDCT) images from the source domain. Experimental results on two simulated datasets and one clinical low-dose imaging dataset show that our proposed method outperforms other methods in quantitative and visualized performance.
Collapse
|
2
|
Dang Y, Ma W, Luo X, Wang H. CAD-Unet: A capsule network-enhanced Unet architecture for accurate segmentation of COVID-19 lung infections from CT images. Med Image Anal 2025; 103:103583. [PMID: 40306203 DOI: 10.1016/j.media.2025.103583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2024] [Revised: 02/17/2025] [Accepted: 04/04/2025] [Indexed: 05/02/2025]
Abstract
Since the outbreak of the COVID-19 pandemic in 2019, medical imaging has emerged as a primary modality for diagnosing COVID-19 pneumonia. In clinical settings, the segmentation of lung infections from computed tomography images enables rapid and accurate quantification and diagnosis of COVID-19. Segmentation of COVID-19 infections in the lungs poses a formidable challenge, primarily due to the indistinct boundaries and limited contrast presented by ground glass opacity manifestations. Moreover, the confounding similarity among infiltrates, lung tissues, and lung walls further complicates this segmentation task. To address these challenges, this paper introduces a novel deep network architecture, called CAD-Unet, for segmenting COVID-19 lung infections. In this architecture, capsule networks are incorporated into the existing Unet framework. Capsule networks represent a novel type of network architecture that differs from traditional convolutional neural networks. They utilize vectors for information transfer among capsules, facilitating the extraction of intricate lesion spatial information. Additionally, we design a capsule encoder path and establish a coupling path between the unet encoder and the capsule encoder. This design maximizes the complementary advantages of both network structures while achieving efficient information fusion. Finally, extensive experiments are conducted on four publicly available datasets, encompassing binary segmentation tasks and multi-class segmentation tasks. The experimental results demonstrate the superior segmentation performance of the proposed model. The code has been released at: https://github.com/AmanoTooko-jie/CAD-Unet.
Collapse
Affiliation(s)
- Yijie Dang
- School of Information Engineering, Ningxia University, Yinchuan, 750021, Ningxia, China
| | - Weijun Ma
- School of Information Engineering, Ningxia University, Yinchuan, 750021, Ningxia, China; Ningxia Key Laboratory of Artificial Intelligence and Information Security for Channeling Computing Resources from the East to the West, Ningxia University, Yinchuan, 750021, Ningxia, China.
| | - Xiaohu Luo
- School of Mathematics and Computer Science, Ningxia Normal University, Guyuan, 756099, China
| | - Huaizhu Wang
- School of Advanced Interdisciplinary Studies, Ningxia University, Zhongwei, 755000, China
| |
Collapse
|
3
|
Agha S, Nazir S, Kaleem M, Najeeb F, Talat R. Performance evaluation of reduced complexity deep neural networks. PLoS One 2025; 20:e0319859. [PMID: 40112278 PMCID: PMC11925470 DOI: 10.1371/journal.pone.0319859] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 02/10/2025] [Indexed: 03/22/2025] Open
Abstract
Deep Neural Networks (DNN) have achieved state-of-the-art performance in medical image classification and are increasingly being used for disease diagnosis. However, these models are quite complex and that necessitates the need to reduce the model complexity for their use in low-power edge applications that are becoming common. The model complexity reduction techniques in most cases comprise of time-consuming operations and are often associated with a loss of model performance in proportion to the model size reduction. In this paper, we propose a simplified model complexity reduction technique based on reducing the number of channels for any DNN and demonstrate the complexity reduction approaches for the ResNet-50 model integration in low-power devices. The model performance of the proposed models was evaluated for multiclass classification of CXR images, as normal, pneumonia, and COVID-19 classes. We demonstrate successive size reductions down to 75%, 87%, and 93% reduction with an acceptable classification performance reduction of 0.5%, 0.5%, and 0.8% respectively. We also provide the results for the model generalization, and visualization with Grad-CAM at an acceptable performance and interpretable level. In addition, a theoretical VLSI architecture for the best performing architecture has been presented.
Collapse
Affiliation(s)
- Shahrukh Agha
- Department of Electrical and Computer Engineering, COMSATS University, Islamabad, Pakistan
| | - Sajid Nazir
- School of Computing, Engineering and Built Environment, Glasgow Caledonian University, Glasgow, Scotland, United Kingdom
| | - Mohammad Kaleem
- Department of Electrical and Computer Engineering, COMSATS University, Islamabad, Pakistan
| | - Faisal Najeeb
- Department of Electrical and Computer Engineering, COMSATS University, Islamabad, Pakistan
| | - Rehab Talat
- Islamic International Medical College, Riphah International University, Rawalpindi, Pakistan
| |
Collapse
|
4
|
Zhou Y, Xu X, Song J, Shen F, Shen HT. MSFlow: Multiscale Flow-Based Framework for Unsupervised Anomaly Detection. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:2437-2450. [PMID: 38194384 DOI: 10.1109/tnnls.2023.3344118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2024]
Abstract
Unsupervised anomaly detection (UAD) attracts a lot of research interest and drives widespread applications, where only anomaly-free samples are available for training. Some UAD applications intend to locate the anomalous regions further even without any anomaly information. Although the absence of anomalous samples and annotations deteriorates the UAD performance, an inconspicuous, yet powerful statistics model, the normalizing flows, is appropriate for anomaly detection (AD) and localization in an unsupervised fashion. The flow-based probabilistic models, only trained on anomaly-free data, can efficiently distinguish unpredictable anomalies by assigning them much lower likelihoods than normal data. Nevertheless, the size variation of unpredictable anomalies introduces another inconvenience to the flow-based methods for high-precision AD and localization. To generalize the anomaly size variation, we propose a novel multiscale flow-based framework (MSFlow) composed of asymmetrical parallel flows followed by a fusion flow to exchange multiscale perceptions. Moreover, different multiscale aggregation strategies are adopted for image-wise AD and pixel-wise anomaly localization according to the discrepancy between them. The proposed MSFlow is evaluated on three AD datasets, significantly outperforming existing methods. Notably, on the challenging MVTec AD benchmark, our MSFlow achieves a new state-of-the-art (SOTA) with a detection AUORC score of up to 99.7%, localization AUCROC score of 98.8% and PRO score of 97.1%.
Collapse
|
5
|
Feng J, Luo H, Ming R. Pointer meters recognition method in the wild based on innovative deep learning techniques. Sci Rep 2025; 15:845. [PMID: 39755689 DOI: 10.1038/s41598-024-81248-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2024] [Accepted: 11/25/2024] [Indexed: 01/06/2025] Open
Abstract
This study presents a novel approach to identifying meters and their pointers in modern industrial scenarios using deep learning. We developed a neural network model that can detect gauges and one or more of their pointers on low-quality images. We use an encoder network, jump connections, and a modified Convolutional Block Attention Module (CBAM) to detect gauge panels and pointer keypoints in images. We also combine the output of the decoder network and the output of the improved CBAM as inputs to the Object Heatmap-Scalarmap Module to find pointer tip heat map peaks and predict pointer pointing. The method proposed in this paper is compared with several deep learning networks. The experimental results show that the model in this paper has the highest recognition correctness, with an average precision of 0.95 and 0.763 for Object Keypoint Similarity and Vector Direction Similarity, and an average recall of 0.951 and 0.856 in the test set, respectively, and achieves the best results in terms of efficiency and accuracy achieve the best trade-off, and performs well in recognizing multiple pointer targets. This demonstrates its robustness in real scenarios and provides a new idea for recognizing pointers in low-quality images more efficiently and accurately in complex industrial scenarios.
Collapse
Affiliation(s)
- Jiajun Feng
- College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou, 350002, China
| | - Haibo Luo
- College of Computer and Data Science, Minjiang University, Fuzhou, 350018, China.
| | - Rui Ming
- College of Computer and Data Science, Minjiang University, Fuzhou, 350018, China
| |
Collapse
|
6
|
Momeni Pour Z, Beheshti Shirazi AA. Identifying COVID-19-Infected Segments in Lung CT Scan Through Two Innovative Artificial Intelligence-Based Transformer Models. ARCHIVES OF ACADEMIC EMERGENCY MEDICINE 2024; 13:e21. [PMID: 39958958 PMCID: PMC11829223 DOI: 10.22037/aaemj.v13i1.2515] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 02/18/2025]
Abstract
Introduction Automatic systems based on Artificial intelligence (AI) algorithms have made significant advancements across various domains, most notably in the field of medicine. This study introduces a novel approach for identifying COVID-19-infected regions in lung computed tomography (CT) scan through the development of two innovative models. Methods In this study we used the Squeeze and Excitation based UNet TRansformers (SE-UNETR) and the Squeeze and Excitation based High-Quality Resolution Swin Transformer Network (SE-HQRSTNet), to develop two three-dimensional segmentation networks for identifying COVID-19-infected regions in lung CT scan. The SE-UNETR model is structured as a 3D UNet architecture with an encoder component built on Vision Transformers (ViTs). This model processes 3D patches directly as input and learns sequential representations of the volumetric data. The encoder connects to the decoder using skip connections, ultimately producing the final semantic segmentation output. Conversely, the SE-HQRSTNet model incorporates High-Resolution Networks (HRNet), Swin Transformer modules, and Squeeze and Excitation (SE) blocks. This architecture is designed to generate features at multiple resolutions, utilizing Multi-Resolution Feature Fusion (MRFF) blocks to effectively integrate semantic features across various scales. The proposed networks were evaluated using a 5-fold cross-validation methodology, along with data augmentation techniques, applied to the COVID-19-CT-Seg and MosMed datasets. Results experimental results demonstrate that the Dice value for the infection masks within the COVID-19-CT-Seg dataset improved by 3.81% and 4.84% with the SE-UNETR and SE-HQRSTNet models, respectively, compared to previously reported work. Furthermore, the Dice value for the MosMed dataset increased from 66.8% to 69.35% and 70.89% for the SE-UNETR and SE-HQRSTNet models, respectively. Conclusion These improvements indicate that the proposed models exhibit superior efficiency and performance relative to existing methodologies.
Collapse
Affiliation(s)
- Zeinab Momeni Pour
- Department of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | | |
Collapse
|
7
|
Oliveira ADS, Costa MGF, Costa JPGF, Costa Filho CFF. Comparing Different Data Partitioning Strategies for Segmenting Areas Affected by COVID-19 in CT Scans. Diagnostics (Basel) 2024; 14:2791. [PMID: 39767152 PMCID: PMC11674714 DOI: 10.3390/diagnostics14242791] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2024] [Revised: 12/09/2024] [Accepted: 12/10/2024] [Indexed: 01/11/2025] Open
Abstract
BACKGROUND/OBJECTIVES According to the World Health Organization, the gold standard for diagnosing COVID-19 is the Reverse Transcription Polymerase Chain Reaction (RT-PCR) test. However, to confirm the diagnosis in patients who have negative results but still show symptoms, imaging tests, especially computed tomography (CT), are used. In this study, using convolutional neural networks, we compared the following topics using manual and automatic lung segmentation methods: (1) the performance of an automatic segmentation of COVID-19 areas using two strategies for data partitioning, CT scans, and slice strategies; (2) the performance of an automatic segmentation method of COVID-19 when there was interobserver agreement between two groups of radiologists; and (3) the performance of the area affected by COVID-19. METHODS Two datasets and two deep neural network architectures are used to evaluate the automatic segmentation of lungs and COVID-19 areas. The performance of the U-Net architecture is compared with the performance of a new architecture proposed by the research group. RESULTS With automatic lung segmentation, the Dice metrics for the segmentation of the COVID-19 area were 73.01 ± 9.47% and 84.66 ± 5.41% for the CT-scan strategy and slice strategy, respectively. With manual lung segmentation, the Dice metrics for the automatic segmentation of COVID-19 were 74.47 ± 9.94% and 85.35 ± 5.41% for the CT-scan and the slice strategy, respectively. CONCLUSIONS The main conclusions were as follows: COVID-19 segmentation was slightly better for the slice strategy than for the CT-scan strategy; a comparison of the performance of the automatic COVID-19 segmentation and the interobserver agreement, in a group of 7 CT scans, revealed that there was no statistically significant difference between any metric.
Collapse
Affiliation(s)
- Anne de Souza Oliveira
- R&D Center in Electronic and Information Technology, Federal University of Amazonas, Manaus 69077-000, Brazil; (A.d.S.O.); (M.G.F.C.)
| | - Marly Guimarães Fernandes Costa
- R&D Center in Electronic and Information Technology, Federal University of Amazonas, Manaus 69077-000, Brazil; (A.d.S.O.); (M.G.F.C.)
| | | | | |
Collapse
|
8
|
Wang C, Xu R, Xu S, Meng W, Xiao J, Zhang X. Accurate Lung Nodule Segmentation With Detailed Representation Transfer and Soft Mask Supervision. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:18381-18393. [PMID: 37824321 DOI: 10.1109/tnnls.2023.3315271] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/14/2023]
Abstract
Accurate lung lesion segmentation from computed tomography (CT) images is crucial to the analysis and diagnosis of lung diseases, such as COVID-19 and lung cancer. However, the smallness and variety of lung nodules and the lack of high-quality labeling make the accurate lung nodule segmentation difficult. To address these issues, we first introduce a novel segmentation mask named " soft mask," which has richer and more accurate edge details description and better visualization, and develop a universal automatic soft mask annotation pipeline to deal with different datasets correspondingly. Then, a novel network with detailed representation transfer and soft mask supervision (DSNet) is proposed to process the input low-resolution images of lung nodules into high-quality segmentation results. Our DSNet contains a special detailed representation transfer module (DRTM) for reconstructing the detailed representation to alleviate the small size of lung nodules images and an adversarial training framework with soft mask for further improving the accuracy of segmentation. Extensive experiments validate that our DSNet outperforms other state-of-the-art methods for accurate lung nodule segmentation, and has strong generalization ability in other accurate medical segmentation tasks with competitive results. Besides, we provide a new challenging lung nodules segmentation dataset for further studies (https://drive.google.com/file/d/15NNkvDTb_0Ku0IoPsNMHezJRTH1Oi1wm/view?usp=sharing).
Collapse
|
9
|
Memon K, Yahya N, Yusoff MZ, Remli R, Mustapha AWMM, Hashim H, Ali SSA, Siddiqui S. Edge Computing for AI-Based Brain MRI Applications: A Critical Evaluation of Real-Time Classification and Segmentation. SENSORS (BASEL, SWITZERLAND) 2024; 24:7091. [PMID: 39517987 PMCID: PMC11548207 DOI: 10.3390/s24217091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/05/2024] [Revised: 10/29/2024] [Accepted: 10/31/2024] [Indexed: 11/16/2024]
Abstract
Medical imaging plays a pivotal role in diagnostic medicine with technologies like Magnetic Resonance Imagining (MRI), Computed Tomography (CT), Positron Emission Tomography (PET), and ultrasound scans being widely used to assist radiologists and medical experts in reaching concrete diagnosis. Given the recent massive uplift in the storage and processing capabilities of computers, and the publicly available big data, Artificial Intelligence (AI) has also started contributing to improving diagnostic radiology. Edge computing devices and handheld gadgets can serve as useful tools to process medical data in remote areas with limited network and computational resources. In this research, the capabilities of multiple platforms are evaluated for the real-time deployment of diagnostic tools. MRI classification and segmentation applications developed in previous studies are used for testing the performance using different hardware and software configurations. Cost-benefit analysis is carried out using a workstation with a NVIDIA Graphics Processing Unit (GPU), Jetson Xavier NX, Raspberry Pi 4B, and Android phone, using MATLAB, Python, and Android Studio. The mean computational times for the classification app on the PC, Jetson Xavier NX, and Raspberry Pi are 1.2074, 3.7627, and 3.4747 s, respectively. On the low-cost Android phone, this time is observed to be 0.1068 s using the Dynamic Range Quantized TFLite version of the baseline model, with slight degradation in accuracy. For the segmentation app, the times are 1.8241, 5.2641, 6.2162, and 3.2023 s, respectively, when using JPEG inputs. The Jetson Xavier NX and Android phone stand out as the best platforms due to their compact size, fast inference times, and affordability.
Collapse
Affiliation(s)
- Khuhed Memon
- Department of Electrical and Electronics Engineering, Universiti Teknologi PETRONAS (UTP), Seri Iskandar 32610, Perak, Malaysia; (K.M.); (M.Z.Y.)
| | - Norashikin Yahya
- Department of Electrical and Electronics Engineering, Universiti Teknologi PETRONAS (UTP), Seri Iskandar 32610, Perak, Malaysia; (K.M.); (M.Z.Y.)
| | - Mohd Zuki Yusoff
- Department of Electrical and Electronics Engineering, Universiti Teknologi PETRONAS (UTP), Seri Iskandar 32610, Perak, Malaysia; (K.M.); (M.Z.Y.)
| | - Rabani Remli
- Faculty of Medicine, Hospital Canselor Tuanku Muhriz UKM, Cheras 56000, Kuala Lumpur, Malaysia; (R.R.); (A.-W.M.M.M.)
| | | | - Hilwati Hashim
- Department of Radiology, Faculty of Medicine, Universiti Teknologi MARA, Sungai Buloh 47000, Selangor, Malaysia;
| | - Syed Saad Azhar Ali
- Aerospace Engineering Department, Interdisciplinary Research Center for Smart Mobility & Logistics, King Fahd University of Petroleum & Minerals, Dhahran 31261, Saudi Arabia;
| | - Shahabuddin Siddiqui
- Department of Radiology, Pakistan Institute of Medical Sciences, Islamabad 44000, Pakistan;
| |
Collapse
|
10
|
Chen Y, Lu W, Qin X, Wang J, Xie X. MetaFed: Federated Learning Among Federations With Cyclic Knowledge Distillation for Personalized Healthcare. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:16671-16682. [PMID: 37506019 DOI: 10.1109/tnnls.2023.3297103] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/30/2023]
Abstract
Federated learning (FL) has attracted increasing attention to building models without accessing raw user data, especially in healthcare. In real applications, different federations can seldom work together due to possible reasons such as data heterogeneity and distrust/inexistence of the central server. In this article, we propose a novel framework called MetaFed to facilitate trustworthy FL between different federations. MetaFed obtains a personalized model for each federation without a central server via the proposed cyclic knowledge distillation. Specifically, MetaFed treats each federation as a meta distribution and aggregates knowledge of each federation in a cyclic manner. The training is split into two parts: common knowledge accumulation and personalization. Comprehensive experiments on seven benchmarks demonstrate that MetaFed without a server achieves better accuracy compared with state-of-the-art methods [e.g., 10%+ accuracy improvement compared with the baseline for physical activity monitoring dataset (PAMAP2)] with fewer communication costs. More importantly, MetaFed shows remarkable performance in real-healthcare-related applications.
Collapse
|
11
|
Yu P, Wang C, Zhang H, Zheng G, Jia C, Liu Z, Wang Q, Mu Y, Yang X, Mao N, Song X. Deep learning-based automatic pipeline system for predicting lateral cervical lymph node metastasis in patients with papillary thyroid carcinoma using computed tomography: A multi-center study. Chin J Cancer Res 2024; 36:545-561. [PMID: 39539818 PMCID: PMC11555202 DOI: 10.21147/j.issn.1000-9604.2024.05.07] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2024] [Accepted: 10/18/2024] [Indexed: 11/16/2024] Open
Abstract
Objective The assessment of lateral lymph node metastasis (LLNM) in patients with papillary thyroid carcinoma (PTC) holds great significance. This study aims to develop and evaluate a deep learning-based automatic pipeline system (DLAPS) for diagnosing LLNM in PTC using computed tomography (CT). Methods A total of 1,266 lateral lymph nodes (LLNs) from 519 PTC patients who underwent CT examinations from January 2019 to November 2022 were included and divided into training and validation set, internal test set, pooled external test set, and prospective test set. The DLAPS consists of an auto-segmentation network based on RefineNet model and a classification network based on ensemble model (ResNet, Xception, and DenseNet). The performance of the DLAPS was compared with that of manually segmented DL models, the clinical model, and Node Reporting and Data System (Node-RADS). The improvement of radiologists' diagnostic performance under the DLAPS-assisted strategy was explored. In addition, bulk RNA-sequencing was conducted based on 12 LLNs to reveal the underlying biological basis of the DLAPS. Results The DLAPS yielded good performance with area under the receiver operating characteristic curve (AUC) of 0.872, 0.910, and 0.822 in the internal, pooled external, and prospective test sets, respectively. The DLAPS significantly outperformed clinical models (AUC 0.731, P<0.001) and Node-RADS (AUC 0.602, P<0.001) in the internal test set. Moreover, the performance of the DLAPS was comparable to that of the manually segmented deep learning (DL) model with AUCs ranging 0.814-0.901 in three test sets. Furthermore, the DLAPS-assisted strategy improved the performance of radiologists and enhanced inter-observer consistency. In clinical situations, the rate of unnecessary LLN dissection decreased from 33.33% to 7.32%. Furthermore, the DLAPS was associated with the cell-cell conjunction in the microenvironment. Conclusions Using CT images from PTC patients, the DLAPS could effectively segment and classify LLNs non-invasively, and this system had a good generalization ability and clinical applicability.
Collapse
Affiliation(s)
- Pengyi Yu
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
| | - Cai Wang
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
- Cardiff China Medical Research Collaborative, Cardiff University School of Medicine, Cardiff CF14 4XN, UK
| | - Haicheng Zhang
- Big data and Artificial Intelligence Laboratory, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
| | - Guibin Zheng
- Department of Thyroid Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
| | - Chuanliang Jia
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
| | - Zhonglu Liu
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
| | - Qi Wang
- Big data and Artificial Intelligence Laboratory, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
| | - Yakui Mu
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
| | - Xin Yang
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
| | - Ning Mao
- Big data and Artificial Intelligence Laboratory, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Department of Radiology, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
| | - Xicheng Song
- Department of Otorhinolaryngology, Head and Neck Surgery, Yantai Yuhuangding Hospital, Qingdao University, Yantai 264000, China
- Shandong Provincial Key Laboratory of Neuroimmune Interaction and Regulation, Yantai 264000, China
- Shandong Provincial Clinical Research Center for Otorhinolaryngologic Diseases, Yantai 264000, China
- Yantai Key Laboratory of Otorhinolaryngologic Diseases, Yantai 264000, China
| |
Collapse
|
12
|
Dong G, Wang Z, Chen Y, Sun Y, Song H, Liu L, Cui H. An efficient segment anything model for the segmentation of medical images. Sci Rep 2024; 14:19425. [PMID: 39169054 PMCID: PMC11339323 DOI: 10.1038/s41598-024-70288-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2024] [Accepted: 08/14/2024] [Indexed: 08/23/2024] Open
Abstract
This paper introduces the efficient medical-images-aimed segment anything model (EMedSAM), addressing the high computational demands and limited adaptability of using SAM for medical image segmentation tasks. We present a novel, compact image encoder, DD-TinyViT, designed to enhance segmentation efficiency through an innovative parameter tuning method called med-adapter. The lightweight DD-TinyViT encoder is derived from the well-known ViT-H using a decoupled distillation approach.The segmentation and recognition capabilities of EMedSAM for specific structures are improved by med-adapter, which dynamically adjusts the model parameters specifically for medical imaging. We conducted extensive testing on EMedSAM using the public FLARE 2022 dataset and datasets from the First Hospital of Zhejiang University School of Medicine. The results demonstrate that our model outperforms existing state-of-the-art models in both multi-organ and lung segmentation tasks.
Collapse
Affiliation(s)
- Guanliang Dong
- School of Information Engineering, Huzhou University, Huzhou, 313000, China
| | - Zhangquan Wang
- College of Information Science and Technology, Zhejiang Shuren University, Hangzhou, 310015, China.
| | - Yourong Chen
- College of Information Science and Technology, Zhejiang Shuren University, Hangzhou, 310015, China
| | - Yuliang Sun
- College of Information Science and Technology, Zhejiang Shuren University, Hangzhou, 310015, China
| | - Hongbo Song
- College of Information Science and Technology, Zhejiang Shuren University, Hangzhou, 310015, China
| | - Liyuan Liu
- Department of Decision and System Sciences, Saint Joseph's University, Philadelphia, 19131, USA
| | - Haidong Cui
- Department of Breast Surgery, First Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, 310003, China
| |
Collapse
|
13
|
Lin Q, Tan W, Cai S, Yan B, Li J, Zhong Y. Lesion-Decoupling-Based Segmentation With Large-Scale Colon and Esophageal Datasets for Early Cancer Diagnosis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:11142-11156. [PMID: 37028330 DOI: 10.1109/tnnls.2023.3248804] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Lesions of early cancers often show flat, small, and isochromatic characteristics in medical endoscopy images, which are difficult to be captured. By analyzing the differences between the internal and external features of the lesion area, we propose a lesion-decoupling-based segmentation (LDS) network for assisting early cancer diagnosis. We introduce a plug-and-play module called self-sampling similar feature disentangling module (FDM) to obtain accurate lesion boundaries. Then, we propose a feature separation loss (FSL) function to separate pathological features from normal ones. Moreover, since physicians make diagnoses with multimodal data, we propose a multimodal cooperative segmentation network with two different modal images as input: white-light images (WLIs) and narrowband images (NBIs). Our FDM and FSL show a good performance for both single-modal and multimodal segmentations. Extensive experiments on five backbones prove that our FDM and FSL can be easily applied to different backbones for a significant lesion segmentation accuracy improvement, and the maximum increase of mean Intersection over Union (mIoU) is 4.58. For colonoscopy, we can achieve up to mIoU of 91.49 on our Dataset A and 84.41 on the three public datasets. For esophagoscopy, mIoU of 64.32 is best achieved on the WLI dataset and 66.31 on the NBI dataset.
Collapse
|
14
|
Wang Z, Yang L, Sun T, Yan W. Fusion PCAM R-CNN of Automatic Segmentation for Magnetic Flux Leakage Defects. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:11424-11435. [PMID: 37027265 DOI: 10.1109/tnnls.2023.3261363] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Magnetic leakage detection technology plays an important role in the long-oil pipeline. Automatic segmentation of defecting images is crucial for the detection of magnetic flux leakage (MFL) works. At present, accurate segmentation for small defects has always been a difficult problem. In contrast to the state-of-the-art MFL detection methodologies based on convolution neural network (CNN), an optimization method is devised in our study by integrating mask region-based CNN (Mask R-CNN) and information entropy constraint (IEC). To be precise, the principal component analysis (PCA) is utilized to improve the feature learning and network segmentation ability of the convolution kernel. The similarity constraint rule of information entropy is proposed to be inserted into the convolution layer in the Mask R-CNN network. The Mask R-CNN optimizes the convolutional kernel with similar weights or higher similarity, meanwhile, the PCA network reduces the dimension of the feature image to reconstruct the original feature vector. As such, the feature extraction of MFL defects is optimized in the convolution check. The research results can be applied in the field of MFL detection.
Collapse
|
15
|
Agarwal S, Saxena S, Carriero A, Chabert GL, Ravindran G, Paul S, Laird JR, Garg D, Fatemi M, Mohanty L, Dubey AK, Singh R, Fouda MM, Singh N, Naidu S, Viskovic K, Kukuljan M, Kalra MK, Saba L, Suri JS. COVLIAS 3.0: cloud-based quantized hybrid UNet3+ deep learning for COVID-19 lesion detection in lung computed tomography. Front Artif Intell 2024; 7:1304483. [PMID: 39006802 PMCID: PMC11240867 DOI: 10.3389/frai.2024.1304483] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Accepted: 06/10/2024] [Indexed: 07/16/2024] Open
Abstract
Background and novelty When RT-PCR is ineffective in early diagnosis and understanding of COVID-19 severity, Computed Tomography (CT) scans are needed for COVID diagnosis, especially in patients having high ground-glass opacities, consolidations, and crazy paving. Radiologists find the manual method for lesion detection in CT very challenging and tedious. Previously solo deep learning (SDL) was tried but they had low to moderate-level performance. This study presents two new cloud-based quantized deep learning UNet3+ hybrid (HDL) models, which incorporated full-scale skip connections to enhance and improve the detections. Methodology Annotations from expert radiologists were used to train one SDL (UNet3+), and two HDL models, namely, VGG-UNet3+ and ResNet-UNet3+. For accuracy, 5-fold cross-validation protocols, training on 3,500 CT scans, and testing on unseen 500 CT scans were adopted in the cloud framework. Two kinds of loss functions were used: Dice Similarity (DS) and binary cross-entropy (BCE). Performance was evaluated using (i) Area error, (ii) DS, (iii) Jaccard Index, (iii) Bland-Altman, and (iv) Correlation plots. Results Among the two HDL models, ResNet-UNet3+ was superior to UNet3+ by 17 and 10% for Dice and BCE loss. The models were further compressed using quantization showing a percentage size reduction of 66.76, 36.64, and 46.23%, respectively, for UNet3+, VGG-UNet3+, and ResNet-UNet3+. Its stability and reliability were proved by statistical tests such as the Mann-Whitney, Paired t-Test, Wilcoxon test, and Friedman test all of which had a p < 0.001. Conclusion Full-scale skip connections of UNet3+ with VGG and ResNet in HDL framework proved the hypothesis showing powerful results improving the detection accuracy of COVID-19.
Collapse
Affiliation(s)
- Sushant Agarwal
- Advanced Knowledge Engineering Center, GBTI, Roseville, CA, United States
- Department of CSE, PSIT, Kanpur, India
| | | | - Alessandro Carriero
- Department of Radiology, “Maggiore della Carità” Hospital, University of Piemonte Orientale (UPO), Novara, Italy
| | | | - Gobinath Ravindran
- Department of Civil Engineering, SR University, Warangal, Telangana, India
| | - Sudip Paul
- Department of Biomedical Engineering, NEHU, Shillong, India
| | - John R. Laird
- Heart and Vascular Institute, Adventist Health St. Helena, St. Helena, CA, United States
| | - Deepak Garg
- School of CS and AI, SR University, Warangal, Telangana, India
| | - Mostafa Fatemi
- Department of Physiology and Biomedical Engineering, Mayo Clinic College of Medicine and Science, Rochester, MN, United States
| | - Lopamudra Mohanty
- Department of Computer Science, ABES Engineering College, Ghaziabad, UP, India
- Department of Computer science, Bennett University, Greater Noida, UP, India
| | - Arun K. Dubey
- Bharati Vidyapeeth’s College of Engineering, New Delhi, India
| | - Rajesh Singh
- Division of Research and Innovation, Uttaranchal Institute of Technology, Uttaranchal University, Dehradun, India
| | - Mostafa M. Fouda
- Department of ECE, Idaho State University, Pocatello, ID, United States
| | - Narpinder Singh
- Department of Food Science and Technology, Graphic Era Deemed to be University, Dehradun, India
| | - Subbaram Naidu
- Department of EE, University of Minnesota, Duluth, MN, United States
| | | | - Melita Kukuljan
- Department of Interventional and Diagnostic Radiology, Clinical Hospital Center Rijeka, Rijeka, Croatia
| | - Manudeep K. Kalra
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
| | - Luca Saba
- Department of Radiology, A.O.U., Cagliari, Italy
| | - Jasjit S. Suri
- Department of ECE, Idaho State University, Pocatello, ID, United States
- Department of Computer Science, Graphic Era Deemed to Be University, Dehradun, Uttarakhand, India
- Symbiosis Institute of Technology, Nagpur Campus, Symbiosis International (Deemed University), Pune, India
- Stroke and Monitoring Division, AtheroPoint LLC, Roseville, CA, United States
| |
Collapse
|
16
|
Bougourzi F, Dornaika F, Distante C, Taleb-Ahmed A. D-TrAttUnet: Toward hybrid CNN-transformer architecture for generic and subtle segmentation in medical images. Comput Biol Med 2024; 176:108590. [PMID: 38763066 DOI: 10.1016/j.compbiomed.2024.108590] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Revised: 04/16/2024] [Accepted: 05/09/2024] [Indexed: 05/21/2024]
Abstract
Over the past two decades, machine analysis of medical imaging has advanced rapidly, opening up significant potential for several important medical applications. As complicated diseases increase and the number of cases rises, the role of machine-based imaging analysis has become indispensable. It serves as both a tool and an assistant to medical experts, providing valuable insights and guidance. A particularly challenging task in this area is lesion segmentation, a task that is challenging even for experienced radiologists. The complexity of this task highlights the urgent need for robust machine learning approaches to support medical staff. In response, we present our novel solution: the D-TrAttUnet architecture. This framework is based on the observation that different diseases often target specific organs. Our architecture includes an encoder-decoder structure with a composite Transformer-CNN encoder and dual decoders. The encoder includes two paths: the Transformer path and the Encoders Fusion Module path. The Dual-Decoder configuration uses two identical decoders, each with attention gates. This allows the model to simultaneously segment lesions and organs and integrate their segmentation losses. To validate our approach, we performed evaluations on the Covid-19 and Bone Metastasis segmentation tasks. We also investigated the adaptability of the model by testing it without the second decoder in the segmentation of glands and nuclei. The results confirmed the superiority of our approach, especially in Covid-19 infections and the segmentation of bone metastases. In addition, the hybrid encoder showed exceptional performance in the segmentation of glands and nuclei, solidifying its role in modern medical image analysis.
Collapse
Affiliation(s)
- Fares Bougourzi
- Junia, UMR 8520, CNRS, Centrale Lille, University of Polytechnique Hauts-de-France, 59000 Lille, France.
| | - Fadi Dornaika
- University of the Basque Country UPV/EHU, San Sebastian, Spain; IKERBASQUE, Basque Foundation for Science, Bilbao, Spain.
| | - Cosimo Distante
- Institute of Applied Sciences and Intelligent Systems, National Research Council of Italy, 73100 Lecce, Italy.
| | - Abdelmalik Taleb-Ahmed
- Université Polytechnique Hauts-de-France, Université de Lille, CNRS, Valenciennes, 59313, Hauts-de-France, France.
| |
Collapse
|
17
|
Qiu Y, Liu Y, Li S, Xu J. MiniSeg: An Extremely Minimum Network Based on Lightweight Multiscale Learning for Efficient COVID-19 Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:8570-8584. [PMID: 37015641 DOI: 10.1109/tnnls.2022.3230821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
The rapid spread of the new pandemic, i.e., coronavirus disease 2019 (COVID-19), has severely threatened global health. Deep-learning-based computer-aided screening, e.g., COVID-19 infected area segmentation from computed tomography (CT) image, has attracted much attention by serving as an adjunct to increase the accuracy of COVID-19 screening and clinical diagnosis. Although lesion segmentation is a hot topic, traditional deep learning methods are usually data-hungry with millions of parameters, easy to overfit under limited available COVID-19 training data. On the other hand, fast training/testing and low computational cost are also necessary for quick deployment and development of COVID-19 screening systems, but traditional methods are usually computationally intensive. To address the above two problems, we propose MiniSeg, a lightweight model for efficient COVID-19 segmentation from CT images. Our efforts start with the design of an attentive hierarchical spatial pyramid (AHSP) module for lightweight, efficient, effective multiscale learning that is essential for image segmentation. Then, we build a two-path (TP) encoder for deep feature extraction, where one path uses AHSP modules for learning multiscale contextual features and the other is a shallow convolutional path for capturing fine details. The two paths interact with each other for learning effective representations. Based on the extracted features, a simple decoder is added for COVID-19 segmentation. For comparing MiniSeg to previous methods, we build a comprehensive COVID-19 segmentation benchmark. Extensive experiments demonstrate that the proposed MiniSeg achieves better accuracy because its only 83k parameters make it less prone to overfitting. Its high efficiency also makes it easy to deploy and develop. The code has been released at https://github.com/yun-liu/MiniSeg.
Collapse
|
18
|
Fu J, Peng H, Li B, Liu Z, Lugu R, Wang J, Ramírez-de-Arellano A. Multitask Adversarial Networks Based on Extensive Nonlinear Spiking Neuron Models. Int J Neural Syst 2024; 34:2450032. [PMID: 38624267 DOI: 10.1142/s0129065724500321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/17/2024]
Abstract
Deep learning technology has been successfully used in Chest X-ray (CXR) images of COVID-19 patients. However, due to the characteristics of COVID-19 pneumonia and X-ray imaging, the deep learning methods still face many challenges, such as lower imaging quality, fewer training samples, complex radiological features and irregular shapes. To address these challenges, this study first introduces an extensive NSNP-like neuron model, and then proposes a multitask adversarial network architecture based on ENSNP-like neurons for chest X-ray images of COVID-19, called MAE-Net. The MAE-Net serves two tasks: (i) converting low-quality CXR images to high-quality images; (ii) classifying CXR images of COVID-19. The adversarial architecture of MAE-Net uses two generators and two discriminators, and two new loss functions have been introduced to guide the optimization of the network. The MAE-Net is tested on four benchmark COVID-19 CXR image datasets and compared them with eight deep learning models. The experimental results show that the proposed MAE-Net can enhance the conversion quality and the accuracy of image classification results.
Collapse
Affiliation(s)
- Jun Fu
- School of Computer and Software Engineering, Xihua University, Chengdu 610039, P. R. China
| | - Hong Peng
- School of Computer and Software Engineering, Xihua University, Chengdu 610039, P. R. China
| | - Bing Li
- School of Computer and Software Engineering, Xihua University, Chengdu 610039, P. R. China
| | - Zhicai Liu
- School of Computer and Software Engineering, Xihua University, Chengdu 610039, P. R. China
| | - Rikong Lugu
- School of Computer and Software Engineering, Xihua University, Chengdu 610039, P. R. China
| | - Jun Wang
- School of Electrical Engineering and Electronic Information, Xihua University, Chengdu 610039, P. R. China
| | - Antonio Ramírez-de-Arellano
- Research Group of Natural Computing, Department of Computer Science and Artificial Intelligence, University of Seville, Sevilla 41012, Spain
| |
Collapse
|
19
|
Yi C, Niu G, Zhang Y, Rao J, Liu G, Yang W, Fei X. Advances in artificial intelligence in thyroid-associated ophthalmopathy. Front Endocrinol (Lausanne) 2024; 15:1356055. [PMID: 38715793 PMCID: PMC11075148 DOI: 10.3389/fendo.2024.1356055] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/15/2023] [Accepted: 04/10/2024] [Indexed: 05/23/2024] Open
Abstract
Thyroid-associated ophthalmopathy (TAO), also referred to as Graves' ophthalmopathy, is a medical condition wherein ocular complications arise due to autoimmune thyroid illness. The diagnosis of TAO, reliant on imaging, typical ocular symptoms, and abnormalities in thyroid function or thyroid-associated antibodies, is generally graded and staged. In recent years, Artificial intelligence(AI), particularly deep learning(DL) technology, has gained widespread use in the diagnosis and treatment of ophthalmic diseases. This paper presents a discussion on specific studies involving AI, specifically DL, in the context of TAO, highlighting their applications in TAO diagnosis, staging, grading, and treatment decisions. Additionally, it addresses certain limitations in AI research on TAO and potential future directions for the field.
Collapse
Affiliation(s)
- Chenyuan Yi
- Guangdong Key Laboratory of Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen, China
| | - Geng Niu
- School of Medical Technology and Nursing, Shenzhen Polytechnic University, Shenzhen, China
| | - Yinghuai Zhang
- Guangdong Key Laboratory of Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen, China
| | - Jing Rao
- Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China
| | - Guiqin Liu
- Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China
| | - Weihua Yang
- Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China
| | - XingZhen Fei
- Department of Endocrinology, First People’s Hospital of Huzhou, Huzhou University, Huzhou, China
| |
Collapse
|
20
|
He S, Li Q, Li X, Zhang M. A Lightweight Convolutional Neural Network Based on Dynamic Level-Set Loss Function for Spine MR Image Segmentation. J Magn Reson Imaging 2024; 59:1438-1453. [PMID: 37382232 DOI: 10.1002/jmri.28877] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Revised: 06/09/2023] [Accepted: 06/09/2023] [Indexed: 06/30/2023] Open
Abstract
BACKGROUND Spine MR image segmentation is important foundation for computer-aided diagnostic (CAD) algorithms of spine disorders. Convolutional neural networks segment effectively, but require high computational costs. PURPOSE To design a lightweight model based on dynamic level-set loss function for high segmentation performance. STUDY TYPE Retrospective. POPULATION Four hundred forty-eight subjects (3163 images) from two separate datasets. Dataset-1: 276 subjects/994 images (53.26% female, mean age 49.02 ± 14.09), all for disc degeneration screening, 188 had disc degeneration, 67 had herniated disc. Dataset-2: public dataset with 172 subjects/2169 images, 142 patients with vertebral degeneration, 163 patients with disc degeneration. FIELD STRENGTH/SEQUENCE T2 weighted turbo spin echo sequences at 3T. ASSESSMENT Dynamic Level-set Net (DLS-Net) was compared with four mainstream (including U-net++) and four lightweight models, and manual label made by five radiologists (vertebrae, discs, spinal fluid) used as segmentation evaluation standard. Five-fold cross-validation are used for all experiments. Based on segmentation, a CAD algorithm of lumbar disc was designed for assessing DLS-Net's practicality, and the text annotation (normal, bulging, or herniated) from medical history data were used as evaluation standard. STATISTICAL TESTS All segmentation models were evaluated with DSC, accuracy, precision, and AUC. The pixel numbers of segmented results were compared with manual label using paired t-tests, with P < 0.05 indicating significance. The CAD algorithm was evaluated with accuracy of lumbar disc diagnosis. RESULTS With only 1.48% parameters of U-net++, DLS-Net achieved similar accuracy in both datasets (Dataset-1: DSC 0.88 vs. 0.89, AUC 0.94 vs. 0.94; Dataset-2: DSC 0.86 vs. 0.86, AUC 0.93 vs. 0.93). The segmentation results of DLS-Net showed no significant differences with manual labels in pixel numbers for discs (Dataset-1: 1603.30 vs. 1588.77, P = 0.22; Dataset-2: 863.61 vs. 886.4, P = 0.14) and vertebrae (Dataset-1: 3984.28 vs. 3961.94, P = 0.38; Dataset-2: 4806.91 vs. 4732.85, P = 0.21). Based on DLS-Net's segmentation results, the CAD algorithm achieved higher accuracy than using non-cropped MR images (87.47% vs. 61.82%). DATA CONCLUSION The proposed DLS-Net has fewer parameters but achieves similar accuracy to U-net++, helps CAD algorithm achieve higher accuracy, which facilitates wider application. EVIDENCE LEVEL 2 TECHNICAL EFFICACY: Stage 1.
Collapse
Affiliation(s)
- Siyuan He
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, China
| | - Qi Li
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, China
- Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, China
| | - Xianda Li
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, China
| | - Mengchao Zhang
- Department of Radiology, China-Japan Union Hospital of Jilin University, Changchun, China
| |
Collapse
|
21
|
Du H, Wang J, Liu M, Wang Y, Meijering E. SwinPA-Net: Swin Transformer-Based Multiscale Feature Pyramid Aggregation Network for Medical Image Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:5355-5366. [PMID: 36121961 DOI: 10.1109/tnnls.2022.3204090] [Citation(s) in RCA: 21] [Impact Index Per Article: 21.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
The precise segmentation of medical images is one of the key challenges in pathology research and clinical practice. However, many medical image segmentation tasks have problems such as large differences between different types of lesions and similar shapes as well as colors between lesions and surrounding tissues, which seriously affects the improvement of segmentation accuracy. In this article, a novel method called Swin Pyramid Aggregation network (SwinPA-Net) is proposed by combining two designed modules with Swin Transformer to learn more powerful and robust features. The two modules, named dense multiplicative connection (DMC) module and local pyramid attention (LPA) module, are proposed to aggregate the multiscale context information of medical images. The DMC module cascades the multiscale semantic feature information through dense multiplicative feature fusion, which minimizes the interference of shallow background noise to improve the feature expression and solves the problem of excessive variation in lesion size and type. Moreover, the LPA module guides the network to focus on the region of interest by merging the global attention and the local attention, which helps to solve similar problems. The proposed network is evaluated on two public benchmark datasets for polyp segmentation task and skin lesion segmentation task as well as a clinical private dataset for laparoscopic image segmentation task. Compared with existing state-of-the-art (SOTA) methods, the SwinPA-Net achieves the most advanced performance and can outperform the second-best method on the mean Dice score by 1.68%, 0.8%, and 1.2% on the three tasks, respectively.
Collapse
|
22
|
Gupta U, Paluru N, Nankani D, Kulkarni K, Awasthi N. A comprehensive review on efficient artificial intelligence models for classification of abnormal cardiac rhythms using electrocardiograms. Heliyon 2024; 10:e26787. [PMID: 38562492 PMCID: PMC10982903 DOI: 10.1016/j.heliyon.2024.e26787] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2024] [Accepted: 02/20/2024] [Indexed: 04/04/2024] Open
Abstract
Deep learning has made many advances in data classification using electrocardiogram (ECG) waveforms. Over the past decade, data science research has focused on developing artificial intelligence (AI) based models that can analyze ECG waveforms to identify and classify abnormal cardiac rhythms accurately. However, the primary drawback of the current AI models is that most of these models are heavy, computationally intensive, and inefficient in terms of cost for real-time implementation. In this review, we first discuss the current state-of-the-art AI models utilized for ECG-based cardiac rhythm classification. Next, we present some of the upcoming modeling methodologies which have the potential to perform real-time implementation of AI-based heart rhythm diagnosis. These models hold significant promise in being lightweight and computationally efficient without compromising the accuracy. Contemporary models predominantly utilize 12-lead ECG for cardiac rhythm classification and cardiovascular status prediction, increasing the computational burden and making real-time implementation challenging. We also summarize research studies evaluating the potential of efficient data setups to reduce the number of ECG leads without affecting classification accuracy. Lastly, we present future perspectives on AI's utility in precision medicine by providing opportunities for accurate prediction and diagnostics of cardiovascular status in patients.
Collapse
Affiliation(s)
- Utkarsh Gupta
- Department of Computational and Data Sciences, Indian Institute of Science, Bengaluru, 560012, India
| | - Naveen Paluru
- Department of Computational and Data Sciences, Indian Institute of Science, Bengaluru, 560012, India
| | - Deepankar Nankani
- Department of Computer Science and Engineering, Indian Institute of Technology, Guwahati, Assam, 781039, India
| | - Kanchan Kulkarni
- IHU-LIRYC, Heart Rhythm Disease Institute, Fondation Bordeaux Université, Pessac, Bordeaux, F-33000, France
- University of Bordeaux, INSERM, Centre de recherche Cardio-Thoracique de Bordeaux, U1045, Bordeaux, F-33000, France
| | - Navchetan Awasthi
- Faculty of Science, Mathematics and Computer Science, Informatics Institute, University of Amsterdam, Amsterdam, 1090 GH, the Netherlands
- Department of Biomedical Engineering and Physics, Amsterdam UMC, Amsterdam, 1081 HV, the Netherlands
| |
Collapse
|
23
|
Roy R, Mazumdar S, Chowdhury AS. ADGAN: Attribute-Driven Generative Adversarial Network for Synthesis and Multiclass Classification of Pulmonary Nodules. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:2484-2495. [PMID: 35853058 DOI: 10.1109/tnnls.2022.3190331] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Lung cancer is the leading cause of cancer-related deaths worldwide. According to the American Cancer Society, early diagnosis of pulmonary nodules in computed tomography (CT) scans can improve the five-year survival rate up to 70% with proper treatment planning. In this article, we propose an attribute-driven Generative Adversarial Network (ADGAN) for synthesis and multiclass classification of Pulmonary Nodules. A self-attention U-Net (SaUN) architecture is proposed to improve the generation mechanism of the network. The generator is designed with two modules, namely, self-attention attribute module (SaAM) and a self-attention spatial module (SaSM). SaAM generates a nodule image based on given attributes whereas SaSM specifies the nodule region of the input image to be altered. A reconstruction loss along with an attention localization loss (AL) is used to produce an attention map prioritizing the nodule regions. To avoid resemblance between a generated image and a real image, we further introduce an adversarial loss containing a regularization term based on KL divergence. The discriminator part of the proposed model is designed to achieve the multiclass nodule classification task. Our proposed approach is validated over two challenging publicly available datasets, namely LIDC-IDRI and LUNGX. Exhaustive experimentation on these two datasets clearly indicate that we have achieved promising classification accuracy as compared to other state-of-the-art methods.
Collapse
|
24
|
Zhang Y, Dong J. MAEF-Net: MLP Attention for Feature Enhancement in U-Net based Medical Image Segmentation Networks. IEEE J Biomed Health Inform 2024; 28:846-857. [PMID: 37976191 DOI: 10.1109/jbhi.2023.3332908] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2023]
Abstract
Medical image segmentation plays an important role in diagnosis. Since the introduction of U-Net, numerous advancements have been implemented to enhance its performance and expand its applicability. The advent of Transformers in computer vision has led to the integration of self-attention mechanisms into U-Net, resulting in significant breakthroughs. However, the inherent complexity of Transformers renders these networks computationally demanding and parameter-heavy. Recent studies have demonstrated that multilayer perceptrons (MLPs), with their simpler architecture, can achieve comparable performance to Transformers in natural language processing and computer vision tasks. Building upon these findings, we have enhanced the previously proposed "Enhanced-Feature-Four-Fold-Net" (EF 3-Net) by introducing an MLP-attention block to learn long-range dependencies and expand the receptive field. This enhanced network is termed "MLP-Attention Enhanced-Feature-four-fold-Net", abbreviated as "MAEF-Net". To further enhance accuracy while reducing computational complexity, the proposed network incorporates additional efficient design elements. MAEF-Net was evaluated against several general and specialized medical image segmentation networks using four challenging medical image datasets. The results demonstrate that the proposed network exhibits high computational efficiency and comparable or superior performance to EF 3-Net and several state-of-the-art methods, particularly in segmenting blurry objects.
Collapse
|
25
|
Lu F, Zhang Z, Liu T, Tang C, Bai H, Zhai G, Chen J, Wu X. A weakly supervised inpainting-based learning method for lung CT image segmentation. PATTERN RECOGNITION 2023; 144:109861. [DOI: 10.1016/j.patcog.2023.109861] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/05/2024]
|
26
|
He S, Li Q, Li X, Zhang M. LSW-Net: Lightweight Deep Neural Network Based on Small-World properties for Spine MR Image Segmentation. J Magn Reson Imaging 2023; 58:1762-1776. [PMID: 37118994 DOI: 10.1002/jmri.28735] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Revised: 03/30/2023] [Accepted: 03/30/2023] [Indexed: 04/30/2023] Open
Abstract
BACKGROUND Segmenting spinal tissues from MR images is important for automatic image analysis. Deep neural network-based segmentation methods are efficient, yet have high computational costs. PURPOSE To design a lightweight model based on small-world properties (LSW-Net) to segment spinal MR images, suitable for low-computing-power embedded devices. STUDY TYPE Retrospective. POPULATION A total of 386 subjects (2948 images) from two independent sources. Dataset I: 214 subjects/779 images, all for disk degeneration screening, 147 had disk degeneration, 52 had herniated disc. Dataset II: 172 subjects/2169 images, 142 patients with vertebral degeneration, 163 patients with disc degeneration. 70% images in each dataset for training, 20% for validation, and 10% for testing. FIELD STRENGTH/SEQUENCE T1- and T2-weighted turbo spin echo sequences at 3 T. ASSESSMENT Segmentation performance of LSW-Net was compared with four mainstream (including U-net and U-net++) and five lightweight models using five radiologists' manual segmentations (vertebrae, disks, spinal fluid) as reference standard. LSW-Net was also deployed on NVIDIA Jetson nano to compare the pixels number in segmented vertebrae and disks. STATISTICAL TESTS All models were evaluated with accuracy, precision, Dice similarity coefficient (DSC), and area under the receiver operating characteristic (AUC). Pixel numbers segmented by LSW-Net on the embedded device were compared with manual segmentation using paired t-tests, with P < 0.05 indicating significance. RESULTS LSW-Net had 98.5% fewer parameters than U-net but achieved similar accuracy in both datasets (dataset I: DSC 0.84 vs. 0.87, AUC 0.92 vs. 0.94; dataset II: DSC 0.82 vs. 0.82, AUC 0.88 vs. 0.88). LSW-Net showed no significant differences in pixel numbers for vertebrae (dataset I: 5893.49 vs. 5752.61, P = 0.21; dataset II: 5073.42 vs. 5137.12, P = 0.56) and disks (dataset I: 1513.07 vs. 1535.69, P = 0.42; dataset II: 1049.74 vs. 1087.88, P = 0.24) segmentation on an embedded device compared to manual segmentation. DATA CONCLUSION Proposed LSW-Net achieves high accuracy with fewer parameters than U-net and can be deployed on embedded device, facilitating wider application. EVIDENCE LEVEL 2. TECHNICAL EFFICACY 1.
Collapse
Affiliation(s)
- Siyuan He
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, China
| | - Qi Li
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, China
- Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, China
| | - Xianda Li
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, China
| | - Mengchao Zhang
- Department of Radiology, China-Japan Union Hospital of Jilin University, Changchun, China
| |
Collapse
|
27
|
Yue G, Yang C, Zhao Z, An Z, Yang Y. ERGPNet: lesion segmentation network for COVID-19 chest X-ray images based on embedded residual convolution and global perception. Front Physiol 2023; 14:1296185. [PMID: 38028767 PMCID: PMC10679680 DOI: 10.3389/fphys.2023.1296185] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 11/02/2023] [Indexed: 12/01/2023] Open
Abstract
The Segmentation of infected areas from COVID-19 chest X-ray (CXR) images is of great significance for the diagnosis and treatment of patients. However, accurately and effectively segmenting infected areas of CXR images is still challenging due to the inherent ambiguity of CXR images and the cross-scale variations in infected regions. To address these issues, this article proposes a ERGPNet based on embedded residuals and global perception, to segment lesion regions in COVID-19 CXR images. First, aiming at the inherent fuzziness of CXR images, an embedded residual convolution structure is proposed to enhance the ability of internal feature extraction. Second, a global information perception module is constructed to guide the network in generating long-distance information flow, alleviating the interferences of cross-scale variations on the algorithm's discrimination ability. Finally, the network's sensitivity to target regions is improved, and the interference of noise information is suppressed through the utilization of parallel spatial and serial channel attention modules. The interactions between each module fully establish the mapping relationship between feature representation and information decision-making and improve the accuracy of lesion segmentation. Extensive experiments on three datasets of COVID-19 CXR images, and the results demonstrate that the proposed method outperforms other state-of-the-art segmentation methods of CXR images.
Collapse
Affiliation(s)
- Gongtao Yue
- School of Computer Science, Xijing University, Xi’an, China
| | - Chen Yang
- School of Computer Science, Xijing University, Xi’an, China
| | - Zhengyang Zhao
- School of Information and Navigation, Air Force Engineering University, Xi’an, China
| | - Ziheng An
- School of Integrated Circuits, Anhui University, Hefei, China
| | - Yongsheng Yang
- School of Computer Science, Xijing University, Xi’an, China
| |
Collapse
|
28
|
Xiang Z, Mao Q, Wang J, Tian Y, Zhang Y, Wang W. Dmbg-Net: Dilated multiresidual boundary guidance network for COVID-19 infection segmentation. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:20135-20154. [PMID: 38052640 DOI: 10.3934/mbe.2023892] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
Accurate segmentation of infected regions in lung computed tomography (CT) images is essential for the detection and diagnosis of coronavirus disease 2019 (COVID-19). However, lung lesion segmentation has some challenges, such as obscure boundaries, low contrast and scattered infection areas. In this paper, the dilated multiresidual boundary guidance network (Dmbg-Net) is proposed for COVID-19 infection segmentation in CT images of the lungs. This method focuses on semantic relationship modelling and boundary detail guidance. First, to effectively minimize the loss of significant features, a dilated residual block is substituted for a convolutional operation, and dilated convolutions are employed to expand the receptive field of the convolution kernel. Second, an edge-attention guidance preservation block is designed to incorporate boundary guidance of low-level features into feature integration, which is conducive to extracting the boundaries of the region of interest. Third, the various depths of features are used to generate the final prediction, and the utilization of a progressive multi-scale supervision strategy facilitates enhanced representations and highly accurate saliency maps. The proposed method is used to analyze COVID-19 datasets, and the experimental results reveal that the proposed method has a Dice similarity coefficient of 85.6% and a sensitivity of 84.2%. Extensive experimental results and ablation studies have shown the effectiveness of Dmbg-Net. Therefore, the proposed method has a potential application in the detection, labeling and segmentation of other lesion areas.
Collapse
Affiliation(s)
- Zhenwu Xiang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Qi Mao
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Jintao Wang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Yi Tian
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Yan Zhang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| | - Wenfeng Wang
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
| |
Collapse
|
29
|
Zhao X, Lin Z, Yu S, Xiao J, Xie L, Xu Y, Tsui CK, Cui K, Zhao L, Zhang G, Zhang S, Lu Y, Lin H, Liang X, Lin D. An artificial intelligence system for the whole process from diagnosis to treatment suggestion of ischemic retinal diseases. Cell Rep Med 2023; 4:101197. [PMID: 37734379 PMCID: PMC10591037 DOI: 10.1016/j.xcrm.2023.101197] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2023] [Revised: 05/29/2023] [Accepted: 08/23/2023] [Indexed: 09/23/2023]
Abstract
Ischemic retinal diseases (IRDs) are a series of common blinding diseases that depend on accurate fundus fluorescein angiography (FFA) image interpretation for diagnosis and treatment. An artificial intelligence system (Ai-Doctor) was developed to interpret FFA images. Ai-Doctor performed well in image phase identification (area under the curve [AUC], 0.991-0.999, range), diabetic retinopathy (DR) and branch retinal vein occlusion (BRVO) diagnosis (AUC, 0.979-0.992), and non-perfusion area segmentation (Dice similarity coefficient [DSC], 89.7%-90.1%) and quantification. The segmentation model was expanded to unencountered IRDs (central RVO and retinal vasculitis), with DSCs of 89.2% and 83.6%, respectively. A clinically applicable ischemia index (CAII) was proposed to evaluate ischemic degree; patients with CAII values exceeding 0.17 in BRVO and 0.08 in DR may be associated with increased possibility for laser therapy. Ai-Doctor is expected to achieve accurate FFA image interpretation for IRDs, potentially reducing the reliance on retinal specialists.
Collapse
Affiliation(s)
- Xinyu Zhao
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China; Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen 518040, China
| | - Zhenzhe Lin
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Shanshan Yu
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Jun Xiao
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Liqiong Xie
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Yue Xu
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Ching-Kit Tsui
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Kaixuan Cui
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Lanqin Zhao
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China
| | - Guoming Zhang
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen 518040, China
| | - Shaochong Zhang
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen 518040, China
| | - Yan Lu
- Foshan Second People's Hospital, Foshan 528001, China
| | - Haotian Lin
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China; Hainan Eye Hospital and Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Haikou 570311, China; Center for Precision Medicine and Department of Genetics and Biomedical Informatics, Zhongshan School of Medicine, Sun Yat-sen University, Guangzhou 510080, China.
| | - Xiaoling Liang
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China.
| | - Duoru Lin
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou 510060, China.
| |
Collapse
|
30
|
Rao Z, Dai Y, Shen Z, He R. Rethinking Training Strategy in Stereo Matching. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:7796-7809. [PMID: 35143404 DOI: 10.1109/tnnls.2022.3146306] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
In stereo matching, various learning-based approaches have shown impressive performance in solving traditional difficulties on multiple datasets. While most progress is obtained on a specific dataset with a dataset-specific network design, the performance on the single dataset and cross dataset affected by training strategy is often ignored. In this article, we analyze the relationship between different training strategies and performance by retraining some representative state-of-the-art methods (e.g., geometry and context network (GC-Net), pyramid stereo matching network (PSM-Net), and guided aggregation network (GA-Net), etc.). According to our research, it is surprising that the performance of networks on single or cross datasets is significantly improved by pre-training and data augmentation without any particular structure acquirement. Based on this discovery, we improve our previous non-local context attention network (NLCA-Net) to NLCA-Net v2 and train it with the novel strategy and rethink the training strategy of stereo matching concurrently. The quantitative experiments demonstrate that: 1) our model is capable of reaching top performance on both the single dataset and the multiple datasets with the same parameters in this study, which also won the 2nd place in the stereo task of the ECCV Robust vision Challenge 2020 (RVC 2020); and 2) on small datasets (e.g., KITTI, ETH3D, and Middlebury), the model's generalization and robustness are significantly affected by pre-training and data augmentation, even exceeding the network structure's influence in some cases. These observations present a challenge to the conventional wisdom of network architectures in this stage. We expect these discoveries to encourage researchers to rethink the current paradigm of "excessive attention on the performance of a single small dataset" in stereo matching.
Collapse
|
31
|
He Z, Wong ANN, Yoo JS. Co-ERA-Net: Co-Supervision and Enhanced Region Attention for Accurate Segmentation in COVID-19 Chest Infection Images. Bioengineering (Basel) 2023; 10:928. [PMID: 37627813 PMCID: PMC10451793 DOI: 10.3390/bioengineering10080928] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Revised: 07/31/2023] [Accepted: 08/03/2023] [Indexed: 08/27/2023] Open
Abstract
Accurate segmentation of infected lesions in chest images remains a challenging task due to the lack of utilization of lung region information, which could serve as a strong location hint for infection. In this paper, we propose a novel segmentation network Co-ERA-Net for infections in chest images that leverages lung region information by enhancing supervised information and fusing multi-scale lung region and infection information at different levels. To achieve this, we introduce a Co-supervision scheme incorporating lung region information to guide the network to accurately locate infections within the lung region. Furthermore, we design an Enhanced Region Attention Module (ERAM) to highlight regions with a high probability of infection by incorporating infection information into the lung region information. The effectiveness of the proposed scheme is demonstrated using COVID-19 CT and X-ray datasets, with the results showing that the proposed schemes and modules are promising. Based on the baseline, the Co-supervision scheme, when integrated with lung region information, improves the Dice coefficient by 7.41% and 2.22%, and the IoU by 8.20% and 3.00% in CT and X-ray datasets respectively. Moreover, when this scheme is combined with the Enhanced Region Attention Module, the Dice coefficient sees further improvement of 14.24% and 2.97%, with the IoU increasing by 28.64% and 4.49% for the same datasets. In comparison with existing approaches across various datasets, our proposed method achieves better segmentation performance in all main metrics and exhibits the best generalization and comprehensive performance.
Collapse
Affiliation(s)
| | | | - Jung Sun Yoo
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China; (Z.H.); (A.N.N.W.)
| |
Collapse
|
32
|
Saha S, Dutta S, Goswami B, Nandi D. ADU-Net: An Attention Dense U-Net based deep supervised DNN for automated lesion segmentation of COVID-19 from chest CT images. Biomed Signal Process Control 2023; 85:104974. [PMID: 37122956 PMCID: PMC10121143 DOI: 10.1016/j.bspc.2023.104974] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Revised: 04/01/2023] [Accepted: 04/15/2023] [Indexed: 05/02/2023]
Abstract
An automatic method for qualitative and quantitative evaluation of chest Computed Tomography (CT) images is essential for diagnosing COVID-19 patients. We aim to develop an automated COVID-19 prediction framework using deep learning. We put forth a novel Deep Neural Network (DNN) composed of an attention-based dense U-Net with deep supervision for COVID-19 lung lesion segmentation from chest CT images. We incorporate dense U-Net where convolution kernel size 5×5 is used instead of 3×3. The dense and transition blocks are introduced to implement a densely connected network on each encoder level. Also, the attention mechanism is applied between the encoder, skip connection, and decoder. These are used to keep both the high and low-level features efficiently. The deep supervision mechanism creates secondary segmentation maps from the features. Deep supervision combines secondary supervision maps from various resolution levels and produces a better final segmentation map. The trained artificial DNN model takes the test data at its input and generates a prediction output for COVID-19 lesion segmentation. The proposed model has been applied to the MedSeg COVID-19 chest CT segmentation dataset. Data pre-processing methods help the training process and improve performance. We compare the performance of the proposed DNN model with state-of-the-art models by computing the well-known metrics: dice coefficient, Jaccard coefficient, accuracy, specificity, sensitivity, and precision. As a result, the proposed model outperforms the state-of-the-art models. This new model may be considered an efficient automated screening system for COVID-19 diagnosis and can potentially improve patient health care and management system.
Collapse
Affiliation(s)
- Sanjib Saha
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, 713209, West Bengal, India
- Department of Computer Science and Engineering, Dr. B. C. Roy Engineering College, Durgapur, 713206, West Bengal, India
| | - Subhadeep Dutta
- Department of Computer Science and Engineering, Dr. B. C. Roy Engineering College, Durgapur, 713206, West Bengal, India
| | - Biswarup Goswami
- Department of Respiratory Medicine, Health and Family Welfare, Government of West Bengal, Kolkata, 700091, West Bengal, India
| | - Debashis Nandi
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, 713209, West Bengal, India
| |
Collapse
|
33
|
Bhosale YH, Patnaik KS. Bio-medical imaging (X-ray, CT, ultrasound, ECG), genome sequences applications of deep neural network and machine learning in diagnosis, detection, classification, and segmentation of COVID-19: a Meta-analysis & systematic review. MULTIMEDIA TOOLS AND APPLICATIONS 2023; 82:1-54. [PMID: 37362676 PMCID: PMC10015538 DOI: 10.1007/s11042-023-15029-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/12/2022] [Revised: 02/01/2023] [Accepted: 02/27/2023] [Indexed: 06/28/2023]
Abstract
This review investigates how Deep Machine Learning (DML) has dealt with the Covid-19 epidemic and provides recommendations for future Covid-19 research. Despite the fact that vaccines for this epidemic have been developed, DL methods have proven to be a valuable asset in radiologists' arsenals for the automated assessment of Covid-19. This detailed review debates the techniques and applications developed for Covid-19 findings using DL systems. It also provides insights into notable datasets used to train neural networks, data partitioning, and various performance measurement metrics. The PRISMA taxonomy has been formed based on pretrained(45 systems) and hybrid/custom(17 systems) models with radiography modalities. A total of 62 systems with respect to X-ray(32), CT(19), ultrasound(7), ECG(2), and genome sequence(2) based modalities as taxonomy are selected from the studied articles. We originate by valuing the present phase of DL and conclude with significant limitations. The restrictions contain incomprehensibility, simplification measures, learning from incomplete labeled data, and data secrecy. Moreover, DML can be utilized to detect and classify Covid-19 from other COPD illnesses. The proposed literature review has found many DL-based systems to fight against Covid19. We expect this article will assist in speeding up the procedure of DL for Covid-19 researchers, including medical, radiology technicians, and data engineers.
Collapse
Affiliation(s)
- Yogesh H. Bhosale
- Computer Science and Engineering Department, Birla Institute of Technology, Mesra, Ranchi, India
| | - K. Sridhar Patnaik
- Computer Science and Engineering Department, Birla Institute of Technology, Mesra, Ranchi, India
| |
Collapse
|
34
|
Rao Y, Lv Q, Zeng S, Yi Y, Huang C, Gao Y, Cheng Z, Sun J. COVID-19 CT ground-glass opacity segmentation based on attention mechanism threshold. Biomed Signal Process Control 2023; 81:104486. [PMID: 36505089 PMCID: PMC9721288 DOI: 10.1016/j.bspc.2022.104486] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2022] [Revised: 11/23/2022] [Accepted: 12/01/2022] [Indexed: 12/12/2022]
Abstract
The ground glass opacity (GGO) of the lung is one of the essential features of COVID-19. The GGO in computed tomography (CT) images has various features and low-intensity contrast between the GGO and edge structures. These problems pose significant challenges for segmenting the GGO. To tackle these problems, we propose a new threshold method for accurate segmentation of GGO. Specifically, we offer a framework for adjusting the threshold parameters according to the image contrast. Three functions include Attention mechanism threshold, Contour equalization, and Lung segmentation (ACL). The lung is divided into three areas using the attention mechanism threshold. Further, the segmentation parameters of the attention mechanism thresholds of the three parts are adaptively adjusted according to the image contrast. Only the segmentation regions restricted by the lung segmentation results are retained. Extensive experiments on four COVID datasets show that ACL can segment GGO images at low contrast well. Compared with the state-of-the-art methods, the similarity Dice of the ACL segmentation results is improved by 8.9%, the average symmetry surface distance ASD is reduced by 23%, and the required computational power F L O P s are only 0.09% of those of deep learning models. For GGO segmentation, ACL is more lightweight, and the accuracy is higher. Code will be released at https://github.com/Lqs-github/ACL.
Collapse
Affiliation(s)
- Yunbo Rao
- School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Qingsong Lv
- School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Shaoning Zeng
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou, 313000, China
| | - Yuling Yi
- School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Cheng Huang
- Fifth Clinical College of Chongqing Medical University, Chongqing, 402177, China
| | - Yun Gao
- Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
| | - Zhanglin Cheng
- Advanced Technology Chinese Academy of Sciences, Shenzhen, 610042, China
| | - Jihong Sun
- Sir Run Run Shaw Hospital, Zhejiang University School of Medicine, Hangzhou, 310014, China
| |
Collapse
|
35
|
Lu F, Tang C, Liu T, Zhang Z, Li L. Multi-Attention Segmentation Networks Combined with the Sobel Operator for Medical Images. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23052546. [PMID: 36904754 PMCID: PMC10007317 DOI: 10.3390/s23052546] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/18/2022] [Revised: 02/10/2023] [Accepted: 02/15/2023] [Indexed: 05/27/2023]
Abstract
Medical images are used as an important basis for diagnosing diseases, among which CT images are seen as an important tool for diagnosing lung lesions. However, manual segmentation of infected areas in CT images is time-consuming and laborious. With its excellent feature extraction capabilities, a deep learning-based method has been widely used for automatic lesion segmentation of COVID-19 CT images. However, the segmentation accuracy of these methods is still limited. To effectively quantify the severity of lung infections, we propose a Sobel operator combined with multi-attention networks for COVID-19 lesion segmentation (SMA-Net). In our SMA-Net method, an edge feature fusion module uses the Sobel operator to add edge detail information to the input image. To guide the network to focus on key regions, SMA-Net introduces a self-attentive channel attention mechanism and a spatial linear attention mechanism. In addition, the Tversky loss function is adopted for the segmentation network for small lesions. Comparative experiments on COVID-19 public datasets show that the average Dice similarity coefficient (DSC) and joint intersection over union (IOU) of the proposed SMA-Net model are 86.1% and 77.8%, respectively, which are better than those in most existing segmentation networks.
Collapse
Affiliation(s)
- Fangfang Lu
- College of Computer Science and Technology, Shanghai University of Electric Power, Shanghai 201399, China
- Department of Electronic Engineering, School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Chi Tang
- College of Computer Science and Technology, Shanghai University of Electric Power, Shanghai 201399, China
| | - Tianxiang Liu
- College of Computer Science and Technology, Shanghai University of Electric Power, Shanghai 201399, China
| | - Zhihao Zhang
- College of Computer Science and Technology, Shanghai University of Electric Power, Shanghai 201399, China
| | - Leida Li
- School of Artificial Intelligence, Xidian University, Xi’an 710000, China
| |
Collapse
|
36
|
Biglari A, Tang W. A Review of Embedded Machine Learning Based on Hardware, Application, and Sensing Scheme. SENSORS (BASEL, SWITZERLAND) 2023; 23:2131. [PMID: 36850729 PMCID: PMC9959746 DOI: 10.3390/s23042131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/16/2022] [Revised: 01/17/2023] [Accepted: 02/09/2023] [Indexed: 06/18/2023]
Abstract
Machine learning is an expanding field with an ever-increasing role in everyday life, with its utility in the industrial, agricultural, and medical sectors being undeniable. Recently, this utility has come in the form of machine learning implementation on embedded system devices. While there have been steady advances in the performance, memory, and power consumption of embedded devices, most machine learning algorithms still have a very high power consumption and computational demand, making the implementation of embedded machine learning somewhat difficult. However, different devices can be implemented for different applications based on their overall processing power and performance. This paper presents an overview of several different implementations of machine learning on embedded systems divided by their specific device, application, specific machine learning algorithm, and sensors. We will mainly focus on NVIDIA Jetson and Raspberry Pi devices with a few different less utilized embedded computers, as well as which of these devices were more commonly used for specific applications in different fields. We will also briefly analyze the specific ML models most commonly implemented on the devices and the specific sensors that were used to gather input from the field. All of the papers included in this review were selected using Google Scholar and published papers in the IEEExplore database. The selection criterion for these papers was the usage of embedded computing systems in either a theoretical study or practical implementation of machine learning models. The papers needed to have provided either one or, preferably, all of the following results in their studies-the overall accuracy of the models on the system, the overall power consumption of the embedded machine learning system, and the inference time of their models on the embedded system. Embedded machine learning is experiencing an explosion in both scale and scope, both due to advances in system performance and machine learning models, as well as greater affordability and accessibility of both. Improvements are noted in quality, power usage, and effectiveness.
Collapse
|
37
|
Kamath V, Renuka A. Deep Learning Based Object Detection for Resource Constrained Devices- Systematic Review, Future Trends and Challenges Ahead. Neurocomputing 2023. [DOI: 10.1016/j.neucom.2023.02.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/12/2023]
|
38
|
Rodrigues Moreira LF, Moreira R, Travençolo BAN, Backes AR. An Artificial Intelligence-as-a-Service Architecture for deep learning model embodiment on low-cost devices: A case study of COVID-19 diagnosis. Appl Soft Comput 2023; 134:110014. [PMID: 36687763 PMCID: PMC9837155 DOI: 10.1016/j.asoc.2023.110014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2022] [Revised: 12/03/2022] [Accepted: 01/07/2023] [Indexed: 01/15/2023]
Abstract
Coronavirus Disease-2019 (COVID-19) causes Severe Acute Respiratory Syndrome-Corona Virus-2 (SARS-CoV-2) and has opened several challenges for research concerning diagnosis and treatment. Chest X-rays and computed tomography (CT) scans are effective and fast alternatives to detect and assess the damage that COVID causes to the lungs at different stages of the disease. Although the CT scan is an accurate exam, the chest X-ray is still helpful due to the cheaper, faster, lower radiation exposure, and is available in low-incoming countries. Computer-aided diagnostic systems based on Artificial Intelligence (AI) and computer vision are an alternative to extract features from X-ray images, providing an accurate COVID-19 diagnosis. However, specialized and expensive computational resources come across as challenging. Also, it needs to be better understood how low-cost devices and smartphones can hold AI models to predict diseases timely. Even using deep learning to support image-based medical diagnosis, challenges still need to be addressed once the known techniques use centralized intelligence on high-performance servers, making it difficult to embed these models in low-cost devices. This paper sheds light on these questions by proposing the Artificial Intelligence as a Service Architecture (AIaaS), a hybrid AI support operation, both centralized and distributed, with the purpose of enabling the embedding of already-trained models on low-cost devices or smartphones. We demonstrated the suitability of our architecture through a case study of COVID-19 diagnosis using a low-cost device. Among the main findings of this paper, we point out the performance evaluation of low-cost devices to handle COVID-19 predicting tasks timely and accurately and the quantitative performance evaluation of CNN models embodiment on low-cost devices.
Collapse
Affiliation(s)
| | - Rodrigo Moreira
- Institute of Exacts and Technological Sciences (IEP), Federal University of Viçosa, Rio Paranaíba, Minas Gerais, Brazil
| | | | - André Ricardo Backes
- Department of Computing (DC), Federal University of São Carlos, São Carlos, São Paulo, Brazil
| |
Collapse
|
39
|
Wang J, Luo Y, Wang Z, Hounye AH, Cao C, Hou M, Zhang J. A cell phone app for facial acne severity assessment. APPL INTELL 2023; 53:7614-7633. [PMID: 35919632 PMCID: PMC9336136 DOI: 10.1007/s10489-022-03774-z] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/15/2022] [Indexed: 11/28/2022]
Abstract
Acne vulgaris, the most common skin disease, can cause substantial economic and psychological impacts to the people it affects, and its accurate grading plays a crucial role in the treatment of patients. In this paper, we firstly proposed an acne grading criterion that considers lesion classifications and a metric for producing accurate severity ratings. Due to similar appearance of acne lesions with comparable severities and difficult-to-count lesions, severity assessment is a challenging task. We cropped facial skin images of several lesion patches and then addressed the acne lesion with a lightweight acne regular network (Acne-RegNet). Acne-RegNet was built by using a median filter and histogram equalization to improve image quality, a channel attention mechanism to boost the representational power of network, a region-based focal loss to handle classification imbalances and a model pruning and feature-based knowledge distillation to reduce model size. After the application of Acne-RegNet, the severity score is calculated, and the acne grading is further optimized by the metadata of the patients. The entire acne assessment procedure was deployed to a mobile device, and a phone app was designed. Compared with state-of-the-art lightweight models, the proposed Acne-RegNet significantly improves the accuracy of lesion classifications. The acne app demonstrated promising results in severity assessments (accuracy: 94.56%) and showed a dermatologist-level diagnosis on the internal clinical dataset.The proposed acne app could be a useful adjunct to assess acne severity in clinical practice and it enables anyone with a smartphone to immediately assess acne, anywhere and anytime.
Collapse
Affiliation(s)
- Jiaoju Wang
- School of Mathematics and Statistics, Central South University, Changsha, 410083 Hunan China
| | - Yan Luo
- Department of dermatology of Xiangya hospital, Central South University, Changsha, 410083 Hunan China
| | - Zheng Wang
- School of Mathematics and Statistics, Central South University, Changsha, 410083 Hunan China.,Science and Engineering School, Hunan First Normal University, Changsha, 410083 Hunan China
| | - Alphonse Houssou Hounye
- School of Mathematics and Statistics, Central South University, Changsha, 410083 Hunan China
| | - Cong Cao
- School of Mathematics and Statistics, Central South University, Changsha, 410083 Hunan China
| | - Muzhou Hou
- School of Mathematics and Statistics, Central South University, Changsha, 410083 Hunan China
| | - Jianglin Zhang
- Department of Dermatology of Shenzhen People's Hospital The Second Clinical Medical College of Jinan Uninversity, The First Affiliated Hospital of Southern University of Science and Technology, Shenzhen, 518020 Guangdong China.,Candidate Branch of National Clinical Research Center for Skin Diseases, Shenzhen, 518020 Guangdong China
| |
Collapse
|
40
|
Chen H, Jiang Y, Ko H, Loew M. A teacher-student framework with Fourier Transform augmentation for COVID-19 infection segmentation in CT images. Biomed Signal Process Control 2023; 79:104250. [PMID: 36188130 PMCID: PMC9510070 DOI: 10.1016/j.bspc.2022.104250] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 08/11/2022] [Accepted: 09/18/2022] [Indexed: 11/23/2022]
Abstract
Automatic segmentation of infected regions in computed tomography (CT) images is necessary for the initial diagnosis of COVID-19. Deep-learning-based methods have the potential to automate this task but require a large amount of data with pixel-level annotations. Training a deep network with annotated lung cancer CT images, which are easier to obtain, can alleviate this problem to some extent. However, this approach may suffer from a reduction in performance when applied to unseen COVID-19 images during the testing phase, caused by the difference in the image intensity and object region distribution between the training set and test set. In this paper, we proposed a novel unsupervised method for COVID-19 infection segmentation that aims to learn the domain-invariant features from lung cancer and COVID-19 images to improve the generalization ability of the segmentation network for use with COVID-19 CT images. First, to address the intensity difference, we proposed a novel data augmentation module based on Fourier Transform, which transfers the annotated lung cancer data into the style of COVID-19 image. Secondly, to reduce the distribution difference, we designed a teacher-student network to learn rotation-invariant features for segmentation. The experiments demonstrated that even without getting access to the annotations of the COVID-19 CT images during the training phase, the proposed network can achieve a state-of-the-art segmentation performance on COVID-19 infection.
Collapse
Affiliation(s)
- Han Chen
- School of Electrical Engineering, Korea University, Seoul, South Korea
| | - Yifan Jiang
- School of Electrical Engineering, Korea University, Seoul, South Korea
| | - Hanseok Ko
- School of Electrical Engineering, Korea University, Seoul, South Korea
| | - Murray Loew
- Biomedical Engineering, George Washington University, Washington D.C., USA
| |
Collapse
|
41
|
A novel automatic reading method of pointer meters based on deep learning. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-08110-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
|
42
|
Lasker A, Ghosh M, Obaidullah SM, Chakraborty C, Roy K. LWSNet - a novel deep-learning architecture to segregate Covid-19 and pneumonia from x-ray imagery. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:21801-21823. [PMID: 36532598 PMCID: PMC9734972 DOI: 10.1007/s11042-022-14247-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/06/2022] [Revised: 08/18/2022] [Accepted: 11/04/2022] [Indexed: 05/27/2023]
Abstract
Automatic detection of lung diseases using AI-based tools became very much necessary to handle the huge number of cases occurring across the globe and support the doctors. This paper proposed a novel deep learning architecture named LWSNet (Light Weight Stacking Network) to separate Covid-19, cold pneumonia, and normal chest x-ray images. This framework is based on single, double, triple, and quadruple stack mechanisms to address the above-mentioned tri-class problem. In this framework, a truncated version of standard deep learning models and a lightweight CNN model was considered to conviniently deploy in resource-constraint devices. An evaluation was conducted on three publicly available datasets alongwith their combination. We received 97.28%, 96.50%, 97.41%, and 98.54% highest classification accuracies using quadruple stack. On further investigation, we found, using LWSNet, the average accuracy got improved from individual model to quadruple model by 2.31%, 2.55%, 2.88%, and 2.26% on four respective datasets.
Collapse
Affiliation(s)
- Asifuzzaman Lasker
- Department of Computer Science & Engineering, Aliah University, Kolkata, India
| | - Mridul Ghosh
- Department of Computer Science, Shyampur Siddheswari Mahavidyalaya, Howrah, India
| | - Sk Md Obaidullah
- Department of Computer Science & Engineering, Aliah University, Kolkata, India
| | | | - Kaushik Roy
- Department of Computer Science, West Bengal State University, Barasat, India
| |
Collapse
|
43
|
Lasker A, Ghosh M, Obaidullah SM, Chakraborty C, Roy K. LWSNet - a novel deep-learning architecture to segregate Covid-19 and pneumonia from x-ray imagery. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:1-23. [PMID: 36532598 DOI: 10.1007/s11042-022-13740-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/06/2022] [Revised: 08/18/2022] [Accepted: 11/04/2022] [Indexed: 05/23/2023]
Abstract
Automatic detection of lung diseases using AI-based tools became very much necessary to handle the huge number of cases occurring across the globe and support the doctors. This paper proposed a novel deep learning architecture named LWSNet (Light Weight Stacking Network) to separate Covid-19, cold pneumonia, and normal chest x-ray images. This framework is based on single, double, triple, and quadruple stack mechanisms to address the above-mentioned tri-class problem. In this framework, a truncated version of standard deep learning models and a lightweight CNN model was considered to conviniently deploy in resource-constraint devices. An evaluation was conducted on three publicly available datasets alongwith their combination. We received 97.28%, 96.50%, 97.41%, and 98.54% highest classification accuracies using quadruple stack. On further investigation, we found, using LWSNet, the average accuracy got improved from individual model to quadruple model by 2.31%, 2.55%, 2.88%, and 2.26% on four respective datasets.
Collapse
Affiliation(s)
- Asifuzzaman Lasker
- Department of Computer Science & Engineering, Aliah University, Kolkata, India
| | - Mridul Ghosh
- Department of Computer Science, Shyampur Siddheswari Mahavidyalaya, Howrah, India
| | - Sk Md Obaidullah
- Department of Computer Science & Engineering, Aliah University, Kolkata, India
| | | | - Kaushik Roy
- Department of Computer Science, West Bengal State University, Barasat, India
| |
Collapse
|
44
|
Hussain MA, Mirikharaji Z, Momeny M, Marhamati M, Neshat AA, Garbi R, Hamarneh G. Active deep learning from a noisy teacher for semi-supervised 3D image segmentation: Application to COVID-19 pneumonia infection in CT. Comput Med Imaging Graph 2022; 102:102127. [PMID: 36257092 PMCID: PMC9540707 DOI: 10.1016/j.compmedimag.2022.102127] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2022] [Revised: 09/23/2022] [Accepted: 09/28/2022] [Indexed: 01/27/2023]
Abstract
Supervised deep learning has become a standard approach to solving medical image segmentation tasks. However, serious difficulties in attaining pixel-level annotations for sufficiently large volumetric datasets in real-life applications have highlighted the critical need for alternative approaches, such as semi-supervised learning, where model training can leverage small expert-annotated datasets to enable learning from much larger datasets without laborious annotation. Most of the semi-supervised approaches combine expert annotations and machine-generated annotations with equal weights within deep model training, despite the latter annotations being relatively unreliable and likely to affect model optimization negatively. To overcome this, we propose an active learning approach that uses an example re-weighting strategy, where machine-annotated samples are weighted (i) based on the similarity of their gradient directions of descent to those of expert-annotated data, and (ii) based on the gradient magnitude of the last layer of the deep model. Specifically, we present an active learning strategy with a query function that enables the selection of reliable and more informative samples from machine-annotated batch data generated by a noisy teacher. When validated on clinical COVID-19 CT benchmark data, our method improved the performance of pneumonia infection segmentation compared to the state of the art.
Collapse
Affiliation(s)
| | - Zahra Mirikharaji
- Medical Image Analysis Lab, Simon Fraser University, Burnaby, BC V5A 1S6, Canada.
| | | | | | | | - Rafeef Garbi
- BiSICL, University of British Columbia, Vancouver, BC V6T 1Z4, Canada.
| | - Ghassan Hamarneh
- Medical Image Analysis Lab, Simon Fraser University, Burnaby, BC V5A 1S6, Canada.
| |
Collapse
|
45
|
Lasker A, Obaidullah SM, Chakraborty C, Roy K. Application of Machine Learning and Deep Learning Techniques for COVID-19 Screening Using Radiological Imaging: A Comprehensive Review. SN COMPUTER SCIENCE 2022; 4:65. [PMID: 36467853 PMCID: PMC9702883 DOI: 10.1007/s42979-022-01464-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/22/2022] [Accepted: 10/18/2022] [Indexed: 11/26/2022]
Abstract
Lung, being one of the most important organs in human body, is often affected by various SARS diseases, among which COVID-19 has been found to be the most fatal disease in recent times. In fact, SARS-COVID 19 led to pandemic that spreads fast among the community causing respiratory problems. Under such situation, radiological imaging-based screening [mostly chest X-ray and computer tomography (CT) modalities] has been performed for rapid screening of the disease as it is a non-invasive approach. Due to scarcity of physician/chest specialist/expert doctors, technology-enabled disease screening techniques have been developed by several researchers with the help of artificial intelligence and machine learning (AI/ML). It can be remarkably observed that the researchers have introduced several AI/ML/DL (deep learning) algorithms for computer-assisted detection of COVID-19 using chest X-ray and CT images. In this paper, a comprehensive review has been conducted to summarize the works related to applications of AI/ML/DL for diagnostic prediction of COVID-19, mainly using X-ray and CT images. Following the PRISMA guidelines, total 265 articles have been selected out of 1715 published articles till the third quarter of 2021. Furthermore, this review summarizes and compares varieties of ML/DL techniques, various datasets, and their results using X-ray and CT imaging. A detailed discussion has been made on the novelty of the published works, along with advantages and limitations.
Collapse
Affiliation(s)
- Asifuzzaman Lasker
- Department of Computer Science & Engineering, Aliah University, Kolkata, India
| | - Sk Md Obaidullah
- Department of Computer Science & Engineering, Aliah University, Kolkata, India
| | - Chandan Chakraborty
- Department of Computer Science & Engineering, National Institute of Technical Teachers’ Training & Research Kolkata, Kolkata, India
| | - Kaushik Roy
- Department of Computer Science, West Bengal State University, Barasat, India
| |
Collapse
|
46
|
Liu L, Liu Y, Zhou J, Guo C, Duan H. A novel MCF-Net: Multi-level context fusion network for 2D medical image segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 226:107160. [PMID: 36191351 DOI: 10.1016/j.cmpb.2022.107160] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Revised: 08/14/2022] [Accepted: 09/25/2022] [Indexed: 06/16/2023]
Abstract
Medical image segmentation is a crucial step in the clinical applications for diagnosis and analysis of some diseases. U-Net-based convolution neural networks have achieved impressive performance in medical image segmentation tasks. However, the multi-level contextual information integration capability and the feature extraction ability are often insufficient. In this paper, we present a novel multi-level context fusion network (MCF-Net) to improve the performance of U-Net on various segmentation tasks by designing three modules, hybrid attention-based residual atrous convolution (HARA) module, multi-scale feature memory (MSFM) module, and multi-receptive field fusion (MRFF) module, to fuse multi-scale contextual information. HARA module was proposed to effectively extract multi-receptive field features by combing atrous spatial pyramid pooling and attention mechanism. We further design the MSFM and MRFF modules to fuse features of different levels and effectively extract contextual information. The proposed MCF-Net was evaluated on the ISIC 2018, DRIVE, BUSI, and Kvasir-SEG datasets, which have challenging images of many sizes and widely varying anatomy. The experimental results show that MCF-Net is very competitive with other U-Net models, and it offers tremendous potential as a general-purpose deep learning model for 2D medical image segmentation.
Collapse
Affiliation(s)
- Lizhu Liu
- Engineering Research Center of Automotive Electrics and Control Technology, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China; National Engineering Laboratory of Robot Visual Perception and Control Technology, School of Robotics, Hunan University, Changsha 410082, China.
| | - Yexin Liu
- Engineering Research Center of Automotive Electrics and Control Technology, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China.
| | - Jian Zhou
- Engineering Research Center of Automotive Electrics and Control Technology, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China.
| | - Cheng Guo
- Engineering Research Center of Automotive Electrics and Control Technology, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China.
| | - Huigao Duan
- Engineering Research Center of Automotive Electrics and Control Technology, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China.
| |
Collapse
|
47
|
Li W, Zhang L, Wu C, Cui Z, Niu C. A new lightweight deep neural network for surface scratch detection. THE INTERNATIONAL JOURNAL, ADVANCED MANUFACTURING TECHNOLOGY 2022; 123:1999-2015. [PMID: 36313192 PMCID: PMC9596349 DOI: 10.1007/s00170-022-10335-8] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/07/2022] [Accepted: 10/15/2022] [Indexed: 06/16/2023]
Abstract
This paper aims to develop a lightweight convolutional neural network, WearNet, to realise automatic scratch detection for components in contact sliding such as those in metal forming. To this end, a large surface scratch dataset obtained from cylinder-on-flat sliding tests was used to train the WearNet with appropriate training parameters such as learning rate, gradient algorithm and mini-batch size. A comprehensive investigation on the network response and decision mechanism was also conducted to show the capability of the developed WearNet. It was found that compared with the existing networks, WearNet can realise an excellent classification accuracy of 94.16% with a much smaller model size and faster detection speed. Besides, WearNet outperformed other state-of-the-art networks when a public image database was used for network evaluation. The application of WearNet in an embedded system further demonstrated such advantages in the detection of surface scratches in sheet metal forming processes.
Collapse
Affiliation(s)
- Wei Li
- School of Mechanical and Manufacturing Engineering, The University of New South Wales, Kensington, NSW 2052 Australia
| | - Liangchi Zhang
- Shenzhen Key Laboratory of Cross-Scale Manufacturing Mechanics, Southern University of Science and Technology, Shenzhen, Guangdong, 518055 China
- SUSTech Institute for Manufacturing Innovation, Southern University of Science and Technology, Shenzhen, Guangdong, 518055 China
- Department of Mechanics and Aerospace Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, 518055 China
| | - Chuhan Wu
- School of Mechanical and Manufacturing Engineering, The University of New South Wales, Kensington, NSW 2052 Australia
| | - Zhenxiang Cui
- Baoshan Iron & Steel Co., Ltd., Shanghai, 200941 China
| | - Chao Niu
- Baoshan Iron & Steel Co., Ltd., Shanghai, 200941 China
| |
Collapse
|
48
|
Suri JS, Agarwal S, Saba L, Chabert GL, Carriero A, Paschè A, Danna P, Mehmedović A, Faa G, Jujaray T, Singh IM, Khanna NN, Laird JR, Sfikakis PP, Agarwal V, Teji JS, R Yadav R, Nagy F, Kincses ZT, Ruzsa Z, Viskovic K, Kalra MK. Multicenter Study on COVID-19 Lung Computed Tomography Segmentation with varying Glass Ground Opacities using Unseen Deep Learning Artificial Intelligence Paradigms: COVLIAS 1.0 Validation. J Med Syst 2022; 46:62. [PMID: 35988110 PMCID: PMC9392994 DOI: 10.1007/s10916-022-01850-y] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Accepted: 08/02/2022] [Indexed: 11/09/2022]
Abstract
Variations in COVID-19 lesions such as glass ground opacities (GGO), consolidations, and crazy paving can compromise the ability of solo-deep learning (SDL) or hybrid-deep learning (HDL) artificial intelligence (AI) models in predicting automated COVID-19 lung segmentation in Computed Tomography (CT) from unseen data leading to poor clinical manifestations. As the first study of its kind, "COVLIAS 1.0-Unseen" proves two hypotheses, (i) contrast adjustment is vital for AI, and (ii) HDL is superior to SDL. In a multicenter study, 10,000 CT slices were collected from 72 Italian (ITA) patients with low-GGO, and 80 Croatian (CRO) patients with high-GGO. Hounsfield Units (HU) were automatically adjusted to train the AI models and predict from test data, leading to four combinations-two Unseen sets: (i) train-CRO:test-ITA, (ii) train-ITA:test-CRO, and two Seen sets: (iii) train-CRO:test-CRO, (iv) train-ITA:test-ITA. COVILAS used three SDL models: PSPNet, SegNet, UNet and six HDL models: VGG-PSPNet, VGG-SegNet, VGG-UNet, ResNet-PSPNet, ResNet-SegNet, and ResNet-UNet. Two trained, blinded senior radiologists conducted ground truth annotations. Five types of performance metrics were used to validate COVLIAS 1.0-Unseen which was further benchmarked against MedSeg, an open-source web-based system. After HU adjustment for DS and JI, HDL (Unseen AI) > SDL (Unseen AI) by 4% and 5%, respectively. For CC, HDL (Unseen AI) > SDL (Unseen AI) by 6%. The COVLIAS-MedSeg difference was < 5%, meeting regulatory guidelines.Unseen AI was successfully demonstrated using automated HU adjustment. HDL was found to be superior to SDL.
Collapse
Affiliation(s)
- Jasjit S Suri
- Stroke Diagnostic and Monitoring Division, AtheroPoint™, Roseville, CA, USA.
- Advanced Knowledge Engineering Centre, GBTI, Roseville, CA, USA.
| | - Sushant Agarwal
- Advanced Knowledge Engineering Centre, GBTI, Roseville, CA, USA
- Department of Computer Science Engineering, Pranveer Singh Institute of Technology, Kanpur, Uttar Pradesh, India
| | - Luca Saba
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Gian Luca Chabert
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Alessandro Carriero
- Depart of Radiology, "Maggiore Della Carità" Hospital, University of Piemonte Orientale, Via Solaroli 17, 28100, Novara, Italy
| | - Alessio Paschè
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Pietro Danna
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | | | - Gavino Faa
- Department of Pathology - AOU of Cagliari, Cagliari, Italy
| | - Tanay Jujaray
- Advanced Knowledge Engineering Centre, GBTI, Roseville, CA, USA
- Dept of Molecular, Cell and Developmental Biology, University of California, Santa Cruz, CA, USA
| | - Inder M Singh
- Stroke Diagnostic and Monitoring Division, AtheroPoint™, Roseville, CA, USA
| | - Narendra N Khanna
- Department of Cardiology, Indraprastha APOLLO Hospitals, New Delhi, India
| | - John R Laird
- Heart and Vascular Institute, Adventist Health St. Helena, St Helena, CA, USA
| | - Petros P Sfikakis
- Rheumatology Unit, National Kapodistrian University of Athens, Athens, Greece
| | | | - Jagjit S Teji
- Ann and Robert H. Lurie Children's Hospital of Chicago, Chicago, USA
| | | | - Ferenc Nagy
- Internal Medicine Department, University of Szeged, Szeged, 6725, Hungary
| | | | - Zoltan Ruzsa
- Invasive Cardiology Division, University of Szeged, Budapest, Hungary
| | | | - Mannudeep K Kalra
- Department of Radiology, Massachusetts General Hospital, 55 Fruit Street, Boston, MA, USA
| |
Collapse
|
49
|
Agarwal M, Agarwal S, Saba L, Chabert GL, Gupta S, Carriero A, Pasche A, Danna P, Mehmedovic A, Faa G, Shrivastava S, Jain K, Jain H, Jujaray T, Singh IM, Turk M, Chadha PS, Johri AM, Khanna NN, Mavrogeni S, Laird JR, Sobel DW, Miner M, Balestrieri A, Sfikakis PP, Tsoulfas G, Misra DP, Agarwal V, Kitas GD, Teji JS, Al-Maini M, Dhanjil SK, Nicolaides A, Sharma A, Rathore V, Fatemi M, Alizad A, Krishnan PR, Yadav RR, Nagy F, Kincses ZT, Ruzsa Z, Naidu S, Viskovic K, Kalra MK, Suri JS. Eight pruning deep learning models for low storage and high-speed COVID-19 computed tomography lung segmentation and heatmap-based lesion localization: A multicenter study using COVLIAS 2.0. Comput Biol Med 2022; 146:105571. [PMID: 35751196 PMCID: PMC9123805 DOI: 10.1016/j.compbiomed.2022.105571] [Citation(s) in RCA: 25] [Impact Index Per Article: 8.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2022] [Revised: 04/05/2022] [Accepted: 04/26/2022] [Indexed: 12/12/2022]
Abstract
BACKGROUND COVLIAS 1.0: an automated lung segmentation was designed for COVID-19 diagnosis. It has issues related to storage space and speed. This study shows that COVLIAS 2.0 uses pruned AI (PAI) networks for improving both storage and speed, wiliest high performance on lung segmentation and lesion localization. METHOD ology: The proposed study uses multicenter ∼9,000 CT slices from two different nations, namely, CroMed from Croatia (80 patients, experimental data), and NovMed from Italy (72 patients, validation data). We hypothesize that by using pruning and evolutionary optimization algorithms, the size of the AI models can be reduced significantly, ensuring optimal performance. Eight different pruning techniques (i) differential evolution (DE), (ii) genetic algorithm (GA), (iii) particle swarm optimization algorithm (PSO), and (iv) whale optimization algorithm (WO) in two deep learning frameworks (i) Fully connected network (FCN) and (ii) SegNet were designed. COVLIAS 2.0 was validated using "Unseen NovMed" and benchmarked against MedSeg. Statistical tests for stability and reliability were also conducted. RESULTS Pruning algorithms (i) FCN-DE, (ii) FCN-GA, (iii) FCN-PSO, and (iv) FCN-WO showed improvement in storage by 92.4%, 95.3%, 98.7%, and 99.8% respectively when compared against solo FCN, and (v) SegNet-DE, (vi) SegNet-GA, (vii) SegNet-PSO, and (viii) SegNet-WO showed improvement by 97.1%, 97.9%, 98.8%, and 99.2% respectively when compared against solo SegNet. AUC > 0.94 (p < 0.0001) on CroMed and > 0.86 (p < 0.0001) on NovMed data set for all eight EA model. PAI <0.25 s per image. DenseNet-121-based Grad-CAM heatmaps showed validation on glass ground opacity lesions. CONCLUSIONS Eight PAI networks that were successfully validated are five times faster, storage efficient, and could be used in clinical settings.
Collapse
Affiliation(s)
- Mohit Agarwal
- Department of Computer Science Engineering, Bennett University, India
| | - Sushant Agarwal
- Department of Computer Science Engineering, PSIT, Kanpur, India; Advanced Knowledge Engineering Centre, Global Biomedical Technologies, Inc., Roseville, CA 95661, USA
| | - Luca Saba
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Gian Luca Chabert
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Suneet Gupta
- Department of Computer Science Engineering, Bennett University, India
| | - Alessandro Carriero
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Alessio Pasche
- Depart of Radiology, "Maggiore della Carità" Hospital, University of Piemonte Orientale, Via Solaroli 17, 28100, Novara, Italy
| | - Pietro Danna
- Depart of Radiology, "Maggiore della Carità" Hospital, University of Piemonte Orientale, Via Solaroli 17, 28100, Novara, Italy
| | | | - Gavino Faa
- Department of Pathology - AOU of Cagliari, Italy
| | - Saurabh Shrivastava
- College of Computing Sciences and IT, Teerthanker Mahaveer University, Moradabad, 244001, India
| | - Kanishka Jain
- College of Computing Sciences and IT, Teerthanker Mahaveer University, Moradabad, 244001, India
| | - Harsh Jain
- College of Computing Sciences and IT, Teerthanker Mahaveer University, Moradabad, 244001, India
| | - Tanay Jujaray
- Dept of Molecular, Cell and Developmental Biology, University of California, Santa Cruz, CA, USA
| | | | - Monika Turk
- The Hanse-Wissenschaftskolleg Institute for Advanced Study, Delmenhorst, Germany
| | | | - Amer M Johri
- Division of Cardiology, Queen's University, Kingston, Ontario, Canada
| | - Narendra N Khanna
- Department of Cardiology, Indraprastha APOLLO Hospitals, New Delhi, India
| | - Sophie Mavrogeni
- Cardiology Clinic, Onassis Cardiac Surgery Center, Athens, Greece
| | - John R Laird
- Heart and Vascular Institute, Adventist Health St. Helena, St Helena, CA, USA
| | - David W Sobel
- Minimally Invasive Urology Institute, Brown University, Providence, RI, USA
| | - Martin Miner
- Men's Health Center, Miriam Hospital Providence, Rhode Island, USA
| | - Antonella Balestrieri
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), Cagliari, Italy
| | - Petros P Sfikakis
- Rheumatology Unit, National Kapodistrian University of Athens, Greece
| | - George Tsoulfas
- Aristoteleion University of Thessaloniki, Thessaloniki, Greece
| | | | | | - George D Kitas
- Academic Affairs, Dudley Group NHS Foundation Trust, Dudley, UK; Arthritis Research UK Epidemiology Unit, Manchester University, Manchester, UK
| | - Jagjit S Teji
- Ann and Robert H. Lurie Children's Hospital of Chicago, Chicago, USA
| | - Mustafa Al-Maini
- Allergy, Clinical Immunology and Rheumatology Institute, Toronto, Canada
| | | | - Andrew Nicolaides
- Vascular Screening and Diagnostic Centre and Univ. of Nicosia Medical School, Cyprus
| | - Aditya Sharma
- Division of Cardiovascular Medicine, University of Virginia, Charlottesville, VA, USA
| | | | - Mostafa Fatemi
- Dept. of Physiology & Biomedical Engg., Mayo Clinic College of Medicine and Science, MN, USA
| | - Azra Alizad
- Dept. of Radiology, Mayo Clinic College of Medicine and Science, MN, USA
| | | | | | - Frence Nagy
- Department of Radiology, University of Szeged, 6725, Hungary
| | | | - Zoltan Ruzsa
- Invasive Cardiology Division, University of Szeged, Budapest, Hungary
| | - Subbaram Naidu
- Electrical Engineering Department, University of Minnesota, Duluth, MN, USA
| | | | - Manudeep K Kalra
- Department of Radiology, Massachusetts General Hospital, Boston, MA, USA
| | - Jasjit S Suri
- College of Computing Sciences and IT, Teerthanker Mahaveer University, Moradabad, 244001, India; Stroke Diagnostic and Monitoring Division, AtheroPoint™, Roseville, CA, USA.
| |
Collapse
|
50
|
Heidari A, Toumaj S, Navimipour NJ, Unal M. A privacy-aware method for COVID-19 detection in chest CT images using lightweight deep conventional neural network and blockchain. Comput Biol Med 2022; 145:105461. [PMID: 35366470 PMCID: PMC8958272 DOI: 10.1016/j.compbiomed.2022.105461] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2022] [Revised: 03/13/2022] [Accepted: 03/24/2022] [Indexed: 12/16/2022]
Abstract
With the global spread of the COVID-19 epidemic, a reliable method is required for identifying COVID-19 victims. The biggest issue in detecting the virus is a lack of testing kits that are both reliable and affordable. Due to the virus's rapid dissemination, medical professionals have trouble finding positive patients. However, the next real-life issue is sharing data with hospitals around the world while considering the organizations' privacy concerns. The primary worries for training a global Deep Learning (DL) model are creating a collaborative platform and personal confidentiality. Another challenge is exchanging data with health care institutions while protecting the organizations' confidentiality. The primary concerns for training a universal DL model are creating a collaborative platform and preserving privacy. This paper provides a model that receives a small quantity of data from various sources, like organizations or sections of hospitals, and trains a global DL model utilizing blockchain-based Convolutional Neural Networks (CNNs). In addition, we use the Transfer Learning (TL) technique to initialize layers rather than initialize randomly and discover which layers should be removed before selection. Besides, the blockchain system verifies the data, and the DL method trains the model globally while keeping the institution's confidentiality. Furthermore, we gather the actual and novel COVID-19 patients. Finally, we run extensive experiments utilizing Python and its libraries, such as Scikit-Learn and TensorFlow, to assess the proposed method. We evaluated works using five different datasets, including Boukan Dr. Shahid Gholipour hospital, Tabriz Emam Reza hospital, Mahabad Emam Khomeini hospital, Maragheh Dr.Beheshti hospital, and Miandoab Abbasi hospital datasets, and our technique outperform state-of-the-art methods on average in terms of precision (2.7%), recall (3.1%), F1 (2.9%), and accuracy (2.8%).
Collapse
Affiliation(s)
- Arash Heidari
- Department of Computer Engineering, Tabriz Branch, Islamic Azad University, Tabriz, Iran; Department of Computer Engineering, Shabestar Branch, Islamic Azad University, Shabestar, Iran
| | - Shiva Toumaj
- Urmia University of Medical Sciences, Urmia, Iran
| | | | - Mehmet Unal
- Department of Computer Engineering, Nisantasi University, Istanbul, Turkey
| |
Collapse
|