1
|
Shang Q, Wang G, Wang X, Li Y, Wang H. S-Net: A novel shallow network for enhanced detail retention in medical image segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 265:108730. [PMID: 40184853 DOI: 10.1016/j.cmpb.2025.108730] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/10/2024] [Revised: 03/17/2025] [Accepted: 03/18/2025] [Indexed: 04/07/2025]
Abstract
BACKGROUND AND OBJECTIVE In recent years, deep U-shaped network architectures have been widely applied to medical image segmentation tasks, achieving notable successes. However, the inherent limitation of this architecture is that multiple down-sampling lead to significant loss of input image detail information. A series of improvements in skip connections designed to enhance information transfer have not fundamentally resolved the issue. Therefore, we consider retaining information in a simpler and more effective way. METHODS In this paper, we propose a novel shallow network, S-Net, which contains only two output resolution stages, allowing for the preservation of more detailed information from the input images. To address the challenge of shallow networks primarily relying on high-resolution feature maps as the main information flow, we propose a Global-Local Feature Fusion (GLFF) module at the network bottleneck layer. This module integrates the superior global contextual information extraction capabilities of Mamba with the local feature capturing abilities of multi-scale depthwise convolutions, enabling the extraction of crucial semantic features from high-resolution feature maps within a shallow network architecture, while maintaining a smaller model size. RESULTS Extensive experiments on four different types of medical image datasets show that S-Net achieves the best segmentation performance compared to existing models, with more refined segmentation details. For example, on ultrasound datasets (BUSI), the IOU is 2.95% higher and DICE is 2.27% higher than the second-best model. Additionally, S-Net has only 1.52M parameters, making it competitive in terms of lightweight design. CONCLUSIONS Comparative and ablation experiments demonstrate the efficiency of the proposed architecture and modules. It shows that we do not need many down-sampling operations to reduce the size of feature maps significantly. This work provides new research ideas for further improving the accuracy of medical image segmentation and expands the research direction for model lightweight design. The code will be available at: https://github.com/qinghua0715/S-Net.
Collapse
Affiliation(s)
- Qinghua Shang
- College of Electronic and Information Engineering, Hebei University, Hebei 071002, PR China
| | - Guanglei Wang
- College of Electronic and Information Engineering, Hebei University, Hebei 071002, PR China; Hebei Key Laboratory of Precise Imaging of Inflammation Related Tumors, Hebei, 071000, PR China.
| | - Xihao Wang
- College of Electronic and Information Engineering, Hebei University, Hebei 071002, PR China
| | - Yan Li
- College of Electronic and Information Engineering, Hebei University, Hebei 071002, PR China
| | - Hongrui Wang
- College of Electronic and Information Engineering, Hebei University, Hebei 071002, PR China
| |
Collapse
|
2
|
Ding J, Xu W, Shu X, Wang W, Chen S, Wu Y. CMSAF-Net: integrative network design with enhanced decoder for precision segmentation of pear leaf diseases. PLANT METHODS 2025; 21:74. [PMID: 40448197 DOI: 10.1186/s13007-025-01392-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/09/2024] [Accepted: 05/14/2025] [Indexed: 06/02/2025]
Abstract
Pear leaf diseases represent one of the major challenges in agriculture, significantly affecting fruit quality and reducing overall yield. With the advancement of precision agriculture, accurate identification and segmentation of diseased areas are critical for targeted disease management and optimizing crop production. To address these issues, this study proposes a novel segmentation model, CMSAF-Net, for pear leaf diseases. CMSAF-Net integrates a Multi-scale Convolutional Attention Module (MBCA), a Self-adaptive Attention-augmented Upsampling Module (SAUP), and a Cross-layer Feature Alignment Module (CGAG) to enhance feature extraction, preserve edge information in complex disease regions, and optimize cross-layer information fusion. Additionally, CMSAF-Net incorporates pre-trained weights to leverage prior knowledge, accelerating convergence and improving segmentation accuracy. On a self-constructed dataset containing three types of pear leaf diseases, experimental results demonstrate that CMSAF-Net achieves 88.65%, 93.36%, and 93.86% in key metrics of MIoU, MPA, and Dice, respectively. Compared with mainstream models such as Unet++, DeepLabv3+, U2 -Net, and TransUNet, CMSAF-Net exhibits significant performance improvements, with MIoU increases of 2.45%, 3.86%, 2.21%, and 8.28%, respectively. This study highlights CMSAF-Net's potential for large-scale disease monitoring in intelligent agriculture, providing an efficient segmentation solution with substantial theoretical and practical implications.
Collapse
Affiliation(s)
- Jie Ding
- School of Information and Artificial Intelligence, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
- Anhui Beidou Precision Agriculture Information Engineering Research Center, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
| | - Wenwen Xu
- School of Information and Artificial Intelligence, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
- Anhui Beidou Precision Agriculture Information Engineering Research Center, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
| | - Xin Shu
- School of Information and Artificial Intelligence, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
- Anhui Beidou Precision Agriculture Information Engineering Research Center, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
| | - Wenyu Wang
- School of Information and Artificial Intelligence, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
- Anhui Beidou Precision Agriculture Information Engineering Research Center, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
| | - Shuxia Chen
- School of Information and Artificial Intelligence, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
- Anhui Beidou Precision Agriculture Information Engineering Research Center, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China
| | - Yunzhi Wu
- School of Information and Artificial Intelligence, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China.
- Anhui Beidou Precision Agriculture Information Engineering Research Center, Anhui Agricultural University, West Changjiang Road, 230036, Hefei, AH, China.
| |
Collapse
|
3
|
Gao Y, Jiang Y, Peng Y, Yuan F, Zhang X, Wang J. Medical Image Segmentation: A Comprehensive Review of Deep Learning-Based Methods. Tomography 2025; 11:52. [PMID: 40423254 DOI: 10.3390/tomography11050052] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2025] [Revised: 04/23/2025] [Accepted: 04/28/2025] [Indexed: 05/28/2025] Open
Abstract
Medical image segmentation is a critical application of computer vision in the analysis of medical images. Its primary objective is to isolate regions of interest in medical images from the background, thereby assisting clinicians in accurately identifying lesions, their sizes, locations, and their relationships with surrounding tissues. However, compared to natural images, medical images present unique challenges, such as low resolution, poor contrast, inconsistency, and scattered target regions. Furthermore, the accuracy and stability of segmentation results are subject to more stringent requirements. In recent years, with the widespread application of Convolutional Neural Networks (CNNs) in computer vision, deep learning-based methods for medical image segmentation have become a focal point of research. This paper categorizes, reviews, and summarizes the current representative methods and research status in the field of medical image segmentation. A comparative analysis of relevant experiments is presented, along with an introduction to commonly used public datasets, performance evaluation metrics, and loss functions in medical image segmentation. Finally, potential future research directions and development trends in this field are predicted and analyzed.
Collapse
Affiliation(s)
- Yuxiao Gao
- College of Artificial Intelligence, Taiyuan University of Technology, Jinzhong 036000, China
| | - Yang Jiang
- College of Mechanical Engineering, Chongqing University of Technology, Chongqing 400054, China
| | - Yanhong Peng
- College of Mechanical Engineering, Chongqing University of Technology, Chongqing 400054, China
| | - Fujiang Yuan
- School of Computer Science and Technology, Taiyuan Normal University, Taiyuan 030619, China
| | - Xinyue Zhang
- College of Artificial Intelligence, Taiyuan University of Technology, Jinzhong 036000, China
| | - Jianfeng Wang
- School of Software, Taiyuan University of Technology, Jinzhong 036000, China
| |
Collapse
|
4
|
Zheng S, Ye X, Yang C, Yu L, Li W, Gao X, Zhao Y. Asymmetric Adaptive Heterogeneous Network for Multi-Modality Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:1836-1852. [PMID: 40031190 DOI: 10.1109/tmi.2025.3526604] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Existing studies of multi-modality medical image segmentation tend to aggregate all modalities without discrimination and employ multiple symmetric encoders or decoders for feature extraction and fusion. They often overlook the different contributions to visual representation and intelligent decisions among multi-modality images. Motivated by this discovery, this paper proposes an asymmetric adaptive heterogeneous network for multi-modality image feature extraction with modality discrimination and adaptive fusion. For feature extraction, it uses a heterogeneous two-stream asymmetric feature-bridging network to extract complementary features from auxiliary multi-modality and leading single-modality images, respectively. For feature adaptive fusion, the proposed Transformer-CNN Feature Alignment and Fusion (T-CFAF) module enhances the leading single-modality information, and the Cross-Modality Heterogeneous Graph Fusion (CMHGF) module further fuses multi-modality features at a high-level semantic layer adaptively. Comparative evaluation with ten segmentation models on six datasets demonstrates significant efficiency gains as well as highly competitive segmentation accuracy. (Our code is publicly available at https://github.com/joker-527/AAHN).
Collapse
|
5
|
Ji Z, Chen Z, Ma X. Grouped multi-scale vision transformer for medical image segmentation. Sci Rep 2025; 15:11122. [PMID: 40169823 PMCID: PMC11961587 DOI: 10.1038/s41598-025-95361-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2025] [Accepted: 03/20/2025] [Indexed: 04/03/2025] Open
Abstract
Medical image segmentation plays a pivotal role in clinical diagnosis and pathological research by delineating regions of interest within medical images. While early approaches based on Convolutional Neural Networks (CNNs) have achieved significant success, their limited receptive field constrains their ability to capture long-range dependencies. Recent advances in Vision Transformers (ViTs) have demonstrated remarkable improvements by leveraging self-attention mechanisms. However, existing ViT-based segmentation models often struggle to effectively capture multi-scale variations within a single attention layer, limiting their capacity to model complex anatomical structures. To address this limitation, we propose Grouped Multi-Scale Attention (GMSA), which enhances multi-scale feature representation by grouping channels and performing self-attention at different scales within a single layer. Additionally, we introduce Inter-Scale Attention (ISA) to facilitate cross-scale feature fusion, further improving segmentation performance. Extensive experiments on the Synapse, ACDC, and ISIC2018 datasets demonstrate the effectiveness of our model, achieving state-of-the-art results in medical image segmentation. Our code is available at: https://github.com/Chen2zheng/ScaleFormer .
Collapse
Affiliation(s)
- Zexuan Ji
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, 210094, China
| | - Zheng Chen
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, 210094, China.
| | - Xiao Ma
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, 210094, China
| |
Collapse
|
6
|
Ji Z, Ye Y, Ma X. BDFormer: Boundary-aware dual-decoder transformer for skin lesion segmentation. Artif Intell Med 2025; 162:103079. [PMID: 39983372 DOI: 10.1016/j.artmed.2025.103079] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2024] [Revised: 01/17/2025] [Accepted: 02/05/2025] [Indexed: 02/23/2025]
Abstract
Segmenting skin lesions from dermatoscopic images is crucial for improving the quantitative analysis of skin cancer. However, automatic segmentation of skin lesions remains a challenging task due to the presence of unclear boundaries, artifacts, and obstacles such as hair and veins, all of which complicate the segmentation process. Transformers have demonstrated superior capabilities in capturing long-range dependencies through self-attention mechanisms and are gradually replacing CNNs in this domain. However, one of their primary limitations is the inability to effectively capture local details, which is crucial for handling unclear boundaries and significantly affects segmentation accuracy. To address this issue, we propose a novel boundary-aware dual-decoder transformer that employs a single encoder and dual-decoder framework for both skin lesion segmentation and dilated boundary segmentation. Within this model, we introduce a shifted window cross-attention block to build the dual-decoder structure and apply multi-task distillation to enable efficient interaction of inter-task information. Additionally, we propose a multi-scale aggregation strategy to refine the extracted features, ensuring optimal predictions. To further enhance boundary details, we incorporate a dilated boundary loss function, which expands the single-pixel boundary mask into planar information. We also introduce a task-wise consistency loss to promote consistency across tasks. Our method is evaluated on three datasets: ISIC2018, ISIC2017, and PH2, yielding promising results with excellent performance compared to state-of-the-art models. The code is available at https://github.com/Yuxuan-Ye/BDFormer.
Collapse
Affiliation(s)
- Zexuan Ji
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China
| | - Yuxuan Ye
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China.
| | - Xiao Ma
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China
| |
Collapse
|
7
|
Zhao L, Wang T, Chen Y, Zhang X, Tang H, Lin F, Li C, Li Q, Tan T, Kang D, Tong T. A novel framework for segmentation of small targets in medical images. Sci Rep 2025; 15:9924. [PMID: 40121297 PMCID: PMC11929788 DOI: 10.1038/s41598-025-94437-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2024] [Accepted: 03/13/2025] [Indexed: 03/25/2025] Open
Abstract
Medical image segmentation represents a pivotal and intricate procedure in the domain of medical image processing and analysis. With the progression of artificial intelligence in recent years, the utilization of deep learning techniques for medical image segmentation has witnessed escalating popularity. Nevertheless, the intricate nature of medical image poses challenges on the segmentation of diminutive targets is still in its early stages. Current networks encounter difficulties in addressing the segmentation of exceedingly small targets, especially when the number of training samples is limited. To overcome this constraint, we have implemented a proficient strategy to enhance lesion images containing small targets and constrained samples. We introduce a segmentation framework termed STS-Net, specifically designed for small target segmentation. This framework leverages the established capacity of convolutional neural networks to acquire effective image representations. The proposed STS-Net network adopts a ResNeXt50-32x4d architecture as the encoder, integrating attention mechanisms during the encoding phase to amplify the feature representation capabilities of the network. We evaluated the proposed network on four publicly available datasets. Experimental results underscore the superiority of our approach in the domain of medical image segmentation, particularly for small target segmentation. The codes are available at https://github.com/zlxokok/STSNet .
Collapse
Affiliation(s)
- Longxuan Zhao
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, 350100, China.
- Fujian Key Lab of Medical Instrumentation and Pharmaceutical Technology, Fuzhou, 350100, China.
| | - Tao Wang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, 350100, China
- Fujian Key Lab of Medical Instrumentation and Pharmaceutical Technology, Fuzhou, 350100, China
| | - Yuanbin Chen
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, 350100, China
- Fujian Key Lab of Medical Instrumentation and Pharmaceutical Technology, Fuzhou, 350100, China
| | - Xinlin Zhang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, 350100, China
- Fujian Key Lab of Medical Instrumentation and Pharmaceutical Technology, Fuzhou, 350100, China
- Imperial Vision Technology, Fuzhou, 350100, China
| | - Hui Tang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, 350100, China
- Fujian Key Lab of Medical Instrumentation and Pharmaceutical Technology, Fuzhou, 350100, China
| | - Fuxin Lin
- Department of Neurosurgery, Neurosurgery Research Institute, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
- Department of Neurosurgery, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
- Department of Neurosurgery, Fujian Institute of Brain Disorders and Brain Science, Fujian Clinical Research Center for Neurological Diseases, The First Affiliated Hospital and Neurosurgery Research Institute, Fujian Medical University, Fuzhou, 350100, China
- Fujian Provincial Clinical Research Center for Neurological Diseases, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
- Clinical Research and Translation Center, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
| | - Chunwang Li
- Department of Neurosurgery, Neurosurgery Research Institute, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
- Department of Neurosurgery, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
| | - Qixuan Li
- Department of Neurosurgery, Neurosurgery Research Institute, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
- Department of Neurosurgery, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China
| | - Tao Tan
- Macao Polytechnic University, Macao, 999078, China
| | - Dezhi Kang
- Department of Neurosurgery, Neurosurgery Research Institute, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China.
- Department of Neurosurgery, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China.
- Department of Neurosurgery, Fujian Institute of Brain Disorders and Brain Science, Fujian Clinical Research Center for Neurological Diseases, The First Affiliated Hospital and Neurosurgery Research Institute, Fujian Medical University, Fuzhou, 350100, China.
- Fujian Provincial Clinical Research Center for Neurological Diseases, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China.
- Clinical Research and Translation Center, The First Affiliated Hospital, Fujian Medical University, Fuzhou, 350100, China.
| | - Tong Tong
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, 350100, China.
- Fujian Key Lab of Medical Instrumentation and Pharmaceutical Technology, Fuzhou, 350100, China.
- Imperial Vision Technology, Fuzhou, 350100, China.
| |
Collapse
|
8
|
Ovi TB, Bashree N, Nyeem H, Wahed MA. FocusU 2Net: Pioneering dual attention with gated U-Net for colonoscopic polyp segmentation. Comput Biol Med 2025; 186:109617. [PMID: 39793349 DOI: 10.1016/j.compbiomed.2024.109617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2024] [Revised: 12/09/2024] [Accepted: 12/22/2024] [Indexed: 01/13/2025]
Abstract
The detection and excision of colorectal polyps, precursors to colorectal cancer (CRC), can improve survival rates by up to 90%. Automated polyp segmentation in colonoscopy images expedites diagnosis and aids in the precise identification of adenomatous polyps, thus mitigating the burden of manual image analysis. This study introduces FocusU2Net, an innovative bi-level nested U-structure integrated with a dual-attention mechanism. The model integrates Focus Gate (FG) modules for spatial and channel-wise attention and Residual U-blocks (RSU) with multi-scale receptive fields for capturing diverse contextual information. Comprehensive evaluations on five benchmark datasets - Kvasir-SEG, CVC-ClinicDB, CVC-ColonDB, ETISLarib, and EndoScene - demonstrate Dice score improvements of 3.14% to 43.59% over state-of-the-art models, with an 85% success rate in cross-dataset validations, significantly surpassing prior competing models with sub-5% success rates. The model combines high segmentation accuracy with computational efficiency, featuring 46.64 million parameters, 78.09 GFLOPs, and 39.02 GMacs, making it suitable for real-time applications. Enhanced with Explainable AI techniques, FocusU2Net provides clear insights into its decision-making process, improving interpretability. This combination of high performance, efficiency, and transparency positions FocusU2Net as a powerful, scalable solution for automated polyp segmentation in clinical practice, advancing medical image analysis and computer-aided diagnosis.
Collapse
Affiliation(s)
- Tareque Bashar Ovi
- Department of EECE, Military Institute of Science and Technology (MIST), Mirpur Cantonment, Dhaka, 1216, Bangladesh.
| | - Nomaiya Bashree
- Department of EECE, Military Institute of Science and Technology (MIST), Mirpur Cantonment, Dhaka, 1216, Bangladesh.
| | - Hussain Nyeem
- Department of EECE, Military Institute of Science and Technology (MIST), Mirpur Cantonment, Dhaka, 1216, Bangladesh.
| | - Md Abdul Wahed
- Department of EECE, Military Institute of Science and Technology (MIST), Mirpur Cantonment, Dhaka, 1216, Bangladesh.
| |
Collapse
|
9
|
Alam MS, Wang D, Arzhaeva Y, Ende JA, Kao J, Silverstone L, Yates D, Salvado O, Sowmya A. Attention-based multi-residual network for lung segmentation in diseased lungs with custom data augmentation. Sci Rep 2024; 14:28983. [PMID: 39578613 PMCID: PMC11584877 DOI: 10.1038/s41598-024-79494-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2024] [Accepted: 11/11/2024] [Indexed: 11/24/2024] Open
Abstract
Lung disease analysis in chest X-rays (CXR) using deep learning presents significant challenges due to the wide variation in lung appearance caused by disease progression and differing X-ray settings. While deep learning models have shown remarkable success in segmenting lungs from CXR images with normal or mildly abnormal findings, their performance declines when faced with complex structures, such as pulmonary opacifications. In this study, we propose AMRU++, an attention-based multi-residual UNet++ network designed for robust and accurate lung segmentation in CXR images with both normal and severe abnormalities. The model incorporates attention modules to capture relevant spatial information and multi-residual blocks to extract rich contextual and discriminative features of lung regions. To further enhance segmentation performance, we introduce a data augmentation technique that simulates the features and characteristics of CXR pathologies, addressing the issue of limited annotated data. Extensive experiments on public and private datasets comprising 350 cases of pneumoconiosis, COVID-19, and tuberculosis validate the effectiveness of our proposed framework and data augmentation technique.
Collapse
Affiliation(s)
- Md Shariful Alam
- School of Computer Science and Engineering, University of New South Wales, Sydney, Australia.
| | | | | | - Jesse Alexander Ende
- Department of Radiology, St Vincent's Hospital Sydney, Darlinghurst, NSW, 2010, Australia
| | - Joanna Kao
- Department of Radiology, St Vincent's Hospital Sydney, Darlinghurst, NSW, 2010, Australia
| | - Liz Silverstone
- Department of Radiology, St Vincent's Hospital Sydney, Darlinghurst, NSW, 2010, Australia
| | - Deborah Yates
- Department of Thoracic Medicine, St Vincent's Hospital Sydney, Darlinghurst, NSW, 2010, Australia
| | - Olivier Salvado
- School of Electrical Engineering & Robotics, Queensland University of Technology, Brisbane, QLD, 4001, Australia
| | - Arcot Sowmya
- School of Computer Science and Engineering, University of New South Wales, Sydney, Australia
| |
Collapse
|