1
|
Zakaria I, Yus TM, Rahman S, Gani A, Ersan MA. Assessing Fracture Detection: A Comparison of Minimal-Resource and Standard-Resource Plain Radiographic Interpretations. Diagnostics (Basel) 2025; 15:876. [PMID: 40218227 PMCID: PMC11988379 DOI: 10.3390/diagnostics15070876] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2025] [Revised: 03/20/2025] [Accepted: 03/23/2025] [Indexed: 04/14/2025] Open
Abstract
Background: The accuracy of fracture diagnosis through radiographic imaging largely depends on image quality and the interpreter's experience. In resource-limited settings (minimal-resource settings), imaging quality is often lower than in standard-resource facilities, potentially affecting diagnostic accuracy. Objective: This study aims to compare the diagnostic accuracy of plain radiograph interpretations between minimal-resource and standard-resource methods and assess the influence of interpreter experience on diagnostic precision. Methods: This cross-sectional study is based on secondary data from patients' medical records at the Dr. Zainoel Abidin General Hospital (RSUDZA) Banda Aceh, Indonesia. Comparisons between minimal-resource and standard-resource interpretations were made and validated using a reference standard (gold standard). Statistical analyses included diagnostic testing, Chi-square tests, and ROC curve analysis to evaluate sensitivity, specificity, and accuracy. Results: The findings indicate that standard-resource radiographs have significantly higher accuracy than minimal-resource radiographs (p < 0.05). Radiologists demonstrated the highest diagnostic accuracy compared to general practitioners and radiology residents. Conclusions: The standard-resource method is superior in detecting fractures compared to the minimal-resource method. Enhancing imaging quality and providing additional training for medical personnel are essential to improve diagnostic accuracy in resource-limited settings.
Collapse
Affiliation(s)
- Iskandar Zakaria
- Department of Radiology, Faculty of Medicine, Universitas Syiah Kuala, Banda Aceh 23111, Indonesia;
| | - Teuku Muhammad Yus
- Department of Radiology, Faculty of Medicine, Universitas Syiah Kuala, Banda Aceh 23111, Indonesia;
| | - Safrizal Rahman
- Department of Orthopedic and Traumatology, Faculty of Medicine, Universitas Syiah Kuala, Banda Aceh 23111, Indonesia;
| | - Azhari Gani
- Department of Internal Medicine, Faculty of Medicine, Universitas Syiah Kuala, Banda Aceh 23111, Indonesia;
| | - Muhammad Ariq Ersan
- Bachelor of Medicine Program, Faculty of Medicine, Universitas Syiah Kuala, Banda Aceh 23111, Indonesia;
| |
Collapse
|
2
|
Yoon SH, Kim J, Kim J, Lee JH, Choi I, Shin CW, Park CM. Improving Image Quality of Chest Radiography with Artificial Intelligence-Supported Dual-Energy X-Ray Imaging System: An Observer Preference Study in Healthy Volunteers. J Clin Med 2025; 14:2091. [PMID: 40142899 PMCID: PMC11942644 DOI: 10.3390/jcm14062091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2025] [Revised: 03/13/2025] [Accepted: 03/17/2025] [Indexed: 03/28/2025] Open
Abstract
Background/Objectives: To compare the image quality of chest radiography with a dual-energy X-ray imaging system using AI technology (DE-AI) to that of conventional chest radiography with a standard protocol. Methods: In this prospective study, 52 healthy volunteers underwent dual-energy chest radiography. Images were obtained using two exposures at 60 kVp and 120 kVp, separated by a 150 ms interval. Four images were generated for each participant: a conventional image, an enhanced standard image, a soft-tissue-selective image, and a bone-selective image. A machine learning model optimized the cancellation parameters for generating soft-tissue and bone-selective images. To enhance image quality, motion artifacts were minimized using Laplacian pyramid diffeomorphic registration, while a wavelet directional cycle-consistent adversarial network (WavCycleGAN) reduced image noise. Four radiologists independently evaluated the visibility of thirteen anatomical regions (eight soft-tissue regions and five bone regions) and the overall image with a five-point scale of preference. Pooled mean values were calculated for each anatomic region through meta-analysis using a random-effects model. Results: Radiologists preferred DE-AI images to conventional chest radiographs in various anatomic regions. The enhanced standard image showed superior quality in 9 of 13 anatomic regions. Preference for the soft-tissue-selective image was statistically significant for three of eight anatomic regions. Preference for the bone-selective image was statistically significant for four of five anatomic regions. Conclusions: Images produced by DE-AI provide better visualization of thoracic structures.
Collapse
Affiliation(s)
- Sung-Hyun Yoon
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam-si 13620, Republic of Korea
| | - Jihang Kim
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam-si 13620, Republic of Korea
| | - Junghoon Kim
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam-si 13620, Republic of Korea
| | - Jong-Hyuk Lee
- Department of Radiology, Seoul National University Hospital, Seoul National University College of Medicine, Seoul 03080, Republic of Korea
| | - Ilwoong Choi
- Research & Development Center, DRTECH Corp., Seongnam-si 13606, Republic of Korea
| | - Choul-Woo Shin
- Research & Development Center, DRTECH Corp., Seongnam-si 13606, Republic of Korea
| | - Chang-Min Park
- Department of Radiology, Seoul National University Hospital, Seoul National University College of Medicine, Seoul 03080, Republic of Korea
| |
Collapse
|
3
|
Ibrahim S, Selim S, Elattar M. Facilitating Radiograph Interpretation: Refined Generative Models for Precise Bone Suppression in Chest X-rays. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01461-2. [PMID: 40082331 DOI: 10.1007/s10278-025-01461-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/28/2024] [Revised: 02/17/2025] [Accepted: 02/18/2025] [Indexed: 03/16/2025]
Abstract
Chest X-ray (CXR) is crucial for diagnosing lung diseases, especially lung nodules. Recent studies indicate that bones, such as ribs and clavicles, obscure 82 to 95% of undiagnosed lung cancers. The development of computer-aided detection (CAD) systems with automated bone suppression is vital to improve detection rates and support early clinical decision-making. Current bone suppression methods face challenges: they often depend on manual subtraction of bone-only images from CXRs, leading to inefficiency and poor generalization; there is significant information loss in data compression within deep convolutional end-to-end architectures; and a balance between model efficiency and accuracy has not been sufficiently achieved in existing research. We introduce a novel end-to-end architecture, the mask-guided model, to address these challenges. Leveraging the Pix2Pix framework, our model enhances computational efficiency by reducing parameter count by 92.5%. It features a rib mask-guided module with a mask encoder and cross-attention mechanism, which provides spatial constraints, reduces information loss during encoder compression, and preserves non-relevant areas. An ablation study evaluates the impact of various factors. The model undergoes initial training on digitally reconstructed radiographs (DRRs) derived from CT projections for bone suppression and is fine-tuned on the JSRT dataset to accelerate convergence. The mask-guided model surpasses previous state-of-the-art methods, showing superior bone suppression performance in terms of structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), and processing speed. It achieves an SSIM of 0.99 ± 0.002 and a PSNR of 36.14 ± 1.13 on the JSRT dataset. This study underscores the proposed model's effectiveness compared to existing methods, showcasing its capability to reduce model size and increase accuracy. This makes it well-suited for deployment in affordable, low-power hardware devices across various clinical settings.
Collapse
Affiliation(s)
- Samar Ibrahim
- Medical Imaging and Image Processing Research Group, Center for Informatics Science (CIS), Nile University, 26th of July Corridor, Sheikh Zayed City, Giza, 12588, Egypt
| | - Sahar Selim
- Medical Imaging and Image Processing Research Group, Center for Informatics Science (CIS), Nile University, 26th of July Corridor, Sheikh Zayed City, Giza, 12588, Egypt.
- School of Information Technology and Computer Science, Nile University, 26th of July Corridor, Sheikh Zayed City, Giza, 12588, Egypt.
| | - Mustafa Elattar
- Medical Imaging and Image Processing Research Group, Center for Informatics Science (CIS), Nile University, 26th of July Corridor, Sheikh Zayed City, Giza, 12588, Egypt
- School of Information Technology and Computer Science, Nile University, 26th of July Corridor, Sheikh Zayed City, Giza, 12588, Egypt
| |
Collapse
|
4
|
Xie W, Gan M, Tan X, Li M, Yang W, Wang W. Efficient labeling for fine-tuning chest X-ray bone-suppression networks for pediatric patients. Med Phys 2025; 52:978-992. [PMID: 39546640 PMCID: PMC11788263 DOI: 10.1002/mp.17516] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2024] [Revised: 10/07/2024] [Accepted: 10/25/2024] [Indexed: 11/17/2024] Open
Abstract
BACKGROUND Pneumonia, a major infectious cause of morbidity and mortality among children worldwide, is typically diagnosed using low-dose pediatric chest X-ray [CXR (chest radiography)]. In pediatric CXR images, bone occlusion leads to a risk of missed diagnosis. Deep learning-based bone-suppression networks relying on training data have enabled considerable progress to be achieved in bone suppression in adult CXR images; however, these networks have poor generalizability to pediatric CXR images because of the lack of labeled pediatric CXR images (i.e., bone images vs. soft-tissue images). Dual-energy subtraction imaging approaches are capable of producing labeled adult CXR images; however, their application is limited because they require specialized equipment, and they are infrequently employed in pediatric settings. Traditional image processing-based models can be used to label pediatric CXR images, but they are semiautomatic and have suboptimal performance. PURPOSE We developed an efficient labeling approach for fine-tuning pediatric CXR bone-suppression networks capable of automatically suppressing bone structures in CXR images for pediatric patients without the need for specialized equipment and technologist training. METHODS Three steps were employed to label pediatric CXR images and fine-tune pediatric bone-suppression networks: distance transform-based bone-edge detection, traditional image processing-based bone suppression, and fully automated pediatric bone suppression. In distance transform-based bone-edge detection, bone edges were automatically detected by predicting bone-edge distance-transform images, which were then used as inputs in traditional image processing. In this processing, pediatric CXR images were labeled by obtaining bone images through a series of traditional image processing techniques. Finally, the pediatric bone-suppression network was fine-tuned using the labeled pediatric CXR images. This network was initially pretrained on a public adult dataset comprising 240 adult CXR images (A240) and then fine-tuned and validated on 40 pediatric CXR images (P260_40labeled) from our customized dataset (named P260) through five-fold cross-validation; finally, the network was tested on 220 pediatric CXR images (P260_220unlabeled dataset). RESULTS The distance transform-based bone-edge detection network achieved a mean boundary distance of 1.029. Moreover, the traditional image processing-based bone-suppression model obtained bone images exhibiting a relative Weber contrast of 93.0%. Finally, the fully automated pediatric bone-suppression network achieved a relative mean absolute error of 3.38%, a peak signal-to-noise ratio of 35.5 dB, a structural similarity index measure of 98.1%, and a bone-suppression ratio of 90.1% on P260_40labeled. CONCLUSIONS The proposed fully automated pediatric bone-suppression network, together with the proposed distance transform-based bone-edge detection network, can automatically acquire bone and soft-tissue images solely from CXR images for pediatric patients and has the potential to help diagnose pneumonia in children.
Collapse
Affiliation(s)
- Weijie Xie
- Information and Data Centre, Guangzhou First People's HospitalGuangzhou Medical UniversityGuangzhouChina
- Information and Data Centre, the Second Affiliated Hospital, School of MedicineSouth China University of TechnologyGuangzhouChina
| | - Mengkun Gan
- Information and Data Centre, Guangzhou First People's HospitalGuangzhou Medical UniversityGuangzhouChina
- Information and Data Centre, the Second Affiliated Hospital, School of MedicineSouth China University of TechnologyGuangzhouChina
| | - Xiaocong Tan
- Information and Data Centre, Guangzhou First People's HospitalGuangzhou Medical UniversityGuangzhouChina
- Information and Data Centre, the Second Affiliated Hospital, School of MedicineSouth China University of TechnologyGuangzhouChina
| | - Mujiao Li
- Information and Data Centre, Guangzhou First People's HospitalGuangzhou Medical UniversityGuangzhouChina
- Information and Data Centre, the Second Affiliated Hospital, School of MedicineSouth China University of TechnologyGuangzhouChina
| | - Wei Yang
- School of Biomedical EngineeringSouthern Medical UniversityGuangzhouChina
- Guangdong Provincial Key Laboratory of Medical Image ProcessingSchool of Biomedical Engineering, Southern Medical UniversityGuangzhouChina
| | - Wenhui Wang
- Information and Data Centre, Guangzhou First People's HospitalGuangzhou Medical UniversityGuangzhouChina
- Information and Data Centre, the Second Affiliated Hospital, School of MedicineSouth China University of TechnologyGuangzhouChina
| |
Collapse
|
5
|
Huang Z, Li H, Shao S, Zhu H, Hu H, Cheng Z, Wang J, Kevin Zhou S. PELE scores: pelvic X-ray landmark detection with pelvis extraction and enhancement. Int J Comput Assist Radiol Surg 2024; 19:939-950. [PMID: 38491244 DOI: 10.1007/s11548-024-03089-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2023] [Accepted: 02/27/2024] [Indexed: 03/18/2024]
Abstract
PURPOSE Pelvic X-ray (PXR) is widely utilized in clinical decision-making associated with the pelvis, the lower part of the trunk that supports and balances the trunk. In particular, PXR-based landmark detection facilitates downstream analysis and computer-assisted diagnosis and treatment of pelvic diseases. Although PXR has the advantages of low radiation and reduced cost compared to computed tomography (CT), it characterizes the 2D pelvis-tissue superposition of 3D structures, which may affect the accuracy of landmark detection in some cases. However, the superposition nature of PXR is implicitly handled by existing deep learning-based landmark detection methods, which mainly design the deep network structures for better detection performances. Explicit handling of the superposition nature of PXR is rarely done. METHODS In this paper, we explicitly focus on the superposition of X-ray images. Specifically, we propose a pelvis extraction (PELE) module that consists of a decomposition network, a domain adaptation network, and an enhancement module, which utilizes 3D prior anatomical knowledge in CT to guide and well isolate the pelvis from PXR, thereby eliminating the influence of soft tissue for landmark detection. The extracted pelvis image, after enhancement, is then used for landmark detection. RESULTS We conduct an extensive evaluation based on two public and one private dataset, totaling 850 PXRs. The experimental results show that the proposed PELE module significantly improves the accuracy of PXRs landmark detection and achieves state-of-the-art performances in several benchmark metrics. CONCLUSION The design of PELE module can improve the accuracy of different pelvic landmark detection baselines, which we believe is obviously conducive to the positioning and inspection of clinical landmarks and critical structures, thus better serving downstream tasks. Our project has been open-sourced at https://github.com/ECNUACRush/PELEscores .
Collapse
Affiliation(s)
- Zhen Huang
- Computer Science Department, University of Science and Technology of China (USTC), Hefei, 230026, China
- Center for Medical Imaging, Robotics, Analytic Computing and Learning (MIRACLE), Suzhou Institute for Advanced Research, USTC, Suzhou, 215123, China
| | - Han Li
- School of Biomedical Engineering, Division of Life Sciences and Medicine, USTC, Hefei, 230026, China
- Center for Medical Imaging, Robotics, Analytic Computing and Learning (MIRACLE), Suzhou Institute for Advanced Research, USTC, Suzhou, 215123, China
| | | | - Heqin Zhu
- School of Biomedical Engineering, Division of Life Sciences and Medicine, USTC, Hefei, 230026, China
- Center for Medical Imaging, Robotics, Analytic Computing and Learning (MIRACLE), Suzhou Institute for Advanced Research, USTC, Suzhou, 215123, China
| | - Huijie Hu
- Computer Science Department, University of Science and Technology of China (USTC), Hefei, 230026, China
- Center for Medical Imaging, Robotics, Analytic Computing and Learning (MIRACLE), Suzhou Institute for Advanced Research, USTC, Suzhou, 215123, China
| | | | - Jianji Wang
- Affiliated Hospital of Guizhou Medical University, Guiyang, 550000, China
| | - S Kevin Zhou
- School of Biomedical Engineering, Division of Life Sciences and Medicine, USTC, Hefei, 230026, China.
- Center for Medical Imaging, Robotics, Analytic Computing and Learning (MIRACLE), Suzhou Institute for Advanced Research, USTC, Suzhou, 215123, China.
- Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing, 100190, China.
| |
Collapse
|
6
|
Nakanishi N, Otake Y, Hiasa Y, Gu Y, Uemura K, Takao M, Sugano N, Sato Y. Decomposition of musculoskeletal structures from radiographs using an improved CycleGAN framework. Sci Rep 2023; 13:8482. [PMID: 37231008 DOI: 10.1038/s41598-023-35075-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2022] [Accepted: 05/12/2023] [Indexed: 05/27/2023] Open
Abstract
This paper presents methods of decomposition of musculoskeletal structures from radiographs into multiple individual muscle and bone structures. While existing solutions require dual-energy scan for the training dataset and are mainly applied to structures with high-intensity contrast, such as bones, we focused on multiple superimposed muscles with subtle contrast in addition to bones. The decomposition problem is formulated as an image translation problem between (1) a real X-ray image and (2) multiple digitally reconstructed radiographs, each of which contains a single muscle or bone structure, and solved using unpaired training based on the CycleGAN framework. The training dataset was created via automatic computed tomography (CT) segmentation of muscle/bone regions and virtually projecting them with geometric parameters similar to the real X-ray images. Two additional features were incorporated into the CycleGAN framework to achieve a high-resolution and accurate decomposition: hierarchical learning and reconstruction loss with the gradient correlation similarity metric. Furthermore, we introduced a new diagnostic metric for muscle asymmetry directly measured from a plain X-ray image to validate the proposed method. Our simulation and real-image experiments using real X-ray and CT images of 475 patients with hip diseases suggested that each additional feature significantly enhanced the decomposition accuracy. The experiments also evaluated the accuracy of muscle volume ratio measurement, which suggested a potential application to muscle asymmetry assessment from an X-ray image for diagnostic and therapeutic assistance. The improved CycleGAN framework can be applied for investigating the decomposition of musculoskeletal structures from single radiographs.
Collapse
Affiliation(s)
- Naoki Nakanishi
- Division of Information Science, Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Nara, 630-0192, Japan
| | - Yoshito Otake
- Division of Information Science, Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Nara, 630-0192, Japan.
| | - Yuta Hiasa
- Division of Information Science, Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Nara, 630-0192, Japan.
| | - Yi Gu
- Division of Information Science, Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Nara, 630-0192, Japan
| | - Keisuke Uemura
- Department of Orthopaedic Medical Engineering, Osaka University Graduate School of Medicine, Suita, Osaka, 565-0871, Japan
| | - Masaki Takao
- Department of Bone and Joint Surgery, Ehime University Graduate School of Medicine, Toon, Ehime, 791-0295, Japan
| | - Nobuhiko Sugano
- Department of Orthopaedic Medical Engineering, Osaka University Graduate School of Medicine, Suita, Osaka, 565-0871, Japan
| | - Yoshinobu Sato
- Division of Information Science, Graduate School of Science and Technology, Nara Institute of Science and Technology, Ikoma, Nara, 630-0192, Japan.
| |
Collapse
|
7
|
van Tulder G, de Bruijne M. Unpaired, unsupervised domain adaptation assumes your domains are already similar. Med Image Anal 2023; 87:102825. [PMID: 37116296 DOI: 10.1016/j.media.2023.102825] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2022] [Revised: 03/30/2023] [Accepted: 04/17/2023] [Indexed: 04/30/2023]
Abstract
Unsupervised domain adaptation is a popular method in medical image analysis, but it can be tricky to make it work: without labels to link the domains, domains must be matched using feature distributions. If there is no additional information, this often leaves a choice between multiple possibilities to map the data that may be equally likely but not equally correct. In this paper we explore the fundamental problems that may arise in unsupervised domain adaptation, and discuss conditions that might still make it work. Focusing on medical image analysis, we argue that images from different domains may have similar class balance, similar intensities, similar spatial structure, or similar textures. We demonstrate how these implicit conditions can affect domain adaptation performance in experiments with synthetic data, MNIST digits, and medical images. We observe that practical success of unsupervised domain adaptation relies on existing similarities in the data, and is anything but guaranteed in the general case. Understanding these implicit assumptions is a key step in identifying potential problems in domain adaptation and improving the reliability of the results.
Collapse
Affiliation(s)
- Gijs van Tulder
- Data Science group, Faculty of Science, Radboud University, Postbus 9010, 6500 GL Nijmegen, The Netherlands; Biomedical Imaging Group, Erasmus MC, Postbus 2040, 3000 CA Rotterdam, The Netherlands.
| | - Marleen de Bruijne
- Biomedical Imaging Group, Erasmus MC, Postbus 2040, 3000 CA Rotterdam, The Netherlands; Department of Computer Science, University of Copenhagen, Universitetsparken 1, 2100 Copenhagen, Denmark.
| |
Collapse
|
8
|
Liu Y, Zeng F, Ma M, Zheng B, Yun Z, Qin G, Yang W, Feng Q. Bone suppression of lateral chest x-rays with imperfect and limited dual-energy subtraction images. Comput Med Imaging Graph 2023; 105:102186. [PMID: 36731328 DOI: 10.1016/j.compmedimag.2023.102186] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2022] [Revised: 01/06/2023] [Accepted: 01/06/2023] [Indexed: 01/22/2023]
Abstract
Bone suppression is to suppress the superimposed bone components over the soft tissues within the lung area of Chest X-ray (CXR), which is potentially useful for the subsequent lung disease diagnosis for radiologists, as well as computer-aided systems. Despite bone suppression methods for frontal CXRs being well studied, it remains challenging for lateral CXRs due to the limited and imperfect DES dataset containing paired lateral CXR and soft-tissue/bone images and more complex anatomical structures in the lateral view. In this work, we propose a bone suppression method for lateral CXRs by leveraging a two-stage distillation learning strategy and a specific data correction method. Specifically, a primary model is first trained on a real DES dataset with limited samples. The bone-suppressed results on a relatively large lateral CXR dataset produced by the primary model are improved by a designed gradient correction method. Secondly, the corrected results serve as training samples to train the distillated model. By automatically learning knowledge from both the primary model and the extra correction procedure, our distillated model is expected to promote the performance of the primary model while omitting the tedious correction procedure. We adopt an ensemble model named MsDd-MAP for the primary and distillated models, which learns the complementary information of Multi-scale and Dual-domain (i.e., intensity and gradient) and fuses them in a maximum-a-posteriori (MAP) framework. Our method is evaluated on a two-exposure lateral DES dataset consisting of 46 subjects and a lateral CXR dataset consisting of 240 subjects. The experimental results suggest that our method is superior to other competing methods regarding the quantitative evaluation metrics. Furthermore, the subjective evaluation by three experienced radiologists also indicates that the distillated model can produce more visually appealing soft-tissue images than the primary model, even comparable to real DES imaging for lateral CXRs.
Collapse
Affiliation(s)
- Yunbi Liu
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; School of Science and Engineering, The Chinese University of Hong Kong, Shenzhen, Shenzhen, Guangdong 518172, China; Shenzhen Research Institute of Big Data, Shenzhen, China; University of Science and Technology of China, Hefei, China
| | - Fengxia Zeng
- Radiology Department, Nanfang Hospital, Southern Medical University, Guangzhou 510515, China
| | - Mengwei Ma
- Radiology Department, Nanfang Hospital, Southern Medical University, Guangzhou 510515, China
| | - Bowen Zheng
- Radiology Department, Nanfang Hospital, Southern Medical University, Guangzhou 510515, China
| | - Zhaoqiang Yun
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China
| | - Genggeng Qin
- Radiology Department, Nanfang Hospital, Southern Medical University, Guangzhou 510515, China.
| | - Wei Yang
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China.
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China
| |
Collapse
|
9
|
Rajaraman S, Yang F, Zamzmi G, Xue Z, Antani S. Assessing the Impact of Image Resolution on Deep Learning for TB Lesion Segmentation on Frontal Chest X-rays. Diagnostics (Basel) 2023; 13:747. [PMID: 36832235 PMCID: PMC9955202 DOI: 10.3390/diagnostics13040747] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2023] [Revised: 02/10/2023] [Accepted: 02/15/2023] [Indexed: 02/18/2023] Open
Abstract
Deep learning (DL) models are state-of-the-art in segmenting anatomical and disease regions of interest (ROIs) in medical images. Particularly, a large number of DL-based techniques have been reported using chest X-rays (CXRs). However, these models are reportedly trained on reduced image resolutions for reasons related to the lack of computational resources. Literature is sparse in discussing the optimal image resolution to train these models for segmenting the tuberculosis (TB)-consistent lesions in CXRs. In this study, we investigated the performance variations with an Inception-V3 UNet model using various image resolutions with/without lung ROI cropping and aspect ratio adjustments and identified the optimal image resolution through extensive empirical evaluations to improve TB-consistent lesion segmentation performance. We used the Shenzhen CXR dataset for the study, which includes 326 normal patients and 336 TB patients. We proposed a combinatorial approach consisting of storing model snapshots, optimizing segmentation threshold and test-time augmentation (TTA), and averaging the snapshot predictions, to further improve performance with the optimal resolution. Our experimental results demonstrate that higher image resolutions are not always necessary; however, identifying the optimal image resolution is critical to achieving superior performance.
Collapse
Affiliation(s)
- Sivaramakrishnan Rajaraman
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894, USA
| | | | | | | | | |
Collapse
|
10
|
Chen C, Qi S, Zhou K, Lu T, Ning H, Xiao R. Pairwise attention-enhanced adversarial model for automatic bone segmentation in CT images. Phys Med Biol 2023; 68. [PMID: 36634367 DOI: 10.1088/1361-6560/acb2ab] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2022] [Accepted: 01/12/2023] [Indexed: 01/14/2023]
Abstract
Objective. Bone segmentation is a critical step in screw placement navigation. Although the deep learning methods have promoted the rapid development for bone segmentation, the local bone separation is still challenging due to irregular shapes and similar representational features.Approach. In this paper, we proposed the pairwise attention-enhanced adversarial model (Pair-SegAM) for automatic bone segmentation in computed tomography images, which includes the two parts of the segmentation model and discriminator. Considering that the distributions of the predictions from the segmentation model contains complicated semantics, we improve the discriminator to strengthen the awareness ability of the target region, improving the parsing of semantic information features. The Pair-SegAM has a pairwise structure, which uses two calculation mechanics to set up pairwise attention maps, then we utilize the semantic fusion to filter unstable regions. Therefore, the improved discriminator provides more refinement information to capture the bone outline, thus effectively enhancing the segmentation models for bone segmentation.Main results. To test the Pair-SegAM, we selected the two bone datasets for assessment. We evaluated our method against several bone segmentation models and latest adversarial models on the both datasets. The experimental results prove that our method not only exhibits superior bone segmentation performance, but also states effective generalization.Significance. Our method provides a more efficient segmentation of specific bones and has the potential to be extended to other semantic segmentation domains.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Siyu Qi
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Tong Lu
- Visual 3D Medical Science and Technology Development Co. Ltd, Beijing 100082, People's Republic of China
| | - Huansheng Ning
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, People's Republic of China.,Shunde Innovation School, University of Science and Technology Beijing, Foshan 100024, People's Republic of China
| |
Collapse
|
11
|
Xie C, Hu Y, Han L, Fu J, Vardhanabhuti V, Yang H. Prediction of Individual Lymph Node Metastatic Status in Esophageal Squamous Cell Carcinoma Using Routine Computed Tomography Imaging: Comparison of Size-Based Measurements and Radiomics-Based Models. Ann Surg Oncol 2022; 29:8117-8126. [PMID: 36018524 DOI: 10.1245/s10434-022-12207-7] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Accepted: 06/08/2022] [Indexed: 12/29/2022]
Abstract
BACKGROUND Lymph node status is vital for prognosis and treatment decisions for esophageal squamous cell carcinoma (ESCC). This study aimed to construct and evaluate an optimal radiomics-based method for a more accurate evaluation of individual regional lymph node status in ESCC and to compare it with traditional size-based measurements. METHODS The study consecutively collected 3225 regional lymph nodes from 530 ESCC patients receiving upfront surgery from January 2011 to October 2015. Computed tomography (CT) scans for individual lymph nodes were analyzed. The study evaluated the predictive performance of machine-learning models trained on features extracted from two-dimensional (2D) and three-dimensional (3D) radiomics by different contouring methods. Robust and important radiomics features were selected, and classification models were further established and validated. RESULTS The lymph node metastasis rate was 13.2% (427/3225). The average short-axis diameter was 6.4 mm for benign lymph nodes and 7.9 mm for metastatic lymph nodes. The division of lymph node stations into five regions according to anatomic lymph node drainage (cervical, upper mediastinal, middle mediastinal, lower mediastinal, and abdominal regions) improved the predictive performance. The 2D radiomics method showed optimal diagnostic results, with more efficient segmentation of nodal lesions. In the test set, this optimal model achieved an area under the receiver operating characteristic curve of 0.841-0.891, an accuracy of 84.2-94.7%, a sensitivity of 65.7-83.3%, and a specificity of 84.4-96.7%. CONCLUSIONS The 2D radiomics-based models noninvasively predicted the metastatic status of an individual lymph node in ESCC and outperformed the conventional size-based measurement. The 2D radiomics-based model could be incorporated into the current clinical workflow to enable better decision-making for treatment strategies.
Collapse
Affiliation(s)
- Chenyi Xie
- Department of Radiology, Guangdong Provincial People's Hospital, Guangdong Academy of Medical Sciences, Guangzhou, China.,Department of Diagnostic Radiology, Li Ka Shing Faculty of Medicine, University of Hong Kong, Pok Fu Lam, Hong Kong SAR, China
| | - Yihuai Hu
- Department of Thoracic Surgery, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Esophageal Cancer Institute, Sun Yat-sen University Cancer Center, Guangzhou, China.,Department of Thoracic Surgery, Guangdong Provincial People's Hospital, Guangdong Academy of Medical Sciences, Guangzhou, China
| | - Lujun Han
- Department of Radiology, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Jianhua Fu
- Department of Thoracic Surgery, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Esophageal Cancer Institute, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Varut Vardhanabhuti
- Department of Diagnostic Radiology, Li Ka Shing Faculty of Medicine, University of Hong Kong, Pok Fu Lam, Hong Kong SAR, China.
| | - Hong Yang
- Department of Thoracic Surgery, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Esophageal Cancer Institute, Sun Yat-sen University Cancer Center, Guangzhou, China.
| |
Collapse
|
12
|
Chen Y, Zhao X, Tang B. Boosting lesion annotation via aggregating explicit relations in external medical knowledge graph. Artif Intell Med 2022; 132:102376. [DOI: 10.1016/j.artmed.2022.102376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2021] [Revised: 05/09/2022] [Accepted: 08/17/2022] [Indexed: 11/29/2022]
|
13
|
Rani G, Misra A, Dhaka VS, Zumpano E, Vocaturo E. Spatial feature and resolution maximization GAN for bone suppression in chest radiographs. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 224:107024. [PMID: 35863123 DOI: 10.1016/j.cmpb.2022.107024] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/22/2022] [Revised: 06/29/2022] [Accepted: 07/12/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Chest radiographs (CXR) are in great demand for visualizing the pathology of the lungs. However, the appearance of bones in the lung region hinders the localization of any lesion or nodule present in the CXR. Thus, bone suppression becomes an important task for the effective screening of lung diseases. Simultaneously, it is equally important to preserve spatial information and image quality because they provide crucial insights on the size and area of infection, color accuracy, structural quality, etc. Many researchers considered bone suppression as an image denoising problem and proposed conditional Generative Adversarial Network-based (cGAN) models for generating bone suppressed images from CXRs. These works do not focus on the retention of spatial features and image quality. The authors of this manuscript developed the Spatial Feature and Resolution Maximization (SFRM) GAN to efficiently minimize the visibility of bones in CXRs while ensuring maximum retention of critical information. METHOD This task is achieved by modifying the architectures of the discriminator and generator of the pix2pix model. The discriminator is combined with the Wasserstein GAN with Gradient Penalty to increase its performance and training stability. For the generator, a combination of different task-specific loss functions, viz., L1, Perceptual, and Sobel loss are employed to capture the intrinsic information in the image. RESULT The proposed model reported as measures of performance a mean PSNR of 43.588, mean NMSE of 0.00025, mean SSIM of 0.989, and mean Entropy of 0.454 bits/pixel on a test size of 100 images. Further, the combination of δ=104, α=1, β=10, and γ=10 are the hyperparameters that provided the best trade-off between image denoising and quality retention. CONCLUSION The degree of bone suppression and spatial information preservation can be improved by adding the Sobel and Perceptual loss respectively. SFRM-GAN not only suppresses bones but also retains the image quality and intrinsic information. Based on the results of student's t-test it is concluded that SFRM-GAN yields statistically significant results at a 0.95 level of confidence and shows its supremacy over the state-of-the-art models. Thus, it may be used for denoising and preprocessing of images.
Collapse
Affiliation(s)
- Geeta Rani
- Department of Computer and Communication Engineering, Manipal University Jaipur, India.
| | - Ankit Misra
- Department of Computer Science and Engineering, Manipal University Jaipur, India; Goergen Institute for Data Science, University of Rochester, USA.
| | - Vijaypal Singh Dhaka
- Department of Computer and Communication Engineering, Manipal University Jaipur, India.
| | - Ester Zumpano
- Department of Computer Engineering, Modeling, Electronics and Systems Engineering, University of Calabria, Italy.
| | - Eugenio Vocaturo
- Department of Computer Engineering, Modeling, Electronics and Systems Engineering, University of Calabria, Italy.
| |
Collapse
|
14
|
Singh A, Lall B, Panigrahi BK, Agrawal A, Agrawal A, Thangakunam B, Christopher DJ. Semantic segmentation of bone structures in chest X-rays including unhealthy radiographs: A robust and accurate approach. Int J Med Inform 2022; 165:104831. [PMID: 35870303 DOI: 10.1016/j.ijmedinf.2022.104831] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2022] [Revised: 06/14/2022] [Accepted: 07/08/2022] [Indexed: 10/17/2022]
Abstract
The chest X-ray is a widely used medical imaging technique for the diagnosis of several lung diseases. Some nodules or other pathologies present in the lungs are difficult to visualize on chest X-rays because they are obscured byoverlying boneshadows. Segmentation of bone structures and suppressing them assist medical professionals in reliable diagnosis and organ morphometry. But segmentation of bone structures is challenging due to fuzzy boundaries of organs and inconsistent shape and size of organs due to health issues, age, and gender. The existing bone segmentation methods do not report their performance on abnormal chest X-rays, where it is even more critical to segment the bones. This work presents a robust encoder-decoder network for semantic segmentation of bone structures on normal as well as abnormal chest X-rays. The novelty here lies in combining techniques from two existing networks (Deeplabv3+ and U-net) to achieve robust and superior performance. The fully connected layers of the pre-trained ResNet50 network have been replaced by an Atrous spatial pyramid pooling block for improving the quality of the embedding in the encoder module. The decoder module includes four times upsampling blocks to connect both low-level and high-level features information enabling us to retain both the edges and detail information of the objects. At each level, the up-sampled decoder features are concatenated with the encoder features at a similar level and further fine-tuned to refine the segmentation output. We construct a diverse chest X-ray dataset with ground truth binary masks of anterior ribs, posterior ribs, and clavicle bone for experimentation. The dataset includes 100 samples of chest X-rays belonging to healthy and confirmed patients of lung diseases to maintain the diversity and test the robustness of our method. We test our method using multiple standard metrics and experimental results indicate an excellent performance on both normal and abnormal chest X-rays.
Collapse
Affiliation(s)
- Anushikha Singh
- Bharti School of Telecommunication Technology and Management, Indian Institute of Technology Delhi, New Delhi, India.
| | - Brejesh Lall
- Department of Electrical Engineering, Indian Institute of Technology Delhi, New Delhi, India.
| | - B K Panigrahi
- Department of Electrical Engineering, Indian Institute of Technology Delhi, New Delhi, India
| | - Anjali Agrawal
- Teleradiology Solutions, Civil Lines, Delhi 110054, India.
| | - Anurag Agrawal
- CSIR-Institute of Genomics and Integrative Biology, New Delhi 110025, India.
| | | | | |
Collapse
|
15
|
Cho K, Seo J, Kyung S, Kim M, Hong GS, Kim N. Bone suppression on pediatric chest radiographs via a deep learning-based cascade model. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 215:106627. [PMID: 35032722 DOI: 10.1016/j.cmpb.2022.106627] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/08/2021] [Revised: 12/05/2021] [Accepted: 01/07/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND AND OBJECTIVE Bone suppression images (BSIs) of chest radiographs (CXRs) have been proven to improve diagnosis of pulmonary diseases. To acquire BSIs, dual-energy subtraction (DES) or a deep-learning-based model trained with DES-based BSIs have been used. However, neither technique could be applied to pediatric patients owing to the harmful effects of DES. In this study, we developed a novel method for bone suppression in pediatric CXRs. METHODS First, a model using digitally reconstructed radiographs (DRRs) of adults, which were used to generate pseudo-CXRs from computed tomography images, was developed by training a 2-channel contrastive-unpaired-image-translation network. Second, this model was applied to 129 pediatric DRRs to generate the paired training data of pseudo-pediatric CXRs. Finally, by training a U-Net with these paired data, a bone suppression model for pediatric CXRs was developed. RESULTS The evaluation metrics were peak signal to noise ratio, root mean absolute error and structural similarity index measure at soft-tissue and bone region of the lung. In addition, an expert radiologist scored the effectiveness of BSIs on a scale of 1-5. The obtained result of 3.31 ± 0.48 indicates that the BSIs show homogeneous bone removal despite subtle residual bone shadow. CONCLUSION Our method shows that the pixel intensity at soft-tissue regions was preserved, and bones were well subtracted; this can be useful for detecting early pulmonary disease in pediatric CXRs.
Collapse
Affiliation(s)
- Kyungjin Cho
- Department of Biomedical Engineering, Asan Medical Institute of Convergence Science and Technology, Asan Medical Center, College of Medicine, University of Ulsan, Seoul, Republic of Korea
| | - Jiyeon Seo
- Department of Biomedical Engineering, Asan Medical Institute of Convergence Science and Technology, Asan Medical Center, College of Medicine, University of Ulsan, Seoul, Republic of Korea
| | - Sunggu Kyung
- Department of Biomedical Engineering, Asan Medical Institute of Convergence Science and Technology, Asan Medical Center, College of Medicine, University of Ulsan, Seoul, Republic of Korea
| | - Mingyu Kim
- Department of Convergence Medicine, University of Ulsan College of Medicine, Asan Medical Center, 88 Olympic-Ro 43-Gil Songpa-Gu, Seoul 05505, Korea
| | - Gil-Sun Hong
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine & Asan Medical Center, 88 Olympic-ro 43-gil, Songpa-gu Seoul 05505, Republic of Korea.
| | - Namkug Kim
- Department of Convergence Medicine, University of Ulsan College of Medicine, Asan Medical Center, 88 Olympic-Ro 43-Gil Songpa-Gu, Seoul 05505, Korea; Department of Radiology, Asan Medical Center, University of Ulsan College of Medicine, Seoul 05505, Republic of Korea.
| |
Collapse
|
16
|
Bi XA, Xing Z, Zhou W, Li L, Xu L. Pathogeny Detection for Mild Cognitive Impairment via Weighted Evolutionary Random Forest with Brain Imaging and Genetic Data. IEEE J Biomed Health Inform 2022; 26:3068-3079. [PMID: 35157601 DOI: 10.1109/jbhi.2022.3151084] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Medical imaging technology and gene sequencing technology have long been widely used to analyze the pathogenesis and make precise diagnoses of mild cognitive impairment (MCI). However, few studies involve the fusion of radiomics data with genomics data to make full use of the complementarity between different omics to detect pathogenic factors of MCI. This paper performs multimodal fusion analysis based on functional magnetic resonance imaging (fMRI) data and single nucleotide polymorphism (SNP) data of MCI patients. In specific, first, using correlation analysis methods on sequence information of regions of interests (ROIs) and digitalized gene sequences, the fusion features of samples are constructed. Then, introducing weighted evolution strategy into ensemble learning, a novel weighted evolutionary random forest (WERF) model is built to eliminate the inefficient features. Consequently, with the help of WERF, an overall multimodal data analysis framework is established to effectively identify MCI patients and extract pathogenic factors. Based on the data of MCI patients from the ADNI database and compared with some existing popular methods, the superiority in performance of the framework is verified. Our study has great potential to be an effective tool for pathogenic factors detection of MCI.
Collapse
|
17
|
GAN-based disentanglement learning for chest X-ray rib suppression. Med Image Anal 2022; 77:102369. [DOI: 10.1016/j.media.2022.102369] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2021] [Revised: 12/09/2021] [Accepted: 01/10/2022] [Indexed: 11/19/2022]
|
18
|
Chest X-ray Bone Suppression for Improving Classification of Tuberculosis-Consistent Findings. Diagnostics (Basel) 2021; 11:diagnostics11050840. [PMID: 34067034 PMCID: PMC8151767 DOI: 10.3390/diagnostics11050840] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2021] [Revised: 05/04/2021] [Accepted: 05/05/2021] [Indexed: 11/16/2022] Open
Abstract
Chest X-rays (CXRs) are the most commonly performed diagnostic examination to detect cardiopulmonary abnormalities. However, the presence of bony structures such as ribs and clavicles can obscure subtle abnormalities, resulting in diagnostic errors. This study aims to build a deep learning (DL)-based bone suppression model that identifies and removes these occluding bony structures in frontal CXRs to assist in reducing errors in radiological interpretation, including DL workflows, related to detecting manifestations consistent with tuberculosis (TB). Several bone suppression models with various deep architectures are trained and optimized using the proposed combined loss function and their performances are evaluated in a cross-institutional test setting using several metrics such as mean absolute error (MAE), peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM), and multiscale structural similarity measure (MS-SSIM). The best-performing model (ResNet-BS) (PSNR = 34.0678; MS-SSIM = 0.9828) is used to suppress bones in the publicly available Shenzhen and Montgomery TB CXR collections. A VGG-16 model is pretrained on a large collection of publicly available CXRs. The CXR-pretrained model is then fine-tuned individually on the non-bone-suppressed and bone-suppressed CXRs of Shenzhen and Montgomery TB CXR collections to classify them as showing normal lungs or TB manifestations. The performances of these models are compared using several performance metrics such as accuracy, the area under the curve (AUC), sensitivity, specificity, precision, F-score, and Matthews correlation coefficient (MCC), analyzed for statistical significance, and their predictions are qualitatively interpreted through class-selective relevance maps (CRMs). It is observed that the models trained on bone-suppressed CXRs (Shenzhen: AUC = 0.9535 ± 0.0186; Montgomery: AUC = 0.9635 ± 0.0106) significantly outperformed (p < 0.05) the models trained on the non-bone-suppressed CXRs (Shenzhen: AUC = 0.8991 ± 0.0268; Montgomery: AUC = 0.8567 ± 0.0870).. Models trained on bone-suppressed CXRs improved detection of TB-consistent findings and resulted in compact clustering of the data points in the feature space signifying that bone suppression improved the model sensitivity toward TB classification.
Collapse
|
19
|
Zhou SK, Greenspan H, Davatzikos C, Duncan JS, van Ginneken B, Madabhushi A, Prince JL, Rueckert D, Summers RM. A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises. PROCEEDINGS OF THE IEEE. INSTITUTE OF ELECTRICAL AND ELECTRONICS ENGINEERS 2021; 109:820-838. [PMID: 37786449 PMCID: PMC10544772 DOI: 10.1109/jproc.2021.3054390] [Citation(s) in RCA: 267] [Impact Index Per Article: 66.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/04/2023]
Abstract
Since its renaissance, deep learning has been widely used in various medical imaging tasks and has achieved remarkable success in many medical imaging applications, thereby propelling us into the so-called artificial intelligence (AI) era. It is known that the success of AI is mostly attributed to the availability of big data with annotations for a single task and the advances in high performance computing. However, medical imaging presents unique challenges that confront deep learning approaches. In this survey paper, we first present traits of medical imaging, highlight both clinical needs and technical challenges in medical imaging, and describe how emerging trends in deep learning are addressing these issues. We cover the topics of network architecture, sparse and noisy labels, federating learning, interpretability, uncertainty quantification, etc. Then, we present several case studies that are commonly found in clinical practice, including digital pathology and chest, brain, cardiovascular, and abdominal imaging. Rather than presenting an exhaustive literature survey, we instead describe some prominent research highlights related to these case study applications. We conclude with a discussion and presentation of promising future directions.
Collapse
Affiliation(s)
- S Kevin Zhou
- School of Biomedical Engineering, University of Science and Technology of China and Institute of Computing Technology, Chinese Academy of Sciences
| | - Hayit Greenspan
- Biomedical Engineering Department, Tel-Aviv University, Israel
| | - Christos Davatzikos
- Radiology Department and Electrical and Systems Engineering Department, University of Pennsylvania, USA
| | - James S Duncan
- Departments of Biomedical Engineering and Radiology & Biomedical Imaging, Yale University
| | | | - Anant Madabhushi
- Department of Biomedical Engineering, Case Western Reserve University and Louis Stokes Cleveland Veterans Administration Medical Center, USA
| | - Jerry L Prince
- Electrical and Computer Engineering Department, Johns Hopkins University, USA
| | - Daniel Rueckert
- Klinikum rechts der Isar, TU Munich, Germany and Department of Computing, Imperial College, UK
| | | |
Collapse
|
20
|
Zhou Y, Zhou T, Zhou T, Fu H, Liu J, Shao L. Contrast-Attentive Thoracic Disease Recognition With Dual-Weighting Graph Reasoning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1196-1206. [PMID: 33406037 DOI: 10.1109/tmi.2021.3049498] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Automatic thoracic disease diagnosis is a rising research topic in the medical imaging community, with many potential applications. However, the inconsistent appearances and high complexities of various lesions in chest X-rays currently hinder the development of a reliable and robust intelligent diagnosis system. Attending to the high-probability abnormal regions and exploiting the priori of a related knowledge graph offers one promising route to addressing these issues. As such, in this paper, we propose two contrastive abnormal attention models and a dual-weighting graph convolution to improve the performance of thoracic multi-disease recognition. First, a left-right lung contrastive network is designed to learn intra-attentive abnormal features to better identify the most common thoracic diseases, whose lesions rarely appear in both sides symmetrically. Moreover, an inter-contrastive abnormal attention model aims to compare the query scan with multiple anchor scans without lesions to compute the abnormal attention map. Once the intra- and inter-contrastive attentions are weighted over the features, in addition to the basic visual spatial convolution, a chest radiology graph is constructed for dual-weighting graph reasoning. Extensive experiments on the public NIH ChestX-ray and CheXpert datasets show that our model achieves consistent improvements over the state-of-the-art methods both on thoracic disease identification and localization.
Collapse
|