1
|
Zhang K, Liang W, Cao P, Mao Z, Yang J, Zaiane OR. CorLabelNet: a comprehensive framework for multi-label chest X-ray image classification with correlation guided discriminant feature learning and oversampling. Med Biol Eng Comput 2025; 63:1045-1058. [PMID: 39609353 DOI: 10.1007/s11517-024-03247-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Accepted: 11/12/2024] [Indexed: 11/30/2024]
Abstract
Recent advancements in deep learning techniques have significantly improved multi-label chest X-ray (CXR) image classification for clinical diagnosis. However, most previous studies neither effectively learn label correlations nor take full advantage of them to improve multi-label classification performance. In addition, different labels of CXR images are usually severely imbalanced, resulting in the model exhibiting a bias towards the majority class. To address these challenges, we introduce a framework that not only learns label correlations but also utilizes them to guide the learning of features and the process of oversampling. In this paper, our approach incorporates self-attention to capture high-order label correlations and considers label correlations from both global and local perspectives. Then, we propose a consistency constraint and a multi-label contrastive loss to enhance feature learning. To alleviate the imbalance issue, we further propose an oversampling approach that exploits the learned label correlation to identify crucial seed samples for oversampling. Our approach repeats 5-fold cross-validation process experiments three times and achieves the best performance on both the CheXpert and ChestX-Ray14 datasets. Learning accurate label correlation is significant for multi-label classification and taking full advantage of label correlations is beneficial for discriminative feature learning and oversampling. A comparative analysis with the state-of-the-art approaches highlights the effectiveness of our proposed methods.
Collapse
Affiliation(s)
- Kai Zhang
- Computer Science and Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China
| | - Wei Liang
- Computer Science and Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China
| | - Peng Cao
- Computer Science and Engineering, Northeastern University, Shenyang, China.
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China.
| | - Zhaoyang Mao
- Computer Science and Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China
| | - Jinzhu Yang
- Computer Science and Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China
| | - Osmar R Zaiane
- Alberta Machine Intelligence Institute, University of Alberta, Edmonton, Canada
| |
Collapse
|
2
|
Liu Y, Peng X, Wei X, Geng L, Zhang F, Xiao Z, Lin JCW. Label-Aware Dual Graph Neural Networks for Multi-Label Fundus Image Classification. IEEE J Biomed Health Inform 2025; 29:2731-2743. [PMID: 39255075 DOI: 10.1109/jbhi.2024.3457232] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/12/2024]
Abstract
Fundus disease is a complex and universal disease involving a variety of pathologies. Its early diagnosis using fundus images can effectively prevent further diseases and provide targeted treatment plans for patients. Recent deep learning models for classification of this disease are gradually emerging as a critical research field, which is attracting widespread attention. However, in practice, most of the existing methods only focus on local visual cues of a single image, and ignore the underlying explicit interaction similarity between subjects and correlation information among pathologies in fundus diseases. In this paper, we propose a novel label-aware dual graph neural networks for multi-label fundus image classification that consists of population-based graph representation learning and pathology-based graph representation learning modules. Specifically, we first construct a population-based graph by integrating image features and non-image information to learn patient's representations by incorporating associations between subjects. Then, we represent pathologies as a sparse graph where its nodes are associated with pathology-based feature vectors and the edges correspond to probability of the co-occurrence of labels to generate a set of classifier scores by the propagation of multi-layer graph information. Finally, our model can adaptively recalibrate multi-label outputs. Detailed experiments and analysis of our results show the effectiveness of our method compared with state-of-the-art multi-label fundus image classification methods.
Collapse
|
3
|
Zhang M, Hu X, Gu L, Liu L, Kobayashi K, Harada T, Yan Y, Summers RM, Zhu Y. A New Benchmark: Clinical Uncertainty and Severity Aware Labeled Chest X-Ray Images With Multi-Relationship Graph Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:338-347. [PMID: 39120990 DOI: 10.1109/tmi.2024.3441494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/11/2024]
Abstract
Chest radiography, commonly known as CXR, is frequently utilized in clinical settings to detect cardiopulmonary conditions. However, even seasoned radiologists might offer different evaluations regarding the seriousness and uncertainty associated with observed abnormalities. Previous research has attempted to utilize clinical notes to extract abnormal labels for training deep-learning models in CXR image diagnosis. However, these methods often neglected the varying degrees of severity and uncertainty linked to different labels. In our study, we initially assembled a comprehensive new dataset of CXR images based on clinical textual data, which incorporated radiologists' assessments of uncertainty and severity. Using this dataset, we introduced a multi-relationship graph learning framework that leverages spatial and semantic relationships while addressing expert uncertainty through a dedicated loss function. Our research showcases a notable enhancement in CXR image diagnosis and the interpretability of the diagnostic model, surpassing existing state-of-the-art methodologies. The dataset address of disease severity and uncertainty we extracted is: https://physionet.org/content/cad-chest/1.0/.
Collapse
|
4
|
Rocha J, Pereira SC, Sousa P, Campilho A, Mendonça AM. CLARE-XR: explainable regression-based classification of chest radiographs with label embeddings. Sci Rep 2024; 14:31024. [PMID: 39730802 DOI: 10.1038/s41598-024-82222-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2024] [Accepted: 12/03/2024] [Indexed: 12/29/2024] Open
Abstract
An automatic system for pathology classification in chest X-ray scans needs more than predictive performance, since providing explanations is deemed essential for fostering end-user trust, improving decision-making, and regulatory compliance. CLARE-XR is a novel methodology that, when presented with an X-ray image, identifies the associated pathologies and provides explanations based on the presentation of similar cases. The diagnosis is achieved using a regression model that maps an image into a 2D latent space containing the reference coordinates of all findings. The references are generated once through label embedding, before the regression step, by converting the original binary ground-truth annotations to 2D coordinates. The classification is inferred minding the distance from the coordinates of an inference image to the reference coordinates. Furthermore, as the regressor is trained on a known set of images, the distance from the coordinates of an inference image to the coordinates of the training set images also allows retrieving similar instances, mimicking the common clinical practice of comparing scans to confirm diagnoses. This inherently interpretable framework discloses specific classification rules and visual explanations through automatic image retrieval methods, outperforming the multi-label ResNet50 classification baseline across multiple evaluation settings on the NIH ChestX-ray14 dataset.
Collapse
Affiliation(s)
- Joana Rocha
- Institute for Systems and Computer Engineering Technology and Science (INESC-TEC), Porto, 4200-465, Portugal.
- Faculty of Engineering, University of Porto, Porto, 4200-465, Portugal.
| | - Sofia Cardoso Pereira
- Institute for Systems and Computer Engineering Technology and Science (INESC-TEC), Porto, 4200-465, Portugal
- Faculty of Engineering, University of Porto, Porto, 4200-465, Portugal
| | - Pedro Sousa
- Hospital Center of Vila Nova de Gaia/Espinho, Vila Nova de Gaia, 4430-000, Portugal
| | - Aurélio Campilho
- Institute for Systems and Computer Engineering Technology and Science (INESC-TEC), Porto, 4200-465, Portugal
- Faculty of Engineering, University of Porto, Porto, 4200-465, Portugal
| | - Ana Maria Mendonça
- Institute for Systems and Computer Engineering Technology and Science (INESC-TEC), Porto, 4200-465, Portugal
- Faculty of Engineering, University of Porto, Porto, 4200-465, Portugal
| |
Collapse
|
5
|
Xiao J, Li S, Lin T, Zhu J, Yuan X, Feng DD, Sheng B. Multi-Label Chest X-Ray Image Classification With Single Positive Labels. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:4404-4418. [PMID: 38949934 DOI: 10.1109/tmi.2024.3421644] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/03/2024]
Abstract
Deep learning approaches for multi-label Chest X-ray (CXR) images classification usually require large-scale datasets. However, acquiring such datasets with full annotations is costly, time-consuming, and prone to noisy labels. Therefore, we introduce a weakly supervised learning problem called Single Positive Multi-label Learning (SPML) into CXR images classification (abbreviated as SPML-CXR), in which only one positive label is annotated per image. A simple solution to SPML-CXR problem is to assume that all the unannotated pathological labels are negative, however, it might introduce false negative labels and decrease the model performance. To this end, we present a Multi-level Pseudo-label Consistency (MPC) framework for SPML-CXR. First, inspired by the pseudo-labeling and consistency regularization in semi-supervised learning, we construct a weak-to-strong consistency framework, where the model prediction on weakly-augmented image is treated as the pseudo label for supervising the model prediction on a strongly-augmented version of the same image, and define an Image-level Perturbation-based Consistency (IPC) regularization to recover the potential mislabeled positive labels. Besides, we incorporate Random Elastic Deformation (RED) as an additional strong augmentation to enhance the perturbation. Second, aiming to expand the perturbation space, we design a perturbation stream to the consistency framework at the feature-level and introduce a Feature-level Perturbation-based Consistency (FPC) regularization as a supplement. Third, we design a Transformer-based encoder module to explore the sample relationship within each mini-batch by a Batch-level Transformer-based Correlation (BTC) regularization. Extensive experiments on the CheXpert and MIMIC-CXR datasets have shown the effectiveness of our MPC framework for solving the SPML-CXR problem.
Collapse
|
6
|
Holste G, Zhou Y, Wang S, Jaiswal A, Lin M, Zhuge S, Yang Y, Kim D, Nguyen-Mau TH, Tran MT, Jeong J, Park W, Ryu J, Hong F, Verma A, Yamagishi Y, Kim C, Seo H, Kang M, Celi LA, Lu Z, Summers RM, Shih G, Wang Z, Peng Y. Towards long-tailed, multi-label disease classification from chest X-ray: Overview of the CXR-LT challenge. Med Image Anal 2024; 97:103224. [PMID: 38850624 PMCID: PMC11365790 DOI: 10.1016/j.media.2024.103224] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Revised: 04/01/2024] [Accepted: 05/27/2024] [Indexed: 06/10/2024]
Abstract
Many real-world image recognition problems, such as diagnostic medical imaging exams, are "long-tailed" - there are a few common findings followed by many more relatively rare conditions. In chest radiography, diagnosis is both a long-tailed and multi-label problem, as patients often present with multiple findings simultaneously. While researchers have begun to study the problem of long-tailed learning in medical image recognition, few have studied the interaction of label imbalance and label co-occurrence posed by long-tailed, multi-label disease classification. To engage with the research community on this emerging topic, we conducted an open challenge, CXR-LT, on long-tailed, multi-label thorax disease classification from chest X-rays (CXRs). We publicly release a large-scale benchmark dataset of over 350,000 CXRs, each labeled with at least one of 26 clinical findings following a long-tailed distribution. We synthesize common themes of top-performing solutions, providing practical recommendations for long-tailed, multi-label medical image classification. Finally, we use these insights to propose a path forward involving vision-language foundation models for few- and zero-shot disease classification.
Collapse
Affiliation(s)
- Gregory Holste
- Department of Electrical and Computer Engineering, The University of Texas at Austin, 78712, Austin, TX, USA
| | - Yiliang Zhou
- Department of Population Health Sciences, Weill Cornell Medicine, 10065, New York, NY, USA
| | - Song Wang
- Department of Electrical and Computer Engineering, The University of Texas at Austin, 78712, Austin, TX, USA
| | - Ajay Jaiswal
- Department of Electrical and Computer Engineering, The University of Texas at Austin, 78712, Austin, TX, USA
| | - Mingquan Lin
- Department of Population Health Sciences, Weill Cornell Medicine, 10065, New York, NY, USA
| | - Sherry Zhuge
- School of Information Systems, Carnegie Mellon University, 15213, Pittsburgh, PA, USA
| | - Yuzhe Yang
- Department of Electrical Engineering and Computer Science, Massachussetts Institute of Technology, 02139, Cambridge, MA, USA
| | - Dongkyun Kim
- School of Computer Science, Carnegie Mellon University, 15213, Pittsburgh, PA, USA
| | | | | | - Jaehyup Jeong
- KT Research & Development Center, KT Corporation, 06763, Seoul, South Korea
| | - Wongi Park
- Department of Software and Computer Engineering, Ajou University, 16499, Suwon, South Korea
| | - Jongbin Ryu
- Department of Software and Computer Engineering, Ajou University, 16499, Suwon, South Korea
| | - Feng Hong
- Cooperative Medianet Innovation Center, Shanghai Jiao Tong University, 200240, Shanghai, China
| | - Arsh Verma
- Wadhwani Institute for Artificial Intelligence, 400079, Mumbai, India
| | - Yosuke Yamagishi
- Division of Radiology and Biomedical Engineering, Graduate School of Medicine, The University of Tokyo, 113-0033, Tokyo, Japan
| | - Changhyun Kim
- BioMedical AI Team, AIX Future R&D Center, SK Telecom, 04539, Seoul, South Korea
| | - Hyeryeong Seo
- Interdisciplinary Program in AI (IPAI), Seoul National University, 02504, Seoul, South Korea
| | - Myungjoo Kang
- Department of Mathematical Sciences, Seoul National University, 02504, Seoul, South Korea
| | - Leo Anthony Celi
- Laboratory for Computational Physiology, Massachusetts Institute of Technology, 02139, Cambridge, MA, USA; Division of Pulmonary, Critical Care and Sleep Medicine, Beth Israel Deaconess Medical Center, 02215, Boston, MA, USA; Department of Biostatistics, Harvard T.H. Chan School of Public Health, 02115, Boston, MA, USA
| | - Zhiyong Lu
- National Center for Biotechnology Information, National Library of Medicine, 20894, Bethesda, MD, USA
| | - Ronald M Summers
- Clinical Center, National Institutes of Health, 20892, Bethesda, MD, USA
| | - George Shih
- Department of Radiology, Weill Cornell Medicine, 10065, New York, NY, USA
| | - Zhangyang Wang
- Department of Electrical and Computer Engineering, The University of Texas at Austin, 78712, Austin, TX, USA.
| | - Yifan Peng
- Department of Population Health Sciences, Weill Cornell Medicine, 10065, New York, NY, USA.
| |
Collapse
|
7
|
Li D, Huo H, Jiao S, Sun X, Chen S. Automated thorax disease diagnosis using multi-branch residual attention network. Sci Rep 2024; 14:11865. [PMID: 38789592 PMCID: PMC11126636 DOI: 10.1038/s41598-024-62813-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2024] [Accepted: 05/21/2024] [Indexed: 05/26/2024] Open
Abstract
Chest X-ray (CXR) is an extensively utilized radiological modality for supporting the diagnosis of chest diseases. However, existing research approaches suffer from limitations in effectively integrating multi-scale CXR image features and are also hindered by imbalanced datasets. Therefore, there is a pressing need for further advancement in computer-aided diagnosis (CAD) of thoracic diseases. To tackle these challenges, we propose a multi-branch residual attention network (MBRANet) for thoracic disease diagnosis. MBRANet comprises three components. Firstly, to address the issue of inadequate extraction of spatial and positional information by the convolutional layer, a novel residual structure incorporating a coordinate attention (CA) module is proposed to extract features at multiple scales. Next, based on the concept of a Feature Pyramid Network (FPN), we perform multi-scale feature fusion in the following manner. Thirdly, we propose a novel Multi-Branch Feature Classifier (MFC) approach, which leverages the class-specific residual attention (CSRA) module for classification instead of relying solely on the fully connected layer. In addition, the designed BCEWithLabelSmoothing loss function improves the generalization ability and mitigates the problem of class imbalance by introducing a smoothing factor. We evaluated MBRANet on the ChestX-Ray14, CheXpert, MIMIC-CXR, and IU X-Ray datasets and achieved average AUCs of 0.841, 0.895, 0.805, and 0.745, respectively. Our method outperformed state-of-the-art baselines on these benchmark datasets.
Collapse
Affiliation(s)
- Dongfang Li
- School of Information Engineering, Henan University of Science and Technology, Luoyang, 471000, Henan, China
| | - Hua Huo
- School of Information Engineering, Henan University of Science and Technology, Luoyang, 471000, Henan, China.
| | - Shupei Jiao
- School of Information Engineering, Henan University of Science and Technology, Luoyang, 471000, Henan, China
| | - Xiaowei Sun
- School of Information Engineering, Henan University of Science and Technology, Luoyang, 471000, Henan, China
| | - Shuya Chen
- School of Information Engineering, Henan University of Science and Technology, Luoyang, 471000, Henan, China
| |
Collapse
|
8
|
Shi Q, Ye M, Huang W, Ruan W, Du B. Label-Aware Calibration and Relation-Preserving in Visual Intention Understanding. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:2627-2638. [PMID: 38536683 DOI: 10.1109/tip.2024.3380250] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Visual intention understanding is a challenging task that explores the hidden intention behind the images of publishers in social media. Visual intention represents implicit semantics, whose ambiguous definition inevitably leads to label shifting and label blemish. The former indicates that the same image delivers intention discrepancies under different data augmentations, while the latter represents that the label of intention data is susceptible to errors or omissions during the annotation process. This paper proposes a novel method, called Label-aware Calibration and Relation-preserving (LabCR) to alleviate the above two problems from both intra-sample and inter-sample views. First, we disentangle the multiple intentions into a single intention for explicit distribution calibration in terms of the overall and the individual. Calibrating the class probability distributions in augmented instance pairs provides consistent inferred intention to address label shifting. Second, we utilize the intention similarity to establish correlations among samples, which offers additional supervision signals to form correlation alignments in instance pairs. This strategy alleviates the effect of label blemish. Extensive experiments have validated the superiority of the proposed method LabCR in visual intention understanding and pedestrian attribute recognition. Code is available at https://github.com/ShiQingHongYa/LabCR.
Collapse
|
9
|
Hung-Nguyen M. Patch-Level Feature Selection for Thoracic Disease Classification by Chest X-ray Images Using Information Bottleneck. Bioengineering (Basel) 2024; 11:316. [PMID: 38671737 PMCID: PMC11047950 DOI: 10.3390/bioengineering11040316] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Revised: 03/20/2024] [Accepted: 03/23/2024] [Indexed: 04/28/2024] Open
Abstract
Chest X-ray (CXR) examination serves as a widely employed clinical test in medical diagnostics. Many studied have tried to apply artificial intelligence (AI) programs to analyze CXR images. Despite numerous positive outcomes, assessing the applicability of AI models for comprehensive diagnostic support remains a formidable challenge. We observed that, even when AI models exhibit high accuracy on one dataset, their performance may deteriorate when tested on another. To address this issue, we propose incorporating a variational information bottleneck (VIB) at the patch level to enhance the generalizability of diagnostic support models. The VIB introduces a probabilistic model aimed at approximating the posterior distribution of latent variables given input data, thereby enhancing the model's generalization capabilities on unseen data. Unlike the conventional VIB approaches that flatten features and use a re-parameterization trick to sample a new latent feature, our method applies the trick to 2D feature maps. This design allows only important pixels to respond, and the model will select important patches in an image. Moreover, the proposed patch-level VIB seamlessly integrates with various convolutional neural networks, offering a versatile solution to improve performance. Experimental results illustrate enhanced accuracy in standard experiment settings. In addition, the method shows robust improvement when training and testing on different datasets.
Collapse
Affiliation(s)
- Manh Hung-Nguyen
- Faculty of Electrical and Electronics Engineering, HCMC University of Technology and Education, Ho Chi Minh City 7000, Vietnam
| |
Collapse
|
10
|
Zhang K, Liang W, Cao P, Liu X, Yang J, Zaiane O. Label correlation guided discriminative label feature learning for multi-label chest image classification. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 245:108032. [PMID: 38244339 DOI: 10.1016/j.cmpb.2024.108032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Revised: 01/02/2024] [Accepted: 01/12/2024] [Indexed: 01/22/2024]
Abstract
BACKGROUND AND OBJECTIVE Multi-label Chest X-ray (CXR) images often contain rich label relationship information, which is beneficial to improve classification performance. However, because of the intricate relationships among labels, most existing works fail to effectively learn and make full use of the label correlations, resulting in limited classification performance. In this study, we propose a multi-label learning framework that learns and leverages the label correlations to improve multi-label CXR image classification. METHODS In this paper, we capture the global label correlations through the self-attention mechanism. Meanwhile, to better utilize label correlations for guiding feature learning, we decompose the image-level features into label-level features. Furthermore, we enhance label-level feature learning in an end-to-end manner by a consistency constraint between global and local label correlations, and a label correlation guided multi-label supervised contrastive loss. RESULTS To demonstrate the superior performance of our proposed approach, we conduct three times 5-fold cross-validation experiments on the CheXpert dataset. Our approach obtains an average F1 score of 44.6% and an AUC of 76.5%, achieving a 7.7% and 1.3% improvement compared to the state-of-the-art results. CONCLUSION More accurate label correlations and full utilization of the learned label correlations help learn more discriminative label-level features. Experimental results demonstrate that our approach achieves exceptionally competitive performance compared to the state-of-the-art algorithms.
Collapse
Affiliation(s)
- Kai Zhang
- Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Wei Liang
- Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Peng Cao
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China; National Frontiers Science Center for Industrial Intelligence and Systems Optimization, Shenyang, China.
| | - Xiaoli Liu
- DAMO Academy, Alibaba Group, Hangzhou, China
| | - Jinzhu Yang
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China; National Frontiers Science Center for Industrial Intelligence and Systems Optimization, Shenyang, China
| | - Osmar Zaiane
- Alberta Machine Intelligence Institute, University of Alberta, Edmonton, Alberta, Canada
| |
Collapse
|
11
|
Gao X, Jiang B, Wang X, Huang L, Tu Z. Chest x-ray diagnosis via spatial-channel high-order attention representation learning. Phys Med Biol 2024; 69:045026. [PMID: 38347732 DOI: 10.1088/1361-6560/ad2014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Accepted: 01/18/2024] [Indexed: 02/15/2024]
Abstract
Objective. Chest x-ray image representation and learning is an important problem in computer-aided diagnostic area. Existing methods usually adopt CNN or Transformers for feature representation learning and focus on learning effective representations for chest x-ray images. Although good performance can be obtained, however, these works are still limited mainly due to the ignorance of mining the correlations of channels and pay little attention on the local context-aware feature representation of chest x-ray image.Approach. To address these problems, in this paper, we propose a novel spatial-channel high-order attention model (SCHA) for chest x-ray image representation and diagnosis. The proposed network architecture mainly contains three modules, i.e. CEBN, SHAM and CHAM. To be specific, firstly, we introduce a context-enhanced backbone network by employing multi-head self-attention to extract initial features for the input chest x-ray images. Then, we develop a novel SCHA which contains both spatial and channel high-order attention learning branches. For the spatial branch, we develop a novel local biased self-attention mechanism which can capture both local and long-range global dependences of positions to learn rich context-aware representation. For the channel branch, we employ Brownian Distance Covariance to encode the correlation information of channels and regard it as the image representation. Finally, the two learning branches are integrated together for the final multi-label diagnosis classification and prediction.Main results. Experiments on the commonly used datasets including ChestX-ray14 and CheXpert demonstrate that our proposed SCHA approach can obtain better performance when comparing many related approaches.Significance. This study obtains a more discriminative method for chest x-ray classification and provides a technique for computer-aided diagnosis.
Collapse
Affiliation(s)
- Xinyue Gao
- The School of Computer Science and Technology, Anhui University, Hefei 230601, People's Republic of China
| | - Bo Jiang
- The School of Computer Science and Technology, Anhui University, Hefei 230601, People's Republic of China
| | - Xixi Wang
- The School of Computer Science and Technology, Anhui University, Hefei 230601, People's Republic of China
| | - Lili Huang
- The School of Computer Science and Technology, Anhui University, Hefei 230601, People's Republic of China
| | - Zhengzheng Tu
- The School of Computer Science and Technology, Anhui University, Hefei 230601, People's Republic of China
| |
Collapse
|
12
|
Hasanah U, Avian C, Darmawan JT, Bachroin N, Faisal M, Prakosa SW, Leu JS, Tsai CT. CheXNet and feature pyramid network: a fusion deep learning architecture for multilabel chest X-Ray clinical diagnoses classification. THE INTERNATIONAL JOURNAL OF CARDIOVASCULAR IMAGING 2023:10.1007/s10554-023-03039-x. [PMID: 38150139 DOI: 10.1007/s10554-023-03039-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Accepted: 12/18/2023] [Indexed: 12/28/2023]
Abstract
The existing multilabel X-Ray image learning tasks generally contain much information on pathology co-occurrence and interdependency, which is very important for clinical diagnosis. However, the challenging part of this subject is to accurately diagnose multiple diseases that occurred in a single X-Ray image since multiple levels of features are generated in the images, and create different features as in single label detection. Various works were developed to address this challenge with proposed deep learning architectures to improve classification performance and enrich diagnosis results with multi-probability disease detection. The objective is to create an accurate result and a faster inference system to support a quick diagnosis in the medical system. To contribute to this state-of-the-art, we designed a fusion architecture, CheXNet and Feature Pyramid Network (FPN), to classify and discriminate multiple thoracic diseases from chest X-Rays. This concept enables the model to extract while creating a pyramid of feature maps with different spatial resolutions that capture low-level and high-level semantic information to encounter multiple features. The model's effectiveness is evaluated using the NIH ChestXray14 dataset, with the Area Under Curve (AUC) and accuracy metrics used to compare the results against other cutting-edge approaches. The overall results demonstrate that our method outperforms other approaches and has become promising for multilabel disease classification in chest X-Rays, with potential applications in clinical practice. The result demonstrated that we achieved an average AUC of 0.846 and an accuracy of 0.914. Further, our proposed architecture diagnoses images in 0.013 s, faster than the latest approaches.
Collapse
Affiliation(s)
- Uswatun Hasanah
- Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan
| | - Cries Avian
- Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan
| | | | - Nabil Bachroin
- Department of Electrical Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan
| | - Muhamad Faisal
- Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan
| | - Setya Widyawan Prakosa
- Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan
| | - Jenq-Shiou Leu
- Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei, Taiwan.
| | - Chia-Ti Tsai
- Department of Internal Medicine, National Taiwan University Hospital, Taipei, Taiwan
| |
Collapse
|
13
|
Wang G, Wang P, Cong J, Wei B. MRChexNet: Multi-modal bridge and relational learning for thoracic disease recognition in chest X-rays. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:21292-21314. [PMID: 38124598 DOI: 10.3934/mbe.2023942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/23/2023]
Abstract
While diagnosing multiple lesion regions in chest X-ray (CXR) images, radiologists usually apply pathological relationships in medicine before making decisions. Therefore, a comprehensive analysis of labeling relationships in different data modes is essential to improve the recognition performance of the model. However, most automated CXR diagnostic methods that consider pathological relationships treat different data modalities as independent learning objects, ignoring the alignment of pathological relationships among different data modalities. In addition, some methods that use undirected graphs to model pathological relationships ignore the directed information, making it difficult to model all pathological relationships accurately. In this paper, we propose a novel multi-label CXR classification model called MRChexNet that consists of three modules: a representation learning module (RLM), a multi-modal bridge module (MBM) and a pathology graph learning module (PGL). RLM captures specific pathological features at the image level. MBM performs cross-modal alignment of pathology relationships in different data modalities. PGL models directed relationships between disease occurrences as directed graphs. Finally, the designed graph learning block in PGL performs the integrated learning of pathology relationships in different data modalities. We evaluated MRChexNet on two large-scale CXR datasets (ChestX-Ray14 and CheXpert) and achieved state-of-the-art performance. The mean area under the curve (AUC) scores for the 14 pathologies were 0.8503 (ChestX-Ray14) and 0.8649 (CheXpert). MRChexNet effectively aligns pathology relationships in different modalities and learns more detailed correlations between pathologies. It demonstrates high accuracy and generalization compared to competing approaches. MRChexNet can contribute to thoracic disease recognition in CXR.
Collapse
Affiliation(s)
- Guoli Wang
- Center for Medical Artificial Intelligence, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
- Qingdao Academy of Chinese Medical Sciences, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
| | - Pingping Wang
- Center for Medical Artificial Intelligence, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
- Qingdao Academy of Chinese Medical Sciences, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
| | - Jinyu Cong
- Center for Medical Artificial Intelligence, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
- Qingdao Academy of Chinese Medical Sciences, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
| | - Benzheng Wei
- Center for Medical Artificial Intelligence, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
- Qingdao Academy of Chinese Medical Sciences, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
| |
Collapse
|
14
|
Ahmad Z, Malik AK, Qamar N, Islam SU. Efficient Thorax Disease Classification and Localization Using DCNN and Chest X-ray Images. Diagnostics (Basel) 2023; 13:3462. [PMID: 37998598 PMCID: PMC10669971 DOI: 10.3390/diagnostics13223462] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2023] [Revised: 11/05/2023] [Accepted: 11/09/2023] [Indexed: 11/25/2023] Open
Abstract
Thorax disease is a life-threatening disease caused by bacterial infections that occur in the lungs. It could be deadly if not treated at the right time, so early diagnosis of thoracic diseases is vital. The suggested study can assist radiologists in more swiftly diagnosing thorax disorders and in the rapid airport screening of patients with a thorax disease, such as pneumonia. This paper focuses on automatically detecting and localizing thorax disease using chest X-ray images. It provides accurate detection and localization using DenseNet-121 which is foundation of our proposed framework, called Z-Net. The proposed framework utilizes the weighted cross-entropy loss function (W-CEL) that manages class imbalance issue in the ChestX-ray14 dataset, which helped in achieving the highest performance as compared to the previous models. The 112,120 images contained in the ChestX-ray14 dataset (60,412 images are normal, and the rest contain thorax diseases) were preprocessed and then trained for classification and localization. This work uses computer-aided diagnosis (CAD) system that supports development of highly accurate and precise computer-aided systems. We aim to develop a CAD system using a deep learning approach. Our quantitative results show high AUC scores in comparison with the latest research works. The proposed approach achieved the highest mean AUC score of 85.8%. This is the highest accuracy documented in the literature for any related model.
Collapse
Affiliation(s)
- Zeeshan Ahmad
- Department of Computer Science, COMSATS University Islamabad, Islamabad 45550, Pakistan
| | - Ahmad Kamran Malik
- Department of Computer Science, COMSATS University Islamabad, Islamabad 45550, Pakistan
| | - Nafees Qamar
- School of Health and Behavioral Sciences, Bryant University, Smithfield, RI 02917, USA
| | - Saif Ul Islam
- Department of Computer Science, Institute of Space Technology, Islamabad 44000, Pakistan
| |
Collapse
|
15
|
Kim K, Lee JH, Je Oh S, Chung MJ. AI-based computer-aided diagnostic system of chest digital tomography synthesis: Demonstrating comparative advantage with X-ray-based AI systems. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 240:107643. [PMID: 37348439 DOI: 10.1016/j.cmpb.2023.107643] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Revised: 05/26/2023] [Accepted: 06/03/2023] [Indexed: 06/24/2023]
Abstract
BACKGROUND Compared with chest X-ray (CXR) imaging, which is a single image projected from the front of the patient, chest digital tomosynthesis (CDTS) imaging can be more advantageous for lung lesion detection because it acquires multiple images projected from multiple angles of the patient. Various clinical comparative analysis and verification studies have been reported to demonstrate this, but there is no artificial intelligence (AI)-based comparative analysis studies. Existing AI-based computer-aided detection (CAD) systems for lung lesion diagnosis have been developed mainly based on CXR images; however, CAD-based on CDTS, which uses multi-angle images of patients in various directions, has not been proposed and verified for its usefulness compared to CXR-based counterparts. BACKGROUND AND OBJECTIVE This study develops and tests a CDTS-based AI CAD system to detect lung lesions to demonstrate performance improvements compared to CXR-based AI CAD. METHODS We used multiple (e.g., five) projection images as input for the CDTS-based AI model and a single-projection image as input for the CXR-based AI model to compare and evaluate the performance between models. Multiple/single projection input images were obtained by virtual projection on the three-dimensional (3D) stack of computed tomography (CT) slices of each patient's lungs from which the bed area was removed. These multiple images result from shooting from the front and left and right 30/60∘. The projected image captured from the front was used as the input for the CXR-based AI model. The CDTS-based AI model used all five projected images. The proposed CDTS-based AI model consisted of five AI models that received images in each of the five directions, and obtained the final prediction result through an ensemble of five models. Each model used WideResNet-50. To train and evaluate CXR- and CDTS-based AI models, 500 healthy data, 206 tuberculosis data, and 242 pneumonia data were used, and three three-fold cross-validation was applied. RESULTS The proposed CDTS-based AI CAD system yielded sensitivities of 0.782 and 0.785 and accuracies of 0.895 and 0.837 for the (binary classification) performance of detecting tuberculosis and pneumonia, respectively, against normal subjects. These results show higher performance than the sensitivity of 0.728 and 0.698 and accuracies of 0.874 and 0.826 for detecting tuberculosis and pneumonia through the CXR-based AI CAD, which only uses a single projection image in the frontal direction. We found that CDTS-based AI CAD improved the sensitivity of tuberculosis and pneumonia by 5.4% and 8.7% respectively, compared to CXR-based AI CAD without loss of accuracy. CONCLUSIONS This study comparatively proves that CDTS-based AI CAD technology can improve performance more than CXR. These results suggest that we can enhance the clinical application of CDTS. Our code is available at https://github.com/kskim-phd/CDTS-CAD-P.
Collapse
Affiliation(s)
- Kyungsu Kim
- Medical AI Research Center, Research Institute for Future Medicine, Samsung Medical Center, Seoul 06351, Republic of Korea; Department of Data Convergence and Future Medicine, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea.
| | - Ju Hwan Lee
- Department of Health Sciences and Technology, SAIHST, Sungkyunkwan University, Seoul 06351, Republic of Korea
| | - Seong Je Oh
- Department of Health Sciences and Technology, SAIHST, Sungkyunkwan University, Seoul 06351, Republic of Korea
| | - Myung Jin Chung
- Medical AI Research Center, Research Institute for Future Medicine, Samsung Medical Center, Seoul 06351, Republic of Korea; Department of Data Convergence and Future Medicine, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea.
| |
Collapse
|
16
|
Liu Z, Cheng Y, Tamura S. Multi-Label Local to Global Learning: A Novel Learning Paradigm for Chest X-Ray Abnormality Classification. IEEE J Biomed Health Inform 2023; 27:4409-4420. [PMID: 37252867 DOI: 10.1109/jbhi.2023.3281466] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Deep neural network (DNN) approaches have shown remarkable progress in automatic Chest X-rays classification. However, existing methods use a training scheme that simultaneously trains all abnormalities without considering their learning priority. Inspired by the clinical practice of radiologists progressively recognizing more abnormalities and the observation that existing curriculum learning (CL) methods based on image difficulty may not be suitable for disease diagnosis, we propose a novel CL paradigm, named multi-label local to global (ML-LGL). This approach iteratively trains DNN models on gradually increasing abnormalities within the dataset, i,e, from fewer abnormalities (local) to more ones (global). At each iteration, we first build the local category by adding high-priority abnormalities for training, and the abnormality's priority is determined by our three proposed clinical knowledge-leveraged selection functions. Then, images containing abnormalities in the local category are gathered to form a new training set. The model is lastly trained on this set using a dynamic loss. Additionally, we demonstrate the superiority of ML-LGL from the perspective of the model's initial stability during training. Experimental results on three open-source datasets, PLCO, ChestX-ray14 and CheXpert show that our proposed learning paradigm outperforms baselines and achieves comparable results to state-of-the-art methods. The improved performance promises potential applications in multi-label Chest X-ray classification.
Collapse
|
17
|
Sun Z, Qu L, Luo J, Song Z, Wang M. Label correlation transformer for automated chest X-ray diagnosis with reliable interpretability. LA RADIOLOGIA MEDICA 2023:10.1007/s11547-023-01647-0. [PMID: 37233906 DOI: 10.1007/s11547-023-01647-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Received: 01/11/2023] [Accepted: 05/10/2023] [Indexed: 05/27/2023]
Abstract
Computer-aided diagnosis of chest X-ray (CXR) images can help reduce the huge workload of radiologists and avoid the inter-observer variability in large-scale early disease screening. Recently, most state-of-the-art studies employ deep learning techniques to address this problem through multi-label classification. However, existing methods still suffer from low classification accuracy and poor interpretability for each diagnostic task. This study aims to propose a novel transformer-based deep learning model for automated CXR diagnosis with high performance and reliable interpretability. We introduce a novel transformer architecture into this problem and utilize the unique query structure of transformer to capture the global and local information of the images and the correlation between labels. In addition, we propose a new loss function to help the model find correlations between the labels in CXR images. To achieve accurate and reliable interpretability, we generate heatmaps using the proposed transformer model and compare with the true pathogenic regions labeled by the physicians. The proposed model achieves a mean AUC of 0.831 on chest X-ray 14 and 0.875 on PadChest dataset, which outperforms existing state-of-the-art methods. The attention heatmaps show that our model could focus on the exact corresponding areas of related truly labeled pathogenic regions. The proposed model effectively improves the performance of CXR multi-label classification and the interpretability of label correlations, thus providing new evidence and methods for automated clinical diagnosis.
Collapse
Affiliation(s)
- Zexuan Sun
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai, 200032, China
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Linhao Qu
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai, 200032, China
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Jiazheng Luo
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai, 200032, China
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Zhijian Song
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai, 200032, China.
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China.
| | - Manning Wang
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai, 200032, China.
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China.
| |
Collapse
|
18
|
Farhan AMQ, Yang S. Automatic lung disease classification from the chest X-ray images using hybrid deep learning algorithm. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1-27. [PMID: 37362647 PMCID: PMC10030349 DOI: 10.1007/s11042-023-15047-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/04/2022] [Revised: 08/30/2022] [Accepted: 02/27/2023] [Indexed: 06/28/2023]
Abstract
The chest X-ray images provide vital information about the congestion cost-effectively. We propose a novel Hybrid Deep Learning Algorithm (HDLA) framework for automatic lung disease classification from chest X-ray images. The model consists of steps including pre-processing of chest X-ray images, automatic feature extraction, and detection. In a pre-processing step, our goal is to improve the quality of raw chest X-ray images using the combination of optimal filtering without data loss. The robust Convolutional Neural Network (CNN) is proposed using the pre-trained model for automatic lung feature extraction. We employed the 2D CNN model for the optimum feature extraction in minimum time and space requirements. The proposed 2D CNN model ensures robust feature learning with highly efficient 1D feature estimation from the input pre-processed image. As the extracted 1D features have suffered from significant scale variations, we optimized them using min-max scaling. We classify the CNN features using the different machine learning classifiers such as AdaBoost, Support Vector Machine (SVM), Random Forest (RM), Backpropagation Neural Network (BNN), and Deep Neural Network (DNN). The experimental results claim that the proposed model improves the overall accuracy by 3.1% and reduces the computational complexity by 16.91% compared to state-of-the-art methods.
Collapse
Affiliation(s)
- Abobaker Mohammed Qasem Farhan
- School of information and Software Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Shangming Yang
- School of information and Software Engineering, University of Electronic Science and Technology of China, Chengdu, China
| |
Collapse
|
19
|
Zhang Y, Luo L, Dou Q, Heng PA. Triplet attention and dual-pool contrastive learning for clinic-driven multi-label medical image classification. Med Image Anal 2023; 86:102772. [PMID: 36822050 DOI: 10.1016/j.media.2023.102772] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2022] [Revised: 11/21/2022] [Accepted: 02/10/2023] [Indexed: 02/18/2023]
Abstract
Multi-label classification (MLC) can attach multiple labels on single image, and has achieved promising results on medical images. But existing MLC methods still face challenging clinical realities in practical use, such as: (1) medical risks arising from misclassification, (2) sample imbalance problem among different diseases, (3) inability to classify the diseases that are not pre-defined (unseen diseases). Here, we design a hybrid label to improve the flexibility of MLC methods and alleviate the sample imbalance problem. Specifically, in the labeled training set, we remain independent labels for high-frequency diseases with enough samples and use a hybrid label to merge low-frequency diseases with fewer samples. The hybrid label can also be used to put unseen diseases in practical use. In this paper, we propose Triplet Attention and Dual-pool Contrastive Learning (TA-DCL) for multi-label medical image classification based on the aforementioned label representation. TA-DCL architecture is a triplet attention network (TAN), which combines category-attention, self-attention and cross-attention together to learn high-quality label embeddings for all disease labels by mining effective information from medical images. DCL includes dual-pool contrastive training (DCT) and dual-pool contrastive inference (DCI). DCT optimizes the clustering centers of label embeddings belonging to different disease labels to improve the discrimination of label embeddings. DCI relieves the error classification of sick cases for reducing the clinical risk and improving the ability to detect unseen diseases by contrast of differences. TA-DCL is validated on two public medical image datasets, ODIR and NIH-ChestXray14, showing superior performance than other state-of-the-art MLC methods. Code is available at https://github.com/ZhangYH0502/TA-DCL.
Collapse
Affiliation(s)
- Yuhan Zhang
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China; Institute of Medical Intelligence and XR, The Chinese University of Hong Kong, Hong Kong, China; Shenzhen Research Institute, The Chinese University of Hong Kong, Hong Kong, China.
| | - Luyang Luo
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China.
| | - Qi Dou
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China; Institute of Medical Intelligence and XR, The Chinese University of Hong Kong, Hong Kong, China.
| | - Pheng-Ann Heng
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Hong Kong, China; Institute of Medical Intelligence and XR, The Chinese University of Hong Kong, Hong Kong, China.
| |
Collapse
|
20
|
Meng Y, Bridge J, Addison C, Wang M, Merritt C, Franks S, Mackey M, Messenger S, Sun R, Fitzmaurice T, McCann C, Li Q, Zhao Y, Zheng Y. Bilateral adaptive graph convolutional network on CT based Covid-19 diagnosis with uncertainty-aware consensus-assisted multiple instance learning. Med Image Anal 2023; 84:102722. [PMID: 36574737 PMCID: PMC9753459 DOI: 10.1016/j.media.2022.102722] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2022] [Revised: 10/17/2022] [Accepted: 12/02/2022] [Indexed: 12/23/2022]
Abstract
Coronavirus disease (COVID-19) has caused a worldwide pandemic, putting millions of people's health and lives in jeopardy. Detecting infected patients early on chest computed tomography (CT) is critical in combating COVID-19. Harnessing uncertainty-aware consensus-assisted multiple instance learning (UC-MIL), we propose to diagnose COVID-19 using a new bilateral adaptive graph-based (BA-GCN) model that can use both 2D and 3D discriminative information in 3D CT volumes with arbitrary number of slices. Given the importance of lung segmentation for this task, we have created the largest manual annotation dataset so far with 7,768 slices from COVID-19 patients, and have used it to train a 2D segmentation model to segment the lungs from individual slices and mask the lungs as the regions of interest for the subsequent analyses. We then used the UC-MIL model to estimate the uncertainty of each prediction and the consensus between multiple predictions on each CT slice to automatically select a fixed number of CT slices with reliable predictions for the subsequent model reasoning. Finally, we adaptively constructed a BA-GCN with vertices from different granularity levels (2D and 3D) to aggregate multi-level features for the final diagnosis with the benefits of the graph convolution network's superiority to tackle cross-granularity relationships. Experimental results on three largest COVID-19 CT datasets demonstrated that our model can produce reliable and accurate COVID-19 predictions using CT volumes with any number of slices, which outperforms existing approaches in terms of learning and generalisation ability. To promote reproducible research, we have made the datasets, including the manual annotations and cleaned CT dataset, as well as the implementation code, available at https://doi.org/10.5281/zenodo.6361963.
Collapse
Affiliation(s)
- Yanda Meng
- Department of Eye and Vision Science, University of Liverpool, Liverpool, United Kingdom
| | - Joshua Bridge
- Department of Eye and Vision Science, University of Liverpool, Liverpool, United Kingdom
| | - Cliff Addison
- Advanced Research Computing, University of Liverpool, Liverpool, United Kingdom
| | - Manhui Wang
- Advanced Research Computing, University of Liverpool, Liverpool, United Kingdom
| | | | - Stu Franks
- Alces Flight Limited, Bicester, United Kingdom
| | - Maria Mackey
- Amazon Web Services, 60 Holborn Viaduct, London, United Kingdom
| | - Steve Messenger
- Amazon Web Services, 60 Holborn Viaduct, London, United Kingdom
| | - Renrong Sun
- Department of Radiology, Hubei Provincial Hospital of Integrated Chinese and Western Medicine, Hubei University of Chinese Medicine, Wuhan, China
| | - Thomas Fitzmaurice
- Adult Cystic Fibrosis Unit, Liverpool Heart and Chest Hospital NHS Foundation Trust, Liverpool, United Kingdom
| | - Caroline McCann
- Radiology, Liverpool Heart and Chest Hospital NHS Foundation Trust, United Kingdom
| | - Qiang Li
- The Affiliated People’s Hospital of Ningbo University, Ningbo, China
| | - Yitian Zhao
- The Affiliated People's Hospital of Ningbo University, Ningbo, China; Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Science, Ningbo, China.
| | - Yalin Zheng
- Department of Eye and Vision Science, University of Liverpool, Liverpool, United Kingdom; Liverpool Centre for Cardiovascular Science, University of Liverpool and Liverpool Heart & Chest Hospital, Liverpool, United Kingdom.
| |
Collapse
|
21
|
Ibragimov B, Arzamasov K, Maksudov B, Kiselev S, Mongolin A, Mustafaev T, Ibragimova D, Evteeva K, Andreychenko A, Morozov S. A 178-clinical-center experiment of integrating AI solutions for lung pathology diagnosis. Sci Rep 2023; 13:1135. [PMID: 36670118 PMCID: PMC9859802 DOI: 10.1038/s41598-023-27397-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2021] [Accepted: 01/02/2023] [Indexed: 01/22/2023] Open
Abstract
In 2020, an experiment testing AI solutions for lung X-ray analysis on a multi-hospital network was conducted. The multi-hospital network linked 178 Moscow state healthcare centers, where all chest X-rays from the network were redirected to a research facility, analyzed with AI, and returned to the centers. The experiment was formulated as a public competition with monetary awards for participating industrial and research teams. The task was to perform the binary detection of abnormalities from chest X-rays. For the objective real-life evaluation, no training X-rays were provided to the participants. This paper presents one of the top-performing AI frameworks from this experiment. First, the framework used two EfficientNets, histograms of gradients, Haar feature ensembles, and local binary patterns to recognize whether an input image represents an acceptable lung X-ray sample, meaning the X-ray is not grayscale inverted, is a frontal chest X-ray, and completely captures both lung fields. Second, the framework extracted the region with lung fields and then passed them to a multi-head DenseNet, where the heads recognized the patient's gender, age and the potential presence of abnormalities, and generated the heatmap with the abnormality regions highlighted. During one month of the experiment from 11.23.2020 to 12.25.2020, 17,888 cases have been analyzed by the framework with 11,902 cases having radiological reports with the reference diagnoses that were unequivocally parsed by the experiment organizers. The performance measured in terms of the area under receiving operator curve (AUC) was 0.77. The AUC for individual diseases ranged from 0.55 for herniation to 0.90 for pneumothorax.
Collapse
Affiliation(s)
- Bulat Ibragimov
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.
| | - Kirill Arzamasov
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
| | - Bulat Maksudov
- School of Electronic Engineering, Dublin City University, Dublin, Ireland
| | | | - Alexander Mongolin
- Innopolis University, Innopolis, Russia
- Nova Information Management School, Universidade Nova de Lisboa, Lisbon, Portugal
| | - Tamerlan Mustafaev
- Innopolis University, Innopolis, Russia
- University Clinic Kazan State University, Kazan, Russia
| | | | - Ksenia Evteeva
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
| | - Anna Andreychenko
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
| | - Sergey Morozov
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
- Osimis SA, Liege, Belgium
| |
Collapse
|
22
|
Han Q, Hou M, Wang H, Wu C, Tian S, Qiu Z, Zhou B. EHDFL: Evolutionary hybrid domain feature learning based on windowed fast Fourier convolution pyramid for medical image classification. Comput Biol Med 2023; 152:106353. [PMID: 36473339 DOI: 10.1016/j.compbiomed.2022.106353] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2022] [Revised: 11/15/2022] [Accepted: 11/23/2022] [Indexed: 11/26/2022]
Abstract
With the development of modern medical technology, medical image classification has played an important role in medical diagnosis and clinical practice. Medical image classification algorithms based on deep learning emerge in endlessly, and have achieved amazing results. However, most of these methods ignore the feature representation based on frequency domain, and only focus on spatial features. To solve this problem, we propose a hybrid domain feature learning (HDFL) module based on windowed fast Fourier convolution pyramid, which combines the global features with a wide range of receptive fields in frequency domain and the local features with multiple scales in spatial domain. In order to prevent frequency leakage, we construct a Windowed Fast Fourier Convolution (WFFC) structure based on Fast Fourier Convolution (FFC). In order to learn hybrid domain features, we combine ResNet, FPN, and attention mechanism to construct a hybrid domain feature learning module. In addition, a super-parametric optimization algorithm is constructed based on genetic algorithm for our classification model, so as to realize the automation of our super-parametric optimization. We evaluated the newly published medical image classification dataset MedMNIST, and the experimental results show that our method can effectively learning the hybrid domain feature information of frequency domain and spatial domain.
Collapse
Affiliation(s)
- Qi Han
- School of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing 401331, China
| | - Mingyang Hou
- School of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing 401331, China.
| | - Hongyi Wang
- School of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing 401331, China
| | - Chen Wu
- School of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing 401331, China
| | - Sheng Tian
- School of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing 401331, China
| | - Zicheng Qiu
- School of Intelligent Technology and Engineering, Chongqing University of Science and Technology, Chongqing 401331, China
| | - Baoping Zhou
- College of Information Engineering, Tarim University, Alar City, China
| |
Collapse
|
23
|
Zhu X, Pang S, Zhang X, Huang J, Zhao L, Tang K, Feng Q. PCAN: Pixel-wise classification and attention network for thoracic disease classification and weakly supervised localization. Comput Med Imaging Graph 2022; 102:102137. [PMID: 36308870 DOI: 10.1016/j.compmedimag.2022.102137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Revised: 09/23/2022] [Accepted: 10/05/2022] [Indexed: 11/06/2022]
Abstract
Automatic chest X-ray (CXR) disease classification has drawn increasing public attention as CXR is widely used in thoracic disease diagnosis. Existing classification networks typically employ a global average pooling layer to produce the final feature for the subsequent classifier. This limits the classification performance owing to the characteristics of lesions in CXR images, including small relative sizes, varied absolute sizes, and different occurrence locations. In this study, we propose a pixel-wise classification and attention network (PCAN) to simultaneously perform disease classification and weakly supervised localization, which provides interpretability for disease classification. The PCAN comprises a backbone network for extracting mid-level features, a pixel-wise classification branch (pc-branch) for generating pixel-wise diagnoses, and a pixel-wise attention branch (pa-branch) for producing pixel-wise weights. The pc-branch is capable of explicitly detecting small lesions, and the pa-branch is capable of adaptively focusing on different regions when classifying different thoracic diseases. Then, the pixel-wise diagnoses are multiplied with the pixel-wise weights to obtain the disease localization map, which provides the sizes and locations of lesions in a manner of weakly supervised learning. The final image-wise diagnosis is obtained by summing up the disease localization map at the spatial dimension. Comprehensive experiments conducted on the ChestX-ray14 and CheXpert datasets demonstrate the effectiveness of the proposed PCAN, which has great potential for thoracic disease diagnosis and treatment. The source codes are available at https://github.com/fzfs/PCAN.
Collapse
Affiliation(s)
- Xiongfeng Zhu
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| | - Shumao Pang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Xiaoxuan Zhang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Junzhang Huang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Lei Zhao
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Kai Tang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| |
Collapse
|
24
|
Han M, Wu H, Chen Z, Li M, Zhang X. A survey of multi-label classification based on supervised and semi-supervised learning. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-022-01658-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
25
|
Chen Y, Zhao X, Tang B. Boosting lesion annotation via aggregating explicit relations in external medical knowledge graph. Artif Intell Med 2022; 132:102376. [DOI: 10.1016/j.artmed.2022.102376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2021] [Revised: 05/09/2022] [Accepted: 08/17/2022] [Indexed: 11/29/2022]
|
26
|
Modeling global and local label correlation with graph convolutional networks for multi-label chest X-ray image classification. Med Biol Eng Comput 2022; 60:2567-2588. [DOI: 10.1007/s11517-022-02604-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Accepted: 06/07/2022] [Indexed: 10/17/2022]
|
27
|
Li Y, Zhang Y, Cui W, Lei B, Kuang X, Zhang T. Dual Encoder-Based Dynamic-Channel Graph Convolutional Network With Edge Enhancement for Retinal Vessel Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1975-1989. [PMID: 35167444 DOI: 10.1109/tmi.2022.3151666] [Citation(s) in RCA: 47] [Impact Index Per Article: 15.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Retinal vessel segmentation with deep learning technology is a crucial auxiliary method for clinicians to diagnose fundus diseases. However, the deep learning approaches inevitably lose the edge information, which contains spatial features of vessels while performing down-sampling, leading to the limited segmentation performance of fine blood vessels. Furthermore, the existing methods ignore the dynamic topological correlations among feature maps in the deep learning framework, resulting in the inefficient capture of the channel characterization. To address these limitations, we propose a novel dual encoder-based dynamic-channel graph convolutional network with edge enhancement (DE-DCGCN-EE) for retinal vessel segmentation. Specifically, we first design an edge detection-based dual encoder to preserve the edge of vessels in down-sampling. Secondly, we investigate a dynamic-channel graph convolutional network to map the image channels to the topological space and synthesize the features of each channel on the topological map, which solves the limitation of insufficient channel information utilization. Finally, we study an edge enhancement block, aiming to fuse the edge and spatial features in the dual encoder, which is beneficial to improve the accuracy of fine blood vessel segmentation. Competitive experimental results on five retinal image datasets validate the efficacy of the proposed DE-DCGCN-EE, which achieves more remarkable segmentation results against the other state-of-the-art methods, indicating its potential clinical application.
Collapse
|
28
|
Zhang S, Ren Y, Wang J, Song B, Li R, Xu Y. GSTCNet: Gated spatio-temporal correlation network for stroke mortality prediction. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:9966-9982. [PMID: 36031978 DOI: 10.3934/mbe.2022465] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Stroke continues to be the most common cause of death in China. It has great significance for mortality prediction for stroke patients, especially in terms of analyzing the complex interactions between non-negligible factors. In this paper, we present a gated spatio-temporal correlation network (GSTCNet) to predict the one-year post-stroke mortality. Based on the four categories of risk factors: vascular event, chronic disease, medical usage and surgery, we designed a gated correlation graph convolution kernel to capture spatial features and enhance the spatial correlation between feature categories. Bi-LSTM represents the temporal features of five timestamps. The novel gated correlation attention mechanism is then connected to the Bi-LSTM to realize the comprehensive mining of spatio-temporal correlations. Using the data on 2275 patients obtained from the neurology department of a local hospital, we constructed a series of sequential experiments. The experimental results show that the proposed model achieves competitive results on each evaluation metric, reaching an AUC of 89.17%, a precision of 97.75%, a recall of 95.33% and an F1-score of 95.19%. The interpretability analysis of the feature categories and timestamps also verified the potential application value of the model for stroke.
Collapse
Affiliation(s)
- Shuo Zhang
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450000, China
- Cooperative Innovation Center of Internet Healthcare, Zhengzhou University, Zhengzhou 450000, China
| | - Yonghao Ren
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450000, China
- Cooperative Innovation Center of Internet Healthcare, Zhengzhou University, Zhengzhou 450000, China
| | - Jing Wang
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450000, China
- Cooperative Innovation Center of Internet Healthcare, Zhengzhou University, Zhengzhou 450000, China
| | - Bo Song
- Department of Neurology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450000, China
- NHC Key Laboratory of Prevention and Treatment of Cerebrovascular Diseases, Zhengzhou 450000, China
| | - Runzhi Li
- Cooperative Innovation Center of Internet Healthcare, Zhengzhou University, Zhengzhou 450000, China
| | - Yuming Xu
- Department of Neurology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450000, China
- NHC Key Laboratory of Prevention and Treatment of Cerebrovascular Diseases, Zhengzhou 450000, China
| |
Collapse
|
29
|
Wang S, Lin M, Ghosal T, Ding Y, Peng Y. Knowledge Graph Applications in Medical Imaging Analysis: A Scoping Review. HEALTH DATA SCIENCE 2022; 2022:9841548. [PMID: 35800847 PMCID: PMC9259200 DOI: 10.34133/2022/9841548] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2021] [Accepted: 05/22/2022] [Indexed: 11/06/2022]
Abstract
Background There is an increasing trend to represent domain knowledge in structured graphs, which provide efficient knowledge representations for many downstream tasks. Knowledge graphs are widely used to model prior knowledge in the form of nodes and edges to represent semantically connected knowledge entities, which several works have adopted into different medical imaging applications. Methods We systematically searched over five databases to find relevant articles that applied knowledge graphs to medical imaging analysis. After screening, evaluating, and reviewing the selected articles, we performed a systematic analysis. Results We looked at four applications in medical imaging analysis, including disease classification, disease localization and segmentation, report generation, and image retrieval. We also identified limitations of current work, such as the limited amount of available annotated data and weak generalizability to other tasks. We further identified the potential future directions according to the identified limitations, including employing semisupervised frameworks to alleviate the need for annotated data and exploring task-agnostic models to provide better generalizability. Conclusions We hope that our article will provide the readers with aggregated documentation of the state-of-the-art knowledge graph applications for medical imaging to encourage future research.
Collapse
Affiliation(s)
- Song Wang
- The University of Texas at Austin, Austin, USA
| | - Mingquan Lin
- Population Health Sciences, Weill Cornell Medicine, New York, USA
| | - Tirthankar Ghosal
- Institute of Formal and Applied Linguistics, Charles University, Czechia, Czech Republic
| | - Ying Ding
- The University of Texas at Austin, Austin, USA
| | - Yifan Peng
- Population Health Sciences, Weill Cornell Medicine, New York, USA
| |
Collapse
|
30
|
Zhang J, Mao Y, Li J, Li Y, Luo J. A metric learning-based method using graph neural network for pancreatic cystic neoplasm classification from CTs. Med Phys 2022; 49:5523-5536. [PMID: 35536056 DOI: 10.1002/mp.15708] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2021] [Revised: 04/01/2022] [Accepted: 04/21/2022] [Indexed: 11/10/2022] Open
Abstract
PURPOSE Pancreatic cystic neoplasms (PCNs) are relatively rare neoplasms and difficult to be classified preoperatively. Ordinary deep learning methods have great potential to provide support for doctors in PCNs classification but require a quantity of labeled samples and exact segmentation of neoplasm. The proposed metric learning-based method using graph neural network aims to overcome the limitations brought by small and imbalanced dataset and get fast and accurate PCNs classification result from computed tomography (CT) images. METHODS The proposed framework applies graph neural network (GNN). GNNs perform well in fusing information and modeling relational data and get better results on dataset with small size. Based on metric learning strategy, model learns distance from the data. The similarity-based algorithm enhances the classification performance, and more characteristic information is found. We use a convolutional neural network (CNN) to extract features from given images. Then GNN is used to find the similarity between each two feature vectors and complete the classification. Several subtasks consisting of randomly selected images are established to improve generalization of the model. The experiments are carried out on the dataset provided by Huashan Hospital. The dataset is labeled by postoperative pathological analysis and contains ROI information calibrated by experts. We set two tasks based on the dataset: benign or malignant diagnosis of PCNs and classification of specific types. RESULTS Our model shows good performance on the 2 tasks with accuracies of 88.926% and 74.497%. The comparison of different methods' F1 scores in the benign or malignant diagnosis shows the proposed GNN-based method effectively reduces the negative impact brought by imbalanced dataset, which is also verified by the macro-average comparison in the 4-class classification task. CONCLUSIONS Compared with existing models, the proposed GNN-based model shows better performance in terms of imbalanced dataset with small size while reducing labeling cost. The result provides a possibility for its application into the computer aided diagnosis of PCNs. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Jiachen Zhang
- School of Information Science and Engineering, East China University of Science and Technology, Shanghai, China
| | - Yishen Mao
- Department of Pancreas Surgery, Huashan Hospital Fudan University, Shanghai, China
| | - Ji Li
- Department of Pancreas Surgery, Huashan Hospital Fudan University, Shanghai, China
| | - Yiru Li
- School of Information Science and Engineering, East China University of Science and Technology, Shanghai, China
| | - Jianxu Luo
- School of Information Science and Engineering, East China University of Science and Technology, Shanghai, China
| |
Collapse
|
31
|
Guo P, Li L, Li C, Huang W, Zhao G, Wang S, Wang M, Lin Y. Multiparametric Magnetic Resonance Imaging Information Fusion Using Graph Convolutional Network for Glioma Grading. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:7315665. [PMID: 35591941 PMCID: PMC9113909 DOI: 10.1155/2022/7315665] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/12/2021] [Revised: 03/31/2022] [Accepted: 04/01/2022] [Indexed: 11/17/2022]
Abstract
Accurate preoperative glioma grading is essential for clinical decision-making and prognostic evaluation. Multiparametric magnetic resonance imaging (mpMRI) serves as an important diagnostic tool for glioma patients due to its superior performance in describing noninvasively the contextual information in tumor tissues. Previous studies achieved promising glioma grading results with mpMRI data utilizing a convolutional neural network (CNN)-based method. However, these studies have not fully exploited and effectively fused the rich tumor contextual information provided in the magnetic resonance (MR) images acquired with different imaging parameters. In this paper, a novel graph convolutional network (GCN)-based mpMRI information fusion module (named MMIF-GCN) is proposed to comprehensively fuse the tumor grading relevant information in mpMRI. Specifically, a graph is constructed according to the characteristics of mpMRI data. The vertices are defined as the glioma grading features of different slices extracted by the CNN, and the edges reflect the distances between the slices in a 3D volume. The proposed method updates the information in each vertex considering the interaction between adjacent vertices. The final glioma grading is conducted by combining the fused information in all vertices. The proposed MMIF-GCN module can introduce an additional nonlinear representation learning step in the process of mpMRI information fusion while maintaining the positional relationship between adjacent slices. Experiments were conducted on two datasets, that is, a public dataset (named BraTS2020) and a private one (named GliomaHPPH2018). The results indicate that the proposed method can effectively fuse the grading information provided in mpMRI data for better glioma grading performance.
Collapse
Affiliation(s)
- Peiying Guo
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou 450052, China
| | - Longfei Li
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou 450052, China
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518000, China
| | - Cheng Li
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518000, China
| | - Weijian Huang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518000, China
| | - Guohua Zhao
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou 450052, China
| | - Shanshan Wang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518000, China
| | - Meiyun Wang
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou 450052, China
- Department of Radiology, Henan Provincial People's Hospital, Zhengzhou 450003, China
| | - Yusong Lin
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou 450052, China
- School of Cyber Science and Engineering, Zhengzhou University, Zhengzhou 450002, China
- Hanwei IoT Institute, Zhengzhou University, Zhengzhou 450002, China
| |
Collapse
|
32
|
Li L, Long Y, Huang B, Chen Z, Liu Z, Yang Z. Research on Chest Disease Recognition Based on Deep Hierarchical Learning Algorithm. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:6996444. [PMID: 35035852 PMCID: PMC8759895 DOI: 10.1155/2022/6996444] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/25/2021] [Revised: 10/24/2021] [Accepted: 11/22/2021] [Indexed: 11/29/2022]
Abstract
Chest X-ray has become one of the most common ways in diagnostic radiology exams, and this technology assists expert radiologists with finding the patients at potential risk of cardiopathy and lung diseases. However, it is still a challenge for expert radiologists to assess thousands of cases in a short period so that deep learning methods are introduced to tackle this problem. Since the diseases have correlations with each other and have hierarchical features, the traditional classification scheme could not achieve a good performance. In order to extract the correlation features among the diseases, some GCN-based models are introduced to combine the features extracted from the images to make prediction. This scheme can work well with the high quality of image features, so backbone with high computation cost plays a vital role in this scheme. However, a fast prediction in diagnostic radiology is also needed especially in case of emergency or region with low computation facilities, so we proposed an efficient convolutional neural network with GCN, which is named SGGCN, to meet the need of efficient computation and considerable accuracy. SGGCN used SGNet-101 as backbone, which is built by ShuffleGhost Block (Huang et al., 2021) to extract features with a low computation cost. In order to make sufficient usage of the information in GCN, a new GCN architecture is designed to combine information from different layers together in GCNM module so that we can utilize various hierarchical features and meanwhile make the GCN scheme faster. The experiment on CheXPert datasets illustrated that SGGCN achieves a considerable performance. Compared with GCN and ResNet-101 (He et al., 2015) backbone (test AUC 0.8080, parameters 4.7M and FLOPs 16.0B), the SGGCN achieves 0.7831 (-3.08%) test AUC with parameters 1.2M (-73.73%) and FLOPs 3.1B (-80.82%), where GCN with MobileNet (Sandler and Howard, 2018) backbone achieves 0.7531 (-6.79%) test AUC with parameters 0.5M (-88.46%) and FLOPs 0.66B (-95.88%).
Collapse
Affiliation(s)
- Lingling Li
- Department of Central Laboratory, Children's Hospital of Shanghai Jiao Tong University, Shanghai, China
| | - Yangyang Long
- School of Computing and Information System, University of Melbourne, Melbourne, Australia
| | - Bangtong Huang
- School of Management, Shanghai University of Engineering Science, Shanghai, China
| | - Zihong Chen
- College of Engineering, Shantou University, Shantou, China
| | - Zheng Liu
- School of Management, Shanghai University of Engineering Science, Shanghai, China
| | - Zekun Yang
- Department of Computer Science and Engineering, The Chinese University of Hong Kong, Shatin, Hong Kong, China
| |
Collapse
|
33
|
Ahmedt-Aristizabal D, Armin MA, Denman S, Fookes C, Petersson L. Graph-Based Deep Learning for Medical Diagnosis and Analysis: Past, Present and Future. SENSORS (BASEL, SWITZERLAND) 2021; 21:4758. [PMID: 34300498 PMCID: PMC8309939 DOI: 10.3390/s21144758] [Citation(s) in RCA: 58] [Impact Index Per Article: 14.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/07/2021] [Revised: 07/05/2021] [Accepted: 07/07/2021] [Indexed: 01/17/2023]
Abstract
With the advances of data-driven machine learning research, a wide variety of prediction problems have been tackled. It has become critical to explore how machine learning and specifically deep learning methods can be exploited to analyse healthcare data. A major limitation of existing methods has been the focus on grid-like data; however, the structure of physiological recordings are often irregular and unordered, which makes it difficult to conceptualise them as a matrix. As such, graph neural networks have attracted significant attention by exploiting implicit information that resides in a biological system, with interacting nodes connected by edges whose weights can be determined by either temporal associations or anatomical junctions. In this survey, we thoroughly review the different types of graph architectures and their applications in healthcare. We provide an overview of these methods in a systematic manner, organized by their domain of application including functional connectivity, anatomical structure, and electrical-based analysis. We also outline the limitations of existing techniques and discuss potential directions for future research.
Collapse
Affiliation(s)
- David Ahmedt-Aristizabal
- Imaging and Computer Vision Group, CSIRO Data61, Canberra 2601, Australia; (M.A.A.); (L.P.)
- Signal Processing, Artificial Intelligence and Vision Technologies (SAIVT) Research Program, Queensland University of Technology, Brisbane 4000, Australia; (S.D.); (C.F.)
| | - Mohammad Ali Armin
- Imaging and Computer Vision Group, CSIRO Data61, Canberra 2601, Australia; (M.A.A.); (L.P.)
| | - Simon Denman
- Signal Processing, Artificial Intelligence and Vision Technologies (SAIVT) Research Program, Queensland University of Technology, Brisbane 4000, Australia; (S.D.); (C.F.)
| | - Clinton Fookes
- Signal Processing, Artificial Intelligence and Vision Technologies (SAIVT) Research Program, Queensland University of Technology, Brisbane 4000, Australia; (S.D.); (C.F.)
| | - Lars Petersson
- Imaging and Computer Vision Group, CSIRO Data61, Canberra 2601, Australia; (M.A.A.); (L.P.)
| |
Collapse
|
34
|
Çallı E, Sogancioglu E, van Ginneken B, van Leeuwen KG, Murphy K. Deep learning for chest X-ray analysis: A survey. Med Image Anal 2021; 72:102125. [PMID: 34171622 DOI: 10.1016/j.media.2021.102125] [Citation(s) in RCA: 126] [Impact Index Per Article: 31.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2021] [Revised: 05/17/2021] [Accepted: 05/27/2021] [Indexed: 12/14/2022]
Abstract
Recent advances in deep learning have led to a promising performance in many medical image analysis tasks. As the most commonly performed radiological exam, chest radiographs are a particularly important modality for which a variety of applications have been researched. The release of multiple, large, publicly available chest X-ray datasets in recent years has encouraged research interest and boosted the number of publications. In this paper, we review all studies using deep learning on chest radiographs published before March 2021, categorizing works by task: image-level prediction (classification and regression), segmentation, localization, image generation and domain adaptation. Detailed descriptions of all publicly available datasets are included and commercial systems in the field are described. A comprehensive discussion of the current state of the art is provided, including caveats on the use of public datasets, the requirements of clinically useful systems and gaps in the current literature.
Collapse
Affiliation(s)
- Erdi Çallı
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands.
| | - Ecem Sogancioglu
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| | - Bram van Ginneken
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| | - Kicky G van Leeuwen
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| | - Keelin Murphy
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| |
Collapse
|
35
|
Zhang R, Yang F, Luo Y, Liu J, Li J, Wang C. Part-Aware Mask-Guided Attention for Thorax Disease Classification. ENTROPY 2021; 23:e23060653. [PMID: 34070982 PMCID: PMC8224595 DOI: 10.3390/e23060653] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/29/2021] [Revised: 05/20/2021] [Accepted: 05/21/2021] [Indexed: 11/24/2022]
Abstract
Thorax disease classification is a challenging task due to complex pathologies and subtle texture changes, etc. It has been extensively studied for years largely because of its wide application in computer-aided diagnosis. Most existing methods directly learn global feature representations from whole Chest X-ray (CXR) images, without considering in depth the richer visual cues lying around informative local regions. Thus, these methods often produce sub-optimal thorax disease classification performance because they ignore the very informative pathological changes around organs. In this paper, we propose a novel Part-Aware Mask-Guided Attention Network (PMGAN) that learns complementary global and local feature representations from all-organ region and multiple single-organ regions simultaneously for thorax disease classification. Specifically, multiple innovative soft attention modules are designed to progressively guide feature learning toward the global informative regions of whole CXR image. A mask-guided attention module is designed to further search for informative regions and visual cues within the all-organ or single-organ images, where attention is elegantly regularized by automatically generated organ masks and without introducing computation during the inference stage. In addition, a multi-task learning strategy is designed, which effectively maximizes the learning of complementary local and global representations. The proposed PMGAN has been evaluated on the ChestX-ray14 dataset and the experimental results demonstrate its superior thorax disease classification performance against the state-of-the-art methods.
Collapse
Affiliation(s)
- Ruihua Zhang
- School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China; (R.Z.); (Y.L.); (C.W.)
- Key Laboratory of Trustworthy Distributed Computing and Service, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Fan Yang
- School of Electronics Engineering and Computer Science, Peking University, Beijing 100871, China;
| | - Yan Luo
- School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China; (R.Z.); (Y.L.); (C.W.)
- Key Laboratory of Trustworthy Distributed Computing and Service, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Jianyi Liu
- School of Cyberspace Security, Beijing University of Posts and Telecommunications, Beijing 100876, China
- Correspondence:
| | - Jinbin Li
- Local Servive Center, National Population Health Data Center, Beijing 100005, China;
| | - Cong Wang
- School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China; (R.Z.); (Y.L.); (C.W.)
- Key Laboratory of Trustworthy Distributed Computing and Service, Beijing University of Posts and Telecommunications, Beijing 100876, China
| |
Collapse
|
36
|
Chen B, Zhang Z, Lin J, Chen Y, Lu G. Two-stream collaborative network for multi-label chest X-ray Image classification with lung segmentation. Pattern Recognit Lett 2020. [DOI: 10.1016/j.patrec.2020.04.016] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
|