1
|
Durgam R, Panduri B, Balaji V, Khadidos AO, Khadidos AO, Selvarajan S. Enhancing lung cancer detection through integrated deep learning and transformer models. Sci Rep 2025; 15:15614. [PMID: 40320438 PMCID: PMC12050323 DOI: 10.1038/s41598-025-00516-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2024] [Accepted: 04/29/2025] [Indexed: 05/08/2025] Open
Abstract
Lung cancer has been stated as one of the prevalent killers of cancer up to this present time and this clearly underlines the rationale for early diagnosis to enhance life expectancy of patients afflicted with the condition. The reasons behind the usage of the transformer and deep learning classifiers for the detection of lung cancer include accuracy, robustness along with the capability to handle and evaluate large data sets and much more. Such models can be more complex and can help to utilize multiple modalities of data to give extensive information that will be critical in ascertaining the right diagnosis at the right time. However, the existing works encounter several limitations including reliance on large annotated data, overfitting, high computation complexity, and interpretability. Third, the issue of the stability of these models' performance when applied to actual clinical datasets is still an open question; this is an even bigger issue that will greatly reduce the actual utilization of these models in clinical practice. To tackle these, we develop a novel Cancer Nexus Synergy (CanNS), which applies of A. Swin-Transformer UNet (SwiNet) Model for segmentation, Xception-LSTM GAN (XLG) CancerNet for classification, and Devilish Levy Optimization (DevLO) for fine-tuning parameters. This paper breaks new ground in that the presented elements are incorporated in a manner that co-operatively elevates the diagnostic capabilities while at the same time being computationally light and resilient. These are SwiNet for segmented analysis, XLG CancerNet for precise classification of the cases, and DevLO that optimizes the parameters of the lung cancer detection system, making the system more sensible and efficient. The performance outcomes indicate that the CanNS framework enhances the detection's accuracy, sensitivity, and specificity compared to the previous approaches.
Collapse
Affiliation(s)
- Revathi Durgam
- Department of Data Science, AVN Institute of Engineering and Technology, Hyderabad, India
| | - Bharathi Panduri
- Department of Information Technology, Gokaraju Rangaraju Institute of Engineering & Technology, Hyderabad, India
| | - V Balaji
- Department of CSE-AIML, Vardhaman College of Engineering, Hyderabad, India
| | - Adil O Khadidos
- Department of Information Technology, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Alaa O Khadidos
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, 21589, Saudi Arabia
- Center of Research Excellence in Artificial Intelligence and Data Science, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Shitharth Selvarajan
- Department of Computer Science and Engineering, Kebri Dehar University, Kebri Dehar, Ethiopia.
- Centre for Research Impact & Outcome, Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, Rajpura, 140401, India.
- Department of Computer Science and Engineering, Chennai Institute of Technology, Chennai, India.
| |
Collapse
|
2
|
Nawabi J, Eminovic S, Hartenstein A, Baumgaertner GL, Schnurbusch N, Rudolph M, Wasilewski D, Onken J, Siebert E, Wiener E, Bohner G, Dell’Orco A, Wattjes MP, Hamm B, Fehrenbach U, Penzkofer T. Bayesian-Optimized Convolutional Neural Networks for Classifying Primary Tumor Origin of Brain Metastases from MRI. Brain Sci 2025; 15:450. [PMID: 40426621 PMCID: PMC12110443 DOI: 10.3390/brainsci15050450] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2025] [Revised: 04/13/2025] [Accepted: 04/16/2025] [Indexed: 05/29/2025] Open
Abstract
Background/Objectives: This study evaluates whether convolutional neural networks (CNNs) can be trained to determine the primary tumor origin from MRI images alone in patients with metastatic brain lesions. Methods: This retrospective, monocentric study involved the segmentation of 1175 brain lesions from MRI scans of 436 patients with histologically confirmed primary tumor origins. The four most common tumor types-lung adenocarcinoma, small cell lung cancer, breast cancer, and melanoma-were selected, and a class-balanced dataset was created through under-sampling. This resulted in 276 training datasets and 88 hold-out test datasets. Bayesian optimization was employed to determine the optimal CNN architecture, the most relevant imaging sequences, and whether the masking of images was necessary. We compared the performance of the CNN with that of two expert radiologists specializing in neuro-oncological imaging. Results: The best-performing CNN from the Bayesian optimization process used masked images across all available MRI sequences. It achieved Area-Under-the-Curve (AUC) values of 0.75 for melanoma, 0.65 for small cell lung cancer, 0.64 for breast cancer, and 0.57 for lung adenocarcinoma. Masked images likely improved performance by focusing the CNN on relevant regions and reducing noise from surrounding tissues. In comparison, Radiologist 1 achieved AUCs of 0.55, 0.52, 0.45, and 0.51, and Radiologist 2 achieved AUCs of 0.68, 0.55, 0.64, and 0.43 for the same tumor types, respectively. The CNN consistently showed higher accuracy, particularly for melanoma and breast cancer. Conclusions: Bayesian optimization enabled the creation of a CNN that outperformed expert radiologists in classifying the primary tumor origin of brain metastases from MRI.
Collapse
Affiliation(s)
- Jawed Nawabi
- Department of Neuroradiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (J.N.); (E.S.); (E.W.); (G.B.); (A.D.); (M.P.W.)
| | - Semil Eminovic
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
| | | | - Georg Lukas Baumgaertner
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
| | - Nils Schnurbusch
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
| | - Madhuri Rudolph
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
| | - David Wasilewski
- Department of Neurosurgery, Medical Faculty & University Hospital Düsseldorf, Heinrich Heine University Dusseldorf, 40225 Dusseldorf, Germany;
| | - Julia Onken
- Department of Neurosurgery, Charité—Universitätsmedizin, 10117 Berlin, Germany;
| | - Eberhard Siebert
- Department of Neuroradiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (J.N.); (E.S.); (E.W.); (G.B.); (A.D.); (M.P.W.)
| | - Edzard Wiener
- Department of Neuroradiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (J.N.); (E.S.); (E.W.); (G.B.); (A.D.); (M.P.W.)
| | - Georg Bohner
- Department of Neuroradiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (J.N.); (E.S.); (E.W.); (G.B.); (A.D.); (M.P.W.)
| | - Andrea Dell’Orco
- Department of Neuroradiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (J.N.); (E.S.); (E.W.); (G.B.); (A.D.); (M.P.W.)
| | - Mike P. Wattjes
- Department of Neuroradiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (J.N.); (E.S.); (E.W.); (G.B.); (A.D.); (M.P.W.)
| | - Bernd Hamm
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
| | - Uli Fehrenbach
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
| | - Tobias Penzkofer
- Department of Radiology, Charité—Universitätsmedizin, 10117 Berlin, Germany; (G.L.B.); (N.S.); (M.R.); (B.H.); (U.F.); (T.P.)
- Berlin Institute of Health (BIH), 10117 Berlin, Germany
| |
Collapse
|
3
|
Aburass S, Dorgham O, Al Shaqsi J, Abu Rumman M, Al-Kadi O. Vision Transformers in Medical Imaging: a Comprehensive Review of Advancements and Applications Across Multiple Diseases. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01481-y. [PMID: 40164818 DOI: 10.1007/s10278-025-01481-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/27/2024] [Revised: 02/15/2025] [Accepted: 03/11/2025] [Indexed: 04/02/2025]
Abstract
The rapid advancement of artificial intelligence techniques, particularly deep learning, has transformed medical imaging. This paper presents a comprehensive review of recent research that leverage vision transformer (ViT) models for medical image classification across various disciplines. The medical fields of focus include breast cancer, skin lesions, magnetic resonance imaging brain tumors, lung diseases, retinal and eye analysis, COVID-19, heart diseases, colon cancer, brain disorders, diabetic retinopathy, skin diseases, kidney diseases, lymph node diseases, and bone analysis. Each work is critically analyzed and interpreted with respect to its performance, data preprocessing methodologies, model architecture, transfer learning techniques, model interpretability, and identified challenges. Our findings suggest that ViT shows promising results in the medical imaging domain, often outperforming traditional convolutional neural networks (CNN). A comprehensive overview is presented in the form of figures and tables summarizing the key findings from each field. This paper provides critical insights into the current state of medical image classification using ViT and highlights potential future directions for this rapidly evolving research area.
Collapse
Affiliation(s)
- Sanad Aburass
- Department of Computer Science, Luther College, Decorah, IA, USA.
| | - Osama Dorgham
- Prince Abdullah Bin Ghazi Faculty of Information and Communication Technology, Al-Balqa Applied University, Al-Salt, Jordan
| | - Jamil Al Shaqsi
- Information Systems Department, Sultan Qaboos University, Seeb, Oman
| | - Maha Abu Rumman
- Prince Abdullah Bin Ghazi Faculty of Information and Communication Technology, Al-Balqa Applied University, Al-Salt, Jordan
| | - Omar Al-Kadi
- Artificial Intelligence Department, King Abdullah II School of Information Technology, University of Jordan, Amman, 11942, Jordan
| |
Collapse
|
4
|
Volovăț SR, Boboc DI, Ostafe MR, Buzea CG, Agop M, Ochiuz L, Rusu DI, Vasincu D, Ungureanu MI, Volovăț CC. Utilizing Vision Transformers for Predicting Early Response of Brain Metastasis to Magnetic Resonance Imaging-Guided Stage Gamma Knife Radiosurgery Treatment. Tomography 2025; 11:15. [PMID: 39997998 PMCID: PMC11860310 DOI: 10.3390/tomography11020015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2024] [Revised: 01/11/2025] [Accepted: 02/01/2025] [Indexed: 02/26/2025] Open
Abstract
BACKGROUND/OBJECTIVES This study explores the application of vision transformers to predict early responses to stereotactic radiosurgery in patients with brain metastases using minimally pre-processed magnetic resonance imaging scans. The objective is to assess the potential of vision transformers as a predictive tool for clinical decision-making, particularly in the context of imbalanced datasets. METHODS We analyzed magnetic resonance imaging scans from 19 brain metastases patients, focusing on axial fluid-attenuated inversion recovery and high-resolution contrast-enhanced T1-weighted sequences. Patients were categorized into responders (complete or partial response) and non-responders (stable or progressive disease). RESULTS Despite the imbalanced nature of the dataset, our results demonstrate that vision transformers can predict early treatment responses with an overall accuracy of 99%. The model exhibited high precision (99% for progression and 100% for regression) and recall (99% for progression and 100% for regression). The use of the attention mechanism in the vision transformers allowed the model to focus on relevant features in the magnetic resonance imaging images, ensuring an unbiased performance even with the imbalanced data. Confusion matrix analysis further confirmed the model's reliability, with minimal misclassifications. Additionally, the model achieved a perfect area under the receiver operator characteristic curve (AUC = 1.00), effectively distinguishing between responders and non-responders. CONCLUSIONS These findings highlight the potential of vision transformers, aided by the attention mechanism, as a non-invasive, predictive tool for early response assessment in clinical oncology. The vision transformer (ViT) model employed in this study processes MRIs as sequences of patches, enabling the capture of localized tumor features critical for early response prediction. By leveraging patch-based feature learning, this approach enhances robustness, interpretability, and clinical applicability, addressing key challenges in tumor progression prediction following stereotactic radiosurgery (SRS). The model's robust performance, despite the dataset imbalance, underscores its ability to provide unbiased predictions. This approach could significantly enhance clinical decision-making and support personalized treatment strategies for brain metastases. Future research should validate these findings in larger, more diverse cohorts and explore the integration of additional data types to further optimize the model's clinical utility.
Collapse
Affiliation(s)
- Simona Ruxandra Volovăț
- Medical Oncology-Radiotherapy Department, “Grigore T. Popa” University of Medicine and Pharmacy Iași, 700115 Iași, Romania; (S.R.V.); (D.-I.B.); (M.-R.O.)
| | - Diana-Ioana Boboc
- Medical Oncology-Radiotherapy Department, “Grigore T. Popa” University of Medicine and Pharmacy Iași, 700115 Iași, Romania; (S.R.V.); (D.-I.B.); (M.-R.O.)
| | - Mădălina-Raluca Ostafe
- Medical Oncology-Radiotherapy Department, “Grigore T. Popa” University of Medicine and Pharmacy Iași, 700115 Iași, Romania; (S.R.V.); (D.-I.B.); (M.-R.O.)
| | - Călin Gheorghe Buzea
- “Prof. Dr. Nicolae Oblu” Clinical Emergency Hospital Iași, 700309 Iași, Romania;
- National Institute of Research and Development for Technical Physics, IFT Iași, 700050 Iași, Romania
| | - Maricel Agop
- Physics Department, “Gheorghe Asachi” Technical University Iași, 700050 Iași, Romania;
| | - Lăcrămioara Ochiuz
- Faculty of Pharmacy, “Grigore T. Popa” University of Medicine and Pharmacy Iași, 700115 Iași, Romania;
| | - Dragoș Ioan Rusu
- Faculty of Science, “V. Alecsandri” University of Bacău, 600115 Bacău, Romania;
| | - Decebal Vasincu
- Surgery Department, “Grigore T. Popa” University of Medicine and Pharmacy Iași, 700115 Iași, Romania;
| | - Monica Iuliana Ungureanu
- Preventive Medicine and Interdisciplinarity Department, “Grigore T. Popa” University of Medicine and Pharmacy Iași, 700115 Iași, Romania
| | | |
Collapse
|
5
|
Wang F, Zou Z, Sakla N, Partyka L, Rawal N, Singh G, Zhao W, Ling H, Huang C, Prasanna P, Chen C. TopoTxR: A topology-guided deep convolutional network for breast parenchyma learning on DCE-MRIs. Med Image Anal 2025; 99:103373. [PMID: 39454312 DOI: 10.1016/j.media.2024.103373] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Revised: 09/28/2024] [Accepted: 10/10/2024] [Indexed: 10/28/2024]
Abstract
Characterization of breast parenchyma in dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI) is a challenging task owing to the complexity of underlying tissue structures. Existing quantitative approaches, like radiomics and deep learning models, lack explicit quantification of intricate and subtle parenchymal structures, including fibroglandular tissue. To address this, we propose a novel topological approach that explicitly extracts multi-scale topological structures to better approximate breast parenchymal structures, and then incorporates these structures into a deep-learning-based prediction model via an attention mechanism. Our topology-informed deep learning model, TopoTxR, leverages topology to provide enhanced insights into tissues critical for disease pathophysiology and treatment response. We empirically validate TopoTxR using the VICTRE phantom breast dataset, showing that the topological structures extracted by our model effectively approximate the breast parenchymal structures. We further demonstrate TopoTxR's efficacy in predicting response to neoadjuvant chemotherapy. Our qualitative and quantitative analyses suggest differential topological behavior of breast tissue in treatment-naïve imaging, in patients who respond favorably to therapy as achieving pathological complete response (pCR) versus those who do not. In a comparative analysis with several baselines on the publicly available I-SPY 1 dataset (N = 161, including 47 patients with pCR and 114 without) and the Rutgers proprietary dataset (N = 120, with 69 patients achieving pCR and 51 not), TopoTxR demonstrates a notable improvement, achieving a 2.6% increase in accuracy and a 4.6% enhancement in AUC compared to the state-of-the-art method.
Collapse
Affiliation(s)
- Fan Wang
- Department of Computer Science, State University of New York at Stony Brook, NY, USA.
| | - Zhilin Zou
- Department of Computer Science, State University of New York at Stony Brook, NY, USA
| | - Nicole Sakla
- Department of Radiology, Newark Beth Israel Medical Center, NJ, USA
| | - Luke Partyka
- Department of Radiology, Newark Beth Israel Medical Center, NJ, USA
| | - Nil Rawal
- Department of Radiology, Newark Beth Israel Medical Center, NJ, USA
| | - Gagandeep Singh
- Department of Radiology, Columbia University Irving Medical Center, NY, USA
| | - Wei Zhao
- Department of Radiology, State University of New York at Stony Brook, NY, USA
| | - Haibin Ling
- Department of Computer Science, State University of New York at Stony Brook, NY, USA
| | - Chuan Huang
- Department of Radiology and Imaging Sciences, Emory University School of Medicine, GA, USA; Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, GA, USA
| | - Prateek Prasanna
- Department of Biomedical Informatics, State University of New York at Stony Brook, NY, USA.
| | - Chao Chen
- Department of Biomedical Informatics, State University of New York at Stony Brook, NY, USA.
| |
Collapse
|
6
|
Chukwujindu E, Faiz H, Ai-Douri S, Faiz K, De Sequeira A. Role of artificial intelligence in brain tumour imaging. Eur J Radiol 2024; 176:111509. [PMID: 38788610 DOI: 10.1016/j.ejrad.2024.111509] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2024] [Revised: 04/29/2024] [Accepted: 05/13/2024] [Indexed: 05/26/2024]
Abstract
Artificial intelligence (AI) is a rapidly evolving field with many neuro-oncology applications. In this review, we discuss how AI can assist in brain tumour imaging, focusing on machine learning (ML) and deep learning (DL) techniques. We describe how AI can help in lesion detection, differential diagnosis, anatomic segmentation, molecular marker identification, prognostication, and pseudo-progression evaluation. We also cover AI applications in non-glioma brain tumours, such as brain metastasis, posterior fossa, and pituitary tumours. We highlight the challenges and limitations of AI implementation in radiology, such as data quality, standardization, and integration. Based on the findings in the aforementioned areas, we conclude that AI can potentially improve the diagnosis and treatment of brain tumours and provide a path towards personalized medicine and better patient outcomes.
Collapse
Affiliation(s)
| | | | | | - Khunsa Faiz
- McMaster University, Department of Radiology, L8S 4L8, Canada.
| | | |
Collapse
|
7
|
Strotzer QD, Wagner T, Angstwurm P, Hense K, Scheuermeyer L, Noeva E, Dinkel J, Stroszczynski C, Fellner C, Riemenschneider MJ, Rosengarth K, Pukrop T, Wiesinger I, Wendl C, Schicho A. Limited capability of MRI radiomics to predict primary tumor histology of brain metastases in external validation. Neurooncol Adv 2024; 6:vdae060. [PMID: 38800697 PMCID: PMC11125388 DOI: 10.1093/noajnl/vdae060] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/29/2024] Open
Abstract
Background Growing research demonstrates the ability to predict histology or genetic information of various malignancies using radiomic features extracted from imaging data. This study aimed to investigate MRI-based radiomics in predicting the primary tumor of brain metastases through internal and external validation, using oversampling techniques to address the class imbalance. Methods This IRB-approved retrospective multicenter study included brain metastases from lung cancer, melanoma, breast cancer, colorectal cancer, and a combined heterogenous group of other primary entities (5-class classification). Local data were acquired between 2003 and 2021 from 231 patients (545 metastases). External validation was performed with 82 patients (280 metastases) and 258 patients (809 metastases) from the publicly available Stanford BrainMetShare and the University of California San Francisco Brain Metastases Stereotactic Radiosurgery datasets, respectively. Preprocessing included brain extraction, bias correction, coregistration, intensity normalization, and semi-manual binary tumor segmentation. Two-thousand five hundred and twenty-eight radiomic features were extracted from T1w (± contrast), fluid-attenuated inversion recovery (FLAIR), and wavelet transforms for each sequence (8 decompositions). Random forest classifiers were trained with selected features on original and oversampled data (5-fold cross-validation) and evaluated on internal/external holdout test sets using accuracy, precision, recall, F1 score, and area under the receiver-operating characteristic curve (AUC). Results Oversampling did not improve the overall unsatisfactory performance on the internal and external test sets. Incorrect data partitioning (oversampling before train/validation/test split) leads to a massive overestimation of model performance. Conclusions Radiomics models' capability to predict histologic or genomic data from imaging should be critically assessed; external validation is essential.
Collapse
Affiliation(s)
- Quirin D Strotzer
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
- Division of Neuroradiology, Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts, USA
| | - Thomas Wagner
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| | - Pia Angstwurm
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| | - Katharina Hense
- Department of Neurosurgery, University Medical Center Regensburg, Regensburg, Germany
| | - Lucca Scheuermeyer
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| | - Ekaterina Noeva
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| | - Johannes Dinkel
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| | | | - Claudia Fellner
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| | | | - Katharina Rosengarth
- Department of Neurosurgery, University Medical Center Regensburg, Regensburg, Germany
| | - Tobias Pukrop
- Department of Internal Medicine III—Hematology and Oncology, University Medical Center Regensburg, Regensburg, Germany
| | - Isabel Wiesinger
- Center of Neuroradiology, medbo District Hospital and University Medical Center Regensburg, Regensburg, Germany
| | - Christina Wendl
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
- Center of Neuroradiology, medbo District Hospital and University Medical Center Regensburg, Regensburg, Germany
| | - Andreas Schicho
- Department of Radiology, University Medical Center Regensburg, Regensburg, Germany
| |
Collapse
|
8
|
Egashira M, Arimura H, Kobayashi K, Moriyama K, Kodama T, Tokuda T, Ninomiya K, Okamoto H, Igaki H. Magnetic resonance-based imaging biopsy with signatures including topological Betti number features for prediction of primary brain metastatic sites. Phys Eng Sci Med 2023; 46:1411-1426. [PMID: 37603131 DOI: 10.1007/s13246-023-01308-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Accepted: 07/20/2023] [Indexed: 08/22/2023]
Abstract
This study incorporated topology Betti number (BN) features into the prediction of primary sites of brain metastases and the construction of magnetic resonance-based imaging biopsy (MRB) models. The significant features of the MRB model were selected from those obtained from gray-scale and three-dimensional wavelet-filtered images, BN and inverted BN (iBN) maps, and clinical variables (age and gender). The primary sites were predicted as either lung cancer or other cancers using MRB models, which were built using seven machine learning methods with significant features chosen by three feature selection methods followed by a combination strategy. Our study dealt with a dataset with relatively smaller brain metastases, which included effective diameters greater than 2 mm, with metastases ranging from 2 to 9 mm accounting for 17% of the dataset. The MRB models were trained by T1-weighted contrast-enhanced images of 494 metastases chosen from 247 patients and applied to 115 metastases from 62 test patients. The most feasible model attained an area under the receiver operating characteristic curve (AUC) of 0.763 for the test patients when using a signature including features of BN and iBN maps, gray-scale and wavelet-filtered images, and clinical variables. The AUCs of the model were 0.744 for non-small cell lung cancer and 0.861 for small cell lung cancer. The results suggest that the BN signature boosted the performance of MRB for the identification of primary sites of brain metastases including small tumors.
Collapse
Affiliation(s)
- Mai Egashira
- Division of Medical Quantum Science, Department of Health Science, Graduate School of Medical Science, Kyushu University, 3-1-1 Maidashi, Higashi-ku, Fukuoka, 812-8582, Japan
| | - Hidetaka Arimura
- Division of Medical Quantum Science, Department of Health Sciences, Faculty of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashi-ku, Fukuoka, 812-8582, Japan.
| | - Kazuma Kobayashi
- Department of Medical AI Research and Development, National Cancer Center Research Institute, 5-1-1 Tsukiji, Chuo-ku, Tokyo, 104-0045, Japan.
| | - Kazutoshi Moriyama
- Division of Medical Quantum Science, Department of Health Science, Graduate School of Medical Science, Kyushu University, 3-1-1 Maidashi, Higashi-ku, Fukuoka, 812-8582, Japan
| | - Takumi Kodama
- Division of Medical Quantum Science, Department of Health Science, Graduate School of Medical Science, Kyushu University, 3-1-1 Maidashi, Higashi-ku, Fukuoka, 812-8582, Japan
| | - Tomoki Tokuda
- Joint Graduate School of Mathematics for Innovation, Kyushu University, 744 Motooka, Nishi-ku, Fukuoka, 819-0395, Japan
| | - Kenta Ninomiya
- Sanford Burnham Prebys Medical Discovery Institute, San Diego, CA, USA
| | - Hiroyuki Okamoto
- Radiation Safety and Quality Assurance Division, National Cancer Center Hospital, 5-1-1 Tsukiji, Chuo-ku, Tokyo, 104-0045, Japan
| | - Hiroshi Igaki
- Department of Radiation Oncology, National Cancer Center Hospital, 5-1-1 Tsukiji, Chuo-ku, Tokyo, 104-0045, Japan
| |
Collapse
|
9
|
Jiao T, Li F, Cui Y, Wang X, Li B, Shi F, Xia Y, Zhou Q, Zeng Q. Deep Learning With an Attention Mechanism for Differentiating the Origin of Brain Metastasis Using MR images. J Magn Reson Imaging 2023; 58:1624-1635. [PMID: 36965182 DOI: 10.1002/jmri.28695] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Revised: 03/10/2023] [Accepted: 03/10/2023] [Indexed: 03/27/2023] Open
Abstract
BACKGROUND Brain metastasis (BM) is a serious neurological complication of cancer of different origins. The value of deep learning (DL) to identify multiple types of primary origins remains unclear. PURPOSE To distinguish primary site of BM and identify the best DL models. STUDY TYPE Retrospective. POPULATION A total of 449 BM derived from 214 patients (49.5% for female, mean age 58 years) (100 from small cell lung cancer [SCLC], 125 from non-small cell lung cancer [NSCLC], 116 from breast cancer [BC], and 108 from gastrointestinal cancer [GIC]) were included. FIELD STRENGTH/SEQUENCE A 3-T, T1 turbo spin echo (T1-TSE), T2-TSE, T2FLAIR-TSE, DWI echo-planar imaging (DWI-EPI) and contrast-enhanced T1-TSE (CE T1-TSE). ASSESSMENT Lesions were divided into training (n = 285, 153 patients), testing (n = 122, 93 patients), and independent testing cohorts (n = 42, 34 patients). Three-dimensional residual network (3D-ResNet), named 3D ResNet6 and 3D ResNet 18, was proposed for identifying the four origins based on single MRI and combined MRI (T1WI + T2-FLAIR + DWI, CE-T1WI + DWI, CE-T1WI + T2WI + DWI). DL model was used to distinguish lung cancer from non-lung cancer; then SCLC vs. NSCLC for lung cancer classification and BC vs. GIC for non-lung cancer classification was performed. A subjective visual analysis was implemented and compared with DL models. Gradient-weighted class activation mapping (Grad-CAM) was used to visualize the model by heatmaps. STATISTICAL TESTS The area under the receiver operating characteristics curve (AUC) assess each classification performance. RESULTS 3D ResNet18 with Grad-CAM and AIC showed better performance than 3DResNet6, 3DResNet18 and the radiologist for distinguishing lung cancer from non-lung cancer, SCLC from NSCLC, and BC from GIC. For single MRI sequence, T1WI, DWI, and CE-T1WI performed best for lung cancer vs. non-lung cancer, SCLC vs. NSCLC, and BC vs. GIC classifications. The AUC ranged from 0.675 to 0.876 and from 0.684 to 0.800 regarding the testing and independent testing datasets, respectively. For combined MRI sequences, the combination of CE-T1WI + T2WI + DWI performed better for BC vs. GIC (AUCs of 0.788 and 0.848 on testing and independent testing datasets, respectively), while the combined MRI approach (T1WI + T2-FLAIR + DWI, CE-T1WI + DWI) could not achieve higher AUCs for lung cancer vs. non-lung cancer, SCLC vs. NSCLC. Grad-CAM helped for model visualization by heatmaps that focused on tumor regions. DATA CONCLUSION DL models may help to distinguish the origins of BM based on MRI data. EVIDENCE LEVEL 3 TECHNICAL EFFICACY: Stage 2.
Collapse
Affiliation(s)
- Tianyu Jiao
- Department of Radiology, The First Affiliated Hospital of Shandong First Medical University & Shandong Provincial Qianfoshan Hospital, Jinan, China
- Shandong First Medical University, Jinan, China
| | - Fuyan Li
- Department of Radiology, Shandong Provincial Hospital affiliated to Shandong First Medical University, Jinan, China
| | - Yi Cui
- Department of Radiology, Qilu Hospital of Shandong University, Jinan, China
| | - Xiao Wang
- Department of Radiology, Jining No. 1 People's Hospital, Jining, China
| | - Butuo Li
- Department of Radiation Oncology, Shandong Cancer Hospital & Institute, Jinan, China
| | - Feng Shi
- Shanghai United Imaging Intelligence, Shanghai, China
| | - Yuwei Xia
- Shanghai United Imaging Intelligence, Shanghai, China
| | - Qing Zhou
- Shanghai United Imaging Intelligence, Shanghai, China
| | - Qingshi Zeng
- Department of Radiology, The First Affiliated Hospital of Shandong First Medical University & Shandong Provincial Qianfoshan Hospital, Jinan, China
| |
Collapse
|
10
|
Al-Hammuri K, Gebali F, Kanan A, Chelvan IT. Vision transformer architecture and applications in digital health: a tutorial and survey. Vis Comput Ind Biomed Art 2023; 6:14. [PMID: 37428360 PMCID: PMC10333157 DOI: 10.1186/s42492-023-00140-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Accepted: 05/30/2023] [Indexed: 07/11/2023] Open
Abstract
The vision transformer (ViT) is a state-of-the-art architecture for image recognition tasks that plays an important role in digital health applications. Medical images account for 90% of the data in digital medicine applications. This article discusses the core foundations of the ViT architecture and its digital health applications. These applications include image segmentation, classification, detection, prediction, reconstruction, synthesis, and telehealth such as report generation and security. This article also presents a roadmap for implementing the ViT in digital health systems and discusses its limitations and challenges.
Collapse
Affiliation(s)
- Khalid Al-Hammuri
- Electrical and Computer Engineering, University of Victoria, Victoria, V8W 2Y2, Canada.
| | - Fayez Gebali
- Electrical and Computer Engineering, University of Victoria, Victoria, V8W 2Y2, Canada
| | - Awos Kanan
- Computer Engineering, Princess Sumaya University for Technology, Amman, 11941, Jordan
| | | |
Collapse
|