1
|
Li C, Li J, Wang YZ. Data integrity of food and machine learning: Strategies, advances and prospective. Food Chem 2025; 480:143831. [PMID: 40120309 DOI: 10.1016/j.foodchem.2025.143831] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2024] [Revised: 03/01/2025] [Accepted: 03/08/2025] [Indexed: 03/25/2025]
Abstract
Data integrity is an emerging concept aimed at recording real food properties in the form of data throughout the food lifecycle. However, due to the one-sided nature of current food control data, the comprehensive implementation of data integrity has not been fully achieved. Cause food data integrity realization is required to establish the connection of data-algorithm-application. Machine learning (ML) provides a possibility for the practical carrier of food data integrity. Despite ML is one of top-trend in food quality and safety, ML applications are floating on the surface. The current review does not reveal the relationships behind different algorithms and data patterns. Similarly, due to the rapid development of ML, the current advanced concepts and data explanation tools have not been systematically reviewed. This paper expounds the feasibility of machine learning to achieve data integrity and looks forward to the future vision brought about by artificial intelligence to data integrity.
Collapse
Affiliation(s)
- Chenming Li
- College of Food Science and Technology, Yunnan Agricultural University, Kunming, 650201, China; Medicinal Plants Research Institute, Yunnan, Academy of Agricultural Sciences, Kunming, 650200, China
| | - Jieqing Li
- College of Food Science and Technology, Yunnan Agricultural University, Kunming, 650201, China.
| | - Yuan-Zhong Wang
- Medicinal Plants Research Institute, Yunnan, Academy of Agricultural Sciences, Kunming, 650200, China.
| |
Collapse
|
2
|
Xia C, Zuo M, Lin Z, Deng L, Rao Y, Chen W, Chen J, Yao W, Hu M. Multimodal Deep Learning Fusing Clinical and Radiomics Scores for Prediction of Early-Stage Lung Adenocarcinoma Lymph Node Metastasis. Acad Radiol 2025; 32:2977-2989. [PMID: 39730249 DOI: 10.1016/j.acra.2024.12.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2024] [Revised: 11/27/2024] [Accepted: 12/09/2024] [Indexed: 12/29/2024]
Abstract
RATIONALE AND OBJECTIVES To develop and validate a multimodal deep learning (DL) model based on computed tomography (CT) images and clinical knowledge to predict lymph node metastasis (LNM) in early lung adenocarcinoma. MATERIALS AND METHODS A total of 724 pathologically confirmed early invasive lung adenocarcinoma patients were retrospectively included from two centers. Clinical and CT semantic features of the patients were collected, and 3D radiomics features were extracted from nonenhanced CT images. We proposed a multimodal feature fusion DL network based on the InceptionResNetV2 architecture, which can effectively extract and integrate image and clinical knowledge to predict LNM. RESULTS A total of 524 lung adenocarcinoma patients from Center 1 were randomly divided into training (n=418) and internal validation (n=106) sets in a 4:1 ratio, while 200 lung adenocarcinoma patients from Center 2 served as the independent test set. Among the 16 collected clinical and imaging features, 8 were selected: gender, serum carcinoembryonic antigen, cytokeratin 19 fragment antigen 21-1, neuron-specific enolase, tumor size, location, density, and centrality. From the 1595 extracted radiomics features, six key features were identified. The CS-RS-DL fusion model achieved the highest area under the receiver operating characteristic curve in both the internal validation set (0.877) and the independent test set (0.906) compared to other models. The Delong test results for the independent test set indicated that the CS-RS-DL model significantly outperformed the clinical model (0.844), radiomics model (0.850), CS-RS model (0.872), single DL model (0.848), and the CS-DL model (0.875) (all P<0.05). Additionally, the CS-RS-DL model exhibited the highest sensitivity (0.941) and average precision (0.642). CONCLUSION The knowledge derived from clinical, radiomics, and DL is complementary in predicting LNM in lung adenocarcinoma. The integration of clinical and radiomics scores through DL can significantly improve the accuracy of lymph node status assessment.
Collapse
Affiliation(s)
- Chengcheng Xia
- School of Public Health, Jiangxi Medical College, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.); Jiangxi Provincial Key Laboratory of Disease Prevention and Public Health, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.)
| | - Minjing Zuo
- Department of Radiology, The Second Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang 330006, China (M.Z.); Intelligent Medical Imaging of Jiangxi Key Laboratory, Nanchang 330006, China (M.Z.)
| | - Ze Lin
- Department of Radiology, Hubei Provincial Hospital of Traditional Chinese Medicine, Wuhan 430022, China (Z.L.); Affiliated Hospital of Hubei University of Chinese Medicine, Wuhan 430022, China (Z.L.)
| | - Libin Deng
- School of Public Health, Jiangxi Medical College, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.); Jiangxi Provincial Key Laboratory of Disease Prevention and Public Health, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.)
| | - Yulian Rao
- Wanli District Center for Disease Control and Prevention of Nanchang, Nanchang 330004, China (Y.R.)
| | - Wenxiang Chen
- School of Public Health, Jiangxi Medical College, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.); Jiangxi Provincial Key Laboratory of Disease Prevention and Public Health, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.)
| | - Jinqin Chen
- Jiangxi Medical College, Nanchang University, Nanchang, China (J.C.)
| | - Weirong Yao
- Department of Oncology, Jiangxi Provincial People's Hospital, The First Affiliated Hospital of Nanchang Medical College, Nanchang, China (W.Y.)
| | - Min Hu
- School of Public Health, Jiangxi Medical College, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.); Jiangxi Provincial Key Laboratory of Disease Prevention and Public Health, Nanchang University, Nanchang 330006, China (C.X., L.D., W.C., M.H.).
| |
Collapse
|
3
|
Zubair M, Owais M, Hassan T, Bendechache M, Hussain M, Hussain I, Werghi N. An interpretable framework for gastric cancer classification using multi-channel attention mechanisms and transfer learning approach on histopathology images. Sci Rep 2025; 15:13087. [PMID: 40240457 PMCID: PMC12003787 DOI: 10.1038/s41598-025-97256-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2024] [Accepted: 04/03/2025] [Indexed: 04/18/2025] Open
Abstract
The importance of gastric cancer (GC) and the role of deep learning techniques in categorizing GC histopathology images have recently increased. Identifying the drawbacks of traditional deep learning models, including lack of interpretability, inability to capture complex patterns, lack of adaptability, and sensitivity to noise. A multi-channel attention mechanism-based framework is proposed that can overcome the limitations of conventional deep learning models by dynamically focusing on relevant features, enhancing extraction, and capturing complex relationships in medical data. The proposed framework uses three different attention mechanism channels and convolutional neural networks to extract multichannel features during the classification process. The proposed framework's strong performance is confirmed by competitive experiments conducted on a publicly available Gastric Histopathology Sub-size Image Database, which yielded remarkable classification accuracies of 99.07% and 98.48% on the validation and testing sets, respectively. Additionally, on the HCRF dataset, the framework achieved high classification accuracy of 99.84% and 99.65% on the validation and testing sets, respectively. The effectiveness and interchangeability of the three channels are further confirmed by ablation and interchangeability experiments, highlighting the remarkable performance of the framework in GC histopathological image classification tasks. This offers an advanced and pragmatic artificial intelligence solution that addresses challenges posed by unique medical image characteristics for intricate image analysis. The proposed approach in artificial intelligence medical engineering demonstrates significant potential for enhancing diagnostic precision by achieving high classification accuracy and treatment outcomes.
Collapse
Affiliation(s)
- Muhammad Zubair
- Interdisciplinary Research Center for Finance and Digital Economy, King Fahd University of Petroleum and Minerals, 31261, Dhahran, Saudi Arabia
| | - Muhammad Owais
- Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates.
| | - Taimur Hassan
- Departement of Electrical and Computer Engineering, Abu Dhabi University, Abu Dhabi, United Arab Emirates
| | - Malika Bendechache
- ADAPT Research Centre, School of Computer Science, University of Galway, H91 TK33, Galway, Ireland
| | - Muzammil Hussain
- Department of Software Engineering, Faculty of Information Technology, Al-Ahliyya Amman University, Amman, Jordan
| | - Irfan Hussain
- Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates
| | - Naoufel Werghi
- Department of Computer Science, Khalifa University, Abu Dhabi, United Arab Emirates
| |
Collapse
|
4
|
Yuan H, Zhu M, Yang R, Liu H, Li I, Hong C. Rethinking Domain-Specific Pretraining by Supervised or Self-Supervised Learning for Chest Radiograph Classification: A Comparative Study Against ImageNet Counterparts in Cold-Start Active Learning. HEALTH CARE SCIENCE 2025; 4:110-143. [PMID: 40241982 PMCID: PMC11997468 DOI: 10.1002/hcs2.70009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/10/2024] [Revised: 01/05/2025] [Accepted: 01/26/2025] [Indexed: 04/18/2025]
Abstract
Objective Deep learning (DL) has become the prevailing method in chest radiograph analysis, yet its performance heavily depends on large quantities of annotated images. To mitigate the cost, cold-start active learning (AL), comprising an initialization followed by subsequent learning, selects a small subset of informative data points for labeling. Recent advancements in pretrained models by supervised or self-supervised learning tailored to chest radiograph have shown broad applicability to diverse downstream tasks. However, their potential in cold-start AL remains unexplored. Methods To validate the efficacy of domain-specific pretraining, we compared two foundation models: supervised TXRV and self-supervised REMEDIS with their general domain counterparts pretrained on ImageNet. Model performance was evaluated at both initialization and subsequent learning stages on two diagnostic tasks: psychiatric pneumonia and COVID-19. For initialization, we assessed their integration with three strategies: diversity, uncertainty, and hybrid sampling. For subsequent learning, we focused on uncertainty sampling powered by different pretrained models. We also conducted statistical tests to compare the foundation models with ImageNet counterparts, investigate the relationship between initialization and subsequent learning, examine the performance of one-shot initialization against the full AL process, and investigate the influence of class balance in initialization samples on initialization and subsequent learning. Results First, domain-specific foundation models failed to outperform ImageNet counterparts in six out of eight experiments on informative sample selection. Both domain-specific and general pretrained models were unable to generate representations that could substitute for the original images as model inputs in seven of the eight scenarios. However, pretrained model-based initialization surpassed random sampling, the default approach in cold-start AL. Second, initialization performance was positively correlated with subsequent learning performance, highlighting the importance of initialization strategies. Third, one-shot initialization performed comparably to the full AL process, demonstrating the potential of reducing experts' repeated waiting during AL iterations. Last, a U-shaped correlation was observed between the class balance of initialization samples and model performance, suggesting that the class balance is more strongly associated with performance at middle budget levels than at low or high budgets. Conclusions In this study, we highlighted the limitations of medical pretraining compared to general pretraining in the context of cold-start AL. We also identified promising outcomes related to cold-start AL, including initialization based on pretrained models, the positive influence of initialization on subsequent learning, the potential for one-shot initialization, and the influence of class balance on middle-budget AL. Researchers are encouraged to improve medical pretraining for versatile DL foundations and explore novel AL methods.
Collapse
Affiliation(s)
- Han Yuan
- Duke‐NUS Medical School, Centre for Quantitative MedicineSingaporeSingapore
| | - Mingcheng Zhu
- Duke‐NUS Medical School, Centre for Quantitative MedicineSingaporeSingapore
- Department of Engineering ScienceUniversity of OxfordOxfordUK
| | - Rui Yang
- Duke‐NUS Medical School, Centre for Quantitative MedicineSingaporeSingapore
| | - Han Liu
- Department of Computer ScienceVanderbilt UniversityNashvilleTennesseeUSA
| | - Irene Li
- Information Technology CenterUniversity of TokyoBunkyo‐kuJapan
| | - Chuan Hong
- Department of Biostatistics and BioinformaticsDuke UniversityDurhamNorth CarolinaUSA
| |
Collapse
|
5
|
Le Vuong TT, Kwak JT. MoMA: Momentum contrastive learning with multi-head attention-based knowledge distillation for histopathology image analysis. Med Image Anal 2025; 101:103421. [PMID: 39671769 DOI: 10.1016/j.media.2024.103421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Revised: 11/07/2024] [Accepted: 11/29/2024] [Indexed: 12/15/2024]
Abstract
There is no doubt that advanced artificial intelligence models and high quality data are the keys to success in developing computational pathology tools. Although the overall volume of pathology data keeps increasing, a lack of quality data is a common issue when it comes to a specific task due to several reasons including privacy and ethical issues with patient data. In this work, we propose to exploit knowledge distillation, i.e., utilize the existing model to learn a new, target model, to overcome such issues in computational pathology. Specifically, we employ a student-teacher framework to learn a target model from a pre-trained, teacher model without direct access to source data and distill relevant knowledge via momentum contrastive learning with multi-head attention mechanism, which provides consistent and context-aware feature representations. This enables the target model to assimilate informative representations of the teacher model while seamlessly adapting to the unique nuances of the target data. The proposed method is rigorously evaluated across different scenarios where the teacher model was trained on the same, relevant, and irrelevant classification tasks with the target model. Experimental results demonstrate the accuracy and robustness of our approach in transferring knowledge to different domains and tasks, outperforming other related methods. Moreover, the results provide a guideline on the learning strategy for different types of tasks and scenarios in computational pathology.
Collapse
Affiliation(s)
- Trinh Thi Le Vuong
- School of Electrical Engineering, Korea University, Seoul, Republic of Korea
| | - Jin Tae Kwak
- School of Electrical Engineering, Korea University, Seoul, Republic of Korea.
| |
Collapse
|
6
|
Jang BK, Kim S, Yu JY, Hong J, Cho HW, Lee HS, Park J, Woo J, Lee YH, Park YR. Classification models for arthropathy grades of multiple joints based on hierarchical continual learning. LA RADIOLOGIA MEDICA 2025:10.1007/s11547-025-01974-4. [PMID: 40126794 DOI: 10.1007/s11547-025-01974-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/28/2024] [Accepted: 02/14/2025] [Indexed: 03/26/2025]
Abstract
PURPOSE To develop a hierarchical continual arthropathy classification model for multiple joints that can be updated continuously for large-scale studies of various anatomical structures. MATERIALS AND METHODS This study included a total of 1371 radiographs of knee, elbow, ankle, shoulder, and hip joints from three tertiary hospitals. For model development, 934 radiographs of the knee, elbow, ankle, and shoulder were gathered from Sinchon Severance Hospital between July 1 and December 31, 2022. For external validation, 125 hip radiographs were collected from Yongin Severance Hospital between January 1 and December 31, 2022, and 312 knee cases were gathered from Gangnam Severance Hospital between January 1 and June 31, 2023. The Hierarchical Dynamically Expandable Representation (Hi-DER) model was trained stepwise on four joints using five-fold cross-validation. Arthropathy classification was evaluated at three hierarchical levels: abnormal classification (L1), low-grade or high-grade classification (L2), and specific grade classification (L3). The model's performance was compared with the grading predictions of two other AI models and three radiologists. For model explainability, gradient-weighted class activation mapping (Grad-CAM) and progressive erasing plus progressive restoration (PEPPR) were employed. RESULTS The model achieved a weighted average AUC of 0.994 (95% CI: 0.985, 0.999) for L1, 0.980 (95% CI: 0.958, 0.996) for L2, and 0.973 (95% CI: 0.943, 0.993) for L3. The model maintained an AUC above 0.800 with 70% of the input regions erased. During external validation on hip joints, the model demonstrated a weighted average AUC of 0.978 (95% CI: 0.952, 0.996) for L1, 0.977 (95% CI: 0.946, 0.996) for L2, and 0.971 (95% CI: 0.934, 0.996) for L3. For external knee data, the model yielded a weighted average AUC of 0.934 (95%: CI 0.904, 0.958), 0.929 (95% CI: 0.900, 0.954), and 0.857 (95% CI: 0.816, 0.894) for L1, L2, and L3, respectively. CONCLUSION The Hi-DER may enhance the efficiency of arthropathy diagnosis through accurate classification of arthropathy grades across multiple joints, potentially enabling early treatment.
Collapse
Affiliation(s)
- Bong Kyung Jang
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Shiwon Kim
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Republic of Korea
- Department of Digital Analytics, College of Computing, Yonsei University, Seoul, Republic of Korea
| | - Jae Yong Yu
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - JaeSeong Hong
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Hee Woo Cho
- Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Hong Seon Lee
- Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Jiwoo Park
- Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Jeesoo Woo
- School of Medicine, CHA University Gyeonggi-do, Pocheon, Republic of Korea
| | - Young Han Lee
- Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, Seoul, Republic of Korea.
- Institute for Innovation in Digital Healthcare, Yonsei University, Seoul, Republic of Korea.
| | - Yu Rang Park
- Department of Biomedical Systems Informatics, Yonsei University College of Medicine, Seoul, Republic of Korea.
- Department of Digital Analytics, College of Computing, Yonsei University, Seoul, Republic of Korea.
- Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, Seoul, Republic of Korea.
- Institute for Innovation in Digital Healthcare, Yonsei University, Seoul, Republic of Korea.
| |
Collapse
|
7
|
Sujeeun LY, Phul IC, Goonoo N, Kotov NA, Bhaw-Luximon A. Predicting inflammatory response of biomimetic nanofibre scaffolds for tissue regeneration using machine learning and graph theory. J Mater Chem B 2025; 13:3304-3318. [PMID: 39869000 DOI: 10.1039/d4tb02494j] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/28/2025]
Abstract
Tissue regeneration after a wound occurs through three main overlapping and interrelated stages namely inflammatory, proliferative, and remodelling phases, respectively. The inflammatory phase is key for successful tissue reconstruction and triggers the proliferative phase. The macrophages in the non-healing wounds remain in the inflammatory loop, but their phenotypes can be changed via interactions with nanofibre-based scaffolds mimicking the organisation of the native structural support of healthy tissues. However, the organisation of extracellular matrix (ECM) is highly complex, combining order and disorder, which makes it difficult to replicate. The possibility of predicting the desirable biomimetic geometry and chemistry of these nanofibre scaffolds would streamline the scaffold design process. Fifteen families of nanofibre scaffolds, electrospun from combinations of polyesters (polylactide, polyhydroxybutyrate), polysaccharides (polysucrose, carrageenan, cellulose), and polyester ether (polydioxanone) were investigated and analysed using machine learning (ML). The Random Forest model had the best performance (92.8%) in predicting inflammatory responses of macrophages on the nanoscaffolds using tumour necrosis factor-alpha as the output. CellProfiler proved to be an effective tool to process scanning electron microscopy (SEM) images of the macrophages on the scaffolds, successfully extracting various features and measurements related to cell phenotypes M0, M1, and M2. Deep learning modelling indicated that convolutional neural network models have the potential to be applied to SEM images to classify macrophage cells according to their phenotypes. The complex organisation of the nanofibre scaffolds can be analysed using graph theory (GT), revealing the underlying connectivity patterns of the nanofibres. Analysis of GT descriptors showed that the electrospun membranes closely mimic the connectivity patterns of the ECM. We conclude that ML-facilitated, GT-quantified engineering of cellular scaffolds has the potential to predict cell interactions, streamlining the pipeline for tissue engineering.
Collapse
Affiliation(s)
- Lakshmi Yaneesha Sujeeun
- Biomaterials Drug Delivery and Nanotechnology Unit, Centre for Biomedical and Biomaterials Research (CBBR), University of Mauritius, Réduit, Mauritius.
- Center for Complex Particle Systems (COMPASS), University of Michigan, Ann Arbor, Michigan 48109, USA
| | - Itisha Chummun Phul
- Biomaterials Drug Delivery and Nanotechnology Unit, Centre for Biomedical and Biomaterials Research (CBBR), University of Mauritius, Réduit, Mauritius.
| | - Nowsheen Goonoo
- Biomaterials Drug Delivery and Nanotechnology Unit, Centre for Biomedical and Biomaterials Research (CBBR), University of Mauritius, Réduit, Mauritius.
| | - Nicholas A Kotov
- Center for Complex Particle Systems (COMPASS), University of Michigan, Ann Arbor, Michigan 48109, USA
| | - Archana Bhaw-Luximon
- Biomaterials Drug Delivery and Nanotechnology Unit, Centre for Biomedical and Biomaterials Research (CBBR), University of Mauritius, Réduit, Mauritius.
| |
Collapse
|
8
|
Sugawara K, Takaya E, Inamori R, Konaka Y, Sato J, Shiratori Y, Hario F, Kobayashi T, Ueda T, Okamoto Y. Breast cancer classification based on breast tissue structures using the Jigsaw puzzle task in self-supervised learning. Radiol Phys Technol 2025; 18:209-218. [PMID: 39760975 PMCID: PMC11876229 DOI: 10.1007/s12194-024-00874-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2024] [Revised: 12/18/2024] [Accepted: 12/20/2024] [Indexed: 01/07/2025]
Abstract
Self-supervised learning (SSL) has gained attention in the medical field as a deep learning approach utilizing unlabeled data. The Jigsaw puzzle task in SSL enables models to learn both features of images and the positional relationships within images. In breast cancer diagnosis, radiologists evaluate not only lesion-specific features but also the surrounding breast structures. However, deep learning models that adopt a diagnostic approach similar to human radiologists are still limited. This study aims to evaluate the effectiveness of the Jigsaw puzzle task in characterizing breast tissue structures for breast cancer classification on mammographic images. Using the Chinese Mammography Database (CMMD), we compared four pre-training pipelines: (1) IN-Jig, pre-trained with both the ImageNet classification task and the Jigsaw puzzle task, (2) Scratch-Jig, pre-trained only with the Jigsaw puzzle task, (3) IN, pre-trained only with the ImageNet classification task, and (4) Scratch, that is trained from random initialization without any pre-training tasks. All pipelines were fine-tuned using binary classification to distinguish between the presence or absence of breast cancer. Performance was evaluated based on the area under the receiver operating characteristic curve (AUC), sensitivity, and specificity. Additionally, detailed analysis was conducted for performance across different radiological findings, breast density, and regions of interest were visualized using gradient-weighted class activation mapping (Grad-CAM). The AUC for the four models were 0.925, 0.921, 0.918, 0.909, respectively. Our results suggest the Jigsaw puzzle task is an effective pre-training method for breast cancer classification, with the potential to enhance diagnostic accuracy with limited data.
Collapse
Affiliation(s)
- Keisuke Sugawara
- Department of Diagnostic Radiology, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Eichi Takaya
- Department of Diagnostic Imaging, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan.
- AI Lab, Tohoku University Hospital, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan.
| | - Ryusei Inamori
- Department of Radiological Imaging and Informatics, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Yuma Konaka
- Department of Diagnostic Radiology, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Jumpei Sato
- Department of Diagnostic Radiology, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Yuta Shiratori
- Department of Diagnostic Imaging, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Fumihito Hario
- Department of Diagnostic Imaging, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Tomoya Kobayashi
- Department of Diagnostic Imaging, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
- AI Lab, Tohoku University Hospital, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Takuya Ueda
- Department of Diagnostic Radiology, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| | - Yoshikazu Okamoto
- Department of Diagnostic Imaging, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
- AI Lab, Tohoku University Hospital, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi, 980-8575, Japan
| |
Collapse
|
9
|
Chang Q, Wang S, Wang F, Gong B, Wang Y, Zuo F, Xie X, Bai Y. Development of a diagnostic classification model for lateral cephalograms based on multitask learning. BMC Oral Health 2025; 25:246. [PMID: 39955570 PMCID: PMC11830185 DOI: 10.1186/s12903-025-05588-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2024] [Accepted: 02/03/2025] [Indexed: 02/17/2025] Open
Abstract
OBJECTIVES This study aimed to develop a cephalometric classification method based on multitask learning for eight diagnostic classifications. METHODS This study was retrospective. A total of 3,310 lateral cephalograms were collected to construct a dataset. Eight clinical classifications were employed, including sagittal and vertical skeletal facial patterns, maxillary and mandibular anteroposterior positions, inclinations of upper and lower incisors, as well as their anteroposterior positions. The images were manually annotated for initially classification, which was verified by senior orthodontists. The data were randomly divided into training, validation, and test sets at a ratio of approximately 8:1:1. The multitask learning classification model was constructed based on the ResNeXt50_32 × 4d network and consisted of shared layers and task-specific layers. The performance of the model was evaluated using classification accuracy, precision, sensitivity, specificity and area under the curve (AUC). RESULTS This model could perform eight clinical diagnostic classifications on cephalograms within an average of 0.0096 s. The accuracy of the six classifications was 0.8-0.9, and the accuracy of the two classifications was 0.75-0.8. The overall AUC values for each classification exceeded 0.9. CONCLUSIONS An automatic diagnostic classification model for lateral cephalograms was established based on multitask learning to achieve simultaneous classification of eight common clinical diagnostic items. The multitask learning model achieved better classification performance and reduced the computational costs, providing a novel perspective and reference for addressing such problems.
Collapse
Affiliation(s)
- Qiao Chang
- Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing, China
| | - Shaofeng Wang
- Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing, China
| | - Fan Wang
- Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing, China
| | - Beiwen Gong
- Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing, China
| | - Yajie Wang
- Department of Engineering Physics, Tsinghua University, Beijing, China
- LargeV Instrument Corporation Limited, Beijing, China
| | - Feifei Zuo
- LargeV Instrument Corporation Limited, Beijing, China
| | - Xianju Xie
- Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing, China.
| | - Yuxing Bai
- Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing, China.
| |
Collapse
|
10
|
Gao R, Peng A, Duan Y, Chen M, Zheng T, Zhang M, Chen L, Sun H. Associations of Postencephalitic Epilepsy Using Multi-Contrast Whole Brain MRI: A Large Self-Supervised Vision Foundation Model Strategy. J Magn Reson Imaging 2025. [PMID: 39898495 DOI: 10.1002/jmri.29734] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2024] [Revised: 01/22/2025] [Accepted: 01/23/2025] [Indexed: 02/04/2025] Open
Abstract
BACKGROUND Postencephalitic epilepsy (PEE) is a severe neurological complication following encephalitis. Early identification of individuals at high risk for PEE is important for timely intervention. PURPOSE To develop a large self-supervised vision foundation model using a big dataset of multi-contrast head MRI scans, followed by fine-tuning with MRI data and follow-up outcomes from patients with PEE to develop a PEE association model. STUDY TYPE Retrospective. POPULATION Fifty-seven thousand six hundred twenty-one contrast-enhanced head MRI scans from 34,871 patients for foundation model construction, and head MRI scans from 144 patients with encephalitis (64 PEE, 80 N-PEE) for the PEE association model. FIELD STRENGTH/SEQUENCE 1.5-T, 3-T, T1-weighted imaging, T2-weighted imaging, fluid attenuated inversion recovery, T1-weighted contrast-enhanced imaging. ASSESSMENT The foundation model was developed using self-supervised learning and cross-contrast context recovery. Patients with encephalitis were monitored for a median of 3.7 years (range 0.7-7.5 years), with epilepsy diagnosed according to International League Against Epilepsy. Occlusion sensitivity mapping highlighted brain regions involved in PEE classifications. Model performance was compared with DenseNet without pre-trained weights. STATISTICAL TESTS Performance was assessed via confusion matrices, accuracy, sensitivity, specificity, precision, F1 score, and area under the receiver operating characteristic curve (AUC). The DeLong test evaluated AUC between the two models (P < 0.05 for statistical significance). RESULTS The PEE association model achieved accuracy, sensitivity, specificity, precision, F1 score, and AUC of 79.3% (95% CI: 0.71-0.92), 92.3% (95% CI: 0.80-1.00), 68.8% (95% CI: 0.55-0.87), 70.6% (95% CI: 0.61-0.90), 80.0% (95% CI: 0.71-0.93), and 81.0% (95% CI: 0.68-0.92), respectively. A significant AUC improvement was found compared to DenseNet (Delong test, P = 0.03). The association model focused on brain regions affected by encephalitis. DATA CONCLUSION Using extensive unlabeled data via self-supervised learning addressed the limitations of supervised tasks with limited data. The fine-tuned foundation model outperformed DenseNet, which was trained exclusively on task data. PLAIN LANGUAGE SUMMARY This research develops a model to assess the occurrence epilepsy after encephalitis, a severe brain inflammation condition. By using over 57,000 brain scans, the study trains a computer program to recognize patterns in brain images. The model analyzes whole-brain scans to identify areas commonly affected by the disease, such as the temporal and frontal lobes. It was tested on data from patients with encephalitis and showed better performance than older methods. The model can assess the risk of secondary epilepsy in patients with encephalitis, allowing doctors to intervene early and improve treatment outcomes for those affected by this condition. EVIDENCE LEVEL 4 TECHNICAL EFFICACY: Stage 1.
Collapse
Affiliation(s)
- Ronghui Gao
- Department of Radiology, West China Hospital, Sichuan University, Chengdu, China
| | - Anjiao Peng
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
| | - Yifei Duan
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
| | - Mengyao Chen
- Huaxi MR Research Center (HMRRC), West China Hospital, Sichuan University, Chengdu, China
| | - Tao Zheng
- IT Center, West China Hospital, Sichuan University, Chengdu, China
| | - Meng Zhang
- NVIDIA Corp, Beijing Representative Office, Beijing, China
| | - Lei Chen
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
- Pazhou Lab, Guangzhou, China
| | - Huaiqiang Sun
- Department of Radiology, West China Hospital, Sichuan University, Chengdu, China
- Huaxi MR Research Center (HMRRC), West China Hospital, Sichuan University, Chengdu, China
| |
Collapse
|
11
|
Zhu Y, Wang J, Xue C, Zhai X, Xiao C, Lu T. Deep Learning and Habitat Radiomics for the Prediction of Glioma Pathology Using Multiparametric MRI: A Multicenter Study. Acad Radiol 2025; 32:963-975. [PMID: 39322536 DOI: 10.1016/j.acra.2024.09.021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Revised: 09/05/2024] [Accepted: 09/06/2024] [Indexed: 09/27/2024]
Abstract
RATIONALE AND OBJECTIVES Recent radiomics studies on predicting pathological outcomes of glioma have shown immense potential. However, the predictive ability remains suboptimal due to the tumor intrinsic heterogeneity. We aimed to achieve better pathological prediction outcomes by combining habitat analysis with deep learning. MATERIALS AND METHODS 387 cases of primary glioma from three hospitals were collected, along with their T1 contrast-enhanced and T2-weighted MR sequences, pathological reports and clinical histories. The training set consisted of 264 patients, 82 patients composed the test set, and 41 patients were used as the validation set for hyperparameter tuning and optimal model selection. All groups were sourced from different centers. Through radiomics, deep learning, habitat analysis and combined analysis, we extracted imaging features separately and jointly modeled them with clinical features. We identified the optimal models for predicting glioma grades, Ki67 expression levels, P53 mutation and IDH1 mutation. RESULTS Using a LightGBM model with DenseNet161 features based on habitat subregions, the best tumor grade prediction model was achieved. A LightGBM model with ResNet50 features based on habitat subregions yielded the best Ki67 expression level prediction model. An SVM model with Radiomics and Inception_v3 features provided the best prediction of P53 mutation. The best model for predicting IDH1 mutation was achieved by an MLP model with Radiomics features based on habitat subregions. Clinical features might be potentially helpful for the prediction with relatively weak evidence. CONCLUSION Habitat+Deep Learning feature extraction methods were optimal for predicting grades and Ki67 levels. Deep Learning is optimal for predicting P53 mutation, while the combination of Habitat+ Radiomics models yielded the best prediction for IDH1 mutation.
Collapse
Affiliation(s)
- Yunyang Zhu
- Department of Neurosurgery, The First Affiliated Hospital of Soochow University, Suzhou, China (Y.Z., J.W., T.L.)
| | - Jing Wang
- Department of Neurosurgery, The First Affiliated Hospital of Soochow University, Suzhou, China (Y.Z., J.W., T.L.)
| | - Chen Xue
- Department of Radiology, the Affiliated Brain Hospital of Nanjing Medical University, Nanjing, China (C.X., C.X.)
| | - Xiaoyang Zhai
- The First Affiliated Hospital of Xinxiang University, Xinxiang, China (X.Z.)
| | - Chaoyong Xiao
- Department of Radiology, the Affiliated Brain Hospital of Nanjing Medical University, Nanjing, China (C.X., C.X.)
| | - Ting Lu
- Department of Neurosurgery, The First Affiliated Hospital of Soochow University, Suzhou, China (Y.Z., J.W., T.L.).
| |
Collapse
|
12
|
Okazaki S, Mine Y, Yoshimi Y, Iwamoto Y, Ito S, Peng TY, Nishimura T, Suehiro T, Koizumi Y, Nomura R, Tanimoto K, Kakimoto N, Murayama T. RadImageNet and ImageNet as Datasets for Transfer Learning in the Assessment of Dental Radiographs: A Comparative Study. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:534-544. [PMID: 39048809 PMCID: PMC11811346 DOI: 10.1007/s10278-024-01204-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/29/2024] [Revised: 07/14/2024] [Accepted: 07/15/2024] [Indexed: 07/27/2024]
Abstract
Transfer learning (TL) is an alternative approach to the full training of deep learning (DL) models from scratch and can transfer knowledge gained from large-scale data to solve different problems. ImageNet, which is a publicly available large-scale dataset, is a commonly used dataset for TL-based image analysis; many studies have applied pre-trained models from ImageNet to clinical prediction tasks and have reported promising results. However, some have questioned the effectiveness of using ImageNet, which consists solely of natural images, for medical image analysis. The aim of this study was to evaluate whether pre-trained models using RadImageNet, which is a large-scale medical image dataset, could achieve superior performance in classification tasks in dental imaging modalities compared with ImageNet pre-trained models. To evaluate the classification performance of RadImageNet and ImageNet pre-trained models for TL, two dental imaging datasets were used. The tasks were (1) classifying the presence or absence of supernumerary teeth from a dataset of panoramic radiographs and (2) classifying sex from a dataset of lateral cephalometric radiographs. Performance was evaluated by comparing the area under the curve (AUC). On the panoramic radiograph dataset, the RadImageNet models gave average AUCs of 0.68 ± 0.15 (p < 0.01), and the ImageNet models had values of 0.74 ± 0.19. In contrast, on the lateral cephalometric dataset, the RadImageNet models demonstrated average AUCs of 0.76 ± 0.09, and the ImageNet models achieved values of 0.75 ± 0.17. The difference in performance between RadImageNet and ImageNet models in TL depends on the dental image dataset used.
Collapse
Affiliation(s)
- Shota Okazaki
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
- Project Research Center for Integrating Digital Dentistry, Hiroshima University, Hiroshima, Japan
| | - Yuichi Mine
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan.
- Project Research Center for Integrating Digital Dentistry, Hiroshima University, Hiroshima, Japan.
| | - Yuki Yoshimi
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Yuko Iwamoto
- Department of Pediatric Dentistry, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Shota Ito
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Tzu-Yu Peng
- School of Dentistry, College of Oral Medicine, Taipei Medical University, Taipei, Taiwan
| | - Taku Nishimura
- Department of Pediatric Dentistry, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Tomoya Suehiro
- Department of Genomic Oncology and Oral Medicine, Graduate School of Biomedical and Health Science, Hiroshima University, Hiroshima, Japan
| | - Yuma Koizumi
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Ryota Nomura
- Department of Pediatric Dentistry, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Kotaro Tanimoto
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Naoya Kakimoto
- Department of Oral and Maxillofacial Radiology, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan
| | - Takeshi Murayama
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University, Hiroshima, Japan.
- Project Research Center for Integrating Digital Dentistry, Hiroshima University, Hiroshima, Japan.
| |
Collapse
|
13
|
Meijerink LM, Dunias ZS, Leeuwenberg AM, de Hond AAH, Jenkins DA, Martin GP, Sperrin M, Peek N, Spijker R, Hooft L, Moons KGM, van Smeden M, Schuit E. Updating methods for artificial intelligence-based clinical prediction models: a scoping review. J Clin Epidemiol 2025; 178:111636. [PMID: 39662644 DOI: 10.1016/j.jclinepi.2024.111636] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2024] [Revised: 12/02/2024] [Accepted: 12/03/2024] [Indexed: 12/13/2024]
Abstract
OBJECTIVES To give an overview of methods for updating artificial intelligence (AI)-based clinical prediction models based on new data. STUDY DESIGN AND SETTING We comprehensively searched Scopus and Embase up to August 2022 for articles that addressed developments, descriptions, or evaluations of prediction model updating methods. We specifically focused on articles in the medical domain involving AI-based prediction models that were updated based on new data, excluding regression-based updating methods as these have been extensively discussed elsewhere. We categorized and described the identified methods used to update the AI-based prediction model as well as the use cases in which they were used. RESULTS We included 78 articles. The majority of the included articles discussed updating for neural network methods (93.6%) with medical images as input data (65.4%). In many articles (51.3%) existing, pretrained models for broad tasks were updated to perform specialized clinical tasks. Other common reasons for model updating were to address changes in the data over time and cross-center differences; however, more unique use cases were also identified, such as updating a model from a broad population to a specific individual. We categorized the identified model updating methods into four categories: neural network-specific methods (described in 92.3% of the articles), ensemble-specific methods (2.5%), model-agnostic methods (9.0%), and other (1.3%). Variations of neural network-specific methods are further categorized based on the following: (1) the part of the original neural network that is kept, (2) whether and how the original neural network is extended with new parameters, and (3) to what extent the original neural network parameters are adjusted to the new data. The most frequently occurring method (n = 30) involved selecting the first layer(s) of an existing neural network, appending new, randomly initialized layers, and then optimizing the entire neural network. CONCLUSION We identified many ways to adjust or update AI-based prediction models based on new data, within a large variety of use cases. Updating methods for AI-based prediction models other than neural networks (eg, random forest) appear to be underexplored in clinical prediction research. PLAIN LANGUAGE SUMMARY AI-based prediction models are increasingly used in health care, helping clinicians with diagnosing diseases, guiding treatment decisions, and informing patients. However, these prediction models do not always work well when applied to hospitals, patient populations, or times different from those used to develop the models. Developing new models for every situation is neither practical nor desired, as it wastes resources, time, and existing knowledge. A more efficient approach is to adjust existing models to new contexts ('updating'), but there is limited guidance on how to do this for AI-based clinical prediction models. To address this, we reviewed 78 studies in detail to understand how researchers are currently updating AI-based clinical prediction models, and the types of situations in which these updating methods are used. Our findings provide a comprehensive overview of the available methods to update existing models. This is intended to serve as guidance and inspiration for researchers. Ultimately, this can lead to better reuse of existing models and improve the quality and efficiency of AI-based prediction models in health care.
Collapse
Affiliation(s)
- Lotta M Meijerink
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands.
| | - Zoë S Dunias
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - Artuur M Leeuwenberg
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - Anne A H de Hond
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - David A Jenkins
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Manchester, United Kingdom
| | - Glen P Martin
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Manchester, United Kingdom
| | - Matthew Sperrin
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Manchester, United Kingdom
| | - Niels Peek
- Department of Public Health and Primary Care, The Healthcare Improvement Studies Institute, University of Cambridge, Cambridge, United Kingdom
| | - René Spijker
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - Lotty Hooft
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - Karel G M Moons
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - Maarten van Smeden
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| | - Ewoud Schuit
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands
| |
Collapse
|
14
|
Fang X, Chong CF, Wong KL, Simões M, Ng BK. Investigating the key principles in two-step heterogeneous transfer learning for early laryngeal cancer identification. Sci Rep 2025; 15:2146. [PMID: 39820368 PMCID: PMC11739633 DOI: 10.1038/s41598-024-84836-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2024] [Accepted: 12/27/2024] [Indexed: 01/19/2025] Open
Abstract
Data scarcity in medical images makes transfer learning a common approach in computer-aided diagnosis. Some disease classification tasks can rely on large homogeneous public datasets to train the transferred model, while others cannot, i.e., endoscopic laryngeal cancer image identification. Distinguished from most current works, this work pioneers exploring a two-step heterogeneous transfer learning (THTL) framework for laryngeal cancer identification and summarizing the fundamental principles for the intermediate domain selection. For heterogeneity and clear vascular representation, diabetic retinopathy images were chosen as THTL's intermediate domain. The experiment results reveal two vital principles in intermediate domain selection for future studies: 1) the size of the intermediate domain is not a sufficient condition to improve the transfer learning performance; 2) even distinct vascular features in the intermediate domain do not guarantee improved performance in the target domain. We observe that radial vascular patterns benefit benign classification, whereas twisted and tangled patterns align more with malignant classification. Additionally, to compensate for the absence of twisted patterns in the intermediate domains, we propose the Step-Wise Fine-Tuning (SWFT) technique, guided by the Layer Class Activate Map (LayerCAM) visualization result, getting 20.4% accuracy increases compared to accuracy from THTL's, even higher than fine-tune all layers.
Collapse
Affiliation(s)
- Xinyi Fang
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China
- Department of Informatics Engineering, Centre for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, 3000, Portugal
| | - Chak Fong Chong
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China
- Department of Informatics Engineering, Centre for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, 3000, Portugal
| | - Kei Long Wong
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China
- Department of Computer Science and Engineering, University of Bologna, Bologna, 40100, Italy
| | - Marco Simões
- Department of Informatics Engineering, Centre for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, 3000, Portugal
| | - Benjamin K Ng
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China.
| |
Collapse
|
15
|
Vach M, Weiss D, Ivan VL, Boschenriedter C, Wolf L, Beez T, Hofmann BB, Rubbert C, Caspers J. Deep-Learning-based Automated Identification of Ventriculoperitoneal-Shunt Valve Models from Skull X-rays. Clin Neuroradiol 2025:10.1007/s00062-024-01490-4. [PMID: 39809871 DOI: 10.1007/s00062-024-01490-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2024] [Accepted: 12/11/2024] [Indexed: 01/16/2025]
Abstract
INTRODUCTION Ventriculoperitoneal shunts (VPS) are an essential part of the treatment of hydrocephalus, with numerous valve models available with different ways of indicating pressure levels. The model types often need to be identified on X‑rays to assess pressure levels using a matching template. Artificial intelligence (AI), in particular deep learning, is ideally suited to automate repetitive tasks such as identifying different VPS valve models. The aim of this work was to investigate whether AI, in particular deep learning, allows the identification of VPS models in cranial X‑rays. METHODS 959 cranial X‑rays of patients with a VPS were included and reviewed for image quality and complete visualization of VPS valves. The images included four VPS model types: Codman Hakim (n = 774, 81%), Codman Certas Plus (n = 117, 12%), Sophysa Sophy Mini SM8 (n = 35, 4%) and proGAV 2.0 (n = 33, 3%). A Convolutional Neural Network (CNN) was trained using stratified five-fold cross-validation to classify the four VPS model types in the dataset. A finetuned CNN pretrained on the ImageNet dataset as well as a model trained from scratch were compared. The averaged performance and uncertainty metrics were evaluated across the cross-validation splits. RESULTS The fine-tuned model identified VPS valve models with a mean accuracy of 0.98 ± 0.01, macro-averaged F1 score of 0.93 ± 0.04, a recall of 0.94 ± 0.03 and a precision of 0.95 ± 0.08 across the five cross-validation splits. CONCLUSION Automatic classification of VPS valve models in skull X‑rays, using fully automatable preprocessing steps and a CNN, is feasible. This is an encouraging finding to further explore the possibility of automating VPS valve model identification and pressure level reading in skull X‑rays.
Collapse
Affiliation(s)
- Marius Vach
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany.
| | - Daniel Weiss
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany
| | - Vivien Lorena Ivan
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany
| | - Christian Boschenriedter
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany
| | - Luisa Wolf
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany
| | - Thomas Beez
- Department of Neurosurgery, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Düsseldorf, Germany
| | - Björn B Hofmann
- Department of Neurosurgery, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Düsseldorf, Germany
| | - Christian Rubbert
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany
| | - Julian Caspers
- Department of Diagnostic and Interventional Radiology, Medical Faculty and University Hospital Düsseldorf, Heinrich-Heine-University Düsseldorf, Moorenstraße 5, 40225, Düsseldorf, Germany
| |
Collapse
|
16
|
Tandel GS, Tiwari A, Kakde OG. Multi-Class Brain Tumor Grades Classification Using a Deep Learning-Based Majority Voting Algorithm and Its Validation Using Explainable-AI. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-024-01368-4. [PMID: 39779641 DOI: 10.1007/s10278-024-01368-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/05/2024] [Revised: 11/14/2024] [Accepted: 12/02/2024] [Indexed: 01/11/2025]
Abstract
Biopsy is considered the gold standard for diagnosing brain tumors, but its invasive nature can pose risks to patients. Additionally, tissue analysis can be cumbersome and inconsistent among observers. This research aims to develop a cost-effective, non-invasive, MRI-based computer-aided diagnosis tool that can reliably, accurately and swiftly identify brain tumor grades. Our system employs ensemble deep learning (EDL) within an MRI multiclass framework that includes five datasets: two-class (C2), three-class (C3), four-class (C4), five-class (C5) and six-class (C6). The EDL utilizes a majority voting algorithm to classify brain tumors by combining seven renowned deep learning (DL) models-EfficientNet, VGG16, ResNet18, GoogleNet, ResNet50, Inception-V3 and DarkNet-and seven machine learning (ML) models, including support vector machine, K-nearest neighbour, Naïve Bayes, decision tree, linear discriminant analysis, artificial neural network and random forest. Additionally, local interpretable model-agnostic explanations (LIME) are employed as an explainable AI algorithm, providing a visual representation of the CNN's internal workings to enhance the credibility of the results. Through extensive five-fold cross-validation experiments, the DL-based majority voting algorithm outperformed the ML-based majority voting algorithm, achieving the highest average accuracies of 100 ± 0.00%, 98.55 ± 0.35%, 98.47 ± 0.63%, 95.34 ± 1.17% and 96.61 ± 0.85% for the C2, C3, C4, C5 and C6 datasets, respectively. Majority voting algorithms typically yield consistent results across different folds of the brain tumor data and enhance performance compared to any individual deep learning and machine learning models.
Collapse
Affiliation(s)
- Gopal Singh Tandel
- Department of Computer Science, Allahabad Degree College, University of Allahabad, Prayagraj, India.
| | - Ashish Tiwari
- Department of Computer Science and Engineering, Visvesvaraya National Institute of Technology, Nagpur, India
| | - Omprakash G Kakde
- Department of Computer Science and Engineering, Visvesvaraya National Institute of Technology, Nagpur, India
| |
Collapse
|
17
|
Chen S, Huang J, Huang J, Wang P, Sun C, Zhang Z, Jiang S. Explainable deep learning identifies patterns and drivers of freshwater harmful algal blooms. ENVIRONMENTAL SCIENCE AND ECOTECHNOLOGY 2025; 23:100522. [PMID: 39897111 PMCID: PMC11786749 DOI: 10.1016/j.ese.2024.100522] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 12/25/2024] [Accepted: 12/26/2024] [Indexed: 02/04/2025]
Abstract
The escalating magnitude, frequency, and duration of harmful algal blooms (HABs) pose significant challenges to freshwater ecosystems worldwide. However, the mechanisms driving HABs remain poorly understood, in part due to the strong regional specificity of algal processes and the uneven data availability. These complexities make it difficult to generalize HAB dynamics and effectively predict their occurrence using traditional models. To address these challenges, we developed an explainable deep learning approach using long short-term memory (LSTM) models combined with explanation techniques that can capture complex patterns and provide explainable insights into key HAB drivers. We applied this approach for algal density modeling at 102 sites in China's lakes and reservoirs over three years. LSTMs effectively captured daily algal dynamics, achieving mean and maximum Nash-Sutcliffe efficiency coefficients of 0.48 and 0.95 during testing phase. Moreover, water temperature emerged as the primary driver of HABs both nationally and in over 30% of localities, with stronger water temperature sensitivity observed in mid-to low-latitudes. We also identified regional similarities that allow for the successful transferability in modeling algal dynamics. Specifically, using fine-tuned transfer learning, we improved the prediction accuracy in over 75% of poorly gauged areas. Overall, LSTM-based explainable deep learning approach effectively addresses key challenges in HAB modeling by tackling both regional specificity and data limitations. By accurately predicting algal dynamics and identifying critical drivers, this approach provides actionable insights into the mechanisms of HABs, ultimately aids in the implementation of effective mitigation measures for nationwide and regional freshwater ecosystems.
Collapse
Affiliation(s)
- Shengyue Chen
- Fujian Key Laboratory of Coastal Pollution Prevention and Control, Xiamen University, Xiamen, 361102, China
- Department of Biogeochemical Integration, Max Planck Institute for Biogeochemistry, Jena, 07745, Germany
| | - Jinliang Huang
- Fujian Key Laboratory of Coastal Pollution Prevention and Control, Xiamen University, Xiamen, 361102, China
| | - Jiacong Huang
- Key Laboratory of Watershed Geographic Sciences, Nanjing Institute of Geography and Limnology, Chinese Academy of Sciences, 73 East Beijing Road, Nanjing, 210008, China
| | - Peng Wang
- Fujian Key Laboratory of Coastal Pollution Prevention and Control, Xiamen University, Xiamen, 361102, China
| | - Changyang Sun
- Fujian Key Laboratory of Coastal Pollution Prevention and Control, Xiamen University, Xiamen, 361102, China
| | - Zhenyu Zhang
- Fujian Key Laboratory of Coastal Pollution Prevention and Control, Xiamen University, Xiamen, 361102, China
- School of Geographical Sciences, Fujian Normal University, Fuzhou, 350007, China
| | - Shijie Jiang
- Department of Biogeochemical Integration, Max Planck Institute for Biogeochemistry, Jena, 07745, Germany
- ELLIS Unit Jena, Jena, 07745, Germany
| |
Collapse
|
18
|
Martins C, Barros H, Moreira A. Transfer learning in spirometry: CNN models for automated flow-volume curve quality control in paediatric populations. Comput Biol Med 2025; 184:109341. [PMID: 39541898 DOI: 10.1016/j.compbiomed.2024.109341] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2024] [Revised: 10/11/2024] [Accepted: 10/30/2024] [Indexed: 11/17/2024]
Abstract
PROBLEM Current spirometers face challenges in evaluating acceptability criteria, often requiring manual visual inspection by trained specialists. Automating this process could improve diagnostic workflows and reduce variability in test assessments. AIM This study aimed to apply transfer learning to convolutional neural networks (CNNs) to automate the classification of spirometry flow-volume curves based on acceptability criteria. METHODS A total of 5287 spirometry flow-volume curves were divided into three categories: (A) all criteria met, (B) early termination, and (C) non-acceptable results. Six CNN models (VGG16, InceptionV3, Xception, ResNet152V2, InceptionResNetV2, DenseNet121) were trained using a balanced dataset after data augmentation. The models' performance was evaluated on part of the original unbalanced dataset with accuracy, precision, recall, and F1-score metrics. RESULTS VGG16 achieved the highest accuracy at 93.9 %, while ResNet152V2 had the lowest at 83.0 %. Non-acceptable curves (Group C) were the easiest to classify, with precision reaching at least 87.7 %. Early termination curves (Group B) were the most challenging, with precision ranging from 75.0 % to 90.3 %. CONCLUSION CNN models, particularly VGG16, show promise in automating spirometry quality control, potentially reducing the need for manual inspection by specialized technicians. This approach can streamline spirometry assessments, offering consistent, high-quality diagnostics even in non-specialized or low-resource environments.
Collapse
Affiliation(s)
- Carla Martins
- EPIUnit - Instituto de Saúde Pública da Universidade do Porto, Laboratório para a Investigação Integrativa e Translacional em Saúde Populacional (ITR), Porto, Portugal; Serviço de Imunoalergologia, Unidade Local de Saúde de São João, Porto, Portugal.
| | - Henrique Barros
- EPIUnit - Instituto de Saúde Pública da Universidade do Porto, Laboratório para a Investigação Integrativa e Translacional em Saúde Populacional (ITR), Porto, Portugal
| | - André Moreira
- EPIUnit - Instituto de Saúde Pública da Universidade do Porto, Laboratório para a Investigação Integrativa e Translacional em Saúde Populacional (ITR), Porto, Portugal; Serviço de Imunoalergologia, Unidade Local de Saúde de São João, Porto, Portugal; Laboratório de Imunologia Básica e Clínica, Departamento de Patologia, Faculdade de Medicina da Universidade do Porto, Portugal
| |
Collapse
|
19
|
Lan Z, Chen R, Zou D, Zhao C. Microfluidic Nanoparticle Separation for Precision Medicine. ADVANCED SCIENCE (WEINHEIM, BADEN-WURTTEMBERG, GERMANY) 2025; 12:e2411278. [PMID: 39632600 PMCID: PMC11775552 DOI: 10.1002/advs.202411278] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/14/2024] [Revised: 11/11/2024] [Indexed: 12/07/2024]
Abstract
A deeper understanding of disease heterogeneity highlights the urgent need for precision medicine. Microfluidics, with its unique advantages, such as high adjustability, diverse material selection, low cost, high processing efficiency, and minimal sample requirements, presents an ideal platform for precision medicine applications. As nanoparticles, both of biological origin and for therapeutic purposes, become increasingly important in precision medicine, microfluidic nanoparticle separation proves particularly advantageous for handling valuable samples in personalized medicine. This technology not only enhances detection, diagnosis, monitoring, and treatment accuracy, but also reduces invasiveness in medical procedures. This review summarizes the fundamentals of microfluidic nanoparticle separation techniques for precision medicine, starting with an examination of nanoparticle properties essential for separation and the core principles that guide various microfluidic methods. It then explores passive, active, and hybrid separation techniques, detailing their principles, structures, and applications. Furthermore, the review highlights their contributions to advancements in liquid biopsy and nanomedicine. Finally, it addresses existing challenges and envisions future development spurred by emerging technologies such as advanced materials science, 3D printing, and artificial intelligence. These interdisciplinary collaborations are anticipated to propel the platformization of microfluidic separation techniques, significantly expanding their potential in precision medicine.
Collapse
Affiliation(s)
- Zhenwei Lan
- School of Chemical Engineering, Faculty of Sciences, Engineering and TechnologyThe University of AdelaideAdelaideSA5005Australia
| | - Rui Chen
- School of Chemical Engineering, Faculty of Sciences, Engineering and TechnologyThe University of AdelaideAdelaideSA5005Australia
| | - Da Zou
- School of Chemical Engineering, Faculty of Sciences, Engineering and TechnologyThe University of AdelaideAdelaideSA5005Australia
| | - Chun‐Xia Zhao
- School of Chemical Engineering, Faculty of Sciences, Engineering and TechnologyThe University of AdelaideAdelaideSA5005Australia
| |
Collapse
|
20
|
Stathopoulos I, Serio L, Karavasilis E, Kouri MA, Velonakis G, Kelekis N, Efstathopoulos E. Evaluating Brain Tumor Detection with Deep Learning Convolutional Neural Networks Across Multiple MRI Modalities. J Imaging 2024; 10:296. [PMID: 39728193 DOI: 10.3390/jimaging10120296] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2024] [Revised: 11/18/2024] [Accepted: 11/20/2024] [Indexed: 12/28/2024] Open
Abstract
Central Nervous System (CNS) tumors represent a significant public health concern due to their high morbidity and mortality rates. Magnetic Resonance Imaging (MRI) has emerged as a critical non-invasive modality for the detection, diagnosis, and management of brain tumors, offering high-resolution visualization of anatomical structures. Recent advancements in deep learning, particularly convolutional neural networks (CNNs), have shown potential in augmenting MRI-based diagnostic accuracy for brain tumor detection. In this study, we evaluate the diagnostic performance of six fundamental MRI sequences in detecting tumor-involved brain slices using four distinct CNN architectures enhanced with transfer learning techniques. Our dataset comprises 1646 MRI slices from the examinations of 62 patients, encompassing both tumor-bearing and normal findings. With our approach, we achieved a classification accuracy of 98.6%, underscoring the high potential of CNN-based models in this context. Additionally, we assessed the performance of each MRI sequence across the different CNN models, identifying optimal combinations of MRI modalities and neural networks to meet radiologists' screening requirements effectively. This study offers critical insights into the integration of deep learning with MRI for brain tumor detection, with implications for improving diagnostic workflows in clinical settings.
Collapse
Affiliation(s)
- Ioannis Stathopoulos
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
- Technology Department, CERN, 1211 Geneva, Switzerland
| | - Luigi Serio
- Technology Department, CERN, 1211 Geneva, Switzerland
| | - Efstratios Karavasilis
- Medical Physics Laboratory, School of Medicine, Democritus University of Thrace, 68100 Alexandroupolis, Greece
| | - Maria Anthi Kouri
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| | - Georgios Velonakis
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| | - Nikolaos Kelekis
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| | - Efstathios Efstathopoulos
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| |
Collapse
|
21
|
Adnan N, Faizan Ahmed SM, Das JK, Aijaz S, Sukhia RH, Hoodbhoy Z, Umer F. Developing an AI-based application for caries index detection on intraoral photographs. Sci Rep 2024; 14:26752. [PMID: 39500993 PMCID: PMC11538444 DOI: 10.1038/s41598-024-78184-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2024] [Accepted: 10/29/2024] [Indexed: 11/08/2024] Open
Abstract
This study evaluates the effectiveness of an Artificial Intelligence (AI)-based smartphone application designed for decay detection on intraoral photographs, comparing its performance to that of junior dentists. Conducted at The Aga Khan University Hospital, Karachi, Pakistan, this study utilized a dataset comprising 7,465 intraoral images, including both primary and secondary dentitions. These images were meticulously annotated by two experienced dentists and further verified by senior dentists. A YOLOv5s model was trained on this dataset and integrated into a smartphone application, while a Detection Transformer was also fine-tuned for comparative purposes. Explainable AI techniques were employed to assess the AI's decision-making processes. A sample of 70 photographs was used to directly compare the application's performance with that of junior dentists. Results showed that the YOLOv5s-based smartphone application achieved a precision of 90.7%, sensitivity of 85.6%, and an F1 score of 88.0% in detecting dental decay. In contrast, junior dentists achieved 83.3% precision, 64.1% sensitivity, and an F1 score of 72.4%. The study concludes that the YOLOv5s algorithm effectively detects dental decay on intraoral photographs and performs comparably to junior dentists. This application holds potential for aiding in the evaluation of the caries index within populations, thus contributing to efforts aimed at reducing the disease burden at the community level.
Collapse
Affiliation(s)
- Niha Adnan
- Department of Surgery, Aga Khan University Hospital, Karachi, Pakistan
- MeDenTec, Karachi, Pakistan
| | | | | | - Sehrish Aijaz
- Dow University of Health Sciences, Karachi, Pakistan
| | | | | | - Fahad Umer
- Department of Surgery, Aga Khan University Hospital, Karachi, Pakistan.
- MeDenTec, Karachi, Pakistan.
| |
Collapse
|
22
|
Sandnes AT, Grimstad B, Kolbjørnsen O. Multi-task neural networks by learned contextual inputs. Neural Netw 2024; 179:106528. [PMID: 39024706 DOI: 10.1016/j.neunet.2024.106528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2023] [Revised: 05/29/2024] [Accepted: 07/07/2024] [Indexed: 07/20/2024]
Abstract
This paper explores learned-context neural networks. It is a multi-task learning architecture based on a fully shared neural network and an augmented input vector containing trainable task parameters. The architecture is interesting due to its powerful task adaption mechanism, which facilitates a low-dimensional task parameter space. Theoretically, we show that a scalar task parameter is sufficient for universal approximation of all tasks, which is not necessarily the case for more common architectures. Empirically it is shown that, for homogeneous tasks, the dimension of the task parameter may vary with the complexity of the tasks, but a small task parameter space is generally viable. The task parameter space is found to be well-behaved, which simplifies workflows related to updating models as new data arrives, and learning new tasks with the shared parameters are frozen. Additionally, the architecture displays robustness towards datasets where tasks have few data points. The architecture's performance is compared to similar neural network architectures on ten datasets, with competitive results.
Collapse
Affiliation(s)
- Anders T Sandnes
- Solution Seeker AS, Oslo, Norway; Department of Mathematics, University of Oslo, Oslo, Norway.
| | - Bjarne Grimstad
- Solution Seeker AS, Oslo, Norway; Department of Engineering Cybernetics, Norwegian University of Science and Technology, Trondheim, Norway
| | | |
Collapse
|
23
|
Wang L, Mao Y, Wang L, Sun Y, Song J, Zhang Y. Suitability of GPT-4o as an evaluator of cardiopulmonary resuscitation skills examinations. Resuscitation 2024; 204:110404. [PMID: 39343124 DOI: 10.1016/j.resuscitation.2024.110404] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2024] [Revised: 09/23/2024] [Accepted: 09/24/2024] [Indexed: 10/01/2024]
Abstract
AIM To assess the accuracy and reliability of GPT-4o for scoring examinees' performance on cardiopulmonary resuscitation (CPR) skills tests. METHODS This study included six experts certified to supervise the national medical licensing examination (three junior and three senior) who reviewed the CPR skills test videos across 103 examinees. All videos reviewed by the experts were subjected to automated assessment by GPT-4o. Both the experts and GPT-4o scored the videos across four sections: patient assessment, chest compressions, rescue breathing, and repeated operations. The experts subsequently rated GPT-4o's reliability on a 5-point Likert scale (1, completely unreliable; 5, completely reliable). GPT-4o's accuracy was evaluated using the intraclass correlation coefficient (for the first three sections) and Fleiss' Kappa (for the last section) to assess the agreement between its scores vs. those of the experts. RESULTS The mean accuracy scores for the patient assessment, chest compressions, rescue breathing, and repeated operation sections were 0.65, 0.58, 0.60, and 0.31, respectively, when comparing the GPT-4o's vs. junior experts' scores and 0.75, 0.65, 0.72, and 0.41, respectively, when comparing the GPT-4o's vs. senior experts' scores. For reliability, the median Likert scale scores were 4.00 (interquartile range [IQR] = 3.66-4.33, mean [standard deviation] = 3.95 [0.55]) and 4.33 (4.00-4.67, 4.29 [0.50]) for the junior and senior experts, respectively. CONCLUSIONS GPT-4o demonstrated a level of accuracy that was similar to that of senior experts in examining CPR skills examination videos. The results demonstrate the potential for deploying this large language model in medical examination settings.
Collapse
Affiliation(s)
- Lu Wang
- Shengjing Hospital of China Medical University, Shenyang, Liaoning 110004, China; School of Health Management, China Medical University, Shenyang, Liaoning 110122, China
| | - Yuqiang Mao
- Department of Thoracic Surgery, Shengjing Hospital of China Medical University, Shenyang, Liaoning 110004, China
| | - Lin Wang
- Department of Emergency Medicine, Shengjing Hospital of China Medical University, Shenyang, Liaoning 110022, China
| | - Yujie Sun
- Center for Clinical Skills Practice and Teaching, China Medical University, Shenyang, Liaoning 110122, China
| | - Jiangdian Song
- School of Health Management, China Medical University, Shenyang, Liaoning 110122, China.
| | - Yang Zhang
- Center for Clinical Skills Practice and Teaching, China Medical University, Shenyang, Liaoning 110122, China
| |
Collapse
|
24
|
Fotouhi M, Shahbandi A, Mehr FSK, Shahla MM, Nouredini SM, Kankam SB, Khorasanizadeh M, Chambless LB. Application of radiomics for diagnosis, subtyping, and prognostication of medulloblastomas: a systematic review. Neurosurg Rev 2024; 47:827. [PMID: 39467891 DOI: 10.1007/s10143-024-03060-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2024] [Revised: 08/20/2024] [Accepted: 10/13/2024] [Indexed: 10/30/2024]
Abstract
Applications of radiomics for clinical management of medulloblastomas are expanding. This systematic review aims to aggregate and overview the current role of radiomics in the diagnosis, subtyping, and prognostication of medulloblastomas. The present systematic review was performed in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. PubMed/MEDLINE were searched using a standardized search term. Articles found within the database from the inception until November 2022 were considered for screening. Retrieved records were screened independently by two authors based on their titles and abstracts. The full text of selected articles was reviewed to finalize the eligibility. Due to the heterogeneity of included studies, no formal data synthesis was conducted. Of the 249 screened citations, 21 studies were included and analyzed. Radiomics demonstrated promising performance for discriminating medulloblastomas from other posterior fossa tumors, particularly ependymomas and pilocytic astrocytomas. It was also efficacious in determining the subtype (i.e., WNT+, SHH+, group 3, and group 4) of medulloblastomas non-invasively. Regarding prognostication, radiomics exhibited some ability to predict overall survival and progression-free survival of patients with medulloblastomas. Our systematic review revealed that radiomics represents a promising tool for diagnosis and prognostication of medulloblastomas. Further prospective research measuring the clinical value of radiomics in this setting is warranted.
Collapse
Affiliation(s)
- Maryam Fotouhi
- Advanced Diagnostic and Interventional Radiology Research Center (ADIR), Cancer Institute, Imam Khomeini Hospital Complex, Tehran University of Medical Sciences, Tehran, Iran
| | | | | | | | | | - Samuel B Kankam
- Department of Neurosurgery, Brigham and Women's Hospital, Harvard Medical School, Cambridge, MA, USA
- T. H. Chan School of Public Health, Harvard University, Cambridge, MA, USA
| | | | - Lola B Chambless
- Department of Neurosurgery, Vanderbilt University Medical Center, Nashville, TN, USA
| |
Collapse
|
25
|
Jones MA, Zhang K, Faiz R, Islam W, Jo J, Zheng B, Qiu Y. Utilizing Pseudo Color Image to Improve the Performance of Deep Transfer Learning-Based Computer-Aided Diagnosis Schemes in Breast Mass Classification. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01237-0. [PMID: 39455542 DOI: 10.1007/s10278-024-01237-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Revised: 07/15/2024] [Accepted: 08/14/2024] [Indexed: 10/28/2024]
Abstract
The purpose of this study is to investigate the impact of using morphological information in classifying suspicious breast lesions. The widespread use of deep transfer learning can significantly improve the performance of the mammogram based CADx schemes. However, digital mammograms are grayscale images, while deep learning models are typically optimized using the natural images containing three channels. Thus, it is needed to convert the grayscale mammograms into three channel images for the input of deep transfer models. This study aims to develop a novel pseudo color image generation method which utilizes the mass contour information to enhance the classification performance. Accordingly, a total of 830 breast cancer cases were retrospectively collected, which contains 310 benign and 520 malignant cases, respectively. For each case, a total of four regions of interest (ROI) are collected from the grayscale images captured for both the CC and MLO views of the two breasts. Meanwhile, a total of seven pseudo color image sets are generated as the input of the deep learning models, which are created through a combination of the original grayscale image, a histogram equalized image, a bilaterally filtered image, and a segmented mass. Accordingly, the output features from four identical pre-trained deep learning models are concatenated and then processed by a support vector machine-based classifier to generate the final benign/malignant labels. The performance of each image set was evaluated and compared. The results demonstrate that the pseudo color sets containing the manually segmented mass performed significantly better than all other pseudo color sets, which achieved an AUC (area under the ROC curve) up to 0.889 ± 0.012 and an overall accuracy up to 0.816 ± 0.020, respectively. At the same time, the performance improvement is also dependent on the accuracy of the mass segmentation. The results of this study support our hypothesis that adding accurately segmented mass contours can provide complementary information, thereby enhancing the performance of the deep transfer model in classifying suspicious breast lesions.
Collapse
Affiliation(s)
- Meredith A Jones
- Stephenson School of Biomedical Engineering, University of Oklahoma, Norman, OK, 73019, USA
| | - Ke Zhang
- Stephenson School of Biomedical Engineering, University of Oklahoma, Norman, OK, 73019, USA
| | - Rowzat Faiz
- School of Electrical and Computer Engineering, University of Oklahoma, Norman, OK, 73019, USA
| | - Warid Islam
- School of Electrical and Computer Engineering, University of Oklahoma, Norman, OK, 73019, USA
| | - Javier Jo
- School of Electrical and Computer Engineering, University of Oklahoma, Norman, OK, 73019, USA
| | - Bin Zheng
- School of Electrical and Computer Engineering, University of Oklahoma, Norman, OK, 73019, USA
| | - Yuchen Qiu
- Stephenson School of Biomedical Engineering, University of Oklahoma, Norman, OK, 73019, USA.
- School of Electrical and Computer Engineering, University of Oklahoma, Norman, OK, 73019, USA.
| |
Collapse
|
26
|
Doganay MT, Chakraborty P, Bommakanti SM, Jammalamadaka S, Battalapalli D, Madabhushi A, Draz MS. Artificial intelligence performance in testing microfluidics for point-of-care. LAB ON A CHIP 2024; 24:4998-5008. [PMID: 39360887 PMCID: PMC11448392 DOI: 10.1039/d4lc00671b] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/13/2024] [Accepted: 09/16/2024] [Indexed: 10/06/2024]
Abstract
Artificial intelligence (AI) is revolutionizing medicine by automating tasks like image segmentation and pattern recognition. These AI approaches support seamless integration with existing platforms, enhancing diagnostics, treatment, and patient care. While recent advancements have demonstrated AI superiority in advancing microfluidics for point of care (POC) diagnostics, a gap remains in comparative evaluations of AI algorithms in testing microfluidics. We conducted a comparative evaluation of AI models specifically for the two-class classification problem of identifying the presence or absence of bubbles in microfluidic channels under various imaging conditions. Using a model microfluidic system with a single channel loaded with 3D transparent objects (bubbles), we challenged each of the tested machine learning (ML) (n = 6) and deep learning (DL) (n = 9) models across different background settings. Evaluation revealed that the random forest ML model achieved 95.52% sensitivity, 82.57% specificity, and 97% AUC, outperforming other ML algorithms. Among DL models suitable for mobile integration, DenseNet169 demonstrated superior performance, achieving 92.63% sensitivity, 92.22% specificity, and 92% AUC. Remarkably, DenseNet169 integration into a mobile POC system demonstrated exceptional accuracy (>0.84) in testing microfluidics at under challenging imaging settings. Our study confirms the transformative potential of AI in healthcare, emphasizing its capacity to revolutionize precision medicine through accurate and accessible diagnostics. The integration of AI into healthcare systems holds promise for enhancing patient outcomes and streamlining healthcare delivery.
Collapse
Affiliation(s)
- Mert Tunca Doganay
- Department of Medicine, Case Western Reserve University School of Medicine, Cleveland, OH, 44106, USA.
| | - Purbali Chakraborty
- Department of Medicine, Case Western Reserve University School of Medicine, Cleveland, OH, 44106, USA.
| | - Sri Moukthika Bommakanti
- Department of Medicine, Case Western Reserve University School of Medicine, Cleveland, OH, 44106, USA.
| | - Soujanya Jammalamadaka
- Department of Medicine, Case Western Reserve University School of Medicine, Cleveland, OH, 44106, USA.
| | | | - Anant Madabhushi
- Department of Biomedical Engineering, Emory University, Atlanta, GA, USA
- Atlanta Veterans Administration Medical Center, Atlanta, GA, USA
| | - Mohamed S Draz
- Department of Medicine, Case Western Reserve University School of Medicine, Cleveland, OH, 44106, USA.
- Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH, USA
- Department of Biomedical Engineering, Cleveland Clinic, Cleveland, OH, 44106, USA
| |
Collapse
|
27
|
Xu Y, Quan R, Xu W, Huang Y, Chen X, Liu F. Advances in Medical Image Segmentation: A Comprehensive Review of Traditional, Deep Learning and Hybrid Approaches. Bioengineering (Basel) 2024; 11:1034. [PMID: 39451409 PMCID: PMC11505408 DOI: 10.3390/bioengineering11101034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2024] [Revised: 10/08/2024] [Accepted: 10/11/2024] [Indexed: 10/26/2024] Open
Abstract
Medical image segmentation plays a critical role in accurate diagnosis and treatment planning, enabling precise analysis across a wide range of clinical tasks. This review begins by offering a comprehensive overview of traditional segmentation techniques, including thresholding, edge-based methods, region-based approaches, clustering, and graph-based segmentation. While these methods are computationally efficient and interpretable, they often face significant challenges when applied to complex, noisy, or variable medical images. The central focus of this review is the transformative impact of deep learning on medical image segmentation. We delve into prominent deep learning architectures such as Convolutional Neural Networks (CNNs), Fully Convolutional Networks (FCNs), U-Net, Recurrent Neural Networks (RNNs), Adversarial Networks (GANs), and Autoencoders (AEs). Each architecture is analyzed in terms of its structural foundation and specific application to medical image segmentation, illustrating how these models have enhanced segmentation accuracy across various clinical contexts. Finally, the review examines the integration of deep learning with traditional segmentation methods, addressing the limitations of both approaches. These hybrid strategies offer improved segmentation performance, particularly in challenging scenarios involving weak edges, noise, or inconsistent intensities. By synthesizing recent advancements, this review provides a detailed resource for researchers and practitioners, offering valuable insights into the current landscape and future directions of medical image segmentation.
Collapse
Affiliation(s)
- Yan Xu
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| | - Rixiang Quan
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| | - Weiting Xu
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| | - Yi Huang
- Bristol Medical School, University of Bristol, Bristol BS8 1UD, UK;
| | - Xiaolong Chen
- Department of Mechanical, Materials and Manufacturing Engineering, University of Nottingham, Nottingham NG7 2RD, UK;
| | - Fengyuan Liu
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| |
Collapse
|
28
|
Chen J, Lin F, Dai Z, Chen Y, Fan Y, Li A, Zhao C. Survival prediction in diffuse large B-cell lymphoma patients: multimodal PET/CT deep features radiomic model utilizing automated machine learning. J Cancer Res Clin Oncol 2024; 150:452. [PMID: 39382750 PMCID: PMC11464575 DOI: 10.1007/s00432-024-05905-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2024] [Accepted: 07/21/2024] [Indexed: 10/10/2024]
Abstract
PURPOSE We sought to develop an effective combined model for predicting the survival of patients with diffuse large B-cell lymphoma (DLBCL) based on the multimodal PET-CT deep features radiomics signature (DFR-signature). METHODS 369 DLBCL patients from two medical centers were included in this study. Their PET and CT images were fused to construct the multimodal PET-CT images using a deep learning fusion network. Then the deep features were extracted from those fused PET-CT images, and the DFR-signature was constructed through an Automated machine learning (AutoML) model. Combined with clinical indexes from the Cox regression analysis, we constructed a combined model to predict the progression-free survival (PFS) and the overall survival (OS) of patients. In addition, the combined model was evaluated in the concordance index (C-index) and the time-dependent area under the ROC curve (tdAUC). RESULTS A total of 1000 deep features were extracted to build a DFR-signature. Besides the DFR-signature, the combined model integrating metabolic and clinical factors performed best in terms of PFS and OS. For PFS, the C-indices are 0.784 and 0.739 in the training cohort and internal validation cohort, respectively. For OS, the C-indices are 0.831 and 0.782 in the training cohort and internal validation cohort. CONCLUSIONS DFR-signature constructed from multimodal images improved the classification accuracy of prognosis for DLBCL patients. Moreover, the constructed DFR-signature combined with NCCN-IPI exhibited excellent potential for risk stratification of DLBCL patients.
Collapse
Affiliation(s)
- Jianxin Chen
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China.
| | - Fengyi Lin
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China
| | - Zhaoyan Dai
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China
| | - Yu Chen
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China
| | - Yawen Fan
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China
| | - Ang Li
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China
| | - Chenyu Zhao
- The Key Laboratory of Broadband Wireless Communication and Sensor Network Technology (Ministry of Education), Nanjing University of Posts and Telecommunications, Nanjing, China
| |
Collapse
|
29
|
Chen M, Zhang M, Yin L, Ma L, Ding R, Zheng T, Yue Q, Lui S, Sun H. Medical image foundation models in assisting diagnosis of brain tumors: a pilot study. Eur Radiol 2024; 34:6667-6679. [PMID: 38627290 DOI: 10.1007/s00330-024-10728-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2023] [Revised: 02/08/2024] [Accepted: 03/04/2024] [Indexed: 04/23/2024]
Abstract
OBJECTIVES To build self-supervised foundation models for multicontrast MRI of the whole brain and evaluate their efficacy in assisting diagnosis of brain tumors. METHODS In this retrospective study, foundation models were developed using 57,621 enhanced head MRI scans through self-supervised learning with a pretext task of cross-contrast context restoration with two different content dropout schemes. Downstream classifiers were constructed based on the pretrained foundation models and fine-tuned for brain tumor detection, discrimination, and molecular status prediction. Metrics including accuracy, sensitivity, specificity, and area under the ROC curve (AUC) were used to evaluate the performance. Convolutional neural networks trained exclusively on downstream task data were employed for comparative analysis. RESULTS The pretrained foundation models demonstrated their ability to extract effective representations from multicontrast whole-brain volumes. The best classifiers, endowed with pretrained weights, showed remarkable performance with accuracies of 94.9, 92.3, and 80.4%, and corresponding AUC values of 0.981, 0.972, and 0.852 on independent test datasets in brain tumor detection, discrimination, and molecular status prediction, respectively. The classifiers with pretrained weights outperformed the convolutional classifiers trained from scratch by approximately 10% in terms of accuracy and AUC across all tasks. The saliency regions in the correctly predicted cases are mainly clustered around the tumors. Classifiers derived from the two dropout schemes differed significantly only in the detection of brain tumors. CONCLUSIONS Foundation models obtained from self-supervised learning have demonstrated encouraging potential for scalability and interpretability in downstream brain tumor-related tasks and hold promise for extension to neurological diseases with diffusely distributed lesions. CLINICAL RELEVANCE STATEMENT The application of our proposed method to the prediction of key molecular status in gliomas is expected to improve treatment planning and patient outcomes. Additionally, the foundation model we developed could serve as a cornerstone for advancing AI applications in the diagnosis of brain-related diseases.
Collapse
Affiliation(s)
- Mengyao Chen
- Department of Radiology, West China Hospital of Sichuan University, Chengdu, China
- Huaxi MR Research Center (HMRRC), West China Hospital of Sichuan University, Chengdu, China
| | | | - Lijuan Yin
- Department of Pathology, West China Hospital of Sichuan University, Chengdu, China
| | - Lu Ma
- Department of Neurosurgery, West China Hospital of Sichuan University, Chengdu, China
| | - Renxing Ding
- IT center, West China Hospital of Sichuan University, Chengdu, China
| | - Tao Zheng
- IT center, West China Hospital of Sichuan University, Chengdu, China
| | - Qiang Yue
- Department of Radiology, West China Hospital of Sichuan University, Chengdu, China
- Huaxi MR Research Center (HMRRC), West China Hospital of Sichuan University, Chengdu, China
| | - Su Lui
- Department of Radiology, West China Hospital of Sichuan University, Chengdu, China
- Huaxi MR Research Center (HMRRC), West China Hospital of Sichuan University, Chengdu, China
| | - Huaiqiang Sun
- Department of Radiology, West China Hospital of Sichuan University, Chengdu, China.
- Huaxi MR Research Center (HMRRC), West China Hospital of Sichuan University, Chengdu, China.
| |
Collapse
|
30
|
Saavedra JP, Droppelmann G, Jorquera C, Feijoo F. Automated segmentation and classification of supraspinatus fatty infiltration in shoulder magnetic resonance image using a convolutional neural network. Front Med (Lausanne) 2024; 11:1416169. [PMID: 39290391 PMCID: PMC11405335 DOI: 10.3389/fmed.2024.1416169] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2024] [Accepted: 08/20/2024] [Indexed: 09/19/2024] Open
Abstract
Background Goutallier's fatty infiltration of the supraspinatus muscle is a critical condition in degenerative shoulder disorders. Deep learning research primarily uses manual segmentation and labeling to detect this condition. Employing unsupervised training with a hybrid framework of segmentation and classification could offer an efficient solution. Aim To develop and assess a two-step deep learning model for detecting the region of interest and categorizing the magnetic resonance image (MRI) supraspinatus muscle fatty infiltration according to Goutallier's scale. Materials and methods A retrospective study was performed from January 1, 2019 to September 20, 2020, using 900 MRI T2-weighted images with supraspinatus muscle fatty infiltration diagnoses. A model with two sequential neural networks was implemented and trained. The first sub-model automatically detects the region of interest using a U-Net model. The second sub-model performs a binary classification using the VGG-19 architecture. The model's performance was computed as the average of five-fold cross-validation processes. Loss, accuracy, Dice coefficient (CI. 95%), AU-ROC, sensitivity, and specificity (CI. 95%) were reported. Results Six hundred and six shoulders MRIs were analyzed. The Goutallier distribution was presented as follows: 0 (66.50%); 1 (18.81%); 2 (8.42%); 3 (3.96%); 4 (2.31%). Segmentation results demonstrate high levels of accuracy (0.9977 ± 0.0002) and Dice score (0.9441 ± 0.0031), while the classification model also results in high levels of accuracy (0.9731 ± 0.0230); sensitivity (0.9000 ± 0.0980); specificity (0.9788 ± 0.0257); and AUROC (0.9903 ± 0.0092). Conclusion The two-step training method proposed using a deep learning model demonstrated strong performance in segmentation and classification tasks.
Collapse
Affiliation(s)
- Juan Pablo Saavedra
- School of Industrial Engineering, Pontificia Universidad Católica de Valparaíso, Valparaíso, Chile
| | - Guillermo Droppelmann
- Clínica MEDS, Santiago, Chile
- Harvard T.H. Chan School of Public Health, Boston, MA, United States
| | - Carlos Jorquera
- Facultad de Ciencias, Escuela de Nutrición y Dietética, Universidad Mayor, Santiago, Chile
| | - Felipe Feijoo
- School of Industrial Engineering, Pontificia Universidad Católica de Valparaíso, Valparaíso, Chile
| |
Collapse
|
31
|
Kesavapillai AR, Aslam SM, Umapathy S, Almutairi F. RA-XTNet: A Novel CNN Model to Predict Rheumatoid Arthritis from Hand Radiographs and Thermal Images: A Comparison with CNN Transformer and Quantum Computing. Diagnostics (Basel) 2024; 14:1911. [PMID: 39272696 PMCID: PMC11394616 DOI: 10.3390/diagnostics14171911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2024] [Revised: 08/17/2024] [Accepted: 08/18/2024] [Indexed: 09/15/2024] Open
Abstract
The aim and objective of the research are to develop an automated diagnosis system for the prediction of rheumatoid arthritis (RA) based on artificial intelligence (AI) and quantum computing for hand radiographs and thermal images. The hand radiographs and thermal images were segmented using a UNet++ model and color-based k-means clustering technique, respectively. The attributes from the segmented regions were generated using the Speeded-Up Robust Features (SURF) feature extractor and classification was performed using k-star and Hoeffding classifiers. For the ground truth and the predicted test image, the study utilizing UNet++ segmentation achieved a pixel-wise accuracy of 98.75%, an intersection over union (IoU) of 0.87, and a dice coefficient of 0.86, indicating a high level of similarity. The custom RA-X-ray thermal imaging (XTNet) surpassed all the models for the detection of RA with a classification accuracy of 90% and 93% for X-ray and thermal imaging modalities, respectively. Furthermore, the study employed quantum support vector machine (QSVM) as a quantum computing approach which yielded an accuracy of 93.75% and 87.5% for the detection of RA from hand X-ray and thermal images. In addition, vision transformer (ViT) was employed to classify RA which obtained an accuracy of 80% for hand X-rays and 90% for thermal images. Thus, depending on the performance measures, the RA-XTNet model can be used as an effective automated diagnostic method to diagnose RA accurately and rapidly in hand radiographs and thermal images.
Collapse
Affiliation(s)
- Ahalya R Kesavapillai
- Department of Biomedical Engineering, SRM Institute of Science and Technology, College of Engineering and Technology, Chennai 603203, India
- Department of Biomedical Engineering, Easwari Engineering College, Ramapuram, Chennai 600089, India
| | - Shabnam M Aslam
- Department of Information Technology, College of Computer and Information Sciences (CCIS), Majmaah University, Al Majmaah 11952, Saudi Arabia
| | - Snekhalatha Umapathy
- Department of Biomedical Engineering, SRM Institute of Science and Technology, College of Engineering and Technology, Chennai 603203, India
| | - Fadiyah Almutairi
- Department of Information System, College of Computer and Information Sciences (CCIS), Majmaah University, Al Majmaah 11952, Saudi Arabia
| |
Collapse
|
32
|
Martis JE, M S S, R B, Mutawa AM, Murugappan M. Novel Hybrid Quantum Architecture-Based Lung Cancer Detection Using Chest Radiograph and Computerized Tomography Images. Bioengineering (Basel) 2024; 11:799. [PMID: 39199758 PMCID: PMC11351577 DOI: 10.3390/bioengineering11080799] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2024] [Revised: 07/28/2024] [Accepted: 08/02/2024] [Indexed: 09/01/2024] Open
Abstract
Lung cancer, the second most common type of cancer worldwide, presents significant health challenges. Detecting this disease early is essential for improving patient outcomes and simplifying treatment. In this study, we propose a hybrid framework that combines deep learning (DL) with quantum computing to enhance the accuracy of lung cancer detection using chest radiographs (CXR) and computerized tomography (CT) images. Our system utilizes pre-trained models for feature extraction and quantum circuits for classification, achieving state-of-the-art performance in various metrics. Not only does our system achieve an overall accuracy of 92.12%, it also excels in other crucial performance measures, such as sensitivity (94%), specificity (90%), F1-score (93%), and precision (92%). These results demonstrate that our hybrid approach can more accurately identify lung cancer signatures compared to traditional methods. Moreover, the incorporation of quantum computing enhances processing speed and scalability, making our system a promising tool for early lung cancer screening and diagnosis. By leveraging the strengths of quantum computing, our approach surpasses traditional methods in terms of speed, accuracy, and efficiency. This study highlights the potential of hybrid computational technologies to transform early cancer detection, paving the way for wider clinical applications and improved patient care outcomes.
Collapse
Affiliation(s)
- Jason Elroy Martis
- Department of ISE, NMAM Institute of Technology, Nitte Deemed to be University, Udupi 574110, Karnataka, India (B.R.)
| | - Sannidhan M S
- Department of CSE, NMAM Institute of Technology, Nitte Deemed to be University, Udupi 574110, Karnataka, India;
| | - Balasubramani R
- Department of ISE, NMAM Institute of Technology, Nitte Deemed to be University, Udupi 574110, Karnataka, India (B.R.)
| | - A. M. Mutawa
- Computer Engineering Department, College of Engineering and Petroleum, Kuwait University, Safat 13060, Kuwait
- Computer Sciences Department, University of Hamburg, 22527 Hamburg, Germany
| | - M. Murugappan
- Intelligent Signal Processing (ISP) Research Lab, Department of Electronics and Communication Engineering, Kuwait College of Science and Technology, Block 4, Doha 13133, Kuwait
- Department of Electronics and Communication Engineering, School of Engineering, Vels Institute of Sciences, Technology, and Advanced Studies, Chennai 600117, Tamil Nadu, India
- Center of Excellence for Unmanned Aerial Systems (CoEUAS), Universiti Malaysia Perlis, Arau 02600, Malaysia
| |
Collapse
|
33
|
Wang J, Liu G, Zhou C, Cui X, Wang W, Wang J, Huang Y, Jiang J, Wang Z, Tang Z, Zhang A, Cui D. Application of artificial intelligence in cancer diagnosis and tumor nanomedicine. NANOSCALE 2024; 16:14213-14246. [PMID: 39021117 DOI: 10.1039/d4nr01832j] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/20/2024]
Abstract
Cancer is a major health concern due to its high incidence and mortality rates. Advances in cancer research, particularly in artificial intelligence (AI) and deep learning, have shown significant progress. The swift evolution of AI in healthcare, especially in tools like computer-aided diagnosis, has the potential to revolutionize early cancer detection. This technology offers improved speed, accuracy, and sensitivity, bringing a transformative impact on cancer diagnosis, treatment, and management. This paper provides a concise overview of the application of artificial intelligence in the realms of medicine and nanomedicine, with a specific emphasis on the significance and challenges associated with cancer diagnosis. It explores the pivotal role of AI in cancer diagnosis, leveraging structured, unstructured, and multimodal fusion data. Additionally, the article delves into the applications of AI in nanomedicine sensors and nano-oncology drugs. The fundamentals of deep learning and convolutional neural networks are clarified, underscoring their relevance to AI-driven cancer diagnosis. A comparative analysis is presented, highlighting the accuracy and efficiency of traditional methods juxtaposed with AI-based approaches. The discussion not only assesses the current state of AI in cancer diagnosis but also delves into the challenges faced by AI in this context. Furthermore, the article envisions the future development direction and potential application of artificial intelligence in cancer diagnosis, offering a hopeful prospect for enhanced cancer detection and improved patient prognosis.
Collapse
Affiliation(s)
- Junhao Wang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Guan Liu
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Cheng Zhou
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Xinyuan Cui
- Imaging Department of Rui Jin Hospital, Medical School of Shanghai Jiao Tong University, Shanghai, China
| | - Wei Wang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Jiulin Wang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Yixin Huang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Jinlei Jiang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Zhitao Wang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Zengyi Tang
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Amin Zhang
- Department of Food Science & Technology, School of Agriculture & Biology, Shanghai Jiao Tong University, Shanghai, China.
| | - Daxiang Cui
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China.
- School of Medicine, Henan University, Henan, China
| |
Collapse
|
34
|
Ali IE, Sumita Y, Wakabayashi N. Advancing maxillofacial prosthodontics by using pre-trained convolutional neural networks: Image-based classification of the maxilla. J Prosthodont 2024; 33:645-654. [PMID: 38566564 DOI: 10.1111/jopr.13853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2023] [Accepted: 03/15/2024] [Indexed: 04/04/2024] Open
Abstract
PURPOSE The study aimed to compare the performance of four pre-trained convolutional neural networks in recognizing seven distinct prosthodontic scenarios involving the maxilla, as a preliminary step in developing an artificial intelligence (AI)-powered prosthesis design system. MATERIALS AND METHODS Seven distinct classes, including cleft palate, dentulous maxillectomy, edentulous maxillectomy, reconstructed maxillectomy, completely dentulous, partially edentulous, and completely edentulous, were considered for recognition. Utilizing transfer learning and fine-tuned hyperparameters, four AI models (VGG16, Inception-ResNet-V2, DenseNet-201, and Xception) were employed. The dataset, consisting of 3541 preprocessed intraoral occlusal images, was divided into training, validation, and test sets. Model performance metrics encompassed accuracy, precision, recall, F1 score, area under the receiver operating characteristic curve (AUC), and confusion matrix. RESULTS VGG16, Inception-ResNet-V2, DenseNet-201, and Xception demonstrated comparable performance, with maximum test accuracies of 0.92, 0.90, 0.94, and 0.95, respectively. Xception and DenseNet-201 slightly outperformed the other models, particularly compared with InceptionResNet-V2. Precision, recall, and F1 scores exceeded 90% for most classes in Xception and DenseNet-201 and the average AUC values for all models ranged between 0.98 and 1.00. CONCLUSIONS While DenseNet-201 and Xception demonstrated superior performance, all models consistently achieved diagnostic accuracy exceeding 90%, highlighting their potential in dental image analysis. This AI application could help work assignments based on difficulty levels and enable the development of an automated diagnosis system at patient admission. It also facilitates prosthesis designing by integrating necessary prosthesis morphology, oral function, and treatment difficulty. Furthermore, it tackles dataset size challenges in model optimization, providing valuable insights for future research.
Collapse
Affiliation(s)
- Islam E Ali
- Department of Advanced Prosthodontics, Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
- Department of Prosthodontics, Faculty of Dentistry, Mansoura University, Mansoura, Egypt
| | - Yuka Sumita
- Division of General Dentistry 4, The Nippon Dental University Hospital, Tokyo, Japan
- Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
| | - Noriyuki Wakabayashi
- Department of Advanced Prosthodontics, Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
| |
Collapse
|
35
|
Choopong P, Kusakunniran W. Selection of pre-trained weights for transfer learning in automated cytomegalovirus retinitis classification. Sci Rep 2024; 14:15899. [PMID: 38987446 PMCID: PMC11237151 DOI: 10.1038/s41598-024-67121-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2024] [Accepted: 07/08/2024] [Indexed: 07/12/2024] Open
Abstract
Cytomegalovirus retinitis (CMVR) is a significant cause of vision loss. Regular screening is crucial but challenging in resource-limited settings. A convolutional neural network is a state-of-the-art deep learning technique to generate automatic diagnoses from retinal images. However, there are limited numbers of CMVR images to train the model properly. Transfer learning (TL) is a strategy to train a model with a scarce dataset. This study explores the efficacy of TL with different pre-trained weights for automated CMVR classification using retinal images. We utilised a dataset of 955 retinal images (524 CMVR and 431 normal) from Siriraj Hospital, Mahidol University, collected between 2005 and 2015. Images were processed using Kowa VX-10i or VX-20 fundus cameras and augmented for training. We employed DenseNet121 as a backbone model, comparing the performance of TL with weights pre-trained on ImageNet, APTOS2019, and CheXNet datasets. The models were evaluated based on accuracy, loss, and other performance metrics, with the depth of fine-tuning varied across different pre-trained weights. The study found that TL significantly enhances model performance in CMVR classification. The best results were achieved with weights sequentially transferred from ImageNet to APTOS2019 dataset before application to our CMVR dataset. This approach yielded the highest mean accuracy (0.99) and lowest mean loss (0.04), outperforming other methods. The class activation heatmaps provided insights into the model's decision-making process. The model with APTOS2019 pre-trained weights offered the best explanation and highlighted the pathologic lesions resembling human interpretation. Our findings demonstrate the potential of sequential TL in improving the accuracy and efficiency of CMVR diagnosis, particularly in settings with limited data availability. They highlight the importance of domain-specific pre-training in medical image classification. This approach streamlines the diagnostic process and paves the way for broader applications in automated medical image analysis, offering a scalable solution for early disease detection.
Collapse
Affiliation(s)
- Pitipol Choopong
- Department of Ophthalmology, Faculty of Medicine Siriraj Hospital, Mahidol University, Bangkok, Thailand
- Faculty of Information and Communication Technology, Mahidol University, Nakhon Pathom, Thailand
| | - Worapan Kusakunniran
- Faculty of Information and Communication Technology, Mahidol University, Nakhon Pathom, Thailand.
| |
Collapse
|
36
|
Chattopadhyay T, Ozarkar SS, Buwa K, Joshy NA, Komandur D, Naik J, Thomopoulos SI, Ver Steeg G, Ambite JL, Thompson PM. Comparison of deep learning architectures for predicting amyloid positivity in Alzheimer's disease, mild cognitive impairment, and healthy aging, from T1-weighted brain structural MRI. Front Neurosci 2024; 18:1387196. [PMID: 39015378 PMCID: PMC11250587 DOI: 10.3389/fnins.2024.1387196] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 06/14/2024] [Indexed: 07/18/2024] Open
Abstract
Abnormal β-amyloid (Aβ) accumulation in the brain is an early indicator of Alzheimer's disease (AD) and is typically assessed through invasive procedures such as PET (positron emission tomography) or CSF (cerebrospinal fluid) assays. As new anti-Alzheimer's treatments can now successfully target amyloid pathology, there is a growing interest in predicting Aβ positivity (Aβ+) from less invasive, more widely available types of brain scans, such as T1-weighted (T1w) MRI. Here we compare multiple approaches to infer Aβ + from standard anatomical MRI: (1) classical machine learning algorithms, including logistic regression, XGBoost, and shallow artificial neural networks, (2) deep learning models based on 2D and 3D convolutional neural networks (CNNs), (3) a hybrid ANN-CNN, combining the strengths of shallow and deep neural networks, (4) transfer learning models based on CNNs, and (5) 3D Vision Transformers. All models were trained on paired MRI/PET data from 1,847 elderly participants (mean age: 75.1 yrs. ± 7.6SD; 863 females/984 males; 661 healthy controls, 889 with mild cognitive impairment (MCI), and 297 with Dementia), scanned as part of the Alzheimer's Disease Neuroimaging Initiative. We evaluated each model's balanced accuracy and F1 scores. While further tests on more diverse data are warranted, deep learning models trained on standard MRI showed promise for estimating Aβ + status, at least in people with MCI. This may offer a potential screening option before resorting to more invasive procedures.
Collapse
Affiliation(s)
- Tamoghna Chattopadhyay
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | - Saket S. Ozarkar
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | - Ketaki Buwa
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | - Neha Ann Joshy
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | - Dheeraj Komandur
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | - Jayati Naik
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | - Sophia I. Thomopoulos
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| | | | - Jose Luis Ambite
- Information Sciences Institute, University of Southern California, Marina del Rey, CA, United States
| | - Paul M. Thompson
- Imaging Genetics Center, Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Marina del Rey, CA, United States
| |
Collapse
|
37
|
Jeong HK, Park C, Jiang SW, Nicholas M, Chen S, Henao R, Kheterpal M. Image Quality Assessment Using Convolutional Neural Network in Clinical Skin Images. JID INNOVATIONS 2024; 4:100285. [PMID: 39036289 PMCID: PMC11260318 DOI: 10.1016/j.xjidi.2024.100285] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2023] [Revised: 12/24/2023] [Accepted: 03/06/2024] [Indexed: 07/23/2024] Open
Abstract
The image quality received for clinical evaluation is often suboptimal. The goal is to develop an image quality analysis tool to assess patient- and primary care physician-derived images using deep learning model. Dataset included patient- and primary care physician-derived images from August 21, 2018 to June 30, 2022 with 4 unique quality labels. VGG16 model was fine tuned with input data, and optimal threshold was determined by Youden's index. Ordinal labels were transformed to binary labels using a majority vote because model distinguishes between 2 categories (good vs bad). At a threshold of 0.587, area under the curve for the test set was 0.885 (95% confidence interval = 0.838-0.933); sensitivity, specificity, positive predictive value, and negative predictive value were 0.829, 0.784, 0.906, and 0.645, respectively. Independent validation of 300 additional images (from patients and primary care physicians) demonstrated area under the curve of 0.864 (95% confidence interval = 0.818-0.909) and area under the curve of 0.902 (95% confidence interval = 0.85-0.95), respectively. The sensitivity, specificity, positive predictive value, and negative predictive value for the 300 images were 0.827, 0.800, 0.959, and 0.450, respectively. We demonstrate a practical approach improving the image quality for clinical workflow. Although users may have to capture additional images, this is offset by the improved workload and efficiency for clinical teams.
Collapse
Affiliation(s)
- Hyeon Ki Jeong
- Department of Biostatistics & Bioinformatics, Duke University School of Medicine, Durham, North Carolina, USA
| | - Christine Park
- Duke University School of Medicine, Durham, North Carolina, USA
| | - Simon W. Jiang
- Duke University School of Medicine, Durham, North Carolina, USA
| | - Matilda Nicholas
- Department of Dermatology, Duke University School of Medicine, Durham, North Carolina, USA
| | - Suephy Chen
- Department of Dermatology, Duke University School of Medicine, Durham, North Carolina, USA
- Durham VA Medical Center, Durham, North Carolina, USA
| | - Ricardo Henao
- Department of Biostatistics & Bioinformatics, Duke University School of Medicine, Durham, North Carolina, USA
| | - Meenal Kheterpal
- Department of Dermatology, Duke University School of Medicine, Durham, North Carolina, USA
| |
Collapse
|
38
|
Kolhar M, Al Rajeh AM, Kazi RNA. Augmenting Radiological Diagnostics with AI for Tuberculosis and COVID-19 Disease Detection: Deep Learning Detection of Chest Radiographs. Diagnostics (Basel) 2024; 14:1334. [PMID: 39001228 PMCID: PMC11240993 DOI: 10.3390/diagnostics14131334] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2024] [Revised: 06/18/2024] [Accepted: 06/21/2024] [Indexed: 07/16/2024] Open
Abstract
In this research, we introduce a network that can identify pneumonia, COVID-19, and tuberculosis using X-ray images of patients' chests. The study emphasizes tuberculosis, COVID-19, and healthy lung conditions, discussing how advanced neural networks, like VGG16 and ResNet50, can improve the detection of lung issues from images. To prepare the images for the model's input requirements, we enhanced them through data augmentation techniques for training purposes. We evaluated the model's performance by analyzing the precision, recall, and F1 scores across training, validation, and testing datasets. The results show that the ResNet50 model outperformed VGG16 with accuracy and resilience. It displayed superior ROC AUC values in both validation and test scenarios. Particularly impressive were ResNet50's precision and recall rates, nearing 0.99 for all conditions in the test set. On the hand, VGG16 also performed well during testing-detecting tuberculosis with a precision of 0.99 and a recall of 0.93. Our study highlights the performance of our deep learning method by showcasing the effectiveness of ResNet50 over traditional approaches like VGG16. This progress utilizes methods to enhance classification accuracy by augmenting data and balancing them. This positions our approach as an advancement in using state-of-the-art deep learning applications in imaging. By enhancing the accuracy and reliability of diagnosing ailments such as COVID-19 and tuberculosis, our models have the potential to transform care and treatment strategies, highlighting their role in clinical diagnostics.
Collapse
Affiliation(s)
- Manjur Kolhar
- Department Health Informatics, College of Applied Medical Sciences, King Faisal University, Al-Hofuf 31982, Saudi Arabia
| | - Ahmed M Al Rajeh
- College of Applied Medical Sciences, King Faisal University, Al-Hofuf 31982, Saudi Arabia
| | - Raisa Nazir Ahmed Kazi
- College of Applied Medical Sciences, King Faisal University, Al-Hofuf 31982, Saudi Arabia
| |
Collapse
|
39
|
Connolly L, Fooladgar F, Jamzad A, Kaufmann M, Syeda A, Ren K, Abolmaesumi P, Rudan JF, McKay D, Fichtinger G, Mousavi P. ImSpect: Image-driven self-supervised learning for surgical margin evaluation with mass spectrometry. Int J Comput Assist Radiol Surg 2024; 19:1129-1136. [PMID: 38600411 DOI: 10.1007/s11548-024-03106-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Accepted: 03/08/2024] [Indexed: 04/12/2024]
Abstract
PURPOSE Real-time assessment of surgical margins is critical for favorable outcomes in cancer patients. The iKnife is a mass spectrometry device that has demonstrated potential for margin detection in cancer surgery. Previous studies have shown that using deep learning on iKnife data can facilitate real-time tissue characterization. However, none of the existing literature on the iKnife facilitate the use of publicly available, state-of-the-art pretrained networks or datasets that have been used in computer vision and other domains. METHODS In a new framework we call ImSpect, we convert 1D iKnife data, captured during basal cell carcinoma (BCC) surgery, into 2D images in order to capitalize on state-of-the-art image classification networks. We also use self-supervision to leverage large amounts of unlabeled, intraoperative data to accommodate the data requirements of these networks. RESULTS Through extensive ablation studies, we show that we can surpass previous benchmarks of margin evaluation in BCC surgery using iKnife data, achieving an area under the receiver operating characteristic curve (AUC) of 81%. We also depict the attention maps of the developed DL models to evaluate the biological relevance of the embedding space CONCLUSIONS: We propose a new method for characterizing tissue at the surgical margins, using mass spectrometry data from cancer surgery.
Collapse
Affiliation(s)
| | | | | | | | | | - Kevin Ren
- Queen's University, Kingston, ON, Canada
| | | | | | - Doug McKay
- Queen's University, Kingston, ON, Canada
| | | | | |
Collapse
|
40
|
Hase H, Mine Y, Okazaki S, Yoshimi Y, Ito S, Peng TY, Sano M, Koizumi Y, Kakimoto N, Tanimoto K, Murayama T. Sex estimation from maxillofacial radiographs using a deep learning approach. Dent Mater J 2024; 43:394-399. [PMID: 38599831 DOI: 10.4012/dmj.2023-253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/12/2024]
Abstract
The purpose of this study was to construct deep learning models for more efficient and reliable sex estimation. Two deep learning models, VGG16 and DenseNet-121, were used in this retrospective study. In total, 600 lateral cephalograms were analyzed. A saliency map was generated by gradient-weighted class activation mapping for each output. The two deep learning models achieved high values in each performance metric according to accuracy, sensitivity (recall), precision, F1 score, and areas under the receiver operating characteristic curve. Both models showed substantial differences in the positions indicated in saliency maps for male and female images. The positions in saliency maps also differed between VGG16 and DenseNet-121, regardless of sex. This analysis of our proposed system suggested that sex estimation from lateral cephalograms can be achieved with high accuracy using deep learning.
Collapse
Affiliation(s)
- Hiroki Hase
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University
| | - Yuichi Mine
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University
- Project Research Center for Integrating Digital Dentistry, Hiroshima University
| | - Shota Okazaki
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University
- Project Research Center for Integrating Digital Dentistry, Hiroshima University
| | - Yuki Yoshimi
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University
| | - Shota Ito
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University
| | - Tzu-Yu Peng
- School of Dentistry, College of Oral Medicine, Taipei Medical University
| | - Mizuho Sano
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University
| | - Yuma Koizumi
- Department of Orthodontics and Craniofacial Developmental Biology, Graduate School of Biomedical and Health Sciences, Hiroshima University
| | - Naoya Kakimoto
- School of Dentistry, College of Oral Medicine, Taipei Medical University
| | - Kotaro Tanimoto
- Department of Oral and Maxillofacial Radiology, Graduate School of Biomedical and Health Sciences, Hiroshima University
| | - Takeshi Murayama
- Department of Medical Systems Engineering, Graduate School of Biomedical and Health Sciences, Hiroshima University
- Project Research Center for Integrating Digital Dentistry, Hiroshima University
| |
Collapse
|
41
|
Han X, Chen Z, Lin G, Lv W, Zheng C, Lu W, Sun Y, Lu L. Semi-supervised model based on implicit neural representation and mutual learning (SIMN) for multi-center nasopharyngeal carcinoma segmentation on MRI. Comput Biol Med 2024; 175:108368. [PMID: 38663351 DOI: 10.1016/j.compbiomed.2024.108368] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Revised: 03/06/2024] [Accepted: 03/24/2024] [Indexed: 05/15/2024]
Abstract
BACKGROUND The issue of using deep learning to obtain accurate gross tumor volume (GTV) and metastatic lymph nodes (MLN) segmentation for nasopharyngeal carcinoma (NPC) on heterogeneous magnetic resonance imaging (MRI) images with limited labeling remains unsolved. METHOD We collected 918 patients with MRI images from three hospitals to develop and validate models and proposed a semi-supervised framework for the fine delineation of multi-center NPC boundaries by integrating uncertainty-based implicit neural representations named SIMN. The framework utilizes the deep mutual learning approach with CNN and Transformer, incorporating dynamic thresholds. Additionally, domain adaptive algorithms are employed to enhance the performance. RESULTS SIMN predictions have a high overlap ratio with the ground truth. Under the 20 % labeled cases, for the internal test cohorts, the average DSC in GTV and MLN are 0.7981 and 0.7804, respectively; for external test cohort Wu Zhou Red Cross Hospital, the average DSC in GTV and MLN are 0.7217 and 0.7581, respectively; for external test cohorts First People Hospital of Foshan, the average DSC in GTV and MLN are 0.7004 and 0.7692, respectively. No significant differences are found in DSC, HD95, ASD, and Recall for patients with different clinical categories. Moreover, SIMN outperformed existing classical semi-supervised methods. CONCLUSIONS SIMN showed a highly accurate GTV and MLN segmentation for NPC on multi-center MRI images under Semi-Supervised Learning (SSL), which can easily transfer to other centers without fine-tuning. It suggests that it has the potential to act as a generalized delineation solution for heterogeneous MRI images with limited labels in clinical deployment.
Collapse
Affiliation(s)
- Xu Han
- School of Biomedical Engineering, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Pazhou Lab, Guangzhou, 510515, China
| | - Zihang Chen
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, 510060, China
| | - Guoyu Lin
- School of Biomedical Engineering, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, 510060, China
| | - Wenbing Lv
- School of Information Science and Engineering, Yunnan University, Kunming, 650504, China
| | - Chundan Zheng
- School of Biomedical Engineering, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Pazhou Lab, Guangzhou, 510515, China
| | - Wantong Lu
- School of Biomedical Engineering, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Pazhou Lab, Guangzhou, 510515, China
| | - Ying Sun
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, 510060, China.
| | - Lijun Lu
- School of Biomedical Engineering, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, 1023 Shatai Road, Guangzhou, 510515, China; Pazhou Lab, Guangzhou, 510515, China.
| |
Collapse
|
42
|
Sheng H, Ma L, Samson JF, Liu D. BarlowTwins-CXR: enhancing chest X-ray abnormality localization in heterogeneous data with cross-domain self-supervised learning. BMC Med Inform Decis Mak 2024; 24:126. [PMID: 38755563 PMCID: PMC11097466 DOI: 10.1186/s12911-024-02529-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 05/07/2024] [Indexed: 05/18/2024] Open
Abstract
BACKGROUND Chest X-ray imaging based abnormality localization, essential in diagnosing various diseases, faces significant clinical challenges due to complex interpretations and the growing workload of radiologists. While recent advances in deep learning offer promising solutions, there is still a critical issue of domain inconsistency in cross-domain transfer learning, which hampers the efficiency and accuracy of diagnostic processes. This study aims to address the domain inconsistency problem and improve autonomic abnormality localization performance of heterogeneous chest X-ray image analysis, particularly in detecting abnormalities, by developing a self-supervised learning strategy called "BarlwoTwins-CXR". METHODS We utilized two publicly available datasets: the NIH Chest X-ray Dataset and the VinDr-CXR. The BarlowTwins-CXR approach was conducted in a two-stage training process. Initially, self-supervised pre-training was performed using an adjusted Barlow Twins algorithm on the NIH dataset with a Resnet50 backbone pre-trained on ImageNet. This was followed by supervised fine-tuning on the VinDr-CXR dataset using Faster R-CNN with Feature Pyramid Network (FPN). The study employed mean Average Precision (mAP) at an Intersection over Union (IoU) of 50% and Area Under the Curve (AUC) for performance evaluation. RESULTS Our experiments showed a significant improvement in model performance with BarlowTwins-CXR. The approach achieved a 3% increase in mAP50 accuracy compared to traditional ImageNet pre-trained models. In addition, the Ablation CAM method revealed enhanced precision in localizing chest abnormalities. The study involved 112,120 images from the NIH dataset and 18,000 images from the VinDr-CXR dataset, indicating robust training and testing samples. CONCLUSION BarlowTwins-CXR significantly enhances the efficiency and accuracy of chest X-ray image-based abnormality localization, outperforming traditional transfer learning methods and effectively overcoming domain inconsistency in cross-domain scenarios. Our experiment results demonstrate the potential of using self-supervised learning to improve the generalizability of models in medical settings with limited amounts of heterogeneous data. This approach can be instrumental in aiding radiologists, particularly in high-workload environments, offering a promising direction for future AI-driven healthcare solutions.
Collapse
Affiliation(s)
- Haoyue Sheng
- Département d'informatique et de recherche opérationnelle, Université de Montréal, 2920 chemin de la Tour, Montréal, H3T 1J4, QC, Canada.
- Mila - Quebec AI Institute, 6666 Rue Saint-Urbain, Montréal, H2S 3H1, QC, Canada.
- Direction des ressources informationnelles, CIUSSS du Centre-Sud-de-l'Île-de-Montréal, 400 Blvd. De Maisonneuve Ouest, Montréal, H3A 1L4, QC, Canada.
| | - Linrui Ma
- Département d'informatique et de recherche opérationnelle, Université de Montréal, 2920 chemin de la Tour, Montréal, H3T 1J4, QC, Canada
- Mila - Quebec AI Institute, 6666 Rue Saint-Urbain, Montréal, H2S 3H1, QC, Canada
| | - Jean-François Samson
- Direction des ressources informationnelles, CIUSSS du Centre-Sud-de-l'Île-de-Montréal, 400 Blvd. De Maisonneuve Ouest, Montréal, H3A 1L4, QC, Canada
| | - Dianbo Liu
- Mila - Quebec AI Institute, 6666 Rue Saint-Urbain, Montréal, H2S 3H1, QC, Canada
- School of Medicine and College of Design and Engineering, National University of Singapore, 21 Lower Kent Ridge Rd, Singapore, 119077, SG, Singapore
| |
Collapse
|
43
|
Shobayo O, Saatchi R, Ramlakhan S. Convolutional Neural Network to Classify Infrared Thermal Images of Fractured Wrists in Pediatrics. Healthcare (Basel) 2024; 12:994. [PMID: 38786405 PMCID: PMC11121475 DOI: 10.3390/healthcare12100994] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2024] [Revised: 05/02/2024] [Accepted: 05/10/2024] [Indexed: 05/25/2024] Open
Abstract
Convolutional neural network (CNN) models were devised and evaluated to classify infrared thermal (IRT) images of pediatric wrist fractures. The images were recorded from 19 participants with a wrist fracture and 21 without a fracture (sprain). The injury diagnosis was by X-ray radiography. For each participant, 299 IRT images of their wrists were recorded. These generated 11,960 images (40 participants × 299 images). For each image, the wrist region of interest (ROI) was selected and fast Fourier transformed (FFT) to obtain a magnitude frequency spectrum. The spectrum was resized to 100 × 100 pixels from its center as this region represented the main frequency components. Image augmentations of rotation, translation and shearing were applied to the 11,960 magnitude frequency spectra to assist with the CNN generalization during training. The CNN had 34 layers associated with convolution, batch normalization, rectified linear unit, maximum pooling and SoftMax and classification. The ratio of images for the training and test was 70:30, respectively. The effects of augmentation and dropout on CNN performance were explored. Wrist fracture identification sensitivity and accuracy of 88% and 76%, respectively, were achieved. The CNN model was able to identify wrist fractures; however, a larger sample size would improve accuracy.
Collapse
Affiliation(s)
- Olamilekan Shobayo
- Department of Computing, Sheffield Hallam University, Sheffield S1 2NU, UK;
| | - Reza Saatchi
- Department of Engineering and Mathematics, Sheffield Hallam University, Sheffield S1 1WB, UK
| | - Shammi Ramlakhan
- Emergency Department, Sheffield Children’s Hospital NHS Foundation Trust, Sheffield S10 2TH, UK;
| |
Collapse
|
44
|
Büyükçakır B, Bertels J, Claes P, Vandermeulen D, de Tobel J, Thevissen PW. OPG-based dental age estimation using a data-technical exploration of deep learning techniques. J Forensic Sci 2024; 69:919-931. [PMID: 38291770 DOI: 10.1111/1556-4029.15473] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2023] [Revised: 01/07/2024] [Accepted: 01/16/2024] [Indexed: 02/01/2024]
Abstract
Dental age estimation, a cornerstone in forensic age assessment, has been extensively tried and tested, yet manual methods are impeded by tedium and interobserver variability. Automated approaches using deep transfer learning encounter challenges like data scarcity, suboptimal training, and fine-tuning complexities, necessitating robust training methods. This study explores the impact of convolutional neural network hyperparameters, model complexity, training batch size, and sample quantity on age estimation. EfficientNet-B4, DenseNet-201, and MobileNet V3 models underwent cross-validation on a dataset of 3896 orthopantomograms (OPGs) with batch sizes escalating from 10 to 160 in a doubling progression, as well as random subsets of this training dataset. Results demonstrate the EfficientNet-B4 model, trained on the complete dataset with a batch size of 160, as the top performer with a mean absolute error of 0.562 years on the test set, notably surpassing the MAE of 1.01 at a batch size of 10. Increasing batch size consistently improved performance for EfficientNet-B4 and DenseNet-201, whereas MobileNet V3 performance peaked at batch size 40. Similar trends emerged in training with reduced sample sizes, though they were outperformed by the complete models. This underscores the critical role of hyperparameter optimization in adopting deep learning for age estimation from complete OPGs. The findings not only highlight the nuanced interplay of hyperparameters and performance but also underscore the potential for accurate age estimation models through optimization. This study contributes to advancing the application of deep learning in forensic age estimation, emphasizing the significance of tailored training methodologies for optimal outcomes.
Collapse
Affiliation(s)
- Barkın Büyükçakır
- ESAT, Center for Processing Speech and Images, KU Leuven, Leuven, Belgium
| | - Jeroen Bertels
- ESAT, Center for Processing Speech and Images, KU Leuven, Leuven, Belgium
| | - Peter Claes
- ESAT, Center for Processing Speech and Images, KU Leuven, Leuven, Belgium
| | - Dirk Vandermeulen
- ESAT, Center for Processing Speech and Images, KU Leuven, Leuven, Belgium
| | - Jannick de Tobel
- Department of Diagnostic Sciences and Radiology, Ghent University, Ghent, Belgium
| | | |
Collapse
|
45
|
Moharrami M, Farmer J, Singhal S, Watson E, Glogauer M, Johnson AEW, Schwendicke F, Quinonez C. Detecting dental caries on oral photographs using artificial intelligence: A systematic review. Oral Dis 2024; 30:1765-1783. [PMID: 37392423 DOI: 10.1111/odi.14659] [Citation(s) in RCA: 16] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Revised: 05/19/2023] [Accepted: 06/15/2023] [Indexed: 07/03/2023]
Abstract
OBJECTIVES This systematic review aimed at evaluating the performance of artificial intelligence (AI) models in detecting dental caries on oral photographs. METHODS Methodological characteristics and performance metrics of clinical studies reporting on deep learning and other machine learning algorithms were assessed. The risk of bias was evaluated using the quality assessment of diagnostic accuracy studies 2 (QUADAS-2) tool. A systematic search was conducted in EMBASE, Medline, and Scopus. RESULTS Out of 3410 identified records, 19 studies were included with six and seven studies having low risk of biases and applicability concerns for all the domains, respectively. Metrics varied widely and were assessed on multiple levels. F1-scores for classification and detection tasks were 68.3%-94.3% and 42.8%-95.4%, respectively. Irrespective of the task, F1-scores were 68.3%-95.4% for professional cameras, 78.8%-87.6%, for intraoral cameras, and 42.8%-80% for smartphone cameras. Limited studies allowed assessing AI performance for lesions of different severity. CONCLUSION Automatic detection of dental caries using AI may provide objective verification of clinicians' diagnoses and facilitate patient-clinician communication and teledentistry. Future studies should consider more robust study designs, employ comparable and standardized metrics, and focus on the severity of caries lesions.
Collapse
Affiliation(s)
- Mohammad Moharrami
- Faculty of Dentistry, University of Toronto, Toronto, Ontario, Canada
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Geneva, Switzerland
| | - Julie Farmer
- Faculty of Dentistry, University of Toronto, Toronto, Ontario, Canada
| | - Sonica Singhal
- Faculty of Dentistry, University of Toronto, Toronto, Ontario, Canada
- Health Promotion, Chronic Disease and Injury Prevention Department, Public Health Ontario, Toronto, Canada
| | - Erin Watson
- Faculty of Dentistry, University of Toronto, Toronto, Ontario, Canada
- Department of Dental Oncology, Princess Margaret Cancer Centre, Toronto, Ontario, Canada
| | - Michael Glogauer
- Faculty of Dentistry, University of Toronto, Toronto, Ontario, Canada
- Department of Dental Oncology, Princess Margaret Cancer Centre, Toronto, Ontario, Canada
- Department of Dentistry, Centre for Advanced Dental Research and Care, Mount Sinai Hospital, Toronto, Ontario, Canada
| | - Alistair E W Johnson
- Program in Child Health Evaluative Sciences, The Hospital for Sick Children, Toronto, Ontario, Canada
| | - Falk Schwendicke
- Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Geneva, Switzerland
- Oral Diagnostics, Digital Health and Health Services Research, Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Carlos Quinonez
- Faculty of Dentistry, University of Toronto, Toronto, Ontario, Canada
- Schulich School of Medicine and Dentistry, Western University, London, Ontario, Canada
| |
Collapse
|
46
|
Wang Z, Song J, Lin K, Hong W, Mao S, Wu X, Zhang J. Automated detection of otosclerosis with interpretable deep learning using temporal bone computed tomography images. Heliyon 2024; 10:e29670. [PMID: 38655358 PMCID: PMC11036044 DOI: 10.1016/j.heliyon.2024.e29670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Revised: 04/10/2024] [Accepted: 04/12/2024] [Indexed: 04/26/2024] Open
Abstract
Objective This study aimed to develop an automated detection schema for otosclerosis with interpretable deep learning using temporal bone computed tomography images. Methods With approval from the institutional review board, we retrospectively analyzed high-resolution computed tomography scans of the temporal bone of 182 participants with otosclerosis (67 male subjects and 115 female subjects; average age, 36.42 years) and 157 participants without otosclerosis (52 male subjects and 102 female subjects; average age, 30.61 years) using deep learning. Transfer learning with the pretrained VGG19, Mask RCNN, and EfficientNet models was used. In addition, 3 clinical experts compared the system's performance by reading the same computed tomography images for a subset of 35 unseen subjects. An area under the receiver operating characteristic curve and a saliency map were used to further evaluate the diagnostic performance. Results In prospective unseen test data, the diagnostic performance of the automatically interpretable otosclerosis detection system at the optimal threshold was 0.97 and 0.98 for sensitivity and specificity, respectively. In comparison with the clinical acumen of otolaryngologists at P < 0.05, the proposed system was not significantly different. Moreover, the area under the receiver operating characteristic curve for the proposed system was 0.99, indicating satisfactory diagnostic accuracy. Conclusion Our research develops and evaluates a deep learning system that detects otosclerosis at a level comparable with clinical otolaryngologists. Our system is an effective schema for the differential diagnosis of otosclerosis in computed tomography examinations.
Collapse
Affiliation(s)
- Zheng Wang
- School of Computer Science, Hunan First Normal University, Changsha, 410205, China
- Key Laboratory of Informalization Technology for Basic Education in Hunan Province, Changsha, 410205, China
| | - Jian Song
- Department of Otorhinolaryngology, Xiangya Hospital Central South University, Changsha, Hunan, China
- Province Key Laboratory of Otolaryngology Critical Diseases, Changsha, Hunan, China
| | - Kaibin Lin
- School of Computer Science, Hunan First Normal University, Changsha, 410205, China
- Key Laboratory of Informalization Technology for Basic Education in Hunan Province, Changsha, 410205, China
| | - Wei Hong
- School of Computer Science, Hunan First Normal University, Changsha, 410205, China
- Key Laboratory of Informalization Technology for Basic Education in Hunan Province, Changsha, 410205, China
| | - Shuang Mao
- Department of Otorhinolaryngology, Xiangya Hospital Central South University, Changsha, Hunan, China
- Province Key Laboratory of Otolaryngology Critical Diseases, Changsha, Hunan, China
| | - Xuewen Wu
- Department of Otorhinolaryngology, Xiangya Hospital Central South University, Changsha, Hunan, China
- Province Key Laboratory of Otolaryngology Critical Diseases, Changsha, Hunan, China
| | - Jianglin Zhang
- Department of Dermatology, Shenzhen People's Hospital, The Second Clinical Medical College, Jinan University. The First Affiliated Hospital, Southern University of Science and Technology, Shenzhen, 518020, Guangdong, China
- Candidate Branch of National Clinical Research Center for Skin Diseases, Shenzhen, 518020, Guangdong, China
- Department of Geriatrics, Shenzhen People's Hospital, The Second Clinical Medical College, Jinan University. The First Affiliated Hospital, Southern University of Science and Technology, Shenzhen, 518020, Guangdong, China
| |
Collapse
|
47
|
Tang Z, Mahmoodi S, Meng D, Darekar A, Vollmer B. Rule-based deep learning method for prognosis of neonatal hypoxic-ischemic encephalopathy by using susceptibility weighted image analysis. MAGMA (NEW YORK, N.Y.) 2024; 37:227-239. [PMID: 38252196 DOI: 10.1007/s10334-023-01139-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/09/2023] [Revised: 12/06/2023] [Accepted: 12/11/2023] [Indexed: 01/23/2024]
Abstract
OBJECTIVE Susceptibility weighted imaging (SWI) of neonatal hypoxic-ischemic brain injury can provide assistance in the prognosis of neonatal hypoxic-ischemic encephalopathy (HIE). We propose a convolutional neural network model to classify SWI images with HIE. MATERIALS AND METHODS Due to the lack of a large dataset, transfer learning method with fine-tuning a pre-trained ResNet 50 is introduced. We randomly select 11 datasets from patients with normal neurology outcomes (n = 31) and patients with abnormal neurology outcomes (n = 11) at 24 months of age to avoid bias in classification due to any imbalance in the data. RESULTS We develop a rule-based system to improve the classification performance, with an accuracy of 0.93 ± 0.09. We also compute heatmaps produced by the Grad-CAM technique to analyze which areas of SWI images contributed more to the classification patients with abnormal neurology outcome. CONCLUSION Such regions that are important in the classification accuracy can interpret the relationship between the brain regions affected by hypoxic-ischemic and neurodevelopmental outcomes of infants with HIE at the age of 2 years.
Collapse
Affiliation(s)
- Zhen Tang
- School of Computer Science and Technology, AnHui University of Technology, Maxiang Street, Maanshan, 243032, Anhui, China.
| | - Sasan Mahmoodi
- School of Electronics and Computer Science, University of Southampton, Southampton, SO17 1BJ, UK
| | - Di Meng
- School of Computer Science and Technology, AnHui University of Technology, Maxiang Street, Maanshan, 243032, Anhui, China
| | - Angela Darekar
- Department of Medical Physics, University Hospital Southampton NHS Foundation Trust, Southampton, SO16 6YD, UK
| | - Brigitte Vollmer
- Clinical Neurosciences and Clinical and Experimental Sciences, Faculty of Medicine, University of Southampton, Southampton, SO17 1BJ, UK
| |
Collapse
|
48
|
Africano G, Arponen O, Rinta-Kiikka I, Pertuz S. Transfer learning for the generalization of artificial intelligence in breast cancer detection: a case-control study. Acta Radiol 2024; 65:334-340. [PMID: 38115699 DOI: 10.1177/02841851231218960] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2023]
Abstract
BACKGROUND Some researchers have questioned whether artificial intelligence (AI) systems maintain their performance when used for women from populations not considered during the development of the system. PURPOSE To evaluate the impact of transfer learning as a way of improving the generalization of AI systems in the detection of breast cancer. MATERIAL AND METHODS This retrospective case-control Finnish study involved 191 women diagnosed with breast cancer and 191 matched healthy controls. We selected a state-of-the-art AI system for breast cancer detection trained using a large US dataset. The selected baseline system was evaluated in two experimental settings. First, we examined our private Finnish sample as an independent test set that had not been considered in the development of the system (unseen population). Second, the baseline system was retrained to attempt to improve its performance in the unseen population by means of transfer learning. To analyze performance, we used areas under the receiver operating characteristic curve (AUCs) with DeLong's test. RESULTS Two versions of the baseline system were considered: ImageOnly and Heatmaps. The ImageOnly and Heatmaps versions yielded mean AUC values of 0.82±0.008 and 0.88±0.003 in the US dataset and 0.56 (95% CI=0.50-0.62) and 0.72 (95% CI=0.67-0.77) when evaluated in the unseen population, respectively. The retrained systems achieved AUC values of 0.61 (95% CI=0.55-0.66) and 0.69 (95% CI=0.64-0.75), respectively. There was no statistical difference between the baseline system and the retrained system. CONCLUSION Transfer learning with a small study sample did not yield a significant improvement in the generalization of the system.
Collapse
Affiliation(s)
- Gerson Africano
- School of Electrical, Electronics and Telecommunications Engineering, Universidad Industrial de Santander, Bucaramanga, Colombia
| | - Otso Arponen
- Department of Radiology, Tampere University Hospital, Tampere, Finland
- Faculty of Medicine and Health Technology, Tampere University, Tampere, Finland
| | - Irina Rinta-Kiikka
- Department of Radiology, Tampere University Hospital, Tampere, Finland
- Faculty of Medicine and Health Technology, Tampere University, Tampere, Finland
| | - Said Pertuz
- School of Electrical, Electronics and Telecommunications Engineering, Universidad Industrial de Santander, Bucaramanga, Colombia
| |
Collapse
|
49
|
Yang J, Lü J, Qiu Z, Zhang M, Yan H. Risk prediction of pulse wave for hypertensive target organ damage based on frequency-domain feature map. Med Eng Phys 2024; 126:104161. [PMID: 38621841 DOI: 10.1016/j.medengphy.2024.104161] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 02/29/2024] [Accepted: 03/27/2024] [Indexed: 04/17/2024]
Abstract
The application of deep learning to the classification of pulse waves in Traditional Chinese Medicine (TCM) related to hypertensive target organ damage (TOD) is hindered by challenges such as low classification accuracy and inadequate generalization performance. To address these challenges, we introduce a lightweight transfer learning model named MobileNetV2SCP. This model transforms time-domain pulse waves into 36-dimensional frequency-domain waveform feature maps and establishes a dedicated pre-training network based on these maps to enhance the learning capability for small samples. To improve global feature correlation, we incorporate a novel fusion attention mechanism (SAS) into the inverted residual structure, along with the utilization of 3 × 3 convolutional layers and BatchNorm layers to mitigate model overfitting. The proposed model is evaluated using cross-validation results from 805 cases of pulse waves associated with hypertensive TOD. The assessment metrics, including Accuracy (92.74 %), F1-score (91.47 %), and Area Under Curve (AUC) (97.12 %), demonstrate superior classification accuracy and generalization performance compared to various state-of-the-art models. Furthermore, this study investigates the correlations between time-domain and frequency-domain features in pulse waves and their classification in hypertensive TOD. It analyzes key factors influencing pulse wave classification, providing valuable insights for the clinical diagnosis of TOD.
Collapse
Affiliation(s)
- Jingdong Yang
- Autonomous Robot Lab, School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China.
| | - Jiangtao Lü
- Autonomous Robot Lab, School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Zehao Qiu
- Autonomous Robot Lab, School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Mengchu Zhang
- Shanghai Key Laboratory of Health Identification and Assessment, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China
| | - Haixia Yan
- Shanghai Key Laboratory of Health Identification and Assessment, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China.
| |
Collapse
|
50
|
Iwata H, Hayashi Y, Koyama T, Hasegawa A, Ohgi K, Kobayashi I, Okuno Y. Feature extraction of particle morphologies of pharmaceutical excipients from scanning electron microscope images using convolutional neural networks. Int J Pharm 2024; 653:123873. [PMID: 38336179 DOI: 10.1016/j.ijpharm.2024.123873] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2023] [Revised: 01/08/2024] [Accepted: 01/30/2024] [Indexed: 02/12/2024]
Abstract
Scanning electron microscopy (SEM) images are the most widely used tool for evaluating particle morphology; however, quantitative evaluation using SEM images is time-consuming and often neglected. In this study, we aimed to extract features related to particle morphology of pharmaceutical excipients from SEM images using a convolutional neural network (CNN). SEM images of 67 excipients were acquired and used as models. A classification CNN model of the excipients was constructed based on the SEM images. Further, features were extracted from the middle layer of this CNN model, and the data was compressed to two dimensions using uniform manifold approximation and projection. Lastly, hierarchical clustering analysis (HCA) was performed to categorize the excipients into several clusters and identify similarities among the samples. The classification CNN model showed high accuracy, allowing each excipient to be identified with a high degree of accuracy. HCA revealed that the 67 excipients were classified into seven clusters. Additionally, the particle morphologies of excipients belonging to the same cluster were found to be very similar. These results suggest that CNN models are useful tools for extracting information and identifying similarities among the particle morphologies of excipients.
Collapse
Affiliation(s)
- Hiroaki Iwata
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan.
| | - Yoshihiro Hayashi
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan; Pharmaceutical Technology Management Department, Production Division, Nichi-Iko Pharmaceutical Co., Ltd., 205-1, Shimoumezawa Namerikawa-shi, Toyama 936-0857, Japan.
| | - Takuto Koyama
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan
| | - Aki Hasegawa
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan
| | - Kosuke Ohgi
- Formulation Development Department, Development & Planning Division, Nichi-Iko Pharmaceutical Co., Ltd., 205-1, Shimoumezawa Namerikawa-shi, Toyama 936-0857, Japan
| | - Ippei Kobayashi
- Formulation Development Department, Development & Planning Division, Nichi-Iko Pharmaceutical Co., Ltd., 205-1, Shimoumezawa Namerikawa-shi, Toyama 936-0857, Japan
| | - Yasushi Okuno
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan; RIKEN Center for Computational Science, Kobe 650-0047, Japan
| |
Collapse
|