1
|
Liu J, Bhadra S, Shafaat O, Mukherjee P, Parnell C, Summers RM. A unified approach to medical image segmentation by leveraging mixed supervision and self and transfer learning (MIST). Comput Med Imaging Graph 2025; 122:102517. [PMID: 40088573 PMCID: PMC12007390 DOI: 10.1016/j.compmedimag.2025.102517] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2024] [Revised: 01/15/2025] [Accepted: 02/22/2025] [Indexed: 03/17/2025]
Abstract
Medical image segmentation is important for quantitative disease diagnosis and treatment but relies on accurate pixel-wise labels, which are costly, time-consuming, and require domain expertise. This work introduces MIST (MIxed supervision, Self, and Transfer learning) to reduce manual labeling in medical image segmentation. A small set of cases was manually annotated ("strong labels"), while the rest used automated, less accurate labels ("weak labels"). Both label types trained a dual-branch network with a shared encoder and two decoders. Self-training iteratively refined weak labels, and transfer learning reduced computational costs by freezing the encoder and fine-tuning the decoders. Applied to segmenting muscle, subcutaneous, and visceral adipose tissue, MIST used only 100 manually labeled slices from 20 CT scans to generate accurate labels for all slices of 102 internal scans, which were then used to train a 3D nnU-Net model. Using MIST to update weak labels significantly improved nnU-Net segmentation accuracy compared to training directly on strong and weak labels. Dice similarity coefficient (DSC) increased for muscle (89.2 ± 4.3% to 93.2 ± 2.1%), subcutaneous (75.1 ± 14.4% to 94.2 ± 2.8%), and visceral adipose tissue (66.6 ± 16.4% to 77.1 ± 19.0% ) on an internal dataset (p<.05). DSC improved for muscle (80.5 ± 6.9% to 86.6 ± 3.9%) and subcutaneous adipose tissue (61.8 ± 12.5% to 82.7 ± 11.1%) on an external dataset (p<.05). MIST reduced the annotation burden by 99%, enabling efficient, accurate pixel-wise labeling for medical image segmentation. Code is available at https://github.com/rsummers11/NIH_CADLab_Body_Composition.
Collapse
Affiliation(s)
- Jianfei Liu
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA.
| | - Sayantan Bhadra
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| | - Omid Shafaat
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| | - Pritam Mukherjee
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| | - Christopher Parnell
- Walter Reed National Military Medical Center, 4494 Palmer Rd N, Bethesda, 20814, MD, USA
| | - Ronald M Summers
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| |
Collapse
|
2
|
Baxter JSH, Eagleson R. Exploring the values underlying machine learning research in medical image analysis. Med Image Anal 2025; 102:103494. [PMID: 40020419 DOI: 10.1016/j.media.2025.103494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 01/26/2025] [Accepted: 02/01/2025] [Indexed: 03/03/2025]
Abstract
Machine learning has emerged as a crucial tool for medical image analysis, largely due to recent developments in deep artificial neural networks addressing numerous, diverse clinical problems. As with any conceptual tool, the effective use of machine learning should be predicated on an understanding of its underlying motivations just as much as algorithms or theory - and to do so, we need to explore its philosophical foundations. One of these foundations is the understanding of how values, despite being non-empirical, nevertheless affect scientific research. This article has three goals: to introduce the reader to values in a way that is specific to medical image analysis; to characterise a particular set of technical decisions (what we call the end-to-end vs. separable learning spectrum) that are fundamental to machine learning for medical image analysis; and to create a simple and structured method to show how these values can be rigorously connected to these technical decisions. This better understanding of how the philosophy of science can clarify fundamental elements of how medical image analysis research is performed and can be improved.
Collapse
Affiliation(s)
- John S H Baxter
- Laboratoire Traitement du Signal et de l'Image (LTSI, INSERM UMR 1099), Université de Rennes, Rennes, France.
| | - Roy Eagleson
- Biomedical Engineering Graduate Program, Western University, London, Canada
| |
Collapse
|
3
|
Uramoto L, Hayashi Y, Oda M, Kitasaka T, Mori K. Semantic segmentation dataset authoring with simplified labels. Int J Comput Assist Radiol Surg 2025:10.1007/s11548-024-03314-9. [PMID: 40186718 DOI: 10.1007/s11548-024-03314-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Accepted: 12/18/2024] [Indexed: 04/07/2025]
Abstract
PURPOSE Semantic segmentation of laparoscopic images is a key problem in surgical scene understanding. Creating ground truth labels for semantic segmentation tasks is time consuming, and in the medical field a need for medical training of annotators adds further complications, leading to reliance on a small pool of experts. Previous research has focused on reducing the time to author datasets, by using spatially weak labels, pseudolabels, and synthetic data. In this paper, we address the difficulties caused by the need for medically trained annotators, hoping to enable non-medical annotators to participate in medical annotation tasks, to ease the creation of large datasets. METHODS We propose simplified labels, labels that are semantically weak. Our labels allow non-medical annotators to participate in medical dataset authoring, by lowering the need for medical expertise. We simulate authoring processes with mixtures of medical and non-medical annotators and measure the impact adding non-medical annotators has on accuracy. We also show that simplified labels offer a simple formulation for multi-dataset training. RESULTS We show that simplified labels are a viable approach to dataset authoring. Including non-medical annotators in the authoring process is beneficial, but medically trained annotators are worth multiple non-medical annotators, with maximal Dice score increases of 9.3% for 1 medically trained annotator and 6.9% for 3 non-medical annotators. We also show that the labels offer a simple formulation for multi-dataset training, even with no overlapping classes. We find that converting the labels of a secondary incompatible dataset into simplified labels and jointly training on both datasets improves performance. CONCLUSION Simplified labels offer a framework that can be applied both to dataset authoring and to multi-dataset training. Using the proposed method, non-medical annotators can participate in semantic segmentation dataset authoring. Labels of incompatible datasets can be converted into simplified datasets, enabling multi-dataset training.
Collapse
Affiliation(s)
- Leo Uramoto
- Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, Aichi, 464-8601, Japan.
| | - Yuichiro Hayashi
- Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, Aichi, 464-8601, Japan
| | - Masahiro Oda
- Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, Aichi, 464-8601, Japan
- Information Technology Center, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, Aichi, 464-8601, Japan
| | - Takayuki Kitasaka
- Aichi Institute of Technology, 1247 Yachigusa, Yakusa-cho, Toyota, Aichi, 470-0356, Japan
| | - Kensaku Mori
- Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, Aichi, 464-8601, Japan.
- Information Technology Center, Nagoya University, Furo-cho, Chikusa-ku, Nagoya, Aichi, 464-8601, Japan.
- Research Center of Medical Big Data, National Institute of Informatics, 2 Chome-1-2 Hitotsubashi, Chiyoda City, Tokyo, 101-0003, Japan.
| |
Collapse
|
4
|
Zhang H, Yang B, Li S, Zhang X, Li X, Liu T, Higashita R, Liu J. Retinal OCT image segmentation with deep learning: A review of advances, datasets, and evaluation metrics. Comput Med Imaging Graph 2025; 123:102539. [PMID: 40203494 DOI: 10.1016/j.compmedimag.2025.102539] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2024] [Revised: 03/07/2025] [Accepted: 03/22/2025] [Indexed: 04/11/2025]
Abstract
Optical coherence tomography (OCT) is a widely used imaging technology in ophthalmic clinical practice, providing non-invasive access to high-resolution retinal images. Segmentation of anatomical structures and pathological lesions in retinal OCT images, directly impacts clinical decisions. While commercial OCT devices segment multiple retinal layers in healthy eyes, their performance degrades severely under pathological conditions. In recent years, the rapid advancements in deep learning have significantly driven research in OCT image segmentation. This review provides a comprehensive overview of the latest developments in deep learning-based segmentation methods for retinal OCT images. Additionally, it summarizes the medical significance, publicly available datasets, and commonly used evaluation metrics in this field. The review also discusses the current challenges faced by the research community and highlights potential future directions.
Collapse
Affiliation(s)
- Huihong Zhang
- Harbin Institute of Technology, No. 92 West Dazhi Street, Nangang District, Harbin, 150001, Heilongjiang, China; Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Bing Yang
- Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Sanqian Li
- Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Xiaoqing Zhang
- Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Xiaoling Li
- Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Tianhang Liu
- Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Risa Higashita
- Research Institute of Trustworthy Autonomous Systems, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China
| | - Jiang Liu
- Department of Computer Science and Engineering, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China; Research Institute of Trustworthy Autonomous Systems, Southern University of Science and Technology, 1088 Xueyuan Avenue, Shenzhen, 518055, Guangdong, China; University of Nottingham Ningbo China, 199 Taikang East Road, 315100, Ningbo, China.
| |
Collapse
|
5
|
Shi G, Lu H, Hui H, Tian J. Benefit from public unlabeled data: A Frangi filter-based pretraining network for 3D cerebrovascular segmentation. Med Image Anal 2025; 101:103442. [PMID: 39837153 DOI: 10.1016/j.media.2024.103442] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2024] [Revised: 11/27/2024] [Accepted: 12/16/2024] [Indexed: 01/23/2025]
Abstract
Precise cerebrovascular segmentation in Time-of-Flight Magnetic Resonance Angiography (TOF-MRA) data is crucial for computer-aided clinical diagnosis. The sparse distribution of cerebrovascular structures within TOF-MRA images often results in high costs for manual data labeling. Leveraging unlabeled TOF-MRA data can significantly enhance model performance. In this study, we have constructed the largest preprocessed unlabeled TOF-MRA dataset to date, comprising 1510 subjects. Additionally, we provide manually annotated segmentation masks for 113 subjects based on existing external image datasets to facilitate evaluation. We propose a simple yet effective pretraining strategy utilizing the Frangi filter, known for its capability to enhance vessel-like structures, to optimize the use of the unlabeled data for 3D cerebrovascular segmentation. This involves a Frangi filter-based preprocessing workflow tailored for large-scale unlabeled datasets and a multi-task pretraining strategy to efficiently utilize the preprocessed data. This approach ensures maximal extraction of useful knowledge from the unlabeled data. The efficacy of the pretrained model is assessed across four cerebrovascular segmentation datasets, where it demonstrates superior performance, improving the clDice metric by approximately 2%-3% compared to the latest semi- and self-supervised methods. Additionally, ablation studies validate the generalizability and effectiveness of our pretraining method across various backbone structures. The code and data have been open source at: https://github.com/shigen-StoneRoot/FFPN.
Collapse
Affiliation(s)
- Gen Shi
- School of Engineering Medicine and School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Key Laboratory of Big DataBased Precision Medicine (Beihang University), Ministry of Industry and Information Technology of China, Beijing, 100191, China; CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China
| | - Hao Lu
- State Key Laboratory for Management and Control of Complex Systems, Institute of Automation, Chinese Academic of Science, Beijing 10086, China
| | - Hui Hui
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China; National Key Laboratory of Kidney Diseases, Beijing, 100853, China.
| | - Jie Tian
- School of Engineering Medicine and School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Key Laboratory of Big DataBased Precision Medicine (Beihang University), Ministry of Industry and Information Technology of China, Beijing, 100191, China; CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China; National Key Laboratory of Kidney Diseases, Beijing, 100853, China.
| |
Collapse
|
6
|
Xue J, Liu H, Jiang L, Yin Q, Chen L, Wang M. Limitations of nomogram models in predicting survival outcomes for glioma patients. Front Immunol 2025; 16:1547506. [PMID: 40170838 PMCID: PMC11959071 DOI: 10.3389/fimmu.2025.1547506] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2024] [Accepted: 02/28/2025] [Indexed: 04/03/2025] Open
Abstract
Purpose Glioma represents a prevalent and malignant tumor of the central nervous system (CNS), and it is essential to accurately predict the survival of glioma patients to optimize their subsequent treatment plans. This review outlines the most recent advancements and viewpoints regarding the application of nomograms in glioma prognosis research. Design With an emphasis on the precision and external applicability of predictive models, we carried out a comprehensive review of the literature on the application of nomograms in glioma and provided a step-by-step guide for developing and evaluating nomograms. Results A summary of thirty-nine articles was produced. The majority of nomogram-building research has used limited patient samples, disregarded the proportional hazards (PH) assumption in Cox regression models, and some of them have failed to incorporate external validation. Furthermore, the predictive capability of nomograms is influenced by the selection of incorporated risk factors. Overall, the current predictive accuracy of nomograms is moderately credible. Conclusion The development and validation of nomogram models ought to adhere to a standardized set of criteria, thereby augmenting their worth in clinical decision-making and clinician-patient communication. Prior to the clinical application of a nomogram, it is imperative to thoroughly scrutinize its statistical foundation, rigorously evaluate its accuracy, and, whenever feasible, assess its external applicability utilizing multicenter databases.
Collapse
Affiliation(s)
- Jihao Xue
- Department of Neurosurgery, The Affiliated Hospital, Southwest Medical University, Luzhou, Sichuan, China
| | - Hang Liu
- Department of Neurosurgery, The Affiliated Hospital, Southwest Medical University, Luzhou, Sichuan, China
| | - Lu Jiang
- Department of Neurosurgery, The Affiliated Hospital, Southwest Medical University, Luzhou, Sichuan, China
| | - Qijia Yin
- Department of Urology or Nursing, Dazhou First People’s Hospital, Dazhou, Sichuan, China
- College of Nursing, Chongqing Medical University, Chongqing, Chongqing, China
| | - Ligang Chen
- Department of Neurosurgery, The Affiliated Hospital, Southwest Medical University, Luzhou, Sichuan, China
- Academician (Expert) Workstation of Sichuan Province, The Affiliated Hospital, Southwest Medical University, Luzhou, China
- Neurological Diseases and Brain Function Laboratory, The Affiliated Hospital, Southwest Medical University, Luzhou, China
| | - Ming Wang
- Department of Neurosurgery, The Affiliated Hospital, Southwest Medical University, Luzhou, Sichuan, China
| |
Collapse
|
7
|
Li H, Wang Y, Qiang Y. A semi-supervised domain adaptive medical image segmentation method based on dual-level multi-scale alignment. Sci Rep 2025; 15:8784. [PMID: 40082549 PMCID: PMC11906615 DOI: 10.1038/s41598-025-93824-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2024] [Accepted: 03/10/2025] [Indexed: 03/16/2025] Open
Abstract
In the actual image segmentation tasks in the medical field, the phenomenon of limited labeled data accompanied by domain shifts often occurs and such domain shifts may exist in homologous or even heterologous data. In the study, a novel method was proposed to deal with this challenging phenomenon. Firstly, a model was trained with labeled data in source and target domains so as to adapt to unlabeled data. Then, the alignment at two main levels was realized. At the style level, based on multi-scale stylistic features, the alignment of unlabeled target images was maximized and unlabeled target image features were enhanced. At the inter-domain level, the similarity of the category centroids between target domain data and mixed image data was also maximized. Additionally, a fused supervised loss and alignment loss computation method was proposed. In validation experiments, two cross-domain medical image datasets were constructed: homologous and heterologous datasets. Experimental results showed that the proposed method had the more advantageous comprehensive performance than common semi-supervised and domain adaptation methods.
Collapse
Affiliation(s)
- Hualing Li
- School of Software, North University of China, Taiyuan, Shanxi, China.
| | - Yaodan Wang
- School of Software, North University of China, Taiyuan, Shanxi, China
| | - Yan Qiang
- School of Software, North University of China, Taiyuan, Shanxi, China
| |
Collapse
|
8
|
Mill L, Aust O, Ackermann JA, Burger P, Pascual M, Palumbo-Zerr K, Krönke G, Uderhardt S, Schett G, Clemen CS, Holtzhausen C, Jabari S, Schröder R, Maier A, Grüneboom A. Deep learning-based image analysis in muscle histopathology using photo-realistic synthetic data. COMMUNICATIONS MEDICINE 2025; 5:64. [PMID: 40050400 PMCID: PMC11885816 DOI: 10.1038/s43856-025-00777-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2024] [Accepted: 02/20/2025] [Indexed: 03/09/2025] Open
Abstract
BACKGROUND Artificial intelligence (AI), specifically Deep learning (DL), has revolutionized biomedical image analysis, but its efficacy is limited by the need for representative, high-quality large datasets with manual annotations. While latest research on synthetic data using AI-based generative models has shown promising results to tackle this problem, several challenges such as lack of interpretability and need for vast amounts of real data remain. This study aims to introduce a new approach-SYNTA-for the generation of photo-realistic synthetic biomedical image data to address the challenges associated with state-of-the art generative models and DL-based image analysis. METHODS The SYNTA method employs a fully parametric approach to create photo-realistic synthetic training datasets tailored to specific biomedical tasks. Its applicability is tested in the context of muscle histopathology and skeletal muscle analysis. This new approach is evaluated for two real-world datasets to validate its applicability to solve complex image analysis tasks on real data. RESULTS Here we show that SYNTA enables expert-level segmentation of unseen real-world biomedical data using only synthetic training data. By addressing the lack of representative and high-quality real-world training data, SYNTA achieves robust performance in muscle histopathology image analysis, offering a scalable, controllable and interpretable alternative to generative models such as Generative Adversarial Networks (GANs) or Diffusion Models. CONCLUSIONS SYNTA demonstrates great potential to accelerate and improve biomedical image analysis. Its ability to generate high-quality photo-realistic synthetic data reduces reliance on extensive collection of data and manual annotations, paving the way for advancements in histopathology and medical research.
Collapse
Affiliation(s)
- Leonid Mill
- MIRA Vision Microscopy GmbH, 73037, Göppingen, Germany.
- Pattern Recognition Lab, Friedrich-Alexander University Erlangen-Nürnberg (FAU), 91058, Erlangen, Germany.
| | - Oliver Aust
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Jochen A Ackermann
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Philipp Burger
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Monica Pascual
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Katrin Palumbo-Zerr
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Gerhard Krönke
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Stefan Uderhardt
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Georg Schett
- Department of Medicine 3 - Rheumatology and Immunology & Deutsches Zentrum für Immuntherapie, Friedrich-Alexander University Erlangen-Nürnberg (FAU) and Universitätsklinikum Erlangen, 91054, Erlangen, Germany
| | - Christoph S Clemen
- Institute of Aerospace Medicine, German Aerospace Center (DLR), Cologne, Germany
- Institute of Vegetative Physiology, Medical Faculty, University of Cologne, Cologne, Germany
| | - Christian Holtzhausen
- Department of Neuropathology, Universitätsklinikum Erlangen, Friedrich-Alexander University Erlangen-Nürnberg (FAU), 91054, Erlangen, Germany
| | - Samir Jabari
- Department of Neuropathology, Universitätsklinikum Erlangen, Friedrich-Alexander University Erlangen-Nürnberg (FAU), 91054, Erlangen, Germany
- Klinikum Nuremberg, Institute of Pathology, Paracelsus Medical University, 90419, Nuremberg, Germany
| | - Rolf Schröder
- Department of Neuropathology, Universitätsklinikum Erlangen, Friedrich-Alexander University Erlangen-Nürnberg (FAU), 91054, Erlangen, Germany
| | - Andreas Maier
- Pattern Recognition Lab, Friedrich-Alexander University Erlangen-Nürnberg (FAU), 91058, Erlangen, Germany
| | - Anika Grüneboom
- Leibniz-Institut für Analytische Wissenschaften - ISAS - e.V, 44139, Dortmund, Germany.
| |
Collapse
|
9
|
Tong MW, Zhou J, Akkaya Z, Majumdar S, Bhattacharjee R. Artificial intelligence in musculoskeletal applications: a primer for radiologists. Diagn Interv Radiol 2025; 31:89-101. [PMID: 39157958 PMCID: PMC11880867 DOI: 10.4274/dir.2024.242830] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2024] [Accepted: 07/11/2024] [Indexed: 08/20/2024]
Abstract
As an umbrella term, artificial intelligence (AI) covers machine learning and deep learning. This review aimed to elaborate on these terms to act as a primer for radiologists to learn more about the algorithms commonly used in musculoskeletal radiology. It also aimed to familiarize them with the common practices and issues in the use of AI in this domain.
Collapse
Affiliation(s)
- Michelle W. Tong
- University of California San Francisco Department of Radiology and Biomedical Imaging, San Francisco, USA
- University of California San Francisco Department of Bioengineering, San Francisco, USA
- University of California Berkeley Department of Bioengineering, Berkeley, USA
| | - Jiamin Zhou
- University of California San Francisco Department of Orthopaedic Surgery, San Francisco, USA
| | - Zehra Akkaya
- University of California San Francisco Department of Radiology and Biomedical Imaging, San Francisco, USA
- Ankara University Faculty of Medicine Department of Radiology, Ankara, Türkiye
| | - Sharmila Majumdar
- University of California San Francisco Department of Radiology and Biomedical Imaging, San Francisco, USA
- University of California San Francisco Department of Bioengineering, San Francisco, USA
| | - Rupsa Bhattacharjee
- University of California San Francisco Department of Radiology and Biomedical Imaging, San Francisco, USA
| |
Collapse
|
10
|
Zou C, Chen R, Wang B, Fei Q, Song H, Zang L. Development of a deep learning radiomics model combining lumbar CT, multi-sequence MRI, and clinical data to predict high-risk cage subsidence after lumbar fusion: a retrospective multicenter study. Biomed Eng Online 2025; 24:27. [PMID: 40025592 PMCID: PMC11872306 DOI: 10.1186/s12938-025-01355-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2024] [Accepted: 02/18/2025] [Indexed: 03/04/2025] Open
Abstract
BACKGROUND To develop and validate a model that integrates clinical data, deep learning radiomics, and radiomic features to predict high-risk patients for cage subsidence (CS) after lumbar fusion. METHODS This study analyzed preoperative CT and MRI data from 305 patients undergoing lumbar fusion surgery from three centers. Using a deep learning model based on 3D vision transformations, the data were divided the dataset into training (n = 214), validation (n = 61), and test (n = 30) groups. Feature selection was performed using LASSO regression, followed by the development of a logistic regression model. The predictive ability of the model was assessed using various machine learning algorithms, and a combined clinical model was also established. RESULTS Ultimately, 11 traditional radiomic features, 5 deep learning radiomic features, and 1 clinical feature were selected. The combined model demonstrated strong predictive performance, with area under the curve (AUC) values of 0.941, 0.832, and 0.935 for the training, validation, and test groups, respectively. Notably, our model outperformed predictions made by two experienced surgeons. CONCLUSIONS This study developed a robust predictive model that integrates clinical features and imaging data to identify high-risk patients for CS following lumbar fusion. This model has the potential to improve clinical decision-making and reduce the need for revision surgeries, easing the burden on healthcare systems.
Collapse
Affiliation(s)
- Congying Zou
- Department of Orthopedics, Beijing Chaoyang Hospital, Capital Medical University, 5 JingYuan Road, Shijingshan District, Beijing, 100043, China
| | - Ruiyuan Chen
- Department of Orthopedics, Beijing Chaoyang Hospital, Capital Medical University, 5 JingYuan Road, Shijingshan District, Beijing, 100043, China
| | - Baodong Wang
- Department of Orthopedics, Beijing Chaoyang Hospital, Capital Medical University, 5 JingYuan Road, Shijingshan District, Beijing, 100043, China
| | - Qi Fei
- Department of Orthopedics, Beijing Friendship Hospital, Capital Medical University, No 95, Yong'an Road, Xicheng District, Beijing, 100050, China
| | - Hongxing Song
- Department of Orthopedics, Beijing Shijitan Hospital, Capital Medical University, Beijing, 100038, China
| | - Lei Zang
- Department of Orthopedic Surgery, Beijing Chao-Yang Hospital, Capital Medical University, 8 Gong Ti Nan Lu, Chaoyang District, Beijing, 100020, China.
- Department of Orthopedics, Beijing Chaoyang Hospital, Capital Medical University, 5 JingYuan Road, Shijingshan District, Beijing, 100043, China.
| |
Collapse
|
11
|
Chung SL, Cheng CT, Liao CH, Chung IF. Patch-based feature mapping with generative adversarial networks for auxiliary hip fracture detection. Comput Biol Med 2025; 186:109627. [PMID: 39793347 DOI: 10.1016/j.compbiomed.2024.109627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Revised: 12/23/2024] [Accepted: 12/23/2024] [Indexed: 01/13/2025]
Abstract
BACKGROUND Hip fractures are a significant public health issue, particularly among the elderly population. Pelvic radiographs (PXRs) play a crucial role in diagnosing hip fractures and are commonly used for their evaluation. Previous research has demonstrated promising performance in classification models for hip fracture detection. However, these models sometimes focus on the images' non-fracture regions, reducing their explainability. This study applies weakly supervised learning techniques to address this issue and improve the model's focus on the fracture region. Additionally, we introduce a method to quantitatively evaluate the model's focus on the region of interest (ROI). METHODS We propose a new auxiliary module called the patch-auxiliary generative adversarial network (PAGAN) for weakly supervised learning tasks. PAGAN can be integrated with any state-of-the-art (SOTA) classification model, such as EfficientNetB0, ResNet50, and DenseNet121, to enhance hip fracture detection. This training strategy incorporates global information (the entire PXR image) and local information (the hip region patch) for more effective learning. Furthermore, we employ GradCAM to generate attention heatmaps, highlighting the focus areas within the classification model. The intersection over union (IOU) and dice coefficient (Dise) are then computed between the attention heatmap and the fracture area, enabling a quantitative assessment of the model's explainability. RESULTS AND CONCLUSIONS Incorporating PAGAN improved the performance of the classification models. The accuracy of EfficientNetB0 increased from 93.61 % to 95.97 %, ResNet50 improved from 90.66 % to 94.89 %, and DenseNet121 saw an increase from 93.51 % to 94.49 %. Regarding model explainability, the integration of PAGAN into classification models led to a more pronounced attention to ROI. The average IOU improved from 0.32 to 0.54 for EfficientNetB0, from 0.28 to 0.40 for ResNet50, and from 0.37 to 0.51 for DenseNet121. These results indicate that PAGAN improves hip fracture classification performance and substantially enhances the model's focus on the fracture region, thereby increasing its explainability.
Collapse
Affiliation(s)
- Shang-Lin Chung
- Institute of Biomedical Informatics, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Chi-Tung Cheng
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital, Linkou, Chang Gung University, Taoyuan, Taiwan
| | - Chien-Hung Liao
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital, Linkou, Chang Gung University, Taoyuan, Taiwan
| | - I-Fang Chung
- Institute of Biomedical Informatics, National Yang Ming Chiao Tung University, Taipei, Taiwan.
| |
Collapse
|
12
|
Yadav A, Welland S, Hoffman JM, Hyun J Kim G, Brown MS, Prosper AE, Aberle DR, McNitt-Gray MF, Hsu W. A comparative analysis of image harmonization techniques in mitigating differences in CT acquisition and reconstruction. Phys Med Biol 2025; 70:055015. [PMID: 39823753 DOI: 10.1088/1361-6560/adabad] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2024] [Accepted: 01/17/2025] [Indexed: 01/20/2025]
Abstract
Objective. The study aims to systematically characterize the effect of CT parameter variations on images and lung radiomic and deep features, and to evaluate the ability of different image harmonization methods to mitigate the observed variations.Approach. A retrospective in-house sinogram dataset of 100 low-dose chest CT scans was reconstructed by varying radiation dose (100%, 25%, 10%) and reconstruction kernels (smooth, medium, sharp). A set of image processing, convolutional neural network (CNNs), and generative adversarial network-based (GANs) methods were trained to harmonize all image conditions to a reference condition (100% dose, medium kernel). Harmonized scans were evaluated for image similarity using peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM), and learned perceptual image patch similarity (LPIPS), and for the reproducibility of radiomic and deep features using concordance correlation coefficient (CCC).Main Results. CNNs consistently yielded higher image similarity metrics amongst others; for Sharp/10%, which exhibited the poorest visual similarity, PSNR increased from a mean ± CI of 17.763 ± 0.492 to 31.925 ± 0.571, SSIM from 0.219 ± 0.009 to 0.754 ± 0.017, and LPIPS decreased from 0.490 ± 0.005 to 0.275 ± 0.016. Texture-based radiomic features exhibited a greater degree of variability across conditions, i.e. a CCC of 0.500 ± 0.332, compared to intensity-based features (0.972 ± 0.045). GANs achieved the highest CCC (0.969 ± 0.009 for radiomic and 0.841 ± 0.070 for deep features) amongst others. CNNs are suitable if downstream applications necessitate visual interpretation of images, whereas GANs are better alternatives for generating reproducible quantitative image features needed for machine learning applications.Significance. Understanding the efficacy of harmonization in addressing multi-parameter variability is crucial for optimizing diagnostic accuracy and a critical step toward building generalizable models suitable for clinical use.
Collapse
Affiliation(s)
- Anil Yadav
- Department of Bioengineering, Samueli School of Engineering, University of California, Los Angeles, CA 90095, United States of America
- Medical & Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - Spencer Welland
- Center for Computer Vision and Imaging Biomarkers, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - John M Hoffman
- Center for Computer Vision and Imaging Biomarkers, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - Grace Hyun J Kim
- Center for Computer Vision and Imaging Biomarkers, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - Matthew S Brown
- Center for Computer Vision and Imaging Biomarkers, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - Ashley E Prosper
- Medical & Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - Denise R Aberle
- Department of Bioengineering, Samueli School of Engineering, University of California, Los Angeles, CA 90095, United States of America
- Medical & Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - Michael F McNitt-Gray
- Center for Computer Vision and Imaging Biomarkers, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| | - William Hsu
- Department of Bioengineering, Samueli School of Engineering, University of California, Los Angeles, CA 90095, United States of America
- Medical & Imaging Informatics Group, Department of Radiological Sciences, David Geffen School of Medicine at UCLA, Los Angeles, CA 90095, United States of America
| |
Collapse
|
13
|
Guo YX, Lan JL, Bu WQ, Tang Y, Wu D, Yang H, Ren JC, Song YX, Yue HY, Guo YC, Meng HT. Automatic maxillary sinus segmentation and age estimation model for the northwestern Chinese Han population. BMC Oral Health 2025; 25:310. [PMID: 40011898 DOI: 10.1186/s12903-025-05618-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2024] [Accepted: 02/10/2025] [Indexed: 02/28/2025] Open
Abstract
BACKGROUND Age estimation is vital in forensic science, with maxillary sinus development serving as a reliable indicator. This study developed an automatic segmentation model for maxillary sinus identification and parameter measurement, combined with regression and machine learning models for age estimation. METHODS Cone Beam Computed Tomography (CBCT) images from 292 Han individuals (ranging from 5 to 53 years) were used to train and validate the segmentation model. Measurements included sinus dimensions (length, width, height), inter-sinus distance, and volume. Age estimation models using multiple linear regression and random forest algorithms were built based on these variables. RESULTS The automatic segmentation model achieved high accuracy, which yielded a Dice similarity coefficient (DSC) of 0.873, an Intersection over Union (IoU) of 0.7753, a Hausdorff Distance 95% (HD95) of 9.8337, and an Average Surface Distance (ASD) of 2.4507. The regression model performed best, with mean absolute errors (MAE) of 1.45 years (under 18) and 3.51 years (aged 18 and above), providing relatively precise age predictions. CONCLUSION The maxillary sinus-based model is a promising tool for age estimation, particularly in adults, and could be enhanced by incorporating additional variables like dental dimensions.
Collapse
Affiliation(s)
- Yu-Xin Guo
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Jun-Long Lan
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Wen-Qing Bu
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Yu Tang
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Di Wu
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Hui Yang
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Jia-Chen Ren
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Yu-Xuan Song
- College of Forensic Science, Xi'an Jiaotong University, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Hong-Ying Yue
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Yu-Cheng Guo
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China
| | - Hao-Tian Meng
- Key Laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China.
- Department of Orthodontics, College of Stomatology, Xi'an Jiaotong University, 98 XiWu Road, Xi'an, 710004, Shaanxi, People's Republic of China.
| |
Collapse
|
14
|
Johnson ET, Bande JK, Thomas J. Retrieval Augmented Medical Diagnosis System. Biol Methods Protoc 2025; 10:bpaf017. [PMID: 40078867 PMCID: PMC11897588 DOI: 10.1093/biomethods/bpaf017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2025] [Revised: 02/14/2025] [Accepted: 02/21/2025] [Indexed: 03/14/2025] Open
Abstract
Subjective variability in human interpretation of diagnostic imaging presents significant clinical limitations, potentially resulting in diagnostic errors and increased healthcare costs. While artificial intelligence (AI) algorithms offer promising solutions to reduce interpreter subjectivity, they frequently demonstrate poor generalizability across different healthcare settings. To address these issues, we introduce Retrieval Augmented Medical Diagnosis System (RAMDS), which integrates an AI classification model with a similar image model. This approach retrieves historical cases and their diagnoses to provide context for the AI predictions. By weighing similar image diagnoses alongside AI predictions, RAMDS produces a final weighted prediction, aiding physicians in understanding the diagnosis process. Moreover, RAMDS does not require complete retraining when applied to new datasets; rather, it simply necessitates re-calibration of the weighing system. When RAMDS fine-tuned for negative predictive value was evaluated on breast ultrasounds for cancer classification, RAMDS improved sensitivity by 21% and negative predictive value by 9% compared to ResNet-34. Offering enhanced metrics, explainability, and adaptability, RAMDS represents a notable advancement in medical AI. RAMDS is a new approach in medical AI that has the potential for pan-pathological uses, though further research is needed to optimize its performance and integrate multimodal data.
Collapse
Affiliation(s)
- Ethan Thomas Johnson
- Central High School, 423 E Central St, Springfield, Missouri, 65802, United States
| | - Jathin Koushal Bande
- Central High School, 423 E Central St, Springfield, Missouri, 65802, United States
| | - Johnson Thomas
- Department of Endocrinology, Mercy Hospital, Springfield, Missouri, 65807, United States
| |
Collapse
|
15
|
Abbas Y, Hadi HJ, Aziz K, Ahmed N, Akhtar MU, Alshara MA, Chakrabarti P. Reinforcement-based leveraging transfer learning for multiclass optical coherence tomography images classification. Sci Rep 2025; 15:6193. [PMID: 39979354 PMCID: PMC11842753 DOI: 10.1038/s41598-025-89831-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2024] [Accepted: 02/07/2025] [Indexed: 02/22/2025] Open
Abstract
The accurate diagnosis of retinal diseases, such as Diabetic Macular Edema (DME) and Age-related Macular Degeneration (AMD), is essential for preventing vision loss. Optical Coherence Tomography (OCT) imaging plays a crucial role in identifying these conditions, especially given the increasing prevalence of AMD. This study introduces a novel Reinforcement-Based Leveraging Transfer Learning (RBLTL) framework, which integrates reinforcement Q-learning with transfer learning using pre-trained models, including InceptionV3, DenseNet201, and InceptionResNetV2. The RBLTL framework dynamically optimizes hyperparameters, improving classification accuracy and generalization while mitigating overfitting. Experimental evaluations demonstrate remarkable performance, achieving testing accuracies of 98.75%, 98.90%, and 99.20% across three scenarios for multiclass OCT image classification. These results highlight the effectiveness of the RBLTL framework in categorizing OCT images for conditions like DME and AMD, establishing it as a reliable and versatile approach for automated medical image classification with significant implications for clinical diagnostics.
Collapse
Affiliation(s)
- Yawar Abbas
- Key Laboratory of Aerospace Information Security and Trusted Computing, Ministry of Education, School of Cyber Science and Engineering, Wuhan University, Wuhan, China.
| | - Hassan Jalil Hadi
- College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
| | - Kamran Aziz
- Key Laboratory of Aerospace Information Security and Trusted Computing, Ministry of Education, School of Cyber Science and Engineering, Wuhan University, Wuhan, China.
| | - Naveed Ahmed
- College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
| | | | - Mohammed Ali Alshara
- College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
| | - Prasun Chakrabarti
- Department of Computer Science and Engineering, Sir Padampat Singhania University, Udaipur, Rajasthan, 313601, India
| |
Collapse
|
16
|
Wen X, Tu H, Zhao B, Zhou W, Yang Z, Li L. Identification of benign and malignant breast nodules on ultrasound: comparison of multiple deep learning models and model interpretation. Front Oncol 2025; 15:1517278. [PMID: 40040727 PMCID: PMC11876547 DOI: 10.3389/fonc.2025.1517278] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2024] [Accepted: 01/30/2025] [Indexed: 03/06/2025] Open
Abstract
Background and Purpose Deep learning (DL) algorithms generally require full supervision of annotating the region of interest (ROI), a process that is both labor-intensive and susceptible to bias. We aimed to develop a weakly supervised algorithm to differentiate between benign and malignant breast tumors in ultrasound images without image annotation. Methods We developed and validated the models using two publicly available datasets: breast ultrasound image (BUSI) and GDPH&SYSUCC breast ultrasound datasets. After removing the poor quality images, a total of 3049 images were included, divided into two classes: benign (N = 1320 images) and malignant (N = 1729 images). Weakly-supervised DL algorithms were implemented with four networks (DenseNet121, ResNet50, EffientNetb0, and Vision Transformer) and trained using 2136 unannotated breast ultrasound images. 609 and 304 images were used for validation and test sets, respectively. Diagnostic performances were calculated as the area under the receiver operating characteristic curve (AUC). Using the class activation map to interpret the prediction results of weakly supervised DL algorithms. Results The DenseNet121 model, utilizing complete image inputs without ROI annotations, demonstrated superior diagnostic performance in distinguishing between benign and malignant breast nodules when compared to ResNet50, EfficientNetb0, and Vision Transformer models. DenseNet121 achieved the highest AUC, with values of 0.94 on the validation set and 0.93 on the test set, significantly surpassing the performance of the other models across both datasets (all P < 0.05). Conclusion The weakly supervised DenseNet121 model developed in this study demonstrated feasibility for ultrasound diagnosis of breast tumor and showed good capabilities in differential diagnosis. This model may help radiologists, especially novice doctors, to improve the accuracy of breast tumor diagnosis using ultrasound.
Collapse
Affiliation(s)
- Xi Wen
- Department of Ultrasound, The Central Hospital of Enshi Tujia And Miao Autonomous Prefecture (Enshi Clinical College of Wuhan University), Enshi, China
| | - Hao Tu
- Department of Ultrasound, The Central Hospital of Enshi Tujia And Miao Autonomous Prefecture (Enshi Clinical College of Wuhan University), Enshi, China
| | - Bingyang Zhao
- Department of Neurology, China-Japan Union Hospital of Jilin University, Changchun, China
| | - Wenbo Zhou
- Department of Stomatology, China-Japan Union Hospital of Jilin University, Changchun, China
| | - Zhuo Yang
- Department of Ultrasound, The Central Hospital of Enshi Tujia And Miao Autonomous Prefecture (Enshi Clinical College of Wuhan University), Enshi, China
| | - Lijuan Li
- Department of Ultrasound, The Central Hospital of Enshi Tujia And Miao Autonomous Prefecture (Enshi Clinical College of Wuhan University), Enshi, China
| |
Collapse
|
17
|
Zhou P, Zhou Q, Xiao X, Fan X, Zou Y, Sun L, Jiang J, Song D, Chen L. Machine Learning in Solid-State Hydrogen Storage Materials: Challenges and Perspectives. ADVANCED MATERIALS (DEERFIELD BEACH, FLA.) 2025; 37:e2413430. [PMID: 39703108 DOI: 10.1002/adma.202413430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/07/2024] [Revised: 11/10/2024] [Indexed: 12/21/2024]
Abstract
Machine learning (ML) has emerged as a pioneering tool in advancing the research application of high-performance solid-state hydrogen storage materials (HSMs). This review summarizes the state-of-the-art research of ML in resolving crucial issues such as low hydrogen storage capacity and unfavorable de-/hydrogenation cycling conditions. First, the datasets, feature descriptors, and prevalent ML models tailored for HSMs are described. Specific examples include the successful application of ML in titanium-based, rare-earth-based, solid solution, magnesium-based, and complex HSMs, showcasing its role in exploiting composition-structure-property relationships and designing novel HSMs for specific applications. One of the representative ML works is the single-phase Ti-based HSM with superior cost-effective and comprehensive properties, tailored to fuel cell hydrogen feeding system at ambient temperature and pressure through high-throughput composition-performance scanning. More importantly, this review also identifies and critically analyzes the key challenges faced by ML in this domain, including poor data quality and availability, and the balance between model interpretability and accuracy, together with feasible countermeasures suggested to ameliorate these problems. In summary, this work outlines a roadmap for enhancing ML's utilization in solid-state hydrogen storage research, promoting more efficient and sustainable energy storage solutions.
Collapse
Affiliation(s)
- Panpan Zhou
- College of Materials Science and Engineering, Hohai University, Changzhou, Jiangsu, 213200, China
- State Key Laboratory of Silicon and Advanced Semiconductor Materials, School of Materials Science and Engineering, Zhejiang University, Hangzhou, Zhejiang, 310058, China
| | - Qianwen Zhou
- State Key Laboratory of Silicon and Advanced Semiconductor Materials, School of Materials Science and Engineering, Zhejiang University, Hangzhou, Zhejiang, 310058, China
| | - Xuezhang Xiao
- State Key Laboratory of Silicon and Advanced Semiconductor Materials, School of Materials Science and Engineering, Zhejiang University, Hangzhou, Zhejiang, 310058, China
- School of Advanced Energy, Sun Yat-Sen University, Shenzhen, 518107, China
| | - Xiulin Fan
- State Key Laboratory of Silicon and Advanced Semiconductor Materials, School of Materials Science and Engineering, Zhejiang University, Hangzhou, Zhejiang, 310058, China
| | - Yongjin Zou
- Guangxi Key Laboratory of Information Materials, Guilin University of Electronic Technology, Guilin, 541004, China
| | - Lixian Sun
- Guangxi Key Laboratory of Information Materials, Guilin University of Electronic Technology, Guilin, 541004, China
| | - Jinghua Jiang
- College of Materials Science and Engineering, Hohai University, Changzhou, Jiangsu, 213200, China
| | - Dan Song
- College of Materials Science and Engineering, Hohai University, Changzhou, Jiangsu, 213200, China
| | - Lixin Chen
- State Key Laboratory of Silicon and Advanced Semiconductor Materials, School of Materials Science and Engineering, Zhejiang University, Hangzhou, Zhejiang, 310058, China
- Key Laboratory of Hydrogen Storage and Transportation Technology of Zhejiang Province, Hangzhou, Zhejiang, 310027, China
| |
Collapse
|
18
|
Lian W, Lindblad J, Runow Stark C, Hirsch JM, Sladoje N. Let it shine: Autofluorescence of Papanicolaou-stain improves AI-based cytological oral cancer detection. Comput Biol Med 2025; 185:109498. [PMID: 39662319 DOI: 10.1016/j.compbiomed.2024.109498] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2024] [Revised: 10/27/2024] [Accepted: 11/26/2024] [Indexed: 12/13/2024]
Abstract
BACKGROUND AND OBJECTIVES Oral cancer is a global health challenge. The disease can be successfully treated if detected early, but the survival rate drops significantly for late stage cases. There is a growing interest in a shift from the current standard of invasive and time-consuming tissue sampling and histological examination, towards non-invasive brush biopsies and cytological examination, facilitating continued risk group monitoring. For cost effective and accurate cytological analysis there is a great need for reliable computer-assisted data-driven approaches. However, infeasibility of accurate cell-level annotation hinders model performance, and limits evaluation and interpretation of the results. This study aims to improve AI-based oral cancer detection by introducing additional information through multimodal imaging and deep multimodal information fusion. METHODS We combine brightfield and fluorescence whole slide microscopy imaging to analyze Papanicolaou-stained liquid-based cytology slides of brush biopsies collected from both healthy and cancer patients. Given the challenge of detailed cytological annotations, we utilize a weakly supervised deep learning approach only relying on patient-level labels. We evaluate various multimodal information fusion strategies, including early, late, and three recent intermediate fusion methods. RESULTS Our experiments demonstrate that: (i) there is substantial diagnostic information to gain from fluorescence imaging of Papanicolaou-stained cytological samples, (ii) multimodal information fusion improves classification performance and cancer detection accuracy, compared to single-modality approaches. Intermediate fusion emerges as the leading method among the studied approaches. Specifically, the Co-Attention Fusion Network (CAFNet) model achieves impressive results, with an F1 score of 83.34% and an accuracy of 91.79% at cell level, surpassing human performance on the task. Additional tests highlight the importance of accurate image registration to maximize the benefits of the multimodal analysis. CONCLUSION This study advances the field of cytopathology by integrating deep learning methods, multimodal imaging and information fusion to enhance non-invasive early detection of oral cancer. Our approach not only improves diagnostic accuracy, but also allows an efficient, yet uncomplicated, clinical workflow. The developed pipeline has potential applications in other cytological analysis settings. We provide a validated open-source analysis framework and share a unique multimodal oral cancer dataset to support further research and innovation.
Collapse
Affiliation(s)
- Wenyi Lian
- Centre for Image Analysis, Department of Information Technology, Uppsala University, Uppsala, Sweden.
| | - Joakim Lindblad
- Centre for Image Analysis, Department of Information Technology, Uppsala University, Uppsala, Sweden.
| | - Christina Runow Stark
- Department of Surgical Sciences, Uppsala University, Uppsala, Sweden; Folktandvården, Region Uppsala, Uppsala, Sweden
| | - Jan-Michaél Hirsch
- Department of Surgical Sciences, Uppsala University, Uppsala, Sweden; Folktandvården Stockholms län AB, Region Stockholm, Stockholm, Sweden
| | - Nataša Sladoje
- Centre for Image Analysis, Department of Information Technology, Uppsala University, Uppsala, Sweden.
| |
Collapse
|
19
|
Wu Y, Zhou Y, Saiyin J, Wei B, Lai M, Shou J, Xu Y. AttriPrompter: Auto-Prompting With Attribute Semantics for Zero-Shot Nuclei Detection via Visual-Language Pre-Trained Models. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:982-993. [PMID: 39361456 DOI: 10.1109/tmi.2024.3473745] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/05/2024]
Abstract
Large-scale visual-language pre-trained models (VLPMs) have demonstrated exceptional performance in downstream object detection through text prompts for natural scenes. However, their application to zero-shot nuclei detection on histopathology images remains relatively unexplored, mainly due to the significant gap between the characteristics of medical images and the web-originated text-image pairs used for pre-training. This paper aims to investigate the potential of the object-level VLPM, Grounded Language-Image Pre-training (GLIP), for zero-shot nuclei detection. Specifically, we propose an innovative auto-prompting pipeline, named AttriPrompter, comprising attribute generation, attribute augmentation, and relevance sorting, to avoid subjective manual prompt design. AttriPrompter utilizes VLPMs' text-to-image alignment to create semantically rich text prompts, which are then fed into GLIP for initial zero-shot nuclei detection. Additionally, we propose a self-trained knowledge distillation framework, where GLIP serves as the teacher with its initial predictions used as pseudo labels, to address the challenges posed by high nuclei density, including missed detections, false positives, and overlapping instances. Our method exhibits remarkable performance in label-free nuclei detection, outperforming all existing unsupervised methods and demonstrating excellent generality. Notably, this work highlights the astonishing potential of VLPMs pre-trained on natural image-text pairs for downstream tasks in the medical field as well. Code will be released at github.com/AttriPrompter.
Collapse
|
20
|
Duta I, Kariuki SM, Ngugi AK, Mwesige AK, Masanja H, Mwanga DM, Owusu-Agyei S, Wagner R, Cross JH, Sander JW, Newton CR, Sen A, Jones GD. Evaluating the generalisability of region-naïve machine learning algorithms for the identification of epilepsy in low-resource settings. PLOS DIGITAL HEALTH 2025; 4:e0000491. [PMID: 39937713 PMCID: PMC11819582 DOI: 10.1371/journal.pdig.0000491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/26/2024] [Accepted: 12/02/2024] [Indexed: 02/14/2025]
Abstract
OBJECTIVES Approximately 80% of people with epilepsy live in low- and middle-income countries (LMICs), where limited resources and stigma hinder accurate diagnosis and treatment. Clinical machine learning models have demonstrated substantial promise in supporting the diagnostic process in LMICs by aiding in preliminary screening and detection of possible epilepsy cases without relying on specialised or trained personnel. How well these models generalise to naïve regions is, however, underexplored. Here, we use a novel approach to assess the suitability and applicability of such clinical tools to aid screening and diagnosis of active convulsive epilepsy in settings beyond their original training contexts. METHODS We sourced data from the Study of Epidemiology of Epilepsy in Demographic Sites dataset, which includes demographic information and clinical variables related to diagnosing epilepsy across five sub-Saharan African sites. For each site, we developed a region-specific (single-site) predictive model for epilepsy and assessed its performance at other sites. We then iteratively added sites to a multi-site model and evaluated model performance on the omitted regions. Model performances and parameters were then compared across every permutation of sites. We used a leave-one-site-out cross-validation analysis to assess the impact of incorporating individual site data in the model. RESULTS Single-site clinical models performed well within their own regions, but generally worse when evaluated in other regions (p<0.05). Model weights and optimal thresholds varied markedly across sites. When the models were trained using data from an increasing number of sites, mean internal performance decreased while external performance improved. CONCLUSIONS Clinical models for epilepsy diagnosis in LMICs demonstrate characteristic traits of ML models, such as limited generalisability and a trade-off between internal and external performance. The relationship between predictors and model outcomes also varies across sites, suggesting the need to update specific model aspects with local data before broader implementation. Variations are likely to be particular to the cultural context of diagnosis. We recommend developing models adapted to the cultures and contexts of their intended deployment and caution against deploying region- and culture-naïve models without thorough prior evaluation.
Collapse
Affiliation(s)
- Ioana Duta
- Oxford Epilepsy Research Group, Nuffield Department of Clinical Neurosciences, John Radcliffe Hospital, Oxford, United Kingdom
- Oxford Digital Health Labs, Nuffield Department of Women’s and Reproductive Health, The University of Oxford, John Radcliffe Hospital, Oxford, United Kingdom
| | - Symon M. Kariuki
- KEMRI/Wellcome Trust Research Programme, Centre for Geographic Medicine Research–Coast, Kilifi, Kenya
- Studies of Epidemiology of Epilepsy in Demographic Surveillance Systems (SEEDS)–INDEPTH Network, Accra, Ghana
- Department of Public Health, Pwani University, Kilifi, Kenya
| | - Anthony K. Ngugi
- Studies of Epidemiology of Epilepsy in Demographic Surveillance Systems (SEEDS)–INDEPTH Network, Accra, Ghana
- Department of Population Health, Aga Khan University, Nairobi, Kenya
| | - Angelina Kakooza Mwesige
- Department of Paediatrics and Child Health, Makerere University College of Health Sciences, Kampala, Uganda
| | | | - Daniel M. Mwanga
- Department of Population Health, Aga Khan University, Nairobi, Kenya
- Department of Mathematics, University of Nairobi, Nairobi, Kenya
| | - Seth Owusu-Agyei
- Kintampo Health Research Centre, Kintampo, Ghana
- Institute of Health Research, University of Health and Allied Sciences, Ho. Ghana
| | - Ryan Wagner
- MRC/Wits Rural Public Health & Health Transitions Research Unit (Agincourt), School of Public Health, Faculty of Health Sciences, University of the Witwatersrand, Johannesburg, South Africa
| | - J Helen Cross
- Developmental Neurosciences, University College London NIHR BRC Great Ormond Street Institute of Child Health, London, United Kingdom
| | - Josemir W. Sander
- Department of Clinical & Experimental Epilepsy, UCL Queen Square Institute of Neurology, London, & Chalfont Centre for Epilepsy, Chalfont St Peter, United Kingdom
- Stichting Epilepsie Instellingen Nederland, Heemstede, Netherlands
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, China
| | - Charles R. Newton
- Oxford Epilepsy Research Group, Nuffield Department of Clinical Neurosciences, John Radcliffe Hospital, Oxford, United Kingdom
- KEMRI/Wellcome Trust Research Programme, Centre for Geographic Medicine Research–Coast, Kilifi, Kenya
- Studies of Epidemiology of Epilepsy in Demographic Surveillance Systems (SEEDS)–INDEPTH Network, Accra, Ghana
- Department of Psychiatry, University of Oxford, Oxford, United Kingdom
| | - Arjune Sen
- Oxford Epilepsy Research Group, Nuffield Department of Clinical Neurosciences, John Radcliffe Hospital, Oxford, United Kingdom
| | - Gabriel Davis Jones
- Oxford Epilepsy Research Group, Nuffield Department of Clinical Neurosciences, John Radcliffe Hospital, Oxford, United Kingdom
- Oxford Digital Health Labs, Nuffield Department of Women’s and Reproductive Health, The University of Oxford, John Radcliffe Hospital, Oxford, United Kingdom
- The Alan Turing Institute, London, United Kingdom
| |
Collapse
|
21
|
Chen Z, Bian Y, Shen E, Fan L, Zhu W, Shi F, Shao C, Chen X, Xiang D. Moment-Consistent Contrastive CycleGAN for Cross-Domain Pancreatic Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:422-435. [PMID: 39167524 DOI: 10.1109/tmi.2024.3447071] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/23/2024]
Abstract
CT and MR are currently the most common imaging techniques for pancreatic cancer diagnosis. Accurate segmentation of the pancreas in CT and MR images can provide significant help in the diagnosis and treatment of pancreatic cancer. Traditional supervised segmentation methods require a large number of labeled CT and MR training data, which is usually time-consuming and laborious. Meanwhile, due to domain shift, traditional segmentation networks are difficult to be deployed on different imaging modality datasets. Cross-domain segmentation can utilize labeled source domain data to assist unlabeled target domains in solving the above problems. In this paper, a cross-domain pancreas segmentation algorithm is proposed based on Moment-Consistent Contrastive Cycle Generative Adversarial Networks (MC-CCycleGAN). MC-CCycleGAN is a style transfer network, in which the encoder of its generator is used to extract features from real images and style transfer images, constrain feature extraction through a contrastive loss, and fully extract structural features of input images during style transfer while eliminate redundant style features. The multi-order central moments of the pancreas are proposed to describe its anatomy in high dimensions and a contrastive loss is also proposed to constrain the moment consistency, so as to maintain consistency of the pancreatic structure and shape before and after style transfer. Multi-teacher knowledge distillation framework is proposed to transfer the knowledge from multiple teachers to a single student, so as to improve the robustness and performance of the student network. The experimental results have demonstrated the superiority of our framework over state-of-the-art domain adaptation methods.
Collapse
|
22
|
Li J, Shi H, Chen W, Liu N, Hwang KS. Semi-Supervised Detection Model Based on Adaptive Ensemble Learning for Medical Images. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:237-248. [PMID: 37339032 DOI: 10.1109/tnnls.2023.3282809] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/22/2023]
Abstract
Introducing deep learning technologies into the medical image processing field requires accuracy guarantee, especially for high-resolution images relayed through endoscopes. Moreover, works relying on supervised learning are powerless in the case of inadequate labeled samples. Therefore, for end-to-end medical image detection with overcritical efficiency and accuracy in endoscope detection, an ensemble-learning-based model with a semi-supervised mechanism is developed in this work. To gain a more accurate result through multiple detection models, we propose a new ensemble mechanism, termed alternative adaptive boosting method (Al-Adaboost), combining the decision-making of two hierarchical models. Specifically, the proposal consists of two modules. One is a local region proposal model with attentive temporal-spatial pathways for bounding box regression and classification, and the other one is a recurrent attention model (RAM) to provide more precise inferences for further classification according to the regression result. The proposal Al-Adaboost will adjust the weights of labeled samples and the two classifiers adaptively, and the nonlabel samples are assigned pseudolabels by our model. We investigate the performance of Al-Adaboost on both the colonoscopy and laryngoscopy data coming from CVC-ClinicDB and the affiliated hospital of Kaohsiung Medical University. The experimental results prove the feasibility and superiority of our model.
Collapse
|
23
|
Askarizadeh M, Morsali A, Nguyen KK. Resource-Constrained Multisource Instance-Based Transfer Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:1029-1043. [PMID: 37930915 DOI: 10.1109/tnnls.2023.3327248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2023]
Abstract
In today's machine learning (ML), the need for vast amounts of training data has become a significant challenge. Transfer learning (TL) offers a promising solution by leveraging knowledge across different domains/tasks, effectively addressing data scarcity. However, TL encounters computational and communication challenges in resource-constrained scenarios, and negative transfer (NT) can arise from specific data distributions. This article presents a novel focus on maximizing the accuracy of instance-based TL in multisource resource-constrained environments while mitigating NT, a key concern in TL. Previous studies have overlooked the impact of resource consumption in addressing the NT problem. To address these challenges, we introduce an optimization model named multisource resource-constrained optimized TL (MSOPTL), which employs a convex combination of empirical sources and target errors while considering feasibility and resource constraints. Moreover, we enhance one of the generalization error upper bounds in domain adaptation setting by demonstrating the potential to substitute the divergence with the Kullback-Leibler (KL) divergence. We utilize this enhanced error upper bound as one of the feasibility constraints of MSOPTL. Our suggested model can be applied as a versatile framework for various ML methods. Our approach is extensively validated in a neural network (NN)-based classification problem, demonstrating the efficiency of MSOPTL in achieving the desired trade-offs between TL's benefits and associated costs. This advancement holds tremendous potential for enhancing edge artificial intelligence (AI) applications in resource-constrained environments.
Collapse
|
24
|
Zeng Q, Xie Y, Lu Z, Lu M, Zhang J, Xia Y. Consistency-Guided Differential Decoding for Enhancing Semi-Supervised Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:44-56. [PMID: 39088492 DOI: 10.1109/tmi.2024.3429340] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/03/2024]
Abstract
Semi-supervised learning (SSL) has been proven beneficial for mitigating the issue of limited labeled data, especially on volumetric medical image segmentation. Unlike previous SSL methods which focus on exploring highly confident pseudo-labels or developing consistency regularization schemes, our empirical findings suggest that differential decoder features emerge naturally when two decoders strive to generate consistent predictions. Based on the observation, we first analyze the treasure of discrepancy in learning towards consistency, under both pseudo-labeling and consistency regularization settings, and subsequently propose a novel SSL method called LeFeD, which learns the feature-level discrepancies obtained from two decoders, by feeding such information as feedback signals to the encoder. The core design of LeFeD is to enlarge the discrepancies by training differential decoders, and then learn from the differential features iteratively. We evaluate LeFeD against eight state-of-the-art (SOTA) methods on three public datasets. Experiments show LeFeD surpasses competitors without any bells and whistles, such as uncertainty estimation and strong constraints, as well as setting a new state of the art for semi-supervised medical image segmentation. Code has been released at https://github.com/maxwell0027/LeFeD.
Collapse
|
25
|
Anaya-Sánchez H, Altamirano-Robles L, Díaz-Hernández R, Zapotecas-Martínez S. WGAN-GP for Synthetic Retinal Image Generation: Enhancing Sensor-Based Medical Imaging for Classification Models. SENSORS (BASEL, SWITZERLAND) 2024; 25:167. [PMID: 39796958 PMCID: PMC11723073 DOI: 10.3390/s25010167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/29/2024] [Revised: 12/19/2024] [Accepted: 12/28/2024] [Indexed: 01/13/2025]
Abstract
Accurate synthetic image generation is crucial for addressing data scarcity challenges in medical image classification tasks, particularly in sensor-derived medical imaging. In this work, we propose a novel method using a Wasserstein Generative Adversarial Network with Gradient Penalty (WGAN-GP) and nearest-neighbor interpolation to generate high-quality synthetic images for diabetic retinopathy classification. Our approach enhances training datasets by generating realistic retinal images that retain critical pathological features. We evaluated the method across multiple retinal image datasets, including Retinal-Lesions, Fine-Grained Annotated Diabetic Retinopathy (FGADR), Indian Diabetic Retinopathy Image Dataset (IDRiD), and the Kaggle Diabetic Retinopathy dataset. The proposed method outperformed traditional generative models, such as conditional GANs and PathoGAN, achieving the best performance on key metrics: a Fréchet Inception Distance (FID) of 15.21, a Mean Squared Error (MSE) of 0.002025, and a Structural Similarity Index (SSIM) of 0.89 in the Kaggle dataset. Additionally, expert evaluations revealed that only 56.66% of synthetic images could be distinguished from real ones, demonstrating the high fidelity and clinical relevance of the generated data. These results highlight the effectiveness of our approach in improving medical image classification by generating realistic and diverse synthetic datasets.
Collapse
Affiliation(s)
- Héctor Anaya-Sánchez
- Computer Science Department, Instituto Nacional de Astrofísica Óptica y Electrónica, Luis Enrrique Erro No. 1, Sta. María Tonantzintla, Puebla 72840, Mexico; (H.A.-S.); (S.Z.-M.)
| | - Leopoldo Altamirano-Robles
- Computer Science Department, Instituto Nacional de Astrofísica Óptica y Electrónica, Luis Enrrique Erro No. 1, Sta. María Tonantzintla, Puebla 72840, Mexico; (H.A.-S.); (S.Z.-M.)
| | - Raquel Díaz-Hernández
- Optics Department, Instituto Nacional de Astrofísica Óptica y Electrónica, Luis Enrrique Erro No. 1, Sta. María Tonantzintla, Puebla 72840, Mexico;
| | - Saúl Zapotecas-Martínez
- Computer Science Department, Instituto Nacional de Astrofísica Óptica y Electrónica, Luis Enrrique Erro No. 1, Sta. María Tonantzintla, Puebla 72840, Mexico; (H.A.-S.); (S.Z.-M.)
| |
Collapse
|
26
|
Vegesana K, Thomas PG. Cracking the code of adaptive immunity: The role of computational tools. Cell Syst 2024; 15:1156-1167. [PMID: 39701033 DOI: 10.1016/j.cels.2024.11.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Revised: 06/14/2024] [Accepted: 11/14/2024] [Indexed: 12/21/2024]
Abstract
In recent years, the advances in high-throughput and deep sequencing have generated a diverse amount of adaptive immune repertoire data. This surge in data has seen a proportional increase in computational methods aimed to characterize T cell receptor (TCR) repertoires. In this perspective, we will provide a brief commentary on the various domains of TCR repertoire analysis, their respective computational methods, and the ongoing challenges. Given the breadth of methods and applications of TCR analysis, we will focus our perspective on sequence-based computational methods.
Collapse
Affiliation(s)
- Kasi Vegesana
- Department of Host-Microbe Interactions, St. Jude Children's Research Hospital, Memphis, TN, USA
| | - Paul G Thomas
- Department of Host-Microbe Interactions, St. Jude Children's Research Hospital, Memphis, TN, USA.
| |
Collapse
|
27
|
Zhang J, Hao F, Liu X, Yao S, Wu Y, Li M, Zheng W. Multi-scale multi-instance contrastive learning for whole slide image classification. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE 2024; 138:109300. [DOI: 10.1016/j.engappai.2024.109300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/03/2025]
|
28
|
Kang H, Kim M, Ko YS, Cho Y, Yi MY. WISE: Efficient WSI selection for active learning in histopathology. Comput Med Imaging Graph 2024; 118:102455. [PMID: 39481146 DOI: 10.1016/j.compmedimag.2024.102455] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2024] [Revised: 10/06/2024] [Accepted: 10/18/2024] [Indexed: 11/02/2024]
Abstract
Deep neural network (DNN) models have been applied to a wide variety of medical image analysis tasks, often with the successful performance outcomes that match those of medical doctors. However, given that even minor errors in a model can impact patients' life, it is critical that these models are continuously improved. Hence, active learning (AL) has garnered attention as an effective and sustainable strategy for enhancing DNN models for the medical domain. Extant AL research in histopathology has primarily focused on patch datasets derived from whole-slide images (WSIs), a standard form of cancer diagnostic images obtained from a high-resolution scanner. However, this approach has failed to address the selection of WSIs, which can impede the performance improvement of deep learning models and increase the number of WSIs needed to achieve the target performance. This study introduces a WSI-level AL method, termed WSI-informative selection (WISE). WISE is designed to select informative WSIs using a newly formulated WSI-level class distance metric. This method aims to identify diverse and uncertain cases of WSIs, thereby contributing to model performance enhancement. WISE demonstrates state-of-the-art performance across the Colon and Stomach datasets, collected in the real world, as well as the public DigestPath dataset, significantly reducing the required number of WSIs by more than threefold compared to the one-pool dataset setting, which has been dominantly used in the field.
Collapse
Affiliation(s)
- Hyeongu Kang
- Graduate School of Data Science, Department of Industrial and Systems Engineering, Korea Advanced Institute of Science and Technology, Daejeon, South Korea
| | - Mujin Kim
- Graduate School of Data Science, Department of Industrial and Systems Engineering, Korea Advanced Institute of Science and Technology, Daejeon, South Korea
| | - Young Sin Ko
- Pathology Center, Seegene Medical Foundation, Seoul, South Korea
| | - Yesung Cho
- Graduate School of Data Science, Department of Industrial and Systems Engineering, Korea Advanced Institute of Science and Technology, Daejeon, South Korea
| | - Mun Yong Yi
- Graduate School of Data Science, Department of Industrial and Systems Engineering, Korea Advanced Institute of Science and Technology, Daejeon, South Korea.
| |
Collapse
|
29
|
Torres HR, Oliveira B, Fritze A, Birdir C, Rudiger M, Fonseca JC, Morais P, Vilaca JL. Deep-DM: Deep-Driven Deformable Model for 3D Image Segmentation Using Limited Data. IEEE J Biomed Health Inform 2024; 28:7287-7299. [PMID: 39110559 DOI: 10.1109/jbhi.2024.3440171] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Objective - Medical image segmentation is essential for several clinical tasks, including diagnosis, surgical and treatment planning, and image-guided interventions. Deep Learning (DL) methods have become the state-of-the-art for several image segmentation scenarios. However, a large and well-annotated dataset is required to effectively train a DL model, which is usually difficult to obtain in clinical practice, especially for 3D images. Methods - In this paper, we proposed Deep-DM, a learning-guided deformable model framework for 3D medical imaging segmentation using limited training data. In the proposed method, an energy function is learned by a Convolutional Neural Network (CNN) and integrated into an explicit deformable model to drive the evolution of an initial surface towards the object to segment. Specifically, the learning-based energy function is iteratively retrieved from localized anatomical representations of the image containing the image information around the evolving surface at each iteration. By focusing on localized regions of interest, this representation excludes irrelevant image information, facilitating the learning process. Results and conclusion - The performance of the proposed method is demonstrated for the tasks of left ventricle and fetal head segmentation in ultrasound, left atrium segmentation in Magnetic Resonance, and bladder segmentation in Computed Tomography, using different numbers of training volumes in each study. The results obtained showed the feasibility of the proposed method to segment different anatomical structures in different imaging modalities. Moreover, the results also showed that the proposed approach is less dependent on the size of the training dataset in comparison with state-of-the-art DL-based segmentation methods, outperforming them for all tasks when a low number of samples is available. Significance - Overall, by offering a more robust and less data-intensive approach to accurately segmenting anatomical structures, the proposed method has the potential to enhance clinical tasks that require image segmentation strategies.
Collapse
|
30
|
Sun Y, Ge X, Niu R, Gao J, Shi Y, Shao X, Wang Y, Shao X. PET/CT radiomics and deep learning in the diagnosis of benign and malignant pulmonary nodules: progress and challenges. Front Oncol 2024; 14:1491762. [PMID: 39582533 PMCID: PMC11581934 DOI: 10.3389/fonc.2024.1491762] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2024] [Accepted: 10/25/2024] [Indexed: 11/26/2024] Open
Abstract
Lung cancer is currently the leading cause of cancer-related deaths, and early diagnosis and screening can significantly reduce its mortality rate. Since some early-stage lung cancers lack obvious clinical symptoms and only present as pulmonary nodules (PNs) in imaging examinations, accurately determining the benign or malignant nature of PNs is crucial for improving patient survival rates. 18F-FDG PET/CT is important in diagnosing PNs, but its specificity needs improvement. Radiomics can provide information beyond traditional visual assessment, overcoming its limitations by extracting high-throughput quantitative features from medical images. Radiomics features based on 18F-FDG PET/CT and deep learning methods have shown great potential in the noninvasive diagnosis of PNs. This paper reviews the latest advancements in these methods and discusses their contributions to improving diagnostic accuracy and the challenges they face.
Collapse
Affiliation(s)
- Yan Sun
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Xinyu Ge
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Rong Niu
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Jianxiong Gao
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Yunmei Shi
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Xiaoliang Shao
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Yuetao Wang
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| | - Xiaonan Shao
- Department of Nuclear Medicine, the Third Affiliated Hospital of Soochow University, Changzhou, China
- Institute of Clinical Translation of Nuclear Medicine and Molecular Imaging, Soochow University, Changzhou, China
- Department of Nuclear Medicine, Changzhou Clinical Medical Center, Changzhou, China
| |
Collapse
|
31
|
Niestrata M, Radia M, Jackson J, Allan B. Global review of publicly available image datasets for the anterior segment of the eye. J Cataract Refract Surg 2024; 50:1184-1190. [PMID: 39150312 DOI: 10.1097/j.jcrs.0000000000001538] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2024] [Accepted: 08/03/2024] [Indexed: 08/17/2024]
Abstract
This study comprehensively reviewed publicly available image datasets for the anterior segment, with a focus on cataract, refractive, and corneal surgeries. The goal was to assess characteristics of existing datasets and identify areas for improvement. PubMED and Google searches were performed using the search terms "refractive surgery," "anterior segment," "cornea," "corneal," "cataract" AND "database," with the related word of "imaging." Results of each of these searches were collated, identifying 26 publicly available anterior segment image datasets. Imaging modalities included optical coherence tomography, photography, and confocal microscopy. Most datasets were small, 80% originated in the U.S., China, or Europe. Over 50% of images were from normal eyes. Disease states represented included keratoconus, corneal ulcers, and Fuchs dystrophy. Most of the datasets were incompletely described. To promote accessibility going forward to 2030, the ESCRS Digital Health Special Interest Group will annually update a list of available image datasets for anterior segment at www.escrs.org .
Collapse
Affiliation(s)
- Magdalena Niestrata
- From the NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL Institute of Ophthalmology, London, United Kingdom (Niestrata, Allan); Moorfields Eye Hospital NHS Foundation Trust, London, United Kingdom (Radia, Allan); Data and Statistics Department, University of East London, London, United Kingdom (Jackson)
| | | | | | | |
Collapse
|
32
|
Wiedenmann M, Barch M, Chang PS, Giltnane J, Risom T, Zijlstra A. An Immunofluorescence-Guided Segmentation Model in Hematoxylin and Eosin Images Is Enabled by Tissue Artifact Correction Using a Cycle-Consistent Generative Adversarial Network. Mod Pathol 2024; 37:100591. [PMID: 39147031 DOI: 10.1016/j.modpat.2024.100591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 08/01/2024] [Indexed: 08/17/2024]
Abstract
Despite recent advances, the adoption of computer vision methods into clinical and commercial applications has been hampered by the limited availability of accurate ground truth tissue annotations required to train robust supervised models. Generating such ground truth can be accelerated by annotating tissue molecularly using immunofluorescence (IF) staining and mapping these annotations to a post-IF hematoxylin and eosin (H&E) (terminal H&E) stain. Mapping the annotations between IF and terminal H&E increases both the scale and accuracy by which ground truth could be generated. However, discrepancies between terminal H&E and conventional H&E caused by IF tissue processing have limited this implementation. We sought to overcome this challenge and achieve compatibility between these parallel modalities using synthetic image generation, in which a cycle-consistent generative adversarial network was applied to transfer the appearance of conventional H&E such that it emulates terminal H&E. These synthetic emulations allowed us to train a deep learning model for the segmentation of epithelium in terminal H&E that could be validated against the IF staining of epithelial-based cytokeratins. The combination of this segmentation model with the cycle-consistent generative adversarial network stain transfer model enabled performative epithelium segmentation in conventional H&E images. The approach demonstrates that the training of accurate segmentation models for the breadth of conventional H&E data can be executed free of human expert annotations by leveraging molecular annotation strategies such as IF, so long as the tissue impacts of the molecular annotation protocol are captured by generative models that can be deployed prior to the segmentation process.
Collapse
Affiliation(s)
- Marcel Wiedenmann
- Department of Computer and Information Science, University of Konstanz, Konstanz, Germany
| | - Mariya Barch
- Department of Research Pathology, Genentech Inc, South San Francisco, California
| | - Patrick S Chang
- Department of Research Pathology, Genentech Inc, South San Francisco, California
| | - Jennifer Giltnane
- Department of Research Pathology, Genentech Inc, South San Francisco, California
| | - Tyler Risom
- Department of Research Pathology, Genentech Inc, South San Francisco, California.
| | - Andries Zijlstra
- Department of Research Pathology, Genentech Inc, South San Francisco, California; Department of Pathology, Microbiology and Immunology, Vanderbilt University Medical Center, Nashville, Tennessee
| |
Collapse
|
33
|
Håkansson S, Tuci M, Bolliger M, Curt A, Jutzeler CR, Brüningk SC. Data-driven prediction of spinal cord injury recovery: An exploration of current status and future perspectives. Exp Neurol 2024; 380:114913. [PMID: 39097073 DOI: 10.1016/j.expneurol.2024.114913] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Revised: 07/24/2024] [Accepted: 07/30/2024] [Indexed: 08/05/2024]
Abstract
Spinal Cord Injury (SCI) presents a significant challenge in rehabilitation medicine, with recovery outcomes varying widely among individuals. Machine learning (ML) is a promising approach to enhance the prediction of recovery trajectories, but its integration into clinical practice requires a thorough understanding of its efficacy and applicability. We systematically reviewed the current literature on data-driven models of SCI recovery prediction. The included studies were evaluated based on a range of criteria assessing the approach, implementation, input data preferences, and the clinical outcomes aimed to forecast. We observe a tendency to utilize routinely acquired data, such as International Standards for Neurological Classification of SCI (ISNCSCI), imaging, and demographics, for the prediction of functional outcomes derived from the Spinal Cord Independence Measure (SCIM) III and Functional Independence Measure (FIM) scores with a focus on motor ability. Although there has been an increasing interest in data-driven studies over time, traditional machine learning architectures, such as linear regression and tree-based approaches, remained the overwhelmingly popular choices for implementation. This implies ample opportunities for exploring architectures addressing the challenges of predicting SCI recovery, including techniques for learning from limited longitudinal data, improving generalizability, and enhancing reproducibility. We conclude with a perspective, highlighting possible future directions for data-driven SCI recovery prediction and drawing parallels to other application fields in terms of diverse data types (imaging, tabular, sequential, multimodal), data challenges (limited, missing, longitudinal data), and algorithmic needs (causal inference, robustness).
Collapse
Affiliation(s)
- Samuel Håkansson
- ETH Zürich, Department of Health Sciences and Technology (D-HEST), Zürich, Switzerland; Swiss Institute of Bioinformatics (SIB), Lausanne, Switzerland.
| | - Miklovana Tuci
- ETH Zürich, Department of Health Sciences and Technology (D-HEST), Zürich, Switzerland; Spinal Cord Injury Center, University Hospital Balgrist, University of Zürich, Switzerland
| | - Marc Bolliger
- Spinal Cord Injury Center, University Hospital Balgrist, University of Zürich, Switzerland
| | - Armin Curt
- Spinal Cord Injury Center, University Hospital Balgrist, University of Zürich, Switzerland
| | - Catherine R Jutzeler
- ETH Zürich, Department of Health Sciences and Technology (D-HEST), Zürich, Switzerland; Swiss Institute of Bioinformatics (SIB), Lausanne, Switzerland
| | - Sarah C Brüningk
- ETH Zürich, Department of Health Sciences and Technology (D-HEST), Zürich, Switzerland; Swiss Institute of Bioinformatics (SIB), Lausanne, Switzerland
| |
Collapse
|
34
|
Maniaci A, Lavalle S, Gagliano C, Lentini M, Masiello E, Parisi F, Iannella G, Cilia ND, Salerno V, Cusumano G, La Via L. The Integration of Radiomics and Artificial Intelligence in Modern Medicine. Life (Basel) 2024; 14:1248. [PMID: 39459547 PMCID: PMC11508875 DOI: 10.3390/life14101248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2024] [Revised: 09/16/2024] [Accepted: 09/18/2024] [Indexed: 10/28/2024] Open
Abstract
With profound effects on patient care, the role of artificial intelligence (AI) in radiomics has become a disruptive force in contemporary medicine. Radiomics, the quantitative feature extraction and analysis from medical images, offers useful imaging biomarkers that can reveal important information about the nature of diseases, how well patients respond to treatment and patient outcomes. The use of AI techniques in radiomics, such as machine learning and deep learning, has made it possible to create sophisticated computer-aided diagnostic systems, predictive models, and decision support tools. The many uses of AI in radiomics are examined in this review, encompassing its involvement of quantitative feature extraction from medical images, the machine learning, deep learning and computer-aided diagnostic (CAD) systems approaches in radiomics, and the effect of radiomics and AI on improving workflow automation and efficiency, optimize clinical trials and patient stratification. This review also covers the predictive modeling improvement by machine learning in radiomics, the multimodal integration and enhanced deep learning architectures, and the regulatory and clinical adoption considerations for radiomics-based CAD. Particular emphasis is given to the enormous potential for enhancing diagnosis precision, treatment personalization, and overall patient outcomes.
Collapse
Affiliation(s)
- Antonino Maniaci
- Faculty of Medicine and Surgery, University of Enna “Kore”, 94100 Enna, Italy; (A.M.); (S.L.); (C.G.)
| | - Salvatore Lavalle
- Faculty of Medicine and Surgery, University of Enna “Kore”, 94100 Enna, Italy; (A.M.); (S.L.); (C.G.)
| | - Caterina Gagliano
- Faculty of Medicine and Surgery, University of Enna “Kore”, 94100 Enna, Italy; (A.M.); (S.L.); (C.G.)
| | - Mario Lentini
- ASP Ragusa, Hospital Giovanni Paolo II, 97100 Ragusa, Italy;
| | - Edoardo Masiello
- Radiology Unit, Department Clinical and Experimental, Experimental Imaging Center, Vita-Salute San Raffaele University, 20132 Milan, Italy
| | - Federica Parisi
- Department of Medical and Surgical Sciences and Advanced Technologies “GF Ingrassia”, ENT Section, University of Catania, Via S. Sofia, 78, 95125 Catania, Italy;
| | - Giannicola Iannella
- Department of ‘Organi di Senso’, University “Sapienza”, Viale dell’Università, 33, 00185 Rome, Italy;
| | - Nicole Dalia Cilia
- Department of Computer Engineering, University of Enna “Kore”, 94100 Enna, Italy;
- Institute for Computing and Information Sciences, Radboud University Nijmegen, 6544 Nijmegen, The Netherlands
| | - Valerio Salerno
- Department of Engineering and Architecture, Kore University of Enna, 94100 Enna, Italy;
| | - Giacomo Cusumano
- University Hospital Policlinico “G. Rodolico—San Marco”, 95123 Catania, Italy;
- Department of General Surgery and Medical-Surgical Specialties, University of Catania, 95123 Catania, Italy
| | - Luigi La Via
- University Hospital Policlinico “G. Rodolico—San Marco”, 95123 Catania, Italy;
| |
Collapse
|
35
|
Peng H, Lin S, King D, Su YH, Abuzeid WM, Bly RA, Moe KS, Hannaford B. Reducing annotating load: Active learning with synthetic images in surgical instrument segmentation. Med Image Anal 2024; 97:103246. [PMID: 38943835 DOI: 10.1016/j.media.2024.103246] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2022] [Revised: 05/28/2024] [Accepted: 06/17/2024] [Indexed: 07/01/2024]
Abstract
Accurate instrument segmentation in the endoscopic vision of minimally invasive surgery is challenging due to complex instruments and environments. Deep learning techniques have shown competitive performance in recent years. However, deep learning usually requires a large amount of labeled data to achieve accurate prediction, which poses a significant workload. To alleviate this workload, we propose an active learning-based framework to generate synthetic images for efficient neural network training. In each active learning iteration, a small number of informative unlabeled images are first queried by active learning and manually labeled. Next, synthetic images are generated based on these selected images. The instruments and backgrounds are cropped out and randomly combined with blending and fusion near the boundary. The proposed method leverages the advantage of both active learning and synthetic images. The effectiveness of the proposed method is validated on two sinus surgery datasets and one intraabdominal surgery dataset. The results indicate a considerable performance improvement, especially when the size of the annotated dataset is small. All the code is open-sourced at: https://github.com/HaonanPeng/active_syn_generator.
Collapse
Affiliation(s)
- Haonan Peng
- University of Washington, 185 E Stevens Way NE AE100R, Seattle, WA 98195, USA.
| | - Shan Lin
- University of California San Diego, 9500 Gilman Dr, La Jolla, CA 92093, USA
| | - Daniel King
- University of Washington, 185 E Stevens Way NE AE100R, Seattle, WA 98195, USA
| | - Yun-Hsuan Su
- Mount Holyoke College, 50 College St, South Hadley, MA 01075, USA
| | - Waleed M Abuzeid
- University of Washington, 185 E Stevens Way NE AE100R, Seattle, WA 98195, USA
| | - Randall A Bly
- University of Washington, 185 E Stevens Way NE AE100R, Seattle, WA 98195, USA
| | - Kris S Moe
- University of Washington, 185 E Stevens Way NE AE100R, Seattle, WA 98195, USA
| | - Blake Hannaford
- University of Washington, 185 E Stevens Way NE AE100R, Seattle, WA 98195, USA
| |
Collapse
|
36
|
de Vente C, van Ginneken B, Hoyng CB, Klaver CCW, Sánchez CI. Uncertainty-aware multiple-instance learning for reliable classification: Application to optical coherence tomography. Med Image Anal 2024; 97:103259. [PMID: 38959721 DOI: 10.1016/j.media.2024.103259] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2023] [Revised: 06/17/2024] [Accepted: 06/24/2024] [Indexed: 07/05/2024]
Abstract
Deep learning classification models for medical image analysis often perform well on data from scanners that were used to acquire the training data. However, when these models are applied to data from different vendors, their performance tends to drop substantially. Artifacts that only occur within scans from specific scanners are major causes of this poor generalizability. We aimed to enhance the reliability of deep learning classification models using a novel method called Uncertainty-Based Instance eXclusion (UBIX). UBIX is an inference-time module that can be employed in multiple-instance learning (MIL) settings. MIL is a paradigm in which instances (generally crops or slices) of a bag (generally an image) contribute towards a bag-level output. Instead of assuming equal contribution of all instances to the bag-level output, UBIX detects instances corrupted due to local artifacts on-the-fly using uncertainty estimation, reducing or fully ignoring their contributions before MIL pooling. In our experiments, instances are 2D slices and bags are volumetric images, but alternative definitions are also possible. Although UBIX is generally applicable to diverse classification tasks, we focused on the staging of age-related macular degeneration in optical coherence tomography. Our models were trained on data from a single scanner and tested on external datasets from different vendors, which included vendor-specific artifacts. UBIX showed reliable behavior, with a slight decrease in performance (a decrease of the quadratic weighted kappa (κw) from 0.861 to 0.708), when applied to images from different vendors containing artifacts; while a state-of-the-art 3D neural network without UBIX suffered from a significant detriment of performance (κw from 0.852 to 0.084) on the same test set. We showed that instances with unseen artifacts can be identified with OOD detection. UBIX can reduce their contribution to the bag-level predictions, improving reliability without retraining on new data. This potentially increases the applicability of artificial intelligence models to data from other scanners than the ones for which they were developed. The source code for UBIX, including trained model weights, is publicly available through https://github.com/qurAI-amsterdam/ubix-for-reliable-classification.
Collapse
Affiliation(s)
- Coen de Vente
- Quantitative Healthcare Analysis (QurAI) Group, Informatics Institute, University of Amsterdam, Amsterdam, Noord-Holland, Netherlands; Department of Biomedical Engineering and Physics, Amsterdam University Medical Center, Amsterdam, Noord-Holland, Netherlands; Diagnostic Image Analysis Group (DIAG), Department of Radiology and Nuclear Medicine, Radboudumc, Nijmegen, Gelderland, Netherlands.
| | - Bram van Ginneken
- Diagnostic Image Analysis Group (DIAG), Department of Radiology and Nuclear Medicine, Radboudumc, Nijmegen, Gelderland, Netherlands
| | - Carel B Hoyng
- Department of Ophthalmology, Radboudumc, Nijmegen, Gelderland, Netherlands
| | - Caroline C W Klaver
- Department of Ophthalmology, Radboudumc, Nijmegen, Gelderland, Netherlands; Ophthalmology & Epidemiology, Erasmus MC, Rotterdam, Zuid-Holland, Netherlands
| | - Clara I Sánchez
- Quantitative Healthcare Analysis (QurAI) Group, Informatics Institute, University of Amsterdam, Amsterdam, Noord-Holland, Netherlands; Department of Biomedical Engineering and Physics, Amsterdam University Medical Center, Amsterdam, Noord-Holland, Netherlands
| |
Collapse
|
37
|
You C, Su GH, Zhang X, Xiao Y, Zheng RC, Sun SY, Zhou JY, Lin LY, Wang ZZ, Wang H, Chen Y, Peng WJ, Jiang YZ, Shao ZM, Gu YJ. Multicenter radio-multiomic analysis for predicting breast cancer outcome and unravelling imaging-biological connection. NPJ Precis Oncol 2024; 8:193. [PMID: 39244594 PMCID: PMC11380684 DOI: 10.1038/s41698-024-00666-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2024] [Accepted: 07/24/2024] [Indexed: 09/09/2024] Open
Abstract
Radiomics offers a noninvasive avenue for predicting clinicopathological factors. However, thorough investigations into a robust breast cancer outcome-predicting model and its biological significance remain limited. This study develops a robust radiomic model for prognosis prediction, and further excavates its biological foundation and transferring prediction performance. We retrospectively collected preoperative dynamic contrast-enhanced MRI data from three distinct breast cancer patient cohorts. In FUSCC cohort (n = 466), Lasso was used to select features correlated with patient prognosis and multivariate Cox regression was utilized to integrate these features and build the radiomic risk model, while multiomic analysis was conducted to investigate the model's biological implications. DUKE cohort (n = 619) and I-SPY1 cohort (n = 128) were used to test the performance of the radiomic signature in outcome prediction. A thirteen-feature radiomic signature was identified in the FUSCC cohort training set and validated in the FUSCC cohort testing set, DUKE cohort and I-SPY1 cohort for predicting relapse-free survival (RFS) and overall survival (OS) (RFS: p = 0.013, p = 0.024 and p = 0.035; OS: p = 0.036, p = 0.005 and p = 0.027 in the three cohorts). Multiomic analysis uncovered metabolic dysregulation underlying the radiomic signature (ATP metabolic process: NES = 1.84, p-adjust = 0.02; cholesterol biosynthesis: NES = 1.79, p-adjust = 0.01). Regarding the therapeutic implications, the radiomic signature exhibited value when combining clinical factors for predicting the pathological complete response to neoadjuvant chemotherapy (DUKE cohort, AUC = 0.72; I-SPY1 cohort, AUC = 0.73). In conclusion, our study identified a breast cancer outcome-predicting radiomic signature in a multicenter radio-multiomic study, along with its correlations with multiomic features in prognostic risk assessment, laying the groundwork for future prospective clinical trials in personalized risk stratification and precision therapy.
Collapse
Affiliation(s)
- Chao You
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Guan-Hua Su
- Key Laboratory of Breast Cancer in Shanghai, Department of Breast Surgery, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Xu Zhang
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Yi Xiao
- Key Laboratory of Breast Cancer in Shanghai, Department of Breast Surgery, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Ren-Cheng Zheng
- Institute of Science and Technology for Brain-inspired Intelligence, Fudan University, Shanghai, China
| | - Shi-Yun Sun
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Jia-Yin Zhou
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Lu-Yi Lin
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Ze-Zhou Wang
- Department of Cancer Prevention, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - He Wang
- Institute of Science and Technology for Brain-inspired Intelligence, Fudan University, Shanghai, China
| | - Yan Chen
- School of Medicine, University of Nottingham, Nottingham, UK
| | - Wei-Jun Peng
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Yi-Zhou Jiang
- Key Laboratory of Breast Cancer in Shanghai, Department of Breast Surgery, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China.
| | - Zhi-Ming Shao
- Key Laboratory of Breast Cancer in Shanghai, Department of Breast Surgery, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China.
| | - Ya-Jia Gu
- Department of Radiology, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China.
| |
Collapse
|
38
|
Yousefirizi F, Shiri I, O JH, Bloise I, Martineau P, Wilson D, Bénard F, Sehn LH, Savage KJ, Zaidi H, Uribe CF, Rahmim A. Semi-supervised learning towards automated segmentation of PET images with limited annotations: application to lymphoma patients. Phys Eng Sci Med 2024; 47:833-849. [PMID: 38512435 DOI: 10.1007/s13246-024-01408-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Accepted: 02/18/2024] [Indexed: 03/23/2024]
Abstract
Manual segmentation poses a time-consuming challenge for disease quantification, therapy evaluation, treatment planning, and outcome prediction. Convolutional neural networks (CNNs) hold promise in accurately identifying tumor locations and boundaries in PET scans. However, a major hurdle is the extensive amount of supervised and annotated data necessary for training. To overcome this limitation, this study explores semi-supervised approaches utilizing unlabeled data, specifically focusing on PET images of diffuse large B-cell lymphoma (DLBCL) and primary mediastinal large B-cell lymphoma (PMBCL) obtained from two centers. We considered 2-[18F]FDG PET images of 292 patients PMBCL (n = 104) and DLBCL (n = 188) (n = 232 for training and validation, and n = 60 for external testing). We harnessed classical wisdom embedded in traditional segmentation methods, such as the fuzzy clustering loss function (FCM), to tailor the training strategy for a 3D U-Net model, incorporating both supervised and unsupervised learning approaches. Various supervision levels were explored, including fully supervised methods with labeled FCM and unified focal/Dice loss, unsupervised methods with robust FCM (RFCM) and Mumford-Shah (MS) loss, and semi-supervised methods combining FCM with supervised Dice loss (MS + Dice) or labeled FCM (RFCM + FCM). The unified loss function yielded higher Dice scores (0.73 ± 0.11; 95% CI 0.67-0.8) than Dice loss (p value < 0.01). Among the semi-supervised approaches, RFCM + αFCM (α = 0.3) showed the best performance, with Dice score of 0.68 ± 0.10 (95% CI 0.45-0.77), outperforming MS + αDice for any supervision level (any α) (p < 0.01). Another semi-supervised approach with MS + αDice (α = 0.2) achieved Dice score of 0.59 ± 0.09 (95% CI 0.44-0.76) surpassing other supervision levels (p < 0.01). Given the time-consuming nature of manual delineations and the inconsistencies they may introduce, semi-supervised approaches hold promise for automating medical imaging segmentation workflows.
Collapse
Affiliation(s)
- Fereshteh Yousefirizi
- Department of Integrative Oncology, BC Cancer Research Institute, Vancouver, BC, Canada.
| | - Isaac Shiri
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland
| | - Joo Hyun O
- College of Medicine, Seoul St. Mary's Hospital, The Catholic University of Korea, Seoul, Republic of Korea
| | | | | | - Don Wilson
- BC Cancer, Vancouver, BC, Canada
- Department of Radiology, University of British Columbia, Vancouver, Canada
| | | | - Laurie H Sehn
- BC Cancer, Vancouver, BC, Canada
- Centre for Lymphoid Cancer, BC Cancer, Vancouver, Canada
| | - Kerry J Savage
- BC Cancer, Vancouver, BC, Canada
- Centre for Lymphoid Cancer, BC Cancer, Vancouver, Canada
| | - Habib Zaidi
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland
- University Medical Center Groningen, University of Groningens, Groningen, Netherlands
- Department of Nuclear Medicine, University of Southern Denmark, Vancouver, Odense, Denmark
- University Research and Innovation Center, Óbuda University, Budapest, Hungary
| | - Carlos F Uribe
- Department of Integrative Oncology, BC Cancer Research Institute, Vancouver, BC, Canada
- BC Cancer, Vancouver, BC, Canada
- Department of Radiology, University of British Columbia, Vancouver, Canada
| | - Arman Rahmim
- Department of Integrative Oncology, BC Cancer Research Institute, Vancouver, BC, Canada
- BC Cancer, Vancouver, BC, Canada
- Department of Radiology, University of British Columbia, Vancouver, Canada
- Departments of Physics and Biomedical Engineering, University of British Columbia, Vancouver, Canada
| |
Collapse
|
39
|
Zhang X, Liu C, Zhu H, Wang T, Du Z, Ding W. A universal multiple instance learning framework for whole slide image analysis. Comput Biol Med 2024; 178:108714. [PMID: 38889627 DOI: 10.1016/j.compbiomed.2024.108714] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Revised: 06/04/2024] [Accepted: 06/04/2024] [Indexed: 06/20/2024]
Abstract
BACKGROUND The emergence of digital whole slide image (WSI) has driven the development of computational pathology. However, obtaining patch-level annotations is challenging and time-consuming due to the high resolution of WSI, which limits the applicability of fully supervised methods. We aim to address the challenges related to patch-level annotations. METHODS We propose a universal framework for weakly supervised WSI analysis based on Multiple Instance Learning (MIL). To achieve effective aggregation of instance features, we design a feature aggregation module from multiple dimensions by considering feature distribution, instances correlation and instance-level evaluation. First, we implement instance-level standardization layer and deep projection unit to improve the separation of instances in the feature space. Then, a self-attention mechanism is employed to explore dependencies between instances. Additionally, an instance-level pseudo-label evaluation method is introduced to enhance the available information during the weak supervision process. Finally, a bag-level classifier is used to obtain preliminary WSI classification results. To achieve even more accurate WSI label predictions, we have designed a key instance selection module that strengthens the learning of local features for instances. Combining the results from both modules leads to an improvement in WSI prediction accuracy. RESULTS Experiments conducted on Camelyon16, TCGA-NSCLC, SICAPv2, PANDA and classical MIL benchmark datasets demonstrate that our proposed method achieves a competitive performance compared to some recent methods, with maximum improvement of 14.6 % in terms of classification accuracy. CONCLUSION Our method can improve the classification accuracy of whole slide images in a weakly supervised way, and more accurately detect lesion areas.
Collapse
Affiliation(s)
- Xueqin Zhang
- College of Information Science and Engineering, East China University of Science and Technology, Shanghai, 200237, China; Shanghai Key Laboratory of Computer Software Evaluating and Testing, Shanghai 201112, China
| | - Chang Liu
- College of Information Science and Engineering, East China University of Science and Technology, Shanghai, 200237, China.
| | - Huitong Zhu
- College of Information Science and Engineering, East China University of Science and Technology, Shanghai, 200237, China
| | - Tianqi Wang
- College of Information Science and Engineering, East China University of Science and Technology, Shanghai, 200237, China
| | - Zunguo Du
- Department of Pathology, Huashan Hospital Affiliated to Fudan University, Shanghai, 200040, China
| | - Weihong Ding
- Department of Urology, Huashan Hospital Affiliated to Fudan University, Shanghai, 200040, China.
| |
Collapse
|
40
|
Liu Z, Kainth K, Zhou A, Deyer TW, Fayad ZA, Greenspan H, Mei X. A review of self-supervised, generative, and few-shot deep learning methods for data-limited magnetic resonance imaging segmentation. NMR IN BIOMEDICINE 2024; 37:e5143. [PMID: 38523402 DOI: 10.1002/nbm.5143] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Revised: 02/15/2024] [Accepted: 02/16/2024] [Indexed: 03/26/2024]
Abstract
Magnetic resonance imaging (MRI) is a ubiquitous medical imaging technology with applications in disease diagnostics, intervention, and treatment planning. Accurate MRI segmentation is critical for diagnosing abnormalities, monitoring diseases, and deciding on a course of treatment. With the advent of advanced deep learning frameworks, fully automated and accurate MRI segmentation is advancing. Traditional supervised deep learning techniques have advanced tremendously, reaching clinical-level accuracy in the field of segmentation. However, these algorithms still require a large amount of annotated data, which is oftentimes unavailable or impractical. One way to circumvent this issue is to utilize algorithms that exploit a limited amount of labeled data. This paper aims to review such state-of-the-art algorithms that use a limited number of annotated samples. We explain the fundamental principles of self-supervised learning, generative models, few-shot learning, and semi-supervised learning and summarize their applications in cardiac, abdomen, and brain MRI segmentation. Throughout this review, we highlight algorithms that can be employed based on the quantity of annotated data available. We also present a comprehensive list of notable publicly available MRI segmentation datasets. To conclude, we discuss possible future directions of the field-including emerging algorithms, such as contrastive language-image pretraining, and potential combinations across the methods discussed-that can further increase the efficacy of image segmentation with limited labels.
Collapse
Affiliation(s)
- Zelong Liu
- BioMedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Komal Kainth
- BioMedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Alexander Zhou
- BioMedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Timothy W Deyer
- East River Medical Imaging, New York, New York, USA
- Department of Radiology, Cornell Medicine, New York, New York, USA
| | - Zahi A Fayad
- BioMedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, New York, USA
- Department of Diagnostic, Molecular, and Interventional Radiology, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Hayit Greenspan
- BioMedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, New York, USA
- Department of Diagnostic, Molecular, and Interventional Radiology, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Xueyan Mei
- BioMedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, New York, USA
- Department of Diagnostic, Molecular, and Interventional Radiology, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| |
Collapse
|
41
|
Wagstyl K, Kobow K, Casillas-Espinosa PM, Cole AJ, Jiménez-Jiménez D, Nariai H, Baulac S, O'Brien T, Henshall DC, Akman O, Sankar R, Galanopoulou AS, Auvin S. WONOEP 2022: Neurotechnology for the diagnosis of epilepsy. Epilepsia 2024; 65:2238-2247. [PMID: 38829313 DOI: 10.1111/epi.18028] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2024] [Revised: 05/10/2024] [Accepted: 05/13/2024] [Indexed: 06/05/2024]
Abstract
Epilepsy's myriad causes and clinical presentations ensure that accurate diagnoses and targeted treatments remain a challenge. Advanced neurotechnologies are needed to better characterize individual patients across multiple modalities and analytical techniques. At the XVIth Workshop on Neurobiology of Epilepsy: Early Onset Epilepsies: Neurobiology and Novel Therapeutic Strategies (WONOEP 2022), the session on "advanced tools" highlighted a range of approaches, from molecular phenotyping of genetic epilepsy models and resected tissue samples to imaging-guided localization of epileptogenic tissue for surgical resection of focal malformations. These tools integrate cutting edge research, clinical data acquisition, and advanced computational methods to leverage the rich information contained within increasingly large datasets. A number of common challenges and opportunities emerged, including the need for multidisciplinary collaboration, multimodal integration, potential ethical challenges, and the multistage path to clinical translation. Despite these challenges, advanced epilepsy neurotechnologies offer the potential to improve our understanding of the underlying causes of epilepsy and our capacity to provide patient-specific treatment.
Collapse
Affiliation(s)
- Konrad Wagstyl
- School of Biomedical Engineering & Imaging Science, King's College London, London, UK
- Developmental Neurosciences, UCL Great Ormond Street for Child Health, UCL, London, UK
| | - Katja Kobow
- Institute of Neuropathology, Universitätsklinikum Erlangen, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
| | - Pablo M Casillas-Espinosa
- Department of Medicine, Royal Melbourne Hospital, University of Melbourne, Parkville, Victoria, Australia
- Department of Neuroscience, Central Clinical School, Monash University, Melbourne, Victoria, Australia
- Department of Neurology, Alfred Hospital, Melbourne, Victoria, Australia
| | - Andrew J Cole
- MGH Epilepsy Service, Division of Clinical Neurophysiology, Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts, USA
| | - Diego Jiménez-Jiménez
- Department of Basic and Clinical Neuroscience, Institute of Psychiatry, Psychology, and Neuroscience, King's College London, London, UK
- Department of Clinical and Experimental Epilepsy, UCL Queen Square Institute of Neurology, London, UK
| | - Hiroki Nariai
- Division of Pediatric Neurology, Department of Pediatrics, UCLA Medical Center, Los Angeles, California, USA
| | - Stéphanie Baulac
- Institut du Cerveau-Paris Brain Institute-ICM, INSERM, CNRS, Sorbonne Université, Paris, France
| | - Terence O'Brien
- Department of Medicine, Royal Melbourne Hospital, University of Melbourne, Parkville, Victoria, Australia
- Department of Neuroscience, Central Clinical School, Monash University, Melbourne, Victoria, Australia
- Department of Neurology, Alfred Hospital, Melbourne, Victoria, Australia
| | - David C Henshall
- FutureNeuro SFI Research Centre, RCSI University of Medicine and Health Sciences, Dublin, Ireland
- Department of Physiology and Medical Physics, RCSI University of Medicine and Health Sciences, Dublin, Ireland
| | - Ozlem Akman
- Department of Physiology, Faculty of Medicine, Demiroglu Bilim University, Istanbul, Turkey
| | - Raman Sankar
- Division of Pediatric Neurology, Department of Pediatrics, UCLA Mattel Children's Hospital, David Geffen School of Medicine, Los Angeles, California, USA
- UCLA Children's Discovery and Innovation Institute, California, Los Angeles, USA
| | - Aristea S Galanopoulou
- Saul R. Korey Department of Neurology, Isabelle Rapin Division of Child Neurology, Laboratory of Developmental Epilepsy, Dominick P. Purpura Department of Neuroscience, Albert Einstein College of Medicine, Bronx, New York, USA
| | - Stéphane Auvin
- Université Paris-Cité, INSERM NeuroDiderot, Paris, France
- Pediatric Neurology Department, APHP, Robert Debré University Hospital, CRMR Epilepsies Rares, EpiCARE member, Paris, France
- Institut Universitaire de France, Paris, France
| |
Collapse
|
42
|
Li W, Guo E, Zhao H, Li Y, Miao L, Liu C, Sun W. Evaluation of transfer ensemble learning-based convolutional neural network models for the identification of chronic gingivitis from oral photographs. BMC Oral Health 2024; 24:814. [PMID: 39020332 PMCID: PMC11256452 DOI: 10.1186/s12903-024-04460-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Accepted: 06/07/2024] [Indexed: 07/19/2024] Open
Abstract
BACKGROUND To evaluate the performances of several advanced deep convolutional neural network models (AlexNet, VGG, GoogLeNet, ResNet) based on ensemble learning for recognizing chronic gingivitis from screening oral images. METHODS A total of 683 intraoral clinical images acquired from 134 volunteers were used to construct the database and evaluate the models. Four deep ConvNet models were developed using ensemble learning and outperformed a single model. The performances of the different models were evaluated by comparing the accuracy and sensitivity for recognizing the existence of gingivitis from intraoral images. RESULTS The ResNet model achieved an area under the curve (AUC) value of 97%, while the AUC values for the GoogLeNet, AlexNet, and VGG models were 94%, 92%, and 89%, respectively. Although the ResNet and GoogLeNet models performed best in classifying gingivitis from images, the sensitivity outcomes were not significantly different among the ResNet, GoogLeNet, and Alexnet models (p>0.05). However, the sensitivity of the VGGNet model differed significantly from those of the other models (p < 0.001). CONCLUSION The ResNet and GoogLeNet models show promise for identifying chronic gingivitis from images. These models can help doctors diagnose periodontal diseases efficiently or based on self-examination of the oral cavity by patients.
Collapse
Affiliation(s)
- Wen Li
- Department of Cariology and Endodontics, Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing, China
| | - Enting Guo
- Division of Computer Science, The University of Aizu, Aizu, Japan
| | - Hong Zhao
- Division of Computer Science, The University of Aizu, Aizu, Japan
| | - Yuyang Li
- Department of Cariology and Endodontics, Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing, China
| | - Leiying Miao
- Department of Cariology and Endodontics, Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing, China
| | - Chao Liu
- Department of Orthodontic, Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing, China.
| | - Weibin Sun
- Department of Periodontics, Nanjing Stomatological Hospital, Affiliated Hospital of Medical School, Research Institute of Stomatology, Nanjing University, Nanjing, China.
| |
Collapse
|
43
|
Guan H, Yap PT, Bozoki A, Liu M. Federated learning for medical image analysis: A survey. PATTERN RECOGNITION 2024; 151:110424. [PMID: 38559674 PMCID: PMC10976951 DOI: 10.1016/j.patcog.2024.110424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Machine learning in medical imaging often faces a fundamental dilemma, namely, the small sample size problem. Many recent studies suggest using multi-domain data pooled from different acquisition sites/centers to improve statistical power. However, medical images from different sites cannot be easily shared to build large datasets for model training due to privacy protection reasons. As a promising solution, federated learning, which enables collaborative training of machine learning models based on data from different sites without cross-site data sharing, has attracted considerable attention recently. In this paper, we conduct a comprehensive survey of the recent development of federated learning methods in medical image analysis. We have systematically gathered research papers on federated learning and its applications in medical image analysis published between 2017 and 2023. Our search and compilation were conducted using databases from IEEE Xplore, ACM Digital Library, Science Direct, Springer Link, Web of Science, Google Scholar, and PubMed. In this survey, we first introduce the background of federated learning for dealing with privacy protection and collaborative learning issues. We then present a comprehensive review of recent advances in federated learning methods for medical image analysis. Specifically, existing methods are categorized based on three critical aspects of a federated learning system, including client end, server end, and communication techniques. In each category, we summarize the existing federated learning methods according to specific research problems in medical image analysis and also provide insights into the motivations of different approaches. In addition, we provide a review of existing benchmark medical imaging datasets and software platforms for current federated learning research. We also conduct an experimental study to empirically evaluate typical federated learning methods for medical image analysis. This survey can help to better understand the current research status, challenges, and potential research opportunities in this promising research field.
Collapse
Affiliation(s)
- Hao Guan
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Pew-Thian Yap
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Andrea Bozoki
- Department of Neurology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Mingxia Liu
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| |
Collapse
|
44
|
Poles I, D'Arnese E, Coggi M, Buccino F, Vergani L, Santambrogio MD. A Multimodal Transfer Learning Approach for Histopathology and SR-microCT Low-Data Regimes Image Segmentation. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-4. [PMID: 40039905 DOI: 10.1109/embc53108.2024.10781540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
Osteocyte-lacunar bone structures are a discerning marker for bone pathophysiology, given their geometric alterations observed during aging and diseases. Deep Learning (DL) image analysis has showcased the potential to comprehend bone health associated with their mechanisms. However, DL examination requires labeled and multimodal datasets, which is arduous with high-dimensional images. Within this context, we propose a method for segmenting osteocytes and lacunae in human bone histopathology and Synchrotron Radiation micro-Computed Tomography (SR-microCT) images, employing a deep U-Net in an intra-domain and multimodal transfer learning setting with a limited number of training images. Our strategy allows achieving 63.92±4.69 and 63.94±4.05 Dice Similarity Coefficient (DSC) osteocytes and lacunae segmentation, while up to 20.38 and 5.86 average DSC improvements over selected baselines even if 44× smaller datasets are employed for training.Clinical relevance-The proposed method analyzes bone histopathologies and SR-microCT images in a multimodal and low-data setting, easing the bone microscale investigations while supporting the study of osteocyte-lacunar pathophysiology.
Collapse
|
45
|
Cadrin-Chênevert A. Navigating Clinical Variability: Transfer Learning's Impact on Imaging Model Performance. Radiol Artif Intell 2024; 6:e240263. [PMID: 38900033 PMCID: PMC11294946 DOI: 10.1148/ryai.240263] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2024] [Revised: 05/04/2024] [Accepted: 05/09/2024] [Indexed: 06/21/2024]
Affiliation(s)
- Alexandre Cadrin-Chênevert
- From the CISSS Lanaudière-Medical Imaging, 200
Louis-Vadeboncoeur, Saint-Charles-Borromee, QC, Canada J6E 6J2
| |
Collapse
|
46
|
Wang J, Qiao L, Zhou S, Zhou J, Wang J, Li J, Ying S, Chang C, Shi J. Weakly Supervised Lesion Detection and Diagnosis for Breast Cancers With Partially Annotated Ultrasound Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2509-2521. [PMID: 38373131 DOI: 10.1109/tmi.2024.3366940] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2024]
Abstract
Deep learning (DL) has proven highly effective for ultrasound-based computer-aided diagnosis (CAD) of breast cancers. In an automatic CAD system, lesion detection is critical for the following diagnosis. However, existing DL-based methods generally require voluminous manually-annotated region of interest (ROI) labels and class labels to train both the lesion detection and diagnosis models. In clinical practice, the ROI labels, i.e. ground truths, may not always be optimal for the classification task due to individual experience of sonologists, resulting in the issue of coarse annotation to limit the diagnosis performance of a CAD model. To address this issue, a novel Two-Stage Detection and Diagnosis Network (TSDDNet) is proposed based on weakly supervised learning to improve diagnostic accuracy of the ultrasound-based CAD for breast cancers. In particular, all the initial ROI-level labels are considered as coarse annotations before model training. In the first training stage, a candidate selection mechanism is then designed to refine manual ROIs in the fully annotated images and generate accurate pseudo-ROIs for the partially annotated images under the guidance of class labels. The training set is updated with more accurate ROI labels for the second training stage. A fusion network is developed to integrate detection network and classification network into a unified end-to-end framework as the final CAD model in the second training stage. A self-distillation strategy is designed on this model for joint optimization to further improves its diagnosis performance. The proposed TSDDNet is evaluated on three B-mode ultrasound datasets, and the experimental results indicate that it achieves the best performance on both lesion detection and diagnosis tasks, suggesting promising application potential.
Collapse
|
47
|
Yuen B, Dong X, Lu T. A 3D ray traced biological neural network learning model. Nat Commun 2024; 15:4693. [PMID: 38824154 PMCID: PMC11525811 DOI: 10.1038/s41467-024-48747-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Accepted: 05/13/2024] [Indexed: 06/03/2024] Open
Abstract
Training large neural networks on big datasets requires significant computational resources and time. Transfer learning reduces training time by pre-training a base model on one dataset and transferring the knowledge to a new model for another dataset. However, current choices of transfer learning algorithms are limited because the transferred models always have to adhere to the dimensions of the base model and can not easily modify the neural architecture to solve other datasets. On the other hand, biological neural networks (BNNs) are adept at rearranging themselves to tackle completely different problems using transfer learning. Taking advantage of BNNs, we design a dynamic neural network that is transferable to any other network architecture and can accommodate many datasets. Our approach uses raytracing to connect neurons in a three-dimensional space, allowing the network to grow into any shape or size. In the Alcala dataset, our transfer learning algorithm trains the fastest across changing environments and input sizes. In addition, we show that our algorithm also outperformance the state of the art in EEG dataset. In the future, this network may be considered for implementation on real biological neural networks to decrease power consumption.
Collapse
Affiliation(s)
- Brosnan Yuen
- Department of Electrical and Computer Engineering, University of Victoria, 3800 Finnerty Road, Victoria, V8P 5C2, BC, Canada
| | - Xiaodai Dong
- Department of Electrical and Computer Engineering, University of Victoria, 3800 Finnerty Road, Victoria, V8P 5C2, BC, Canada.
| | - Tao Lu
- Department of Electrical and Computer Engineering, University of Victoria, 3800 Finnerty Road, Victoria, V8P 5C2, BC, Canada.
| |
Collapse
|
48
|
Qi L, Jiang Z, Shi W, Qu F, Feng G. GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling. Comput Biol Med 2024; 176:108547. [PMID: 38728994 DOI: 10.1016/j.compbiomed.2024.108547] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 04/07/2024] [Accepted: 04/28/2024] [Indexed: 05/12/2024]
Abstract
Self-supervised pre-training and fully supervised fine-tuning paradigms have received much attention to solve the data annotation problem in deep learning fields. Compared with traditional pre-training on large natural image datasets, medical self-supervised learning methods learn rich representations derived from unlabeled data itself thus avoiding the distribution shift between different image domains. However, nowadays state-of-the-art medical pre-training methods were specifically designed for downstream tasks making them less flexible and difficult to apply to new tasks. In this paper, we propose grid mask image modeling, a flexible and general self-supervised method to pre-train medical vision transformers for 3D medical image segmentation. Our goal is to guide networks to learn the correlations between organs and tissues by reconstructing original images based on partial observations. The relationships are consistent within the human body and invariant to disease type or imaging modality. To achieve this, we design a Siamese framework consisting of an online branch and a target branch. An adaptive and hierarchical masking strategy is employed in the online branch to (1) learn the boundaries or small contextual mutation regions within images; (2) to learn high-level semantic representations from deeper layers of the multiscale encoder. In addition, the target branch provides representations for contrastive learning to further reduce representation redundancy. We evaluate our method through segmentation performance on two public datasets. The experimental results demonstrate our method outperforms other self-supervised methods. Codes are available at https://github.com/mobiletomb/Gmim.
Collapse
Affiliation(s)
- Liangce Qi
- Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China.
| | - Zhengang Jiang
- Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China; Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, 528400, Guangzhou, China.
| | - Weili Shi
- Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China; Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, 528400, Guangzhou, China
| | - Feng Qu
- Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China
| | - Guanyuan Feng
- Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China
| |
Collapse
|
49
|
Li S, Wang H, Meng Y, Zhang C, Song Z. Multi-organ segmentation: a progressive exploration of learning paradigms under scarce annotation. Phys Med Biol 2024; 69:11TR01. [PMID: 38479023 DOI: 10.1088/1361-6560/ad33b5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2023] [Accepted: 03/13/2024] [Indexed: 05/21/2024]
Abstract
Precise delineation of multiple organs or abnormal regions in the human body from medical images plays an essential role in computer-aided diagnosis, surgical simulation, image-guided interventions, and especially in radiotherapy treatment planning. Thus, it is of great significance to explore automatic segmentation approaches, among which deep learning-based approaches have evolved rapidly and witnessed remarkable progress in multi-organ segmentation. However, obtaining an appropriately sized and fine-grained annotated dataset of multiple organs is extremely hard and expensive. Such scarce annotation limits the development of high-performance multi-organ segmentation models but promotes many annotation-efficient learning paradigms. Among these, studies on transfer learning leveraging external datasets, semi-supervised learning including unannotated datasets and partially-supervised learning integrating partially-labeled datasets have led the dominant way to break such dilemmas in multi-organ segmentation. We first review the fully supervised method, then present a comprehensive and systematic elaboration of the 3 abovementioned learning paradigms in the context of multi-organ segmentation from both technical and methodological perspectives, and finally summarize their challenges and future trends.
Collapse
Affiliation(s)
- Shiman Li
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, People's Republic of China
| | - Haoran Wang
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, People's Republic of China
| | - Yucong Meng
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, People's Republic of China
| | - Chenxi Zhang
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, People's Republic of China
| | - Zhijian Song
- Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, People's Republic of China
| |
Collapse
|
50
|
Zhang J, Xia L, Tang J, Xia J, Liu Y, Zhang W, Liu J, Liang Z, Zhang X, Zhang L, Tang G. Constructing a Deep Learning Radiomics Model Based on X-ray Images and Clinical Data for Predicting and Distinguishing Acute and Chronic Osteoporotic Vertebral Fractures: A Multicenter Study. Acad Radiol 2024; 31:2011-2026. [PMID: 38016821 DOI: 10.1016/j.acra.2023.10.061] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 09/13/2023] [Accepted: 10/31/2023] [Indexed: 11/30/2023]
Abstract
RATIONALE AND OBJECTIVES To construct and validate a deep learning radiomics (DLR) model based on X-ray images for predicting and distinguishing acute and chronic osteoporotic vertebral fractures (OVFs). METHODS A total of 942 cases (1076 vertebral bodies) with both vertebral X-ray examination and MRI scans were included in this study from three hospitals. They were divided into a training cohort (n = 712), an internal validation cohort (n = 178), an external validation cohort (n = 111), and a prospective validation cohort (n = 75). The ResNet-50 model architecture was used for deep transfer learning (DTL), with pre-training performed on RadImageNet and ImageNet datasets. DTL features and radiomics features were extracted from lateral X-ray images of OVFs patients and fused together. A logistic regression model with the least absolute shrinkage and selection operator was established, with MRI showing bone marrow edema as the gold standard for acute OVFs. The performance of the model was evaluated using receiver operating characteristic curves. Eight machine learning classification models were evaluated for their ability to distinguish between acute and chronic OVFs. The Nomogram was constructed by combining clinical baseline data to achieve visualized classification assessment. The predictive performance of the best RadImageNet model and ImageNet model was compared using the Delong test. The clinical value of the Nomogram was evaluated using decision curve analysis (DCA). RESULTS Pre-training resulted in 34 and 39 fused features after feature selection and fusion. The most effective machine learning algorithm in both DLR models was Light Gradient Boosting Machine. Using the Delong test, the area under the curve (AUC) for distinguishing between acute and chronic OVFs in the training cohort was 0.979 and 0.972 for the RadImageNet and ImageNet models, respectively, with no statistically significant difference between them (P = 0.235). In the internal validation cohort, external validation cohort, and prospective validation cohort, the AUCs for the two models were 0.967 vs 0.629, 0.886 vs 0.817, and 0.933 vs 0.661, respectively, with statistically significant differences in all comparisons (P < 0.05). The deep learning radiomics nomogram (DLRN) was constructed by combining the predictive model of RadImageNet with clinical baseline features, resulting in AUCs of 0.981, 0.974, 0.895, and 0.902 in the training cohort, internal validation cohort, external validation cohort, and prospective validation cohort, respectively. Using the Delong test, the AUCs for the fused feature model and the DLRN in the training cohort were 0.979 and 0.981, respectively, with no statistically significant difference between them (P = 0.169). In the internal validation cohort, external validation cohort, and prospective validation cohort, the AUCs for the two models were 0.967 vs 0.974, 0.886 vs 0.895, and 0.933 vs 0.902, respectively, with statistically significant differences in all comparisons (P < 0.05). The Nomogram showed a slight improvement in predictive performance in the internal and external validation cohort, but a slight decrease in the prospective validation cohort (0.933 vs 0.902). DCA showed that the Nomogram provided more benefits to patients compared to the DLR models. CONCLUSION Compared to the ImageNet model, the RadImageNet model has higher diagnostic value in distinguishing between acute and chronic OVFs. Furthermore, the diagnostic performance of the model is further improved when combined with clinical baseline features to construct the Nomogram.
Collapse
Affiliation(s)
- Jun Zhang
- Department of Radiology, Shanghai Tenth People's Hospital, Clinical Medical College of Nanjing Medical University, 301 Middle Yanchang Road, Shanghai, 200072, PR China (J.Z., G.T.); Department of Radiology, Sir RunRun Hospital affiliated to Nanjing Medical University, 109 Longmian Road, Nanjing, Jiangsu, 211002, PR China (J.Z., L.X., W.Z., J.L., Z.L.)
| | - Liang Xia
- Department of Radiology, Sir RunRun Hospital affiliated to Nanjing Medical University, 109 Longmian Road, Nanjing, Jiangsu, 211002, PR China (J.Z., L.X., W.Z., J.L., Z.L.)
| | - Jun Tang
- Department of Radiology, The Affiliated Taizhou People's Hospital of Nanjing Medical University, 366 Taihu Road, Taizhou, Jiangsu, 225300, PR China (J.T., J.X.)
| | - Jianguo Xia
- Department of Radiology, The Affiliated Taizhou People's Hospital of Nanjing Medical University, 366 Taihu Road, Taizhou, Jiangsu, 225300, PR China (J.T., J.X.)
| | - Yongkang Liu
- Department of Radiology, The Affiliated Hospital of Nanjing University of Chinese Medicine, 155 Hanzhong Road, Nanjing, Jiangsu, 210004, PR China (Y.L.)
| | - Weixiao Zhang
- Department of Radiology, Sir RunRun Hospital affiliated to Nanjing Medical University, 109 Longmian Road, Nanjing, Jiangsu, 211002, PR China (J.Z., L.X., W.Z., J.L., Z.L.)
| | - Jiayi Liu
- Department of Radiology, Sir RunRun Hospital affiliated to Nanjing Medical University, 109 Longmian Road, Nanjing, Jiangsu, 211002, PR China (J.Z., L.X., W.Z., J.L., Z.L.)
| | - Zhipeng Liang
- Department of Radiology, Sir RunRun Hospital affiliated to Nanjing Medical University, 109 Longmian Road, Nanjing, Jiangsu, 211002, PR China (J.Z., L.X., W.Z., J.L., Z.L.)
| | - Xueli Zhang
- Department of Radiology, Shanghai Tenth People's Hospital, Tongji University School of Medicine, 301 Middle Yanchang Road, Shanghai, 200072, PR China (X.Z., L.Z., G.T.)
| | - Lin Zhang
- Department of Radiology, Shanghai Tenth People's Hospital, Tongji University School of Medicine, 301 Middle Yanchang Road, Shanghai, 200072, PR China (X.Z., L.Z., G.T.).
| | - Guangyu Tang
- Department of Radiology, Shanghai Tenth People's Hospital, Clinical Medical College of Nanjing Medical University, 301 Middle Yanchang Road, Shanghai, 200072, PR China (J.Z., G.T.); Department of Radiology, Shanghai Tenth People's Hospital, Tongji University School of Medicine, 301 Middle Yanchang Road, Shanghai, 200072, PR China (X.Z., L.Z., G.T.)
| |
Collapse
|