51
|
Shetty S, Talaat W, AlKawas S, Al-Rawi N, Reddy S, Hamdoon Z, Kheder W, Acharya A, Ozsahin DU, David LR. Application of artificial intelligence-based detection of furcation involvement in mandibular first molar using cone beam tomography images- a preliminary study. BMC Oral Health 2024; 24:1476. [PMID: 39633335 PMCID: PMC11619149 DOI: 10.1186/s12903-024-05268-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2024] [Accepted: 11/27/2024] [Indexed: 12/07/2024] Open
Abstract
BACKGROUND Radiographs play a key role in diagnosis of periodontal diseases. Deep learning models have been explored for image analysis in periodontal diseases. However, there is lacuna of research in the deep learning model-based detection of furcation involvements [FI]. The objective of this study was to determine the accuracy of deep learning model in the detection of FI in axial CBCT images. METHODOLOGY We obtained initial dataset 285 axial CBCT images among which 143 were normal (without FI) and 142 were abnormal (with FI). Data augmentation technique was used to create 600(300 normal and 300 abnormal) images by using 200 images from the training dataset. Remaining 85(43 normal and 42 abnormal) images were kept for testing of model. ResNet101V2 with transfer learning was used employed for the analysis of images. RESULTS Training accuracy of model is 98%, valid accuracy is 97% and test accuracy is 91%. The precision and F1 score were 0.98 and 0.98 respectively. The Area under curve (AUC) was reported at 0.98. The test loss was reported at 0.2170. CONCLUSION The deep learning model (ResNet101V2) can accurately detect the FI in axial CBCT images. However, since our study was preliminary in nature and carried out with relatively smaller dataset, a study with larger dataset will further confirm the accuracy of deep learning models.
Collapse
Affiliation(s)
- Shishir Shetty
- Department of Oral and Craniofacial Health Sciences, College of Dental Medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Wael Talaat
- Department of Oral and Craniofacial Health Sciences, College of Dental Medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Sausan AlKawas
- Department of Oral and Craniofacial Health Sciences, College of Dental Medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Natheer Al-Rawi
- Department of Oral and Craniofacial Health Sciences, College of Dental Medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Sesha Reddy
- College of Dentistry, Gulf Medical University, Ajman, United Arab Emirates
| | - Zaid Hamdoon
- Department of Oral and Craniofacial Health Sciences, College of Dental Medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Waad Kheder
- Department of Preventive and Restorative Dentistry College of Dental medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Anirudh Acharya
- Department of Preventive and Restorative Dentistry College of Dental medicine, University of Sharjah, Sharjah, United Arab Emirates
| | - Dilber Uzun Ozsahin
- Department of Medical Diagnostic Imaging, College of Health Sciences, University of Sharjah, Sharjah, United Arab Emirates.
| | - Leena R David
- Department of Medical Diagnostic Imaging, College of Health Sciences, University of Sharjah, Sharjah, United Arab Emirates.
| |
Collapse
|
52
|
Li S, Xu M, Meng Y, Sun H, Zhang T, Yang H, Li Y, Ma X. The application of the combination between artificial intelligence and endoscopy in gastrointestinal tumors. MEDCOMM – ONCOLOGY 2024; 3. [DOI: 10.1002/mog2.91] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/29/2023] [Accepted: 09/03/2024] [Indexed: 01/04/2025]
Abstract
AbstractGastrointestinal (GI) tumors have always been a major type of malignant tumor and a leading cause of tumor‐related deaths worldwide. The main principles of modern medicine for GI tumors are early prevention, early diagnosis, and early treatment, with early diagnosis being the most effective measure. Endoscopy, due to its ability to visualize lesions, has been one of the primary modalities for screening, diagnosing, and treating GI tumors. However, a qualified endoscopist often requires long training and extensive experience, which to some extent limits the wider use of endoscopy. With advances in data science, artificial intelligence (AI) has brought a new development direction for the endoscopy of GI tumors. AI can quickly process large quantities of data and images and improve diagnostic accuracy with some training, greatly reducing the workload of endoscopists and assisting them in early diagnosis. Therefore, this review focuses on the combined application of endoscopy and AI in GI tumors in recent years, describing the latest research progress on the main types of tumors and their performance in clinical trials, the application of multimodal AI in endoscopy, the development of endoscopy, and the potential applications of AI within it, with the aim of providing a reference for subsequent research.
Collapse
Affiliation(s)
- Shen Li
- Department of Biotherapy Cancer Center, West China Hospital, West China Medical School Sichuan University Chengdu China
| | - Maosen Xu
- Laboratory of Aging Research and Cancer Drug Target, State Key Laboratory of Biotherapy, West China Hospital, National Clinical Research, Sichuan University Chengdu Sichuan China
| | - Yuanling Meng
- West China School of Stomatology Sichuan University Chengdu Sichuan China
| | - Haozhen Sun
- College of Life Sciences Sichuan University Chengdu Sichuan China
| | - Tao Zhang
- Department of Biotherapy Cancer Center, West China Hospital, West China Medical School Sichuan University Chengdu China
| | - Hanle Yang
- Department of Biotherapy Cancer Center, West China Hospital, West China Medical School Sichuan University Chengdu China
| | - Yueyi Li
- Department of Biotherapy Cancer Center, West China Hospital, West China Medical School Sichuan University Chengdu China
| | - Xuelei Ma
- Department of Biotherapy Cancer Center, West China Hospital, West China Medical School Sichuan University Chengdu China
| |
Collapse
|
53
|
Boers TGW, Fockens KN, van der Putten JA, Jaspers TJM, Kusters CHJ, Jukema JB, Jong MR, Struyvenberg MR, de Groof J, Bergman JJ, de With PHN, van der Sommen F. Foundation models in gastrointestinal endoscopic AI: Impact of architecture, pre-training approach and data efficiency. Med Image Anal 2024; 98:103298. [PMID: 39173410 DOI: 10.1016/j.media.2024.103298] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 07/18/2024] [Accepted: 08/06/2024] [Indexed: 08/24/2024]
Abstract
Pre-training deep learning models with large data sets of natural images, such as ImageNet, has become the standard for endoscopic image analysis. This approach is generally superior to training from scratch, due to the scarcity of high-quality medical imagery and labels. However, it is still unknown whether the learned features on natural imagery provide an optimal starting point for the downstream medical endoscopic imaging tasks. Intuitively, pre-training with imagery closer to the target domain could lead to better-suited feature representations. This study evaluates whether leveraging in-domain pre-training in gastrointestinal endoscopic image analysis has potential benefits compared to pre-training on natural images. To this end, we present a dataset comprising of 5,014,174 gastrointestinal endoscopic images from eight different medical centers (GastroNet-5M), and exploit self-supervised learning with SimCLRv2, MoCov2 and DINO to learn relevant features for in-domain downstream tasks. The learned features are compared to features learned on natural images derived with multiple methods, and variable amounts of data and/or labels (e.g. Billion-scale semi-weakly supervised learning and supervised learning on ImageNet-21k). The effects of the evaluation is performed on five downstream data sets, particularly designed for a variety of gastrointestinal tasks, for example, GIANA for angiodyplsia detection and Kvasir-SEG for polyp segmentation. The findings indicate that self-supervised domain-specific pre-training, specifically using the DINO framework, results into better performing models compared to any supervised pre-training on natural images. On the ResNet50 and Vision-Transformer-small architectures, utilizing self-supervised in-domain pre-training with DINO leads to an average performance boost of 1.63% and 4.62%, respectively, on the downstream datasets. This improvement is measured against the best performance achieved through pre-training on natural images within any of the evaluated frameworks. Moreover, the in-domain pre-trained models also exhibit increased robustness against distortion perturbations (noise, contrast, blur, etc.), where the in-domain pre-trained ResNet50 and Vision-Transformer-small with DINO achieved on average 1.28% and 3.55% higher on the performance metrics, compared to the best performance found for pre-trained models on natural images. Overall, this study highlights the importance of in-domain pre-training for improving the generic nature, scalability and performance of deep learning for medical image analysis. The GastroNet-5M pre-trained weights are made publicly available in our repository: huggingface.co/tgwboers/GastroNet-5M_Pretrained_Weights.
Collapse
Affiliation(s)
- Tim G W Boers
- Eindhoven University of Technology, Groene Loper 3, 5612 AE Eindhoven, The Netherlands.
| | - Kiki N Fockens
- Amsterdam UMC, Location VUmc, De Boelelaan 1117, 1081 HV Amsterdam, The Netherlands
| | | | - Tim J M Jaspers
- Eindhoven University of Technology, Groene Loper 3, 5612 AE Eindhoven, The Netherlands
| | - Carolus H J Kusters
- Eindhoven University of Technology, Groene Loper 3, 5612 AE Eindhoven, The Netherlands
| | - Jelmer B Jukema
- Amsterdam UMC, Location VUmc, De Boelelaan 1117, 1081 HV Amsterdam, The Netherlands
| | - Martijn R Jong
- Amsterdam UMC, Location VUmc, De Boelelaan 1117, 1081 HV Amsterdam, The Netherlands
| | | | - Jeroen de Groof
- Amsterdam UMC, Location VUmc, De Boelelaan 1117, 1081 HV Amsterdam, The Netherlands
| | - Jacques J Bergman
- Amsterdam UMC, Location VUmc, De Boelelaan 1117, 1081 HV Amsterdam, The Netherlands
| | - Peter H N de With
- Eindhoven University of Technology, Groene Loper 3, 5612 AE Eindhoven, The Netherlands
| | - Fons van der Sommen
- Eindhoven University of Technology, Groene Loper 3, 5612 AE Eindhoven, The Netherlands
| |
Collapse
|
54
|
Chau M. Ethical, legal, and regulatory landscape of artificial intelligence in Australian healthcare and ethical integration in radiography: A narrative review. J Med Imaging Radiat Sci 2024; 55:101733. [PMID: 39111223 DOI: 10.1016/j.jmir.2024.101733] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2024] [Revised: 07/15/2024] [Accepted: 07/16/2024] [Indexed: 12/02/2024]
Abstract
This narrative review explores the ethical, legal, and regulatory landscape of AI integration in Australian healthcare, focusing on radiography. It examines the current legislative framework, assesses the trust and reliability of AI tools, and proposes future directions for ethical AI integration in radiography. AI systems significantly enhance diagnostic radiography by improving diagnostic accuracy and efficiency in stroke detection, brain imaging, and chest reporting. However, AI raises substantial ethical concerns due to its 'black-box' nature and potential biases in training data. The Therapeutic Goods Administration's reforms in Australia, though comprehensive, fall short of fully addressing issues related to the trustworthiness and legal liabilities of AI tools. Adopting a comprehensive research strategy that includes doctrinal, comparative, and public policy analyses will facilitate an understanding of international practices, particularly from countries with similar legal systems, and help guide Australia in refining its regulatory framework. For an ethical future in radiography, a robust, multi-disciplinary approach is required to prioritize patient safety, data privacy, and equitable AI use. A framework that balances technological innovation with ethical and legal integrity is essential for advancing healthcare while preserving trust and transparency. Healthcare professionals, policymakers, and AI developers must collaborate to establish a resilient, equitable, and transparent healthcare system. Future research should focus on multi-disciplinary methodologies, combining doctrinal, comparative, and public policy research to provide comprehensive insights. This approach will guide Australia in creating a more inclusive and ethically sound legal framework for AI in healthcare, ensuring its ethical and beneficial integration into radiography.
Collapse
Affiliation(s)
- Minh Chau
- Faculty of Science and Health, Charles Sturt University, Level 5, 250 Boorooma St, Charles Sturt University NSW 2678, Australia; South Australia Medical Imaging, Flinders Medical Centre, 1 Flinders Drive, Bedford Park, SA 5042, Australia.
| |
Collapse
|
55
|
Kulkarni P, Sarwe N, Pingale A, Sarolkar Y, Patil RR, Shinde G, Kaur G. Exploring the efficacy of various CNN architectures in diagnosing oral cancer from squamous cell carcinoma. MethodsX 2024; 13:103034. [PMID: 39610794 PMCID: PMC11603122 DOI: 10.1016/j.mex.2024.103034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2024] [Accepted: 11/04/2024] [Indexed: 11/30/2024] Open
Abstract
Oral cancer can result from mutations in cells located in the lips or mouth. Diagnosing oral cavity squamous cell carcinoma (OCSCC) is particularly challenging, often occurring at advanced stages. To address this, computer-aided diagnosis methods are increasingly being used. In this work, a deep learning-based approach utilizing models such as VGG16, ResNet50, LeNet-5, MobileNetV2, and Inception V3 is presented. NEOR and OCSCC datasets were used for feature extraction, with virtual slide images divided into tiles and classified as normal or squamous cell cancer. Performance metrics like accuracy, F1-score, AUC, precision, and recall were analyzed to determine the prerequisites for optimal CNN performance. The proposed CNN approaches were effective for classifying OCSCC and oral dysplasia, with the highest accuracy of 95.41 % achieved using MobileNetV2. Key findings Deep learning models, particularly MobileNetV2, achieved high classification accuracy (95.41 %) for OCSCC.CNN-based methods show promise for early-stage OCSCC and oral dysplasia diagnosis. Performance parameters like precision, recall, and F1-score help optimize CNN model selection for this task.
Collapse
Affiliation(s)
- Prerna Kulkarni
- Department of CSE (AIML), Vishwakarma Institute of Information Technology, Kondhwa (Budruk) Pune, Maharashtra 411048, India
| | - Nidhi Sarwe
- Department of CSE (AIML), Vishwakarma Institute of Information Technology, Kondhwa (Budruk) Pune, Maharashtra 411048, India
| | - Abhishek Pingale
- Department of CSE (AIML), Vishwakarma Institute of Information Technology, Kondhwa (Budruk) Pune, Maharashtra 411048, India
| | - Yash Sarolkar
- Department of CSE (AIML), Vishwakarma Institute of Information Technology, Kondhwa (Budruk) Pune, Maharashtra 411048, India
| | - Rutuja Rajendra Patil
- Department of CSE (AIML), Vishwakarma Institute of Information Technology, Kondhwa (Budruk) Pune, Maharashtra 411048, India
| | - Gitanjali Shinde
- Department of CSE (AIML), Vishwakarma Institute of Information Technology, Kondhwa (Budruk) Pune, Maharashtra 411048, India
| | - Gagandeep Kaur
- CSE Department, Symbiosis Institute of Technology, Nagpur Campus, Symbiosis International (Deemed University), Pune, India
| |
Collapse
|
56
|
Grossetête L, Marcelot C, Gatel C, Pauchet S, Hytch M. Principle of TEM alignment using convolutional neural networks: Case study on condenser aperture alignment. Ultramicroscopy 2024; 267:114047. [PMID: 39413637 DOI: 10.1016/j.ultramic.2024.114047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 08/07/2024] [Accepted: 09/11/2024] [Indexed: 10/18/2024]
Abstract
The possibility of automatically aligning the transmission electron microscope (TEM) is explored using an approach based on artificial intelligence (AI). After presenting the general concept, we test the method on the first step of the alignment process which involves centering the condenser aperture. We propose using a convolutional neural network (CNN) that learns to predict the x and y-shifts needed to realign the aperture in one step. The learning data sets were acquired automatically on the microscope by using a simplified digital twin. Different models were tested and analysed to choose the optimal design. We have developed a human-level estimator and intend to use it safely on all apertures. A similar process could be used for most steps of the alignment process with minimal changes, allowing microscopists to reduce the time and training required to perform this task. The method is also compatible with continuous correction of alignment drift during lengthy experiments or to ensure uniformity of illumination conditions during data acquisition.
Collapse
Affiliation(s)
- Loïc Grossetête
- CEMES-CNRS, 29 rue Jeanne Marvig, Toulouse, 31055, France; Fédération ENAC ISAE-SUPAERO ONERA, 7 Avenue Edouard Belin, Toulouse, 31055, France.
| | | | | | - Sylvain Pauchet
- Fédération ENAC ISAE-SUPAERO ONERA, 7 Avenue Edouard Belin, Toulouse, 31055, France
| | - Martin Hytch
- CEMES-CNRS, 29 rue Jeanne Marvig, Toulouse, 31055, France
| |
Collapse
|
57
|
Jun E, Jeong S, Heo DW, Suk HI. Medical Transformer: Universal Encoder for 3-D Brain MRI Analysis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:17779-17789. [PMID: 37738193 DOI: 10.1109/tnnls.2023.3308712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/24/2023]
Abstract
Transfer learning has attracted considerable attention in medical image analysis because of the limited number of annotated 3-D medical datasets available for training data-driven deep learning models in the real world. We propose Medical Transformer, a novel transfer learning framework that effectively models 3-D volumetric images as a sequence of 2-D image slices. To improve the high-level representation in 3-D-form empowering spatial relations, we use a multiview approach that leverages information from three planes of the 3-D volume, while providing parameter-efficient training. For building a source model generally applicable to various tasks, we pretrain the model using self-supervised learning (SSL) for masked encoding vector prediction as a proxy task, using a large-scale normal, healthy brain magnetic resonance imaging (MRI) dataset. Our pretrained model is evaluated on three downstream tasks: 1) brain disease diagnosis; 2) brain age prediction; and 3) brain tumor segmentation, which are widely studied in brain MRI research. Experimental results demonstrate that our Medical Transformer outperforms the state-of-the-art (SOTA) transfer learning methods, efficiently reducing the number of parameters by up to approximately 92% for classification and regression tasks and 97% for segmentation task, and it also achieves good performance in scenarios where only partial training samples are used.
Collapse
|
58
|
Hosseini MS, Bejnordi BE, Trinh VQH, Chan L, Hasan D, Li X, Yang S, Kim T, Zhang H, Wu T, Chinniah K, Maghsoudlou S, Zhang R, Zhu J, Khaki S, Buin A, Chaji F, Salehi A, Nguyen BN, Samaras D, Plataniotis KN. Computational pathology: A survey review and the way forward. J Pathol Inform 2024; 15:100357. [PMID: 38420608 PMCID: PMC10900832 DOI: 10.1016/j.jpi.2023.100357] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2023] [Revised: 12/21/2023] [Accepted: 12/23/2023] [Indexed: 03/02/2024] Open
Abstract
Computational Pathology (CPath) is an interdisciplinary science that augments developments of computational approaches to analyze and model medical histopathology images. The main objective for CPath is to develop infrastructure and workflows of digital diagnostics as an assistive CAD system for clinical pathology, facilitating transformational changes in the diagnosis and treatment of cancer that are mainly address by CPath tools. With evergrowing developments in deep learning and computer vision algorithms, and the ease of the data flow from digital pathology, currently CPath is witnessing a paradigm shift. Despite the sheer volume of engineering and scientific works being introduced for cancer image analysis, there is still a considerable gap of adopting and integrating these algorithms in clinical practice. This raises a significant question regarding the direction and trends that are undertaken in CPath. In this article we provide a comprehensive review of more than 800 papers to address the challenges faced in problem design all-the-way to the application and implementation viewpoints. We have catalogued each paper into a model-card by examining the key works and challenges faced to layout the current landscape in CPath. We hope this helps the community to locate relevant works and facilitate understanding of the field's future directions. In a nutshell, we oversee the CPath developments in cycle of stages which are required to be cohesively linked together to address the challenges associated with such multidisciplinary science. We overview this cycle from different perspectives of data-centric, model-centric, and application-centric problems. We finally sketch remaining challenges and provide directions for future technical developments and clinical integration of CPath. For updated information on this survey review paper and accessing to the original model cards repository, please refer to GitHub. Updated version of this draft can also be found from arXiv.
Collapse
Affiliation(s)
- Mahdi S. Hosseini
- Department of Computer Science and Software Engineering (CSSE), Concordia Univeristy, Montreal, QC H3H 2R9, Canada
| | | | - Vincent Quoc-Huy Trinh
- Institute for Research in Immunology and Cancer of the University of Montreal, Montreal, QC H3T 1J4, Canada
| | - Lyndon Chan
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Danial Hasan
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Xingwen Li
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Stephen Yang
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Taehyo Kim
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Haochen Zhang
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Theodore Wu
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Kajanan Chinniah
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Sina Maghsoudlou
- Department of Computer Science and Software Engineering (CSSE), Concordia Univeristy, Montreal, QC H3H 2R9, Canada
| | - Ryan Zhang
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Jiadai Zhu
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Samir Khaki
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| | - Andrei Buin
- Huron Digitial Pathology, St. Jacobs, ON N0B 2N0, Canada
| | - Fatemeh Chaji
- Department of Computer Science and Software Engineering (CSSE), Concordia Univeristy, Montreal, QC H3H 2R9, Canada
| | - Ala Salehi
- Department of Electrical and Computer Engineering, University of New Brunswick, Fredericton, NB E3B 5A3, Canada
| | - Bich Ngoc Nguyen
- University of Montreal Hospital Center, Montreal, QC H2X 0C2, Canada
| | - Dimitris Samaras
- Department of Computer Science, Stony Brook University, Stony Brook, NY 11794, United States
| | - Konstantinos N. Plataniotis
- The Edward S. Rogers Sr. Department of Electrical & Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 3G4, Canada
| |
Collapse
|
59
|
Kim S, Kang Y, Shin H, Lee EB, Ham BJ, Choi Y. Liquid Biopsy-Based Detection and Response Prediction for Depression. ACS NANO 2024; 18:32498-32507. [PMID: 39501510 PMCID: PMC11604100 DOI: 10.1021/acsnano.4c08233] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/20/2024] [Revised: 10/25/2024] [Accepted: 10/29/2024] [Indexed: 11/27/2024]
Abstract
Proactively predicting antidepressant treatment response before medication failures is crucial, as it reduces unsuccessful attempts and facilitates the development of personalized therapeutic strategies, ultimately enhancing treatment efficacy. The current decision-making process, which heavily depends on subjective indicators, underscores the need for an objective, indicator-based approach. This study developed a method for detecting depression and predicting treatment response through deep learning-based spectroscopic analysis of extracellular vesicles (EVs) from plasma. EVs were isolated from the plasma of both nondepressed and depressed groups, followed by Raman signal acquisition, which was used for AI algorithm development. The algorithm successfully distinguished depression patients from healthy individuals and those with panic disorder, achieving an AUC accuracy of 0.95. This demonstrates the model's capability to selectively diagnose depression within a nondepressed group, including those with other mental health disorders. Furthermore, the algorithm identified depression-diagnosed patients likely to respond to antidepressants, classifying responders and nonresponders with an AUC accuracy of 0.91. To establish a diagnostic foundation, the algorithm applied explainable AI (XAI), enabling personalized medicine for companion diagnostics and highlighting its potential for the development of liquid biopsy-based mental disorder diagnosis.
Collapse
Affiliation(s)
- Seungmin Kim
- Department
of Biomedical Engineering, Korea University, Seoul 02841, Republic of Korea
- Interdisciplinary
Program in Precision Public Health, Korea
University, Seoul 02841, Republic of Korea
| | - Youbin Kang
- Department
of Biomedical Sciences, Korea University
College of Medicine, Seoul 02841, Republic
of Korea
| | - Hyunku Shin
- Exopert
Corporation, Seoul 02841, Republic of Korea
| | - Eun Byul Lee
- Exopert
Corporation, Seoul 02841, Republic of Korea
| | - Byung-Joo Ham
- Department
of Biomedical Sciences, Korea University
College of Medicine, Seoul 02841, Republic
of Korea
| | - Yeonho Choi
- Department
of Biomedical Engineering, Korea University, Seoul 02841, Republic of Korea
- Interdisciplinary
Program in Precision Public Health, Korea
University, Seoul 02841, Republic of Korea
- Exopert
Corporation, Seoul 02841, Republic of Korea
- School
of Bioengineering, Korea University, Seoul 02841, Republic of Korea
| |
Collapse
|
60
|
Lin L, Zheng Y, Li Y, Jiang D, Cao J, Wang J, Xiao Y, Mao X, Zheng C, Wang Y. Automatic vessel segmentation and reformation of non-contrast coronary magnetic resonance angiography using transfer learning-based three-dimensional U-net with attention mechanism. J Cardiovasc Magn Reson 2024; 27:101126. [PMID: 39581550 DOI: 10.1016/j.jocmr.2024.101126] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Revised: 10/27/2024] [Accepted: 11/18/2024] [Indexed: 11/26/2024] Open
Abstract
BACKGROUND Coronary magnetic resonance angiography (CMRA) presents distinct advantages, but its reliance on manual image post-processing is labor-intensive and requires specialized knowledge. This study aims to design and test an efficient artificial intelligence (AI) model capable of automating coronary artery segmentation and reformation from CMRA images for coronary artery disease (CAD) diagnosis. METHODS By leveraging transfer learning from a pre-existing coronary computed tomography angiography model, a three-dimensional attention-aware U-Net was established, trained, and validated on a dataset of 104 subjects' CMRA. Furthermore, an independent clinical evaluation was conducted on an additional cohort of 70 patients. The AI model's performance in segmenting coronary arteries was assessed using the Dice similarity coefficient (DSC) and recall. The comparison between the AI model and manual processing by experienced radiologists on vessel reformation was based on reformatted image quality (rIQ) scoring, post-processing time, and the number of necessary user interactions. The diagnostic performance of AI-segmented CMRA for significant stenosis (≥50% diameter reduction) was evaluated using conventional coronary angiography (CAG) as a reference in sub-set data. RESULTS The DSC of the AI model achieved on the training and validation sets were 0.952 and 0.944, with recalls of 0.936 and 0.923, respectively. In the clinical evaluation, the model outperformed manual processes by reducing vessel post-processing time, from 632.6±17.0 s to 77.4±8.9 s, and the number of user interactions from 221±59 to 8±2. The AI post-processed images maintained high rIQ scores comparable to those processed manually (2.7±0.8 vs 2.7±0.6; P = 0.4806). In subjects with CAG, the prevalence of CAD was 71%. The sensitivity, specificity, and accuracy at patient-based analysis were 94%, 71%, and 88%, respectively, by AI post-processed whole-heart CMRA. CONCLUSION The AI auto-segmentation system can effectively facilitate CMRA vessel reformation and reduce the time consumption for radiologists. It has the potential to become a standard component of daily workflows, optimizing the clinical application of CMRA in the future.
Collapse
Affiliation(s)
- Lu Lin
- Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Yijia Zheng
- Center for Biomedical Imaging Research, School of Biomedical Engineering, Tsinghua University, Beijing, China
| | - Yanyu Li
- Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | | | - Jian Cao
- Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Jian Wang
- Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | | | | | - Chao Zheng
- Shukun Technology Co., Ltd, Beijing, China
| | - Yining Wang
- Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China.
| |
Collapse
|
61
|
Stathopoulos I, Serio L, Karavasilis E, Kouri MA, Velonakis G, Kelekis N, Efstathopoulos E. Evaluating Brain Tumor Detection with Deep Learning Convolutional Neural Networks Across Multiple MRI Modalities. J Imaging 2024; 10:296. [PMID: 39728193 DOI: 10.3390/jimaging10120296] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2024] [Revised: 11/18/2024] [Accepted: 11/20/2024] [Indexed: 12/28/2024] Open
Abstract
Central Nervous System (CNS) tumors represent a significant public health concern due to their high morbidity and mortality rates. Magnetic Resonance Imaging (MRI) has emerged as a critical non-invasive modality for the detection, diagnosis, and management of brain tumors, offering high-resolution visualization of anatomical structures. Recent advancements in deep learning, particularly convolutional neural networks (CNNs), have shown potential in augmenting MRI-based diagnostic accuracy for brain tumor detection. In this study, we evaluate the diagnostic performance of six fundamental MRI sequences in detecting tumor-involved brain slices using four distinct CNN architectures enhanced with transfer learning techniques. Our dataset comprises 1646 MRI slices from the examinations of 62 patients, encompassing both tumor-bearing and normal findings. With our approach, we achieved a classification accuracy of 98.6%, underscoring the high potential of CNN-based models in this context. Additionally, we assessed the performance of each MRI sequence across the different CNN models, identifying optimal combinations of MRI modalities and neural networks to meet radiologists' screening requirements effectively. This study offers critical insights into the integration of deep learning with MRI for brain tumor detection, with implications for improving diagnostic workflows in clinical settings.
Collapse
Affiliation(s)
- Ioannis Stathopoulos
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
- Technology Department, CERN, 1211 Geneva, Switzerland
| | - Luigi Serio
- Technology Department, CERN, 1211 Geneva, Switzerland
| | - Efstratios Karavasilis
- Medical Physics Laboratory, School of Medicine, Democritus University of Thrace, 68100 Alexandroupolis, Greece
| | - Maria Anthi Kouri
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| | - Georgios Velonakis
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| | - Nikolaos Kelekis
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| | - Efstathios Efstathopoulos
- 2nd Department of Radiology, Medical School, Attikon University Hospital, National and Kapodistrian University of Athens, 11527 Athens, Greece
| |
Collapse
|
62
|
Tan TWK, Nguyen KN, Zhang C, Kong R, Cheng SF, Ji F, Chong JSX, Yi Chong EJ, Venketasubramanian N, Orban C, Chee MWL, Chen C, Zhou JH, Yeo BTT. Evaluation of Brain Age as a Specific Marker of Brain Health. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.11.16.623903. [PMID: 39605400 PMCID: PMC11601463 DOI: 10.1101/2024.11.16.623903] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 11/29/2024]
Abstract
Brain age is a powerful marker of general brain health. Furthermore, brain age models are trained on large datasets, thus giving them a potential advantage in predicting specific outcomes - much like the success of finetuning large language models for specific applications. However, it is also well-accepted in machine learning that models trained to directly predict specific outcomes (i.e., direct models) often perform better than those trained on surrogate outcomes. Therefore, despite their much larger training data, it is unclear whether brain age models outperform direct models in predicting specific brain health outcomes. Here, we compare large-scale brain age models and direct models for predicting specific health outcomes in the context of Alzheimer's Disease (AD) dementia. Using anatomical T1 scans from three continents (N = 1,848), we find that direct models outperform brain age models without finetuning. Finetuned brain age models yielded similar performance as direct models, but importantly, did not outperform direct models although the brain age models were pretrained on 1000 times more data than the direct models: N = 53,542 vs N = 50. Overall, our results do not discount brain age as a useful marker of general brain health. However, in this era of large-scale brain age models, our results suggest that small-scale, targeted approaches for extracting specific brain health markers still hold significant value.
Collapse
Affiliation(s)
- Trevor Wei Kiat Tan
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Department of Medicine, Healthy Longevity Translational Research Programme, Human Potential Translational Research Programme & Institute for Digital Medicine (WisDM), Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- N.1 Institute for Health, National University of Singapore, Singapore
- Integrative Sciences and Engineering Programme (ISEP), National University of Singapore, Singapore
| | - Kim-Ngan Nguyen
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | - Chen Zhang
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Department of Medicine, Healthy Longevity Translational Research Programme, Human Potential Translational Research Programme & Institute for Digital Medicine (WisDM), Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- N.1 Institute for Health, National University of Singapore, Singapore
| | - Ru Kong
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Department of Medicine, Healthy Longevity Translational Research Programme, Human Potential Translational Research Programme & Institute for Digital Medicine (WisDM), Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- N.1 Institute for Health, National University of Singapore, Singapore
| | - Susan F Cheng
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Integrative Sciences and Engineering Programme (ISEP), National University of Singapore, Singapore
| | - Fang Ji
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
| | - Joanna Su Xian Chong
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
| | - Eddie Jun Yi Chong
- Memory, Aging and Cognition Centre, National University Health System, Singapore
- Department of Psychological Medicine, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | | | - Csaba Orban
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Department of Medicine, Healthy Longevity Translational Research Programme, Human Potential Translational Research Programme & Institute for Digital Medicine (WisDM), Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- N.1 Institute for Health, National University of Singapore, Singapore
- Integrative Sciences and Engineering Programme (ISEP), National University of Singapore, Singapore
| | - Michael W L Chee
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Medicine, Healthy Longevity Translational Research Programme, Human Potential Translational Research Programme & Institute for Digital Medicine (WisDM), Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | - Christopher Chen
- Memory, Aging and Cognition Centre, National University Health System, Singapore
- Department of Psychological Medicine, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Pharmacology, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | - Juan Helen Zhou
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Integrative Sciences and Engineering Programme (ISEP), National University of Singapore, Singapore
| | - B T Thomas Yeo
- Centre for Sleep and Cognition & Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Department of Electrical and Computer Engineering, National University of Singapore, Singapore
- Department of Medicine, Healthy Longevity Translational Research Programme, Human Potential Translational Research Programme & Institute for Digital Medicine (WisDM), Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- N.1 Institute for Health, National University of Singapore, Singapore
- Integrative Sciences and Engineering Programme (ISEP), National University of Singapore, Singapore
- Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA, USA
| |
Collapse
|
63
|
Rundo L, Militello C. Image biomarkers and explainable AI: handcrafted features versus deep learned features. Eur Radiol Exp 2024; 8:130. [PMID: 39560820 PMCID: PMC11576747 DOI: 10.1186/s41747-024-00529-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2024] [Accepted: 10/16/2024] [Indexed: 11/20/2024] Open
Abstract
Feature extraction and selection from medical data are the basis of radiomics and image biomarker discovery for various architectures, including convolutional neural networks (CNNs). We herein describe the typical radiomics steps and the components of a CNN for both deep feature extraction and end-to-end approaches. We discuss the curse of dimensionality, along with dimensionality reduction techniques. Despite the outstanding performance of deep learning (DL) approaches, the use of handcrafted features instead of deep learned features needs to be considered for each specific study. Dataset size is a key factor: large-scale datasets with low sample diversity could lead to overfitting; limited sample sizes can provide unstable models. The dataset must be representative of all the "facets" of the clinical phenomenon/disease investigated. The access to high-performance computational resources from graphics processing units is another key factor, especially for the training phase of deep architectures. The advantages of multi-institutional federated/collaborative learning are described. When large language models are used, high stability is needed to avoid catastrophic forgetting in complex domain-specific tasks. We highlight that non-DL approaches provide model explainability superior to that provided by DL approaches. To implement explainability, the need for explainable AI arises, also through post hoc mechanisms. RELEVANCE STATEMENT: This work aims to provide the key concepts for processing the imaging features to extract reliable and robust image biomarkers. KEY POINTS: The key concepts for processing the imaging features to extract reliable and robust image biomarkers are provided. The main differences between radiomics and representation learning approaches are highlighted. The advantages and disadvantages of handcrafted versus learned features are given without losing sight of the clinical purpose of artificial intelligence models.
Collapse
Affiliation(s)
- Leonardo Rundo
- Department of Information and Electrical Engineering and Applied Mathematics (DIEM), University of Salerno, Fisciano, Salerno, Italy.
| | - Carmelo Militello
- High Performance Computing and Networking Institute (ICAR-CNR), Italian National Research Council, Palermo, Italy
| |
Collapse
|
64
|
Wang S, Liu J, Li S, He P, Zhou X, Zhao Z, Zheng L. ResNet-Transformer deep learning model-aided detection of dens evaginatus. Int J Paediatr Dent 2024. [PMID: 39545506 DOI: 10.1111/ipd.13282] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/01/2024] [Revised: 09/29/2024] [Accepted: 10/22/2024] [Indexed: 11/17/2024]
Abstract
BACKGROUND Dens evaginatus is a dental morphological developmental anomaly. Failing to detect it may lead to tubercles fracture and pulpal/periapical disease. Consequently, early detection and intervention of dens evaginatus are significant to preserve vital pulp. AIM This study aimed to develop a deep learning model to assist dentists in early diagnosing dens evaginatus, thereby supporting early intervention and mitigating the risk of severe consequences. DESIGN In this study, a deep learning model was developed utilizing panoramic radiograph images sourced from 1410 patients aged 3-16 years, with high-quality annotations to enable the automatic detection of dens evaginatus. Model performance and model's efficacy in aiding dentists were evaluated. RESULTS The findings indicated that the current deep learning model demonstrated commendable sensitivity (0.8600) and specificity (0.9200), outperforming dentists in detecting dens evaginatus with an F1-score of 0.8866 compared to their average F1-score of 0.8780, indicating that the model could detect dens evaginatus with greater precision. Furthermore, with its support, young dentists heightened their focus on dens evaginatus in tooth germs and achieved improved diagnostic accuracy. CONCLUSION Based on these results, the integration of deep learning for dens evaginatus detection holds significance and can augment dentists' proficiency in identifying such anomaly.
Collapse
Affiliation(s)
- Siwei Wang
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Pediatric Dentistry, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| | - Jialing Liu
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| | - Shihao Li
- Department of Biotherapy, Cancer Center, West China Hospital, Sichuan University, Chengdu, Sichuan, China
| | - Pengcheng He
- Pediatric Dentistry, West China Second University Hospital, Sichuan University, Chengdu, Sichuan, China
| | - Xin Zhou
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Pediatric Dentistry, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| | - Zhihe Zhao
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Orthodontics, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| | - Liwei Zheng
- State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Research Center for Oral Diseases & Department of Pediatric Dentistry, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China
| |
Collapse
|
65
|
Quan R, Qiu Z, Wan H, Yang Z, Li X. Dung beetle optimization algorithm-based hybrid deep learning model for ultra-short-term PV power prediction. iScience 2024; 27:111126. [PMID: 39524344 PMCID: PMC11546542 DOI: 10.1016/j.isci.2024.111126] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2024] [Revised: 08/10/2024] [Accepted: 10/04/2024] [Indexed: 11/16/2024] Open
Abstract
A hybrid model combining self-attention temporal convolutional networks (SATCN) with bidirectional long short-term memory (BiLSTM) networks was developed to improve the accuracy of ultra-short-term photovoltaic (PV) power prediction. The self-attention mechanism and SATCN were used to extract temporal and correlation features, which were then linked to BiLSTM networks. The model's hyperparameters were optimized using the dung beetle optimization algorithm. The model was tested on a year-long dataset of PV power and outperformed convolutional neural networks, BiLSTM networks, temporal convolutional networks, and other hybrid models. It reduced the root-mean-square error (RMSE) by 33.1% compared to the other models. The model achieved a mean absolute error (MAE) of 0.175, a weighted mean absolute percentage error (wMAPE) of 4.821, and a coefficient of determination (R2) of 0.997. These results highlight the model's superior accuracy and its potential applications in solar energy development.
Collapse
Affiliation(s)
- Rui Quan
- Hubei Key Laboratory for High-efficiency Utilization of Solar Energy and Operation Control of Energy Storage System, Hubei University of Technology, Wuhan 430068, China
- Hubei Engineering Research Center for Safety Monitoring of New Energy and Power Grid Equipment, Hubei University of Technology, Wuhan 430068, China
| | - Zhizhuo Qiu
- Hubei Key Laboratory for High-efficiency Utilization of Solar Energy and Operation Control of Energy Storage System, Hubei University of Technology, Wuhan 430068, China
- Hubei Engineering Research Center for Safety Monitoring of New Energy and Power Grid Equipment, Hubei University of Technology, Wuhan 430068, China
| | - Hang Wan
- Hubei Key Laboratory for High-efficiency Utilization of Solar Energy and Operation Control of Energy Storage System, Hubei University of Technology, Wuhan 430068, China
- Hubei Engineering Research Center for Safety Monitoring of New Energy and Power Grid Equipment, Hubei University of Technology, Wuhan 430068, China
| | - Zhiyu Yang
- Hubei Key Laboratory for High-efficiency Utilization of Solar Energy and Operation Control of Energy Storage System, Hubei University of Technology, Wuhan 430068, China
- Hubei Engineering Research Center for Safety Monitoring of New Energy and Power Grid Equipment, Hubei University of Technology, Wuhan 430068, China
| | - Xuerong Li
- Hubei Key Laboratory for High-efficiency Utilization of Solar Energy and Operation Control of Energy Storage System, Hubei University of Technology, Wuhan 430068, China
- Hubei Engineering Research Center for Safety Monitoring of New Energy and Power Grid Equipment, Hubei University of Technology, Wuhan 430068, China
| |
Collapse
|
66
|
Korkmaz OE, Guller H, Miloglu O, Ozbek İY, Oral EA, Guller MT. The detection of distomolar teeth on panoramic radiographs using different artificial intelligence models. JOURNAL OF STOMATOLOGY, ORAL AND MAXILLOFACIAL SURGERY 2024; 126:102151. [PMID: 39550006 DOI: 10.1016/j.jormas.2024.102151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/14/2024] [Accepted: 11/13/2024] [Indexed: 11/18/2024]
Abstract
PURPOSES One notable anomaly, presence of distomolars, arises beyond the typical sequence of the human dental system. In this study, convolutional neural networks (CNNs) based machine learning methods were employed to classify distomolar tooth existence using panoramic radiography (PR). METHODS PRs dataset, composed of 117 subjects with distomolar teeth and 146 subjects without distomolar teeth, was constructed. These images were assessed using AlexNet, DarkNet, DenseNet, EfficientNet, GoogLeNet, ResNet, MobileNet, NasNet-Mobile, VGG, and XceptionNet frameworks for distomolar teeth existence. Considering the moderate number dataset samples, transfer learning was also utilized to improve the performance of these CNN based networks along with 5-fold cross-validation. The final classification was obtained through the fusion of the classifiers results. RESULTS Performance of the experimental studies was assessed utilizing accuracy (Acc), sensitivity (sen), specificity (spe) and precision (pre) metrics. Best accuracy value of 96.2 % was obtained for the fusion of DarkNet, DenseNet, and ResNet, three best individual performing architectures, in distomolar teeth classification problem. CONCLUSION AND PRACTICAL IMPLICATIONS In summary, this study has demonstrated the significant potential of CNNs in accurately detecting distomolar teeth in dental radiographs, a critical task for clinical diagnosis and treatment planning. The fusion of CNN architectures, particularly ResNet, Darknet, and DenseNet, has shown exceptional performance, pointing towards the future of artificial intelligence (AI) driven dental diagnostics. Our findings showed that these systems can help clinicians during radiologic examinations.
Collapse
Affiliation(s)
- Onur Erdem Korkmaz
- Department of Electrical Electronic Engineering, Faculty of Engineering, Ataturk University, Erzurum, Turkey
| | - Hatice Guller
- Department of Oral, Dental and Maxillofacial Radiology, Faculty of Dentistry, Ataturk University, Erzurum, Turkey
| | - Ozkan Miloglu
- Department of Oral, Dental and Maxillofacial Radiology, Faculty of Dentistry, Ataturk University, Erzurum, Turkey.
| | - İbrahim Yucel Ozbek
- Department of Electrical Electronic Engineering (High Performance Comp Applicat & Res Ctr), Ataturk University, Erzurum, Turkey
| | - Emin Argun Oral
- Department of Electrical Electronic Engineering, Faculty of Engineering, Ataturk University, Erzurum, Turkey
| | - Mustafa Taha Guller
- Department of Oral, Dental and Maxillofacial Radiology, Faculty of Dentistry, Giresun University, Giresun, Turkey
| |
Collapse
|
67
|
Liu K, Zhang J. Development of a Cost-Efficient and Glaucoma-Specialized OD/OC Segmentation Model for Varying Clinical Scenarios. SENSORS (BASEL, SWITZERLAND) 2024; 24:7255. [PMID: 39599032 PMCID: PMC11597940 DOI: 10.3390/s24227255] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/01/2024] [Revised: 10/31/2024] [Accepted: 11/11/2024] [Indexed: 11/29/2024]
Abstract
Most existing optic disc (OD) and cup (OC) segmentation models are biased to the dominant size and easy class (normal class), resulting in suboptimal performances on glaucoma-confirmed samples. Thus, these models are not optimal choices for assisting in tracking glaucoma progression and prognosis. Moreover, fully supervised models employing annotated glaucoma samples can achieve superior performances, although restricted by the high cost of collecting and annotating the glaucoma samples. Therefore, in this paper, we are dedicated to developing a glaucoma-specialized model by exploiting low-cost annotated normal fundus images, simultaneously adapting various common scenarios in clinical practice. We employ a contrastive learning and domain adaptation-based model by exploiting shared knowledge from normal samples. To capture glaucoma-related features, we utilize a Gram matrix to encode style information and the domain adaptation strategy to encode domain information, followed by narrowing the style and domain gaps between normal and glaucoma samples by contrastive and adversarial learning, respectively. To validate the efficacy of our proposed model, we conducted experiments utilizing two public datasets to mimic various common scenarios. The results demonstrate the superior performance of our proposed model across multi-scenarios, showcasing its proficiency in both the segmentation- and glaucoma-related metrics. In summary, our study illustrates a concerted effort to target confirmed glaucoma samples, mitigating the inherent bias issue in most existing models. Moreover, we propose an annotation-efficient strategy that exploits low-cost, normal-labeled fundus samples, mitigating the economic- and labor-related burdens by employing a fully supervised strategy. Simultaneously, our approach demonstrates its adaptability across various scenarios, highlighting its potential utility in both assisting in the monitoring of glaucoma progression and assessing glaucoma prognosis.
Collapse
Affiliation(s)
- Kai Liu
- School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China;
- Beijing Advanced Innovation Centre for Biomedical Engineering, Beihang University, Beijing 100083, China
- Department of Computer Science, City University of Hong Kong, Hong Kong 98121, China
| | - Jicong Zhang
- School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China;
- Beijing Advanced Innovation Centre for Biomedical Engineering, Beihang University, Beijing 100083, China
- Hefei Innovation Research Institute, Beihang University, Hefei 230012, China
| |
Collapse
|
68
|
Lv E, Kang X, Wen P, Tian J, Zhang M. A novel benign and malignant classification model for lung nodules based on multi-scale interleaved fusion integrated network. Sci Rep 2024; 14:27506. [PMID: 39528563 PMCID: PMC11555393 DOI: 10.1038/s41598-024-79058-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Accepted: 11/06/2024] [Indexed: 11/16/2024] Open
Abstract
One of the precursors of lung cancer is the presence of lung nodules, and accurate identification of their benign or malignant nature is important for the long-term survival of patients. With the development of artificial intelligence, deep learning has become the main method for lung nodule classification. However, successful deep learning models usually require large number of parameters and carefully annotated data. In the field of medical images, the availability of such data is usually limited, which makes deep networks often perform poorly on new test data. In addition, the model based on the linear stacked single branch structure hinders the extraction of multi-scale features and reduces the classification performance. In this paper, to address this problem, we propose a lightweight interleaved fusion integration network with multi-scale feature learning modules, called MIFNet. The MIFNet consists of a series of MIF blocks that efficiently combine multiple convolutional layers containing 1 × 1 and 3 × 3 convolutional kernels with shortcut links to extract multiscale features at different levels and preserving them throughout the block. The model has only 0.7 M parameters and requires low computational cost and memory space compared to many ImageNet pretrained CNN architectures. The proposed MIFNet conducted exhaustive experiments on the reconstructed LUNA16 dataset, achieving impressive results with 94.82% accuracy, 97.34% F1 value, 96.74% precision, 97.10% sensitivity, and 84.75% specificity. The results show that our proposed deep integrated network achieves higher performance than pre-trained deep networks and state-of-the-art methods. This provides an objective and efficient auxiliary method for accurately classifying the type of lung nodule in medical images.
Collapse
Affiliation(s)
- Enhui Lv
- School of Medical Information & Engineering, Xuzhou Medical University, Xuzhou, Jiangsu, China
| | - Xingxing Kang
- School of Medical Information & Engineering, Xuzhou Medical University, Xuzhou, Jiangsu, China
| | - Pengbo Wen
- School of Medical Information & Engineering, Xuzhou Medical University, Xuzhou, Jiangsu, China
| | - Jiaqi Tian
- School of Medical Information & Engineering, Xuzhou Medical University, Xuzhou, Jiangsu, China.
| | - Mengying Zhang
- School of Medical Information & Engineering, Xuzhou Medical University, Xuzhou, Jiangsu, China.
| |
Collapse
|
69
|
Zhang M, Zhou J, Wang X, Wang X, Ge F. DeepBP: Ensemble deep learning strategy for bioactive peptide prediction. BMC Bioinformatics 2024; 25:352. [PMID: 39528950 PMCID: PMC11556071 DOI: 10.1186/s12859-024-05974-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2024] [Accepted: 11/04/2024] [Indexed: 11/16/2024] Open
Abstract
BACKGROUND Bioactive peptides are important bioactive molecules composed of short-chain amino acids that play various crucial roles in the body, such as regulating physiological processes and promoting immune responses and antibacterial effects. Due to their significance, bioactive peptides have broad application potential in drug development, food science, and biotechnology. Among them, understanding their biological mechanisms will contribute to new ideas for drug discovery and disease treatment. RESULTS This study employs generative adversarial capsule networks (CapsuleGAN), gated recurrent units (GRU), and convolutional neural networks (CNN) as base classifiers to achieve ensemble learning through voting methods, which not only obtains high-precision prediction results on the angiotensin-converting enzyme (ACE) inhibitory peptides dataset and the anticancer peptides (ACP) dataset but also demonstrates effective model performance. For this method, we first utilized the protein language model-evolutionary scale modeling (ESM-2)-to extract relevant features for the ACE inhibitory peptides and ACP datasets. Following feature extraction, we trained three deep learning models-CapsuleGAN, GRU, and CNN-while continuously adjusting the model parameters throughout the training process. Finally, during the voting stage, different weights were assigned to the models based on their prediction accuracy, allowing full utilization of the model's performance. Experimental results show that on the ACE inhibitory peptide dataset, the balanced accuracy is 0.926, the Matthews correlation coefficient (MCC) is 0.831, and the area under the curve is 0.966; on the ACP dataset, the accuracy (ACC) is 0.779, and the MCC is 0.558. The experimental results on both datasets are superior to existing methods, demonstrating the effectiveness of the experimental approach. CONCLUSION In this study, CapsuleGAN, GRU, and CNN were successfully employed as base classifiers to implement ensemble learning, which not only achieved good results in the prediction of two datasets but also surpassed existing methods. The ability to predict peptides with strong ACE inhibitory activity and ACPs more accurately and quickly is significant, and this work provides valuable insights for predicting other functional peptides. The source code and dataset for this experiment are publicly available at https://github.com/Zhou-Jianren/bioactive-peptides .
Collapse
Affiliation(s)
- Ming Zhang
- School of Computer, Jiangsu University of Science and Technology, 666 Changhui Road, Zhenjiang, 212100, China.
| | - Jianren Zhou
- School of Computer, Jiangsu University of Science and Technology, 666 Changhui Road, Zhenjiang, 212100, China
| | - Xiaohua Wang
- School of Computer, Jiangsu University of Science and Technology, 666 Changhui Road, Zhenjiang, 212100, China
| | - Xun Wang
- School of Computer, Jiangsu University of Science and Technology, 666 Changhui Road, Zhenjiang, 212100, China
| | - Fang Ge
- State Key Laboratory of Organic Electronics and Information Displays & Institute of Advanced Materials (IAM), Nanjing University of Posts & Telecommunications, 9 Wenyuan Road, Nanjing, 210023, China.
| |
Collapse
|
70
|
Liu Z, Zhang H, Zhang M, Qu C, Li L, Sun Y, Ma X. Compare three deep learning-based artificial intelligence models for classification of calcified lumbar disc herniation: a multicenter diagnostic study. Front Surg 2024; 11:1458569. [PMID: 39569028 PMCID: PMC11576459 DOI: 10.3389/fsurg.2024.1458569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2024] [Accepted: 10/21/2024] [Indexed: 11/22/2024] Open
Abstract
Objective To develop and validate an artificial intelligence diagnostic model for identifying calcified lumbar disc herniation based on lateral lumbar magnetic resonance imaging(MRI). Methods During the period from January 2019 to March 2024, patients meeting the inclusion criteria were collected. All patients had undergone both lumbar spine MRI and computed tomography(CT) examinations, with regions of interest (ROI) clearly marked on the lumbar sagittal MRI images. The participants were then divided into separate sets for training, testing, and external validation. Ultimately, we developed a deep learning model using the ResNet-34 algorithm model and evaluated its diagnostic efficacy. Results A total of 1,224 eligible patients were included in this study, consisting of 610 males and 614 females, with an average age of 53.34 ± 10.61 years. Notably, the test datasets displayed an impressive classification accuracy rate of 91.67%, whereas the external validation datasets achieved a classification accuracy rate of 88.76%. Among the test datasets, the ResNet34 model outperformed other models, yielding the highest area under the curve (AUC) of 0.96 (95% CI: 0.93, 0.99). Additionally, the ResNet34 model also exhibited superior performance in the external validation datasets, exhibiting an AUC of 0.88 (95% CI: 0.80, 0.93). Conclusion In this study, we established a deep learning model with excellent performance in identifying calcified intervertebral discs, thereby offering a valuable and efficient diagnostic tool for clinical surgeons.
Collapse
Affiliation(s)
- Zhiming Liu
- Department of Spine Surgery, The Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Hao Zhang
- Department of Spine Surgery, The Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Min Zhang
- Department of Neonatology, The Second Affiliated Hospital and Yuying Children's Hospital of Wenzhou Medical University, Wenzhou, Zhejiang, China
| | - Changpeng Qu
- Department of Spine Surgery, The Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Lei Li
- Department of Spine Surgery, The Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Yihao Sun
- Department of Spine Surgery, The Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Xuexiao Ma
- Department of Spine Surgery, The Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| |
Collapse
|
71
|
Hsieh HC, Chen CY, Chou CH, Peng BY, Sun YC, Lin TW, Chien Y, Chiou SH, Hung KF, Lu HHS. Deep learning-based automatic image classification of oral cancer cells acquiring chemoresistance in vitro. PLoS One 2024; 19:e0310304. [PMID: 39485749 PMCID: PMC11530068 DOI: 10.1371/journal.pone.0310304] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2024] [Accepted: 08/29/2024] [Indexed: 11/03/2024] Open
Abstract
Cell shape reflects the spatial configuration resulting from the equilibrium of cellular and environmental signals and is considered a highly relevant indicator of its function and biological properties. For cancer cells, various physiological and environmental challenges, including chemotherapy, cause a cell state transition, which is accompanied by a continuous morphological alteration that is often extremely difficult to recognize even by direct microscopic inspection. To determine whether deep learning-based image analysis enables the detection of cell shape reflecting a crucial cell state alteration, we used the oral cancer cell line resistant to chemotherapy but having cell morphology nearly indiscernible from its non-resistant parental cells. We then implemented the automatic approach via deep learning methods based on EfficienNet-B3 models, along with over- and down-sampling techniques to determine whether image analysis of the Convolutional Neural Network (CNN) can accomplish three-class classification of non-cancer cells vs. cancer cells with and without chemoresistance. We also examine the capability of CNN-based image analysis to approximate the composition of chemoresistant cancer cells within a population. We show that the classification model achieves at least 98.33% accuracy by the CNN model trained with over- and down-sampling techniques. For heterogeneous populations, the best model can approximate the true proportions of non-chemoresistant and chemoresistant cancer cells with Root Mean Square Error (RMSE) reduced to 0.16 by Ensemble Learning (EL). In conclusion, our study demonstrates the potential of CNN models to identify altered cell shapes that are visually challenging to recognize, thus supporting future applications with this automatic approach to image analysis.
Collapse
Affiliation(s)
- Hsing-Chuan Hsieh
- Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu, Taiwan
| | - Cho-Yi Chen
- Institute of Biomedical Informatics, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Chung-Hsien Chou
- Institute of Oral Biology, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Bou-Yue Peng
- Department of Dentistry, Taipei Medical University Hospital, Taipei, Taiwan
- School of Dentistry, College of Oral Medicine, Taipei Medical University, Taipei, Taiwan
| | - Yi-Chen Sun
- College of Medicine, Tzu-Chi University, Hualien, Taiwan
- Department of Ophthalmology, Taipei Tzu Chi Hospital, The Buddhist Tzu Chi Medical Foundation, New Taipei, Taiwan
| | - Tzu-Wei Lin
- Department of Medical Research, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Yueh Chien
- Department of Medical Research, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Shih-Hwa Chiou
- Department of Medical Research, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Kai-Feng Hung
- Department of Medical Research, Taipei Veterans General Hospital, Taipei, Taiwan
- Department of Dentistry, School of Dentistry, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Henry Horng-Shing Lu
- Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu, Taiwan
- Department of Medical Research, Taipei Veterans General Hospital, Taipei, Taiwan
- School of Post-Baccalaureate Medicine, Kaohsiung Medical University, Kaohsiung, Taiwan
- Department of Statistics and Data Science, Cornell University, Ithaca, New York, United States of America
| |
Collapse
|
72
|
Ichiuji Y, Mabu S, Hatta S, Inai K, Higuchi S, Kido S. Domain transformation using semi-supervised CycleGAN for improving performance of classifying thyroid tissue images. Int J Comput Assist Radiol Surg 2024; 19:2153-2163. [PMID: 38238492 DOI: 10.1007/s11548-024-03061-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Accepted: 01/05/2024] [Indexed: 11/07/2024]
Abstract
PURPOSE A large number of research has been conducted on the classification of medical images using deep learning. The thyroid tissue images can be also classified by cancer types. Deep learning requires a large amount of data, but every medical institution cannot collect sufficient number of data for deep learning. In that case, we can consider a case where a classifier trained at a certain medical institution that has a sufficient number of data is reused at other institutions. However, when using data from multiple institutions, it is necessary to unify the feature distribution because the feature of the data differs due to differences in data acquisition conditions. METHODS To unify the feature distribution, the data from Institution T are transformed to have the closer distribution to that from Institution S by applying a domain transformation using semi-supervised CycleGAN. The proposed method enhances CycleGAN considering the feature distribution of classes for making appropriate domain transformation for classification. In addition, to address the problem of imbalanced data with different numbers of data for each cancer type, several methods dealing with imbalanced data are applied to semi-supervised CycleGAN. RESULTS The experimental results showed that the classification performance was enhanced when the dataset from Institution S was used as training data and the testing dataset from Institution T was classified after applying domain transformation. In addition, focal loss contributed to improving the mean F1 score the best as a method that addresses the class imbalance. CONCLUSION The proposed method achieved the domain transformation of thyroid tissue images between two domains, where it retained the important features related to the classes across domains and showed the best F1 score with significant differences compared with other methods. In addition, the proposed method was further enhanced by addressing the class imbalance of the dataset.
Collapse
Affiliation(s)
- Yoshihito Ichiuji
- Graduate School of Sciences and Technology for Innovation, Yamaguchi University, 2-16-1, Tokiwadai, Ube, Yamaguchi, 755-8611, Japan
| | - Shingo Mabu
- Graduate School of Sciences and Technology for Innovation, Yamaguchi University, 2-16-1, Tokiwadai, Ube, Yamaguchi, 755-8611, Japan.
| | - Satomi Hatta
- Division of Molecular Pathology, Department of Pathological Sciences, University of Fukui, 23-3, Matsuoka-shimoaizuki, Eiheiji-cho, Yoshida-gun, Fukui, 910-1193, Japan
- Division of Diagnostic/Surgical Pathology, University of Fukui Hospital, 23-3, Matsuoka-shimoaizuki, Eiheiji-cho, Yoshida-gun, Fukui, 910-1193, Japan
| | - Kunihiro Inai
- Division of Molecular Pathology, Department of Pathological Sciences, University of Fukui, 23-3, Matsuoka-shimoaizuki, Eiheiji-cho, Yoshida-gun, Fukui, 910-1193, Japan
| | - Shohei Higuchi
- Division of Molecular Pathology, Department of Pathological Sciences, University of Fukui, 23-3, Matsuoka-shimoaizuki, Eiheiji-cho, Yoshida-gun, Fukui, 910-1193, Japan
- Department of Pathological Diagnosis, Fukui Prefectural Hospital, 2-8-1, Yotsui, Fukui, Fukui, 910-0846, Japan
| | - Shoji Kido
- Department of Artificial Intelligence Diagnostic Radiology, Graduate School of Medicine, Osaka University, 2-2, Yamadaoka, Suita, Osaka, 565-0871, Japan
| |
Collapse
|
73
|
Salari E, Wang J, Wynne JF, Chang C, Wu Y, Yang X. Artificial intelligence-based motion tracking in cancer radiotherapy: A review. J Appl Clin Med Phys 2024; 25:e14500. [PMID: 39194360 PMCID: PMC11540048 DOI: 10.1002/acm2.14500] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Revised: 07/13/2024] [Accepted: 07/27/2024] [Indexed: 08/29/2024] Open
Abstract
Radiotherapy aims to deliver a prescribed dose to the tumor while sparing neighboring organs at risk (OARs). Increasingly complex treatment techniques such as volumetric modulated arc therapy (VMAT), stereotactic radiosurgery (SRS), stereotactic body radiotherapy (SBRT), and proton therapy have been developed to deliver doses more precisely to the target. While such technologies have improved dose delivery, the implementation of intra-fraction motion management to verify tumor position at the time of treatment has become increasingly relevant. Artificial intelligence (AI) has recently demonstrated great potential for real-time tracking of tumors during treatment. However, AI-based motion management faces several challenges, including bias in training data, poor transparency, difficult data collection, complex workflows and quality assurance, and limited sample sizes. This review presents the AI algorithms used for chest, abdomen, and pelvic tumor motion management/tracking for radiotherapy and provides a literature summary on the topic. We will also discuss the limitations of these AI-based studies and propose potential improvements.
Collapse
Affiliation(s)
- Elahheh Salari
- Department of Radiation OncologyEmory UniversityAtlantaGeorgiaUSA
| | - Jing Wang
- Radiation OncologyIcahn School of Medicine at Mount SinaiNew YorkNew YorkUSA
| | | | - Chih‐Wei Chang
- Department of Radiation OncologyEmory UniversityAtlantaGeorgiaUSA
| | - Yizhou Wu
- School of Electrical and Computer EngineeringGeorgia Institute of TechnologyAtlantaGeorgiaUSA
| | - Xiaofeng Yang
- Department of Radiation OncologyEmory UniversityAtlantaGeorgiaUSA
| |
Collapse
|
74
|
Wang Z, Li X, Zhang H, Duan T, Zhang C, Zhao T. Deep learning Radiomics Based on Two-Dimensional Ultrasound for Predicting the Efficacy of Neoadjuvant Chemotherapy in Breast Cancer. ULTRASONIC IMAGING 2024; 46:357-366. [PMID: 39257175 DOI: 10.1177/01617346241276168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/12/2024]
Abstract
We investigate the predictive value of a comprehensive model based on preoperative ultrasound radiomics, deep learning, and clinical features for pathological complete response (pCR) after neoadjuvant chemotherapy (NAC) for the breast cancer. We enrolled 155 patients with pathologically confirmed breast cancer who underwent NAC. The patients were randomly divided into the training set and the validation set in the ratio of 7:3. The deep learning and radiomics features of pre-treatment ultrasound images were extracted, and the random forest recursive elimination algorithm and the least absolute shrinkage and selection operator were used for feature screening and DL-Score and Rad-Score construction. According to multifactorial logistic regression, independent clinical predictors, DL-Score, and Rad-Score were selected to construct the comprehensive prediction model DLRC. The performance of the model was evaluated in terms of its predictive effect, and clinical practicability. Compared to the clinical, radiomics (Rad-Score), and deep learning (DL-Score) models, the DLRC accurately predicted the pCR status, with an area under the curve (AUC) of 0.937 (95%CI: 0.895-0.970) in the training set and 0.914 (95%CI: 0.838-0.973) in the validation set. Moreover, decision curve analysis confirmed that the DLRC had the highest clinical value among all models. The comprehensive model DLRC based on ultrasound radiomics, deep learning, and clinical features can effectively and accurately predict the pCR status of breast cancer after NAC, which is conducive to assisting clinical personalized diagnosis and treatment plan.
Collapse
Affiliation(s)
- Zhan Wang
- Jintan Peoples Hospital, Jiangsu, Changzhou, China
| | - Xiaoqin Li
- Changzhou Second People's Hospital Affiliated to Nanjing Medical University, Jiangsu, Changzhou, China
| | - Heng Zhang
- Changzhou Second People's Hospital Affiliated to Nanjing Medical University, Jiangsu, Changzhou, China
| | - Tongtong Duan
- Changzhou Second People's Hospital Affiliated to Nanjing Medical University, Jiangsu, Changzhou, China
| | - Chao Zhang
- Changzhou Second People's Hospital Affiliated to Nanjing Medical University, Jiangsu, Changzhou, China
| | - Tong Zhao
- Changzhou Second People's Hospital Affiliated to Nanjing Medical University, Jiangsu, Changzhou, China
| |
Collapse
|
75
|
Huang X, Wang Q, He J, Ban C, Zheng H, Chen H, Zhu X. Fast Multiphoton Microscopic Imaging Joint Image Super-Resolution for Automated Gleason Grading of Prostate Cancers. JOURNAL OF BIOPHOTONICS 2024; 17:e202400233. [PMID: 39262127 DOI: 10.1002/jbio.202400233] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 08/07/2024] [Accepted: 08/08/2024] [Indexed: 09/13/2024]
Abstract
Gleason grading system is dependable for quantifying prostate cancer. This paper introduces a fast multiphoton microscopic imaging method via deep learning for automatic Gleason grading. Due to the contradiction between multiphoton microscopy (MPM) imaging speed and quality, a deep learning architecture (SwinIR) is used for image super-resolution to address this issue. The quality of low-resolution image is improved, which increased the acquisition speed from 7.55 s per frame to 0.24 s per frame. A classification network (Swin Transformer) was introduced for automated Gleason grading. The classification accuracy and Macro-F1 achieved by training on high-resolution images are respectively 90.9% and 90.9%. For training on super-resolution images, the classification accuracy and Macro-F1 are respectively 89.9% and 89.9%. It shows that super-resolution image can provide a comparable performance to high-resolution image. Our results suggested that MPM joint image super-resolution and automatic classification methods hold the potential to be a real-time clinical diagnostic tool for prostate cancer diagnosis.
Collapse
Affiliation(s)
- Xinpeng Huang
- Institute of Laser and Optoelectronics Technology, Fujian Provincial Key Laboratory for Photonics Technology, Key Laboratory of Optoelectronic Science and Technology for Medicine of Ministry of Education, Fujian Normal University, Fuzhou, China
| | - Qianqiong Wang
- Institute of Laser and Optoelectronics Technology, Fujian Provincial Key Laboratory for Photonics Technology, Key Laboratory of Optoelectronic Science and Technology for Medicine of Ministry of Education, Fujian Normal University, Fuzhou, China
| | - Jia He
- Institute of Laser and Optoelectronics Technology, Fujian Provincial Key Laboratory for Photonics Technology, Key Laboratory of Optoelectronic Science and Technology for Medicine of Ministry of Education, Fujian Normal University, Fuzhou, China
| | - Chaoran Ban
- Department of Pathology, The First Affiliated Hospital of Fujian Medical University, Fuzhou, China
| | - Hua Zheng
- Institute of Laser and Optoelectronics Technology, Fujian Provincial Key Laboratory for Photonics Technology, Key Laboratory of Optoelectronic Science and Technology for Medicine of Ministry of Education, Fujian Normal University, Fuzhou, China
| | - Hong Chen
- Department of Pathology, The First Affiliated Hospital of Fujian Medical University, Fuzhou, China
| | - Xiaoqin Zhu
- Institute of Laser and Optoelectronics Technology, Fujian Provincial Key Laboratory for Photonics Technology, Key Laboratory of Optoelectronic Science and Technology for Medicine of Ministry of Education, Fujian Normal University, Fuzhou, China
| |
Collapse
|
76
|
Wang CT, Chen TM, Lee NT, Fang SH. AI Detection of Glottic Neoplasm Using Voice Signals, Demographics, and Structured Medical Records. Laryngoscope 2024; 134:4585-4592. [PMID: 38864282 DOI: 10.1002/lary.31563] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 04/16/2024] [Accepted: 05/21/2024] [Indexed: 06/13/2024]
Abstract
OBJECTIVE This study investigated whether artificial intelligence (AI) models combining voice signals, demographics, and structured medical records can detect glottic neoplasm from benign voice disorders. METHODS We used a primary dataset containing 2-3 s of vowel "ah", demographics, and 26 items of structured medical records (e.g., symptoms, comorbidity, smoking and alcohol consumption, vocal demand) from 60 patients with pathology-proved glottic neoplasm (i.e., squamous cell carcinoma, carcinoma in situ, and dysplasia) and 1940 patients with benign voice disorders. The validation dataset comprised data from 23 patients with glottic neoplasm and 1331 patients with benign disorders. The AI model combined convolutional neural networks, gated recurrent units, and attention layers. We used 10-fold cross-validation (training-validation-testing: 8-1-1) and preserved the percentage between neoplasm and benign disorders in each fold. RESULTS Results from the AI model using voice signals reached an area under the ROC curve (AUC) value of 0.631, and additional demographics increased this to 0.807. The highest AUC of 0.878 was achieved when combining voice, demographics, and medical records (sensitivity: 0.783, specificity: 0.816, accuracy: 0.815). External validation yielded an AUC value of 0.785 (voice plus demographics; sensitivity: 0.739, specificity: 0.745, accuracy: 0.745). Subanalysis showed that AI had higher sensitivity but lower specificity than human assessment (p < 0.01). The accuracy of AI detection with additional medical records was comparable with human assessment (82% vs. 83%, p = 0.78). CONCLUSIONS Voice signal alone was insufficient for AI differentiation between glottic neoplasm and benign voice disorders, but additional demographics and medical records notably improved AI performance and approximated the prediction accuracy of humans. LEVEL OF EVIDENCE NA Laryngoscope, 134:4585-4592, 2024.
Collapse
Affiliation(s)
- Chi-Te Wang
- Department of Otolaryngology Head and Neck Surgery, Far Eastern Memorial Hospital, Taipei, Taiwan
- Center of Artificial Intelligence, Far Eastern Memorial Hospital, Taipei, Taiwan
- Department of Electrical Engineering, Yuan Ze University, Taoyuan, Taiwan
| | - Tsai-Min Chen
- Graduate Program of Data Science, National Taiwan University and Academia Sinica, Taipei, Taiwan
- Research Center for Information Technology Innovation, Academia Sinica, Taipei, Taiwan
| | - Nien-Ting Lee
- Center of Artificial Intelligence, Far Eastern Memorial Hospital, Taipei, Taiwan
| | - Shih-Hau Fang
- Department of Electrical Engineering, Yuan Ze University, Taoyuan, Taiwan
- Department of Electrical Engineering, National Taiwan Normal University, Taipei, Taiwan
| |
Collapse
|
77
|
Karthiga R, Narasimhan K, V T, M H, Amirtharajan R. Review of AI & XAI-based breast cancer diagnosis methods using various imaging modalities. MULTIMEDIA TOOLS AND APPLICATIONS 2024. [DOI: 10.1007/s11042-024-20271-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 08/27/2024] [Accepted: 09/11/2024] [Indexed: 01/02/2025]
|
78
|
Maiora J, Rezola-Pardo C, García G, Sanz B, Graña M. Older Adult Fall Risk Prediction with Deep Learning and Timed Up and Go (TUG) Test Data. Bioengineering (Basel) 2024; 11:1000. [PMID: 39451376 PMCID: PMC11504430 DOI: 10.3390/bioengineering11101000] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2024] [Revised: 09/23/2024] [Accepted: 09/30/2024] [Indexed: 10/26/2024] Open
Abstract
Falls are a major health hazard for older adults; therefore, in the context of an aging population, predicting the risk of a patient suffering falls in the near future is of great impact for health care systems. Currently, the standard prospective fall risk assessment instrument relies on a set of clinical and functional mobility assessment tools, one of them being the Timed Up and Go (TUG) test. Recently, wearable inertial measurement units (IMUs) have been proposed to capture motion data that would allow for the building of estimates of fall risk. The hypothesis of this study is that the data gathered from IMU readings while the patient is performing the TUG test can be used to build a predictive model that would provide an estimate of the probability of suffering a fall in the near future, i.e., assessing prospective fall risk. This study applies deep learning convolutional neural networks (CNN) and recurrent neural networks (RNN) to build such predictive models based on features extracted from IMU data acquired during TUG test realizations. Data were obtained from a cohort of 106 older adults wearing wireless IMU sensors with sampling frequencies of 100 Hz while performing the TUG test. The dependent variable is a binary variable that is true if the patient suffered a fall in the six-month follow-up period. This variable was used as the output variable for the supervised training and validations of the deep learning architectures and competing machine learning approaches. A hold-out validation process using 75 subjects for training and 31 subjects for testing was repeated one hundred times to obtain robust estimations of model performances At each repetition, 5-fold cross-validation was carried out to select the best model over the training subset. Best results were achieved by a bidirectional long short-term memory (BLSTM), obtaining an accuracy of 0.83 and AUC of 0.73 with good sensitivity and specificity values.
Collapse
Affiliation(s)
- Josu Maiora
- Electronic Technology Department, Faculty of Engineering of Gipuzkoa, University of the Basque Country, 20018 San Sebastian, Spain
- Computational Intelligence Group, Department of CCIA, University of the Basque Country, 20018 San Sebastian, Spain;
| | - Chloe Rezola-Pardo
- Department of Physiology, University of the Basque Country, 48940 Leioa, Spain; (C.R.-P.); (B.S.)
| | - Guillermo García
- Systems and Automation Department, Faculty of Engineering of Gipuzkoa, University of the Basque Country, 20018 San Sebastian, Spain;
| | - Begoña Sanz
- Department of Physiology, University of the Basque Country, 48940 Leioa, Spain; (C.R.-P.); (B.S.)
- Biobizkaia Health Research Institute, 48903 Barakaldo, Spain
| | - Manuel Graña
- Computational Intelligence Group, Department of CCIA, University of the Basque Country, 20018 San Sebastian, Spain;
| |
Collapse
|
79
|
Zhicheng H, Yipeng W, Xiao L. Deep Learning-Based Detection of Impacted Teeth on Panoramic Radiographs. Biomed Eng Comput Biol 2024; 15:11795972241288319. [PMID: 39372969 PMCID: PMC11456186 DOI: 10.1177/11795972241288319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2024] [Accepted: 09/16/2024] [Indexed: 10/08/2024] Open
Abstract
Objective The aim is to detect impacted teeth in panoramic radiology by refining the pretrained MedSAM model. Study design Impacted teeth are dental issues that can cause complications and are diagnosed via radiographs. We modified SAM model for individual tooth segmentation using 1016 X-ray images. The dataset was split into training, validation, and testing sets, with a ratio of 16:3:1. We enhanced the SAM model to automatically detect impacted teeth by focusing on the tooth's center for more accurate results. Results With 200 epochs, batch size equals to 1, and a learning rate of 0.001, random images trained the model. Results on the test set showcased performance up to an accuracy of 86.73%, F1-score of 0.5350, and IoU of 0.3652 on SAM-related models. Conclusion This study fine-tunes MedSAM for impacted tooth segmentation in X-ray images, aiding dental diagnoses. Further improvements on model accuracy and selection are essential for enhancing dental practitioners' diagnostic capabilities.
Collapse
Affiliation(s)
- He Zhicheng
- School of Computer and Information Technology, Beijing Jiaotong University, Beijing, PR China
| | - Wang Yipeng
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, PR China
| | - Li Xiao
- School of Computer and Information Technology, Beijing Jiaotong University, Beijing, PR China
| |
Collapse
|
80
|
Rao D, Singh R, Koteshwara P, Vijayananda J. Exploring the Impact of Model Complexity on Laryngeal Cancer Detection. Indian J Otolaryngol Head Neck Surg 2024; 76:4036-4042. [PMID: 39376269 PMCID: PMC11455748 DOI: 10.1007/s12070-024-04776-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Accepted: 05/26/2024] [Indexed: 10/09/2024] Open
Abstract
Background: Laryngeal cancer accounts for a third of all head and neck malignancies, necessitating timely detection for effective treatment and enhanced patient outcomes. Machine learning shows promise in medical diagnostics, but the impact of model complexity on diagnostic efficacy in laryngeal cancer detection can be ambiguous. Methods: In this study, we examine the relationship between model sophistication and diagnostic efficacy by evaluating three approaches: Logistic Regression, a small neural network with 4 layers of neurons and a more complex convolutional neural network with 50 layers and examine their efficacy on laryngeal cancer detection on computed tomography images. Results: Logistic regression achieved 82.5% accuracy. The 4-Layer NN reached 87.2% accuracy, while ResNet-50, a deep learning architecture, achieved the highest accuracy at 92.6%. Its deep learning capabilities excelled in discerning fine-grained CT image features. Conclusion: Our study highlights the choices involved in selecting a laryngeal cancer detection model. Logistic regression is interpretable but may struggle with complex patterns. The 4-Layer NN balances complexity and accuracy. ResNet-50 excels in image classification but demands resources. This research advances understanding affect machine learning model complexity could have on learning features of laryngeal tumor features in contrast CT images for purposes of disease prediction.
Collapse
Affiliation(s)
- Divya Rao
- Department of Information and Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, 576104 India
| | - Rohit Singh
- Department of Otorhinolaryngology, Kasturba Medical College, Manipal Academy of Higher Education, Manipal, 576104 India
| | - Prakashini Koteshwara
- Department of Radiodiagnosis and Imaging, Kasturba Medical College, Manipal Academy of Higher Education, Manipal, 576104 India
| | - J. Vijayananda
- Data Science and Artificial Intelligence, Philips, Bangalore, 560045 India
| |
Collapse
|
81
|
Lou Z, Min X, Li G, Avery J, Stewart R. Advancing Sensing Resolution of Impedance Hand Gesture Recognition Devices. IEEE J Biomed Health Inform 2024; 28:5855-5864. [PMID: 38905093 DOI: 10.1109/jbhi.2024.3417616] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/23/2024]
Abstract
Gestures are composed of motion information (e.g. movements of fingers) and force information (e.g. the force exerted on fingers when interacting with other objects). Current hand gesture recognition solutions such as cameras and strain sensors primarily focus on correlating hand gestures with motion information and force information is seldom addressed. Here we propose a bio-impedance wearable that can recognize hand gestures utilizing both motion information and force information. Compared with previous impedance-based gesture recognition devices that can only recognize a few multi-degrees-of-freedom gestures, the proposed device can recognize 6 single-degree-of-freedom gestures and 20 multiple-degrees-of-freedom gestures, including 8 gestures in 2 force levels. The device uses textile electrodes, is benchmarked over a selected frequency spectrum, and uses a new drive pattern. Experimental results show that 179 kHz achieves the highest signal-to-noise ratio (SNR) and reveals the most distinct features. By analyzing the 49,920 samples from 6 participants, the device is demonstrated to have an average recognition accuracy of 98.96%. As a comparison, the medical electrodes achieved an accuracy of 98.05%.
Collapse
|
82
|
Kinger S. Deep Learning for Automatic Knee Osteoarthritis Severity Grading and Classification. Indian J Orthop 2024; 58:1458-1473. [PMID: 39324090 PMCID: PMC11420401 DOI: 10.1007/s43465-024-01259-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Accepted: 08/22/2024] [Indexed: 09/27/2024]
Abstract
Introduction Knee osteoarthritis (OA) is a prevalent condition that significantly impacts the quality of life, often leading to the need for knee replacement surgery. Accurate and timely identification of knee degeneration is crucial for effective treatment and management. Traditional methods of diagnosing OA rely heavily on radiological assessments, which can be time-consuming and subjective. This study aims to address these challenges by developing a deep learning-based method to predict the likelihood of knee replacement and the Kellgren-Lawrence (KL) grade of knee OA from X-ray images. Methodology We employed the Osteoarthritis Initiative (OAI) dataset and utilized a transfer learning approach with the Inception V3 architecture to enhance the accuracy of OA detection. Our approach involved training 14 different models-Xception, VGG16, VGG19, ResNet50, ResNet101, ResNet152, ResNet50V2, ResNet101V2, ResNet152V2, Inception V3, Inception, ResNetV2, DenseNet121, DenseNet169, DenseNet201-and comparing their performance. Results The study incorporated pixel ratio computation and picture pre-processing, alongside a decision tree model for prediction. Our experiments revealed that the Inception V3 model achieved the highest training accuracy of 91% and testing accuracy of 67%, with notable performance in both training and validation phases. This model effectively identified the presence and severity of OA, correlating with the Kellgren-Lawrence scale and facilitating the assessment of knee replacement needs. Conclusion By integrating advanced deep learning techniques with radiological diagnostics, our methodology supports radiologists in making more accurate and prompt decisions regarding knee degeneration. The Inception V3 model stands out as the optimal choice for knee X-ray analysis, contributing to more efficient and timely healthcare delivery for patients with knee osteoarthritis.
Collapse
Affiliation(s)
- Shakti Kinger
- Dr. Vishwanath Karad MIT World Peace University, Pune, Maharashtra India
| |
Collapse
|
83
|
Qiao B, Wang S, Hou M, Chen H, Zhou Z, Xie X, Pang S, Yang C, Yang F, Zou Q, Sun S. Identifying nucleotide-binding leucine-rich repeat receptor and pathogen effector pairing using transfer-learning and bilinear attention network. Bioinformatics 2024; 40:btae581. [PMID: 39331576 PMCID: PMC11969219 DOI: 10.1093/bioinformatics/btae581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2024] [Revised: 08/24/2024] [Accepted: 09/25/2024] [Indexed: 09/29/2024] Open
Abstract
MOTIVATION Nucleotide-binding leucine-rich repeat (NLR) family is a class of immune receptors capable of detecting and defending against pathogen invasion. They have been widely used in crop breeding. Notably, the correspondence between NLRs and effectors (CNE) determines the applicability and effectiveness of NLRs. Unfortunately, CNE data is very scarce. In fact, we've found a substantial 91 291 NLRs confirmed via wet experiments and bioinformatics methods but only 387 CNEs are recognized, which greatly restricts the potential application of NLRs. RESULTS We propose a deep learning algorithm called ProNEP to identify NLR-effector pairs in a high-throughput manner. Specifically, we conceptualized the CNE prediction task as a protein-protein interaction (PPI) prediction task. Then, ProNEP predicts the interaction between NLRs and effectors by combining the transfer learning with a bilinear attention network. ProNEP achieves superior performance against state-of-the-art models designed for PPI predictions. Based on ProNEP, we conduct extensive identification of potential CNEs for 91 291 NLRs. With the rapid accumulation of genomic data, we expect that this tool will be widely used to predict CNEs in new species, advancing biology, immunology, and breeding. AVAILABILITY AND IMPLEMENTATION The ProNEP is available at http://nerrd.cn/#/prediction. The project code is available at https://github.com/QiaoYJYJ/ProNEP.
Collapse
Affiliation(s)
- Baixue Qiao
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
- State Key Laboratory of Tree Genetics and Breeding, Northeast Forestry University, Harbin 150001, China
| | - Shuda Wang
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
- State Key Laboratory of Tree Genetics and Breeding, Northeast Forestry University, Harbin 150001, China
| | - Mingjun Hou
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
| | - Haodi Chen
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
| | - Zhengwenyang Zhou
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
| | - Xueying Xie
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
| | - Shaozi Pang
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
| | - Chunxue Yang
- College of Landscape Architecture, Northeast Forestry University, Harbin 150001, China
| | - Fenglong Yang
- Department of Bioinformatics, Fujian Key Laboratory of Medical Bioinformatics, School of Medical Technology and Engineering, Fujian Medical University, Fuzhou 350122, China
- Key Laboratory of Ministry of Education for Gastrointestinal Cancer, School of Basic Medical Sciences, Fujian Medical University, Fuzhou 350122, China
| | - Quan Zou
- Institute of Fundamental and Frontier Sciences, University of Electronic Science and Technology of China, Chengdu 611731, China
| | - Shanwen Sun
- Key Laboratory of Saline-Alkali Vegetation Ecology Restoration, Ministry of Education (Northeast Forestry University), Harbin 150001, China
- State Key Laboratory of Tree Genetics and Breeding, Northeast Forestry University, Harbin 150001, China
| |
Collapse
|
84
|
Rai S, Bhatt JS, Patra SK. An AI-Based Low-Risk Lung Health Image Visualization Framework Using LR-ULDCT. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2047-2062. [PMID: 38491236 PMCID: PMC11522248 DOI: 10.1007/s10278-024-01062-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/23/2023] [Revised: 01/18/2024] [Accepted: 02/12/2024] [Indexed: 03/18/2024]
Abstract
In this article, we propose an AI-based low-risk visualization framework for lung health monitoring using low-resolution ultra-low-dose CT (LR-ULDCT). We present a novel deep cascade processing workflow to achieve diagnostic visualization on LR-ULDCT (<0.3 mSv) at par high-resolution CT (HRCT) of 100 mSV radiation technology. To this end, we build a low-risk and affordable deep cascade network comprising three sequential deep processes: restoration, super-resolution (SR), and segmentation. Given degraded LR-ULDCT, the first novel network unsupervisedly learns restoration function from augmenting patch-based dictionaries and residuals. The restored version is then super-resolved (SR) for target (sensor) resolution. Here, we combine perceptual and adversarial losses in novel GAN to establish the closeness between probability distributions of generated SR-ULDCT and restored LR-ULDCT. Thus SR-ULDCT is presented to the segmentation network that first separates the chest portion from SR-ULDCT followed by lobe-wise colorization. Finally, we extract five lobes to account for the presence of ground glass opacity (GGO) in the lung. Hence, our AI-based system provides low-risk visualization of input degraded LR-ULDCT to various stages, i.e., restored LR-ULDCT, restored SR-ULDCT, and segmented SR-ULDCT, and achieves diagnostic power of HRCT. We perform case studies by experimenting on real datasets of COVID-19, pneumonia, and pulmonary edema/congestion while comparing our results with state-of-the-art. Ablation experiments are conducted for better visualizing different operating pipelines. Finally, we present a verification report by fourteen (14) experienced radiologists and pulmonologists.
Collapse
Affiliation(s)
- Swati Rai
- Indian Institute of Information Technology Vadodara, Vadodara, India.
| | - Jignesh S Bhatt
- Indian Institute of Information Technology Vadodara, Vadodara, India
| | | |
Collapse
|
85
|
Esmaeil-Zadeh M, Fattahi M, Soltani-Gol M, Rostami R, Soltanian-Zadeh H. Effective Connectivity Estimation by a Hybrid Neural Network, Empirical Wavelet Transform, and Bayesian Optimization. IEEE J Biomed Health Inform 2024; 28:5696-5707. [PMID: 37883255 DOI: 10.1109/jbhi.2023.3327734] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2023]
Abstract
Accurately measuring nonlinear effective connectivity is a crucial step in investigating brain functions. Brain signals like EEG is nonstationary. Many effective connectivity methods have been proposed but they have drawbacks in their models such as a weakness in proposing a way for hyperparameter and time lag selection as well as dealing with non-stationarity of the time series. This paper proposes an effective connectivity model based on a hybrid neural network model which uses Empirical Wavelet Transform (EWT) and a long short-term memory network (LSTM). The best hyperparameters and time lag are selected using Bayesian Optimization (BO). Due to the importance of generalizability in neural networks and calculating GC, an algorithm was proposed to choose the best generalizable weights. The model was evaluated using simulated and real EEG data consisting of attention deficit hyperactivity disorder (ADHD) and healthy subjects. The proposed model's performance on simulated data was evaluated by comparing it with other neural networks, including LSTM, CNN-LSTM, GRU, RNN, and MLP, using a Blocked cross-validation approach. GC of the simulated data was compared with GRU, linear Granger causality (LGC), Kernel Granger Causality (KGC), Partial Directed Coherence (PDC), and Directed Transfer Function (DTF). Our results demonstrated that the proposed model was superior to the mentioned models. Another advantage of our model is robustness against noise. The results showed that the proposed model can identify the connections in noisy conditions. The comparison of the effective connectivity of ADHD and the healthy group showed that the results are in accordance with previous studies.
Collapse
|
86
|
Kaderuppan SS, Sharma A, Saifuddin MR, Wong WLE, Woo WL. Θ-Net: A Deep Neural Network Architecture for the Resolution Enhancement of Phase-Modulated Optical Micrographs In Silico. SENSORS (BASEL, SWITZERLAND) 2024; 24:6248. [PMID: 39409287 PMCID: PMC11478931 DOI: 10.3390/s24196248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/30/2024] [Revised: 09/23/2024] [Accepted: 09/23/2024] [Indexed: 10/20/2024]
Abstract
Optical microscopy is widely regarded to be an indispensable tool in healthcare and manufacturing quality control processes, although its inability to resolve structures separated by a lateral distance under ~200 nm has culminated in the emergence of a new field named fluorescence nanoscopy, while this too is prone to several caveats (namely phototoxicity, interference caused by exogenous probes and cost). In this regard, we present a triplet string of concatenated O-Net ('bead') architectures (termed 'Θ-Net' in the present study) as a cost-efficient and non-invasive approach to enhancing the resolution of non-fluorescent phase-modulated optical microscopical images in silico. The quality of the afore-mentioned enhanced resolution (ER) images was compared with that obtained via other popular frameworks (such as ANNA-PALM, BSRGAN and 3D RCAN), with the Θ-Net-generated ER images depicting an increased level of detail (unlike previous DNNs). In addition, the use of cross-domain (transfer) learning to enhance the capabilities of models trained on differential interference contrast (DIC) datasets [where phasic variations are not as prominently manifested as amplitude/intensity differences in the individual pixels unlike phase-contrast microscopy (PCM)] has resulted in the Θ-Net-generated images closely approximating that of the expected (ground truth) images for both the DIC and PCM datasets. This thus demonstrates the viability of our current Θ-Net architecture in attaining highly resolved images under poor signal-to-noise ratios while eliminating the need for a priori PSF and OTF information, thereby potentially impacting several engineering fronts (particularly biomedical imaging and sensing, precision engineering and optical metrology).
Collapse
Affiliation(s)
- Shiraz S. Kaderuppan
- Faculty of Science, Agriculture & Engineering (SAgE), Newcastle University, Newcastle upon Tyne NE1 7RU, UK; (A.S.); (M.R.S.)
| | - Anurag Sharma
- Faculty of Science, Agriculture & Engineering (SAgE), Newcastle University, Newcastle upon Tyne NE1 7RU, UK; (A.S.); (M.R.S.)
| | - Muhammad Ramadan Saifuddin
- Faculty of Science, Agriculture & Engineering (SAgE), Newcastle University, Newcastle upon Tyne NE1 7RU, UK; (A.S.); (M.R.S.)
| | - Wai Leong Eugene Wong
- Engineering Cluster, Singapore Institute of Technology, 10 Dover Drive, Singapore 138683, Singapore;
| | - Wai Lok Woo
- Computer and Information Sciences, Sutherland Building, Northumbria University, Northumberland Road, Newcastle upon Tyne NE1 8ST, UK;
| |
Collapse
|
87
|
Liu N, Yuan Y, Wu G, Zhang S, Leng J, Wan L. Multi-label remote sensing classification with self-supervised gated multi-modal transformers. Front Comput Neurosci 2024; 18:1404623. [PMID: 39380741 PMCID: PMC11458396 DOI: 10.3389/fncom.2024.1404623] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Accepted: 09/03/2024] [Indexed: 10/10/2024] Open
Abstract
Introduction With the great success of Transformers in the field of machine learning, it is also gradually attracting widespread interest in the field of remote sensing (RS). However, the research in the field of remote sensing has been hampered by the lack of large labeled data sets and the inconsistency of data modes caused by the diversity of RS platforms. With the rise of self-supervised learning (SSL) algorithms in recent years, RS researchers began to pay attention to the application of "pre-training and fine-tuning" paradigm in RS. However, there are few researches on multi-modal data fusion in remote sensing field. Most of them choose to use only one of the modal data or simply splice multiple modal data roughly. Method In order to study a more efficient multi-modal data fusion scheme, we propose a multi-modal fusion mechanism based on gated unit control (MGSViT). In this paper, we pretrain the ViT model based on BigEarthNet dataset by combining two commonly used SSL algorithms, and propose an intra-modal and inter-modal gated fusion unit for feature learning by combining multispectral (MS) and synthetic aperture radar (SAR). Our method can effectively combine different modal data to extract key feature information. Results and discussion After fine-tuning and comparison experiments, we outperform the most advanced algorithms in all downstream classification tasks. The validity of our proposed method is verified.
Collapse
Affiliation(s)
- Na Liu
- University of Shanghai for Science and Technology, Institute of Machine Intelligence, Shanghai, China
| | - Ye Yuan
- University of Shanghai for Science and Technology, Institute of Machine Intelligence, Shanghai, China
| | - Guodong Wu
- Origin Dynamics Intelligent Robot Co., Ltd., Zhengzhou, China
| | - Sai Zhang
- Origin Dynamics Intelligent Robot Co., Ltd., Zhengzhou, China
| | - Jie Leng
- Origin Dynamics Intelligent Robot Co., Ltd., Zhengzhou, China
| | - Lihong Wan
- Origin Dynamics Intelligent Robot Co., Ltd., Zhengzhou, China
| |
Collapse
|
88
|
Xiao Z, Zhu M, Chen J, You Z. Integrated Transfer Learning and Multitask Learning Strategies to Construct Graph Neural Network Models for Predicting Bioaccumulation Parameters of Chemicals. ENVIRONMENTAL SCIENCE & TECHNOLOGY 2024; 58:15650-15660. [PMID: 39051472 DOI: 10.1021/acs.est.4c02421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/27/2024]
Abstract
Accurate prediction of parameters related to the environmental exposure of chemicals is crucial for the sound management of chemicals. However, the lack of large data sets for training models may result in poor prediction accuracy and robustness. Herein, integrated transfer learning (TL) and multitask learning (MTL) was proposed for constructing a graph neural network (GNN) model (abbreviated as TL-MTL-GNN model) using n-octanol/water partition coefficients as a source domain. The TL-MTL-GNN model was trained to predict three bioaccumulation parameters based on enlarged data sets that cover 2496 compounds with at least one bioaccumulation parameter. Results show that the TL-MTL-GNN model outperformed single-task GNN models with and without the TL, as well as conventional machine learning models trained with molecular descriptors or fingerprints. Applicability domains were characterized by a state-of-the-art structure-activity landscape-based (abbreviated as ADSAL) methodology. The TL-MTL-GNN model coupled with the optimal ADSAL was employed to predict bioaccumulation parameters for around 60,000 chemicals, with more than 13,000 compounds identified as bioaccumulative chemicals. The high predictive accuracy and robustness of the TL-MTL-GNN model demonstrate the feasibility of integrating the TL and MTL strategy in modeling small-sized data sets. The strategy holds significant potential for addressing small data challenges in modeling environmental chemicals.
Collapse
Affiliation(s)
- Zijun Xiao
- Key Laboratory of Industrial Ecology and Environmental Engineering (Ministry of Education), Dalian Key Laboratory on Chemicals Risk Control and Pollution Prevention Technology, School of Environmental Science and Technology, Dalian University of Technology, Dalian 116024, China
| | - Minghua Zhu
- Key Laboratory of Industrial Ecology and Environmental Engineering (Ministry of Education), Dalian Key Laboratory on Chemicals Risk Control and Pollution Prevention Technology, School of Environmental Science and Technology, Dalian University of Technology, Dalian 116024, China
- Key Laboratory of Integrated Regulation and Resources Development of Shallow Lakes of Ministry of Education, College of Environment, Hohai University, Nanjing 210098, China
| | - Jingwen Chen
- Key Laboratory of Industrial Ecology and Environmental Engineering (Ministry of Education), Dalian Key Laboratory on Chemicals Risk Control and Pollution Prevention Technology, School of Environmental Science and Technology, Dalian University of Technology, Dalian 116024, China
| | - Zecang You
- Key Laboratory of Industrial Ecology and Environmental Engineering (Ministry of Education), Dalian Key Laboratory on Chemicals Risk Control and Pollution Prevention Technology, School of Environmental Science and Technology, Dalian University of Technology, Dalian 116024, China
| |
Collapse
|
89
|
Nieman K, García-García HM, Hideo-Kajita A, Collet C, Dey D, Pugliese F, Weissman G, Tijssen JGP, Leipsic J, Opolski MP, Ferencik M, Lu MT, Williams MC, Bruining N, Blanco PJ, Maurovich-Horvat P, Achenbach S. Standards for quantitative assessments by coronary computed tomography angiography (CCTA): An expert consensus document of the society of cardiovascular computed tomography (SCCT). J Cardiovasc Comput Tomogr 2024; 18:429-443. [PMID: 38849237 DOI: 10.1016/j.jcct.2024.05.232] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/31/2023] [Revised: 05/18/2024] [Accepted: 05/23/2024] [Indexed: 06/09/2024]
Abstract
In current clinical practice, qualitative or semi-quantitative measures are primarily used to report coronary artery disease on cardiac CT. With advancements in cardiac CT technology and automated post-processing tools, quantitative measures of coronary disease severity have become more broadly available. Quantitative coronary CT angiography has great potential value for clinical management of patients, but also for research. This document aims to provide definitions and standards for the performance and reporting of quantitative measures of coronary artery disease by cardiac CT.
Collapse
Affiliation(s)
- Koen Nieman
- Stanford University School of Medicine and Cardiovascular Institute, Stanford, CA, United States.
| | - Hector M García-García
- Section of Interventional Cardiology, MedStar Washington Hospital Center, Washington, DC, United States.
| | | | - Carlos Collet
- Onze Lieve Vrouwziekenhuis, Cardiovascular Center Aalst, Aalst, Belgium
| | - Damini Dey
- Biomedical Imaging Research Institute, Cedars-Sinai Medical Center, Los Angeles, CA, United States
| | - Francesca Pugliese
- NIHR Cardiovascular Biomedical Research Unit at Barts, William Harvey Research Institute, Barts and The London School of Medicine and Dentistry, Queen Mary University of London & Department of Cardiology, Barts Health NHS Trust, London, UK
| | - Gaby Weissman
- Section of Interventional Cardiology, MedStar Washington Hospital Center, Washington, DC, United States
| | - Jan G P Tijssen
- Department of Cardiology, Academic Medical Center, Room G4-230, Meibergdreef 9, 1105 AZ, Amsterdam, the Netherlands
| | - Jonathon Leipsic
- Department of Radiology and Medicine (Cardiology), University of British Columbia, Vancouver, BC, Canada
| | - Maksymilian P Opolski
- Department of Interventional Cardiology and Angiology, National Institute of Cardiology, Warsaw, Poland
| | - Maros Ferencik
- Knight Cardiovascular Institute, Oregon Health & Science University, Portland, OR, United States
| | - Michael T Lu
- Cardiovascular Imaging Research Center, Massachusetts General Hospital & Harvard Medical School, Boston, MA, United States
| | - Michelle C Williams
- BHF Centre for Cardiovascular Science, University of Edinburgh, Edinburgh, UK
| | - Nico Bruining
- Department of Cardiology, Erasmus University Medical Center, Rotterdam, the Netherlands
| | | | - Pal Maurovich-Horvat
- MTA-SE Cardiovascular Imaging Research Group, Medical Imaging Centre, Semmelweis University, Budapest, Hungary
| | - Stephan Achenbach
- Department of Cardiology, Friedrich-Alexander-University Erlangen-Nürnberg, Erlangen, Germany
| |
Collapse
|
90
|
Ketsekioulafis I, Filandrianos G, Katsos K, Thomas K, Spiliopoulou C, Stamou G, Sakelliadis EI. Artificial Intelligence in Forensic Sciences: A Systematic Review of Past and Current Applications and Future Perspectives. Cureus 2024; 16:e70363. [PMID: 39469392 PMCID: PMC11513614 DOI: 10.7759/cureus.70363] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/27/2024] [Indexed: 10/30/2024] Open
Abstract
The aim of this study is to review the available knowledge concerning the use of artificial Intelligence (AI) in general in different areas of Forensic Sciences from human identification to postmortem interval estimation and the estimation of different causes of death. This paper aims to emphasize the different uses of AI, especially in Forensic Medicine, and elucidate its technical part. This will be achieved through an explanation of different technologies that have been so far employed and through new ideas that may contribute as a first step to the adoption of new practices and to the development of new technologies. A systematic literature search was performed in accordance with the Preferred Reported Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines in the PubMed Database and Cochrane Central Library. Neither time nor regional constrictions were adopted, and all the included papers were written in English. Terms used were MACHINE AND LEARNING AND FORENSIC AND PATHOLOGY and ARTIFICIAL AND INTELIGENCE AND FORENSIC AND PATHOLOGY. Quality control was performed using the Joanna Briggs Institute critical appraisal tools. A search of 224 articles was performed. Seven more articles were extracted from the references of the initial selection. After excluding all non-relevant articles, the remaining 45 articles were thoroughly reviewed through the whole text. A final number of 33 papers were identified as relevant to the subject, in accordance with the criteria previously established. It must be clear that AI is not meant to replace forensic experts but to assist them in their everyday work life.
Collapse
Affiliation(s)
- Ioannis Ketsekioulafis
- Department of Forensic Medicine and Toxicology, National and Kapodistrian University of Athens School of Medicine, Athens, GRC
| | - Giorgos Filandrianos
- Artificial Intelligence and Learning Systems Laboratory, School of Electrical and Computer Engineering, National Technical University of Athens, Athens, GRC
| | - Konstantinos Katsos
- Department of Forensic Medicine and Toxicology, National and Kapodistrian University of Athens School of Medicine, Athens, GRC
| | - Konstantinos Thomas
- Artificial Intelligence and Learning Systems Laboratory, School of Electrical and Computer Engineering, National Technical University of Athens, Athens, GRC
| | - Chara Spiliopoulou
- Department of Forensic Medicine and Toxicology, National and Kapodistrian University of Athens School of Medicine, Athens, GRC
| | - Giorgos Stamou
- Artificial Intelligence and Learning Systems Laboratory, School of Electrical and Computer Engineering, National Technical University of Athens, Athens, GRC
| | - Emmanouil I Sakelliadis
- Department of Forensic Medicine and Toxicology, National and Kapodistrian University of Athens School of Medicine, Athens, GRC
| |
Collapse
|
91
|
Liu D, Liu Z, Xia Y, Wang Z, Song J, Yu DJ. TransC-ac4C: Identification of N4-Acetylcytidine (ac4C) Sites in mRNA Using Deep Learning. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2024; 21:1403-1412. [PMID: 38607721 DOI: 10.1109/tcbb.2024.3386972] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/14/2024]
Abstract
N4-acetylcytidine (ac4C) is a post-transcriptional modification in mRNA that is critical in mRNA translation in terms of stability and regulation. In the past few years, numerous approaches employing convolutional neural networks (CNN) and Transformer have been proposed for the identification of ac4C sites, with each variety of approaches processing distinct characteristics. CNN-based methods excel at extracting local features and positional information, whereas Transformer-based ones stands out in establishing long-range dependencies and generating global representations. Given the importance of both local and global features in mRNA ac4C sites identification, we propose a novel method termed TransC-ac4C which combines CNN and Transformer together for enhancing the feature extraction capability and improving the identification accuracy. Five different feature encoding strategies (One-hot, NCP, ND, EIIP, and K-mer) are employed to generate the mRNA sequence representations, in which way the sequence attributes and physical and chemical properties of the sequences can be embedded. To strengthen the relevance of features, we construct a novel feature fusion method. Firstly, the CNN is employed to process five single features, stitch them together and feed them to the Transformer layer. Then, our approach employs CNN to extract local features and Transformer subsequently to establish global long-range dependencies among extracted features. We use 5-fold cross-validation to evaluate the model, and the evaluation indicators are significantly improved. The prediction accuracy of the two datasets is as high as 81.42% and 80.69%, respectively. It demonstrates the stronger competitiveness and generalization performance of our model.
Collapse
|
92
|
Mutnuri MK, Stelfox HT, Forkert ND, Lee J. Using Domain Adaptation and Inductive Transfer Learning to Improve Patient Outcome Prediction in the Intensive Care Unit: Retrospective Observational Study. J Med Internet Res 2024; 26:e52730. [PMID: 39167442 PMCID: PMC11375375 DOI: 10.2196/52730] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Revised: 03/22/2024] [Accepted: 07/08/2024] [Indexed: 08/23/2024] Open
Abstract
BACKGROUND Accurate patient outcome prediction in the intensive care unit (ICU) can potentially lead to more effective and efficient patient care. Deep learning models are capable of learning from data to accurately predict patient outcomes, but they typically require large amounts of data and computational resources. Transfer learning (TL) can help in scenarios where data and computational resources are scarce by leveraging pretrained models. While TL has been widely used in medical imaging and natural language processing, it has been rare in electronic health record (EHR) analysis. Furthermore, domain adaptation (DA) has been the most common TL method in general, whereas inductive transfer learning (ITL) has been rare. To the best of our knowledge, DA and ITL have never been studied in-depth in the context of EHR-based ICU patient outcome prediction. OBJECTIVE This study investigated DA, as well as rarely researched ITL, in EHR-based ICU patient outcome prediction under simulated, varying levels of data scarcity. METHODS Two patient cohorts were used in this study: (1) eCritical, a multicenter ICU data from 55,689 unique admission records from 48,672 unique patients admitted to 15 medical-surgical ICUs in Alberta, Canada, between March 2013 and December 2019, and (2) Medical Information Mart for Intensive Care III, a single-center, publicly available ICU data set from Boston, Massachusetts, acquired between 2001 and 2012 containing 61,532 admission records from 46,476 patients. We compared DA and ITL models with baseline models (without TL) of fully connected neural networks, logistic regression, and lasso regression in the prediction of 30-day mortality, acute kidney injury, ICU length of stay, and hospital length of stay. Random subsets of training data, ranging from 1% to 75%, as well as the full data set, were used to compare the performances of DA and ITL with the baseline models at various levels of data scarcity. RESULTS Overall, the ITL models outperformed the baseline models in 55 of 56 comparisons (all P values <.001). The DA models outperformed the baseline models in 45 of 56 comparisons (all P values <.001). ITL resulted in better performance than DA in terms of the number of times and the margin with which it outperformed the baseline models. In 11 of 16 cases (8 of 8 for ITL and 3 of 8 for DA), TL models outperformed baseline models when trained using 1% data subset. CONCLUSIONS TL-based ICU patient outcome prediction models are useful in data-scarce scenarios. The results of this study can be used to estimate ICU outcome prediction performance at different levels of data scarcity, with and without TL. The publicly available pretrained models from this study can serve as building blocks in further research for the development and validation of models in other ICU cohorts and outcomes.
Collapse
Affiliation(s)
- Maruthi Kumar Mutnuri
- Data Intelligence for Health Lab, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
- Department of Biomedical Engineering, Schulich School of Engineering, University of Calgary, Calgary, AB, Canada
| | - Henry Thomas Stelfox
- Department of Critical Care Medicine, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
- O'Brien Institute for Public Health, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Nils Daniel Forkert
- Department of Radiology, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
- Alberta Children's Hospital Research Institute, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Joon Lee
- Data Intelligence for Health Lab, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
- Department of Cardiac Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
- Department of Preventive Medicine, School of Medicine, Kyung Hee University, Seoul, Republic of Korea
| |
Collapse
|
93
|
Zhou Z, Li X, Ji H, Xu X, Chang Z, Wu K, Song Y, Kao M, Chen H, Wu D, Zhang T. Application of improved Unet network in the recognition and segmentation of lung CT images in patients with pneumoconiosis. BMC Med Imaging 2024; 24:220. [PMID: 39160488 PMCID: PMC11331615 DOI: 10.1186/s12880-024-01377-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Accepted: 07/22/2024] [Indexed: 08/21/2024] Open
Abstract
BACKGROUND Pneumoconiosis has a significant impact on the quality of patient survival. This study aims to evaluate the performance and application value of improved Unet network technology in the recognition and segmentation of lesion areas of lung CT images in patients with pneumoconiosis. METHODS A total of 1212 lung CT images of patients with pneumoconiosis were retrospectively included. The improved Unet network was used to identify and segment the CT image regions of the patients' lungs, and the image data of the granular regions of the lungs were processed by the watershed and region growing algorithms. After random sorting, 848 data were selected into the training set and 364 data into the validation set. The experimental dataset underwent data augmentation and were used for model training and validation to evaluate segmentation performance. The segmentation results were compared with FCN-8s, Unet network (Base), Unet (Squeeze-and-Excitation, SE + Rectified Linear Unit, ReLU), and Unet + + networks. RESULTS In the segmentation of lung CT granular region with the improved Unet network, the four evaluation indexes of Dice similarity coefficient, positive prediction value (PPV), sensitivity coefficient (SC) and mean intersection over union (MIoU) reached 0.848, 0.884, 0.895 and 0.885, respectively, increasing by 7.6%, 13.3%, 3.9% and 6.4%, respectively, compared with those of Unet network (Base), and increasing by 187.5%, 249.4%, 131.9% and 51.0%, respectively, compared with those of FCN-8s, and increasing by 14.0%, 31.2%, 4.7% and 9.7%, respectively, compared with those of Unet network (SE + ReLU), while the segmentation performance was also not inferior to that of the Unet + + network. CONCLUSIONS The improved Unet network proposed shows good performance in the recognition and segmentation of abnormal regions in lung CT images in patients with pneumoconiosis, showing potential application value for assisting clinical decision-making.
Collapse
Affiliation(s)
- Zhengsong Zhou
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Xin Li
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China
| | - Hongbo Ji
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Xuanhan Xu
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Zongqi Chang
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China
| | - Keda Wu
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Yangyang Song
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Mingkun Kao
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Hongjun Chen
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Dongsheng Wu
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China.
| | - Tao Zhang
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China.
| |
Collapse
|
94
|
Khan MK, Houran MA, Kauhaniemi K, Zafar MH, Mansoor M, Rashid S. Efficient state of charge estimation of lithium-ion batteries in electric vehicles using evolutionary intelligence-assisted GLA-CNN-Bi-LSTM deep learning model. Heliyon 2024; 10:e35183. [PMID: 39170306 PMCID: PMC11336464 DOI: 10.1016/j.heliyon.2024.e35183] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Revised: 07/23/2024] [Accepted: 07/24/2024] [Indexed: 08/23/2024] Open
Abstract
The battery's performance heavily influences the safety, dependability, and operational efficiency of electric vehicles (EVs). This paper introduces an innovative hybrid deep learning architecture that dramatically enhances the estimation of the state of charge (SoC) of lithium-ion (Li-ion) batteries, crucial for efficient EV operation. Our model uniquely integrates a convolutional neural network (CNN) with bidirectional long short-term memory (Bi-LSTM), optimized through evolutionary intelligence, enabling an advanced level of precision in SoC estimation. A novel aspect of this work is the application of the Group Learning Algorithm (GLA) to tune the hyperparameters of the CNN-Bi-LSTM network meticulously. This approach not only refines the model's accuracy but also significantly enhances its efficiency by optimizing each parameter to best capture and integrate both spatial and temporal information from the battery data. This is in stark contrast to conventional models that typically focus on either spatial or temporal data, but not both effectively. The model's robustness is further demonstrated through its training across six diverse datasets that represent a range of EV discharge profiles, including the Highway Fuel Economy Test (HWFET), the US06 test, the Beijing Dynamic Stress Test (BJDST), the dynamic stress test (DST), the federal urban driving schedule (FUDS), and the urban development driving schedule (UDDS). These tests are crucial for ensuring that the model can perform under various real-world conditions. Experimentally, our hybrid model not only surpasses the performance of existing LSTM and CNN frameworks in tracking SoC estimation but also achieves an impressively quick convergence to true SoC values, maintaining an average root mean square error (RMSE) of less than 1 %. Furthermore, the experimental outcomes suggest that this new deep learning methodology outstrips conventional approaches in both convergence speed and estimation accuracy, thus promising to significantly enhance battery life and overall EV efficiency.
Collapse
Affiliation(s)
| | - Mohamad Abou Houran
- School of Electrical Engineering, Xi'an Jiaotong University, No. 28, West Xianning Road, Xi'an, 710049, China
| | - Kimmo Kauhaniemi
- School of Technology and Innovation, University of Vaasa, Finland
| | - Muhammad Hamza Zafar
- Department of Engineering Sciences, University of Agder, NO-4879, Grimstad, Norway
| | - Majad Mansoor
- Department of Automation, University of Science and Technology of China, Hefei, China
| | - Saad Rashid
- Department of Electrical Engineering, Hamdard University, Islamabad Campus, Islamabad, Pakistan
| |
Collapse
|
95
|
Ozkan EE, Serel TA, Soyupek AS, Kaymak ZA. Utilization of machine learning methods for prediction of acute and late rectal toxicity due to curative prostate radiotherapy. RADIATION PROTECTION DOSIMETRY 2024; 200:1244-1250. [PMID: 38932433 DOI: 10.1093/rpd/ncae154] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Revised: 04/17/2024] [Accepted: 06/13/2024] [Indexed: 06/28/2024]
Abstract
OBJECTIVE Rectal toxicity is one of the primary dose-limiting side effects of prostate cancer radiotherapy, and consequential impairment on quality of life in these patients with long survival is an important problem. In this study, we aimed to evaluate the possibility of predicting rectal toxicity with artificial intelligence model which was including certain dosimetric parameters. MATERIALS AND METHODS One hundred and thirty-seven patients with a diagnosis of prostate cancer who received curative radiotherapy for prostate +/- pelvic lymphatics were included in the study. The association of the clinical data and dosimetric data between early and late rectal toxicity reported during follow-up was evaluated. The sample size was increased to 274 patients by synthetic data generation method. To determine suitable models, 15 models were studied with machine learning algorithms using Python 2.3, Pycaret library. Random forest classifier was used with to detect active variables. RESULTS The area under the curve and accuracy were found to be 0.89-0.97 and 95%-99%, respectively, with machine learning algorithms. The sensitivity values for acute and toxicity were found to be 0.95 and 0.99, respectively. CONCLUSION Early or late rectal toxicity can be predicted with a high probability via dosimetric and physical data and machine learning algorithms of patients who underwent prostate +/- pelvic radiotherapy. The fact that rectal toxicity can be predicted before treatment, which may result in limiting the dose and duration of treatment, makes us think that artificial intelligence can enter our daily practice in a short time in this sense.
Collapse
Affiliation(s)
- Emine Elif Ozkan
- Department of Radiation Oncology, Suleyman Demirel University, Isparta, 32260, Türkiye
| | - Tekin Ahmet Serel
- Department of Urology, Suleyman Demirel University, Isparta, 32260, Türkiye
| | - Arap Sedat Soyupek
- Department of Urology, Suleyman Demirel University, Isparta, 32260, Türkiye
| | - Zumrut Arda Kaymak
- Department of Radiation Oncology, Suleyman Demirel University, Isparta, 32260, Türkiye
| |
Collapse
|
96
|
Nijiati M, Tuerdi M, Damola M, Yimit Y, Yang J, Abulaiti A, Mutailifu A, Aihait D, Wang Y, Zou X. A deep learning radiomics model based on CT images for predicting the biological activity of hepatic cystic echinococcosis. Front Physiol 2024; 15:1426468. [PMID: 39175611 PMCID: PMC11338923 DOI: 10.3389/fphys.2024.1426468] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Accepted: 07/15/2024] [Indexed: 08/24/2024] Open
Abstract
Introduction: Hepatic cystic echinococcosis (HCE) is a widely seen parasitic infection. Biological activity is crucial for treatment planning. This work aims to explore the potential applications of a deep learning radiomics (DLR) model, based on CT images, in predicting the biological activity grading of hepatic cystic echinococcosis. Methods: A retrospective analysis of 160 patients with hepatic echinococcosis was performed (127 and 33 in training and validation sets). Volume of interests (VOIs) were drawn, and radiomics features and deep neural network features were extracted. Feature selection was performed on the training set, and radiomics score (Rad Score) and deep learning score (Deep Score) were calculated. Seven diagnostics models (based on logistic regression algorithm) for the biological activity grading were constructed using the selected radiomics features and two deep model features respectively. All models were evaluated using the receiver operating characteristic curve, and the area under the curve (AUC) was calculated. A nomogram was constructed using the combined model, and its calibration, discriminatory ability, and clinical utility were assessed. Results: 12, 6 and 10 optimal radiomics features, deep learning features were selected from two deep learning network (DLN) features, respectively. For biological activity grading of hepatic cystic echinococcosis, the combined model demonstrated strong diagnostic performance, with an AUC value of 0.888 (95% CI: 0.837-0.936) in the training set and 0.876 (0.761-0.964) in the validation set. The clinical decision analysis curve indicated promising results, while the calibration curve revealed that the nomogram's prediction result was highly compatible with the actual result. Conclusion: The DLR model can be used for predicting the biological activity grading of hepatic echinococcosis.
Collapse
Affiliation(s)
- Mayidili Nijiati
- Department of Radiology, The Fourth Affiliated Hospital of Xinjiang Medical UniversityÜrümqi, Xinjiang, China
- Department Xinjiang Key Laboratory of Artificial Intelligence Assisted Imaging Diagnosis, Kashi, China
| | - Mireayi Tuerdi
- Department of Infectious Diseases, The First People’s Hospital of Kashi Prefecture, Kashi, China
| | - Maihemitijiang Damola
- Department of Radiology, The First People’s Hospital of Kashi Prefecture, Kashi, China
| | - Yasen Yimit
- Department Xinjiang Key Laboratory of Artificial Intelligence Assisted Imaging Diagnosis, Kashi, China
- Department of Radiology, The First People’s Hospital of Kashi Prefecture, Kashi, China
| | - Jing Yang
- Huiying Medical Imaging Technology, The Fourth Affiliated Hospital of Xinjiang Medical University, Beijing, China
| | - Adilijiang Abulaiti
- Department of Radiology, The First People’s Hospital of Kashi Prefecture, Kashi, China
| | | | - Diliaremu Aihait
- Department of Radiology, The First People’s Hospital of Kashi Prefecture, Kashi, China
| | - Yunling Wang
- Department of Imaging Center, The First Affiliated Hospital of Xinjiang Medical University, Ürümqi, China
| | - Xiaoguang Zou
- Department Xinjiang Key Laboratory of Artificial Intelligence Assisted Imaging Diagnosis, Kashi, China
- Clinical Medical Research Center, The First People’s Hospital of Kashi Prefecture, Kashi, China
| |
Collapse
|
97
|
Eun NL, Lee E, Park AY, Son EJ, Kim JA, Youk JH. Artificial intelligence for ultrasound microflow imaging in breast cancer diagnosis. ULTRASCHALL IN DER MEDIZIN (STUTTGART, GERMANY : 1980) 2024; 45:412-417. [PMID: 38593859 DOI: 10.1055/a-2230-2455] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/11/2024]
Abstract
PURPOSE To develop and evaluate artificial intelligence (AI) algorithms for ultrasound (US) microflow imaging (MFI) in breast cancer diagnosis. MATERIALS AND METHODS We retrospectively collected a dataset consisting of 516 breast lesions (364 benign and 152 malignant) in 471 women who underwent B-mode US and MFI. The internal dataset was split into training (n = 410) and test datasets (n = 106) for developing AI algorithms from deep convolutional neural networks from MFI. AI algorithms were trained to provide malignancy risk (0-100%). The developed AI algorithms were further validated with an independent external dataset of 264 lesions (229 benign and 35 malignant). The diagnostic performance of B-mode US, AI algorithms, or their combinations was evaluated by calculating the area under the receiver operating characteristic curve (AUROC). RESULTS The AUROC of the developed three AI algorithms (0.955-0.966) was higher than that of B-mode US (0.842, P < 0.0001). The AUROC of the AI algorithms on the external validation dataset (0.892-0.920) was similar to that of the test dataset. Among the AI algorithms, no significant difference was found in all performance metrics combined with or without B-mode US. Combined B-mode US and AI algorithms had a higher AUROC (0.963-0.972) than that of B-mode US (P < 0.0001). Combining B-mode US and AI algorithms significantly decreased the false-positive rate of BI-RADS category 4A lesions from 87% to 13% (P < 0.0001). CONCLUSION AI-based MFI diagnosed breast cancers with better performance than B-mode US, eliminating 74% of false-positive diagnoses in BI-RADS category 4A lesions.
Collapse
Affiliation(s)
- Na Lae Eun
- Radiology, Gangnam Severance Hospital, Seoul, Korea (the Republic of)
| | - Eunjung Lee
- Computational Science and Engineering, Yonsei University, Seoul, Korea (the Republic of)
| | - Ah Young Park
- Radiology, Bundang CHA Medical Center, Seongnam, Korea (the Republic of)
| | - Eun Ju Son
- Radiology, Gangnam Severance Hospital, Seoul, Korea (the Republic of)
| | - Jeong-Ah Kim
- Radiology, Gangnam Severance Hospital, Seoul, Korea (the Republic of)
| | - Ji Hyun Youk
- Department of Radiology, Yonsei University College of Medicine, Seoul, Korea, Republic of
- Radiology, Gangnam Severance Hospital, Seoul, Korea (the Republic of)
| |
Collapse
|
98
|
Ali IE, Sumita Y, Wakabayashi N. Advancing maxillofacial prosthodontics by using pre-trained convolutional neural networks: Image-based classification of the maxilla. J Prosthodont 2024; 33:645-654. [PMID: 38566564 DOI: 10.1111/jopr.13853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2023] [Accepted: 03/15/2024] [Indexed: 04/04/2024] Open
Abstract
PURPOSE The study aimed to compare the performance of four pre-trained convolutional neural networks in recognizing seven distinct prosthodontic scenarios involving the maxilla, as a preliminary step in developing an artificial intelligence (AI)-powered prosthesis design system. MATERIALS AND METHODS Seven distinct classes, including cleft palate, dentulous maxillectomy, edentulous maxillectomy, reconstructed maxillectomy, completely dentulous, partially edentulous, and completely edentulous, were considered for recognition. Utilizing transfer learning and fine-tuned hyperparameters, four AI models (VGG16, Inception-ResNet-V2, DenseNet-201, and Xception) were employed. The dataset, consisting of 3541 preprocessed intraoral occlusal images, was divided into training, validation, and test sets. Model performance metrics encompassed accuracy, precision, recall, F1 score, area under the receiver operating characteristic curve (AUC), and confusion matrix. RESULTS VGG16, Inception-ResNet-V2, DenseNet-201, and Xception demonstrated comparable performance, with maximum test accuracies of 0.92, 0.90, 0.94, and 0.95, respectively. Xception and DenseNet-201 slightly outperformed the other models, particularly compared with InceptionResNet-V2. Precision, recall, and F1 scores exceeded 90% for most classes in Xception and DenseNet-201 and the average AUC values for all models ranged between 0.98 and 1.00. CONCLUSIONS While DenseNet-201 and Xception demonstrated superior performance, all models consistently achieved diagnostic accuracy exceeding 90%, highlighting their potential in dental image analysis. This AI application could help work assignments based on difficulty levels and enable the development of an automated diagnosis system at patient admission. It also facilitates prosthesis designing by integrating necessary prosthesis morphology, oral function, and treatment difficulty. Furthermore, it tackles dataset size challenges in model optimization, providing valuable insights for future research.
Collapse
Affiliation(s)
- Islam E Ali
- Department of Advanced Prosthodontics, Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
- Department of Prosthodontics, Faculty of Dentistry, Mansoura University, Mansoura, Egypt
| | - Yuka Sumita
- Division of General Dentistry 4, The Nippon Dental University Hospital, Tokyo, Japan
- Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
| | - Noriyuki Wakabayashi
- Department of Advanced Prosthodontics, Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
| |
Collapse
|
99
|
Huang J, Saw SN, He T, Yang R, Qin Y, Chen Y, Kiong LC. DCNNLFS: A Dilated Convolutional Neural Network With Late Fusion Strategy for Intelligent Classification of Gastric Histopathology Images. IEEE J Biomed Health Inform 2024; 28:4534-4543. [PMID: 37983160 DOI: 10.1109/jbhi.2023.3334709] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2023]
Abstract
Gastric cancer has a high incidence rate, significantly threatening patients' health. Gastric histopathology images can reliably diagnose related diseases. Still, the data volume of histopathology images is too large, making misdiagnosis or missed diagnosis easy. The classification model based on deep learning has made some progress on gastric histopathology images. However, traditional convolutional neural networks (CNNs) generally use pooling operations, which will reduce the spatial resolution of the image, resulting in poor prediction results. The image feature in previous CNN has a poor perception of details. Therefore, we design a dilated CNN with a late fusion strategy (DCNNLFS) for gastric histopathology image classification. The DCNNLFS model utilizes dilated convolutions, enabling it to expand the receptive field. The dilated convolutions can learn the different contextual information by adjusting the dilation rate. The DCNNLFS model uses a late fusion strategy to enhance the classification ability of DCNNLFS. We run related experiments on a gastric histopathology image dataset to verify the excellence of the DCNNLFS model, where the three metrics Precision, Accuracy, and F1-Score are 0.938, 0.935, and 0.959.
Collapse
|
100
|
Tourani R, Murphree DH, Sheka A, Melton GB, Kor DJ, Simon GJ. Consensus modeling: Safer transfer learning for small health systems. Artif Intell Med 2024; 154:102899. [PMID: 38843692 DOI: 10.1016/j.artmed.2024.102899] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2021] [Revised: 03/25/2022] [Accepted: 05/21/2024] [Indexed: 08/09/2024]
Abstract
Predictive modeling is becoming an essential tool for clinical decision support, but health systems with smaller sample sizes may construct suboptimal or overly specific models. Models become over-specific when beside true physiological effects, they also incorporate potentially volatile site-specific artifacts. These artifacts can change suddenly and can render the model unsafe. To obtain safer models, health systems with inadequate sample sizes may adopt one of the following options. First, they can use a generic model, such as one purchased from a vendor, but often such a model is not sufficiently specific to the patient population and is thus suboptimal. Second, they can participate in a research network. Paradoxically though, sites with smaller datasets contribute correspondingly less to the joint model, again rendering the final model suboptimal. Lastly, they can use transfer learning, starting from a model trained on a large data set and updating this model to the local population. This strategy can also result in a model that is over-specific. In this paper we present the consensus modeling paradigm, which uses the help of a large site (source) to reach a consensus model at the small site (target). We evaluate the approach on predicting postoperative complications at two health systems with 9,044 and 38,045 patients (rare outcomes at about 1% positive rate), and conduct a simulation study to understand the performance of consensus modeling relative to the other three approaches as a function of the available training sample size at the target site. We found that consensus modeling exhibited the least over-specificity at either the source or target site and achieved the highest combined predictive performance.
Collapse
Affiliation(s)
- Roshan Tourani
- Institute for Health Informatics, University of Minnesota, Twin Cities, MN, United States of America.
| | - Dennis H Murphree
- Department of Health Sciences Research, Mayo Clinic, MN, United States of America.
| | - Adam Sheka
- Department of Surgery, University of Minnesota, Twin Cities, MN, United States of America.
| | - Genevieve B Melton
- Institute for Health Informatics, University of Minnesota, Twin Cities, MN, United States of America; Department of Surgery, University of Minnesota, Twin Cities, MN, United States of America.
| | - Daryl J Kor
- Department of Anesthesia, Mayo Clinic, MN, United States of America.
| | - Gyorgy J Simon
- Institute for Health Informatics, University of Minnesota, Twin Cities, MN, United States of America; Department of Medicine, University of Minnesota, MN, United States of America.
| |
Collapse
|