1
|
Aruk I, Pacal I, Toprak AN. A novel hybrid ConvNeXt-based approach for enhanced skin lesion classification. EXPERT SYSTEMS WITH APPLICATIONS 2025; 283:127721. [DOI: 10.1016/j.eswa.2025.127721] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
2
|
Pacal I, Attallah O. Hybrid deep learning model for automated colorectal cancer detection using local and global feature extraction. Knowl Based Syst 2025; 319:113625. [DOI: 10.1016/j.knosys.2025.113625] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
3
|
Pacal I, Ozdemir B, Zeynalov J, Gasimov H, Pacal N. A novel CNN-ViT-based deep learning model for early skin cancer diagnosis. Biomed Signal Process Control 2025; 104:107627. [DOI: 10.1016/j.bspc.2025.107627] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
4
|
Alom MR, Farid FA, Rahaman MA, Rahman A, Debnath T, Miah ASM, Mansor S. An explainable AI-driven deep neural network for accurate breast cancer detection from histopathological and ultrasound images. Sci Rep 2025; 15:17531. [PMID: 40394112 PMCID: PMC12092800 DOI: 10.1038/s41598-025-97718-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2024] [Accepted: 04/07/2025] [Indexed: 05/22/2025] Open
Abstract
Breast cancer represents a significant global health challenge, which makes it essential to detect breast cancer early and accurately to improve patient prognosis and reduce mortality rates. However, traditional diagnostic processes relying on manual analysis of medical images are inherently complex and subject to variability between observers, highlighting the urgent need for robust automated breast cancer detection systems. While deep learning has demonstrated potential, many current models struggle with limited accuracy and lack of interpretability. This research introduces the Deep Neural Breast Cancer Detection (DNBCD) model, an explainable AI-based framework that utilizes deep learning methods for classifying breast cancer using histopathological and ultrasound images. The proposed model employs Densenet121 as a foundation, integrating customized Convolutional Neural Network (CNN) layers including GlobalAveragePooling2D, Dense, and Dropout layers along with transfer learning to achieve both high accuracy and interpretability for breast cancer diagnosis. The proposed DNBCD model integrates several preprocessing techniques, including image normalization and resizing, and augmentation techniques to enhance the model's robustness and address class imbalances using class weight. It employs Grad-CAM (Gradient-weighted Class Activation Mapping) to offer visual justifications for its predictions, increasing trust and transparency among healthcare providers. The model was assessed using two benchmark datasets: Breakhis-400x (B-400x) and Breast Ultrasound Images Dataset (BUSI) containing 1820 and 1578 images, respectively. We systematically divided the datasets into training (70%), testing (20%,) and validation (10%) sets, ensuring efficient model training and evaluation obtaining accuracies of 93.97% for B-400x dataset having benign and malignant classes and 89.87% for BUSI dataset having benign, malignant, and normal classes for breast cancer detection. Experimental results demonstrate that the proposed DNBCD model significantly outperforms existing state-of-the-art approaches with potential uses in clinical environments. We also made all the materials publicly accessible for the research community at: https://github.com/romzanalom/XAI-Based-Deep-Neural-Breast-Cancer-Detection .
Collapse
Affiliation(s)
- Md Romzan Alom
- Department of Computer Science and Engineering, Green University of Bangladesh (GUB), Purbachal American City, Kanchon, Dhaka, 1460, Bangladesh
| | - Fahmid Al Farid
- Faculty of Artificial Intelligence and Engineering, Multimedia University, 63100, Cyberjaya, Malaysia
| | - Muhammad Aminur Rahaman
- Department of Computer Science and Engineering, Green University of Bangladesh (GUB), Purbachal American City, Kanchon, Dhaka, 1460, Bangladesh.
| | - Anichur Rahman
- Department of Computer Science and Engineering, National Institute of Textile Engineering and Research (NITER), Constituent Institute of the University of Dhaka, Savar, Dhaka, 1350, Bangladesh.
- Department of Computer Science and Engineering, Mawlana Bhashani Science and Technology University, Tangail, Bangladesh.
| | - Tanoy Debnath
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Abu Saleh Musa Miah
- Department of Computer Science and Engineering, Bangladesh Army University of Science and Technology (BAUST), Nilphamari, Bangladesh
| | - Sarina Mansor
- Faculty of Artificial Intelligence and Engineering, Multimedia University, 63100, Cyberjaya, Malaysia.
| |
Collapse
|
5
|
Atteia G, Alabdulhafith M, Abdallah HA, Abdel Samee N, Alayed W. Deep learning-based decision support system for cervical cancer identification in liquid-based cytology pap smears. Technol Health Care 2025:9287329251330081. [PMID: 40302490 DOI: 10.1177/09287329251330081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/02/2025]
Abstract
BackgroundCervical cancer is the fourth most common cause of women cancer deaths worldwide. The primary etiology of cervical cancer is the persistent infection of specific high-risk strains of the human papillomavirus. Liquid-based cytology is the established method for early detection of cervical cancer. The evaluation of cellular abnormalities at a microscopic level allows for the identification of malignant or precancerous features in liquid-based cytology pap smears. This technique is characterized by its time-consuming nature and susceptibility to both inter- and intra-observer variability. Hence, the utilization of Artificial Intelligence in computer-assisted diagnosis can reduce the duration needed for diagnosing this ailment, thereby eliminating delayed diagnosis and facilitating the implementation of an efficient treatment.ObjectiveThis research presents a new deep learning-based cervical cancer identification decision support system in liquid-based cytology smear images.MethodsThe proposed diagnosis support system incorporates a novel hybrid feature reduction and optimization module, which integrates a sparse Autoencoder with the Binary Harris Hawk metaheuristic optimization algorithm to select the most informative features from a supplemented feature set of the input images. The supplemented feature set is retrieved by three pretrained Convolutional Neural Networks. The module utilizes an improved feature set to conduct a Bayesian-optimized K Nearest Neighbors machine learning classification of cervical cancer in input Pap smears.ResultsThe introduced approach achieves a classification accuracy of 99.9% and demonstrates an improved ability to detect the stages of cervical cancer, with a sensitivity of 99.8%. In addition, the system has the ability to identify the lack of cervical cancer stages with a specificity rate of 99.9%.ConclusionThe proposed system outpaces recent deep learning-based cervical cancer identification systems.
Collapse
Affiliation(s)
- Ghada Atteia
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Maali Alabdulhafith
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Hanaa A Abdallah
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Nagwan Abdel Samee
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Walaa Alayed
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| |
Collapse
|
6
|
Asif RN, Naseem MT, Ahmad M, Mazhar T, Khan MA, Khan MA, Al-Rasheed A, Hamam H. Brain tumor detection empowered with ensemble deep learning approaches from MRI scan images. Sci Rep 2025; 15:15002. [PMID: 40301625 PMCID: PMC12041211 DOI: 10.1038/s41598-025-99576-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2025] [Accepted: 04/21/2025] [Indexed: 05/01/2025] Open
Abstract
Brain tumor detection is essential for early diagnosis and successful treatment, both of which can significantly enhance patient outcomes. To evaluate brain MRI scans and categorize them into four types-pituitary, meningioma, glioma, and normal-this study investigates a potent artificial intelligence (AI) technique. Even though AI has been utilized in the past to detect brain tumors, current techniques still have issues with accuracy and dependability. Our study presents a novel AI technique that combines two distinct deep learning models to enhance this. When combined, these models improve accuracy and yield more trustworthy outcomes than when used separately. Key performance metrics including accuracy, precision, and dependability are used to assess the system once it has been trained using MRI scan pictures. Our results show that this combined AI approach works better than individual models, particularly in identifying different types of brain tumors. Specifically, the InceptionV3 + Xception combination hit an accuracy level of 98.50% in training and 98.30% in validation. Such results further argue the potential application for advanced AI techniques in medical imaging while speaking even more strongly to the fact that multiple AI models used concurrently are able to enhance brain tumor detection.
Collapse
Affiliation(s)
- Rizwana Naz Asif
- School of Computer Science, National College of Business Administration and Economics, Lahore, 54000, Pakistan
| | - Muhammad Tahir Naseem
- Department of Electronic Engineering, Yeungnam University, Gyeongsan-si, 38541, Republic of Korea
| | - Munir Ahmad
- University College, Korea University, Seoul, 02841, Republic of Korea
| | - Tehseen Mazhar
- School of Computer Science, National College of Business Administration and Economics, Lahore, 54000, Pakistan.
- Department of Computer Science,School Education Department, Government of Punjab, Layyah, 31200, Pakistan.
| | - Muhammad Adnan Khan
- Department of Software, Faculty of Artificial Intelligence and Software, Gachon University, Seongnam-si, 13557, Republic of Korea.
| | - Muhammad Amir Khan
- School of Computing Sciences, College of Computing, Informatics and Mathematics, Universiti Teknologi MARA, Shah Alam, 40450, Selangor, Malaysia.
| | - Amal Al-Rasheed
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, 11671, Saudi Arabia
| | - Habib Hamam
- Faculty of Engineering, Université de Moncton, Moncton, NB E1 A3E9, Canada
- School of Electrical Engineering, University of Johannesburg, Johannesburg, 2006, South Africa
- International Institute of Technology and Management (IITG), Av. Grandes Ecoles, Libreville BP 1989, Gabon
- Bridges for Academic Excellence, Spectrum, Tunis, Center-ville, Tunisia
| |
Collapse
|
7
|
Zubair M, Owais M, Hassan T, Bendechache M, Hussain M, Hussain I, Werghi N. An interpretable framework for gastric cancer classification using multi-channel attention mechanisms and transfer learning approach on histopathology images. Sci Rep 2025; 15:13087. [PMID: 40240457 PMCID: PMC12003787 DOI: 10.1038/s41598-025-97256-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2024] [Accepted: 04/03/2025] [Indexed: 04/18/2025] Open
Abstract
The importance of gastric cancer (GC) and the role of deep learning techniques in categorizing GC histopathology images have recently increased. Identifying the drawbacks of traditional deep learning models, including lack of interpretability, inability to capture complex patterns, lack of adaptability, and sensitivity to noise. A multi-channel attention mechanism-based framework is proposed that can overcome the limitations of conventional deep learning models by dynamically focusing on relevant features, enhancing extraction, and capturing complex relationships in medical data. The proposed framework uses three different attention mechanism channels and convolutional neural networks to extract multichannel features during the classification process. The proposed framework's strong performance is confirmed by competitive experiments conducted on a publicly available Gastric Histopathology Sub-size Image Database, which yielded remarkable classification accuracies of 99.07% and 98.48% on the validation and testing sets, respectively. Additionally, on the HCRF dataset, the framework achieved high classification accuracy of 99.84% and 99.65% on the validation and testing sets, respectively. The effectiveness and interchangeability of the three channels are further confirmed by ablation and interchangeability experiments, highlighting the remarkable performance of the framework in GC histopathological image classification tasks. This offers an advanced and pragmatic artificial intelligence solution that addresses challenges posed by unique medical image characteristics for intricate image analysis. The proposed approach in artificial intelligence medical engineering demonstrates significant potential for enhancing diagnostic precision by achieving high classification accuracy and treatment outcomes.
Collapse
Affiliation(s)
- Muhammad Zubair
- Interdisciplinary Research Center for Finance and Digital Economy, King Fahd University of Petroleum and Minerals, 31261, Dhahran, Saudi Arabia
| | - Muhammad Owais
- Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates.
| | - Taimur Hassan
- Departement of Electrical and Computer Engineering, Abu Dhabi University, Abu Dhabi, United Arab Emirates
| | - Malika Bendechache
- ADAPT Research Centre, School of Computer Science, University of Galway, H91 TK33, Galway, Ireland
| | - Muzammil Hussain
- Department of Software Engineering, Faculty of Information Technology, Al-Ahliyya Amman University, Amman, Jordan
| | - Irfan Hussain
- Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates
| | - Naoufel Werghi
- Department of Computer Science, Khalifa University, Abu Dhabi, United Arab Emirates
| |
Collapse
|
8
|
Chauhan NK, Singh K, Kumar A, Mishra A, Gupta SK, Mahajan S, Kadry S, Kim J. A hybrid learning network with progressive resizing and PCA for diagnosis of cervical cancer on WSI slides. Sci Rep 2025; 15:12801. [PMID: 40229435 PMCID: PMC11997219 DOI: 10.1038/s41598-025-97719-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2024] [Accepted: 04/07/2025] [Indexed: 04/16/2025] Open
Abstract
Current artificial intelligence (AI) trends are revolutionizing medical image processing, greatly improving cervical cancer diagnosis. Machine learning (ML) algorithms can discover patterns and anomalies in medical images, whereas deep learning (DL) methods, specifically convolutional neural networks (CNNs), are extremely accurate at identifying malignant lesions. Deep models that have been pre-trained and tailored through transfer learning and fine-tuning become faster and more effective, even when data is scarce. This paper implements a state-of-the-art Hybrid Learning Network that combines the Progressive Resizing approach and Principal Component Analysis (PCA) for enhanced cervical cancer diagnostics of whole slide images (WSI) slides. ResNet-152 and VGG-16, two fine-tuned DL models, are employed together with transfer learning to train on augmented and progressively resized training data with dimensions of 224 × 224, 512 × 512, and 1024 × 1024 pixels for enhanced feature extraction. Principal component analysis (PCA) is subsequently employed to process the combined features extracted from two DL models and reduce the dimensional space of the feature set. Furthermore, two ML methods, Support Vector Machine (SVM) and Random Forest (RF) models, are trained on this reduced feature set, and their predictions are integrated using a majority voting approach for evaluating the final classification results, thereby enhancing overall accuracy and reliability. The accuracy of the suggested framework on SIPaKMeD data is 99.29% for two-class classification and 98.47% for five-class classification. Furthermore, it achieves 100% accuracy for four-class categorization on the LBC dataset.
Collapse
Affiliation(s)
- Nitin Kumar Chauhan
- Department of ECE, Indore Institute of Science & Technology, Indore, 453331, India
| | - Krishna Singh
- DSEU Okhla Campus-I, Formerly G. B. Pant Engineering College, New Delhi, 110020, India
| | - Amit Kumar
- Department of Electronics Engineering, Indian Institute of Technology (BHU), Varanasi, 221005, India
| | - Ashutosh Mishra
- Department of Electrical and Electronics Engineering, Birla Institute of Technology and Science (BITS), Pilani Dubai Campus, 345055, Dubai International Academic City, Dubai, United Arab Emirates
| | - Sachin Kumar Gupta
- Department of Electronics and Communication Engineering, Central University of Jammu, Samba, Jammu, 181143, India
| | - Shubham Mahajan
- Amity School of Engineering & Technology, Amity University, Haryana, India.
| | - Seifedine Kadry
- Department of Computer Science and Mathematics, Lebanese American University, Beirut, Lebanon
- Noroff University College, Kristiansand, Norway
| | - Jungeun Kim
- Department of Computer Engineering, Inha University, Incheon, Republic of Korea.
| |
Collapse
|
9
|
Ekinci F, Ugurlu G, Ozcan GS, Acici K, Asuroglu T, Kumru E, Guzel MS, Akata I. Classification of Mycena and Marasmius Species Using Deep Learning Models: An Ecological and Taxonomic Approach. SENSORS (BASEL, SWITZERLAND) 2025; 25:1642. [PMID: 40292694 PMCID: PMC11945257 DOI: 10.3390/s25061642] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2025] [Revised: 02/25/2025] [Accepted: 03/05/2025] [Indexed: 04/30/2025]
Abstract
Fungi play a critical role in ecosystems, contributing to biodiversity and providing economic and biotechnological value. In this study, we developed a novel deep learning-based framework for the classification of seven macrofungi species from the genera Mycena and Marasmius, leveraging their unique ecological and morphological characteristics. The proposed approach integrates a custom convolutional neural network (CNN) with a self-organizing map (SOM) adapted for supervised learning and a Kolmogorov-Arnold Network (KAN) layer to enhance classification performance. The experimental results demonstrate significant improvements in classification metrics when using the CNN-SOM and CNN-KAN architectures. Additionally, advanced pretrained models such as MaxViT-S and ResNetV2-50 achieved high accuracy rates, with MaxViT-S achieving 98.9% accuracy. Statistical analyses using the chi-square test confirmed the reliability of the results, emphasizing the importance of validating evaluation metrics statistically. This research represents the first application of SOM in fungal classification and highlights the potential of deep learning in advancing fungal taxonomy. Future work will focus on optimizing the KAN architecture and expanding the dataset to include more fungal classes, further enhancing classification accuracy and ecological understanding.
Collapse
Affiliation(s)
- Fatih Ekinci
- Institute of Artificial Intelligence, Ankara University, Ankara 06100, Türkiye; (F.E.); (K.A.); (M.S.G.)
| | - Guney Ugurlu
- Department of Computer Engineering, Faculty of Engineering, Başkent University, Ankara 06790, Türkiye; (G.U.); (G.S.O.)
| | - Giray Sercan Ozcan
- Department of Computer Engineering, Faculty of Engineering, Başkent University, Ankara 06790, Türkiye; (G.U.); (G.S.O.)
| | - Koray Acici
- Institute of Artificial Intelligence, Ankara University, Ankara 06100, Türkiye; (F.E.); (K.A.); (M.S.G.)
- Department of Artificial Intelligence and Data Engineering, Faculty of Engineering, Ankara University, Ankara 06830, Türkiye
| | - Tunc Asuroglu
- Faculty of Medicine and Health Technology, Tampere University, 33720 Tampere, Finland
- VTT Technical Research Centre of Finland, 33101 Tampere, Finland
| | - Eda Kumru
- Graduate School of Natural and Applied Sciences, Ankara University, Ankara 06830, Türkiye;
| | - Mehmet Serdar Guzel
- Institute of Artificial Intelligence, Ankara University, Ankara 06100, Türkiye; (F.E.); (K.A.); (M.S.G.)
- Department of Computer Engineering, Faculty of Engineering, Ankara University, Ankara 06830, Türkiye
| | - Ilgaz Akata
- Department of Biology, Faculty of Science, Ankara University, Ankara 06100, Türkiye;
| |
Collapse
|
10
|
Bayram B, Kunduracioglu I, Ince S, Pacal I. A systematic review of deep learning in MRI-based cerebral vascular occlusion-based brain diseases. Neuroscience 2025; 568:76-94. [PMID: 39805420 DOI: 10.1016/j.neuroscience.2025.01.020] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 01/09/2025] [Accepted: 01/10/2025] [Indexed: 01/16/2025]
Abstract
Neurological disorders, including cerebral vascular occlusions and strokes, present a major global health challenge due to their high mortality rates and long-term disabilities. Early diagnosis, particularly within the first hours, is crucial for preventing irreversible damage and improving patient outcomes. Although neuroimaging techniques like magnetic resonance imaging (MRI) have advanced significantly, traditional methods often fail to fully capture the complexity of brain lesions. Deep learning has recently emerged as a powerful tool in medical imaging, offering high accuracy in detecting and segmenting brain anomalies. This review examines 61 MRI-based studies published between 2020 and 2024, focusing on the role of deep learning in diagnosing cerebral vascular occlusion-related conditions. It evaluates the successes and limitations of these studies, including the adequacy and diversity of datasets, and addresses challenges such as data privacy and algorithm explainability. Comparisons between convolutional neural network (CNN)-based and Vision Transformer (ViT)-based approaches reveal distinct advantages and limitations. The findings emphasize the importance of ethically secure frameworks, the inclusion of diverse datasets, and improved model interpretability. Advanced architectures like U-Net variants and transformer-based models are highlighted as promising tools to enhance reliability in clinical applications. By automating complex neuroimaging tasks and improving diagnostic accuracy, deep learning facilitates personalized treatment strategies. This review provides a roadmap for integrating technical advancements into clinical practice, underscoring the transformative potential of deep learning in managing neurological disorders and improving healthcare outcomes globally.
Collapse
Affiliation(s)
- Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| | - Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| |
Collapse
|
11
|
Ozdemir B, Pacal I. An innovative deep learning framework for skin cancer detection employing ConvNeXtV2 and focal self-attention mechanisms. RESULTS IN ENGINEERING 2025; 25:103692. [DOI: 10.1016/j.rineng.2024.103692] [Citation(s) in RCA: 11] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
12
|
Alzahrani AA, Alsamri J, Maashi M, Negm N, Asklany SA, Alkharashi A, Alkhiri H, Obayya M. Deep structured learning with vision intelligence for oral carcinoma lesion segmentation and classification using medical imaging. Sci Rep 2025; 15:6610. [PMID: 39994267 PMCID: PMC11850820 DOI: 10.1038/s41598-025-89971-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2024] [Accepted: 02/10/2025] [Indexed: 02/26/2025] Open
Abstract
Oral carcinoma (OC) is a toxic illness among the most general malignant cancers globally, and it has developed a gradually significant public health concern in emerging and low-to-middle-income states. Late diagnosis, high incidence, and inadequate treatment strategies remain substantial challenges. Analysis at an initial phase is significant for good treatment, prediction, and existence. Despite the current growth in the perception of molecular devices, late analysis and methods near precision medicine for OC patients remain a challenge. A machine learning (ML) model was employed to improve early detection in medicine, aiming to reduce cancer-specific mortality and disease progression. Recent advancements in this approach have significantly enhanced the extraction and diagnosis of critical information from medical images. This paper presents a Deep Structured Learning with Vision Intelligence for Oral Carcinoma Lesion Segmentation and Classification (DSLVI-OCLSC) model for medical imaging. Using medical imaging, the DSLVI-OCLSC model aims to enhance OC's classification and recognition outcomes. To accomplish this, the DSLVI-OCLSC model utilizes wiener filtering (WF) as a pre-processing technique to eliminate the noise. In addition, the ShuffleNetV2 method is used for the group of higher-level deep features from an input image. The convolutional bidirectional long short-term memory network with a multi-head attention mechanism (MA-CNN-BiLSTM) approach is utilized for oral carcinoma recognition and identification. Moreover, the Unet3 + is employed to segment abnormal regions from the classified images. Finally, the sine cosine algorithm (SCA) approach is utilized to hyperparameter-tune the DL model. A wide range of simulations is implemented to ensure the enhanced performance of the DSLVI-OCLSC method under the OC images dataset. The experimental analysis of the DSLVI-OCLSC method portrayed a superior accuracy value of 98.47% over recent approaches.
Collapse
Affiliation(s)
- Ahmad A Alzahrani
- Department of Computer Science and Artificial Intelligence, College of Computing, Umm-AlQura University, Mecca, Saudi Arabia
| | - Jamal Alsamri
- Department of Biomedical Engineering, College of Engineering, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, PO Box 103786, 11543, Riyadh, Saudi Arabia
| | - Noha Negm
- Department of Computer Science, Applied College at Mahayil, King Khalid University, Abha, Saudi Arabia
| | - Somia A Asklany
- Department of Computer Science and Information Technology, Faculty of Sciences and Arts, Turaif, Northern Border University, 91431, Arar, Saudi Arabia.
| | - Abdulwhab Alkharashi
- Department of Computer Science, College of Computing and Informatics, Saudi Electronic University, Riyadh, Saudi Arabia
| | - Hassan Alkhiri
- Department of Computer Science, Faculty of Computing and Information Technology, Al-Baha University, Al-Baha, Saudi Arabia
| | - Marwa Obayya
- Department of Computer Science and Artificial Intelligence, College of Computing, Umm-AlQura University, Mecca, Saudi Arabia
| |
Collapse
|
13
|
Yadav DP, Sharma B, Noonia A, Mehbodniya A. Explainable label guided lightweight network with axial transformer encoder for early detection of oral cancer. Sci Rep 2025; 15:6391. [PMID: 39984521 PMCID: PMC11845714 DOI: 10.1038/s41598-025-87627-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2024] [Accepted: 01/21/2025] [Indexed: 02/23/2025] Open
Abstract
Oral cavity cancer exhibits high morbidity and mortality rates. Therefore, it is essential to diagnose the disease at an early stage. Machine learning and convolution neural networks (CNN) are powerful tools for diagnosing mouth and oral cancer. In this study, we design a lightweight explainable network (LWENet) with label-guided attention (LGA) to provide a second opinion to the expert. The LWENet contains depth-wise separable convolution layers to reduce the computation costs. Moreover, the LGA module provides label consistency to the neighbor pixel and improves the spatial features. Furthermore, AMSA (axial multi-head self-attention) based ViT encoder incorporated in the model to provide global attention. Our ViT (vision transformer) encoder is computationally efficient compared to the classical ViT encoder. We tested LWRNet performance on the MOD (mouth and oral disease) and OCI (oral cancer image) datasets, and results are compared with the other CNN and ViT (vision transformer) based methods. The LWENet achieved a precision and F1-scores of 96.97% and 98.90% on the MOD dataset, and 99.48% and 98.23% on the OCI dataset, respectively. By incorporating Grad-CAM, we visualize the decision-making process, enhancing model interpretability. This work demonstrates the potential of LWENet with LGA in facilitating early oral cancer detection.
Collapse
Affiliation(s)
- Dhirendra Prasad Yadav
- Department of Computer Engineering & Applications, GLA University Mathura, Mathura, India
| | - Bhisham Sharma
- Centre for Research Impact & Outcome, Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab, 140401, India.
| | - Ajit Noonia
- Department of Computer Science and Engineering, Manipal University Jaipur, Jaipur, Rajasthan, India
| | - Abolfazl Mehbodniya
- Department of Electronics and Communication Engineering, Kuwait College of Science and Technology (KCST), Doha Area, 7th Ring Road, Kuwait City, Kuwait
| |
Collapse
|
14
|
Ozdemir B, Pacal I. A robust deep learning framework for multiclass skin cancer classification. Sci Rep 2025; 15:4938. [PMID: 39930026 PMCID: PMC11811178 DOI: 10.1038/s41598-025-89230-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2024] [Accepted: 02/04/2025] [Indexed: 02/13/2025] Open
Abstract
Skin cancer represents a significant global health concern, where early and precise diagnosis plays a pivotal role in improving treatment efficacy and patient survival rates. Nonetheless, the inherent visual similarities between benign and malignant lesions pose substantial challenges to accurate classification. To overcome these obstacles, this study proposes an innovative hybrid deep learning model that combines ConvNeXtV2 blocks and separable self-attention mechanisms, tailored to enhance feature extraction and optimize classification performance. The inclusion of ConvNeXtV2 blocks in the initial two stages is driven by their ability to effectively capture fine-grained local features and subtle patterns, which are critical for distinguishing between visually similar lesion types. Meanwhile, the adoption of separable self-attention in the later stages allows the model to selectively prioritize diagnostically relevant regions while minimizing computational complexity, addressing the inefficiencies often associated with traditional self-attention mechanisms. The model was comprehensively trained and validated on the ISIC 2019 dataset, which includes eight distinct skin lesion categories. Advanced methodologies such as data augmentation and transfer learning were employed to further enhance model robustness and reliability. The proposed architecture achieved exceptional performance metrics, with 93.48% accuracy, 93.24% precision, 90.70% recall, and a 91.82% F1-score, outperforming over ten Convolutional Neural Network (CNN) based and over ten Vision Transformer (ViT) based models tested under comparable conditions. Despite its robust performance, the model maintains a compact design with only 21.92 million parameters, making it highly efficient and suitable for model deployment. The Proposed Model demonstrates exceptional accuracy and generalizability across diverse skin lesion classes, establishing a reliable framework for early and accurate skin cancer diagnosis in clinical practice.
Collapse
Affiliation(s)
- Burhanettin Ozdemir
- Department of Operations and Project Management, College of Business, Alfaisal University, Riyadh, 11533, Saudi Arabia.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, Igdir, 76000, Turkey
- Department of Electronics and Information Technologies, Faculty of Architecture and Engineering, Nakhchivan State University, AZ 7012, Nakhchivan, Azerbaijan
| |
Collapse
|
15
|
Alotaibi SR, Alohali MA, Maashi M, Alqahtani H, Alotaibi M, Mahmud A. Advances in colorectal cancer diagnosis using optimal deep feature fusion approach on biomedical images. Sci Rep 2025; 15:4200. [PMID: 39905104 PMCID: PMC11794880 DOI: 10.1038/s41598-024-83466-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Accepted: 12/16/2024] [Indexed: 02/06/2025] Open
Abstract
Colorectal cancer (CRC) is the second popular cancer in females and third in males, with an increased number of cases. Pathology diagnoses complemented with predictive and prognostic biomarker information is the first step for personalized treatment. Histopathological image (HI) analysis is the benchmark for pathologists to rank colorectal cancer of various kinds. However, pathologists' diagnoses are highly subjective and susceptible to inaccurate diagnoses. The improved diagnosis load in the pathology laboratory, incorporated with the reported intra- and inter-variability in the biomarker assessment, has prompted the quest for consistent machine-based techniques to be integrated into routine practice. In the healthcare field, artificial intelligence (AI) has achieved extraordinary achievements in healthcare applications. Lately, computer-aided diagnosis (CAD) based on HI has progressed rapidly with the increase of machine learning (ML) and deep learning (DL) based models. This study introduces a novel Colorectal Cancer Diagnosis using the Optimal Deep Feature Fusion Approach on Biomedical Images (CCD-ODFFBI) method. The primary objective of the CCD-ODFFBI technique is to examine the biomedical images to identify colorectal cancer (CRC). In the CCD-ODFFBI technique, the median filtering (MF) approach is initially utilized for noise elimination. The CCD-ODFFBI technique utilizes a fusion of three DL models, MobileNet, SqueezeNet, and SE-ResNet, for feature extraction. Moreover, the DL models' hyperparameter selection is performed using the Osprey optimization algorithm (OOA). Finally, the deep belief network (DBN) model is employed to classify CRC. A series of simulations is accomplished to highlight the significant results of the CCD-ODFFBI method under the Warwick-QU dataset. The comparison of the CCD-ODFFBI method showed a superior accuracy value of 99.39% over existing techniques.
Collapse
Affiliation(s)
- Sultan Refa Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Manal Abdullah Alohali
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh, 11671, Saudi Arabia.
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, PO Box 103786, Riyadh, 11543, Saudi Arabia
| | - Hamed Alqahtani
- Department of Information Systems, College of Computer Science, Center of Artificial Intelligence, Unit of Cybersecurity, King Khalid University, Abha, Saudi Arabia
| | - Moneerah Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Ahmed Mahmud
- Research Center, Future University in Egypt, New Cairo, 11835, Egypt
| |
Collapse
|
16
|
Pacal I, Işık G. Utilizing convolutional neural networks and vision transformers for precise corn leaf disease identification. Neural Comput Appl 2025; 37:2479-2496. [DOI: 10.1007/s00521-024-10769-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2024] [Accepted: 11/05/2024] [Indexed: 05/14/2025]
|
17
|
Kaur H, Sharma R, Kaur J. Comparison of deep transfer learning models for classification of cervical cancer from pap smear images. Sci Rep 2025; 15:3945. [PMID: 39890842 PMCID: PMC11785805 DOI: 10.1038/s41598-024-74531-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Accepted: 09/26/2024] [Indexed: 02/03/2025] Open
Abstract
Cervical cancer is one of the most commonly diagnosed cancers worldwide, and it is particularly prevalent among women living in developing countries. Traditional classification algorithms often require segmentation and feature extraction techniques to detect cervical cancer. In contrast, convolutional neural networks (CNN) models require large datasets to reduce overfitting and poor generalization. Based on limited datasets, transfer learning was applied directly to pap smear images to perform a classification task. A comprehensive comparison of 16 pre-trained models (VGG16, VGG19, ResNet50, ResNet50V2, ResNet101, ResNet101V2, ResNet152, ResNet152V2, DenseNet121, DenseNet169, DenseNet201, MobileNet, XceptionNet, InceptionV3, and InceptionResNetV2) were carried out for cervical cancer classification by relying on the Herlev dataset and Sipakmed dataset. A comparison of the results revealed that ResNet50 achieved 95% accuracy both for 2-class classification and for 7-class classification using the Herlev dataset. Based on the Sipakmed dataset, VGG16 obtained an accuracy of 99.95% for 2-class and 5-class classification, DenseNet121 achieved an accuracy of 97.65% for 3-class classification. Our findings indicate that DTL models are suitable for automating cervical cancer screening, providing more accurate and efficient results than manual screening.
Collapse
Affiliation(s)
- Harmanpreet Kaur
- Department of Computer Science & Engineering, Punjabi University, Patiala, India.
| | - Reecha Sharma
- Department of Electronics and Communication Engineering, Punjabi University, Patiala, India
| | - Jagroop Kaur
- Department of Computer Science & Engineering, Punjabi University, Patiala, India
| |
Collapse
|
18
|
Ozdemir B, Aslan E, Pacal I. Attention Enhanced InceptionNeXt-Based Hybrid Deep Learning Model for Lung Cancer Detection. IEEE ACCESS 2025; 13:27050-27069. [DOI: 10.1109/access.2025.3539122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
Affiliation(s)
- Burhanettin Ozdemir
- Department of Operations and Project Management, College of Business, Alfaisal University, Riyadh, Saudi Arabia
| | - Emrah Aslan
- Department of Computer Engineering, Faculty of Engineering and Architecture, Mardin Artuklu University, Mardin, Türkiye
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, Iğdır, Türkiye
| |
Collapse
|
19
|
Voon W, Hum YC, Tee YK, Yap WS, Lai KW, Nisar H, Mokayed H. IMAML-IDCG: Optimization-based meta-learning with ImageNet feature reusing for few-shot invasive ductal carcinoma grading. EXPERT SYSTEMS WITH APPLICATIONS 2024; 257:124969. [DOI: 10.1016/j.eswa.2024.124969] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2025]
|
20
|
Pacal I, Alaftekin M, Zengul FD. Enhancing Skin Cancer Diagnosis Using Swin Transformer with Hybrid Shifted Window-Based Multi-head Self-attention and SwiGLU-Based MLP. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:3174-3192. [PMID: 38839675 PMCID: PMC11612041 DOI: 10.1007/s10278-024-01140-8] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Revised: 05/07/2024] [Accepted: 05/08/2024] [Indexed: 06/07/2024]
Abstract
Skin cancer is one of the most frequently occurring cancers worldwide, and early detection is crucial for effective treatment. Dermatologists often face challenges such as heavy data demands, potential human errors, and strict time limits, which can negatively affect diagnostic outcomes. Deep learning-based diagnostic systems offer quick, accurate testing and enhanced research capabilities, providing significant support to dermatologists. In this study, we enhanced the Swin Transformer architecture by implementing the hybrid shifted window-based multi-head self-attention (HSW-MSA) in place of the conventional shifted window-based multi-head self-attention (SW-MSA). This adjustment enables the model to more efficiently process areas of skin cancer overlap, capture finer details, and manage long-range dependencies, while maintaining memory usage and computational efficiency during training. Additionally, the study replaces the standard multi-layer perceptron (MLP) in the Swin Transformer with a SwiGLU-based MLP, an upgraded version of the gated linear unit (GLU) module, to achieve higher accuracy, faster training speeds, and better parameter efficiency. The modified Swin model-base was evaluated using the publicly accessible ISIC 2019 skin dataset with eight classes and was compared against popular convolutional neural networks (CNNs) and cutting-edge vision transformer (ViT) models. In an exhaustive assessment on the unseen test dataset, the proposed Swin-Base model demonstrated exceptional performance, achieving an accuracy of 89.36%, a recall of 85.13%, a precision of 88.22%, and an F1-score of 86.65%, surpassing all previously reported research and deep learning models documented in the literature.
Collapse
Affiliation(s)
- Ishak Pacal
- Department of Computer Engineering, Igdir University, 76000, Igdir, Turkey
| | - Melek Alaftekin
- Department of Computer Engineering, Igdir University, 76000, Igdir, Turkey
| | - Ferhat Devrim Zengul
- Department of Health Services Administration, The University of Alabama at Birmingham, Birmingham, AL, USA.
- Center for Integrated System, School of Engineering, The University of Alabama at Birmingham, Birmingham, AL, USA.
- Department of Biomedical Informatics and Data Science, School of Medicine, The University of Alabama, Birmingham, USA.
| |
Collapse
|
21
|
Mehedi MHK, Khandaker M, Ara S, Alam MA, Mridha MF, Aung Z. A lightweight deep learning method to identify different types of cervical cancer. Sci Rep 2024; 14:29446. [PMID: 39604499 PMCID: PMC11603366 DOI: 10.1038/s41598-024-79840-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2024] [Accepted: 11/12/2024] [Indexed: 11/29/2024] Open
Abstract
Cervical cancer is the second most common cancer in women's bodies after breast cancer. Cervical cancer develops from dysplasia or cervical intraepithelial neoplasm (CIN), the early stage of the disease, and is characterized by the aberrant growth of cells in the cervix lining. It is primarily caused by Human Papillomavirus (HPV) infection, which spreads through sexual activity. This study focuses on detecting cervical cancer types efficiently using a novel lightweight deep learning model named CCanNet, which combines squeeze block, residual blocks, and skip layer connections. SipakMed, which is not only popular but also publicly available dataset, was used in this study. We conducted a comparative analysis between several transfer learning and transformer models such as VGG19, VGG16, MobileNetV2, AlexNet, ConvNeXT, DeiT_tiny, MobileViT, and Swin Transformer with the proposed CCanNet. Our proposed model outperformed other state-of-the-art models, with 98.53% accuracy and the lowest number of parameters, which is 1,274,663. In addition, accuracy, precision, recall, and the F1 score were used to evaluate the performance of the models. Finally, explainable AI (XAI) was applied to analyze the performance of CCanNet and ensure the results were trustworthy.
Collapse
Affiliation(s)
| | - Moumita Khandaker
- Department of Computer Science and Engineering, BRAC University, Dhaka, Bangladesh
| | - Shaneen Ara
- Department of Computer Science and Engineering, Bangladesh University of Business and Technology, Dhaka, Bangladesh
| | - Md Ashraful Alam
- Department of Computer Science and Engineering, BRAC University, Dhaka, Bangladesh
| | - M F Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka, Bangladesh.
| | - Zeyar Aung
- Department of Computer Science, Khalifa University of Science and Technology, Abu Dhabi, UAE.
| |
Collapse
|
22
|
Xu H, Zhang X, Shen W, Lin Z, Liu S, Jia Q, Li H, Zheng J, Zhong F. Improved CSW-YOLO Model for Bitter Melon Phenotype Detection. PLANTS (BASEL, SWITZERLAND) 2024; 13:3329. [PMID: 39683122 DOI: 10.3390/plants13233329] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/28/2024] [Revised: 11/25/2024] [Accepted: 11/26/2024] [Indexed: 12/18/2024]
Abstract
As a crop with significant medicinal value and nutritional components, the market demand for bitter melon continues to grow. The diversity of bitter melon shapes has a direct impact on its market acceptance and consumer preferences, making precise identification of bitter melon germplasm resources crucial for breeding work. To address the limitations of time-consuming and less accurate traditional manual identification methods, there is a need to enhance the automation and intelligence of bitter melon phenotype detection. This study developed a bitter melon phenotype detection model named CSW-YOLO. By incorporating the ConvNeXt V2 module to replace the backbone network of YOLOv8, the model's focus on critical target features is enhanced. Additionally, the SimAM attention mechanism was introduced to compute attention weights for neurons without increasing the parameter count, further enhancing the model's recognition accuracy. Finally, WIoUv3 was introduced as the bounding box loss function to improve the model's convergence speed and positioning capabilities. The model was trained and tested on a bitter melon image dataset, achieving a precision of 94.6%, a recall of 80.6%, a mAP50 of 96.7%, and an F1 score of 87.04%. These results represent improvements of 8.5%, 0.4%, 11.1%, and 4% in precision, recall, mAP50, and F1 score, respectively, over the original YOLOv8 model. Furthermore, the effectiveness of the improvements was validated through heatmap analysis and ablation experiments, demonstrating that the CSW-YOLO model can more accurately focus on target features, reduce false detection rates, and enhance generalization capabilities. Comparative tests with various mainstream deep learning models also proved the superior performance of CSW-YOLO in bitter melon phenotype detection tasks. This research provides an accurate and reliable method for bitter melon phenotype identification and also offers technical support for the visual detection technologies of other agricultural products.
Collapse
Affiliation(s)
- Haobin Xu
- College of Horticulture, Fujian Agriculture and Forestry University, Fuzhou 350002, China
| | - Xianhua Zhang
- College of Horticulture, Fujian Agriculture and Forestry University, Fuzhou 350002, China
| | - Weilin Shen
- College of Horticulture, Fujian Agriculture and Forestry University, Fuzhou 350002, China
| | - Zhiqiang Lin
- Fujian Agricultural Machinery Extension Station, Fuzhou 350002, China
| | - Shuang Liu
- College of Horticulture, Fujian Agriculture and Forestry University, Fuzhou 350002, China
| | - Qi Jia
- Jiuquan Academy of Agriculture Sciences, Jiuquan 735099, China
| | - Honglong Li
- Fujian Tianmei Seed Industry Technology Co., Fuzhou 350109, China
| | - Jingyuan Zheng
- Institute of Vegetables, Hunan Academy of Agricultural Sciences, Changsha 410125, China
| | - Fenglin Zhong
- College of Horticulture, Fujian Agriculture and Forestry University, Fuzhou 350002, China
| |
Collapse
|
23
|
Maman A, Pacal I, Bati F. Can deep learning effectively diagnose cardiac amyloidosis with 99mTc-PYP scintigraphy? J Radioanal Nucl Chem 2024. [DOI: 10.1007/s10967-024-09879-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Accepted: 11/07/2024] [Indexed: 05/14/2025]
|
24
|
Muksimova S, Umirzakova S, Shoraimov K, Baltayev J, Cho YI. Novelty Classification Model Use in Reinforcement Learning for Cervical Cancer. Cancers (Basel) 2024; 16:3782. [PMID: 39594737 PMCID: PMC11592902 DOI: 10.3390/cancers16223782] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2024] [Revised: 10/29/2024] [Accepted: 11/06/2024] [Indexed: 11/28/2024] Open
Abstract
PURPOSE Cervical cancer significantly impacts global health, where early detection is piv- otal for improving patient outcomes. This study aims to enhance the accuracy of cervical cancer diagnosis by addressing class imbalance through a novel hybrid deep learning model. METHODS The proposed model, RL-CancerNet, integrates EfficientNetV2 and Vision Transformers (ViTs) within a Reinforcement Learning (RL) framework. EfficientNetV2 extracts local features from cervical cytology images to capture fine-grained details, while ViTs analyze these features to recognize global dependencies across image patches. To address class imbalance, an RL agent dynamically adjusts the focus towards minority classes, thus reducing the common bias towards majority classes in medical image classification. Additionally, a Supporter Module incorporating Conv3D and BiLSTM layers with an attention mechanism enhances contextual learning. RESULTS RL-CancerNet was evaluated on the benchmark cervical cytology datasets Herlev and SipaKMeD, achieving an exceptional accuracy of 99.7%. This performance surpasses several state-of-the-art models, demonstrating the model's effectiveness in identifying subtle diagnostic features in complex backgrounds. CONCLUSIONS The integration of CNNs, ViTs, and RL into RL-CancerNet significantly improves the diagnostic accuracy of cervical cancer screenings. This model not only advances the field of automated medical screening but also provides a scalable framework adaptable to other medical imaging tasks, potentially enhancing diagnostic processes across various medical domains.
Collapse
Affiliation(s)
- Shakhnoza Muksimova
- Department of Computer Engineering, Gachon University, Sujeong-gu, Seongnam-si 461-701, Republic of Korea;
| | - Sabina Umirzakova
- Department of Computer Engineering, Gachon University, Sujeong-gu, Seongnam-si 461-701, Republic of Korea;
| | - Khusanboy Shoraimov
- Department of Systematic and Practical Programming, Tashkent University of Information Technologies Named After Muhammad Al-Khwarizmi, Tashkent 100200, Uzbekistan;
| | - Jushkin Baltayev
- Department of Information Systems and Technologies, Tashkent State University of Economic, Tashkent 100066, Uzbekistan;
| | - Young-Im Cho
- Department of Computer Engineering, Gachon University, Sujeong-gu, Seongnam-si 461-701, Republic of Korea;
| |
Collapse
|
25
|
Pacal I, Celik O, Bayram B, Cunha A. Enhancing EfficientNetv2 with global and efficient channel attention mechanisms for accurate MRI-Based brain tumor classification. CLUSTER COMPUTING 2024; 27:11187-11212. [DOI: 10.1007/s10586-024-04532-1] [Citation(s) in RCA: 15] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Revised: 04/08/2024] [Accepted: 04/22/2024] [Indexed: 05/14/2025]
Abstract
AbstractThe early and accurate diagnosis of brain tumors is critical for effective treatment planning, with Magnetic Resonance Imaging (MRI) serving as a key tool in the non-invasive examination of such conditions. Despite the advancements in Computer-Aided Diagnosis (CADx) systems powered by deep learning, the challenge of accurately classifying brain tumors from MRI scans persists due to the high variability of tumor appearances and the subtlety of early-stage manifestations. This work introduces a novel adaptation of the EfficientNetv2 architecture, enhanced with Global Attention Mechanism (GAM) and Efficient Channel Attention (ECA), aimed at overcoming these hurdles. This enhancement not only amplifies the model’s ability to focus on salient features within complex MRI images but also significantly improves the classification accuracy of brain tumors. Our approach distinguishes itself by meticulously integrating attention mechanisms that systematically enhance feature extraction, thereby achieving superior performance in detecting a broad spectrum of brain tumors. Demonstrated through extensive experiments on a large public dataset, our model achieves an exceptional high-test accuracy of 99.76%, setting a new benchmark in MRI-based brain tumor classification. Moreover, the incorporation of Grad-CAM visualization techniques sheds light on the model’s decision-making process, offering transparent and interpretable insights that are invaluable for clinical assessment. By addressing the limitations inherent in previous models, this study not only advances the field of medical imaging analysis but also highlights the pivotal role of attention mechanisms in enhancing the interpretability and accuracy of deep learning models for brain tumor diagnosis. This research sets the stage for advanced CADx systems, enhancing patient care and treatment outcomes.
Collapse
|
26
|
Bhuvaneshwari KV, Lahza H, Sreenivasa BR, Lahza HFM, Shawly T, Poornima B. Optimising ovarian tumor classification using a novel CT sequence selection algorithm. Sci Rep 2024; 14:25010. [PMID: 39443517 PMCID: PMC11500337 DOI: 10.1038/s41598-024-75555-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2023] [Accepted: 10/07/2024] [Indexed: 10/25/2024] Open
Abstract
Gynaecological cancers, especially ovarian cancer, remain a critical public health issue, particularly in regions like India, where there are challenges related to cancer awareness, variable pathology, and limited access to screening facilities. These challenges often lead to the diagnosis of cancer at advanced stages, resulting in poorer outcomes for patients. The goal of this study is to enhance the accuracy of classifying ovarian tumours, with a focus on distinguishing between malignant and early-stage cases, by applying advanced deep learning methods. In our approach, we utilized three pre-trained deep learning models-Xception, ResNet50V2, and ResNet50V2FPN-to classify ovarian tumors using publicly available Computed Tomography (CT) scan data. To further improve the model's performance, we developed a novel CT Sequence Selection Algorithm, which optimises the use of CT images for a more precise classification of ovarian tumours. The models were trained and evaluated on selected TIFF images, comparing the performance of the ResNet50V2FPN model with and without the CT Sequence Selection Algorithm. Our experimental results show the Comparative evaluation against the ResNet50V2 FPN model, both with and without the CT Sequence Selection Algorithm, demonstrates the superiority of the proposed algorithm over existing state-of-the-art methods. This research presents a promising approach for improving the early detection and management of gynecological cancers, with potential benefits for patient outcomes, especially in areas with limited healthcare resources.
Collapse
Affiliation(s)
- K V Bhuvaneshwari
- Department of Information Science & Engineering, Bapuji Institute of Engineering & Technology, Davanagere, Karnataka, India
| | - Husam Lahza
- Department of Information Technology, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| | - B R Sreenivasa
- Department of Computer Science & Design, Bapuji Institute of Engineering & Technology, Davanagere, Karnataka, India.
| | - Hassan Fareed M Lahza
- Cybersecurity Department, Faculty of Computing, Umm Al-Qura University, Makkah, Saudi Arabia
| | - Tawfeeq Shawly
- Department of Electrical Engineering, Faculty of Engineering at Rabigh, King Abdulaziz University, Jeddah, Saudi Arabia
| | - B Poornima
- Department of Information Science & Engineering, Bapuji Institute of Engineering & Technology, Davanagere, Karnataka, India
| |
Collapse
|
27
|
Kumar Y, Shrivastav S, Garg K, Modi N, Wiltos K, Woźniak M, Ijaz MF. Automating cancer diagnosis using advanced deep learning techniques for multi-cancer image classification. Sci Rep 2024; 14:25006. [PMID: 39443621 PMCID: PMC11499884 DOI: 10.1038/s41598-024-75876-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2024] [Accepted: 10/08/2024] [Indexed: 10/25/2024] Open
Abstract
Cancer detection poses a significant challenge for researchers and clinical experts due to its status as the leading cause of global mortality. Early detection is crucial, but traditional cancer detection methods often rely on invasive procedures and time-consuming analyses, creating a demand for more efficient and accurate solutions. This paper addresses these challenges by utilizing automated cancer detection through AI-based techniques, specifically focusing on deep learning models. Convolutional Neural Networks (CNNs), including DenseNet121, DenseNet201, Xception, InceptionV3, MobileNetV2, NASNetLarge, NASNetMobile, InceptionResNetV2, VGG19, and ResNet152V2, are evaluated on image datasets for seven types of cancer: brain, oral, breast, kidney, Acute Lymphocytic Leukemia, lung and colon, and cervical cancer. Initially, images undergo segmentation techniques, proceeded by contour feature extraction where parameters such as perimeter, area, and epsilon are computed. The models are rigorously evaluated, with DenseNet121 achieving the highest validation accuracy as 99.94%, 0.0017 as loss, and the lowest Root Mean Square Error (RMSE) values as 0.036056 for training and 0.045826 for validation. These results revealed the capability of AI-based techniques in improving cancer detection accuracy, with DenseNet121 emerging as the most effective model in this study.
Collapse
Affiliation(s)
- Yogesh Kumar
- Department of Computer Science and Engineering, School of Technology, PDEU, Gandhinagar, Gujarat, 382426, India
| | | | - Kinny Garg
- Department of ECE, AMC Engineering College, Bengaluru, Karnataka, India
| | - Nandini Modi
- Department of Computer Science and Engineering, School of Technology, PDEU, Gandhinagar, Gujarat, 382426, India.
| | - Katarzyna Wiltos
- Faculty of Applied Mathematics, Silesian University of Technology, Kaszubska 23, Gliwice, 44100, Poland
| | - Marcin Woźniak
- Faculty of Applied Mathematics, Silesian University of Technology, Kaszubska 23, Gliwice, 44100, Poland.
| | - Muhammad Fazal Ijaz
- School of IT and Engineering, Melbourne Institute of Technology, Melbourne, 3000, Australia.
| |
Collapse
|
28
|
Lubbad MAH, Kurtulus IL, Karaboga D, Kilic K, Basturk A, Akay B, Nalbantoglu OU, Yilmaz OMD, Ayata M, Yilmaz S, Pacal I. A Comparative Analysis of Deep Learning-Based Approaches for Classifying Dental Implants Decision Support System. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2559-2580. [PMID: 38565730 PMCID: PMC11522249 DOI: 10.1007/s10278-024-01086-x] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/20/2024] [Revised: 02/28/2024] [Accepted: 03/12/2024] [Indexed: 04/04/2024]
Abstract
This study aims to provide an effective solution for the autonomous identification of dental implant brands through a deep learning-based computer diagnostic system. It also seeks to ascertain the system's potential in clinical practices and to offer a strategic framework for improving diagnosis and treatment processes in implantology. This study employed a total of 28 different deep learning models, including 18 convolutional neural network (CNN) models (VGG, ResNet, DenseNet, EfficientNet, RegNet, ConvNeXt) and 10 vision transformer models (Swin and Vision Transformer). The dataset comprises 1258 panoramic radiographs from patients who received implant treatments at Erciyes University Faculty of Dentistry between 2012 and 2023. It is utilized for the training and evaluation process of deep learning models and consists of prototypes from six different implant systems provided by six manufacturers. The deep learning-based dental implant system provided high classification accuracy for different dental implant brands using deep learning models. Furthermore, among all the architectures evaluated, the small model of the ConvNeXt architecture achieved an impressive accuracy rate of 94.2%, demonstrating a high level of classification success.This study emphasizes the effectiveness of deep learning-based systems in achieving high classification accuracy in dental implant types. These findings pave the way for integrating advanced deep learning tools into clinical practice, promising significant improvements in patient care and treatment outcomes.
Collapse
Affiliation(s)
- Mohammed A H Lubbad
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey.
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey.
| | | | - Dervis Karaboga
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | - Kerem Kilic
- Department of Prosthodontics, Dentistry Faculty, Erciyes University, Kayseri, Turkey
| | - Alper Basturk
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | - Bahriye Akay
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | - Ozkan Ufuk Nalbantoglu
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | | | - Mustafa Ayata
- Department of Prosthodontics, Dentistry Faculty, Erciyes University, Kayseri, Turkey
| | - Serkan Yilmaz
- Department of Dentomaxillofacial Radiology, Dentistry Faculty, Erciyes University, Kayseri, Turkey
| | - Ishak Pacal
- Department of Computer Engineering, Engineering Faculty, Igdir University, Igdir, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| |
Collapse
|
29
|
Attallah O. Skin cancer classification leveraging multi-directional compact convolutional neural network ensembles and gabor wavelets. Sci Rep 2024; 14:20637. [PMID: 39232043 PMCID: PMC11375051 DOI: 10.1038/s41598-024-69954-8] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2024] [Accepted: 08/12/2024] [Indexed: 09/06/2024] Open
Abstract
Skin cancer (SC) is an important medical condition that necessitates prompt identification to ensure timely treatment. Although visual evaluation by dermatologists is considered the most reliable method, its efficacy is subjective and laborious. Deep learning-based computer-aided diagnostic (CAD) platforms have become valuable tools for supporting dermatologists. Nevertheless, current CAD tools frequently depend on Convolutional Neural Networks (CNNs) with huge amounts of deep layers and hyperparameters, single CNN model methodologies, large feature space, and exclusively utilise spatial image information, which restricts their effectiveness. This study presents SCaLiNG, an innovative CAD tool specifically developed to address and surpass these constraints. SCaLiNG leverages a collection of three compact CNNs and Gabor Wavelets (GW) to acquire a comprehensive feature vector consisting of spatial-textural-frequency attributes. SCaLiNG gathers a wide range of image details by breaking down these photos into multiple directional sub-bands using GW, and then learning several CNNs using those sub-bands and the original picture. SCaLiNG also combines attributes taken from various CNNs trained with the actual images and subbands derived from GW. This fusion process correspondingly improves diagnostic accuracy due to the thorough representation of attributes. Furthermore, SCaLiNG applies a feature selection approach which further enhances the model's performance by choosing the most distinguishing features. Experimental findings indicate that SCaLiNG maintains a classification accuracy of 0.9170 in categorising SC subcategories, surpassing conventional single-CNN models. The outstanding performance of SCaLiNG underlines its ability to aid dermatologists in swiftly and precisely recognising and classifying SC, thereby enhancing patient outcomes.
Collapse
Affiliation(s)
- Omneya Attallah
- Department of Electronics and Communications Engineering, College of Engineering and Technology, Arab Academy for Science, Technology and Maritime Transport, Alexandria, 21937, Egypt.
- Wearables, Biosensing, and Biosignal Processing Laboratory, Arab Academy for Science, Technology, and Maritime Transport, Alexandria, 21937, Egypt.
| |
Collapse
|
30
|
Alotaibi M, Alshardan A, Maashi M, Asiri MM, Alotaibi SR, Yafoz A, Alsini R, Khadidos AO. Exploiting histopathological imaging for early detection of lung and colon cancer via ensemble deep learning model. Sci Rep 2024; 14:20434. [PMID: 39227664 PMCID: PMC11372073 DOI: 10.1038/s41598-024-71302-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2024] [Accepted: 08/27/2024] [Indexed: 09/05/2024] Open
Abstract
Cancer seems to have a vast number of deaths due to its heterogeneity, aggressiveness, and significant propensity for metastasis. The predominant categories of cancer that may affect males and females and occur worldwide are colon and lung cancer. A precise and on-time analysis of this cancer can increase the survival rate and improve the appropriate treatment characteristics. An efficient and effective method for the speedy and accurate recognition of tumours in the colon and lung areas is provided as an alternative to cancer recognition methods. Earlier diagnosis of the disease on the front drastically reduces the chance of death. Machine learning (ML) and deep learning (DL) approaches can accelerate this cancer diagnosis, facilitating researcher workers to study a vast majority of patients in a limited period and at a low cost. This research presents Histopathological Imaging for the Early Detection of Lung and Colon Cancer via Ensemble DL (HIELCC-EDL) model. The HIELCC-EDL technique utilizes histopathological images to identify lung and colon cancer (LCC). To achieve this, the HIELCC-EDL technique uses the Wiener filtering (WF) method for noise elimination. In addition, the HIELCC-EDL model uses the channel attention Residual Network (CA-ResNet50) model for learning complex feature patterns. Moreover, the hyperparameter selection of the CA-ResNet50 model is performed using the tuna swarm optimization (TSO) technique. Finally, the detection of LCC is achieved by using the ensemble of three classifiers such as extreme learning machine (ELM), competitive neural networks (CNNs), and long short-term memory (LSTM). To illustrate the promising performance of the HIELCC-EDL model, a complete set of experimentations was performed on a benchmark dataset. The experimental validation of the HIELCC-EDL model portrayed a superior accuracy value of 99.60% over recent approaches.
Collapse
Affiliation(s)
- Moneerah Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Amal Alshardan
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, 11671, Riyadh, Saudi Arabia
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 103786, 11543, Riyadh, Saudi Arabia
| | - Mashael M Asiri
- Department of Computer Science, Applied College at Mahayil, King Khalid University, Abha, Saudi Arabia.
| | - Sultan Refa Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Ayman Yafoz
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Raed Alsini
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Alaa O Khadidos
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| |
Collapse
|
31
|
Zhao Z, Hu B, Xu K, Jiang Y, Xu X, Liu Y. A quantitative analysis of artificial intelligence research in cervical cancer: a bibliometric approach utilizing CiteSpace and VOSviewer. Front Oncol 2024; 14:1431142. [PMID: 39296978 PMCID: PMC11408476 DOI: 10.3389/fonc.2024.1431142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2024] [Accepted: 08/16/2024] [Indexed: 09/21/2024] Open
Abstract
Background Cervical cancer, a severe threat to women's health, is experiencing a global increase in incidence, notably among younger demographics. With artificial intelligence (AI) making strides, its integration into medical research is expanding, particularly in cervical cancer studies. This bibliometric study aims to evaluate AI's role, highlighting research trends and potential future directions in the field. Methods This study systematically retrieved literature from the Web of Science Core Collection (WoSCC), employing VOSviewer and CiteSpace for analysis. This included examining collaborations and keyword co-occurrences, with a focus on the relationship between citing and cited journals and authors. A burst ranking analysis identified research hotspots based on citation frequency. Results The study analyzed 927 articles from 2008 to 2024 by 5,299 authors across 81 regions. China, the U.S., and India were the top contributors, with key institutions like the Chinese Academy of Sciences and the NIH leading in publications. Schiffman, Mark, featured among the top authors, while Jemal, A, was the most cited. 'Diagnostics' and 'IEEE Access' stood out for publication volume and citation impact, respectively. Keywords such as 'cervical cancer,' 'deep learning,' 'classification,' and 'machine learning' were dominant. The most cited article was by Berner, ES; et al., published in 2008. Conclusions AI's application in cervical cancer research is expanding, with a growing scholarly community. The study suggests that AI, especially deep learning and machine learning, will remain a key research area, focusing on improving diagnostics and treatment. There is a need for increased international collaboration to maximize AI's potential in advancing cervical cancer research and patient care.
Collapse
Affiliation(s)
- Ziqi Zhao
- School of Basic Medicine, Zhejiang Chinese Medical University, Hangzhou, China
| | - Boqian Hu
- Hebei Provincial Hospital of Traditional Chinese Medicine, Hebei University of Chinese Medicine, Shijiazhuang, Hebei, China
| | - Kun Xu
- School of Basic Medicine, Zhejiang Chinese Medical University, Hangzhou, China
| | - Yizhuo Jiang
- School of Basic Medicine, Zhejiang Chinese Medical University, Hangzhou, China
| | - Xisheng Xu
- School of Basic Medicine, Zhejiang Chinese Medical University, Hangzhou, China
| | - Yuliang Liu
- School of Basic Medicine, Zhejiang Chinese Medical University, Hangzhou, China
| |
Collapse
|
32
|
Zhou Z, Li X, Ji H, Xu X, Chang Z, Wu K, Song Y, Kao M, Chen H, Wu D, Zhang T. Application of improved Unet network in the recognition and segmentation of lung CT images in patients with pneumoconiosis. BMC Med Imaging 2024; 24:220. [PMID: 39160488 PMCID: PMC11331615 DOI: 10.1186/s12880-024-01377-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Accepted: 07/22/2024] [Indexed: 08/21/2024] Open
Abstract
BACKGROUND Pneumoconiosis has a significant impact on the quality of patient survival. This study aims to evaluate the performance and application value of improved Unet network technology in the recognition and segmentation of lesion areas of lung CT images in patients with pneumoconiosis. METHODS A total of 1212 lung CT images of patients with pneumoconiosis were retrospectively included. The improved Unet network was used to identify and segment the CT image regions of the patients' lungs, and the image data of the granular regions of the lungs were processed by the watershed and region growing algorithms. After random sorting, 848 data were selected into the training set and 364 data into the validation set. The experimental dataset underwent data augmentation and were used for model training and validation to evaluate segmentation performance. The segmentation results were compared with FCN-8s, Unet network (Base), Unet (Squeeze-and-Excitation, SE + Rectified Linear Unit, ReLU), and Unet + + networks. RESULTS In the segmentation of lung CT granular region with the improved Unet network, the four evaluation indexes of Dice similarity coefficient, positive prediction value (PPV), sensitivity coefficient (SC) and mean intersection over union (MIoU) reached 0.848, 0.884, 0.895 and 0.885, respectively, increasing by 7.6%, 13.3%, 3.9% and 6.4%, respectively, compared with those of Unet network (Base), and increasing by 187.5%, 249.4%, 131.9% and 51.0%, respectively, compared with those of FCN-8s, and increasing by 14.0%, 31.2%, 4.7% and 9.7%, respectively, compared with those of Unet network (SE + ReLU), while the segmentation performance was also not inferior to that of the Unet + + network. CONCLUSIONS The improved Unet network proposed shows good performance in the recognition and segmentation of abnormal regions in lung CT images in patients with pneumoconiosis, showing potential application value for assisting clinical decision-making.
Collapse
Affiliation(s)
- Zhengsong Zhou
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Xin Li
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China
| | - Hongbo Ji
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Xuanhan Xu
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Zongqi Chang
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China
| | - Keda Wu
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Yangyang Song
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Mingkun Kao
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Hongjun Chen
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Dongsheng Wu
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China.
| | - Tao Zhang
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China.
| |
Collapse
|
33
|
Kunduracioglu I, Pacal I. Advancements in deep learning for accurate classification of grape leaves and diagnosis of grape diseases. JOURNAL OF PLANT DISEASES AND PROTECTION 2024; 131:1061-1080. [DOI: 10.1007/s41348-024-00896-z] [Citation(s) in RCA: 18] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Accepted: 02/20/2024] [Indexed: 05/14/2025]
Abstract
AbstractPlant diseases cause significant agricultural losses, demanding accurate detection methods. Traditional approaches relying on expert knowledge may be biased, but advancements in computing, particularly deep learning, offer non-experts effective tools. This study focuses on fine-tuning cutting-edge pre-trained CNN and vision transformer models to classify grape leaves and diagnose grape leaf diseases through digital images. Our research examined a PlantVillage dataset, which comprises 4062 leaf images distributed across four categories. Additionally, we utilized the Grapevine dataset, consisting of 500 leaf images. This dataset is organized into five distinct groups, with each group containing 100 images corresponding to one of the five grape types. The PlantVillage dataset focuses on four classes related to grape diseases, namely Black Rot, Leaf Blight, Healthy, and Esca leaves. On the other hand, the Grapevine dataset includes five classes for leaf recognition, specifically Ak, Alaidris, Buzgulu, Dimnit, and Nazli. In experiments with 14 CNN and 17 vision transformer models, deep learning demonstrated high accuracy in distinguishing grape diseases and recognizing leaves. Notably, four models achieved 100% accuracy on PlantVillage and Grapevine datasets, with Swinv2-Base standing out. This approach holds promise for enhancing crop productivity through early disease detection and providing insights into grape variety characterization in agriculture.
Collapse
|
34
|
Allogmani AS, Mohamed RM, Al-Shibly NM, Ragab M. Enhanced cervical precancerous lesions detection and classification using Archimedes Optimization Algorithm with transfer learning. Sci Rep 2024; 14:12076. [PMID: 38802525 PMCID: PMC11130149 DOI: 10.1038/s41598-024-62773-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Accepted: 05/21/2024] [Indexed: 05/29/2024] Open
Abstract
Cervical cancer (CC) ranks as the fourth most common form of cancer affecting women, manifesting in the cervix. CC is caused by the Human papillomavirus (HPV) infection and is eradicated by vaccinating women from an early age. However, limited medical facilities present a significant challenge in mid- or low-income countries. It can improve the survivability rate and be successfully treated if the CC is detected at earlier stages. Current technological improvements allow for cost-effective, more sensitive, and rapid screening and treatment measures for CC. DL techniques are widely adopted for the automated detection of CC. DL techniques and architectures are used to detect CC and provide higher detection performance. This study offers the design of Enhanced Cervical Precancerous Lesions Detection and Classification using the Archimedes Optimization Algorithm with Transfer Learning (CPLDC-AOATL) algorithm. The CPLDC-AOATL algorithm aims to diagnose cervical cancer using medical images. At the preliminary stage, the CPLDC-AOATL technique involves a bilateral filtering (BF) technique to eliminate the noise in the input images. Besides, the CPLDC-AOATL technique applies the Inception-ResNetv2 model for the feature extraction process, and the use of AOA chose the hyperparameters. The CPLDC-AOATL technique involves a bidirectional long short-term memory (BiLSTM) model for the cancer detection process. The experimental outcome of the CPLDC-AOATL technique emphasized the superior accuracy outcome of 99.53% over other existing approaches under a benchmark dataset.
Collapse
Affiliation(s)
- Ayed S Allogmani
- University of Jeddah, College of Science and Arts at Khulis, Department of Biology, Jeddah, Saudi Arabia
| | - Roushdy M Mohamed
- University of Jeddah, College of Science and Arts at Khulis, Department of Biology, Jeddah, Saudi Arabia.
| | - Nasser M Al-Shibly
- Physiotherapy Department, College of Applied Health Sciences, Jerash University, Jerash, Jordan
| | - Mahmoud Ragab
- Information Technology Department, Faculty of Computing and Information Technology, King Abdulaziz University, 21589, Jeddah, Saudi Arabia.
| |
Collapse
|
35
|
Lindroth H, Nalaie K, Raghu R, Ayala IN, Busch C, Bhattacharyya A, Moreno Franco P, Diedrich DA, Pickering BW, Herasevich V. Applied Artificial Intelligence in Healthcare: A Review of Computer Vision Technology Application in Hospital Settings. J Imaging 2024; 10:81. [PMID: 38667979 PMCID: PMC11050909 DOI: 10.3390/jimaging10040081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 03/08/2024] [Accepted: 03/11/2024] [Indexed: 04/28/2024] Open
Abstract
Computer vision (CV), a type of artificial intelligence (AI) that uses digital videos or a sequence of images to recognize content, has been used extensively across industries in recent years. However, in the healthcare industry, its applications are limited by factors like privacy, safety, and ethical concerns. Despite this, CV has the potential to improve patient monitoring, and system efficiencies, while reducing workload. In contrast to previous reviews, we focus on the end-user applications of CV. First, we briefly review and categorize CV applications in other industries (job enhancement, surveillance and monitoring, automation, and augmented reality). We then review the developments of CV in the hospital setting, outpatient, and community settings. The recent advances in monitoring delirium, pain and sedation, patient deterioration, mechanical ventilation, mobility, patient safety, surgical applications, quantification of workload in the hospital, and monitoring for patient events outside the hospital are highlighted. To identify opportunities for future applications, we also completed journey mapping at different system levels. Lastly, we discuss the privacy, safety, and ethical considerations associated with CV and outline processes in algorithm development and testing that limit CV expansion in healthcare. This comprehensive review highlights CV applications and ideas for its expanded use in healthcare.
Collapse
Affiliation(s)
- Heidi Lindroth
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- Center for Aging Research, Regenstrief Institute, School of Medicine, Indiana University, Indianapolis, IN 46202, USA
- Center for Health Innovation and Implementation Science, School of Medicine, Indiana University, Indianapolis, IN 46202, USA
| | - Keivan Nalaie
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Roshini Raghu
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
| | - Ivan N. Ayala
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
| | - Charles Busch
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- College of Engineering, University of Wisconsin-Madison, Madison, WI 53705, USA
| | | | - Pablo Moreno Franco
- Department of Transplantation Medicine, Mayo Clinic, Jacksonville, FL 32224, USA
| | - Daniel A. Diedrich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Brian W. Pickering
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| |
Collapse
|