1
|
Awasthi V, Tiwari M, Yadav A, Thakur G, Panda MM, Kumar H, Tripathi S. Optimizing brain tumor detection in MRI scans through InceptionResNetV2 and deep stacked Autoencoders with SwiGLU activation and sparsity regularization. MethodsX 2025; 14:103255. [PMID: 40144141 PMCID: PMC11938147 DOI: 10.1016/j.mex.2025.103255] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2024] [Accepted: 03/05/2025] [Indexed: 03/28/2025] Open
Abstract
This study presents an automated framework for brain tumor classification aimed at accurately distinguishing tumor types in MRI images. The proposed model integrates InceptionResNetV2 for feature extraction with Deep Stacked Autoencoders (DSAEs) for classification, enhanced by sparsity regularization and the SwiGLU activation function. InceptionResNetV2, pre-trained on ImageNet, was fine-tuned to extract multi-scale features, while the DSAE structure compressed these features to highlight critical attributes essential for classification. The approach achieved high performance, reaching an overall accuracy of 99.53 %, precision of 98.27 %, recall of 99.21 %, specificity of 98.73 %, and an F1-score of 98.74 %. These results demonstrate the model's efficacy in accurately categorizing glioma, meningioma, pituitary tumors, and non-tumor cases, with minimal misclassifications. Despite its success, limitations include the model's dependency on pre-trained weights and significant computational resources. Future studies should address these limitations by enhancing interpretability, exploring domain-specific transfer learning, and validating on diverse datasets to strengthen the model's utility in real-world settings. Overall, the InceptionResNetV2 integrated with DSAEs, sparsity regularization, and SwiGLU offers a promising solution for reliable and efficient brain tumor diagnosis in clinical environments.•Leveraging a pre-trained InceptionResNetV2 model to capture multi-scale features from MRI data.•Utilizing Deep Stacked Autoencoders with sparsity regularization to emphasize critical attributes for precise classification.•Incorporating the SwiGLU activation function to capture complex, non-linear patterns within the data.
Collapse
Affiliation(s)
- Vishal Awasthi
- Department of Electronics and Communication Engineering, Chhatrapati Shahu Ji Maharaj University, Kanpur, India
| | - Mamta Tiwari
- Department of Computer Application, Chhatrapati Shahu Ji Maharaj University, Kanpur, India
| | - Amit Yadav
- Department of Computer Application, PSIT College of Higher Education, Kanpur, India
| | | | - Mamata Mayee Panda
- Department of Computer Application, Chhatrapati Shahu Ji Maharaj University, Kanpur, India
| | - Hemant Kumar
- Department of Information Technology, Chhatrapati Shahu Ji Maharaj University, Kanpur, India
| | - Shivneet Tripathi
- Department of Computer Application, Chhatrapati Shahu Ji Maharaj University, Kanpur, India
| |
Collapse
|
2
|
Pacal I, Ozdemir B, Zeynalov J, Gasimov H, Pacal N. A novel CNN-ViT-based deep learning model for early skin cancer diagnosis. Biomed Signal Process Control 2025; 104:107627. [DOI: 10.1016/j.bspc.2025.107627] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
3
|
Alom MR, Farid FA, Rahaman MA, Rahman A, Debnath T, Miah ASM, Mansor S. An explainable AI-driven deep neural network for accurate breast cancer detection from histopathological and ultrasound images. Sci Rep 2025; 15:17531. [PMID: 40394112 PMCID: PMC12092800 DOI: 10.1038/s41598-025-97718-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2024] [Accepted: 04/07/2025] [Indexed: 05/22/2025] Open
Abstract
Breast cancer represents a significant global health challenge, which makes it essential to detect breast cancer early and accurately to improve patient prognosis and reduce mortality rates. However, traditional diagnostic processes relying on manual analysis of medical images are inherently complex and subject to variability between observers, highlighting the urgent need for robust automated breast cancer detection systems. While deep learning has demonstrated potential, many current models struggle with limited accuracy and lack of interpretability. This research introduces the Deep Neural Breast Cancer Detection (DNBCD) model, an explainable AI-based framework that utilizes deep learning methods for classifying breast cancer using histopathological and ultrasound images. The proposed model employs Densenet121 as a foundation, integrating customized Convolutional Neural Network (CNN) layers including GlobalAveragePooling2D, Dense, and Dropout layers along with transfer learning to achieve both high accuracy and interpretability for breast cancer diagnosis. The proposed DNBCD model integrates several preprocessing techniques, including image normalization and resizing, and augmentation techniques to enhance the model's robustness and address class imbalances using class weight. It employs Grad-CAM (Gradient-weighted Class Activation Mapping) to offer visual justifications for its predictions, increasing trust and transparency among healthcare providers. The model was assessed using two benchmark datasets: Breakhis-400x (B-400x) and Breast Ultrasound Images Dataset (BUSI) containing 1820 and 1578 images, respectively. We systematically divided the datasets into training (70%), testing (20%,) and validation (10%) sets, ensuring efficient model training and evaluation obtaining accuracies of 93.97% for B-400x dataset having benign and malignant classes and 89.87% for BUSI dataset having benign, malignant, and normal classes for breast cancer detection. Experimental results demonstrate that the proposed DNBCD model significantly outperforms existing state-of-the-art approaches with potential uses in clinical environments. We also made all the materials publicly accessible for the research community at: https://github.com/romzanalom/XAI-Based-Deep-Neural-Breast-Cancer-Detection .
Collapse
Affiliation(s)
- Md Romzan Alom
- Department of Computer Science and Engineering, Green University of Bangladesh (GUB), Purbachal American City, Kanchon, Dhaka, 1460, Bangladesh
| | - Fahmid Al Farid
- Faculty of Artificial Intelligence and Engineering, Multimedia University, 63100, Cyberjaya, Malaysia
| | - Muhammad Aminur Rahaman
- Department of Computer Science and Engineering, Green University of Bangladesh (GUB), Purbachal American City, Kanchon, Dhaka, 1460, Bangladesh.
| | - Anichur Rahman
- Department of Computer Science and Engineering, National Institute of Textile Engineering and Research (NITER), Constituent Institute of the University of Dhaka, Savar, Dhaka, 1350, Bangladesh.
- Department of Computer Science and Engineering, Mawlana Bhashani Science and Technology University, Tangail, Bangladesh.
| | - Tanoy Debnath
- Department of Computer Science, Stony Brook University, Stony Brook, NY, USA
| | - Abu Saleh Musa Miah
- Department of Computer Science and Engineering, Bangladesh Army University of Science and Technology (BAUST), Nilphamari, Bangladesh
| | - Sarina Mansor
- Faculty of Artificial Intelligence and Engineering, Multimedia University, 63100, Cyberjaya, Malaysia.
| |
Collapse
|
4
|
Hussain I, Ching KB, Uttraphan C, Tay KG, Noor A. Evaluating machine learning algorithms for energy consumption prediction in electric vehicles: A comparative study. Sci Rep 2025; 15:16124. [PMID: 40341692 PMCID: PMC12062255 DOI: 10.1038/s41598-025-94946-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2024] [Accepted: 03/18/2025] [Indexed: 05/10/2025] Open
Abstract
An accurate energy consumption prediction becomes crucial with increasing electric vehicle usage for effective power grid management. This research examined the performance of eleven machine learning models for this purpose: Ridge Regression, Lasso Regression, K-Nearest Neighbors, Gradient Boosting, Support Vector Regression, Multi-Layer Perceptron, XGBoost, CatBoost, LightGBM, Gaussian Processes for Regression(GPR) and Extra Trees Regressor, considering real historical data from Colorado. The models were evaluated using different metrics: Mean Absolute Error (MAE), Mean Squared Error (MSE), R², Root Mean Squared Error(RMSE) and Normalized Root Mean Squared Error(NRMSE), with visual analyses through scatter plots and time series plots. The best model observed was the Extra Trees Regressor, which had an MAE of 0.5888, an MSE of 3.2683, R² value of 0.9592, RMSE of 1.8078 and NRMSE of 0.020. Gradient Boosting and KNN also returned good results, although they were slightly more dispersed. Nevertheless, while non-linear models like MLP, XGBoost, CatBoost, LightGBM and linear models such as Ridge and Lasso Regression offer valuable insights, they exhibit shortcomings in estimating energy, especially at extreme levels, highlighting limitations in capturing complex non-linear interactions. This study focuses on their applicability to energy projections to demonstrate how well ensemble and non-linear models may capture intricate patterns in time series. These cutting-edge machine learning techniques might greatly enhance energy demand predictions.
Collapse
Affiliation(s)
- Izhar Hussain
- Departement of Electrical Engineering, Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat, Johar, 86400, Malaysia.
- Department of Electronics Engineering Benazir Bhutto Shaheed, University of Technology and Skill Development, Khairpure mirs, Khairpur, 66020, Pakistan.
| | - Kok Boon Ching
- Departement of Electrical Engineering, Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat, Johar, 86400, Malaysia.
| | - Chessda Uttraphan
- Department of Computer Engineering, Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, 86400, Parit Raja, Batu Pahat, Johor, Malaysia
| | - Kim Gaik Tay
- Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat, Johar, 86400, Malaysia
| | - Adil Noor
- Departement of Electrical Engineering, Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat, Johar, 86400, Malaysia
| |
Collapse
|
5
|
Asif RN, Naseem MT, Ahmad M, Mazhar T, Khan MA, Khan MA, Al-Rasheed A, Hamam H. Brain tumor detection empowered with ensemble deep learning approaches from MRI scan images. Sci Rep 2025; 15:15002. [PMID: 40301625 PMCID: PMC12041211 DOI: 10.1038/s41598-025-99576-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2025] [Accepted: 04/21/2025] [Indexed: 05/01/2025] Open
Abstract
Brain tumor detection is essential for early diagnosis and successful treatment, both of which can significantly enhance patient outcomes. To evaluate brain MRI scans and categorize them into four types-pituitary, meningioma, glioma, and normal-this study investigates a potent artificial intelligence (AI) technique. Even though AI has been utilized in the past to detect brain tumors, current techniques still have issues with accuracy and dependability. Our study presents a novel AI technique that combines two distinct deep learning models to enhance this. When combined, these models improve accuracy and yield more trustworthy outcomes than when used separately. Key performance metrics including accuracy, precision, and dependability are used to assess the system once it has been trained using MRI scan pictures. Our results show that this combined AI approach works better than individual models, particularly in identifying different types of brain tumors. Specifically, the InceptionV3 + Xception combination hit an accuracy level of 98.50% in training and 98.30% in validation. Such results further argue the potential application for advanced AI techniques in medical imaging while speaking even more strongly to the fact that multiple AI models used concurrently are able to enhance brain tumor detection.
Collapse
Affiliation(s)
- Rizwana Naz Asif
- School of Computer Science, National College of Business Administration and Economics, Lahore, 54000, Pakistan
| | - Muhammad Tahir Naseem
- Department of Electronic Engineering, Yeungnam University, Gyeongsan-si, 38541, Republic of Korea
| | - Munir Ahmad
- University College, Korea University, Seoul, 02841, Republic of Korea
| | - Tehseen Mazhar
- School of Computer Science, National College of Business Administration and Economics, Lahore, 54000, Pakistan.
- Department of Computer Science,School Education Department, Government of Punjab, Layyah, 31200, Pakistan.
| | - Muhammad Adnan Khan
- Department of Software, Faculty of Artificial Intelligence and Software, Gachon University, Seongnam-si, 13557, Republic of Korea.
| | - Muhammad Amir Khan
- School of Computing Sciences, College of Computing, Informatics and Mathematics, Universiti Teknologi MARA, Shah Alam, 40450, Selangor, Malaysia.
| | - Amal Al-Rasheed
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, 11671, Saudi Arabia
| | - Habib Hamam
- Faculty of Engineering, Université de Moncton, Moncton, NB E1 A3E9, Canada
- School of Electrical Engineering, University of Johannesburg, Johannesburg, 2006, South Africa
- International Institute of Technology and Management (IITG), Av. Grandes Ecoles, Libreville BP 1989, Gabon
- Bridges for Academic Excellence, Spectrum, Tunis, Center-ville, Tunisia
| |
Collapse
|
6
|
Kaur P, Mahajan P. Detection of brain tumors using a transfer learning-based optimized ResNet152 model in MR images. Comput Biol Med 2025; 188:109790. [PMID: 39951980 DOI: 10.1016/j.compbiomed.2025.109790] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2024] [Revised: 01/28/2025] [Accepted: 01/31/2025] [Indexed: 02/17/2025]
Abstract
Brain tumors are incredibly harmful and can drastically reduce life expectancy. Most researchers use magnetic resonance (MR) scans to detect tumors because they can provide detailed images of the affected area. Recently, AI-based deep learning methods have emerged to enhance diagnostic accuracy through efficient data processing. This study investigates the effectiveness of deep transfer learning techniques for accurate brain tumor diagnosis. A preprocessing pipeline is used to enhance the image quality. This pipeline includes morphological operations such as erosion and dilation for shape refinement, Gaussian blurring for noise reduction, and thresholding for image cropping. Principal Component Analysis (PCA) is applied for dimensionality reduction, and data augmentation enriches the dataset. The dataset is partitioned into training (80 %) and testing (20 %). Pretrained ResNet152 and GoogleNet extract meaningful features from the images. These extracted features are then classified using conventional machine learning classifiers: Support Vector Machine (SVM), K-Nearest Neighbors (KNN), Classification and Regression Trees (CART), and Gaussian Naive Bayes (GNB). This study compares the performance of two pre-trained models for medical image analysis. Performance metrics such as accuracy, sensitivity, recall, and F1-Score evaluate the final classification results. ResNet152 outperforms GoogleNet, achieving a 98.53 % accuracy, an F1 score of 97.4 %, and a sensitivity of 96.52 %. This study highlights integrating deep learning and traditional machine-learning techniques in medical image analysis for effective brain tumor detection.
Collapse
Affiliation(s)
- Prabhpreet Kaur
- Department of Computer Engineering and Technology, Guru Nanak Dev University, Amritsar, India.
| | - Priyanka Mahajan
- Department of Computer Engineering and Technology, Guru Nanak Dev University, Amritsar, India.
| |
Collapse
|
7
|
Zafar J, Koc V, Zafar H. Dual-Stream Contrastive Latent Learning Generative Adversarial Network for Brain Image Synthesis and Tumor Classification. J Imaging 2025; 11:101. [PMID: 40278017 PMCID: PMC12027838 DOI: 10.3390/jimaging11040101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2025] [Revised: 03/24/2025] [Accepted: 03/26/2025] [Indexed: 04/26/2025] Open
Abstract
Generative adversarial networks (GANs) prioritize pixel-level attributes over capturing the entire image distribution, which is critical in image synthesis. To address this challenge, we propose a dual-stream contrastive latent projection generative adversarial network (DSCLPGAN) for the robust augmentation of MRI images. The dual-stream generator in our architecture incorporates two specialized processing pathways: one is dedicated to local feature variation modeling, while the other captures global structural transformations, ensuring a more comprehensive synthesis of medical images. We used a transformer-based encoder-decoder framework for contextual coherence and the contrastive learning projection (CLP) module integrates contrastive loss into the latent space for generating diverse image samples. The generated images undergo adversarial refinement using an ensemble of specialized discriminators, where discriminator 1 (D1) ensures classification consistency with real MRI images, discriminator 2 (D2) produces a probability map of localized variations, and discriminator 3 (D3) preserves structural consistency. For validation, we utilized a publicly available MRI dataset which contains 3064 T1-weighted contrast-enhanced images with three types of brain tumors: meningioma (708 slices), glioma (1426 slices), and pituitary tumor (930 slices). The experimental results demonstrate state-of-the-art performance, achieving an SSIM of 0.99, classification accuracy of 99.4% for an augmentation diversity level of 5, and a PSNR of 34.6 dB. Our approach has the potential of generating high-fidelity augmentations for reliable AI-driven clinical decision support systems.
Collapse
Affiliation(s)
- Junaid Zafar
- Faculty of Engineering, Government College University, Lahore 54000, Pakistan;
| | - Vincent Koc
- Independent Researcher, Kuala Lumpur 55100, Malaysia;
| | - Haroon Zafar
- Lambe Institute for Translational Research, University of Galway, H91 YR71 Galway, Ireland
| |
Collapse
|
8
|
Bayram B, Kunduracioglu I, Ince S, Pacal I. A systematic review of deep learning in MRI-based cerebral vascular occlusion-based brain diseases. Neuroscience 2025; 568:76-94. [PMID: 39805420 DOI: 10.1016/j.neuroscience.2025.01.020] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 01/09/2025] [Accepted: 01/10/2025] [Indexed: 01/16/2025]
Abstract
Neurological disorders, including cerebral vascular occlusions and strokes, present a major global health challenge due to their high mortality rates and long-term disabilities. Early diagnosis, particularly within the first hours, is crucial for preventing irreversible damage and improving patient outcomes. Although neuroimaging techniques like magnetic resonance imaging (MRI) have advanced significantly, traditional methods often fail to fully capture the complexity of brain lesions. Deep learning has recently emerged as a powerful tool in medical imaging, offering high accuracy in detecting and segmenting brain anomalies. This review examines 61 MRI-based studies published between 2020 and 2024, focusing on the role of deep learning in diagnosing cerebral vascular occlusion-related conditions. It evaluates the successes and limitations of these studies, including the adequacy and diversity of datasets, and addresses challenges such as data privacy and algorithm explainability. Comparisons between convolutional neural network (CNN)-based and Vision Transformer (ViT)-based approaches reveal distinct advantages and limitations. The findings emphasize the importance of ethically secure frameworks, the inclusion of diverse datasets, and improved model interpretability. Advanced architectures like U-Net variants and transformer-based models are highlighted as promising tools to enhance reliability in clinical applications. By automating complex neuroimaging tasks and improving diagnostic accuracy, deep learning facilitates personalized treatment strategies. This review provides a roadmap for integrating technical advancements into clinical practice, underscoring the transformative potential of deep learning in managing neurological disorders and improving healthcare outcomes globally.
Collapse
Affiliation(s)
- Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| | - Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| |
Collapse
|
9
|
Dorathi Jayaseeli JD, Briskilal J, Fancy C, Vaitheeshwaran V, Patibandla RSML, Syed K, Swain AK. An intelligent framework for skin cancer detection and classification using fusion of Squeeze-Excitation-DenseNet with Metaheuristic-driven ensemble deep learning models. Sci Rep 2025; 15:7425. [PMID: 40033075 PMCID: PMC11876321 DOI: 10.1038/s41598-025-92293-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2024] [Accepted: 02/26/2025] [Indexed: 03/05/2025] Open
Abstract
Skin cancer is the most dominant and critical method of cancer, which arises all over the world. Its damaging effects can range from disfigurement to major medical expenditures and even death if not analyzed and preserved timely. Conventional models of skin cancer recognition require a complete physical examination by a specialist, which is time-wasting in a few cases. Computer-aided medicinal analytical methods have gained massive popularity due to their efficiency and effectiveness. This model can assist dermatologists in the initial recognition of skin cancer, which is significant for early diagnosis. An automatic classification model utilizing deep learning (DL) can help doctors perceive the kind of skin lesion and improve the patient's health. The classification of skin cancer is one of the hot topics in the research field, along with the development of DL structure. This manuscript designs and develops a Detection of Skin Cancer Using an Ensemble Deep Learning Model and Gray Wolf Optimization (DSC-EDLMGWO) method. The proposed DSC-EDLMGWO model relies on the recognition and classification of skin cancer in biomedical imaging. The presented DSC-EDLMGWO model initially involves the image preprocessing stage at two levels: contract enhancement using the CLAHE method and noise removal using the wiener filter (WF) model. Furthermore, the proposed DSC-EDLMGWO model utilizes the SE-DenseNet method, which is the fusion of the squeeze-and-excitation (SE) module and DenseNet to extract features. For the classification process, the ensemble of DL models, namely the long short-term memory (LSTM) technique, extreme learning machine (ELM) model, and stacked sparse denoising autoencoder (SSDA) method, is employed. Finally, the gray wolf optimization (GWO) method optimally adjusts the ensemble DL models' hyperparameter values, resulting in more excellent classification performance. The effectiveness of the DSC-EDLMGWO approach is evaluated using a benchmark image database, with outcomes measured across various performance metrics. The experimental validation of the DSC-EDLMGWO approach portrayed a superior accuracy value of 98.38% and 98.17% under HAM10000 and ISIC datasets across other techniques.
Collapse
Affiliation(s)
- J D Dorathi Jayaseeli
- Department of Computing Technologies, SRM Institute of Science and Technology, Kattankulathur, 603203, India
| | - J Briskilal
- Department of Computing Technologies, Faculty of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - C Fancy
- Department of Networking and Communications, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - V Vaitheeshwaran
- Department of Computer Science and Engineering, Aditya University, Surampalem, Kakinada, India
| | - R S M Lakshmi Patibandla
- Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Guntur, Andhra Pradesh, India
| | - Khasim Syed
- School of Computer Science & Engineering, VIT - AP University, Amaravati, 522237, Andhra Pradesh, India.
| | - Anil Kumar Swain
- KIIT Deemed to be University, Bhubaneswar, 751024, Odisha, India
| |
Collapse
|
10
|
Ozdemir B, Pacal I. An innovative deep learning framework for skin cancer detection employing ConvNeXtV2 and focal self-attention mechanisms. RESULTS IN ENGINEERING 2025; 25:103692. [DOI: 10.1016/j.rineng.2024.103692] [Citation(s) in RCA: 11] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
11
|
Alzahrani AA, Alsamri J, Maashi M, Negm N, Asklany SA, Alkharashi A, Alkhiri H, Obayya M. Deep structured learning with vision intelligence for oral carcinoma lesion segmentation and classification using medical imaging. Sci Rep 2025; 15:6610. [PMID: 39994267 PMCID: PMC11850820 DOI: 10.1038/s41598-025-89971-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2024] [Accepted: 02/10/2025] [Indexed: 02/26/2025] Open
Abstract
Oral carcinoma (OC) is a toxic illness among the most general malignant cancers globally, and it has developed a gradually significant public health concern in emerging and low-to-middle-income states. Late diagnosis, high incidence, and inadequate treatment strategies remain substantial challenges. Analysis at an initial phase is significant for good treatment, prediction, and existence. Despite the current growth in the perception of molecular devices, late analysis and methods near precision medicine for OC patients remain a challenge. A machine learning (ML) model was employed to improve early detection in medicine, aiming to reduce cancer-specific mortality and disease progression. Recent advancements in this approach have significantly enhanced the extraction and diagnosis of critical information from medical images. This paper presents a Deep Structured Learning with Vision Intelligence for Oral Carcinoma Lesion Segmentation and Classification (DSLVI-OCLSC) model for medical imaging. Using medical imaging, the DSLVI-OCLSC model aims to enhance OC's classification and recognition outcomes. To accomplish this, the DSLVI-OCLSC model utilizes wiener filtering (WF) as a pre-processing technique to eliminate the noise. In addition, the ShuffleNetV2 method is used for the group of higher-level deep features from an input image. The convolutional bidirectional long short-term memory network with a multi-head attention mechanism (MA-CNN-BiLSTM) approach is utilized for oral carcinoma recognition and identification. Moreover, the Unet3 + is employed to segment abnormal regions from the classified images. Finally, the sine cosine algorithm (SCA) approach is utilized to hyperparameter-tune the DL model. A wide range of simulations is implemented to ensure the enhanced performance of the DSLVI-OCLSC method under the OC images dataset. The experimental analysis of the DSLVI-OCLSC method portrayed a superior accuracy value of 98.47% over recent approaches.
Collapse
Affiliation(s)
- Ahmad A Alzahrani
- Department of Computer Science and Artificial Intelligence, College of Computing, Umm-AlQura University, Mecca, Saudi Arabia
| | - Jamal Alsamri
- Department of Biomedical Engineering, College of Engineering, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, PO Box 103786, 11543, Riyadh, Saudi Arabia
| | - Noha Negm
- Department of Computer Science, Applied College at Mahayil, King Khalid University, Abha, Saudi Arabia
| | - Somia A Asklany
- Department of Computer Science and Information Technology, Faculty of Sciences and Arts, Turaif, Northern Border University, 91431, Arar, Saudi Arabia.
| | - Abdulwhab Alkharashi
- Department of Computer Science, College of Computing and Informatics, Saudi Electronic University, Riyadh, Saudi Arabia
| | - Hassan Alkhiri
- Department of Computer Science, Faculty of Computing and Information Technology, Al-Baha University, Al-Baha, Saudi Arabia
| | - Marwa Obayya
- Department of Computer Science and Artificial Intelligence, College of Computing, Umm-AlQura University, Mecca, Saudi Arabia
| |
Collapse
|
12
|
Yadav DP, Sharma B, Noonia A, Mehbodniya A. Explainable label guided lightweight network with axial transformer encoder for early detection of oral cancer. Sci Rep 2025; 15:6391. [PMID: 39984521 PMCID: PMC11845714 DOI: 10.1038/s41598-025-87627-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2024] [Accepted: 01/21/2025] [Indexed: 02/23/2025] Open
Abstract
Oral cavity cancer exhibits high morbidity and mortality rates. Therefore, it is essential to diagnose the disease at an early stage. Machine learning and convolution neural networks (CNN) are powerful tools for diagnosing mouth and oral cancer. In this study, we design a lightweight explainable network (LWENet) with label-guided attention (LGA) to provide a second opinion to the expert. The LWENet contains depth-wise separable convolution layers to reduce the computation costs. Moreover, the LGA module provides label consistency to the neighbor pixel and improves the spatial features. Furthermore, AMSA (axial multi-head self-attention) based ViT encoder incorporated in the model to provide global attention. Our ViT (vision transformer) encoder is computationally efficient compared to the classical ViT encoder. We tested LWRNet performance on the MOD (mouth and oral disease) and OCI (oral cancer image) datasets, and results are compared with the other CNN and ViT (vision transformer) based methods. The LWENet achieved a precision and F1-scores of 96.97% and 98.90% on the MOD dataset, and 99.48% and 98.23% on the OCI dataset, respectively. By incorporating Grad-CAM, we visualize the decision-making process, enhancing model interpretability. This work demonstrates the potential of LWENet with LGA in facilitating early oral cancer detection.
Collapse
Affiliation(s)
- Dhirendra Prasad Yadav
- Department of Computer Engineering & Applications, GLA University Mathura, Mathura, India
| | - Bhisham Sharma
- Centre for Research Impact & Outcome, Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab, 140401, India.
| | - Ajit Noonia
- Department of Computer Science and Engineering, Manipal University Jaipur, Jaipur, Rajasthan, India
| | - Abolfazl Mehbodniya
- Department of Electronics and Communication Engineering, Kuwait College of Science and Technology (KCST), Doha Area, 7th Ring Road, Kuwait City, Kuwait
| |
Collapse
|
13
|
Ozdemir B, Pacal I. A robust deep learning framework for multiclass skin cancer classification. Sci Rep 2025; 15:4938. [PMID: 39930026 PMCID: PMC11811178 DOI: 10.1038/s41598-025-89230-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2024] [Accepted: 02/04/2025] [Indexed: 02/13/2025] Open
Abstract
Skin cancer represents a significant global health concern, where early and precise diagnosis plays a pivotal role in improving treatment efficacy and patient survival rates. Nonetheless, the inherent visual similarities between benign and malignant lesions pose substantial challenges to accurate classification. To overcome these obstacles, this study proposes an innovative hybrid deep learning model that combines ConvNeXtV2 blocks and separable self-attention mechanisms, tailored to enhance feature extraction and optimize classification performance. The inclusion of ConvNeXtV2 blocks in the initial two stages is driven by their ability to effectively capture fine-grained local features and subtle patterns, which are critical for distinguishing between visually similar lesion types. Meanwhile, the adoption of separable self-attention in the later stages allows the model to selectively prioritize diagnostically relevant regions while minimizing computational complexity, addressing the inefficiencies often associated with traditional self-attention mechanisms. The model was comprehensively trained and validated on the ISIC 2019 dataset, which includes eight distinct skin lesion categories. Advanced methodologies such as data augmentation and transfer learning were employed to further enhance model robustness and reliability. The proposed architecture achieved exceptional performance metrics, with 93.48% accuracy, 93.24% precision, 90.70% recall, and a 91.82% F1-score, outperforming over ten Convolutional Neural Network (CNN) based and over ten Vision Transformer (ViT) based models tested under comparable conditions. Despite its robust performance, the model maintains a compact design with only 21.92 million parameters, making it highly efficient and suitable for model deployment. The Proposed Model demonstrates exceptional accuracy and generalizability across diverse skin lesion classes, establishing a reliable framework for early and accurate skin cancer diagnosis in clinical practice.
Collapse
Affiliation(s)
- Burhanettin Ozdemir
- Department of Operations and Project Management, College of Business, Alfaisal University, Riyadh, 11533, Saudi Arabia.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, Igdir, 76000, Turkey
- Department of Electronics and Information Technologies, Faculty of Architecture and Engineering, Nakhchivan State University, AZ 7012, Nakhchivan, Azerbaijan
| |
Collapse
|
14
|
Alotaibi SR, Alohali MA, Maashi M, Alqahtani H, Alotaibi M, Mahmud A. Advances in colorectal cancer diagnosis using optimal deep feature fusion approach on biomedical images. Sci Rep 2025; 15:4200. [PMID: 39905104 PMCID: PMC11794880 DOI: 10.1038/s41598-024-83466-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Accepted: 12/16/2024] [Indexed: 02/06/2025] Open
Abstract
Colorectal cancer (CRC) is the second popular cancer in females and third in males, with an increased number of cases. Pathology diagnoses complemented with predictive and prognostic biomarker information is the first step for personalized treatment. Histopathological image (HI) analysis is the benchmark for pathologists to rank colorectal cancer of various kinds. However, pathologists' diagnoses are highly subjective and susceptible to inaccurate diagnoses. The improved diagnosis load in the pathology laboratory, incorporated with the reported intra- and inter-variability in the biomarker assessment, has prompted the quest for consistent machine-based techniques to be integrated into routine practice. In the healthcare field, artificial intelligence (AI) has achieved extraordinary achievements in healthcare applications. Lately, computer-aided diagnosis (CAD) based on HI has progressed rapidly with the increase of machine learning (ML) and deep learning (DL) based models. This study introduces a novel Colorectal Cancer Diagnosis using the Optimal Deep Feature Fusion Approach on Biomedical Images (CCD-ODFFBI) method. The primary objective of the CCD-ODFFBI technique is to examine the biomedical images to identify colorectal cancer (CRC). In the CCD-ODFFBI technique, the median filtering (MF) approach is initially utilized for noise elimination. The CCD-ODFFBI technique utilizes a fusion of three DL models, MobileNet, SqueezeNet, and SE-ResNet, for feature extraction. Moreover, the DL models' hyperparameter selection is performed using the Osprey optimization algorithm (OOA). Finally, the deep belief network (DBN) model is employed to classify CRC. A series of simulations is accomplished to highlight the significant results of the CCD-ODFFBI method under the Warwick-QU dataset. The comparison of the CCD-ODFFBI method showed a superior accuracy value of 99.39% over existing techniques.
Collapse
Affiliation(s)
- Sultan Refa Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Manal Abdullah Alohali
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh, 11671, Saudi Arabia.
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, PO Box 103786, Riyadh, 11543, Saudi Arabia
| | - Hamed Alqahtani
- Department of Information Systems, College of Computer Science, Center of Artificial Intelligence, Unit of Cybersecurity, King Khalid University, Abha, Saudi Arabia
| | - Moneerah Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Ahmed Mahmud
- Research Center, Future University in Egypt, New Cairo, 11835, Egypt
| |
Collapse
|
15
|
Pacal I, Işık G. Utilizing convolutional neural networks and vision transformers for precise corn leaf disease identification. Neural Comput Appl 2025; 37:2479-2496. [DOI: 10.1007/s00521-024-10769-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2024] [Accepted: 11/05/2024] [Indexed: 05/14/2025]
|
16
|
Kaur H, Sharma R, Kaur J. Comparison of deep transfer learning models for classification of cervical cancer from pap smear images. Sci Rep 2025; 15:3945. [PMID: 39890842 PMCID: PMC11785805 DOI: 10.1038/s41598-024-74531-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Accepted: 09/26/2024] [Indexed: 02/03/2025] Open
Abstract
Cervical cancer is one of the most commonly diagnosed cancers worldwide, and it is particularly prevalent among women living in developing countries. Traditional classification algorithms often require segmentation and feature extraction techniques to detect cervical cancer. In contrast, convolutional neural networks (CNN) models require large datasets to reduce overfitting and poor generalization. Based on limited datasets, transfer learning was applied directly to pap smear images to perform a classification task. A comprehensive comparison of 16 pre-trained models (VGG16, VGG19, ResNet50, ResNet50V2, ResNet101, ResNet101V2, ResNet152, ResNet152V2, DenseNet121, DenseNet169, DenseNet201, MobileNet, XceptionNet, InceptionV3, and InceptionResNetV2) were carried out for cervical cancer classification by relying on the Herlev dataset and Sipakmed dataset. A comparison of the results revealed that ResNet50 achieved 95% accuracy both for 2-class classification and for 7-class classification using the Herlev dataset. Based on the Sipakmed dataset, VGG16 obtained an accuracy of 99.95% for 2-class and 5-class classification, DenseNet121 achieved an accuracy of 97.65% for 3-class classification. Our findings indicate that DTL models are suitable for automating cervical cancer screening, providing more accurate and efficient results than manual screening.
Collapse
Affiliation(s)
- Harmanpreet Kaur
- Department of Computer Science & Engineering, Punjabi University, Patiala, India.
| | - Reecha Sharma
- Department of Electronics and Communication Engineering, Punjabi University, Patiala, India
| | - Jagroop Kaur
- Department of Computer Science & Engineering, Punjabi University, Patiala, India
| |
Collapse
|
17
|
Abdusalomov A, Mirzakhalilov S, Umirzakova S, Shavkatovich Buriboev A, Meliboev A, Muminov B, Jeon HS. Accessible AI Diagnostics and Lightweight Brain Tumor Detection on Medical Edge Devices. Bioengineering (Basel) 2025; 12:62. [PMID: 39851336 PMCID: PMC11759171 DOI: 10.3390/bioengineering12010062] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2024] [Revised: 12/30/2024] [Accepted: 01/10/2025] [Indexed: 01/26/2025] Open
Abstract
The timely and accurate detection of brain tumors is crucial for effective medical intervention, especially in resource-constrained settings. This study proposes a lightweight and efficient RetinaNet variant tailored for medical edge device deployment. The model reduces computational overhead while maintaining high detection accuracy by replacing the computationally intensive ResNet backbone with MobileNet and leveraging depthwise separable convolutions. The modified RetinaNet achieves an average precision (AP) of 32.1, surpassing state-of-the-art models in small tumor detection (APS: 14.3) and large tumor localization (APL: 49.7). Furthermore, the model significantly reduces computational costs, making real-time analysis feasible on low-power hardware. Clinical relevance is a key focus of this work. The proposed model addresses the diagnostic challenges of small, variable-sized tumors often overlooked by existing methods. Its lightweight architecture enables accurate and timely tumor localization on portable devices, bridging the gap in diagnostic accessibility for underserved regions. Extensive experiments on the BRATS dataset demonstrate the model robustness across tumor sizes and configurations, with confidence scores consistently exceeding 81%. This advancement holds the potential for improving early tumor detection, particularly in remote areas lacking advanced medical infrastructure, thereby contributing to better patient outcomes and broader accessibility to AI-driven diagnostic tools.
Collapse
Affiliation(s)
- Akmalbek Abdusalomov
- Department of Computer Engineering, Gachon University Sujeong-Gu, Seongnam-si 13120, Republic of Korea; (A.A.); (S.U.)
| | - Sanjar Mirzakhalilov
- Department of Computer Systems, Tashkent University of Information Technologies Named after Muhammad Al-Khwarizmi, Tashkent 100200, Uzbekistan;
| | - Sabina Umirzakova
- Department of Computer Engineering, Gachon University Sujeong-Gu, Seongnam-si 13120, Republic of Korea; (A.A.); (S.U.)
| | | | - Azizjon Meliboev
- Department of Digital Technologies and Mathematics, Kokand University, Fergana 150700, Uzbekistan;
| | - Bahodir Muminov
- Department of Artificial Intelligence, Tashkent State University of Economics, Tashkent 100066, Uzbekistan
| | - Heung Seok Jeon
- Department of Software Technology, Konkuk University, Chungju 27478, Republic of Korea
| |
Collapse
|
18
|
Pacal I, Alaftekin M, Zengul FD. Enhancing Skin Cancer Diagnosis Using Swin Transformer with Hybrid Shifted Window-Based Multi-head Self-attention and SwiGLU-Based MLP. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:3174-3192. [PMID: 38839675 PMCID: PMC11612041 DOI: 10.1007/s10278-024-01140-8] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Revised: 05/07/2024] [Accepted: 05/08/2024] [Indexed: 06/07/2024]
Abstract
Skin cancer is one of the most frequently occurring cancers worldwide, and early detection is crucial for effective treatment. Dermatologists often face challenges such as heavy data demands, potential human errors, and strict time limits, which can negatively affect diagnostic outcomes. Deep learning-based diagnostic systems offer quick, accurate testing and enhanced research capabilities, providing significant support to dermatologists. In this study, we enhanced the Swin Transformer architecture by implementing the hybrid shifted window-based multi-head self-attention (HSW-MSA) in place of the conventional shifted window-based multi-head self-attention (SW-MSA). This adjustment enables the model to more efficiently process areas of skin cancer overlap, capture finer details, and manage long-range dependencies, while maintaining memory usage and computational efficiency during training. Additionally, the study replaces the standard multi-layer perceptron (MLP) in the Swin Transformer with a SwiGLU-based MLP, an upgraded version of the gated linear unit (GLU) module, to achieve higher accuracy, faster training speeds, and better parameter efficiency. The modified Swin model-base was evaluated using the publicly accessible ISIC 2019 skin dataset with eight classes and was compared against popular convolutional neural networks (CNNs) and cutting-edge vision transformer (ViT) models. In an exhaustive assessment on the unseen test dataset, the proposed Swin-Base model demonstrated exceptional performance, achieving an accuracy of 89.36%, a recall of 85.13%, a precision of 88.22%, and an F1-score of 86.65%, surpassing all previously reported research and deep learning models documented in the literature.
Collapse
Affiliation(s)
- Ishak Pacal
- Department of Computer Engineering, Igdir University, 76000, Igdir, Turkey
| | - Melek Alaftekin
- Department of Computer Engineering, Igdir University, 76000, Igdir, Turkey
| | - Ferhat Devrim Zengul
- Department of Health Services Administration, The University of Alabama at Birmingham, Birmingham, AL, USA.
- Center for Integrated System, School of Engineering, The University of Alabama at Birmingham, Birmingham, AL, USA.
- Department of Biomedical Informatics and Data Science, School of Medicine, The University of Alabama, Birmingham, USA.
| |
Collapse
|
19
|
Mehedi MHK, Khandaker M, Ara S, Alam MA, Mridha MF, Aung Z. A lightweight deep learning method to identify different types of cervical cancer. Sci Rep 2024; 14:29446. [PMID: 39604499 PMCID: PMC11603366 DOI: 10.1038/s41598-024-79840-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2024] [Accepted: 11/12/2024] [Indexed: 11/29/2024] Open
Abstract
Cervical cancer is the second most common cancer in women's bodies after breast cancer. Cervical cancer develops from dysplasia or cervical intraepithelial neoplasm (CIN), the early stage of the disease, and is characterized by the aberrant growth of cells in the cervix lining. It is primarily caused by Human Papillomavirus (HPV) infection, which spreads through sexual activity. This study focuses on detecting cervical cancer types efficiently using a novel lightweight deep learning model named CCanNet, which combines squeeze block, residual blocks, and skip layer connections. SipakMed, which is not only popular but also publicly available dataset, was used in this study. We conducted a comparative analysis between several transfer learning and transformer models such as VGG19, VGG16, MobileNetV2, AlexNet, ConvNeXT, DeiT_tiny, MobileViT, and Swin Transformer with the proposed CCanNet. Our proposed model outperformed other state-of-the-art models, with 98.53% accuracy and the lowest number of parameters, which is 1,274,663. In addition, accuracy, precision, recall, and the F1 score were used to evaluate the performance of the models. Finally, explainable AI (XAI) was applied to analyze the performance of CCanNet and ensure the results were trustworthy.
Collapse
Affiliation(s)
| | - Moumita Khandaker
- Department of Computer Science and Engineering, BRAC University, Dhaka, Bangladesh
| | - Shaneen Ara
- Department of Computer Science and Engineering, Bangladesh University of Business and Technology, Dhaka, Bangladesh
| | - Md Ashraful Alam
- Department of Computer Science and Engineering, BRAC University, Dhaka, Bangladesh
| | - M F Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka, Bangladesh.
| | - Zeyar Aung
- Department of Computer Science, Khalifa University of Science and Technology, Abu Dhabi, UAE.
| |
Collapse
|
20
|
Maman A, Pacal I, Bati F. Can deep learning effectively diagnose cardiac amyloidosis with 99mTc-PYP scintigraphy? J Radioanal Nucl Chem 2024. [DOI: 10.1007/s10967-024-09879-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Accepted: 11/07/2024] [Indexed: 05/14/2025]
|
21
|
Pacal I, Celik O, Bayram B, Cunha A. Enhancing EfficientNetv2 with global and efficient channel attention mechanisms for accurate MRI-Based brain tumor classification. CLUSTER COMPUTING 2024; 27:11187-11212. [DOI: 10.1007/s10586-024-04532-1] [Citation(s) in RCA: 15] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Revised: 04/08/2024] [Accepted: 04/22/2024] [Indexed: 05/14/2025]
Abstract
AbstractThe early and accurate diagnosis of brain tumors is critical for effective treatment planning, with Magnetic Resonance Imaging (MRI) serving as a key tool in the non-invasive examination of such conditions. Despite the advancements in Computer-Aided Diagnosis (CADx) systems powered by deep learning, the challenge of accurately classifying brain tumors from MRI scans persists due to the high variability of tumor appearances and the subtlety of early-stage manifestations. This work introduces a novel adaptation of the EfficientNetv2 architecture, enhanced with Global Attention Mechanism (GAM) and Efficient Channel Attention (ECA), aimed at overcoming these hurdles. This enhancement not only amplifies the model’s ability to focus on salient features within complex MRI images but also significantly improves the classification accuracy of brain tumors. Our approach distinguishes itself by meticulously integrating attention mechanisms that systematically enhance feature extraction, thereby achieving superior performance in detecting a broad spectrum of brain tumors. Demonstrated through extensive experiments on a large public dataset, our model achieves an exceptional high-test accuracy of 99.76%, setting a new benchmark in MRI-based brain tumor classification. Moreover, the incorporation of Grad-CAM visualization techniques sheds light on the model’s decision-making process, offering transparent and interpretable insights that are invaluable for clinical assessment. By addressing the limitations inherent in previous models, this study not only advances the field of medical imaging analysis but also highlights the pivotal role of attention mechanisms in enhancing the interpretability and accuracy of deep learning models for brain tumor diagnosis. This research sets the stage for advanced CADx systems, enhancing patient care and treatment outcomes.
Collapse
|
22
|
Xiang Q, Li D, Hu Z, Yuan Y, Sun Y, Zhu Y, Fu Y, Jiang Y, Hua X. Quantum classical hybrid convolutional neural networks for breast cancer diagnosis. Sci Rep 2024; 14:24699. [PMID: 39433779 PMCID: PMC11494181 DOI: 10.1038/s41598-024-74778-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Accepted: 09/30/2024] [Indexed: 10/23/2024] Open
Abstract
The World Health Organization states that early diagnosis is essential to increasing the cure rate for breast cancer, which poses a danger to women's health worldwide. However, the efficacy and cost limitations of conventional diagnostic techniques increase the possibility of misdiagnosis. In this work, we present a quantum hybrid classical convolutional neural network (QCCNN) based breast cancer diagnosis approach with the goal of utilizing quantum computing's high-dimensional data processing power and parallelism to increase diagnosis efficiency and accuracy. When working with large-scale and complicated datasets, classical convolutional neural network (CNN) and other machine learning techniques generally demand a large amount of computational resources and time. Their restricted capacity for generalization makes it challenging to maintain consistent performance across multiple data sets. To address these issues, this paper adds a quantum convolutional layer to the classical convolutional neural network to take advantage of quantum computing to improve learning efficiency and processing speed. Simulation experiments on three breast cancer datasets, GBSG, SEER and WDBC, validate the robustness and generalization of QCCNN and significantly outperform CNN and logistic regression models in classification accuracy. This study not only provides a novel method for breast cancer diagnosis but also achieves a breakthrough in breast cancer diagnosis and promotes the development of medical diagnostic technology.
Collapse
Affiliation(s)
- Qiuyu Xiang
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - Dongfen Li
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China.
| | - Zhikang Hu
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - Yuhang Yuan
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - Yuchen Sun
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - Yonghao Zhu
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - You Fu
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - Yangyang Jiang
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| | - Xiaoyu Hua
- College of Computer Science And Cyber Security(Pilot Software College), Chengdu University of Technology, Chengdu, 610059, China
| |
Collapse
|
23
|
Alotaibi M, Alshardan A, Maashi M, Asiri MM, Alotaibi SR, Yafoz A, Alsini R, Khadidos AO. Exploiting histopathological imaging for early detection of lung and colon cancer via ensemble deep learning model. Sci Rep 2024; 14:20434. [PMID: 39227664 PMCID: PMC11372073 DOI: 10.1038/s41598-024-71302-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2024] [Accepted: 08/27/2024] [Indexed: 09/05/2024] Open
Abstract
Cancer seems to have a vast number of deaths due to its heterogeneity, aggressiveness, and significant propensity for metastasis. The predominant categories of cancer that may affect males and females and occur worldwide are colon and lung cancer. A precise and on-time analysis of this cancer can increase the survival rate and improve the appropriate treatment characteristics. An efficient and effective method for the speedy and accurate recognition of tumours in the colon and lung areas is provided as an alternative to cancer recognition methods. Earlier diagnosis of the disease on the front drastically reduces the chance of death. Machine learning (ML) and deep learning (DL) approaches can accelerate this cancer diagnosis, facilitating researcher workers to study a vast majority of patients in a limited period and at a low cost. This research presents Histopathological Imaging for the Early Detection of Lung and Colon Cancer via Ensemble DL (HIELCC-EDL) model. The HIELCC-EDL technique utilizes histopathological images to identify lung and colon cancer (LCC). To achieve this, the HIELCC-EDL technique uses the Wiener filtering (WF) method for noise elimination. In addition, the HIELCC-EDL model uses the channel attention Residual Network (CA-ResNet50) model for learning complex feature patterns. Moreover, the hyperparameter selection of the CA-ResNet50 model is performed using the tuna swarm optimization (TSO) technique. Finally, the detection of LCC is achieved by using the ensemble of three classifiers such as extreme learning machine (ELM), competitive neural networks (CNNs), and long short-term memory (LSTM). To illustrate the promising performance of the HIELCC-EDL model, a complete set of experimentations was performed on a benchmark dataset. The experimental validation of the HIELCC-EDL model portrayed a superior accuracy value of 99.60% over recent approaches.
Collapse
Affiliation(s)
- Moneerah Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Amal Alshardan
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, 11671, Riyadh, Saudi Arabia
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 103786, 11543, Riyadh, Saudi Arabia
| | - Mashael M Asiri
- Department of Computer Science, Applied College at Mahayil, King Khalid University, Abha, Saudi Arabia.
| | - Sultan Refa Alotaibi
- Department of Computer Science, College of Science and Humanities Dawadmi, Shaqra University, Shaqra, Saudi Arabia
| | - Ayman Yafoz
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Raed Alsini
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Alaa O Khadidos
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| |
Collapse
|
24
|
Zhou Z, Li X, Ji H, Xu X, Chang Z, Wu K, Song Y, Kao M, Chen H, Wu D, Zhang T. Application of improved Unet network in the recognition and segmentation of lung CT images in patients with pneumoconiosis. BMC Med Imaging 2024; 24:220. [PMID: 39160488 PMCID: PMC11331615 DOI: 10.1186/s12880-024-01377-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Accepted: 07/22/2024] [Indexed: 08/21/2024] Open
Abstract
BACKGROUND Pneumoconiosis has a significant impact on the quality of patient survival. This study aims to evaluate the performance and application value of improved Unet network technology in the recognition and segmentation of lesion areas of lung CT images in patients with pneumoconiosis. METHODS A total of 1212 lung CT images of patients with pneumoconiosis were retrospectively included. The improved Unet network was used to identify and segment the CT image regions of the patients' lungs, and the image data of the granular regions of the lungs were processed by the watershed and region growing algorithms. After random sorting, 848 data were selected into the training set and 364 data into the validation set. The experimental dataset underwent data augmentation and were used for model training and validation to evaluate segmentation performance. The segmentation results were compared with FCN-8s, Unet network (Base), Unet (Squeeze-and-Excitation, SE + Rectified Linear Unit, ReLU), and Unet + + networks. RESULTS In the segmentation of lung CT granular region with the improved Unet network, the four evaluation indexes of Dice similarity coefficient, positive prediction value (PPV), sensitivity coefficient (SC) and mean intersection over union (MIoU) reached 0.848, 0.884, 0.895 and 0.885, respectively, increasing by 7.6%, 13.3%, 3.9% and 6.4%, respectively, compared with those of Unet network (Base), and increasing by 187.5%, 249.4%, 131.9% and 51.0%, respectively, compared with those of FCN-8s, and increasing by 14.0%, 31.2%, 4.7% and 9.7%, respectively, compared with those of Unet network (SE + ReLU), while the segmentation performance was also not inferior to that of the Unet + + network. CONCLUSIONS The improved Unet network proposed shows good performance in the recognition and segmentation of abnormal regions in lung CT images in patients with pneumoconiosis, showing potential application value for assisting clinical decision-making.
Collapse
Affiliation(s)
- Zhengsong Zhou
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Xin Li
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China
| | - Hongbo Ji
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Xuanhan Xu
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Zongqi Chang
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China
| | - Keda Wu
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Yangyang Song
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Mingkun Kao
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Hongjun Chen
- Department of Electronic Information Engineering, Chengdu Jincheng College, Chengdu, China
| | - Dongsheng Wu
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China.
| | - Tao Zhang
- West China School of Public Health and West China Fourth Hospital, Sichuan University, Chengdu, China.
| |
Collapse
|
25
|
Allogmani AS, Mohamed RM, Al-Shibly NM, Ragab M. Enhanced cervical precancerous lesions detection and classification using Archimedes Optimization Algorithm with transfer learning. Sci Rep 2024; 14:12076. [PMID: 38802525 PMCID: PMC11130149 DOI: 10.1038/s41598-024-62773-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Accepted: 05/21/2024] [Indexed: 05/29/2024] Open
Abstract
Cervical cancer (CC) ranks as the fourth most common form of cancer affecting women, manifesting in the cervix. CC is caused by the Human papillomavirus (HPV) infection and is eradicated by vaccinating women from an early age. However, limited medical facilities present a significant challenge in mid- or low-income countries. It can improve the survivability rate and be successfully treated if the CC is detected at earlier stages. Current technological improvements allow for cost-effective, more sensitive, and rapid screening and treatment measures for CC. DL techniques are widely adopted for the automated detection of CC. DL techniques and architectures are used to detect CC and provide higher detection performance. This study offers the design of Enhanced Cervical Precancerous Lesions Detection and Classification using the Archimedes Optimization Algorithm with Transfer Learning (CPLDC-AOATL) algorithm. The CPLDC-AOATL algorithm aims to diagnose cervical cancer using medical images. At the preliminary stage, the CPLDC-AOATL technique involves a bilateral filtering (BF) technique to eliminate the noise in the input images. Besides, the CPLDC-AOATL technique applies the Inception-ResNetv2 model for the feature extraction process, and the use of AOA chose the hyperparameters. The CPLDC-AOATL technique involves a bidirectional long short-term memory (BiLSTM) model for the cancer detection process. The experimental outcome of the CPLDC-AOATL technique emphasized the superior accuracy outcome of 99.53% over other existing approaches under a benchmark dataset.
Collapse
Affiliation(s)
- Ayed S Allogmani
- University of Jeddah, College of Science and Arts at Khulis, Department of Biology, Jeddah, Saudi Arabia
| | - Roushdy M Mohamed
- University of Jeddah, College of Science and Arts at Khulis, Department of Biology, Jeddah, Saudi Arabia.
| | - Nasser M Al-Shibly
- Physiotherapy Department, College of Applied Health Sciences, Jerash University, Jerash, Jordan
| | - Mahmoud Ragab
- Information Technology Department, Faculty of Computing and Information Technology, King Abdulaziz University, 21589, Jeddah, Saudi Arabia.
| |
Collapse
|