1
|
Aruk I, Pacal I, Toprak AN. A novel hybrid ConvNeXt-based approach for enhanced skin lesion classification. EXPERT SYSTEMS WITH APPLICATIONS 2025; 283:127721. [DOI: 10.1016/j.eswa.2025.127721] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
2
|
Ince S, Kunduracioglu I, Algarni A, Bayram B, Pacal I. Deep learning for cerebral vascular occlusion segmentation: A novel ConvNeXtV2 and GRN-integrated U-Net framework for diffusion-weighted imaging. Neuroscience 2025; 574:42-53. [PMID: 40204150 DOI: 10.1016/j.neuroscience.2025.04.010] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2025] [Revised: 03/26/2025] [Accepted: 04/05/2025] [Indexed: 04/11/2025]
Abstract
Cerebral vascular occlusion is a serious condition that can lead to stroke and permanent neurological damage due to insufficient oxygen and nutrients reaching brain tissue. Early diagnosis and accurate segmentation are critical for effective treatment planning. Due to its high soft tissue contrast, Magnetic Resonance Imaging (MRI) is commonly used for detecting these occlusions such as ischemic stroke. However, challenges such as low contrast, noise, and heterogeneous lesion structures in MRI images complicate manual segmentation and often lead to misinterpretations. As a result, deep learning-based Computer-Aided Diagnosis (CAD) systems are essential for faster and more accurate diagnosis and treatment methods, although they can sometimes face challenges such as high computational costs and difficulties in segmenting small or irregular lesions. This study proposes a novel U-Net architecture enhanced with ConvNeXtV2 blocks and GRN-based Multi-Layer Perceptrons (MLP) to address these challenges in cerebral vascular occlusion segmentation. This is the first application of ConvNeXtV2 in this domain. The proposed model significantly improves segmentation accuracy, even in low-contrast regions, while maintaining high computational efficiency, which is crucial for real-world clinical applications. To reduce false positives and improve overall accuracy, small lesions (≤5 pixels) were removed in the preprocessing step with the support of expert clinicians. Experimental results on the ISLES 2022 dataset showed superior performance with an Intersection over Union (IoU) of 0.8015 and a Dice coefficient of 0.8894. Comparative analyses indicate that the proposed model achieves higher segmentation accuracy than existing U-Net variants and other methods, offering a promising solution for clinical use.
Collapse
Affiliation(s)
- Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000 Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000 Igdir, Turkey.
| | - Ali Algarni
- Department of Informatics and Computer Systems, College of Computer Science, King Khalid University, Abha 61421, Saudi Arabia.
| | - Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000 Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000 Igdir, Turkey; Department of Electronics and Information Technologies, Faculty of Architecture and Engineering, Nakhchivan State University, AZ 7012 Nakhchivan, Azerbaijan.
| |
Collapse
|
3
|
Lilhore UK, Sharma YK, Simaiya S, Alroobaea R, Baqasah AM, Alsafyani M, Alhazmi A. SkinEHDLF a hybrid deep learning approach for accurate skin cancer classification in complex systems. Sci Rep 2025; 15:14913. [PMID: 40295588 PMCID: PMC12037884 DOI: 10.1038/s41598-025-98205-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2025] [Accepted: 04/10/2025] [Indexed: 04/30/2025] Open
Abstract
Skin cancer represents a significant global public health issue, and prompt and precise detection is essential for effective treatment. This study introduces SkinEHDLF, an innovative deep-learning model that enhances skin cancer classification. SkinEHDLF utilizes the advantages of several advanced models, i.e., ConvNeXt, EfficientNetV2, and Swin Transformer, while integrating an adaptive attention-based feature fusion mechanism to enhance the synthesis of acquired features. This hybrid methodology combines ConvNeXt's proficient feature extraction capabilities, EfficientNetV2's scalability, and Swin Transformer's long-range attention mechanisms, resulting in a highly accurate and dependable model. The adaptive attention mechanism dynamically optimizes feature fusion, enabling the model to focus on the most relevant information, enhancing accuracy and reducing false positives. We trained and evaluated SkinEHDLF using the ISIC 2024 dataset, which comprises 401,059 skin lesion images extracted from 3D total-body photography. The dataset is divided into three categories: melanoma, benign lesions, and noncancerous skin anomalies. The findings indicate the superiority of SkinEHDLF compared to current models. In binary skin cancer classification, SkinEHDLF surpassed baseline models, achieving an AUROC of 99.8% and an accuracy of 98.76%. The model attained 98.6% accuracy, 97.9% precision, 97.3% recall, and 99.7% AUROC across all lesion categories in multi-class classification. SkinEHDLF demonstrates a 7.9% enhancement in accuracy and a 28% decrease in false positives, outperforming leading models including ResNet-50, EfficientNet-B3, ViT-B16, and hybrid methodologies such as ResNet-50 + EfficientNet and ViT + CNN, thereby positioning itself as a more precise and reliable solution for automated skin cancer detection. These findings underscore SkinEHDLF's capacity to transform dermatological diagnostics by providing a scalable and accurate method for classifying skin cancer.
Collapse
Affiliation(s)
- Umesh Kumar Lilhore
- Department of Computer Science and Engineering, School of Computing Science and Engineering, Galgotias University, Greater Noida, 203201, Uttar Pradesh, India
- Galgotias Multidisciplinary Research & Development Cell (G-MRDC), Galgotias University, Greater Noida, 203201, Uttar Pradesh, India
| | - Yogesh Kumar Sharma
- Department of Computer Science, Koneru Lakshmaiah University Deemed to be University, Vijayawada, 520 002, Andhra Pradesh, India
| | - Sarita Simaiya
- Department of Computer Science and Engineering, School of Computing Science and Engineering, Galgotias University, Greater Noida, 203201, Uttar Pradesh, India.
- Arba Minch University, Arba Minch, Ethiopia.
| | - Roobaea Alroobaea
- Department of Computer Science, College of Computers and Information Technology, Taif University, P. O. Box 11099, Taif, 21944, Saudi Arabia
| | - Abdullah M Baqasah
- Department of Information Technology, College of Computers and Information Technology, Taif University, P. O. Box 11099, Taif, 21974, Saudi Arabia
| | - Majed Alsafyani
- Department of Computer Science, College of Computers and Information Technology, Taif University, P. O. Box 11099, Taif, 21944, Saudi Arabia
| | - Afnan Alhazmi
- Department of Information Technology, College of Computers and Information Technology, Taif University, P. O. Box 11099, Taif, 21974, Saudi Arabia
| |
Collapse
|
4
|
Rabee A, Anwar Z, AbdelMoety A, Abdelsallam A, Ali M. Comparative analysis of automated foul detection in football using deep learning architectures. Sci Rep 2025; 15:14236. [PMID: 40274843 PMCID: PMC12022285 DOI: 10.1038/s41598-025-96945-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2025] [Accepted: 04/01/2025] [Indexed: 04/26/2025] Open
Abstract
Automated foul detection in football represents a challenging task due to the dynamic nature of the game, the variability in player movements, and the ambiguity in differentiating fouls from regular physical contact. This study presents a comprehensive comparative evaluation of eight state-of-the-art Deep Learning (DL) architectures - EfficientNetV2, ResNet50, VGG16, Xception, InceptionV3, MobileNetV2, InceptionResNetV2, and DenseNet121 - applied to the task of automated foul detection in football. The models were trained and evaluated using a curated dataset comprising 7000 images, which was split into 70% for training (4,900 images), 20% for validation (1,400 images), and 10% for testing (700 images). To ensure fair evaluation, the test set was balanced to contain 350 images depicting foul events and 350 images representing non-foul scenarios, although perfect balance was subject to class distribution constraints. Performance was assessed across multiple metrics, including test accuracy, precision, recall, F1-score, and Area Under the Receiver Operating Characteristic Curve (AUC). The results demonstrate that InceptionResNetV2 achieved the highest test accuracy of 87.57% and a strong F1-score of 0.8966, closely followed by DenseNet121, which attained the highest precision of 0.9786 and an AUC of 0.9641, indicating superior discriminatory power. Lightweight models such as MobileNetV2 also performed competitively, highlighting their potential for real-time deployment. The findings highlight the strengths and trade-offs between model complexity, accuracy, and generalizability, underscoring the viability of integrating DL architectures into existing football officiating systems, such as the Video Assistant Referee (VAR). Furthermore, the study emphasizes the importance of model explainability through techniques such as Gradient-weighted Class Activation Mapping++ (GradCAM++), ensuring that automated decisions can be accompanied by interpretable visual evidence. This comparative evaluation serves as a foundation for future research aimed at enhancing real-time foul detection through multimodal data fusion, temporal modeling, and improved domain adaptation techniques.
Collapse
Affiliation(s)
- Abdallah Rabee
- Sports Health Sciences Department, Faculty of Physical Education, South Valley University, Qena, 1464091, Egypt
| | - Zakaria Anwar
- Sports Training and Kinesiology Sciences Department, Faculty of Physical Education, Suez University, Suez, 43512, Egypt
| | - Ahmed AbdelMoety
- Electrical Engineering Department, Faculty of Engineering, South Valley University, Qena, 83523, Egypt
| | - Ahmed Abdelsallam
- Sports Health Sciences Department, Faculty of Physical Education, South Valley University, Qena, 1464091, Egypt
| | - Mahmoud Ali
- Sports Health Sciences Department, Faculty of Physical Education, South Valley University, Qena, 1464091, Egypt.
| |
Collapse
|
5
|
Zubair M, Owais M, Hassan T, Bendechache M, Hussain M, Hussain I, Werghi N. An interpretable framework for gastric cancer classification using multi-channel attention mechanisms and transfer learning approach on histopathology images. Sci Rep 2025; 15:13087. [PMID: 40240457 PMCID: PMC12003787 DOI: 10.1038/s41598-025-97256-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2024] [Accepted: 04/03/2025] [Indexed: 04/18/2025] Open
Abstract
The importance of gastric cancer (GC) and the role of deep learning techniques in categorizing GC histopathology images have recently increased. Identifying the drawbacks of traditional deep learning models, including lack of interpretability, inability to capture complex patterns, lack of adaptability, and sensitivity to noise. A multi-channel attention mechanism-based framework is proposed that can overcome the limitations of conventional deep learning models by dynamically focusing on relevant features, enhancing extraction, and capturing complex relationships in medical data. The proposed framework uses three different attention mechanism channels and convolutional neural networks to extract multichannel features during the classification process. The proposed framework's strong performance is confirmed by competitive experiments conducted on a publicly available Gastric Histopathology Sub-size Image Database, which yielded remarkable classification accuracies of 99.07% and 98.48% on the validation and testing sets, respectively. Additionally, on the HCRF dataset, the framework achieved high classification accuracy of 99.84% and 99.65% on the validation and testing sets, respectively. The effectiveness and interchangeability of the three channels are further confirmed by ablation and interchangeability experiments, highlighting the remarkable performance of the framework in GC histopathological image classification tasks. This offers an advanced and pragmatic artificial intelligence solution that addresses challenges posed by unique medical image characteristics for intricate image analysis. The proposed approach in artificial intelligence medical engineering demonstrates significant potential for enhancing diagnostic precision by achieving high classification accuracy and treatment outcomes.
Collapse
Affiliation(s)
- Muhammad Zubair
- Interdisciplinary Research Center for Finance and Digital Economy, King Fahd University of Petroleum and Minerals, 31261, Dhahran, Saudi Arabia
| | - Muhammad Owais
- Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates.
| | - Taimur Hassan
- Departement of Electrical and Computer Engineering, Abu Dhabi University, Abu Dhabi, United Arab Emirates
| | - Malika Bendechache
- ADAPT Research Centre, School of Computer Science, University of Galway, H91 TK33, Galway, Ireland
| | - Muzammil Hussain
- Department of Software Engineering, Faculty of Information Technology, Al-Ahliyya Amman University, Amman, Jordan
| | - Irfan Hussain
- Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates
| | - Naoufel Werghi
- Department of Computer Science, Khalifa University, Abu Dhabi, United Arab Emirates
| |
Collapse
|
6
|
İnce S, Kunduracioglu I, Bayram B, Pacal I. U-Net-Based Models for Precise Brain Stroke Segmentation. CHAOS THEORY AND APPLICATIONS 2025; 7:50-60. [DOI: 10.51537/chaos.1605529] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
Abstract
Ischemic stroke, a widespread neurological condition with a substantial mortality rate, necessitates accurate delineation of affected regions to enable proper evaluation of patient outcomes. However, such precision is complicated by factors like variable lesion sizes, noise interference, and the overlapping intensity characteristics of different tissue structures. This research addresses these issues by focusing on the segmentation of Diffusion Weighted Imaging (DWI) scans from the ISLES 2022 dataset and conducting a comparative assessment of three advanced deep learning models: the U-Net framework, its U-Net++ extension, and the Attention U-Net. Applying consistent evaluation criteria specifically, Intersection over Union (IoU), Dice Similarity Coefficient (DSC), and recall the Attention U-Net emerged as the superior choice, establishing record high values for IoU (0.8223) and DSC (0.9021). Although U-Net achieved commendable recall, its performance lagged behind that of U-Net++ in other critical measures. These findings underscore the value of integrating attention mechanisms to achieve more precise segmentation. Moreover, they highlight that the Attention U-Net model is a reliable candidate for medical imaging tasks where both accuracy and efficiency hold paramount importance, while U Net and U Net++ may still prove suitable in certain niche scenarios.
Collapse
Affiliation(s)
- Suat İnce
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital
| | | | - Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital
| | - Ishak Pacal
- IGDIR UNIVERSITY, FACULTY OF ENGINEERING, DEPARTMENT OF COMPUTER ENGINEERING, COMPUTER ENGINEERING PR
| |
Collapse
|
7
|
Bayram B, Kunduracioglu I, Ince S, Pacal I. A systematic review of deep learning in MRI-based cerebral vascular occlusion-based brain diseases. Neuroscience 2025; 568:76-94. [PMID: 39805420 DOI: 10.1016/j.neuroscience.2025.01.020] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 01/09/2025] [Accepted: 01/10/2025] [Indexed: 01/16/2025]
Abstract
Neurological disorders, including cerebral vascular occlusions and strokes, present a major global health challenge due to their high mortality rates and long-term disabilities. Early diagnosis, particularly within the first hours, is crucial for preventing irreversible damage and improving patient outcomes. Although neuroimaging techniques like magnetic resonance imaging (MRI) have advanced significantly, traditional methods often fail to fully capture the complexity of brain lesions. Deep learning has recently emerged as a powerful tool in medical imaging, offering high accuracy in detecting and segmenting brain anomalies. This review examines 61 MRI-based studies published between 2020 and 2024, focusing on the role of deep learning in diagnosing cerebral vascular occlusion-related conditions. It evaluates the successes and limitations of these studies, including the adequacy and diversity of datasets, and addresses challenges such as data privacy and algorithm explainability. Comparisons between convolutional neural network (CNN)-based and Vision Transformer (ViT)-based approaches reveal distinct advantages and limitations. The findings emphasize the importance of ethically secure frameworks, the inclusion of diverse datasets, and improved model interpretability. Advanced architectures like U-Net variants and transformer-based models are highlighted as promising tools to enhance reliability in clinical applications. By automating complex neuroimaging tasks and improving diagnostic accuracy, deep learning facilitates personalized treatment strategies. This review provides a roadmap for integrating technical advancements into clinical practice, underscoring the transformative potential of deep learning in managing neurological disorders and improving healthcare outcomes globally.
Collapse
Affiliation(s)
- Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| | - Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| |
Collapse
|
8
|
Dorathi Jayaseeli JD, Briskilal J, Fancy C, Vaitheeshwaran V, Patibandla RSML, Syed K, Swain AK. An intelligent framework for skin cancer detection and classification using fusion of Squeeze-Excitation-DenseNet with Metaheuristic-driven ensemble deep learning models. Sci Rep 2025; 15:7425. [PMID: 40033075 PMCID: PMC11876321 DOI: 10.1038/s41598-025-92293-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2024] [Accepted: 02/26/2025] [Indexed: 03/05/2025] Open
Abstract
Skin cancer is the most dominant and critical method of cancer, which arises all over the world. Its damaging effects can range from disfigurement to major medical expenditures and even death if not analyzed and preserved timely. Conventional models of skin cancer recognition require a complete physical examination by a specialist, which is time-wasting in a few cases. Computer-aided medicinal analytical methods have gained massive popularity due to their efficiency and effectiveness. This model can assist dermatologists in the initial recognition of skin cancer, which is significant for early diagnosis. An automatic classification model utilizing deep learning (DL) can help doctors perceive the kind of skin lesion and improve the patient's health. The classification of skin cancer is one of the hot topics in the research field, along with the development of DL structure. This manuscript designs and develops a Detection of Skin Cancer Using an Ensemble Deep Learning Model and Gray Wolf Optimization (DSC-EDLMGWO) method. The proposed DSC-EDLMGWO model relies on the recognition and classification of skin cancer in biomedical imaging. The presented DSC-EDLMGWO model initially involves the image preprocessing stage at two levels: contract enhancement using the CLAHE method and noise removal using the wiener filter (WF) model. Furthermore, the proposed DSC-EDLMGWO model utilizes the SE-DenseNet method, which is the fusion of the squeeze-and-excitation (SE) module and DenseNet to extract features. For the classification process, the ensemble of DL models, namely the long short-term memory (LSTM) technique, extreme learning machine (ELM) model, and stacked sparse denoising autoencoder (SSDA) method, is employed. Finally, the gray wolf optimization (GWO) method optimally adjusts the ensemble DL models' hyperparameter values, resulting in more excellent classification performance. The effectiveness of the DSC-EDLMGWO approach is evaluated using a benchmark image database, with outcomes measured across various performance metrics. The experimental validation of the DSC-EDLMGWO approach portrayed a superior accuracy value of 98.38% and 98.17% under HAM10000 and ISIC datasets across other techniques.
Collapse
Affiliation(s)
- J D Dorathi Jayaseeli
- Department of Computing Technologies, SRM Institute of Science and Technology, Kattankulathur, 603203, India
| | - J Briskilal
- Department of Computing Technologies, Faculty of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - C Fancy
- Department of Networking and Communications, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - V Vaitheeshwaran
- Department of Computer Science and Engineering, Aditya University, Surampalem, Kakinada, India
| | - R S M Lakshmi Patibandla
- Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Guntur, Andhra Pradesh, India
| | - Khasim Syed
- School of Computer Science & Engineering, VIT - AP University, Amaravati, 522237, Andhra Pradesh, India.
| | - Anil Kumar Swain
- KIIT Deemed to be University, Bhubaneswar, 751024, Odisha, India
| |
Collapse
|
9
|
Alzahrani AA, Alsamri J, Maashi M, Negm N, Asklany SA, Alkharashi A, Alkhiri H, Obayya M. Deep structured learning with vision intelligence for oral carcinoma lesion segmentation and classification using medical imaging. Sci Rep 2025; 15:6610. [PMID: 39994267 PMCID: PMC11850820 DOI: 10.1038/s41598-025-89971-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2024] [Accepted: 02/10/2025] [Indexed: 02/26/2025] Open
Abstract
Oral carcinoma (OC) is a toxic illness among the most general malignant cancers globally, and it has developed a gradually significant public health concern in emerging and low-to-middle-income states. Late diagnosis, high incidence, and inadequate treatment strategies remain substantial challenges. Analysis at an initial phase is significant for good treatment, prediction, and existence. Despite the current growth in the perception of molecular devices, late analysis and methods near precision medicine for OC patients remain a challenge. A machine learning (ML) model was employed to improve early detection in medicine, aiming to reduce cancer-specific mortality and disease progression. Recent advancements in this approach have significantly enhanced the extraction and diagnosis of critical information from medical images. This paper presents a Deep Structured Learning with Vision Intelligence for Oral Carcinoma Lesion Segmentation and Classification (DSLVI-OCLSC) model for medical imaging. Using medical imaging, the DSLVI-OCLSC model aims to enhance OC's classification and recognition outcomes. To accomplish this, the DSLVI-OCLSC model utilizes wiener filtering (WF) as a pre-processing technique to eliminate the noise. In addition, the ShuffleNetV2 method is used for the group of higher-level deep features from an input image. The convolutional bidirectional long short-term memory network with a multi-head attention mechanism (MA-CNN-BiLSTM) approach is utilized for oral carcinoma recognition and identification. Moreover, the Unet3 + is employed to segment abnormal regions from the classified images. Finally, the sine cosine algorithm (SCA) approach is utilized to hyperparameter-tune the DL model. A wide range of simulations is implemented to ensure the enhanced performance of the DSLVI-OCLSC method under the OC images dataset. The experimental analysis of the DSLVI-OCLSC method portrayed a superior accuracy value of 98.47% over recent approaches.
Collapse
Affiliation(s)
- Ahmad A Alzahrani
- Department of Computer Science and Artificial Intelligence, College of Computing, Umm-AlQura University, Mecca, Saudi Arabia
| | - Jamal Alsamri
- Department of Biomedical Engineering, College of Engineering, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, PO Box 103786, 11543, Riyadh, Saudi Arabia
| | - Noha Negm
- Department of Computer Science, Applied College at Mahayil, King Khalid University, Abha, Saudi Arabia
| | - Somia A Asklany
- Department of Computer Science and Information Technology, Faculty of Sciences and Arts, Turaif, Northern Border University, 91431, Arar, Saudi Arabia.
| | - Abdulwhab Alkharashi
- Department of Computer Science, College of Computing and Informatics, Saudi Electronic University, Riyadh, Saudi Arabia
| | - Hassan Alkhiri
- Department of Computer Science, Faculty of Computing and Information Technology, Al-Baha University, Al-Baha, Saudi Arabia
| | - Marwa Obayya
- Department of Computer Science and Artificial Intelligence, College of Computing, Umm-AlQura University, Mecca, Saudi Arabia
| |
Collapse
|
10
|
Ozdemir B, Pacal I. A robust deep learning framework for multiclass skin cancer classification. Sci Rep 2025; 15:4938. [PMID: 39930026 PMCID: PMC11811178 DOI: 10.1038/s41598-025-89230-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2024] [Accepted: 02/04/2025] [Indexed: 02/13/2025] Open
Abstract
Skin cancer represents a significant global health concern, where early and precise diagnosis plays a pivotal role in improving treatment efficacy and patient survival rates. Nonetheless, the inherent visual similarities between benign and malignant lesions pose substantial challenges to accurate classification. To overcome these obstacles, this study proposes an innovative hybrid deep learning model that combines ConvNeXtV2 blocks and separable self-attention mechanisms, tailored to enhance feature extraction and optimize classification performance. The inclusion of ConvNeXtV2 blocks in the initial two stages is driven by their ability to effectively capture fine-grained local features and subtle patterns, which are critical for distinguishing between visually similar lesion types. Meanwhile, the adoption of separable self-attention in the later stages allows the model to selectively prioritize diagnostically relevant regions while minimizing computational complexity, addressing the inefficiencies often associated with traditional self-attention mechanisms. The model was comprehensively trained and validated on the ISIC 2019 dataset, which includes eight distinct skin lesion categories. Advanced methodologies such as data augmentation and transfer learning were employed to further enhance model robustness and reliability. The proposed architecture achieved exceptional performance metrics, with 93.48% accuracy, 93.24% precision, 90.70% recall, and a 91.82% F1-score, outperforming over ten Convolutional Neural Network (CNN) based and over ten Vision Transformer (ViT) based models tested under comparable conditions. Despite its robust performance, the model maintains a compact design with only 21.92 million parameters, making it highly efficient and suitable for model deployment. The Proposed Model demonstrates exceptional accuracy and generalizability across diverse skin lesion classes, establishing a reliable framework for early and accurate skin cancer diagnosis in clinical practice.
Collapse
Affiliation(s)
- Burhanettin Ozdemir
- Department of Operations and Project Management, College of Business, Alfaisal University, Riyadh, 11533, Saudi Arabia.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, Igdir, 76000, Turkey
- Department of Electronics and Information Technologies, Faculty of Architecture and Engineering, Nakhchivan State University, AZ 7012, Nakhchivan, Azerbaijan
| |
Collapse
|
11
|
Ozdemir B, Aslan E, Pacal I. Attention Enhanced InceptionNeXt-Based Hybrid Deep Learning Model for Lung Cancer Detection. IEEE ACCESS 2025; 13:27050-27069. [DOI: 10.1109/access.2025.3539122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
Affiliation(s)
- Burhanettin Ozdemir
- Department of Operations and Project Management, College of Business, Alfaisal University, Riyadh, Saudi Arabia
| | - Emrah Aslan
- Department of Computer Engineering, Faculty of Engineering and Architecture, Mardin Artuklu University, Mardin, Türkiye
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, Iğdır, Türkiye
| |
Collapse
|