1
|
Wang Z, Fu S, Zhang H, Wang C, Xia C, Hou P, Shun C, Shun G. Dual-branch dynamic hierarchical U-Net with multi-layer space fusion attention for medical image segmentation. Sci Rep 2025; 15:8194. [PMID: 40065006 PMCID: PMC11894187 DOI: 10.1038/s41598-025-92715-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Accepted: 03/03/2025] [Indexed: 03/14/2025] Open
Abstract
Accurate segmentation of organs or lesions from medical images is essential for accurate disease diagnosis and organ morphometrics. Previously, most researchers mainly added feature extraction modules and simply aggregated the semantic features to U-Net network to improve the segmentation accuracy of medical images. However, these improved U-Net networks ignore the semantic differences of different organs in medical images and lack the fusion of high-level semantic features and low-level semantic features, which will lead to blurred or miss boundaries between similar organs and diseased areas. To solve this problem, we propose Dual-branch dynamic hierarchical U-Net with multi-layer space fusion attention (D2HU-Net). Firstly, we propose a multi-layer spatial attention fusion module, which makes the shallow decoding path provide predictive graph supplement to the deep decoding path. Under the guidance of higher semantic features, useful context features are selected from lower semantic features to obtain deeper useful spatial information, which makes up for the semantic differences between organs in different medical images. Secondly, we propose a dynamic multi-scale layered module that enhances the multi-scale representation of the network at a finer granularity level and selectively refines single-scale features. Finally, the network provides guiding optimization for subsequent decoding based on multi-scale loss functions. The experimental results on four medical data sets show D2HU-Net enables the most advanced segmentation capabilities on different medical image datasets, which can help doctors diagnose and treat diseases.
Collapse
Affiliation(s)
- Zhen Wang
- School of Public Health, Qiqihar Medical University, Qiqihar, 161003, China
| | - Shuang Fu
- College of Pharmacy, Qiqihar Medical University, Qiqihar, 161003, China.
| | - Hongguang Zhang
- College of Pharmacy, Qiqihar Medical University, Qiqihar, 161003, China
| | | | - Chunhui Xia
- College of Pharmacy, Qiqihar Medical University, Qiqihar, 161003, China
| | - Pen Hou
- College of Pharmacy, Qiqihar Medical University, Qiqihar, 161003, China
| | - Chunxue Shun
- College of Pharmacy, Qiqihar Medical University, Qiqihar, 161003, China
| | - Ge Shun
- College of Pharmacy, Qiqihar Medical University, Qiqihar, 161003, China
| |
Collapse
|
2
|
Al-Mnayyis AM, Gharaibeh H, Amin M, Anakreh D, Akhdar HF, Alshdaifat EH, Nahar KMO, Nasayreh A, Gharaibeh M, Alsalman N, Alomar A, Gharaibeh M, Abu Mhanna HY. (KAUH-BCMD) dataset: advancing mammographic breast cancer classification with multi-fusion preprocessing and residual depth-wise network. Front Big Data 2025; 8:1529848. [PMID: 40115240 PMCID: PMC11922913 DOI: 10.3389/fdata.2025.1529848] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2024] [Accepted: 02/17/2025] [Indexed: 03/23/2025] Open
Abstract
The categorization of benign and malignant patterns in digital mammography is a critical step in the diagnosis of breast cancer, facilitating early detection and potentially saving many lives. Diverse breast tissue architectures often obscure and conceal breast issues. Classifying worrying regions (benign and malignant patterns) in digital mammograms is a significant challenge for radiologists. Even for specialists, the first visual indicators are nuanced and irregular, complicating identification. Therefore, radiologists want an advanced classifier to assist in identifying breast cancer and categorizing regions of concern. This study presents an enhanced technique for the classification of breast cancer using mammography images. The collection comprises real-world data from King Abdullah University Hospital (KAUH) at Jordan University of Science and Technology, consisting of 7,205 photographs from 5,000 patients aged 18-75. After being classified as benign or malignant, the pictures underwent preprocessing by rescaling, normalization, and augmentation. Multi-fusion approaches, such as high-boost filtering and contrast-limited adaptive histogram equalization (CLAHE), were used to improve picture quality. We created a unique Residual Depth-wise Network (RDN) to enhance the precision of breast cancer detection. The suggested RDN model was compared with many prominent models, including MobileNetV2, VGG16, VGG19, ResNet50, InceptionV3, Xception, and DenseNet121. The RDN model exhibited superior performance, achieving an accuracy of 97.82%, precision of 96.55%, recall of 99.19%, specificity of 96.45%, F1 score of 97.85%, and validation accuracy of 96.20%. The findings indicate that the proposed RDN model is an excellent instrument for early diagnosis using mammography images and significantly improves breast cancer detection when integrated with multi-fusion and efficient preprocessing approaches.
Collapse
Affiliation(s)
| | - Hasan Gharaibeh
- Artificial Intelligence and Data Innovation Office, King Hussein Cancer Center, Amman, Jordan
| | - Mohammad Amin
- Computer Science Department, Faculty of Information Technology and Computer Sciences, Yarmouk University, Irbid, Jordan
| | - Duha Anakreh
- Department of Medicine, Faculty of Medicine, Jordan University of Science and Technology, Irbid, Jordan
| | - Hanan Fawaz Akhdar
- Physics Department, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - Eman Hussein Alshdaifat
- Department of Obstetrics and Gynecology, Faculty of Medicine, Yarmouk University, Irbid, Jordan
| | - Khalid M O Nahar
- Computer Science Department, Faculty of Information Technology and Computer Sciences, Yarmouk University, Irbid, Jordan
| | - Ahmad Nasayreh
- Artificial Intelligence and Data Innovation Office, King Hussein Cancer Center, Amman, Jordan
| | - Mohammad Gharaibeh
- Department of Medicine, Faculty of Medicine, Hashemite University, Zarqa, Jordan
| | - Neda'a Alsalman
- Department of Computer Science, Faculty of Information Technology, Jordan University of Science and Technology, Irbid, Jordan
| | - Alaa Alomar
- Department of Computer Science, Faculty of Information Technology, Jordan University of Science and Technology, Irbid, Jordan
| | - Maha Gharaibeh
- Radiology Department, Jordan University of Science and Technology, Irbid, Jordan
| | | |
Collapse
|
3
|
Wang M, Liu R, Luttrell IV J, Zhang C, Xie J. Detection of Masses in Mammogram Images Based on the Enhanced RetinaNet Network With INbreast Dataset. J Multidiscip Healthc 2025; 18:675-695. [PMID: 39935433 PMCID: PMC11812562 DOI: 10.2147/jmdh.s493873] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2024] [Accepted: 01/10/2025] [Indexed: 02/13/2025] Open
Abstract
Purpose Breast cancer is the most common major public health problems of women in the world. Until now, analyzing mammogram images is still the main method used by doctors to diagnose and detect breast cancers. However, this process usually depends on the experience of radiologists and is always very time consuming. Patients and Methods We propose to introduce deep learning technology into the process for the facilitation of computer-aided diagnosis (CAD), and address the challenges of class imbalance, enhance the detection of small masses and multiple targets, and reduce false positives and negatives in mammogram analysis. Therefore, we adopted and enhanced RetinaNet to detect masses in mammogram images. Specifically, we introduced a novel modification to the network structure, where the feature map M5 is processed by the ReLU function prior to the original convolution kernel. This strategic adjustment was designed to prevent the loss of resolution for small mass features. Additionally, we introduced transfer learning techniques into training process through leveraging pre-trained weights from other RetinaNet applications, and fine-tuned our improved model using the INbreast dataset. Results The aforementioned innovations facilitate superior performance of the enhanced RetiaNet model on the public dataset INbreast, as evidenced by a mAP (mean average precision) of 1.0000 and TPR (true positive rate) of 1.00 at 0.00 FPPI (false positive per image) on the INbreast dataset. Conclusion The experimental results demonstrate that our enhanced RetinaNet model defeats the existing models by having more generalization performance than other published studies, and it can also be applied to other types of patients to assist doctors in making a proper diagnosis.
Collapse
Affiliation(s)
- Mingzhao Wang
- School of Computer Science, Shaanxi Normal University, Xian, People’s Republic of China
| | - Ran Liu
- School of Computer Science, Shaanxi Normal University, Xian, People’s Republic of China
| | - Joseph Luttrell IV
- School of Computing Sciences and Computer Engineering, University of Southern Mississippi, Hattiesburg, MS, USA
| | - Chaoyang Zhang
- School of Computing Sciences and Computer Engineering, University of Southern Mississippi, Hattiesburg, MS, USA
| | - Juanying Xie
- School of Computer Science, Shaanxi Normal University, Xian, People’s Republic of China
| |
Collapse
|
4
|
Sha Y, Zhang Q, Zhai X, Hou M, Lu J, Meng W, Wang Y, Li K, Ma J. CerviFusionNet: A multi-modal, hybrid CNN-transformer-GRU model for enhanced cervical lesion multi-classification. iScience 2024; 27:111313. [PMID: 39634563 PMCID: PMC11615576 DOI: 10.1016/j.isci.2024.111313] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2024] [Revised: 06/10/2024] [Accepted: 10/30/2024] [Indexed: 12/07/2024] Open
Abstract
Cervical lesions pose a significant threat to women's health worldwide. Colposcopy is essential for screening and treating cervical lesions, but its effectiveness depends on the doctor's experience. Artificial intelligence-based solutions via colposcopy images have shown great potential in cervical lesions screening. However, some challenges still need to be addressed, such as low algorithm performance and lack of high-quality multi-modal datasets. Here, we established a multi-modal colposcopy dataset of 2,273 HPV+ patients, comprising original colposcopy images, acetic acid reactions at 60s and 120s, iodine staining, diagnostic reports, and pathological results. Utilizing this dataset, we developed CerviFusionNet, a hybrid architecture that merges convolutional neural networks and vision transformers to learn robust representations. We designed a temporal module to capture dynamic changes in acetic acid sequences, which can boost the model performance without sacrificing inference speed. Compared with several existing methods, CerviFusionNet demonstrated excellent accuracy and efficiency.
Collapse
Affiliation(s)
- Yuyang Sha
- Center for Artificial Intelligence Driven Drug Discovery, Faculty of Applied Sciences, Macao Polytechnic University, Macau SAR 999078, China
| | - Qingyue Zhang
- First Teaching Hospital of Tianjin University of Traditional Chinese Medicine, Tianjin 300381, China
- National Clinical Research Center for Chinese Medicine Acupuncture and Moxibustion, Tianjin 300381, China
| | - Xiaobing Zhai
- Center for Artificial Intelligence Driven Drug Discovery, Faculty of Applied Sciences, Macao Polytechnic University, Macau SAR 999078, China
| | - Menghui Hou
- First Teaching Hospital of Tianjin University of Traditional Chinese Medicine, Tianjin 300381, China
- National Clinical Research Center for Chinese Medicine Acupuncture and Moxibustion, Tianjin 300381, China
| | - Jingtao Lu
- Beijing University of Technology, School of Mathematical Statistics and Mechanics, Beijing 100124, China
| | - Weiyu Meng
- Center for Artificial Intelligence Driven Drug Discovery, Faculty of Applied Sciences, Macao Polytechnic University, Macau SAR 999078, China
| | - Yuefei Wang
- National Key Laboratory of Chinese Medicine Modernization, State Key Laboratory of Component-based Chinese Medicine, Tianjin University of Traditional Chinese Medicine, Tianjin 301617, China
- Haihe Laboratory of Modern Chinese Medicine, Tianjin 301617, China
| | - Kefeng Li
- Center for Artificial Intelligence Driven Drug Discovery, Faculty of Applied Sciences, Macao Polytechnic University, Macau SAR 999078, China
| | - Jing Ma
- First Teaching Hospital of Tianjin University of Traditional Chinese Medicine, Tianjin 300381, China
- National Clinical Research Center for Chinese Medicine Acupuncture and Moxibustion, Tianjin 300381, China
| |
Collapse
|
5
|
Sun Y, Li P, Xu H, Wang R. Structural prior-driven feature extraction with gradient-momentum combined optimization for convolutional neural network image classification. Neural Netw 2024; 179:106511. [PMID: 39146718 DOI: 10.1016/j.neunet.2024.106511] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Revised: 06/12/2024] [Accepted: 07/03/2024] [Indexed: 08/17/2024]
Abstract
Recent image classification efforts have achieved certain success by incorporating prior information such as labels and logical rules to learn discriminative features. However, these methods overlook the variability of features, resulting in feature inconsistency and fluctuations in model parameter updates, which further contribute to decreased image classification accuracy and model instability. To address this issue, this paper proposes a novel method combining structural prior-driven feature extraction with gradient-momentum (SPGM), from the perspectives of consistent feature learning and precise parameter updates, to enhance the accuracy and stability of image classification. Specifically, SPGM leverages a structural prior-driven feature extraction (SPFE) approach to calculate gradients of multi-level features and original images to construct structural information, which is then transformed into prior knowledge to drive the network to learn features consistent with the original images. Additionally, an optimization strategy integrating gradients and momentum (GMO) is introduced, dynamically adjusting the direction and step size of parameter updates based on the angle and norm of the sum of gradients and momentum, enabling precise model parameter updates. Extensive experiments on CIFAR10 and CIFAR100 datasets demonstrate that the SPGM method significantly reduces the top-1 error rate in image classification, enhances the classification performance, and outperforms state-of-the-art methods.
Collapse
Affiliation(s)
- Yunyun Sun
- School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing, 210023, Jiangsu, China.
| | - Peng Li
- School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing, 210023, Jiangsu, China; Jiangsu High Technology Research Key Laboratory for Wireless Sensor Networks, Nanjing, 210023, Jiangsu, China.
| | - He Xu
- School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing, 210023, Jiangsu, China; Jiangsu High Technology Research Key Laboratory for Wireless Sensor Networks, Nanjing, 210023, Jiangsu, China.
| | - Ruchuan Wang
- School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing, 210023, Jiangsu, China; Jiangsu High Technology Research Key Laboratory for Wireless Sensor Networks, Nanjing, 210023, Jiangsu, China.
| |
Collapse
|
6
|
Esfandiari A, Nasiri N. Gene selection and cancer classification using interaction-based feature clustering and improved-binary Bat algorithm. Comput Biol Med 2024; 181:109071. [PMID: 39205342 DOI: 10.1016/j.compbiomed.2024.109071] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2024] [Revised: 08/13/2024] [Accepted: 08/22/2024] [Indexed: 09/04/2024]
Abstract
In high-dimensional gene expression data, selecting an optimal subset of genes is crucial for achieving high classification accuracy and reliable diagnosis of diseases. This paper proposes a two-stage hybrid model for gene selection based on clustering and a swarm intelligence algorithm to identify the most informative genes with high accuracy. First, a clustering-based multivariate filter approach is performed to explore the interactions between the features and eliminate any redundant or irrelevant ones. Then, by controlling for the problem of premature convergence in the binary Bat algorithm, the optimal gene subset is determined using different classifiers with the Monte Carlo cross-validation data partitioning model. The effectiveness of our proposed framework is evaluated using eight gene expression datasets, by comparison with other recently published algorithms in the literature. Experiments confirm that in seven out of eight datasets, the proposed method can achieve superior results in terms of classification accuracy and gene subset size. In particular, it achieves a classification accuracy of 100% in Lymphoma and Ovarian datasets and above 97.4% in the rest with a minimum number of genes. The results demonstrate that our proposed algorithm has the potential to solve the feature selection problem in different applications with high-dimensional datasets.
Collapse
Affiliation(s)
- Ahmad Esfandiari
- Department of Computer Engineering, Sari Branch, Islamic Azad University, Sari, Iran.
| | - Niki Nasiri
- Pediatric Infectious Diseases Research Center, Communicable Diseases Institute, Mazandaran University of Medical Sciences, Sari, Iran
| |
Collapse
|
7
|
Ponrani MA, Anand M, Alsaadi M, Dutta AK, Fayaz R, Mathew S, Chaurasia MA, Sunila, Bhende M. Brain-computer interfaces inspired spiking neural network model for depression stage identification. J Neurosci Methods 2024; 409:110203. [PMID: 38880343 DOI: 10.1016/j.jneumeth.2024.110203] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2024] [Revised: 05/30/2024] [Accepted: 06/13/2024] [Indexed: 06/18/2024]
Abstract
BACKGROUND Depression is a global mental disorder, and traditional diagnostic methods mainly rely on scales and subjective evaluations by doctors, which cannot effectively identify symptoms and even carry the risk of misdiagnosis. Brain-Computer Interfaces inspired deep learning-assisted diagnosis based on physiological signals holds promise for improving traditional methods lacking physiological basis and leads next generation neuro-technologies. However, traditional deep learning methods rely on immense computational power and mostly involve end-to-end network learning. These learning methods also lack physiological interpretability, limiting their clinical application in assisted diagnosis. METHODOLOGY A brain-like learning model for diagnosing depression using electroencephalogram (EEG) is proposed. The study collects EEG data using 128-channel electrodes, producing a 128×128 brain adjacency matrix. Given the assumption of undirected connectivity, the upper half of the 128×128 matrix is chosen in order to minimise the input parameter size, producing 8,128-dimensional data. After eliminating 28 components derived from irrelevant or reference electrodes, a 90×90 matrix is produced, which can be used as an input for a single-channel brain-computer interface image. RESULT At the functional level, a spiking neural network is constructed to classify individuals with depression and healthy individuals, achieving an accuracy exceeding 97.5 %. COMPARISON WITH EXISTING METHODS Compared to deep convolutional methods, the spiking method reduces energy consumption. CONCLUSION At the structural level, complex networks are utilized to establish spatial topology of brain connections and analyse their graph features, identifying potential abnormal brain functional connections in individuals with depression.
Collapse
Affiliation(s)
- M Angelin Ponrani
- Department of ECE, St. Joseph's College of Engineering, Chennai -119, India.
| | - Monika Anand
- Computer Science & Engineering, Chandigarh University, Mohali, India
| | - Mahmood Alsaadi
- Department of computer science, Al-Maarif University College, Al Anbar 31001, Iraq
| | - Ashit Kumar Dutta
- Department of Computer Science and Information Systems, College of Applied Sciences, AlMaarefa University, Ad Diriyah, Riyadh 13713, Saudi Arabia
| | - Roma Fayaz
- Dapartmemt of computer science, college of computer science and information technology, Jazan university, Jazan, Saudi Arabia
| | | | - Mousmi Ajay Chaurasia
- Dept of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Hyderabad, India
| | - Sunila
- Guru Jambheshwar University of Science and Technology, Hisar, Haryana, India
| | - Manisha Bhende
- Dr. D. Y. Patil Vidyapeeth, Pune, Dr. D. Y. Patil School of Science & Technology, Tathawade, Pune, India
| |
Collapse
|
8
|
Lamba K, Rani S. A novel approach of brain-computer interfacing (BCI) and Grad-CAM based explainable artificial intelligence: Use case scenario for smart healthcare. J Neurosci Methods 2024; 408:110159. [PMID: 38723868 DOI: 10.1016/j.jneumeth.2024.110159] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Revised: 04/02/2024] [Accepted: 04/29/2024] [Indexed: 05/19/2024]
Abstract
BACKGROUND In order to push the frontiers of brain-computer interface (BCI) and neuron-electronics, this research presents a novel framework that combines cutting-edge technologies for improved brain-related diagnostics in smart healthcare. This research offers a ground-breaking application of transparent strategies to BCI, promoting openness and confidence in brain-computer interactions and taking inspiration from Grad-CAM (Gradient-weighted Class Activation Mapping) based Explainable Artificial Intelligence (XAI) methodology. The landscape of healthcare diagnostics is about to be redefined by the integration of various technologies, especially when it comes to illnesses related to the brain. NEW METHOD A novel approach has been proposed in this study comprising of Xception architecture which is trained on imagenet database following transfer learning process for extraction of significant features from magnetic resonance imaging dataset acquired from publicly available distinct sources as an input and linear support vector machine has been used for distinguishing distinct classes.Afterwards, gradient-weighted class activation mapping has been deployed as the foundation for explainable artificial intelligence (XAI) for generating informative heatmaps, representing spatial localization of features which were focused to achieve model's predictions. RESULTS Thus, the proposed model not only provides accurate outcomes but also provides transparency for the predictions generated by the Xception network to diagnose presence of abnormal tissues and avoids overfitting issues. Hyperparameters along with performance-metrics are also obtained while validating the proposed network on unseen brain MRI scans to ensure effectiveness of the proposed network. COMPARISON WITH EXISTING METHODS AND CONCLUSIONS The integration of Grad-CAM based explainable artificial intelligence with deep neural network namely Xception offers a significant impact in diagnosing brain tumor disease while highlighting the specific regions of input brain MRI images responsible for making predictions. In this study, the proposed network results in 98.92% accuracy, 98.15% precision, 99.09% sensitivity, 98.18% specificity and 98.91% dice-coefficient while identifying presence of abnormal tissues in the brain. Thus, Xception model trained on distinct dataset following transfer learning process offers remarkable diagnostic accuracy and linear support vector act as a classifier to provide efficient classification among distinct classes. In addition, the deployed explainable artificial intelligence approach helps in revealing the reasoning behind predictions made by deep neural network having black-box nature and provides a clear perspective to assist medical experts in achieving trustworthiness and transparency while diagnosing brain tumor disease in the smart healthcare.
Collapse
Affiliation(s)
- Kamini Lamba
- Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab 140401, India.
| | - Shalli Rani
- Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab 140401, India.
| |
Collapse
|
9
|
Anand V, Gupta S, Koundal D, Alghamdi WY, Alsharbi BM. Deep learning-based image annotation for leukocyte segmentation and classification of blood cell morphology. BMC Med Imaging 2024; 24:83. [PMID: 38589793 PMCID: PMC11003052 DOI: 10.1186/s12880-024-01254-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Accepted: 03/19/2024] [Indexed: 04/10/2024] Open
Abstract
The research focuses on the segmentation and classification of leukocytes, a crucial task in medical image analysis for diagnosing various diseases. The leukocyte dataset comprises four classes of images such as monocytes, lymphocytes, eosinophils, and neutrophils. Leukocyte segmentation is achieved through image processing techniques, including background subtraction, noise removal, and contouring. To get isolated leukocytes, background mask creation, Erythrocytes mask creation, and Leukocytes mask creation are performed on the blood cell images. Isolated leukocytes are then subjected to data augmentation including brightness and contrast adjustment, flipping, and random shearing, to improve the generalizability of the CNN model. A deep Convolutional Neural Network (CNN) model is employed on augmented dataset for effective feature extraction and classification. The deep CNN model consists of four convolutional blocks having eleven convolutional layers, eight batch normalization layers, eight Rectified Linear Unit (ReLU) layers, and four dropout layers to capture increasingly complex patterns. For this research, a publicly available dataset from Kaggle consisting of a total of 12,444 images of four types of leukocytes was used to conduct the experiments. Results showcase the robustness of the proposed framework, achieving impressive performance metrics with an accuracy of 97.98% and precision of 97.97%. These outcomes affirm the efficacy of the devised segmentation and classification approach in accurately identifying and categorizing leukocytes. The combination of advanced CNN architecture and meticulous pre-processing steps establishes a foundation for future developments in the field of medical image analysis.
Collapse
Affiliation(s)
- Vatsala Anand
- Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab, India
| | - Sheifali Gupta
- Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab, India
| | - Deepika Koundal
- School of Computer Science, University of Petroleum & Energy Studies, Dehradun, India.
- Ho Chi Minh City Open University, Ho Chi Minh City, Vietnam.
| | - Wael Y Alghamdi
- Department of Computer Science, College of Computers and Information Technology, Taif University, P. O. Box 11099, 21944, Taif, Saudi Arabia
| | - Bayan M Alsharbi
- Department of Information Technology, College of Computers and Information Technology, Taif University, P. O. Box 11099, 21944, Taif, Saudi Arabia
| |
Collapse
|