1
|
Maqsood S, Damaševičius R, Maskeliūnas R, Forkert ND, Haider S, Latif S. Csec-net: a novel deep features fusion and entropy-controlled firefly feature selection framework for leukemia classification. Health Inf Sci Syst 2025; 13:9. [PMID: 39736875 PMCID: PMC11682032 DOI: 10.1007/s13755-024-00327-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2024] [Accepted: 12/10/2024] [Indexed: 01/01/2025] Open
Abstract
Leukemia, a life-threatening form of cancer, poses a significant global health challenge affecting individuals of all age groups, including both children and adults. Currently, the diagnostic process relies on manual analysis of microscopic images of blood samples. In recent years, machine learning employing deep learning approaches has emerged as cutting-edge solutions for image classification problems. Thus, the aim of this work was to develop and evaluate deep learning methods to enable a computer-aided leukemia diagnosis. The proposed method is composed of multiple stages: Firstly, the given dataset images undergo preprocessing. Secondly, five pre-trained convolutional neural network models, namely MobileNetV2, EfficientNetB0, ConvNeXt-V2, EfficientNetV2, and DarkNet-19, are modified and transfer learning is used for training. Thirdly, deep feature vectors are extracted from each of the convolutional neural network and combined using a convolutional sparse image decomposition fusion strategy. Fourthly, the proposed approach employs an entropy-controlled firefly feature selection technique, which selects the most optimal features for subsequent classification. Finally, the selected features are fed into a multi-class support vector machine for the final classification. The proposed algorithm was applied to a total of 15562 images having four datasets, namely ALLID_B1, ALLID_B2, C_NMC 2019, and ASH and demonstrated superior accuracies of 99.64%, 98.96%, 96.67%, and 98.89%, respectively, surpassing the performance of previous works in the field.
Collapse
Affiliation(s)
- Sarmad Maqsood
- Centre of Real Time Computer Systems, Faculty of Informatics, Kaunas University of Technology, LT-51386 Kaunas, Lithuania
- Department of Radiology, Cumming School of Medicine, University of Calgary, Calgary, AB T2N 4N1 Canada
| | - Robertas Damaševičius
- Centre of Real Time Computer Systems, Faculty of Informatics, Kaunas University of Technology, LT-51386 Kaunas, Lithuania
| | - Rytis Maskeliūnas
- Centre of Real Time Computer Systems, Faculty of Informatics, Kaunas University of Technology, LT-51386 Kaunas, Lithuania
| | - Nils D. Forkert
- Department of Radiology, Cumming School of Medicine, University of Calgary, Calgary, AB T2N 4N1 Canada
| | - Shahab Haider
- Faculty of Computer Science and Engineering, GIK Institute of Engineering Sciences and Technology, Topi, 23640 Pakistan
| | - Shahid Latif
- Department of Electrical Engineering, Iqra National University, Peshawar, 25000 Pakistan
| |
Collapse
|
2
|
Ahmad I, Singh VP, Gore MM. Detection of Diabetic Retinopathy Using Discrete Wavelet-Based Center-Symmetric Local Binary Pattern and Statistical Features. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:1184-1211. [PMID: 39237836 PMCID: PMC11950458 DOI: 10.1007/s10278-024-01243-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/04/2024] [Revised: 07/19/2024] [Accepted: 08/19/2024] [Indexed: 09/07/2024]
Abstract
Computer-aided diagnosis (CAD) system assists ophthalmologists in early diabetic retinopathy (DR) detection by automating the analysis of retinal images, enabling timely intervention and treatment. This paper introduces a novel CAD system based on the global and multi-resolution analysis of retinal images. As a first step, we enhance the quality of the retinal images by applying a sequence of preprocessing techniques, which include the median filter, contrast limited adaptive histogram equalization (CLAHE), and the unsharp filter. These preprocessing steps effectively eliminate noise and enhance the contrast in the retinal images. Further, these images are represented at multi-scales using discrete wavelet transform (DWT), and center symmetric local binary pattern (CSLBP) features are extracted from each scale. The extracted CSLBP features from decomposed images capture the fine and coarse details of the retinal fundus images. Also, statistical features are extracted to capture the global characteristics and provide a comprehensive representation of retinal fundus images. The detection performances of these features are evaluated on a benchmark dataset using two machine learning models, i.e., SVM and k-NN, and found that the performance of the proposed work is considerably more encouraging than other existing methods. Furthermore, the results demonstrate that when wavelet-based CSLBP features are combined with statistical features, they yield notably improved detection performance compared to using these features individually.
Collapse
Affiliation(s)
- Imtiyaz Ahmad
- Department of Computer Science and Engineering, Motilal Nehru National Institute of Technology Allahabad, Prayagraj, 211004, UP, India.
| | - Vibhav Prakash Singh
- Department of Computer Science and Engineering, Motilal Nehru National Institute of Technology Allahabad, Prayagraj, 211004, UP, India
| | - Manoj Madhava Gore
- Department of Computer Science and Engineering, Motilal Nehru National Institute of Technology Allahabad, Prayagraj, 211004, UP, India
| |
Collapse
|
3
|
Chandhakanond P, Aimmanee P. Diabetic retinopathy detection via exudates and hemorrhages segmentation using iterative NICK thresholding, watershed, and Chi 2 feature ranking. Sci Rep 2025; 15:5541. [PMID: 39953248 PMCID: PMC11829032 DOI: 10.1038/s41598-025-90048-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2024] [Accepted: 02/10/2025] [Indexed: 02/17/2025] Open
Abstract
Diabetic retinopathy (DR) is a common eye condition that affects one-third of patients with diabetes, leading to vision loss in both working-age and elderly populations. Early detection and intervention can improve patient outcomes and reduce the burden on healthcare. By developing robust computational techniques, we can advance automated systems for screening and managing diabetic retinopathy. Our specific goal is to detect and segment exudates and hemorrhages in fundus images. In this study, we used the iterative NICK thresholding region growing (INRG) method as a basis. To further improve our results in different applications, we incorporated the watershed separation algorithm (WS) and the Chi2 feature selection method (Chi2) on expanded feature sets. These algorithms were combined with the INRG method to segment hemorrhages and exudates. The segmentation results were used to detect the hemorrhages and exudates, which in turn were used to detect diabetic retinopathy. To evaluate our approach, we compared the results against two traditional methods and two state-of-the-art methods, including the original INRG-HSV model. In terms of hemorrhage segmentation, the INRG with WS (INRG-WS) achieved the highest F-measure of 64.76%, outperforming all other comparative methods. For exudate segmentation, the model INRG-WS- Chi2, which used the combined INRG method with WS and Chi2 ranking on expanded feature sets, performed the best. When it came to hemorrhage detection, the INRG method without WS and using only hue, saturation, and brightness (INRG-HSV) achieved the highest accuracy of 90.27% with the lowest false negative rate (FNR) of 9.39%. For exudate detection, the model INRG-WS-HSV, which used the combined INRG method with WS and only hue, saturation, and brightness, offered the highest accuracy rate of 88.14% and the lowest FNR rate of 8.75%. To detect diabetic retinopathy, we compared the performance of our best hemorrhage detection model (INRG-HSV) and exudate detection model (INRG-WS-HSV) against a state-of-the-art method. Our models significantly outperformed the state-of-the-art method (DT-HSVE), achieving an accuracy of 89.89% and an FNR of 3.66%.
Collapse
Affiliation(s)
- Patsaphon Chandhakanond
- Department of Information Computer and Communication Technology, Sirindhorn International Institute of Technology, Thammasat University, 131 Moo 5, Tivanont Rd, Bangkadi, Meung, 12000, Patumthani, Thailand
| | - Pakinee Aimmanee
- Department of Information Computer and Communication Technology, Sirindhorn International Institute of Technology, Thammasat University, 131 Moo 5, Tivanont Rd, Bangkadi, Meung, 12000, Patumthani, Thailand.
| |
Collapse
|
4
|
Nisar KS, Anjum MW, Raja MAZ, Shoaib M. Design of a novel intelligent computing framework for predictive solutions of malaria propagation model. PLoS One 2024; 19:e0298451. [PMID: 38635576 PMCID: PMC11025872 DOI: 10.1371/journal.pone.0298451] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2023] [Accepted: 01/23/2024] [Indexed: 04/20/2024] Open
Abstract
The paper presents an innovative computational framework for predictive solutions for simulating the spread of malaria. The structure incorporates sophisticated computing methods to improve the reliability of predicting malaria outbreaks. The study strives to provide a strong and effective tool for forecasting the propagation of malaria via the use of an AI-based recurrent neural network (RNN). The model is classified into two groups, consisting of humans and mosquitoes. To develop the model, the traditional Ross-Macdonald model is expanded upon, allowing for a more comprehensive analysis of the intricate dynamics at play. To gain a deeper understanding of the extended Ross model, we employ RNN, treating it as an initial value problem involving a system of first-order ordinary differential equations, each representing one of the seven profiles. This method enables us to obtain valuable insights and elucidate the complexities inherent in the propagation of malaria. Mosquitoes and humans constitute the two cohorts encompassed within the exposition of the mathematical dynamical model. Human dynamics are comprised of individuals who are susceptible, exposed, infectious, and in recovery. The mosquito population, on the other hand, is divided into three categories: susceptible, exposed, and infected. For RNN, we used the input of 0 to 300 days with an interval length of 3 days. The evaluation of the precision and accuracy of the methodology is conducted by superimposing the estimated solution onto the numerical solution. In addition, the outcomes obtained from the RNN are examined, including regression analysis, assessment of error autocorrelation, examination of time series response plots, mean square error, error histogram, and absolute error. A reduced mean square error signifies that the model's estimates are more accurate. The result is consistent with acquiring an approximate absolute error close to zero, revealing the efficacy of the suggested strategy. This research presents a novel approach to solving the malaria propagation model using recurrent neural networks. Additionally, it examines the behavior of various profiles under varying initial conditions of the malaria propagation model, which consists of a system of ordinary differential equations.
Collapse
Affiliation(s)
- Kottakkaran Sooppy Nisar
- Department of Mathematics, College of Science and Humanities in Alkharj, Prince Sattam Bin Abdulaziz, University, Alkharj, Saudi Arabia
| | | | - Muhammad Asif Zahoor Raja
- Future Technology Research Center, National Yunlin University of Science and Technology, Douliou, Yunlin, Taiwan, R.O.C
| | | |
Collapse
|
5
|
Ma F, Liu X, Wang S, Li S, Dai C, Meng J. CSANet: a lightweight channel and spatial attention neural network for grading diabetic retinopathy with optical coherence tomography angiography. Quant Imaging Med Surg 2024; 14:1820-1834. [PMID: 38415109 PMCID: PMC10895115 DOI: 10.21037/qims-23-1270] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Accepted: 12/12/2023] [Indexed: 02/29/2024]
Abstract
Background Diabetic retinopathy (DR) is one of the most common eye diseases. Convolutional neural networks (CNNs) have proven to be a powerful tool for learning DR features; however, accurate DR grading remains challenging due to the small lesions in optical coherence tomography angiography (OCTA) images and the small number of samples. Methods In this article, we developed a novel deep-learning framework to achieve the fine-grained classification of DR; that is, the lightweight channel and spatial attention network (CSANet). Our CSANet comprises two modules: the baseline model, and the hybrid attention module (HAM) based on spatial attention and channel attention. The spatial attention module is used to mine small lesions and obtain a set of spatial position weights to address the problem of small lesions being ignored during the convolution process. The channel attention module uses a set of channel weights to focus on useful features and suppress irrelevant features. Results The extensive experimental results for the OCTA-DR and diabetic retinopathy analysis challenge (DRAC) 2022 data sets showed that the CSANet achieved state-of-the-art DR grading results, showing the effectiveness of the proposed model. The CSANet had an accuracy rate of 97.41% for the OCTA-DR data set and 85.71% for the DRAC 2022 data set. Conclusions Extensive experiments using the OCTA-DR and DRAC 2022 data sets showed that the proposed model effectively mitigated the problems of mutual confusion between DRs of different severity and small lesions being neglected in the convolution process, and thus improved the accuracy of DR classification.
Collapse
Affiliation(s)
- Fei Ma
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Xiao Liu
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Shengbo Wang
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Sien Li
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Cuixia Dai
- College Science, Shanghai Institute of Technology, Shanghai, China
| | - Jing Meng
- School of Computer Science, Qufu Normal University, Rizhao, China
| |
Collapse
|
6
|
Bai Z, Zhu R, He D, Wang S, Huang Z. Adulteration Detection of Pork in Mutton Using Smart Phone with the CBAM-Invert-ResNet and Multiple Parts Feature Fusion. Foods 2023; 12:3594. [PMID: 37835247 PMCID: PMC10572890 DOI: 10.3390/foods12193594] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 09/15/2023] [Accepted: 09/26/2023] [Indexed: 10/15/2023] Open
Abstract
To achieve accurate detection the content of multiple parts pork adulterated in mutton under the effect of mutton flavor essence and colorant by RGB images, the improved CBAM-Invert-ResNet50 network based on the attention mechanism and the inversion residual was used to detect the content of pork from the back, front leg, and hind leg in adulterated mutton. The deep features of different parts extracted by the CBAM-Invert-ResNet50 were fused by feature, stitched, and combined with transfer learning, and the content of pork from mixed parts in adulterated mutton was detected. The results showed that the R2 of the CBAM-Invert-ResNet50 for the back, front leg, and hind leg datasets were 0.9373, 0.8876, and 0.9055, respectively, and the RMSE values were 0.0268 g·g-1, 0.0378 g·g-1, and 0.0316 g·g-1, respectively. The R2 and RMSE of the mixed dataset were 0.9264 and 0.0290 g·g-1, respectively. When the features of different parts were fused, the R2 and RMSE of the CBAM-Invert-ResNet50 for the mixed dataset were 0.9589 and 0.0220 g·g-1, respectively. Compared with the model built before feature fusion, the R2 of the mixed dataset increased by 0.0325, and the RMSE decreased by 0.0070 g·g-1. The above results indicated that the CBAM-Invert-ResNet50 model could effectively detect the content of pork from different parts in adulterated mutton as additives. Feature fusion combined with transfer learning can effectively improve the detection accuracy for the content of mixed parts of pork in adulterated mutton. The results of this study can provide technical support and a basis for maintaining the mutton market order and protecting mutton food safety supervision.
Collapse
Affiliation(s)
- Zongxiu Bai
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China; (Z.B.); (D.H.); (S.W.); (Z.H.)
| | - Rongguang Zhu
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China; (Z.B.); (D.H.); (S.W.); (Z.H.)
- Key Laboratory of Northwest Agricultural Equipment, Ministry of Agriculture and Rural Affairs, Shihezi University, Shihezi 832003, China
- Engineering Research Center for Production Mechanization of Oasis Characteristic Cash Crop, Ministry of Education, Shihezi University, Shihezi 832003, China
| | - Dongyu He
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China; (Z.B.); (D.H.); (S.W.); (Z.H.)
| | - Shichang Wang
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China; (Z.B.); (D.H.); (S.W.); (Z.H.)
| | - Zhongtao Huang
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China; (Z.B.); (D.H.); (S.W.); (Z.H.)
| |
Collapse
|
7
|
Pan J, Ho S, Ly A, Kalloniatis M, Sowmya A. Drusen-aware model for age-related macular degeneration recognition. Ophthalmic Physiol Opt 2023; 43:668-679. [PMID: 36786498 PMCID: PMC10946718 DOI: 10.1111/opo.13108] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2022] [Revised: 01/19/2023] [Accepted: 01/30/2023] [Indexed: 02/15/2023]
Abstract
INTRODUCTION The purpose of this study was to build an automated age-related macular degeneration (AMD) colour fundus photography (CFP) recognition method that incorporates confounders (other ocular diseases) and normal age-related changes by using drusen masks for spatial feature supervision. METHODS A range of clinical sources were used to acquire 7588 CFPs. Contrast limited adaptive histogram equalisation was used for pre-processing. ResNet50 was used as the backbone network, and a spatial attention block was added to integrate prior knowledge of drusen features into the backbone. The evaluation metrics used were sensitivity, specificity and F1 score, which is the harmonic mean of precision and recall (sensitivity) and area under the receiver-operating characteristic (AUC). Fivefold cross-validation was performed, and the results compared with four other methods. RESULTS Excellent discrimination results were obtained with the algorithm. On the public dataset (n = 6565), the proposed method achieved a mean (SD) sensitivity of 0.54 (0.09), specificity of 0.99 (0.00), F1 score of 0.62 (0.06) and AUC of 0.92 (0.02). On the private dataset (n = 1023), the proposed method achieved a sensitivity of 0.92 (0.02), specificity of 0.98 (0.01), F1 score of 0.92 (0.01) and AUC of 0.98 (0.01). CONCLUSION The proposed drusen-aware model outperformed baseline and other vessel feature-based methods in F1 and AUC on the AMD/normal CFP classification task and achieved comparable results on datasets that included other diseases that often confound classification. The method also improved results when a five-category grading protocol was used, thereby reflecting discriminative ability of the algorithm within a real-life clinical setting.
Collapse
Affiliation(s)
- Junjun Pan
- School of Computer Science and EngineeringUniversity of New South WalesKensingtonNew South WalesAustralia
| | - Sharon Ho
- Centre for Eye HealthUniversity of New South WalesKensingtonNew South WalesAustralia
- School of Optometry and Vision ScienceUniversity of New South WalesKensingtonNew South WalesAustralia
| | - Angelica Ly
- School of Optometry and Vision ScienceUniversity of New South WalesKensingtonNew South WalesAustralia
| | - Michael Kalloniatis
- School of Optometry and Vision ScienceUniversity of New South WalesKensingtonNew South WalesAustralia
- School of Medicine (Optometry)Deakin UniversityWaurn PondsVictoriaAustralia
| | - Arcot Sowmya
- School of Computer Science and EngineeringUniversity of New South WalesKensingtonNew South WalesAustralia
| |
Collapse
|
8
|
Xia Y, Yun H, Liu Y. MFEFNet: Multi-scale feature enhancement and Fusion Network for polyp segmentation. Comput Biol Med 2023; 157:106735. [PMID: 36965326 DOI: 10.1016/j.compbiomed.2023.106735] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Revised: 02/19/2023] [Accepted: 02/28/2023] [Indexed: 03/06/2023]
Abstract
The polyp segmentation technology based on computer-aided can effectively avoid the deterioration of polyps and prevent colorectal cancer. To segment the polyp target precisely, the Multi-Scale Feature Enhancement and Fusion Network (MFEFNet) is proposed. First of all, to balance the network's predictive ability and complexity, ResNet50 is designed as the backbone network, and the Shift Channel Block (SCB) is used to unify the spatial location of feature mappings and emphasize local information. Secondly, to further improve the network's feature-extracting ability, the Feature Enhancement Block (FEB) is added, which decouples features, reinforces features by multiple perspectives and reconstructs features. Meanwhile, to weaken the semantic gap in the feature fusion process, we propose strong associated couplers, the Multi-Scale Feature Fusion Block (MSFFB) and the Reducing Difference Block (RDB), which are mainly composed of multiple cross-complementary information interaction modes and reinforce the long-distance dependence between features. Finally, to further refine local regions, the Polarized Self-Attention (PSA) and the Balancing Attention Module (BAM) are introduced for better exploration of detailed information between foreground and background boundaries. Experiments have been conducted under five benchmark datasets (Kvasir-SEG, CVC-ClinicDB, CVC-ClinicDB, CVC300 and CVC-ColonDB) and compared with state-of-the-art polyp segmentation algorithms. The experimental result shows that the proposed network improves Dice and mean intersection over union (mIoU) by an average score of 3.4% and 4%, respectively. Therefore, extensive experiments demonstrate that the proposed network performs favorably against more than a dozen state-of-the-art methods on five popular polyp segmentation benchmarks.
Collapse
Affiliation(s)
- Yang Xia
- School of the Graduate, Changchun University, Changchun, 130022, Jilin, China; School of Electronic Information Engineering, Changchun University, Changchun, 130022, Jilin, China
| | - Haijiao Yun
- School of Electronic Information Engineering, Changchun University, Changchun, 130022, Jilin, China.
| | - Yanjun Liu
- School of the Graduate, Changchun University, Changchun, 130022, Jilin, China; School of Electronic Information Engineering, Changchun University, Changchun, 130022, Jilin, China
| |
Collapse
|
9
|
Alhussan AA, Eid MM, Towfek SK, Khafaga DS. Breast Cancer Classification Depends on the Dynamic Dipper Throated Optimization Algorithm. Biomimetics (Basel) 2023; 8:163. [PMID: 37092415 PMCID: PMC10123690 DOI: 10.3390/biomimetics8020163] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2023] [Revised: 04/12/2023] [Accepted: 04/14/2023] [Indexed: 04/25/2023] Open
Abstract
According to the American Cancer Society, breast cancer is the second largest cause of mortality among women after lung cancer. Women's death rates can be decreased if breast cancer is diagnosed and treated early. Due to the lengthy duration of manual breast cancer diagnosis, an automated approach is necessary for early cancer identification. This research proposes a novel framework integrating metaheuristic optimization with deep learning and feature selection for robustly classifying breast cancer from ultrasound images. The structure of the proposed methodology consists of five stages, namely, data augmentation to improve the learning of convolutional neural network (CNN) models, transfer learning using GoogleNet deep network for feature extraction, selection of the best set of features using a novel optimization algorithm based on a hybrid of dipper throated and particle swarm optimization algorithms, and classification of the selected features using CNN optimized using the proposed optimization algorithm. To prove the effectiveness of the proposed approach, a set of experiments were conducted on a breast cancer dataset, freely available on Kaggle, to evaluate the performance of the proposed feature selection method and the performance of the optimized CNN. In addition, statistical tests were established to study the stability and difference of the proposed approach compared to state-of-the-art approaches. The achieved results confirmed the superiority of the proposed approach with a classification accuracy of 98.1%, which is better than the other approaches considered in the conducted experiments.
Collapse
Affiliation(s)
- Amel Ali Alhussan
- Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Marwa M. Eid
- Faculty of Artificial Intelligence, Delta University for Science and Technology, Mansoura 35712, Egypt
| | - S. K. Towfek
- Delta Higher Institute for Engineering and Technology, Mansoura 35111, Egypt
- Computer Science and Intelligent Systems Research Center, Blacksburg, VA 24060, USA
| | - Doaa Sami Khafaga
- Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| |
Collapse
|
10
|
Alwakid G, Gouda W, Humayun M. Deep Learning-Based Prediction of Diabetic Retinopathy Using CLAHE and ESRGAN for Enhancement. Healthcare (Basel) 2023; 11:863. [PMID: 36981520 PMCID: PMC10048517 DOI: 10.3390/healthcare11060863] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Revised: 03/11/2023] [Accepted: 03/13/2023] [Indexed: 03/17/2023] Open
Abstract
Vision loss can be avoided if diabetic retinopathy (DR) is diagnosed and treated promptly. The main five DR stages are none, moderate, mild, proliferate, and severe. In this study, a deep learning (DL) model is presented that diagnoses all five stages of DR with more accuracy than previous methods. The suggested method presents two scenarios: case 1 with image enhancement using a contrast limited adaptive histogram equalization (CLAHE) filtering algorithm in conjunction with an enhanced super-resolution generative adversarial network (ESRGAN), and case 2 without image enhancement. Augmentation techniques were then performed to generate a balanced dataset utilizing the same parameters for both cases. Using Inception-V3 applied to the Asia Pacific Tele-Ophthalmology Society (APTOS) datasets, the developed model achieved an accuracy of 98.7% for case 1 and 80.87% for case 2, which is greater than existing methods for detecting the five stages of DR. It was demonstrated that using CLAHE and ESRGAN improves a model's performance and learning ability.
Collapse
Affiliation(s)
- Ghadah Alwakid
- Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakakah 72341, Al Jouf, Saudi Arabia;
| | - Walaa Gouda
- Department of Electrical Engineering, Faculty of Engineering at Shoubra, Benha University, Cairo 11672, Egypt;
| | - Mamoona Humayun
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakakah 72341, Al Jouf, Saudi Arabia
| |
Collapse
|
11
|
Rayavel P, Murukesh C. Comparative analysis of deep learning classifiers for diabetic retinopathy identification and detection. THE IMAGING SCIENCE JOURNAL 2023. [DOI: 10.1080/13682199.2023.2168851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/18/2023]
Affiliation(s)
- P. Rayavel
- Department of Computer Science and Engineering (Cybersecurity), Sri Sairam Institute of Technology, Chennai, Tamil Nadu, India
| | - C. Murukesh
- Department of Electronics and Communication Engineering, Velammal Engineering College, Chennai, Tamil Nadu, India
| |
Collapse
|
12
|
Maqsood S, Damaševičius R. Multiclass skin lesion localization and classification using deep learning based features fusion and selection framework for smart healthcare. Neural Netw 2023; 160:238-258. [PMID: 36701878 DOI: 10.1016/j.neunet.2023.01.022] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2022] [Revised: 11/13/2022] [Accepted: 01/19/2023] [Indexed: 01/27/2023]
Abstract
BACKGROUND The idea of smart healthcare has gradually gained attention as a result of the information technology industry's rapid development. Smart healthcare uses next-generation technologies i.e., artificial intelligence (AI) and Internet of Things (IoT), to intelligently transform current medical methods to make them more efficient, dependable and individualized. One of the most prominent uses of telemedicine and e-health in medical image analysis is teledermatology. Telecommunications technologies are used in this industry to send medical information to professionals. Teledermatology is a useful method for the identification of skin lesions, particularly in rural locations, because the skin is visually perceptible. One of the most recent tools for diagnosing skin cancer is dermoscopy. To classify skin malignancies, numerous computational approaches have been proposed in the literature. However, difficulties still exist i.e., lesions with low contrast, imbalanced datasets, high level of memory complexity, and the extraction of redundant features. METHODS In this work, a unified CAD model is proposed based on a deep learning framework for skin lesion segmentation and classification. In the proposed approach, the source dermoscopic images are initially pre-processed using a contrast enhancement based modified bio-inspired multiple exposure fusion approach. In the second stage, a custom 26-layered convolutional neural network (CNN) architecture is designed to segment the skin lesion regions. In the third stage, four pre-trained CNN models (Xception, ResNet-50, ResNet-101 and VGG16) are modified and trained using transfer learning on the segmented lesion images. In the fourth stage, the deep features vectors are extracted from all the CNN models and fused using the convolutional sparse image decomposition fusion approach. In the fifth stage, the univariate measurement and Poisson distribution feature selection approach is used for the best features selection for classification. Finally, the selected features are fed to the multi-class support vector machine (MC-SVM) for the final classification. RESULTS The proposed approach employed to the HAM10000, ISIC2018, ISIC2019, and PH2 datasets and achieved an accuracy of 98.57%, 98.62%, 93.47%, and 98.98% respectively which are better than previous works. CONCLUSION When compared to renowned state-of-the-art methods, experimental results show that the proposed skin lesion detection and classification approach achieved higher performance in terms of both visually and enhanced quantitative evaluation with enhanced accuracy.
Collapse
Affiliation(s)
- Sarmad Maqsood
- Department of Software Engineering, Faculty of Informatics Engineering, Kaunas University of Technology, LT-51386 Kaunas, Lithuania.
| | - Robertas Damaševičius
- Department of Software Engineering, Faculty of Informatics Engineering, Kaunas University of Technology, LT-51386 Kaunas, Lithuania.
| |
Collapse
|
13
|
Effective Handling of Missing Values in Datasets for Classification Using Machine Learning Methods. INFORMATION 2023. [DOI: 10.3390/info14020092] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/05/2023] Open
Abstract
The existence of missing values reduces the amount of knowledge learned by the machine learning models in the training stage thus affecting the classification accuracy negatively. To address this challenge, we introduce the use of Support Vector Machine (SVM) regression for imputing the missing values. Additionally, we propose a two-level classification process to reduce the number of false classifications. Our evaluation of the proposed method was conducted using the PIMA Indian dataset for diabetes classification. We compared the performance of five different machine learning models: Naive Bayes (NB), Support Vector Machine (SVM), k-Nearest Neighbours (KNN), Random Forest (RF), and Linear Regression (LR). The results of our experiments show that the SVM classifier achieved the highest accuracy of 94.89%. The RF classifier had the highest precision (98.80%) and the SVM classifier had the highest recall (85.48%). The NB model had the highest F1-Score (95.59%). Our proposed method provides a promising solution for detecting diabetes at an early stage by addressing the issue of missing values in the dataset. Our results show that the use of SVM regression and a two-level classification process can notably improve the performance of machine learning models for diabetes classification. This work provides a valuable contribution to the field of diabetes research and highlights the importance of addressing missing values in machine learning applications.
Collapse
|
14
|
Wang C, Zhang Y, Xu S, Liu Y, Xie L, Wu C, Yang Q, Chu Y, Ye Q. Research on Assistant Diagnosis of Fundus Optic Neuropathy Based on Deep Learning. Curr Eye Res 2023; 48:51-59. [PMID: 36264060 DOI: 10.1080/02713683.2022.2138917] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022]
Abstract
PURPOSE The purpose of this study was to use the neural network to distinguish optic edema (ODE), and optic atrophy from normal fundus images and try to use visualization to explain the artificial intelligence methods. METHODS Three hundred and sixty-seven images of ODE, 206 images of optic atrophy, and 231 images of normal fundus were used, which were provided by two hospitals. A set of image preprocessing and data enhancement methods were created and a variety of different neural network models, such as VGG16, VGG19, Inception V3, and 50-layer Deep Residual Learning (ResNet50) were used. The accuracy, recall, F1-score, and ROC curve under different networks were analyzed to evaluate the performance of models. Besides, CAM (class activation mapping) was utilized to find the focus of neural network and visualization of neural network with feature fusion. RESULTS Our image preprocessing and data enhancement method significantly improved the accuracy of model performance by about 10%. Among the networks, VGG16 had the best effect, as the accuracy of ODE, optic atrophy and normal fundus were 98, 90, and 95%, respectively. The macro-average and micro-average of VGG16 both reached 0.98. From CAM we can clearly find out that the focus area of the network is near the optic cup. From feature fusion images, we can find out the difference between the three types fundus images. CONCLUSION Through image preprocessing, data enhancement, and neural network training, we applied artificial intelligence to identify ophthalmic diseases, acquired the focus area through CAM, and identified the difference between the three ophthalmic diseases through neural network middle layers visualization. With the help of assistant diagnosis, ophthalmologists can evaluate cases more precisely and more clearly.
Collapse
Affiliation(s)
- Chengjin Wang
- Key Laboratory of Weak-Light Nonlinear Photonics, School of Physics and TEDA Applied Physics, Ministry of Education, Nankai University, Tianjin, China
| | - Yuwei Zhang
- Key Laboratory of Weak-Light Nonlinear Photonics, School of Physics and TEDA Applied Physics, Ministry of Education, Nankai University, Tianjin, China
| | - Shuai Xu
- Key Laboratory of Weak-Light Nonlinear Photonics, School of Physics and TEDA Applied Physics, Ministry of Education, Nankai University, Tianjin, China
| | - Yuyan Liu
- Tianjin Key Lab of Ophthalmology and Visual Science, Tianjin Eye Hospital and Eye Institute, Nankai University Affiliated Eye Hospital, Clinical College of Ophthalmology Tianjin Medical University, Tianjin, China
| | - Lindan Xie
- Tianjin Key Lab of Ophthalmology and Visual Science, Tianjin Eye Hospital and Eye Institute, Nankai University Affiliated Eye Hospital, Clinical College of Ophthalmology Tianjin Medical University, Tianjin, China
| | - Changlong Wu
- Ophthalmology, Jinan Second People's Hospital, Jinan City, Shandong Province, China
| | - Qianhui Yang
- Tianjin Key Laboratory of Retinal Functions and Diseases, Tianjin Branch of National Clinical Research Center for Ocular Disease, Eye Institute and School of Optometry, Tianjin Medical University Eye Hospital, Tianjin, China
| | - Yanhua Chu
- Tianjin Key Lab of Ophthalmology and Visual Science, Tianjin Eye Hospital and Eye Institute, Nankai University Affiliated Eye Hospital, Clinical College of Ophthalmology Tianjin Medical University, Tianjin, China
| | - Qing Ye
- Key Laboratory of Weak-Light Nonlinear Photonics, School of Physics and TEDA Applied Physics, Ministry of Education, Nankai University, Tianjin, China.,Nankai University Eye Institute, Nankai University Afflicted Eye Hospital, Nankai University, Tianjin, China
| |
Collapse
|
15
|
Alwakid G, Gouda W, Humayun M, Jhanjhi NZ. Deep learning-enhanced diabetic retinopathy image classification. Digit Health 2023; 9:20552076231194942. [PMID: 37588156 PMCID: PMC10426308 DOI: 10.1177/20552076231194942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/28/2023] [Indexed: 08/18/2023] Open
Abstract
Objective Diabetic retinopathy (DR) can sometimes be treated and prevented from causing irreversible vision loss if caught and treated properly. In this work, a deep learning (DL) model is employed to accurately identify all five stages of DR. Methods The suggested methodology presents two examples, one with and one without picture augmentation. A balanced dataset meeting the same criteria in both cases is then generated using augmentative methods. The DenseNet-121-rendered model on the Asia Pacific Tele-Ophthalmology Society (APTOS) and dataset for diabetic retinopathy (DDR) datasets performed exceptionally well when compared to other methods for identifying the five stages of DR. Results Our propose model achieved the highest test accuracy of 98.36%, top-2 accuracy of 100%, and top-3 accuracy of 100% for the APTOS dataset, and the highest test accuracy of 79.67%, top-2 accuracy of 92.%76, and top-3 accuracy of 98.94% for the DDR dataset. Additional criteria (precision, recall, and F1-score) for gauging the efficacy of the proposed model were established with the help of APTOS and DDR. Conclusions It was discovered that feeding a model with higher-quality photographs increased its efficiency and ability for learning, as opposed to both state-of-the-art technology and the other, non-enhanced model.
Collapse
Affiliation(s)
- Ghadah Alwakid
- Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakakah, Saudi Arabia
| | - Walaa Gouda
- Department of Electrical Engineering, Faculty of Engineering at Shoubra, Benha University, Cairo, Egypt
| | - Mamoona Humayun
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakakah, Saudi Arabia
| | | |
Collapse
|
16
|
Chandhakanond P, Aimmanee P. Hemorrhage segmentation in mobile-phone retinal images using multiregion contrast enhancement and iterative NICK thresholding region growing. Sci Rep 2022; 12:21513. [PMID: 36513802 PMCID: PMC9747926 DOI: 10.1038/s41598-022-26073-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2022] [Accepted: 12/08/2022] [Indexed: 12/15/2022] Open
Abstract
Hemorrhage segmentation in retinal images is challenging because the sizes and shapes vary for each hemorrhage, the intensity is close to the blood vessels and macula, and the intensity is often nonuniform, especially for large hemorrhages. Hemorrhage segmentation in mobile-phone retinal images is even more challenging because mobile-phone retinal images usually have poorer contrast, more shadows, and uneven illumination compared to those obtained from the table-top ophthalmoscope. In this work, the proposed KMMRC-INRG method enhances the hemorrhage segmentation performance with nonuniform intensity in poor lighting conditions on mobile-phone images. It improves the uneven illumination of mobile-phone retinal images using a proposed method, K-mean multiregion contrast enhancement (KMMRC). It also enhances the boundary segmentation of the hemorrhage blobs using a novel iterative NICK thresholding region growing (INRG) method before applying an SVM classifier based on hue, saturation, and brightness features. This approach can achieve as high as 80.18%, 91.26%, 85.36%, and 80.08% for recall, precision, F1-measure, and IoU, respectively. The F1-measure score improves up to 19.02% compared to a state-of-the-art method DT-HSVE tested on the same full dataset and as much as 58.88% when considering only images with large-size hemorrhages.
Collapse
Affiliation(s)
- Patsaphon Chandhakanond
- grid.412434.40000 0004 1937 1127School of Information, Computer, and Communication Technology, Sirindhorn International Institute of Technology, Thammasat University, 131 Moo 5, Tivanont Rd, Bangkadi, Meung, Patumthani, 12000 Thailand
| | - Pakinee Aimmanee
- grid.412434.40000 0004 1937 1127School of Information, Computer, and Communication Technology, Sirindhorn International Institute of Technology, Thammasat University, 131 Moo 5, Tivanont Rd, Bangkadi, Meung, Patumthani, 12000 Thailand
| |
Collapse
|
17
|
Kusakunniran W, Karnjanapreechakorn S, Choopong P, Siriapisith T, Tesavibul N, Phasukkijwatana N, Prakhunhungsit S, Boonsopon S. Detecting and staging diabetic retinopathy in retinal images using multi-branch CNN. APPLIED COMPUTING AND INFORMATICS 2022. [DOI: 10.1108/aci-06-2022-0150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
Abstract
PurposeThis paper aims to propose a solution for detecting and grading diabetic retinopathy (DR) in retinal images using a convolutional neural network (CNN)-based approach. It could classify input retinal images into a normal class or an abnormal class, which would be further split into four stages of abnormalities automatically.Design/methodology/approachThe proposed solution is developed based on a newly proposed CNN architecture, namely, DeepRoot. It consists of one main branch, which is connected by two side branches. The main branch is responsible for the primary feature extractor of both high-level and low-level features of retinal images. Then, the side branches further extract more complex and detailed features from the features outputted from the main branch. They are designed to capture details of small traces of DR in retinal images, using modified zoom-in/zoom-out and attention layers.FindingsThe proposed method is trained, validated and tested on the Kaggle dataset. The regularization of the trained model is evaluated using unseen data samples, which were self-collected from a real scenario from a hospital. It achieves a promising performance with a sensitivity of 98.18% under the two classes scenario.Originality/valueThe new CNN-based architecture (i.e. DeepRoot) is introduced with the concept of a multi-branch network. It could assist in solving a problem of an unbalanced dataset, especially when there are common characteristics across different classes (i.e. four stages of DR). Different classes could be outputted at different depths of the network.
Collapse
|
18
|
Dubey S, Dixit M. Recent developments on computer aided systems for diagnosis of diabetic retinopathy: a review. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:14471-14525. [PMID: 36185322 PMCID: PMC9510498 DOI: 10.1007/s11042-022-13841-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Revised: 04/27/2022] [Accepted: 09/06/2022] [Indexed: 06/16/2023]
Abstract
Diabetes is a long-term condition in which the pancreas quits producing insulin or the body's insulin isn't utilised properly. One of the signs of diabetes is Diabetic Retinopathy. Diabetic retinopathy is the most prevalent type of diabetes, if remains unaddressed, diabetic retinopathy can affect all diabetics and become very serious, raising the chances of blindness. It is a chronic systemic condition that affects up to 80% of patients for more than ten years. Many researchers believe that if diabetes individuals are diagnosed early enough, they can be rescued from the condition in 90% of cases. Diabetes damages the capillaries, which are microscopic blood vessels in the retina. On images, blood vessel damage is usually noticeable. Therefore, in this study, several traditional, as well as deep learning-based approaches, are reviewed for the classification and detection of this particular diabetic-based eye disease known as diabetic retinopathy, and also the advantage of one approach over the other is also described. Along with the approaches, the dataset and the evaluation metrics useful for DR detection and classification are also discussed. The main finding of this study is to aware researchers about the different challenges occurs while detecting diabetic retinopathy using computer vision, deep learning techniques. Therefore, a purpose of this review paper is to sum up all the major aspects while detecting DR like lesion identification, classification and segmentation, security attacks on the deep learning models, proper categorization of datasets and evaluation metrics. As deep learning models are quite expensive and more prone to security attacks thus, in future it is advisable to develop a refined, reliable and robust model which overcomes all these aspects which are commonly found while designing deep learning models.
Collapse
Affiliation(s)
- Shradha Dubey
- Madhav Institute of Technology & Science (Department of Computer Science and Engineering), Gwalior, M.P. India
| | - Manish Dixit
- Madhav Institute of Technology & Science (Department of Computer Science and Engineering), Gwalior, M.P. India
| |
Collapse
|
19
|
Maqsood S, Damaševičius R, Maskeliūnas R. Multi-Modal Brain Tumor Detection Using Deep Neural Network and Multiclass SVM. Medicina (B Aires) 2022; 58:medicina58081090. [PMID: 36013557 PMCID: PMC9413317 DOI: 10.3390/medicina58081090] [Citation(s) in RCA: 25] [Impact Index Per Article: 8.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2022] [Revised: 08/03/2022] [Accepted: 08/06/2022] [Indexed: 02/05/2023] Open
Abstract
Background and Objectives: Clinical diagnosis has become very significant in today's health system. The most serious disease and the leading cause of mortality globally is brain cancer which is a key research topic in the field of medical imaging. The examination and prognosis of brain tumors can be improved by an early and precise diagnosis based on magnetic resonance imaging. For computer-aided diagnosis methods to assist radiologists in the proper detection of brain tumors, medical imagery must be detected, segmented, and classified. Manual brain tumor detection is a monotonous and error-prone procedure for radiologists; hence, it is very important to implement an automated method. As a result, the precise brain tumor detection and classification method is presented. Materials and Methods: The proposed method has five steps. In the first step, a linear contrast stretching is used to determine the edges in the source image. In the second step, a custom 17-layered deep neural network architecture is developed for the segmentation of brain tumors. In the third step, a modified MobileNetV2 architecture is used for feature extraction and is trained using transfer learning. In the fourth step, an entropy-based controlled method was used along with a multiclass support vector machine (M-SVM) for the best features selection. In the final step, M-SVM is used for brain tumor classification, which identifies the meningioma, glioma and pituitary images. Results: The proposed method was demonstrated on BraTS 2018 and Figshare datasets. Experimental study shows that the proposed brain tumor detection and classification method outperforms other methods both visually and quantitatively, obtaining an accuracy of 97.47% and 98.92%, respectively. Finally, we adopt the eXplainable Artificial Intelligence (XAI) method to explain the result. Conclusions: Our proposed approach for brain tumor detection and classification has outperformed prior methods. These findings demonstrate that the proposed approach obtained higher performance in terms of both visually and enhanced quantitative evaluation with improved accuracy.
Collapse
|
20
|
Applying supervised contrastive learning for the detection of diabetic retinopathy and its severity levels from fundus images. Comput Biol Med 2022; 146:105602. [DOI: 10.1016/j.compbiomed.2022.105602] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2022] [Revised: 04/26/2022] [Accepted: 05/06/2022] [Indexed: 01/02/2023]
|
21
|
Andersen JKH, Hubel MS, Rasmussen ML, Grauslund J, Savarimuthu TR. Automatic Detection of Abnormalities and Grading of Diabetic Retinopathy in 6-Field Retinal Images: Integration of Segmentation Into Classification. Transl Vis Sci Technol 2022; 11:19. [PMID: 35731541 PMCID: PMC9233290 DOI: 10.1167/tvst.11.6.19] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Purpose Classification of diabetic retinopathy (DR) is traditionally based on severity grading, given by the most advanced lesion, but potentially leaving out relevant information for risk stratification. In this study, we aimed to develop a deep learning model able to individually segment seven different DR-lesions, in order to test if this would improve a subsequently developed classification model. Methods First, manual segmentation of 34,075 different DR-lesions was used to construct a segmentation model, with performance subsequently compared to another retinal specialist. Second, we constructed a 5-step classification model using a data set of 31,325 expert-annotated retinal 6-field images and evaluated if performance was improved with the integration of presegmentation given by the segmentation model. Results The segmentation model had higher average sensitivity across all abnormalities compared to the retinal expert (0.68 and 0.62) at a comparable average F1-score (0.60 and 0.62). Model sensitivity for microaneurysms, retinal hemorrhages and intraretinal microvascular abnormalities was higher by 42.5%, 8.8%, and 67.5% and F1-scores by 15.8%, 6.5%, and 12.5%, respectively. When presegmentation was included, grading performance increased by 29.7%, 6.0%, and 4.5% for average per class accuracy, quadratic weighted kappa, and multiclass macro area under the curve, with values of 70.4%, 0.90, and 0.92, respectively. Conclusions The segmentation model matched an expert in detecting retinal abnormalities, and presegmentation substantially improved accuracy of the automated classification model. Translational Relevance Presegmentation may yield more accurate automated DR grading models and increase interpretability and trust in model decisions.
Collapse
Affiliation(s)
- Jakob K H Andersen
- The Maersk Mc-Kinney Moeller Institute, SDU Robotics, University of Southern Denmark, Odense, Denmark.,Steno Diabetes Center Odense, Odense University Hospital, Odense, Denmark
| | - Martin S Hubel
- The Maersk Mc-Kinney Moeller Institute, SDU Robotics, University of Southern Denmark, Odense, Denmark
| | - Malin L Rasmussen
- Department of Ophthalmology, Odense University Hospital, Odense, Denmark.,Department of Clinical Research, University of Southern Denmark, Odense, Denmark
| | - Jakob Grauslund
- Department of Ophthalmology, Odense University Hospital, Odense, Denmark.,Department of Clinical Research, University of Southern Denmark, Odense, Denmark.,Steno Diabetes Center Odense, Odense University Hospital, Odense, Denmark
| | - Thiusius R Savarimuthu
- The Maersk Mc-Kinney Moeller Institute, SDU Robotics, University of Southern Denmark, Odense, Denmark
| |
Collapse
|
22
|
TTCNN: A Breast Cancer Detection and Classification towards Computer-Aided Diagnosis Using Digital Mammography in Early Stages. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12073273] [Citation(s) in RCA: 30] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Breast cancer is a major research area in the medical image analysis field; it is a dangerous disease and a major cause of death among women. Early and accurate diagnosis of breast cancer based on digital mammograms can enhance disease detection accuracy. Medical imagery must be detected, segmented, and classified for computer-aided diagnosis (CAD) systems to help the radiologists for accurate diagnosis of breast lesions. Therefore, an accurate breast cancer detection and classification approach is proposed for screening of mammograms. In this paper, we present a deep learning system that can identify breast cancer in mammogram screening images using an “end-to-end” training strategy that efficiently uses mammography images for computer-aided breast cancer recognition in the early stages. First, the proposed approach implements the modified contrast enhancement method in order to refine the detail of edges from the source mammogram images. Next, the transferable texture convolutional neural network (TTCNN) is presented to enhance the performance of classification and the energy layer is integrated in this work to extract the texture features from the convolutional layer. The proposed approach consists of only three layers of convolution and one energy layer, rather than the pooling layer. In the third stage, we analyzed the performance of TTCNN based on deep features of convolutional neural network models (InceptionResNet-V2, Inception-V3, VGG-16, VGG-19, GoogLeNet, ResNet-18, ResNet-50, and ResNet-101). The deep features are extracted by determining the best layers which enhance the classification accuracy. In the fourth stage, by using the convolutional sparse image decomposition approach, all the extracted feature vectors are fused and, finally, the best features are selected by using the entropy controlled firefly method. The proposed approach employed on DDSM, INbreast, and MIAS datasets and attained the average accuracy of 97.49%. Our proposed transferable texture CNN-based method for classifying screening mammograms has outperformed prior methods. These findings demonstrate that automatic deep learning algorithms can be easily trained to achieve high accuracy in diverse mammography images, and can offer great potential to improve clinical tools to minimize false positive and false negative screening mammography results.
Collapse
|
23
|
Jabeen K, Khan MA, Alhaisoni M, Tariq U, Zhang YD, Hamza A, Mickus A, Damaševičius R. Breast Cancer Classification from Ultrasound Images Using Probability-Based Optimal Deep Learning Feature Fusion. SENSORS (BASEL, SWITZERLAND) 2022; 22:807. [PMID: 35161552 PMCID: PMC8840464 DOI: 10.3390/s22030807] [Citation(s) in RCA: 81] [Impact Index Per Article: 27.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/25/2021] [Revised: 01/12/2022] [Accepted: 01/17/2022] [Indexed: 12/11/2022]
Abstract
After lung cancer, breast cancer is the second leading cause of death in women. If breast cancer is detected early, mortality rates in women can be reduced. Because manual breast cancer diagnosis takes a long time, an automated system is required for early cancer detection. This paper proposes a new framework for breast cancer classification from ultrasound images that employs deep learning and the fusion of the best selected features. The proposed framework is divided into five major steps: (i) data augmentation is performed to increase the size of the original dataset for better learning of Convolutional Neural Network (CNN) models; (ii) a pre-trained DarkNet-53 model is considered and the output layer is modified based on the augmented dataset classes; (iii) the modified model is trained using transfer learning and features are extracted from the global average pooling layer; (iv) the best features are selected using two improved optimization algorithms known as reformed differential evaluation (RDE) and reformed gray wolf (RGW); and (v) the best selected features are fused using a new probability-based serial approach and classified using machine learning algorithms. The experiment was conducted on an augmented Breast Ultrasound Images (BUSI) dataset, and the best accuracy was 99.1%. When compared with recent techniques, the proposed framework outperforms them.
Collapse
Affiliation(s)
- Kiran Jabeen
- Department of Computer Science, HITEC University Taxila, Taxila 47080, Pakistan; (K.J.); (M.A.K.); (A.H.)
| | - Muhammad Attique Khan
- Department of Computer Science, HITEC University Taxila, Taxila 47080, Pakistan; (K.J.); (M.A.K.); (A.H.)
| | - Majed Alhaisoni
- College of Computer Science and Engineering, University of Ha’il, Ha’il 55211, Saudi Arabia;
| | - Usman Tariq
- College of Computer Engineering and Science, Prince Sattam Bin Abdulaziz University, Al-Kharaj 11942, Saudi Arabia;
| | - Yu-Dong Zhang
- Department of Informatics, University of Leicester, Leicester LE1 7RH, UK;
| | - Ameer Hamza
- Department of Computer Science, HITEC University Taxila, Taxila 47080, Pakistan; (K.J.); (M.A.K.); (A.H.)
| | - Artūras Mickus
- Department of Applied Informatics, Vytautas Magnus University, LT-44404 Kaunas, Lithuania;
| | - Robertas Damaševičius
- Department of Applied Informatics, Vytautas Magnus University, LT-44404 Kaunas, Lithuania;
| |
Collapse
|
24
|
Breast Cancer Detection Using Mammogram Images with Improved Multi-Fractal Dimension Approach and Feature Fusion. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app112412122] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/11/2022]
Abstract
Breast cancer detection using mammogram images at an early stage is an important step in disease diagnostics. We propose a new method for the classification of benign or malignant breast cancer from mammogram images. Hybrid thresholding and the machine learning method are used to derive the region of interest (ROI). The derived ROI is then separated into five different blocks. The wavelet transform is applied to suppress noise from each produced block based on BayesShrink soft thresholding by capturing high and low frequencies within different sub-bands. An improved fractal dimension (FD) approach, called multi-FD (M-FD), is proposed to extract multiple features from each denoised block. The number of features extracted is then reduced by a genetic algorithm. Five classifiers are trained and used with the artificial neural network (ANN) to classify the extracted features from each block. Lastly, the fusion process is performed on the results of five blocks to obtain the final decision. The proposed approach is tested and evaluated on four benchmark mammogram image datasets (MIAS, DDSM, INbreast, and BCDR). We present the results of single- and double-dataset evaluations. Only one dataset is used for training and testing in the single-dataset evaluation, whereas two datasets (one for training, and one for testing) are used in the double-dataset evaluation. The experiment results show that the proposed method yields better results on the INbreast dataset in the single-dataset evaluation, whilst better results are obtained on the remaining datasets in the double-dataset evaluation. The proposed approach outperforms other state-of-the-art models on the Mini-MIAS dataset.
Collapse
|
25
|
Predicting COVID-19 Cases in South Korea with All K-Edited Nearest Neighbors Noise Filter and Machine Learning Techniques. INFORMATION 2021. [DOI: 10.3390/info12120528] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022] Open
Abstract
The application of machine learning techniques to the epidemiology of COVID-19 is a necessary measure that can be exploited to curtail the further spread of this endemic. Conventional techniques used to determine the epidemiology of COVID-19 are slow and costly, and data are scarce. We investigate the effects of noise filters on the performance of machine learning algorithms on the COVID-19 epidemiology dataset. Noise filter algorithms are used to remove noise from the datasets utilized in this study. We applied nine machine learning techniques to classify the epidemiology of COVID-19, which are bagging, boosting, support vector machine, bidirectional long short-term memory, decision tree, naïve Bayes, k-nearest neighbor, random forest, and multinomial logistic regression. Data from patients who contracted coronavirus disease were collected from the Kaggle database between 23 January 2020 and 24 June 2020. Noisy and filtered data were used in our experiments. As a result of denoising, machine learning models have produced high results for the prediction of COVID-19 cases in South Korea. For isolated cases after performing noise filtering operations, machine learning techniques achieved an accuracy between 98–100%. The results indicate that filtering noise from the dataset can improve the accuracy of COVID-19 case prediction algorithms.
Collapse
|
26
|
Accurate Diagnosis of Diabetic Retinopathy and Glaucoma Using Retinal Fundus Images Based on Hybrid Features and Genetic Algorithm. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11136178] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/12/2023]
Abstract
Diabetic retinopathy (DR) and glaucoma can both be incurable if they are not detected early enough. Therefore, ophthalmologists worldwide are striving to detect them by personally screening retinal fundus images. However, this procedure is not only tedious, subjective, and labor-intensive, but also error-prone. Worse yet, it may not even be attainable in some countries where ophthalmologists are in short supply. A practical solution to this complicated problem is a computer-aided diagnosis (CAD) system—the objective of this work. We propose an accurate system to detect at once any of the two diseases from retinal fundus images. The accuracy stems from two factors. First, we calculate a large set of hybrid features belonging to three groups: first-order statistics (FOS), higher-order statistics (HOS), and histogram of oriented gradient (HOG). Then, these features are skillfully reduced using a genetic algorithm scheme that selects only the most relevant and significant of them. Finally, the selected features are fed to a classifier to detect one of three classes: DR, glaucoma, or normal. Four classifiers are tested for this job: decision tree (DT), naive Bayes (NB), k-nearest neighbor (kNN), and linear discriminant analysis (LDA). The experimental work, conducted on three publicly available datasets, two of them merged into one, shows impressive performance in terms of four standard classification metrics, each computed using k-fold crossvalidation for added credibility. The highest accuracy has been provided by DT—96.67% for DR, 100% for glaucoma, and 96.67% for normal.
Collapse
|