1
|
Rani R, Sahoo J, Bellamkonda S, Kumar S. Attention-enhanced corn disease diagnosis using few-shot learning and VGG16. MethodsX 2025; 14:103172. [PMID: 39911906 PMCID: PMC11795141 DOI: 10.1016/j.mex.2025.103172] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2024] [Accepted: 01/14/2025] [Indexed: 02/07/2025] Open
Abstract
Plant Disease Detection in the early stage is paramount. Traditionally, it was done manually by the farmers, which is a laborious and time-intensive task. With the advent of AI, Machine Learning and Deep Learning methods are used to detect and categorize plant diseases. However, they rely on extensive datasets for accurate prediction, which is impracticable to acquire and annotate. Thus, Few Shot Learning is the state-of-the-art model in machine learning, which requires minimum examples to train the model for generalization. As humans need a few examples to recognize things, Few-shot Learning mimics the same human brain process. The proposed work uses a pre-trained convolution neural network, VGG16, as the backbone, fine-tuned on the corn disease dataset. An attention module is integrated with the backbone, and further, prototypical few-shot learning is used for corn disease prediction and classification with an accuracy of 98.25 %.•The proposed model intends to identify the diseases early, so the insights generated would be relevant for farmers, and probable losses can be reduced.•By applying Few-Shot Learning, the system avoids the significant challenges of requiring extensively annotated datasets, making it feasible for real-world agricultural applications.•Incorporating a fine-tuned VGG16 backbone along with an attention mechanism and prototypical Few-Shot Learning results in a robust and scalable solution with high accuracy for classifying corn diseases.
Collapse
Affiliation(s)
- Ruchi Rani
- Department of Computer Science and Engineering, Indian Institute of Information Technology Kottayam, 686635, Kerala, India
- Department of Computer Engineering and Technology, School of Computer Engineering and Technology, Dr.Vishwanath Karad MIT World Peace University, Pune, 411038, Maharashtra, India
| | - Jayakrushna Sahoo
- Department of Computer Science and Engineering, Indian Institute of Information Technology Kottayam, 686635, Kerala, India
| | - Sivaiah Bellamkonda
- Department of Computer Science and Engineering, Indian Institute of Information Technology Kottayam, 686635, Kerala, India
| | - Sumit Kumar
- Symbiosis Institute of Technology, Pune Campus, Symbiosis International (Deemed University), Pune, Maharashtra, 412115, India
| |
Collapse
|
2
|
Pacal I, Ozdemir B, Zeynalov J, Gasimov H, Pacal N. A novel CNN-ViT-based deep learning model for early skin cancer diagnosis. Biomed Signal Process Control 2025; 104:107627. [DOI: 10.1016/j.bspc.2025.107627] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
3
|
Askale GT, Yibel AB, Taye BM, Wubneh GD. Mobile based deep CNN model for maize leaf disease detection and classification. PLANT METHODS 2025; 21:72. [PMID: 40442806 PMCID: PMC12121153 DOI: 10.1186/s13007-025-01386-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/12/2025] [Accepted: 05/09/2025] [Indexed: 06/02/2025]
Abstract
Maize is the most produced crop in the world, exceeding wheat and rice production. However, its yield is often affected by various leaf diseases. Early identification of maize leaf disease through easily accessible tool is required to increase the yield of maize. Recently, researchers have attempted to detect and classify maize leaf diseases using Deep Learning algorithms. However, to the best of the researcher's knowledge, nearly all the studies are concentrated on developing an offline model that can detect maize diseases. But, those models are not easily accessible to individuals and don't provide immediate feedback and monitoring. Thus, in this study, we developed a novel real-time, user-friendly maize leaf disease detection and classification mobile application. The VGG16, AlexNet, and ResNet50 models were implemented and compared their performance on maize disease detection and classification. A total of 4188 images of blight, common_rust, grey_leaf_spot, and healthy were used to train each model. Data augmentation techniques were applied to the dataset to increase the size of the dataset, which can also reduce model overfitting. Weighted cross-entropy loss was also employed to mitigate class-imbalance problems. After training, VGG16 achieved 95% of testing accuracy, AlexNet achieved 91%, and ResNet50 achieved 72% of testing accuracy. The VGG16 model outperformed the other models in terms of accuracy. Consequently, we deployed the VGG16 model into a mobile application to provide real-time disease detection and classification tool for farmers, extension officers, agribusiness managers, and policy-makers. The developed application will enhance early disease detection, decision making, and contribute to better crop management and food security.
Collapse
|
4
|
Pei G, Qian X, Zhou B, Liu Z, Wu W. Research on agricultural disease recognition methods based on very large Kernel convolutional network-RepLKNet. Sci Rep 2025; 15:16843. [PMID: 40374696 PMCID: PMC12081735 DOI: 10.1038/s41598-025-01553-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2025] [Accepted: 05/07/2025] [Indexed: 05/17/2025] Open
Abstract
Agricultural diseases pose significant challenges to plant production. With the rapid advancement of deep learning, the accuracy and efficiency of plant disease identification have substantially improved. However, conventional convolutional neural networks that rely on multi-layer small-kernel structures are limited in capturing long-range dependencies and global contextual information due to their constrained receptive fields. To overcome these limitations, this study proposes a plant disease recognition method based on RepLKNet, a convolutional architecture with large kernel designs that significantly expand the receptive field and enhance feature representation. Transfer learning is incorporated to further improve training efficiency and model performance. Experiments conducted on the Plant Diseases Training Dataset, comprising 95,865 images across 61 disease categories, demonstrate the effectiveness of the proposed method. Under five-fold cross-validation, the model achieved an overall accuracy (OA) of 96.03%, an average accuracy (AA) of 94.78%, and a Kappa coefficient of 95.86%. Compared with ResNet50 (OA: 95.62%) and GoogleNet (OA: 94.98%), the proposed model demonstrates competitive or superior performance. Ablation experiments reveal that replacing large kernels with 3×3 or 5×5 convolutions results in accuracy reductions of up to 1.1% in OA and 1.3% in AA, confirming the effectiveness of the large kernel design. These results demonstrate the robustness and superior capability of RepLKNet in plant disease recognition tasks.
Collapse
Affiliation(s)
- Guoquan Pei
- College of Big Data, Yunnan Agricultural University, Kunming, 650201, China
| | - Xueying Qian
- College of Big Data, Yunnan Agricultural University, Kunming, 650201, China
| | - Bing Zhou
- College of Science, Yunnan Agricultural University, Kunming, 650201, China
| | - Zigao Liu
- Yunnan Traceability Technology Co. Ltd., Kunming, 650201, China
| | - Wendou Wu
- College of Big Data, Yunnan Agricultural University, Kunming, 650201, China.
| |
Collapse
|
5
|
Subbarayudu C, Kubendiran M. Segmentation-based lightweight multi-class classification model for crop disease detection, classification, and severity assessment using DCNN. PLoS One 2025; 20:e0322705. [PMID: 40367226 PMCID: PMC12077804 DOI: 10.1371/journal.pone.0322705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2024] [Accepted: 03/26/2025] [Indexed: 05/16/2025] Open
Abstract
Leaf diseases in Zea mays crops have a significant impact on both the calibre and volume of maize yield, eventually impacting the market. Prior detection of the intensity of an infection would enable the efficient allocation of treatment resources and prevent the infection from spreading across the entire area. In this study, deep saliency map segmentation-based CNN is utilized for the detection, multi-class classification, and severity assessment of maize crop leaf diseases has been proposed. The proposed model involves seven different maize crop diseases such as Northern Leaf Blight Exserohilum turcicum, Eye Spot Oculimacula yallundae, Common Rust Puccinia sorghi, Goss's Bacterial Wilt Clavibacter michiganensis subsp. nebraskensis, Downy Mildew Pseudoperonospora, Phaeosphaeria leaf spot Phaeosphaeria maydis, Gray Leaf Spot Cercospora zeae-maydis, and Healthy are selected from publicly available datasets obtained from PlantVillage. After the disease-affected regions are identified, the features are extracted by using the EffiecientNet-B7. To classify the maize infection, a hybrid harris hawks' optimization (HHHO) is utilized for feature selection. Finally, from the optimized features obtained, classification and severity assessment are carried out with the help of Fuzzy SVM. Experimental analysis has been carried out to demonstrate the effectiveness of the proposed approach in detecting maize crop leaf diseases and assessing their severity. The proposed strategy was able to obtain an accuracy rate of around 99.47% on average. The work contributes to advancing automated disease diagnosis in agriculture, thereby supporting efforts for sustainable crop yield improvement and food security.
Collapse
Affiliation(s)
- Chatla Subbarayudu
- School of Computer Science and Engineering, Vellore Institute of Technology, Vellore, India
| | - Mohan Kubendiran
- School of Computer Science and Engineering, Vellore Institute of Technology, Vellore, India
| |
Collapse
|
6
|
Sharma J, Al-Huqail AA, Almogren A, Doshi H, Jayaprakash B, Bharathi B, Ur Rehman A, Hussen S. Deep learning based ensemble model for accurate tomato leaf disease classification by leveraging ResNet50 and MobileNetV2 architectures. Sci Rep 2025; 15:13904. [PMID: 40263518 PMCID: PMC12015254 DOI: 10.1038/s41598-025-98015-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2025] [Accepted: 04/08/2025] [Indexed: 04/24/2025] Open
Abstract
Global food security depends on tomato growing, but several fungal, bacterial, and viral illnesses seriously reduce productivity and quality, therefore causing major financial losses. Reducing these impacts depends on early, exact diagnosis of diseases. This work provides a deep learning-based ensemble model for tomato leaf disease classification combining MobileNetV2 and ResNet50. To improve feature extraction, the models were tweaked by changing their output layers with GlobalAverage Pooling2D, Batch Normalization, Dropout, and Dense layers. To take use of their complimentary qualities, the feature maps from both models were combined. This study uses a publicly available dataset from Kaggle for tomato leaf disease classification. Training on a dataset of 11,000 annotated pictures spanning 10 disease categories, including bacterial spot, early blight, late blight, leaf mold, septoria leaf spot, spider mites, target spot, yellow leaf curl virus, mosaic virus, and healthy leaves. Data preprocessing included image resizing and splitting, along with an 80-10-10 split, allocating 80% for training, 10% for testing, and 10% for validation to ensure a balanced evaluation. The proposed model with a 99.91% test accuracy, the suggested model was quite remarkable. Furthermore, guaranteeing strong classification performance across all disease categories, the model showed great precision (99.92%), recall (99.90%), and an F1-score of 99.91%. With few misclassifications, the confusion matrix verified almost flawless classification even further. These findings show how well deep learning can automate tomato disease diagnosis, therefore providing a scalable and quite accurate solution for smart agriculture. By means of early intervention and precision agriculture techniques, the suggested strategy has the potential to improve crop health monitoring, reduce economic losses, and encourage sustainable farming practices.
Collapse
Affiliation(s)
- Jatin Sharma
- Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab, India
| | - Asma A Al-Huqail
- Department of Botany and Microbiology, College of Science, King Saud University, P.O. Box 2455, Riyadh, 11451, Saudi Arabia
| | - Ahmad Almogren
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh, 11633, Saudi Arabia
| | - Hardik Doshi
- Marwadi University Research Center, Department of Computer Engineering, Faculty of Engineering & Technology Marwadi University, Rajkot, Gujarat, 360003, India
| | - B Jayaprakash
- Department of Computer Science & IT, School of Sciences, Jain (Deemed to be University), Bangalore, Karnataka, India
| | - B Bharathi
- Department of Computer Science and Engineering, Sathyabama Institute of Science and Technology, Chennai, Tamil Nadu, India
| | - Ateeq Ur Rehman
- School of Computing, Gachon University, Seongnam-si, 13120, Republic of Korea.
| | - Seada Hussen
- Department of Electrical Power, Adama Science and Technology University, Adama, 1888, Ethiopia.
| |
Collapse
|
7
|
Zhang J, Zhou H, Liu K, Xu Y. ED-Swin Transformer: A Cassava Disease Classification Model Integrated with UAV Images. SENSORS (BASEL, SWITZERLAND) 2025; 25:2432. [PMID: 40285122 PMCID: PMC12031189 DOI: 10.3390/s25082432] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/07/2025] [Revised: 04/01/2025] [Accepted: 04/10/2025] [Indexed: 04/29/2025]
Abstract
The outbreak of cassava diseases poses a serious threat to agricultural economic security and food production systems in tropical regions. Traditional manual monitoring methods are limited by efficiency bottlenecks and insufficient spatial coverage. Although low-altitude drone technology offers advantages such as high resolution and strong timeliness, it faces dual challenges in the field of disease identification, such as complex background interference and irregular disease morphology. To address these issues, this study proposes an intelligent classification method for cassava diseases based on drone imagery and an ED-Swin Transformer. Firstly, we introduced the EMAGE (Efficient Multi-Scale Attention with Grouping and Expansion) module, which integrates the global distribution features and local texture details of diseased leaves in drone imagery through a multi-scale grouped attention mechanism, effectively mitigating the interference of complex background noise on feature extraction. Secondly, the DASPP (Deformable Atrous Spatial Pyramid Pooling) module was designed to use deformable atrous convolution to adaptively match the irregular boundaries of diseased areas, enhancing the model's robustness to morphological variations caused by angles and occlusions in low-altitude drone photography. The results show that the ED-Swin Transformer model achieved excellent performance across five evaluation metrics, with scores of 94.32%, 94.56%, 98.56%, 89.22%, and 96.52%, representing improvements of 1.28%, 2.32%, 0.38%, 3.12%, and 1.4%, respectively. These experiments demonstrate the superior performance of the ED-Swin Transformer model in cassava classification networks.
Collapse
Affiliation(s)
- Jing Zhang
- College of Artificial Intelligence & Computer Science, Xi’an University of Science and Technology, Xi’an 710600, China; (H.Z.); (Y.X.)
| | - Hao Zhou
- College of Artificial Intelligence & Computer Science, Xi’an University of Science and Technology, Xi’an 710600, China; (H.Z.); (Y.X.)
| | - Kunyu Liu
- School of Economics and Management, Xidian University, Xi’an 710126, China;
| | - Yuguang Xu
- College of Artificial Intelligence & Computer Science, Xi’an University of Science and Technology, Xi’an 710600, China; (H.Z.); (Y.X.)
| |
Collapse
|
8
|
Wang Y, Wang Q, Su Y, Jing B, Feng M. Detection of kidney bean leaf spot disease based on a hybrid deep learning model. Sci Rep 2025; 15:11185. [PMID: 40169647 PMCID: PMC11961604 DOI: 10.1038/s41598-025-93742-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2025] [Accepted: 03/10/2025] [Indexed: 04/03/2025] Open
Abstract
Rapid diagnosis of kidney bean leaf spot disease is crucial for ensuring crop health and increasing yield. However, traditional machine learning methods face limitations in feature extraction, while deep learning approaches, despite their advantages, are computationally expensive and do not always yield optimal results. Moreover, reliable datasets for kidney bean leaf spot disease remain scarce. To address these challenges, this study constructs the first-ever kidney bean leaf spot disease (KBLD) dataset, filling a significant gap in the field. Based on this dataset, a novel hybrid deep learning model framework is proposed, which integrates deep learning models (EfficientNet-B7, MobileNetV3, ResNet50, and VGG16) for feature extraction with machine learning algorithms (Logistic Regression, Random Forest, AdaBoost, and Stochastic Gradient Boosting) for classification. By leveraging the Optuna tool for hyperparameter optimization, 16 combined models were evaluated. Experimental results show that the hybrid model combining EfficientNet-B7 and Stochastic Gradient Boosting achieves the highest detection accuracy of 96.26% on the KBLD dataset, with an F1-score of 0.97. The innovations of this study lie in the construction of a high-quality KBLD dataset and the development of a novel framework combining deep learning and machine learning, significantly improving the detection efficiency and accuracy of kidney bean leaf spot disease. This research provides a new approach for intelligent diagnosis and management of crop diseases in precision agriculture, contributing to increased agricultural productivity and ensuring food security.
Collapse
Affiliation(s)
- Yiwei Wang
- College of Agriculture, Shanxi Agricultural University, Jinzhong, China
| | - Qianyu Wang
- College of Agriculture, Shanxi Agricultural University, Jinzhong, China
| | - Yue Su
- College of Agriculture, Shanxi Agricultural University, Jinzhong, China
| | - Binghan Jing
- College of Agriculture, Shanxi Agricultural University, Jinzhong, China
| | - Meichen Feng
- College of Agriculture, Shanxi Agricultural University, Jinzhong, China.
| |
Collapse
|
9
|
Wang C, Xia Y, Xia L, Wang Q, Gu L. Dual discriminator GAN-based synthetic crop disease image generation for precise crop disease identification. PLANT METHODS 2025; 21:46. [PMID: 40159478 PMCID: PMC11955132 DOI: 10.1186/s13007-025-01361-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/02/2025] [Accepted: 03/09/2025] [Indexed: 04/02/2025]
Abstract
Deep learning-based computer vision technology significantly improves the accuracy and efficiency of crop disease detection. However, the scarcity of crop disease images leads to insufficient training data, limiting the accuracy of disease recognition and the generalization ability of deep learning models. Therefore, increasing the number and diversity of high-quality disease images is crucial for enhancing disease monitoring performance. We design a frequency-domain and wavelet image augmentation network with a dual discriminator structure (FHWD). The first discriminator distinguishes between real and generated images, while the second high-frequency discriminator is specifically used to distinguish between the high-frequency components of both. High-frequency details play a crucial role in the sharpness, texture, and fine-grained structures of an image, which are essential for realistic image generation. During training, we combine the proposed wavelet loss and Fast Fourier Transform loss functions. These loss functions guide the model to focus on image details through multi-band constraints and frequency domain transformation, improving the authenticity of lesions and textures, thereby enhancing the visual quality of the generated images. We compare the generation performance of different models on ten crop diseases from the PlantVillage dataset. The experimental results show that the images generated by FHWD contain more realistic leaf disease lesions, with higher image quality that better aligns with human visual perception. Additionally, in classification tasks involving nine types of tomato leaf diseases from the PlantVillage dataset, FHWD-enhanced data improve classification accuracy by an average of 7.25% for VGG16, GoogleNet, and ResNet18 models.Our results show that FHWD is an effective image augmentation tool that effectively addresses the scarcity of crop disease images and provides more diverse and enriched training data for disease recognition models.
Collapse
Affiliation(s)
- Chao Wang
- School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei, 230036, China
- Anhui Provincial Engineering Research Center for Agricultural Information Perception and Intelligent Computing, Hefei, China
- Key Laboratory of Agricultural Electronic Commerce of the Ministry of Agriculture, Hefei, China
| | - Yuting Xia
- School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei, 230036, China
- Anhui Provincial Engineering Research Center for Agricultural Information Perception and Intelligent Computing, Hefei, China
- Key Laboratory of Agricultural Electronic Commerce of the Ministry of Agriculture, Hefei, China
| | - Lunlong Xia
- School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei, 230036, China
- Anhui Provincial Engineering Research Center for Agricultural Information Perception and Intelligent Computing, Hefei, China
- Key Laboratory of Agricultural Electronic Commerce of the Ministry of Agriculture, Hefei, China
| | - Qingyong Wang
- School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei, 230036, China
- Anhui Provincial Engineering Research Center for Agricultural Information Perception and Intelligent Computing, Hefei, China
- Key Laboratory of Agricultural Electronic Commerce of the Ministry of Agriculture, Hefei, China
| | - Lichuan Gu
- School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei, 230036, China.
- Anhui Provincial Engineering Research Center for Agricultural Information Perception and Intelligent Computing, Hefei, China.
- Key Laboratory of Agricultural Electronic Commerce of the Ministry of Agriculture, Hefei, China.
| |
Collapse
|
10
|
Faisal HM, Aqib M, Rehman SU, Mahmood K, Obregon SA, Iglesias RC, Ashraf I. Detection of cotton crops diseases using customized deep learning model. Sci Rep 2025; 15:10766. [PMID: 40155421 PMCID: PMC11953249 DOI: 10.1038/s41598-025-94636-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2024] [Accepted: 03/17/2025] [Indexed: 04/01/2025] Open
Abstract
The agricultural industry is experiencing revolutionary changes through the latest advances in artificial intelligence and deep learning-based technologies. These powerful tools are being used for a variety of tasks including crop yield estimation, crop maturity assessment, and disease detection. The cotton crop is an essential source of revenue for many countries highlighting the need to protect it from deadly diseases that can drastically reduce yields. Early and accurate disease detection is quite crucial for preventing economic losses in the agricultural sector. Thanks to deep learning algorithms, researchers have developed innovative disease detection approaches that can help safeguard the cotton crop and promote economic growth. This study presents dissimilar state-of-the-art deep learning models for disease recognition including VGG16, DenseNet, EfficientNet, InceptionV3, MobileNet, NasNet, and ResNet models. For this purpose, real cotton disease data is collected from fields and preprocessed using different well-known techniques before using as input to deep learning models. Experimental analysis reveals that the ResNet152 model outperforms all other deep learning models, making it a practical and efficient approach for cotton disease recognition. By harnessing the power of deep learning and artificial intelligence, we can help protect the cotton crop and ensure a prosperous future for the agricultural sector.
Collapse
Affiliation(s)
- Hafiz Muhammad Faisal
- University Institute of Information Technology (UIIT), PMAS-Arid Agriculture University Rawalpindi, Rawalpindi, 46300, Pakistan
| | - Muhammad Aqib
- University Institute of Information Technology (UIIT), PMAS-Arid Agriculture University Rawalpindi, Rawalpindi, 46300, Pakistan
| | - Saif Ur Rehman
- University Institute of Information Technology (UIIT), PMAS-Arid Agriculture University Rawalpindi, Rawalpindi, 46300, Pakistan.
| | - Khalid Mahmood
- Institute of Computational Intelligence, Faculty of Computing, Gomal University, D.I. Khan, 29220, Pakistan
| | - Silvia Aparicio Obregon
- Universidad Europea del Atlántico, Isabel Torres 21, Santander, 39011, Spain
- Universidad Internacional Iberoamericana, Campeche, 24560, Mexico
- Universidad Internacional Iberoamericana Arecibo, Puerto Rico, 00613, USA
| | - Rubén Calderón Iglesias
- Universidad Europea del Atlántico, Isabel Torres 21, Santander, 39011, Spain
- Universidade Internacional do Cuanza, Cuito, Bie, Angola
- Universidad de La Romana, La Romana, Dominican Republic
| | - Imran Ashraf
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan, 38541, Republic of Korea.
| |
Collapse
|
11
|
Ekinci F, Ugurlu G, Ozcan GS, Acici K, Asuroglu T, Kumru E, Guzel MS, Akata I. Classification of Mycena and Marasmius Species Using Deep Learning Models: An Ecological and Taxonomic Approach. SENSORS (BASEL, SWITZERLAND) 2025; 25:1642. [PMID: 40292694 PMCID: PMC11945257 DOI: 10.3390/s25061642] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2025] [Revised: 02/25/2025] [Accepted: 03/05/2025] [Indexed: 04/30/2025]
Abstract
Fungi play a critical role in ecosystems, contributing to biodiversity and providing economic and biotechnological value. In this study, we developed a novel deep learning-based framework for the classification of seven macrofungi species from the genera Mycena and Marasmius, leveraging their unique ecological and morphological characteristics. The proposed approach integrates a custom convolutional neural network (CNN) with a self-organizing map (SOM) adapted for supervised learning and a Kolmogorov-Arnold Network (KAN) layer to enhance classification performance. The experimental results demonstrate significant improvements in classification metrics when using the CNN-SOM and CNN-KAN architectures. Additionally, advanced pretrained models such as MaxViT-S and ResNetV2-50 achieved high accuracy rates, with MaxViT-S achieving 98.9% accuracy. Statistical analyses using the chi-square test confirmed the reliability of the results, emphasizing the importance of validating evaluation metrics statistically. This research represents the first application of SOM in fungal classification and highlights the potential of deep learning in advancing fungal taxonomy. Future work will focus on optimizing the KAN architecture and expanding the dataset to include more fungal classes, further enhancing classification accuracy and ecological understanding.
Collapse
Affiliation(s)
- Fatih Ekinci
- Institute of Artificial Intelligence, Ankara University, Ankara 06100, Türkiye; (F.E.); (K.A.); (M.S.G.)
| | - Guney Ugurlu
- Department of Computer Engineering, Faculty of Engineering, Başkent University, Ankara 06790, Türkiye; (G.U.); (G.S.O.)
| | - Giray Sercan Ozcan
- Department of Computer Engineering, Faculty of Engineering, Başkent University, Ankara 06790, Türkiye; (G.U.); (G.S.O.)
| | - Koray Acici
- Institute of Artificial Intelligence, Ankara University, Ankara 06100, Türkiye; (F.E.); (K.A.); (M.S.G.)
- Department of Artificial Intelligence and Data Engineering, Faculty of Engineering, Ankara University, Ankara 06830, Türkiye
| | - Tunc Asuroglu
- Faculty of Medicine and Health Technology, Tampere University, 33720 Tampere, Finland
- VTT Technical Research Centre of Finland, 33101 Tampere, Finland
| | - Eda Kumru
- Graduate School of Natural and Applied Sciences, Ankara University, Ankara 06830, Türkiye;
| | - Mehmet Serdar Guzel
- Institute of Artificial Intelligence, Ankara University, Ankara 06100, Türkiye; (F.E.); (K.A.); (M.S.G.)
- Department of Computer Engineering, Faculty of Engineering, Ankara University, Ankara 06830, Türkiye
| | - Ilgaz Akata
- Department of Biology, Faculty of Science, Ankara University, Ankara 06100, Türkiye;
| |
Collapse
|
12
|
Bayram B, Kunduracioglu I, Ince S, Pacal I. A systematic review of deep learning in MRI-based cerebral vascular occlusion-based brain diseases. Neuroscience 2025; 568:76-94. [PMID: 39805420 DOI: 10.1016/j.neuroscience.2025.01.020] [Citation(s) in RCA: 15] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 01/09/2025] [Accepted: 01/10/2025] [Indexed: 01/16/2025]
Abstract
Neurological disorders, including cerebral vascular occlusions and strokes, present a major global health challenge due to their high mortality rates and long-term disabilities. Early diagnosis, particularly within the first hours, is crucial for preventing irreversible damage and improving patient outcomes. Although neuroimaging techniques like magnetic resonance imaging (MRI) have advanced significantly, traditional methods often fail to fully capture the complexity of brain lesions. Deep learning has recently emerged as a powerful tool in medical imaging, offering high accuracy in detecting and segmenting brain anomalies. This review examines 61 MRI-based studies published between 2020 and 2024, focusing on the role of deep learning in diagnosing cerebral vascular occlusion-related conditions. It evaluates the successes and limitations of these studies, including the adequacy and diversity of datasets, and addresses challenges such as data privacy and algorithm explainability. Comparisons between convolutional neural network (CNN)-based and Vision Transformer (ViT)-based approaches reveal distinct advantages and limitations. The findings emphasize the importance of ethically secure frameworks, the inclusion of diverse datasets, and improved model interpretability. Advanced architectures like U-Net variants and transformer-based models are highlighted as promising tools to enhance reliability in clinical applications. By automating complex neuroimaging tasks and improving diagnostic accuracy, deep learning facilitates personalized treatment strategies. This review provides a roadmap for integrating technical advancements into clinical practice, underscoring the transformative potential of deep learning in managing neurological disorders and improving healthcare outcomes globally.
Collapse
Affiliation(s)
- Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| | - Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| |
Collapse
|
13
|
Ozdemir B, Pacal I. An innovative deep learning framework for skin cancer detection employing ConvNeXtV2 and focal self-attention mechanisms. RESULTS IN ENGINEERING 2025; 25:103692. [DOI: 10.1016/j.rineng.2024.103692] [Citation(s) in RCA: 12] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
14
|
Pacal I, Işık G. Utilizing convolutional neural networks and vision transformers for precise corn leaf disease identification. Neural Comput Appl 2025; 37:2479-2496. [DOI: 10.1007/s00521-024-10769-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2024] [Accepted: 11/05/2024] [Indexed: 05/14/2025]
|
15
|
Ozdemir B, Aslan E, Pacal I. Attention Enhanced InceptionNeXt-Based Hybrid Deep Learning Model for Lung Cancer Detection. IEEE ACCESS 2025; 13:27050-27069. [DOI: 10.1109/access.2025.3539122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
Affiliation(s)
- Burhanettin Ozdemir
- Department of Operations and Project Management, College of Business, Alfaisal University, Riyadh, Saudi Arabia
| | - Emrah Aslan
- Department of Computer Engineering, Faculty of Engineering and Architecture, Mardin Artuklu University, Mardin, Türkiye
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, Iğdır, Türkiye
| |
Collapse
|
16
|
Pacal I, Alaftekin M, Zengul FD. Enhancing Skin Cancer Diagnosis Using Swin Transformer with Hybrid Shifted Window-Based Multi-head Self-attention and SwiGLU-Based MLP. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:3174-3192. [PMID: 38839675 PMCID: PMC11612041 DOI: 10.1007/s10278-024-01140-8] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Revised: 05/07/2024] [Accepted: 05/08/2024] [Indexed: 06/07/2024]
Abstract
Skin cancer is one of the most frequently occurring cancers worldwide, and early detection is crucial for effective treatment. Dermatologists often face challenges such as heavy data demands, potential human errors, and strict time limits, which can negatively affect diagnostic outcomes. Deep learning-based diagnostic systems offer quick, accurate testing and enhanced research capabilities, providing significant support to dermatologists. In this study, we enhanced the Swin Transformer architecture by implementing the hybrid shifted window-based multi-head self-attention (HSW-MSA) in place of the conventional shifted window-based multi-head self-attention (SW-MSA). This adjustment enables the model to more efficiently process areas of skin cancer overlap, capture finer details, and manage long-range dependencies, while maintaining memory usage and computational efficiency during training. Additionally, the study replaces the standard multi-layer perceptron (MLP) in the Swin Transformer with a SwiGLU-based MLP, an upgraded version of the gated linear unit (GLU) module, to achieve higher accuracy, faster training speeds, and better parameter efficiency. The modified Swin model-base was evaluated using the publicly accessible ISIC 2019 skin dataset with eight classes and was compared against popular convolutional neural networks (CNNs) and cutting-edge vision transformer (ViT) models. In an exhaustive assessment on the unseen test dataset, the proposed Swin-Base model demonstrated exceptional performance, achieving an accuracy of 89.36%, a recall of 85.13%, a precision of 88.22%, and an F1-score of 86.65%, surpassing all previously reported research and deep learning models documented in the literature.
Collapse
Affiliation(s)
- Ishak Pacal
- Department of Computer Engineering, Igdir University, 76000, Igdir, Turkey
| | - Melek Alaftekin
- Department of Computer Engineering, Igdir University, 76000, Igdir, Turkey
| | - Ferhat Devrim Zengul
- Department of Health Services Administration, The University of Alabama at Birmingham, Birmingham, AL, USA.
- Center for Integrated System, School of Engineering, The University of Alabama at Birmingham, Birmingham, AL, USA.
- Department of Biomedical Informatics and Data Science, School of Medicine, The University of Alabama, Birmingham, USA.
| |
Collapse
|
17
|
Gülmez B. Advancements in maize disease detection: A comprehensive review of convolutional neural networks. Comput Biol Med 2024; 183:109222. [PMID: 39388838 DOI: 10.1016/j.compbiomed.2024.109222] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2024] [Revised: 08/25/2024] [Accepted: 09/26/2024] [Indexed: 10/12/2024]
Abstract
This review article provides a comprehensive examination of the state-of-the-art in maize disease detection leveraging Convolutional Neural Networks (CNNs). Beginning with the intrinsic significance of plants and the pivotal role of maize in global agriculture, the increasing importance of detecting and mitigating maize diseases for ensuring food security is explored. The transformative potential of artificial intelligence, particularly CNNs, in automating the identification and diagnosis of maize diseases is investigated. Various aspects of the existing research landscape, including data sources, datasets, and the diversity of maize diseases covered, are scrutinized. A detailed analysis of data preprocessing strategies and data collection zones is conducted to add depth to the understanding of the field. The spectrum of algorithms and models employed in maize disease detection is comprehensively outlined, shedding light on their unique contributions and performance outcomes. The role of hyperparameter optimization techniques in refining model performance is explored across multiple studies. Performance metrics such as accuracy, precision, recall, F1 score, IoU, and mAP are systematically presented, offering insights into the efficacy of different CNN-based approaches. Challenges faced in maize disease detection are critically examined, emerging opportunities are identified, and future research directions are outlined. The review concludes by emphasizing the transformative impact of CNNs in revolutionizing maize disease detection while highlighting the need for ongoing research to address existing challenges and unlock the full potential of this technology.
Collapse
Affiliation(s)
- Burak Gülmez
- Department of Industrial Engineering, Mudanya University, 16940, Mudanya, Bursa, Turkiye; Leiden Institute of Advanced Computer Science, Leiden University, Leiden, Netherlands.
| |
Collapse
|
18
|
Maman A, Pacal I, Bati F. Can deep learning effectively diagnose cardiac amyloidosis with 99mTc-PYP scintigraphy? J Radioanal Nucl Chem 2024. [DOI: 10.1007/s10967-024-09879-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Accepted: 11/07/2024] [Indexed: 05/14/2025]
|
19
|
Fahim-Ul-Islam M, Chakrabarty A, Rahman R, Moon H, Piran MJ. Advancing mango leaf variant identification with a robust multi-layer perceptron model. Sci Rep 2024; 14:27406. [PMID: 39521776 PMCID: PMC11550809 DOI: 10.1038/s41598-024-74612-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2024] [Accepted: 09/27/2024] [Indexed: 11/16/2024] Open
Abstract
Mango, often regarded as the "king of fruits," holds a significant position in Bangladesh's agricultural landscape due to its popularity among the general population. However, identifying different types of mangoes, especially from mango leaves, poses a challenge for most people. While some studies have focused on mango type identification using fruit images, limited work has been done on classifying mango types based on leaf images. Early identification of mango types through leaf analysis is crucial for taking proactive steps in the cultivation process. This research introduces a novel multi-layer perceptron model called WaveVisionNet, designed to address this challenge using mango leaf datasets collected from five regions in Bangladesh. The MangoFolioBD dataset, comprising 16,646 annotated high-resolution images of mango leaves, has been curated and augmented to enhance robustness in real-world conditions. To validate the model, WaveVisionNet is evaluated on both the publicly available dataset and the MangoFolioBD dataset, achieving accuracy rates of 96.11% and 95.21%, respectively, outperforming state-of-the-art models such as Vision Transformer and transfer learning models. The model effectively combines the strengths of lightweight Convolutional Neural Networks and noise-resistant techniques, allowing for accurate analysis of mango leaf images while minimizing the impact of noise and environmental factors. The application of the WaveVisionNet model for automated mango leaf identification offers significant benefits to farmers, agricultural experts, agri-tech companies, government agencies, and consumers by enabling precise diagnosis of plant health, enhancing agricultural practices, and ultimately improving crop yields and quality.
Collapse
Affiliation(s)
- Md Fahim-Ul-Islam
- Department of Computer Science and Engineering, Brac University, Dhaka, Bangladesh
| | - Amitabha Chakrabarty
- Department of Computer Science and Engineering, Brac University, Dhaka, Bangladesh.
| | - Rafeed Rahman
- Department of Computer Science and Engineering, Brac University, Dhaka, Bangladesh
| | - Hyeonjoon Moon
- Department of Computer Science and Engineering, Sejong University, Seoul, 05006, South Korea
| | - Md Jalil Piran
- Department of Computer Science and Engineering, Sejong University, Seoul, 05006, South Korea.
| |
Collapse
|
20
|
Pacal I, Celik O, Bayram B, Cunha A. Enhancing EfficientNetv2 with global and efficient channel attention mechanisms for accurate MRI-Based brain tumor classification. CLUSTER COMPUTING 2024; 27:11187-11212. [DOI: 10.1007/s10586-024-04532-1] [Citation(s) in RCA: 15] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Revised: 04/08/2024] [Accepted: 04/22/2024] [Indexed: 05/14/2025]
Abstract
AbstractThe early and accurate diagnosis of brain tumors is critical for effective treatment planning, with Magnetic Resonance Imaging (MRI) serving as a key tool in the non-invasive examination of such conditions. Despite the advancements in Computer-Aided Diagnosis (CADx) systems powered by deep learning, the challenge of accurately classifying brain tumors from MRI scans persists due to the high variability of tumor appearances and the subtlety of early-stage manifestations. This work introduces a novel adaptation of the EfficientNetv2 architecture, enhanced with Global Attention Mechanism (GAM) and Efficient Channel Attention (ECA), aimed at overcoming these hurdles. This enhancement not only amplifies the model’s ability to focus on salient features within complex MRI images but also significantly improves the classification accuracy of brain tumors. Our approach distinguishes itself by meticulously integrating attention mechanisms that systematically enhance feature extraction, thereby achieving superior performance in detecting a broad spectrum of brain tumors. Demonstrated through extensive experiments on a large public dataset, our model achieves an exceptional high-test accuracy of 99.76%, setting a new benchmark in MRI-based brain tumor classification. Moreover, the incorporation of Grad-CAM visualization techniques sheds light on the model’s decision-making process, offering transparent and interpretable insights that are invaluable for clinical assessment. By addressing the limitations inherent in previous models, this study not only advances the field of medical imaging analysis but also highlights the pivotal role of attention mechanisms in enhancing the interpretability and accuracy of deep learning models for brain tumor diagnosis. This research sets the stage for advanced CADx systems, enhancing patient care and treatment outcomes.
Collapse
|
21
|
Lubbad MAH, Kurtulus IL, Karaboga D, Kilic K, Basturk A, Akay B, Nalbantoglu OU, Yilmaz OMD, Ayata M, Yilmaz S, Pacal I. A Comparative Analysis of Deep Learning-Based Approaches for Classifying Dental Implants Decision Support System. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2559-2580. [PMID: 38565730 PMCID: PMC11522249 DOI: 10.1007/s10278-024-01086-x] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/20/2024] [Revised: 02/28/2024] [Accepted: 03/12/2024] [Indexed: 04/04/2024]
Abstract
This study aims to provide an effective solution for the autonomous identification of dental implant brands through a deep learning-based computer diagnostic system. It also seeks to ascertain the system's potential in clinical practices and to offer a strategic framework for improving diagnosis and treatment processes in implantology. This study employed a total of 28 different deep learning models, including 18 convolutional neural network (CNN) models (VGG, ResNet, DenseNet, EfficientNet, RegNet, ConvNeXt) and 10 vision transformer models (Swin and Vision Transformer). The dataset comprises 1258 panoramic radiographs from patients who received implant treatments at Erciyes University Faculty of Dentistry between 2012 and 2023. It is utilized for the training and evaluation process of deep learning models and consists of prototypes from six different implant systems provided by six manufacturers. The deep learning-based dental implant system provided high classification accuracy for different dental implant brands using deep learning models. Furthermore, among all the architectures evaluated, the small model of the ConvNeXt architecture achieved an impressive accuracy rate of 94.2%, demonstrating a high level of classification success.This study emphasizes the effectiveness of deep learning-based systems in achieving high classification accuracy in dental implant types. These findings pave the way for integrating advanced deep learning tools into clinical practice, promising significant improvements in patient care and treatment outcomes.
Collapse
Affiliation(s)
- Mohammed A H Lubbad
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey.
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey.
| | | | - Dervis Karaboga
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | - Kerem Kilic
- Department of Prosthodontics, Dentistry Faculty, Erciyes University, Kayseri, Turkey
| | - Alper Basturk
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | - Bahriye Akay
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | - Ozkan Ufuk Nalbantoglu
- Department of Computer Engineering, Engineering Faculty, Erciyes University, 38039, Kayseri, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| | | | - Mustafa Ayata
- Department of Prosthodontics, Dentistry Faculty, Erciyes University, Kayseri, Turkey
| | - Serkan Yilmaz
- Department of Dentomaxillofacial Radiology, Dentistry Faculty, Erciyes University, Kayseri, Turkey
| | - Ishak Pacal
- Department of Computer Engineering, Engineering Faculty, Igdir University, Igdir, Turkey
- Artificial Intelligence and Big Data Application and Research Center, Erciyes University, Kayseri, Turkey
| |
Collapse
|
22
|
Pacal I, Kunduracioglu I, Alma MH, Deveci M, Kadry S, Nedoma J, Slany V, Martinek R. A systematic review of deep learning techniques for plant diseases. Artif Intell Rev 2024; 57:304. [DOI: 10.1007/s10462-024-10944-7] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/10/2024] [Indexed: 05/14/2025]
|
23
|
Pacal I. A novel Swin transformer approach utilizing residual multi-layer perceptron for diagnosing brain tumors in MRI images. INT J MACH LEARN CYB 2024; 15:3579-3597. [DOI: 10.1007/s13042-024-02110-w] [Citation(s) in RCA: 24] [Impact Index Per Article: 24.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 01/24/2024] [Indexed: 05/14/2025]
Abstract
AbstractSerious consequences due to brain tumors necessitate a timely and accurate diagnosis. However, obstacles such as suboptimal imaging quality, issues with data integrity, varying tumor types and stages, and potential errors in interpretation hinder the achievement of precise and prompt diagnoses. The rapid identification of brain tumors plays a pivotal role in ensuring patient safety. Deep learning-based systems hold promise in aiding radiologists to make diagnoses swiftly and accurately. In this study, we present an advanced deep learning approach based on the Swin Transformer. The proposed method introduces a novel Hybrid Shifted Windows Multi-Head Self-Attention module (HSW-MSA) along with a rescaled model. This enhancement aims to improve classification accuracy, reduce memory usage, and simplify training complexity. The Residual-based MLP (ResMLP) replaces the traditional MLP in the Swin Transformer, thereby improving accuracy, training speed, and parameter efficiency. We evaluate the Proposed-Swin model on a publicly available brain MRI dataset with four classes, using only test data. Model performance is enhanced through the application of transfer learning and data augmentation techniques for efficient and robust training. The Proposed-Swin model achieves a remarkable accuracy of 99.92%, surpassing previous research and deep learning models. This underscores the effectiveness of the Swin Transformer with HSW-MSA and ResMLP improvements in brain tumor diagnosis. This method introduces an innovative diagnostic approach using HSW-MSA and ResMLP in the Swin Transformer, offering potential support to radiologists in timely and accurate brain tumor diagnosis, ultimately improving patient outcomes and reducing risks.
Collapse
|
24
|
Li R, Su X, Zhang H, Zhang X, Yao Y, Zhou S, Zhang B, Ye M, Lv C. Integration of Diffusion Transformer and Knowledge Graph for Efficient Cucumber Disease Detection in Agriculture. PLANTS (BASEL, SWITZERLAND) 2024; 13:2435. [PMID: 39273919 PMCID: PMC11396938 DOI: 10.3390/plants13172435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/29/2024] [Revised: 08/15/2024] [Accepted: 08/30/2024] [Indexed: 09/15/2024]
Abstract
In this study, a deep learning method combining knowledge graph and diffusion Transformer has been proposed for cucumber disease detection. By incorporating the diffusion attention mechanism and diffusion loss function, the research aims to enhance the model's ability to recognize complex agricultural disease features and to address the issue of sample imbalance efficiently. Experimental results demonstrate that the proposed method outperforms existing deep learning models in cucumber disease detection tasks. Specifically, the method achieved a precision of 93%, a recall of 89%, an accuracy of 92%, and a mean average precision (mAP) of 91%, with a frame rate of 57 frames per second (FPS). Additionally, the study successfully implemented model lightweighting, enabling effective operation on mobile devices, which supports rapid on-site diagnosis of cucumber diseases. The research not only optimizes the performance of cucumber disease detection, but also opens new possibilities for the application of deep learning in the field of agricultural disease detection.
Collapse
Affiliation(s)
- Ruiheng Li
- China Agricultural University, Beijing 100083, China
| | - Xiaotong Su
- China Agricultural University, Beijing 100083, China
| | - Hang Zhang
- China Agricultural University, Beijing 100083, China
| | - Xiyan Zhang
- China Agricultural University, Beijing 100083, China
| | - Yifan Yao
- China Agricultural University, Beijing 100083, China
| | - Shutian Zhou
- China Agricultural University, Beijing 100083, China
| | - Bohan Zhang
- China Agricultural University, Beijing 100083, China
| | - Muyang Ye
- China Agricultural University, Beijing 100083, China
| | - Chunli Lv
- China Agricultural University, Beijing 100083, China
| |
Collapse
|
25
|
Yang J, Zhu W, Liu G, Dai W, Xu Z, Wan L, Zhou G. ICPNet: Advanced Maize Leaf Disease Detection with Multidimensional Attention and Coordinate Depthwise Convolution. PLANTS (BASEL, SWITZERLAND) 2024; 13:2277. [PMID: 39204713 PMCID: PMC11359242 DOI: 10.3390/plants13162277] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/22/2024] [Revised: 08/13/2024] [Accepted: 08/13/2024] [Indexed: 09/04/2024]
Abstract
Maize is an important crop, and the detection of maize diseases is critical for ensuring food security and improving agricultural production efficiency. To address the challenges of difficult feature extraction due to the high similarity among maize leaf disease species, the blurring of image edge features, and the susceptibility of maize leaf images to noise during acquisition and transmission, we propose a maize disease detection method based on ICPNet (Integrated multidimensional attention coordinate depthwise convolution PSO (Particle Swarm Optimization)-Integrated lion optimisation algorithm network). Firstly, we introduce a novel attention mechanism called Integrated Multidimensional Attention (IMA), which enhances the stability and responsiveness of the model in detecting small speckled disease features by combining cross-attention and spatial channel reconstruction methods. Secondly, we propose Coordinate Depthwise Convolution (CDC) to enhance the accuracy of feature maps through multi-scale convolutional processing, allowing for better differentiation of the fuzzy edges of maize leaf disease regions. To further optimize model performance, we introduce the PSO-Integrated Lion Optimisation Algorithm (PLOA), which leverages the exploratory stochasticity and annealing mechanism of the particle swarm algorithm to enhance the model's ability to handle mutation points while maintaining training stability and robustness. The experimental results demonstrate that ICPNet achieved an average accuracy of 88.4% and a precision of 87.3% on the self-constructed dataset. This method effectively extracts the tiny and fuzzy edge features of maize leaf diseases, providing a valuable reference for disease control in large-scale maize production.
Collapse
Affiliation(s)
- Jin Yang
- College of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China; (J.Y.); (G.L.); (W.D.); (L.W.); (G.Z.)
| | - Wenke Zhu
- College of Bangor, Central South University of Forestry and Technology, Changsha 410004, China;
| | - Guanqi Liu
- College of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China; (J.Y.); (G.L.); (W.D.); (L.W.); (G.Z.)
| | - Weisi Dai
- College of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China; (J.Y.); (G.L.); (W.D.); (L.W.); (G.Z.)
| | - Zhuonong Xu
- College of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China; (J.Y.); (G.L.); (W.D.); (L.W.); (G.Z.)
| | - Li Wan
- College of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China; (J.Y.); (G.L.); (W.D.); (L.W.); (G.Z.)
| | - Guoxiong Zhou
- College of Electronic Information and Physics, Central South University of Forestry and Technology, Changsha 410004, China; (J.Y.); (G.L.); (W.D.); (L.W.); (G.Z.)
| |
Collapse
|
26
|
Kunduracioglu I, Pacal I. Advancements in deep learning for accurate classification of grape leaves and diagnosis of grape diseases. JOURNAL OF PLANT DISEASES AND PROTECTION 2024; 131:1061-1080. [DOI: 10.1007/s41348-024-00896-z] [Citation(s) in RCA: 18] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Accepted: 02/20/2024] [Indexed: 05/14/2025]
Abstract
AbstractPlant diseases cause significant agricultural losses, demanding accurate detection methods. Traditional approaches relying on expert knowledge may be biased, but advancements in computing, particularly deep learning, offer non-experts effective tools. This study focuses on fine-tuning cutting-edge pre-trained CNN and vision transformer models to classify grape leaves and diagnose grape leaf diseases through digital images. Our research examined a PlantVillage dataset, which comprises 4062 leaf images distributed across four categories. Additionally, we utilized the Grapevine dataset, consisting of 500 leaf images. This dataset is organized into five distinct groups, with each group containing 100 images corresponding to one of the five grape types. The PlantVillage dataset focuses on four classes related to grape diseases, namely Black Rot, Leaf Blight, Healthy, and Esca leaves. On the other hand, the Grapevine dataset includes five classes for leaf recognition, specifically Ak, Alaidris, Buzgulu, Dimnit, and Nazli. In experiments with 14 CNN and 17 vision transformer models, deep learning demonstrated high accuracy in distinguishing grape diseases and recognizing leaves. Notably, four models achieved 100% accuracy on PlantVillage and Grapevine datasets, with Swinv2-Base standing out. This approach holds promise for enhancing crop productivity through early disease detection and providing insights into grape variety characterization in agriculture.
Collapse
|
27
|
Allogmani AS, Mohamed RM, Al-Shibly NM, Ragab M. Enhanced cervical precancerous lesions detection and classification using Archimedes Optimization Algorithm with transfer learning. Sci Rep 2024; 14:12076. [PMID: 38802525 PMCID: PMC11130149 DOI: 10.1038/s41598-024-62773-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Accepted: 05/21/2024] [Indexed: 05/29/2024] Open
Abstract
Cervical cancer (CC) ranks as the fourth most common form of cancer affecting women, manifesting in the cervix. CC is caused by the Human papillomavirus (HPV) infection and is eradicated by vaccinating women from an early age. However, limited medical facilities present a significant challenge in mid- or low-income countries. It can improve the survivability rate and be successfully treated if the CC is detected at earlier stages. Current technological improvements allow for cost-effective, more sensitive, and rapid screening and treatment measures for CC. DL techniques are widely adopted for the automated detection of CC. DL techniques and architectures are used to detect CC and provide higher detection performance. This study offers the design of Enhanced Cervical Precancerous Lesions Detection and Classification using the Archimedes Optimization Algorithm with Transfer Learning (CPLDC-AOATL) algorithm. The CPLDC-AOATL algorithm aims to diagnose cervical cancer using medical images. At the preliminary stage, the CPLDC-AOATL technique involves a bilateral filtering (BF) technique to eliminate the noise in the input images. Besides, the CPLDC-AOATL technique applies the Inception-ResNetv2 model for the feature extraction process, and the use of AOA chose the hyperparameters. The CPLDC-AOATL technique involves a bidirectional long short-term memory (BiLSTM) model for the cancer detection process. The experimental outcome of the CPLDC-AOATL technique emphasized the superior accuracy outcome of 99.53% over other existing approaches under a benchmark dataset.
Collapse
Affiliation(s)
- Ayed S Allogmani
- University of Jeddah, College of Science and Arts at Khulis, Department of Biology, Jeddah, Saudi Arabia
| | - Roushdy M Mohamed
- University of Jeddah, College of Science and Arts at Khulis, Department of Biology, Jeddah, Saudi Arabia.
| | - Nasser M Al-Shibly
- Physiotherapy Department, College of Applied Health Sciences, Jerash University, Jerash, Jordan
| | - Mahmoud Ragab
- Information Technology Department, Faculty of Computing and Information Technology, King Abdulaziz University, 21589, Jeddah, Saudi Arabia.
| |
Collapse
|
28
|
Pacal I. MaxCerVixT: A novel lightweight vision transformer-based Approach for precise cervical cancer detection. Knowl Based Syst 2024; 289:111482. [DOI: 10.1016/j.knosys.2024.111482] [Citation(s) in RCA: 35] [Impact Index Per Article: 35.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/14/2025]
|
29
|
Sikkandar MY, Sundaram SG, Alassaf A, AlMohimeed I, Alhussaini K, Aleid A, Alolayan SA, Ramkumar P, Almutairi MK, Begum SS. Utilizing adaptive deformable convolution and position embedding for colon polyp segmentation with a visual transformer. Sci Rep 2024; 14:7318. [PMID: 38538774 PMCID: PMC11377543 DOI: 10.1038/s41598-024-57993-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Accepted: 03/24/2024] [Indexed: 09/07/2024] Open
Abstract
Polyp detection is a challenging task in the diagnosis of Colorectal Cancer (CRC), and it demands clinical expertise due to the diverse nature of polyps. The recent years have witnessed the development of automated polyp detection systems to assist the experts in early diagnosis, considerably reducing the time consumption and diagnostic errors. In automated CRC diagnosis, polyp segmentation is an important step which is carried out with deep learning segmentation models. Recently, Vision Transformers (ViT) are slowly replacing these models due to their ability to capture long range dependencies among image patches. However, the existing ViTs for polyp do not harness the inherent self-attention abilities and incorporate complex attention mechanisms. This paper presents Polyp-Vision Transformer (Polyp-ViT), a novel Transformer model based on the conventional Transformer architecture, which is enhanced with adaptive mechanisms for feature extraction and positional embedding. Polyp-ViT is tested on the Kvasir-seg and CVC-Clinic DB Datasets achieving segmentation accuracies of 0.9891 ± 0.01 and 0.9875 ± 0.71 respectively, outperforming state-of-the-art models. Polyp-ViT is a prospective tool for polyp segmentation which can be adapted to other medical image segmentation tasks as well due to its ability to generalize well.
Collapse
Affiliation(s)
- Mohamed Yacin Sikkandar
- Department of Medical Equipment Technology, College of Applied Medical Sciences, Majmaah University, Al Majmaah, 11952, Saudi Arabia.
| | - Sankar Ganesh Sundaram
- Department of Artificial Intelligence and Data Science, KPR Institute of Engineering and Technology, Coimbatore, 641407, India
| | - Ahmad Alassaf
- Department of Medical Equipment Technology, College of Applied Medical Sciences, Majmaah University, Al Majmaah, 11952, Saudi Arabia
| | - Ibrahim AlMohimeed
- Department of Medical Equipment Technology, College of Applied Medical Sciences, Majmaah University, Al Majmaah, 11952, Saudi Arabia
| | - Khalid Alhussaini
- Department of Biomedical Technology, College of Applied Medical Sciences, King Saud University, Riyadh, 12372, Saudi Arabia
| | - Adham Aleid
- Department of Biomedical Technology, College of Applied Medical Sciences, King Saud University, Riyadh, 12372, Saudi Arabia
| | - Salem Ali Alolayan
- Department of Medical Equipment Technology, College of Applied Medical Sciences, Majmaah University, Al Majmaah, 11952, Saudi Arabia
| | - P Ramkumar
- Department of Computer Science and Engineering, Sri Sairam College of Engineering, Anekal, Bengaluru, 562106, Karnataka, India
| | - Meshal Khalaf Almutairi
- Department of Medical Equipment Technology, College of Applied Medical Sciences, Majmaah University, Al Majmaah, 11952, Saudi Arabia
| | - S Sabarunisha Begum
- Department of Biotechnology, P.S.R. Engineering College, Sivakasi, 626140, India
| |
Collapse
|
30
|
Kutluyarov RV, Zakoyan AG, Voronkov GS, Grakhova EP, Butt MA. Neuromorphic Photonics Circuits: Contemporary Review. NANOMATERIALS (BASEL, SWITZERLAND) 2023; 13:3139. [PMID: 38133036 PMCID: PMC10745993 DOI: 10.3390/nano13243139] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/13/2023] [Revised: 12/08/2023] [Accepted: 12/10/2023] [Indexed: 12/23/2023]
Abstract
Neuromorphic photonics is a cutting-edge fusion of neuroscience-inspired computing and photonics technology to overcome the constraints of conventional computing architectures. Its significance lies in the potential to transform information processing by mimicking the parallelism and efficiency of the human brain. Using optics and photonics principles, neuromorphic devices can execute intricate computations swiftly and with impressive energy efficiency. This innovation holds promise for advancing artificial intelligence and machine learning while addressing the limitations of traditional silicon-based computing. Neuromorphic photonics could herald a new era of computing that is more potent and draws inspiration from cognitive processes, leading to advancements in robotics, pattern recognition, and advanced data processing. This paper reviews the recent developments in neuromorphic photonic integrated circuits, applications, and current challenges.
Collapse
Affiliation(s)
- Ruslan V. Kutluyarov
- School of Photonics Engineering and Research Advances (SPhERA), Ufa University of Science and Technology, 32, Z. Validi St., 450076 Ufa, Russia
| | - Aida G. Zakoyan
- School of Photonics Engineering and Research Advances (SPhERA), Ufa University of Science and Technology, 32, Z. Validi St., 450076 Ufa, Russia
| | - Grigory S. Voronkov
- School of Photonics Engineering and Research Advances (SPhERA), Ufa University of Science and Technology, 32, Z. Validi St., 450076 Ufa, Russia
| | - Elizaveta P. Grakhova
- School of Photonics Engineering and Research Advances (SPhERA), Ufa University of Science and Technology, 32, Z. Validi St., 450076 Ufa, Russia
| | | |
Collapse
|