1
|
Sinamenye JH, Chatterjee A, Shrestha R. Potato plant disease detection: leveraging hybrid deep learning models. BMC PLANT BIOLOGY 2025; 25:647. [PMID: 40380088 PMCID: PMC12082912 DOI: 10.1186/s12870-025-06679-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/09/2024] [Accepted: 05/05/2025] [Indexed: 05/19/2025]
Abstract
Agriculture, a crucial sector for global economic development and sustainable food production, faces significant challenges in detecting and managing crop diseases. These diseases can greatly impact yield and productivity, making early and accurate detection vital, especially in staple crops like potatoes. Traditional manual methods, as well as some existing machine learning and deep learning techniques, often lack accuracy and generalizability due to factors such as variability in real-world conditions. This study proposes a novel approach to improve potato plant disease detection and identification using a hybrid deep-learning model, EfficientNetV2B3+ViT. This model combines the strengths of a Convolutional Neural Network - EfficientNetV2B3 and a Vision Transformer (ViT). It has been trained on a diverse potato leaf image dataset, the "Potato Leaf Disease Dataset", which reflects real-world agricultural conditions. The proposed model achieved an accuracy of 85.06 % , representing an 11.43 % improvement over the results of the previous study. These results highlight the effectiveness of the hybrid model in complex agricultural settings and its potential to improve potato plant disease detection and identification.
Collapse
Affiliation(s)
| | - Ayan Chatterjee
- Department of Digital Technology, STIFTELSEN NILU, Kjeller, Norway
| | - Raju Shrestha
- Department of Computer Science, Oslo Metropolitan University (OsloMet), Oslo, Norway
| |
Collapse
|
2
|
Zhao Y, Wang S, Zeng Q, Ni W, Duan H, Xie N, Xiao F. Informed-Learning-Guided Visual Question Answering Model of Crop Disease. PLANT PHENOMICS (WASHINGTON, D.C.) 2024; 6:0277. [PMID: 39687877 PMCID: PMC11649200 DOI: 10.34133/plantphenomics.0277] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/24/2024] [Revised: 10/18/2024] [Accepted: 11/12/2024] [Indexed: 12/18/2024]
Abstract
In contemporary agriculture, experts develop preventative and remedial strategies for various disease stages in diverse crops. Decision-making regarding the stages of disease occurrence exceeds the capabilities of single-image tasks, such as image classification and object detection. Consequently, research now focuses on training visual question answering (VQA) models. However, existing studies concentrate on identifying disease species rather than formulating questions that encompass crucial multiattributes. Additionally, model performance is susceptible to the model structure and dataset biases. To address these challenges, we construct the informed-learning-guided VQA model of crop disease (ILCD). ILCD improves model performance by integrating coattention, a multimodal fusion model (MUTAN), and a bias-balancing (BiBa) strategy. To facilitate the investigation of various visual attributes of crop diseases and the determination of disease occurrence stages, we construct a new VQA dataset called the Crop Disease Multi-attribute VQA with Prior Knowledge (CDwPK-VQA). This dataset contains comprehensive information on various visual attributes such as shape, size, status, and color. We expand the dataset by integrating prior knowledge into CDwPK-VQA to address performance challenges. Comparative experiments are conducted by ILCD on the VQA-v2, VQA-CP v2, and CDwPK-VQA datasets, achieving accuracies of 68.90%, 49.75%, and 86.06%, respectively. Ablation experiments are conducted on CDwPK-VQA to evaluate the effectiveness of various modules, including coattention, MUTAN, and BiBa. These experiments demonstrate that ILCD exhibits the highest level of accuracy, performance, and value in the field of agriculture. The source codes can be accessed at https://github.com/SdustZYP/ILCD-master/tree/main.
Collapse
Affiliation(s)
- Yunpeng Zhao
- College of Computer Science and Engineering,
Shandong University of Science and Technology, Qingdao 266590, China
| | - Shansong Wang
- College of Computer Science and Engineering,
Shandong University of Science and Technology, Qingdao 266590, China
| | - Qingtian Zeng
- College of Computer Science and Engineering,
Shandong University of Science and Technology, Qingdao 266590, China
| | - Weijian Ni
- College of Computer Science and Engineering,
Shandong University of Science and Technology, Qingdao 266590, China
| | - Hua Duan
- College of Computer Science and Engineering,
Shandong University of Science and Technology, Qingdao 266590, China
| | - Nengfu Xie
- Agricultural Information Institute of CAAS, Beijing 100081, China
| | | |
Collapse
|
3
|
Cui Y, Yang Y, Xia Y, Li Y, Feng Z, Liu S, Yuan G, Lv C. An Efficient Weed Detection Method Using Latent Diffusion Transformer for Enhanced Agricultural Image Analysis and Mobile Deployment. PLANTS (BASEL, SWITZERLAND) 2024; 13:3192. [PMID: 39599401 PMCID: PMC11598004 DOI: 10.3390/plants13223192] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/27/2024] [Revised: 11/05/2024] [Accepted: 11/07/2024] [Indexed: 11/29/2024]
Abstract
This paper presents an efficient weed detection method based on the latent diffusion transformer, aimed at enhancing the accuracy and applicability of agricultural image analysis. The experimental results demonstrate that the proposed model achieves a precision of 0.92, a recall of 0.89, an accuracy of 0.91, a mean average precision (mAP) of 0.91, and an F1 score of 0.90, indicating its outstanding performance in complex scenarios. Additionally, ablation experiments reveal that the latent-space-based diffusion subnetwork outperforms traditional models, such as the the residual diffusion network, which has a precision of only 0.75. By combining latent space feature extraction with self-attention mechanisms, the constructed lightweight model can respond quickly on mobile devices, showcasing the significant potential of deep learning technologies in agricultural applications. Future research will focus on data diversity and model interpretability to further enhance the model's adaptability and user trust.
Collapse
Affiliation(s)
- Yuzhuo Cui
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| | - Yingqiu Yang
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| | - Yuqing Xia
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| | - Yan Li
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| | - Zhaoxi Feng
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| | - Shiya Liu
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| | - Guangqi Yuan
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
- School of English and International Studies, Beijing Foreign Studies University, Beijing 100089, China
| | - Chunli Lv
- College of Electrical and Information Engineering, China Agricultural University, Beijing 100083, China
| |
Collapse
|
4
|
Ji M, Zhou Z, Wang X, Tang W, Li Y, Wang Y, Zhou C, Lv C. Implementing Real-Time Image Processing for Radish Disease Detection Using Hybrid Attention Mechanisms. PLANTS (BASEL, SWITZERLAND) 2024; 13:3001. [PMID: 39519918 PMCID: PMC11548246 DOI: 10.3390/plants13213001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/11/2024] [Revised: 10/17/2024] [Accepted: 10/25/2024] [Indexed: 11/16/2024]
Abstract
This paper developed a radish disease detection system based on a hybrid attention mechanism, significantly enhancing the precision and real-time performance in identifying disease characteristics. By integrating spatial and channel attentions, this system demonstrated superior performance across numerous metrics, particularly achieving 93% precision and 91% accuracy in detecting radish virus disease, outperforming existing technologies. Additionally, the introduction of the hybrid attention mechanism proved its superiority in ablation experiments, showing higher performance compared to standard self-attention and the convolutional block attention module. The study also introduced a hybrid loss function that combines cross-entropy loss and Dice loss, effectively addressing the issue of class imbalance and further enhancing the detection capability for rare diseases. These experimental results not only validate the effectiveness of the proposed method, but also provide robust technical support for the rapid and accurate detection of radish diseases, demonstrating its vast potential in agricultural applications. Future research will continue to optimize the model structure and computational efficiency to accommodate a broader range of agricultural disease detection needs.
Collapse
Affiliation(s)
- Mengxue Ji
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| | - Zizhe Zhou
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| | - Xinyue Wang
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| | - Weidong Tang
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| | - Yan Li
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| | - Yilin Wang
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| | - Chaoyu Zhou
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
- School of Computer and Cyberspace Security, Communication University of China, Beijing 100024, China
| | - Chunli Lv
- China Agricultural University, Beijing 100083, China; (M.J.); (Z.Z.); (X.W.); (W.T.); (Y.L.); (Y.W.); (C.Z.)
| |
Collapse
|
5
|
Khan IU, Khan HA, Lee JW. Dual-Stream Architecture Enhanced by Soft-Attention Mechanism for Plant Species Classification. PLANTS (BASEL, SWITZERLAND) 2024; 13:2655. [PMID: 39339630 PMCID: PMC11435159 DOI: 10.3390/plants13182655] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/12/2024] [Revised: 09/14/2024] [Accepted: 09/18/2024] [Indexed: 09/30/2024]
Abstract
Plants play a vital role in numerous domains, including medicine, agriculture, and environmental balance. Furthermore, they contribute to the production of oxygen and the retention of carbon dioxide, both of which are necessary for living beings. Numerous researchers have conducted thorough research in the classification of plant species where certain studies have focused on limited numbers of classes, while others have employed conventional machine-learning and deep-learning models to classify them. To address these limitations, this paper introduces a novel dual-stream neural architecture embedded with a soft-attention mechanism specifically developed for accurately classifying plant species. The proposed model utilizes residual and inception blocks enhanced with dilated convolutional layers for acquiring both local and global information. Following the extraction of features, both streams are combined, and a soft-attention technique is used to improve the distinct characteristics. The efficacy of the model is shown via extensive experimentation on varied datasets, including several plant species. Moreover, we have contributed a novel dataset that comprises 48 classes of different plant species. The results demonstrate a higher level of performance when compared to current models, emphasizing the capability of the dual-stream design in improving accuracy and model generalization. The integration of a dual-stream architecture, dilated convolutions, and soft attention provides a strong and reliable foundation for the botanical community, supporting advancement in the field of plant species classification.
Collapse
Affiliation(s)
- Imran Ullah Khan
- Mixed Reality and Interaction Laboratory, Department of Software, Sejong University, Seoul 05006, Republic of Korea
| | - Haseeb Ali Khan
- Mixed Reality and Interaction Laboratory, Department of Software, Sejong University, Seoul 05006, Republic of Korea
| | - Jong Weon Lee
- Mixed Reality and Interaction Laboratory, Department of Software, Sejong University, Seoul 05006, Republic of Korea
| |
Collapse
|
6
|
Marković D, Stamenković Z, Đorđević B, Ranđić S. Image Processing for Smart Agriculture Applications Using Cloud-Fog Computing. SENSORS (BASEL, SWITZERLAND) 2024; 24:5965. [PMID: 39338710 PMCID: PMC11435679 DOI: 10.3390/s24185965] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/11/2024] [Revised: 09/05/2024] [Accepted: 09/12/2024] [Indexed: 09/30/2024]
Abstract
The widespread use of IoT devices has led to the generation of a huge amount of data and driven the need for analytical solutions in many areas of human activities, such as the field of smart agriculture. Continuous monitoring of crop growth stages enables timely interventions, such as control of weeds and plant diseases, as well as pest control, ensuring optimal development. Decision-making systems in smart agriculture involve image analysis with the potential to increase productivity, efficiency and sustainability. By applying Convolutional Neural Networks (CNNs), state recognition and classification can be performed based on images from specific locations. Thus, we have developed a solution for early problem detection and resource management optimization. The main concept of the proposed solution relies on a direct connection between Cloud and Edge devices, which is achieved through Fog computing. The goal of our work is creation of a deep learning model for image classification that can be optimized and adapted for implementation on devices with limited hardware resources at the level of Fog computing. This could increase the importance of image processing in the reduction of agricultural operating costs and manual labor. As a result of the off-load data processing at Edge and Fog devices, the system responsiveness can be improved, the costs associated with data transmission and storage can be reduced, and the overall system reliability and security can be increased. The proposed solution can choose classification algorithms to find a trade-off between size and accuracy of the model optimized for devices with limited hardware resources. After testing our model for tomato disease classification compiled for execution on FPGA, it was found that the decrease in test accuracy is as small as 0.83% (from 96.29% to 95.46%).
Collapse
Affiliation(s)
- Dušan Marković
- Faculty of Agronomy in Čačak, University of Kragujevac, Cara Dušana 34, 32102 Čačak, Serbia
| | - Zoran Stamenković
- Institute of Computer Science, University of Potsdam, An der Bahn 2, 14476 Potsdam, Germany
- IHP—Leibniz-Institutfür innovative Mikroelektronik, ImTechnologiepark 25, 15236 Frankfurt, Germany
| | | | - Siniša Ranđić
- Faculty of Technical Sciences Čačak, University of Kragujevac, Svetog Save 65, 32102 Čačak, Serbia;
| |
Collapse
|
7
|
Li R, Su X, Zhang H, Zhang X, Yao Y, Zhou S, Zhang B, Ye M, Lv C. Integration of Diffusion Transformer and Knowledge Graph for Efficient Cucumber Disease Detection in Agriculture. PLANTS (BASEL, SWITZERLAND) 2024; 13:2435. [PMID: 39273919 PMCID: PMC11396938 DOI: 10.3390/plants13172435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/29/2024] [Revised: 08/15/2024] [Accepted: 08/30/2024] [Indexed: 09/15/2024]
Abstract
In this study, a deep learning method combining knowledge graph and diffusion Transformer has been proposed for cucumber disease detection. By incorporating the diffusion attention mechanism and diffusion loss function, the research aims to enhance the model's ability to recognize complex agricultural disease features and to address the issue of sample imbalance efficiently. Experimental results demonstrate that the proposed method outperforms existing deep learning models in cucumber disease detection tasks. Specifically, the method achieved a precision of 93%, a recall of 89%, an accuracy of 92%, and a mean average precision (mAP) of 91%, with a frame rate of 57 frames per second (FPS). Additionally, the study successfully implemented model lightweighting, enabling effective operation on mobile devices, which supports rapid on-site diagnosis of cucumber diseases. The research not only optimizes the performance of cucumber disease detection, but also opens new possibilities for the application of deep learning in the field of agricultural disease detection.
Collapse
Affiliation(s)
- Ruiheng Li
- China Agricultural University, Beijing 100083, China
| | - Xiaotong Su
- China Agricultural University, Beijing 100083, China
| | - Hang Zhang
- China Agricultural University, Beijing 100083, China
| | - Xiyan Zhang
- China Agricultural University, Beijing 100083, China
| | - Yifan Yao
- China Agricultural University, Beijing 100083, China
| | - Shutian Zhou
- China Agricultural University, Beijing 100083, China
| | - Bohan Zhang
- China Agricultural University, Beijing 100083, China
| | - Muyang Ye
- China Agricultural University, Beijing 100083, China
| | - Chunli Lv
- China Agricultural University, Beijing 100083, China
| |
Collapse
|
8
|
Isinkaye FO, Olusanya MO, Singh PK. Deep learning and content-based filtering techniques for improving plant disease identification and treatment recommendations: A comprehensive review. Heliyon 2024; 10:e29583. [PMID: 38737274 PMCID: PMC11088271 DOI: 10.1016/j.heliyon.2024.e29583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Revised: 03/30/2024] [Accepted: 04/10/2024] [Indexed: 05/14/2024] Open
Abstract
The importance of identifying plant diseases has risen recently due to the adverse effect they have on agricultutal production. Plant diseases have been a big concern in agriculture, as they affect crop production, and constitute a major threat to global food security. In the domain of modern agriculture, effective plant disease management is vital to ensure healthy crop yields and sustainable practices. Traditional means of identifying plant disease are faced with lots of challenges and the need for better and efficient detection methods cannot be overemphazised. The emergence of advanced technologies, particularly deep learning and content-based filtering techniques, if integrated together can changed the way plant diseases are identified and treated. Such as speedy and correct identification of plant diseases and efficient treatment recommendations which are keys for sustainable food production. In this work, We try to investigate the current state of research, identified gaps and limitations in knowledge, and suggests future directions for researchers, experts and farmers that could help to provide better ways of mitigating plant disease problems.
Collapse
Affiliation(s)
- Folasade Olubusola Isinkaye
- Department of Computer Science and Information Technology, Sol Plaatje University Kimberley, 8301, South Africa
| | - Michael Olusoji Olusanya
- Department of Computer Science and Information Technology, Sol Plaatje University Kimberley, 8301, South Africa
| | - Pramod Kumar Singh
- Department of Computer Science and Engineering, ABV-Indian Institute of Information Technology and Management Gwalior, Gwalior, 474015, MP, India
| |
Collapse
|
9
|
Lu Y, Lu X, Zheng L, Sun M, Chen S, Chen B, Wang T, Yang J, Lv C. Application of Multimodal Transformer Model in Intelligent Agricultural Disease Detection and Question-Answering Systems. PLANTS (BASEL, SWITZERLAND) 2024; 13:972. [PMID: 38611501 PMCID: PMC11013167 DOI: 10.3390/plants13070972] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/30/2024] [Revised: 03/22/2024] [Accepted: 03/24/2024] [Indexed: 04/14/2024]
Abstract
In this study, an innovative approach based on multimodal data and the transformer model was proposed to address challenges in agricultural disease detection and question-answering systems. This method effectively integrates image, text, and sensor data, utilizing deep learning technologies to profoundly analyze and process complex agriculture-related issues. The study achieved technical breakthroughs and provides new perspectives and tools for the development of intelligent agriculture. In the task of agricultural disease detection, the proposed method demonstrated outstanding performance, achieving a precision, recall, and accuracy of 0.95, 0.92, and 0.94, respectively, significantly outperforming the other conventional deep learning models. These results indicate the method's effectiveness in identifying and accurately classifying various agricultural diseases, particularly excelling in handling subtle features and complex data. In the task of generating descriptive text from agricultural images, the method also exhibited impressive performance, with a precision, recall, and accuracy of 0.92, 0.88, and 0.91, respectively. This demonstrates that the method can not only deeply understand the content of agricultural images but also generate accurate and rich descriptive texts. The object detection experiment further validated the effectiveness of our approach, where the method achieved a precision, recall, and accuracy of 0.96, 0.91, and 0.94. This achievement highlights the method's capability for accurately locating and identifying agricultural targets, especially in complex environments. Overall, the approach in this study not only demonstrated exceptional performance in multiple tasks such as agricultural disease detection, image captioning, and object detection but also showcased the immense potential of multimodal data and deep learning technologies in the application of intelligent agriculture.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | | | - Chunli Lv
- China Agricultural University, Beijing 100083, China
| |
Collapse
|
10
|
Debnath A, Hasan MM, Raihan M, Samrat N, Alsulami MM, Masud M, Bairagi AK. A Smartphone-Based Detection System for Tomato Leaf Disease Using EfficientNetV2B2 and Its Explainability with Artificial Intelligence (AI). SENSORS (BASEL, SWITZERLAND) 2023; 23:8685. [PMID: 37960385 PMCID: PMC10648786 DOI: 10.3390/s23218685] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 10/19/2023] [Accepted: 10/20/2023] [Indexed: 11/15/2023]
Abstract
The occurrence of tomato diseases has substantially reduced agricultural output and financial losses. The timely detection of diseases is crucial to effectively manage and mitigate the impact of episodes. Early illness detection can improve output, reduce chemical use, and boost a nation's economy. A complete system for plant disease detection using EfficientNetV2B2 and deep learning (DL) is presented in this paper. This research aims to develop a precise and effective automated system for identifying several illnesses that impact tomato plants. This will be achieved by analyzing tomato leaf photos. A dataset of high-resolution photographs of healthy and diseased tomato leaves was created to achieve this goal. The EfficientNetV2B2 model is the foundation of the deep learning system and excels at picture categorization. Transfer learning (TF) trains the model on a tomato leaf disease dataset using EfficientNetV2B2's pre-existing weights and a 256-layer dense layer. Tomato leaf diseases can be identified using the EfficientNetV2B2 model and a dense layer of 256 nodes. An ideal loss function and algorithm train and tune the model. Next, the concept is deployed in smartphones and online apps. The user can accurately diagnose tomato leaf diseases with this application. Utilizing an automated system facilitates the rapid identification of diseases, assisting in making informed decisions on disease management and promoting sustainable tomato cultivation practices. The 5-fold cross-validation method achieved 99.02% average weighted training accuracy, 99.22% average weighted validation accuracy, and 98.96% average weighted test accuracy. The split method achieved 99.93% training accuracy and 100% validation accuracy. Using the DL approach, tomato leaf disease identification achieves nearly 100% accuracy on a test dataset.
Collapse
Affiliation(s)
- Anjan Debnath
- Department of Computer Science and Engineering, North Western University, Khulna 9100, Bangladesh; (A.D.); (M.M.H.); (N.S.)
| | - Md. Mahedi Hasan
- Department of Computer Science and Engineering, North Western University, Khulna 9100, Bangladesh; (A.D.); (M.M.H.); (N.S.)
| | - M. Raihan
- Department of Computer Science and Engineering, North Western University, Khulna 9100, Bangladesh; (A.D.); (M.M.H.); (N.S.)
| | - Nadim Samrat
- Department of Computer Science and Engineering, North Western University, Khulna 9100, Bangladesh; (A.D.); (M.M.H.); (N.S.)
| | - Mashael M. Alsulami
- Department of Information Technology, College of Computers and Information Technology, Taif University, Taif 21944, Saudi Arabia;
| | - Mehedi Masud
- Department of Computer Science, College of Computers and Information Technology, Taif University, Taif 21944, Saudi Arabia;
| | - Anupam Kumar Bairagi
- Computer Science and Engineering Discipline, Khulna University, Khulna 9208, Bangladesh
| |
Collapse
|
11
|
Apacionado BV, Ahamed T. Sooty Mold Detection on Citrus Tree Canopy Using Deep Learning Algorithms. SENSORS (BASEL, SWITZERLAND) 2023; 23:8519. [PMID: 37896610 PMCID: PMC10610784 DOI: 10.3390/s23208519] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Revised: 10/05/2023] [Accepted: 10/13/2023] [Indexed: 10/29/2023]
Abstract
Sooty mold is a common disease found in citrus plants and is characterized by black fungi growth on fruits, leaves, and branches. This mold reduces the plant's ability to carry out photosynthesis. In small leaves, it is very difficult to detect sooty mold at the early stages. Deep learning-based image recognition techniques have the potential to identify and diagnose pest damage and diseases such as sooty mold. Recent studies used advanced and expensive hyperspectral or multispectral cameras attached to UAVs to examine the canopy of the plants and mid-range cameras to capture close-up infected leaf images. To bridge the gap on capturing canopy level images using affordable camera sensors, this study used a low-cost home surveillance camera to monitor and detect sooty mold infection on citrus canopy combined with deep learning algorithms. To overcome the challenges posed by varying light conditions, the main reason for using specialized cameras, images were collected at night, utilizing the camera's built-in night vision feature. A total of 4200 sliced night-captured images were used for training, 200 for validation, and 100 for testing, employed on the YOLOv5m, YOLOv7, and CenterNet models for comparison. The results showed that YOLOv7 was the most accurate in detecting sooty molds at night, with 74.4% mAP compared to YOLOv5m (72%) and CenterNet (70.3%). The models were also tested using preprocessed (unsliced) night images and day-captured sliced and unsliced images. The testing on preprocessed (unsliced) night images demonstrated the same trend as the training results, with YOLOv7 performing best compared to YOLOv5m and CenterNet. In contrast, testing on the day-captured images had underwhelming outcomes for both sliced and unsliced images. In general, YOLOv7 performed best in detecting sooty mold infections at night on citrus canopy and showed promising potential in real-time orchard disease monitoring and detection. Moreover, this study demonstrated that utilizing a cost-effective surveillance camera and deep learning algorithms can accurately detect sooty molds at night, enabling growers to effectively monitor and identify occurrences of the disease at the canopy level.
Collapse
Affiliation(s)
- Bryan Vivas Apacionado
- Graduate School of Science and Technology, University of Tsukuba, 1-1-1 Tennodai, Tsukuba 305-8577, Japan;
| | - Tofael Ahamed
- Institute of Life and Environmental Sciences, University of Tsukuba, 1-1-1 Tennodai, Tsukuba 305-8577, Japan
| |
Collapse
|