1
|
Feng J, Blair SW, Ayanlade TT, Balu A, Ganapathysubramanian B, Singh A, Sarkar S, Singh AK. Robust soybean seed yield estimation using high-throughput ground robot videos. FRONTIERS IN PLANT SCIENCE 2025; 16:1554193. [PMID: 40230608 PMCID: PMC11994694 DOI: 10.3389/fpls.2025.1554193] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/01/2025] [Accepted: 03/03/2025] [Indexed: 04/16/2025]
Abstract
We present a novel method for soybean [Glycine max (L.) Merr.] yield estimation leveraging high-throughput seed counting via computer vision and deep learning techniques. Traditional methods for collecting yield data are labor-intensive, costly, and prone to equipment failures at critical data collection times and require transportation of equipment across field sites. Computer vision, the field of teaching computers to interpret visual data, allows us to extract detailed yield information directly from images. By treating it as a computer vision task, we report a more efficient alternative, employing a ground robot equipped with fisheye cameras to capture comprehensive videos of soybean plots from which images are extracted in a variety of development programs. These images are processed through the P2PNet-Yield model, a deep learning framework, where we combined a feature extraction module (the backbone of the P2PNet-Soy) and a yield regression module to estimate seed yields of soybean plots. Our results are built on 2 years of yield testing plot data-8,500 plots in 2021 and 650 plots in 2023. With these datasets, our approach incorporates several innovations to further improve the accuracy and generalizability of the seed counting and yield estimation architecture, such as the fisheye image correction and data augmentation with random sensor effects. The P2PNet-Yield model achieved a genotype ranking accuracy score of up to 83%. It demonstrates up to a 32% reduction in time to collect yield data as well as costs associated with traditional yield estimation, offering a scalable solution for breeding programs and agricultural productivity enhancement.
Collapse
Affiliation(s)
- Jiale Feng
- Department of Computer Science, Iowa State University, Ames, IA, United States
| | - Samuel W. Blair
- Department of Agronomy, Iowa State University, Ames, IA, United States
| | - Timilehin T. Ayanlade
- Department of Mechanical Engineering, Iowa State University, Ames, IA, United States
| | - Aditya Balu
- Department of Mechanical Engineering, Iowa State University, Ames, IA, United States
| | | | - Arti Singh
- Department of Agronomy, Iowa State University, Ames, IA, United States
| | - Soumik Sarkar
- Department of Computer Science, Iowa State University, Ames, IA, United States
- Department of Mechanical Engineering, Iowa State University, Ames, IA, United States
| | - Asheesh K. Singh
- Department of Agronomy, Iowa State University, Ames, IA, United States
| |
Collapse
|
2
|
Sun Z, Yang Z, Ding Y, Sun B, Li S, Guo Z, Zhu L. Adaptive spatial-channel feature fusion and self-calibrated convolution for early maize seedlings counting in UAV images. FRONTIERS IN PLANT SCIENCE 2025; 15:1496801. [PMID: 39980762 PMCID: PMC11841422 DOI: 10.3389/fpls.2024.1496801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/15/2024] [Accepted: 12/13/2024] [Indexed: 02/22/2025]
Abstract
Accurate counting of crop plants is essential for agricultural science, particularly for yield forecasting, field management, and experimental studies. Traditional methods are labor-intensive and prone to errors. Unmanned Aerial Vehicle (UAV) technology offers a promising alternative; however, varying UAV altitudes can impact image quality, leading to blurred features and reduced accuracy in early maize seedling counts. To address these challenges, we developed RC-Dino, a deep learning methodology based on DINO, specifically designed to enhance the precision of seedling counts from UAV-acquired images. RC-Dino introduces two innovative components: a novel self-calibrating convolutional layer named RSCconv and an adaptive spatial feature fusion module called ASCFF. The RSCconv layer improves the representation of early maize seedlings compared to non-seedling elements within feature maps by calibrating spatial domain features. The ASCFF module enhances the discriminability of early maize seedlings by adaptively fusing feature maps extracted from different layers of the backbone network. Additionally, transfer learning was employed to integrate pre-trained weights with RSCconv, facilitating faster convergence and improved accuracy. The efficacy of our approach was validated using the Early Maize Seedlings Dataset (EMSD), comprising 1,233 annotated images of early maize seedlings, totaling 83,404 individual annotations. Testing on this dataset demonstrated that RC-Dino outperformed existing models, including DINO, Faster R-CNN, RetinaNet, YOLOX, and Deformable DETR. Specifically, RC-Dino achieved improvements of 16.29% in Average Precision (AP) and 8.19% in Recall compared to the DINO model. Our method also exhibited superior coefficient of determination (R²) values across different datasets for seedling counting. By integrating RSCconv and ASCFF into other detection frameworks such as Faster R-CNN, RetinaNet, and Deformable DETR, we observed enhanced detection and counting accuracy, further validating the effectiveness of our proposed method. These advancements make RC-Dino particularly suitable for accurate early maize seedling counting in the field. The source code for RSCconv and ASCFF is publicly available at https://github.com/collapser-AI/RC-Dino, promoting further research and practical applications.
Collapse
Affiliation(s)
- Zhenyuan Sun
- School of Civil and Hydraulic Engineering, Ningxia University, Yinchuan, China
- Key Laboratory of Digital Water Governance for Yellow River Water Network, Yinchuan, China
- Field Scientific Observation and Research Station of Agricultural Irrigation in Ningxia Diversion Yellow Irrigation District, Ministry of Water Resources, Yinchuan, China
| | - Zhi Yang
- Field Scientific Observation and Research Station of Agricultural Irrigation in Ningxia Diversion Yellow Irrigation District, Ministry of Water Resources, Yinchuan, China
- The Scientific Research Institute of the Water Conservancy of Ningxia, Yinchuan, China
| | - Yimin Ding
- School of Civil and Hydraulic Engineering, Ningxia University, Yinchuan, China
- Key Laboratory of Digital Water Governance for Yellow River Water Network, Yinchuan, China
| | - Boyan Sun
- School of Civil and Hydraulic Engineering, Ningxia University, Yinchuan, China
- Key Laboratory of Digital Water Governance for Yellow River Water Network, Yinchuan, China
| | - Saiju Li
- School of Civil and Hydraulic Engineering, Ningxia University, Yinchuan, China
- Key Laboratory of Digital Water Governance for Yellow River Water Network, Yinchuan, China
| | - Zhen Guo
- School of Civil and Hydraulic Engineering, Ningxia University, Yinchuan, China
- Key Laboratory of Digital Water Governance for Yellow River Water Network, Yinchuan, China
| | - Lei Zhu
- School of Civil and Hydraulic Engineering, Ningxia University, Yinchuan, China
- Key Laboratory of Digital Water Governance for Yellow River Water Network, Yinchuan, China
| |
Collapse
|
3
|
Li T, Blok PM, Burridge J, Kaga A, Guo W. Multi-Scale Attention Network for Vertical Seed Distribution in Soybean Breeding Fields. PLANT PHENOMICS (WASHINGTON, D.C.) 2024; 6:0260. [PMID: 39525982 PMCID: PMC11550408 DOI: 10.34133/plantphenomics.0260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/06/2024] [Revised: 09/12/2024] [Accepted: 09/14/2024] [Indexed: 11/16/2024]
Abstract
The increase in the global population is leading to a doubling of the demand for protein. Soybean (Glycine max), a key contributor to global plant-based protein supplies, requires ongoing yield enhancements to keep pace with increasing demand. Precise, on-plant seed counting and localization may catalyze breeding selection of shoot architectures and seed localization patterns related to superior performance in high planting density and contribute to increased yield. Traditional manual counting and localization methods are labor-intensive and prone to error, necessitating more efficient approaches for yield prediction and seed distribution analysis. To solve this, we propose MSANet: a novel deep learning framework tailored for counting and localization of soybean seeds on mature field-grown soy plants. A multi-scale attention map mechanism was applied to maximize model performance in seed counting and localization in soybean breeding fields. We compared our model with a previous state-of-the-art model using the benchmark dataset and an enlarged dataset, including various soybean genotypes. Our model outperforms previous state-of-the-art methods on all datasets across various soybean genotypes on both counting and localization tasks. Furthermore, our model also performed well on in-canopy 360° video, dramatically increasing data collection efficiency. We also propose a technique that enables previously inaccessible insights into the phenotypic and genetic diversity of single plant vertical seed distribution, which may accelerate the breeding process. To accelerate further research in this domain, we have made our dataset and software publicly available: https://github.com/UTokyo-FieldPhenomics-Lab/MSANet.
Collapse
Affiliation(s)
- Tang Li
- Graduate School of Agricultural and Life Sciences,
The University of Tokyo, Tokyo, Japan
| | - Pieter M. Blok
- Graduate School of Agricultural and Life Sciences,
The University of Tokyo, Tokyo, Japan
| | - James Burridge
- Graduate School of Agricultural and Life Sciences,
The University of Tokyo, Tokyo, Japan
| | - Akito Kaga
- Institute of Crop Sciences, National Agriculture and Food Research Organization, Tsukuba, Ibaraki, Japan
| | - Wei Guo
- Graduate School of Agricultural and Life Sciences,
The University of Tokyo, Tokyo, Japan
| |
Collapse
|
4
|
He J, Weng L, Xu X, Chen R, Peng B, Li N, Xie Z, Sun L, Han Q, He P, Wang F, Yu H, Bhat JA, Feng X. DEKR-SPrior: An Efficient Bottom-Up Keypoint Detection Model for Accurate Pod Phenotyping in Soybean. PLANT PHENOMICS (WASHINGTON, D.C.) 2024; 6:0198. [PMID: 38939747 PMCID: PMC11209727 DOI: 10.34133/plantphenomics.0198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/08/2024] [Accepted: 05/16/2024] [Indexed: 06/29/2024]
Abstract
The pod and seed counts are important yield-related traits in soybean. High-precision soybean breeders face the major challenge of accurately phenotyping the number of pods and seeds in a high-throughput manner. Recent advances in artificial intelligence, especially deep learning (DL) models, have provided new avenues for high-throughput phenotyping of crop traits with increased precision. However, the available DL models are less effective for phenotyping pods that are densely packed and overlap in in situ soybean plants; thus, accurate phenotyping of the number of pods and seeds in soybean plant is an important challenge. To address this challenge, the present study proposed a bottom-up model, DEKR-SPrior (disentangled keypoint regression with structural prior), for in situ soybean pod phenotyping, which considers soybean pods and seeds analogous to human people and joints, respectively. In particular, we designed a novel structural prior (SPrior) module that utilizes cosine similarity to improve feature discrimination, which is important for differentiating closely located seeds from highly similar seeds. To further enhance the accuracy of pod location, we cropped full-sized images into smaller and high-resolution subimages for analysis. The results on our image datasets revealed that DEKR-SPrior outperformed multiple bottom-up models, viz., Lightweight-OpenPose, OpenPose, HigherHRNet, and DEKR, reducing the mean absolute error from 25.81 (in the original DEKR) to 21.11 (in the DEKR-SPrior) in pod phenotyping. This paper demonstrated the great potential of DEKR-SPrior for plant phenotyping, and we hope that DEKR-SPrior will help future plant phenotyping.
Collapse
Affiliation(s)
- Jingjing He
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Lin Weng
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Xiaogang Xu
- School of Computer Science and Technology,
Zhejiang Gongshang University, Hangzhou 310012, Zhejiang, China
| | - Ruochen Chen
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Bo Peng
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Nannan Li
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Zhengchao Xie
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Lijian Sun
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Qiang Han
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Pengfei He
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Fangfang Wang
- Zhejiang Laboratory, Hangzhou 311100, Zhejiang, China
| | - Hui Yu
- Key Laboratory of Soybean Molecular Design Breeding, Northeast Institute of Geography and Agroecology,
Chinese Academy of Sciences, Changchun 130102, Jilin, China
| | | | - Xianzhong Feng
- Key Laboratory of Soybean Molecular Design Breeding, Northeast Institute of Geography and Agroecology,
Chinese Academy of Sciences, Changchun 130102, Jilin, China
| |
Collapse
|
5
|
Jiang T, Yu Q, Zhong Y, Shao M. PlantSR: Super-Resolution Improves Object Detection in Plant Images. J Imaging 2024; 10:137. [PMID: 38921614 PMCID: PMC11204869 DOI: 10.3390/jimaging10060137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2024] [Revised: 06/01/2024] [Accepted: 06/04/2024] [Indexed: 06/27/2024] Open
Abstract
Recent advancements in computer vision, especially deep learning models, have shown considerable promise in tasks related to plant image object detection. However, the efficiency of these deep learning models heavily relies on input image quality, with low-resolution images significantly hindering model performance. Therefore, reconstructing high-quality images through specific techniques will help extract features from plant images, thus improving model performance. In this study, we explored the value of super-resolution technology for improving object detection model performance on plant images. Firstly, we built a comprehensive dataset comprising 1030 high-resolution plant images, named the PlantSR dataset. Subsequently, we developed a super-resolution model using the PlantSR dataset and benchmarked it against several state-of-the-art models designed for general image super-resolution tasks. Our proposed model demonstrated superior performance on the PlantSR dataset, indicating its efficacy in enhancing the super-resolution of plant images. Furthermore, we explored the effect of super-resolution on two specific object detection tasks: apple counting and soybean seed counting. By incorporating super-resolution as a pre-processing step, we observed a significant reduction in mean absolute error. Specifically, with the YOLOv7 model employed for apple counting, the mean absolute error decreased from 13.085 to 5.71. Similarly, with the P2PNet-Soy model utilized for soybean seed counting, the mean absolute error decreased from 19.159 to 15.085. These findings underscore the substantial potential of super-resolution technology in improving the performance of object detection models for accurately detecting and counting specific plants from images. The source codes and associated datasets related to this study are available at Github.
Collapse
Affiliation(s)
- Tianyou Jiang
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (T.J.); (Y.Z.); (M.S.)
| | - Qun Yu
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (T.J.); (Y.Z.); (M.S.)
- Huanghuaihai Key Laboratory of Smart Agricultural Technology, Ministry of Agriculture and Rural Affairs, Tai’an 271018, China
| | - Yang Zhong
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (T.J.); (Y.Z.); (M.S.)
| | - Mingshun Shao
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (T.J.); (Y.Z.); (M.S.)
| |
Collapse
|
6
|
Yu Z, Wang Y, Ye J, Liufu S, Lu D, Zhu X, Yang Z, Tan Q. Accurate and fast implementation of soybean pod counting and localization from high-resolution image. FRONTIERS IN PLANT SCIENCE 2024; 15:1320109. [PMID: 38444529 PMCID: PMC10913015 DOI: 10.3389/fpls.2024.1320109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Accepted: 01/30/2024] [Indexed: 03/07/2024]
Abstract
Introduction Soybean pod count is one of the crucial indicators of soybean yield. Nevertheless, due to the challenges associated with counting pods, such as crowded and uneven pod distribution, existing pod counting models prioritize accuracy over efficiency, which does not meet the requirements for lightweight and real-time tasks. Methods To address this goal, we have designed a deep convolutional network called PodNet. It employs a lightweight encoder and an efficient decoder that effectively decodes both shallow and deep information, alleviating the indirect interactions caused by information loss and degradation between non-adjacent levels. Results We utilized a high-resolution dataset of soybean pods from field harvesting to evaluate the model's generalization ability. Through experimental comparisons between manual counting and model yield estimation, we confirmed the effectiveness of the PodNet model. The experimental results indicate that PodNet achieves an R2 of 0.95 for the prediction of soybean pod quantities compared to ground truth, with only 2.48M parameters, which is an order of magnitude lower than the current SOTA model YOLO POD, and the FPS is much higher than YOLO POD. Discussion Compared to advanced computer vision methods, PodNet significantly enhances efficiency with almost no sacrifice in accuracy. Its lightweight architecture and high FPS make it suitable for real-time applications, providing a new solution for counting and locating dense objects.
Collapse
Affiliation(s)
- Zhenghong Yu
- College of Robotics, Guangdong Polytechnic of Science and Technology, Zhuhai, China
| | - Yangxu Wang
- College of Robotics, Guangdong Polytechnic of Science and Technology, Zhuhai, China
- Department of Network Technology, Guangzhou Institute of Software Engineering, Conghua, China
| | - Jianxiong Ye
- College of Robotics, Guangdong Polytechnic of Science and Technology, Zhuhai, China
- School of Electronics and Information Engineering, Wuyi University, Jiangmen, China
| | - Shengjie Liufu
- College of Business, Guangzhou College of Technology and Business, Foshan, China
| | - Dunlu Lu
- College of Robotics, Guangdong Polytechnic of Science and Technology, Zhuhai, China
| | - Xiuli Zhu
- College of Robotics, Guangdong Polytechnic of Science and Technology, Zhuhai, China
| | - Zhongming Yang
- College of Robotics, Guangdong Polytechnic of Science and Technology, Zhuhai, China
| | - Qingji Tan
- School of Mechanical Engineering, Guangxi University, Nanning, China
| |
Collapse
|
7
|
Qiao Y, Liao Q, Zhang M, Han B, Peng C, Huang Z, Wang S, Zhou G, Xu S. Point clouds segmentation of rapeseed siliques based on sparse-dense point clouds mapping. FRONTIERS IN PLANT SCIENCE 2023; 14:1188286. [PMID: 37521934 PMCID: PMC10375295 DOI: 10.3389/fpls.2023.1188286] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Accepted: 06/30/2023] [Indexed: 08/01/2023]
Abstract
In this study, we propose a high-throughput and low-cost automatic detection method based on deep learning to replace the inefficient manual counting of rapeseed siliques. First, a video is captured with a smartphone around the rapeseed plants in the silique stage. Feature point detection and matching based on SIFT operators are applied to the extracted video frames, and sparse point clouds are recovered using epipolar geometry and triangulation principles. The depth map is obtained by calculating the disparity of the matched images, and the dense point cloud is fused. The plant model of the whole rapeseed plant in the silique stage is reconstructed based on the structure-from-motion (SfM) algorithm, and the background is removed by using the passthrough filter. The downsampled 3D point cloud data is processed by the DGCNN network, and the point cloud is divided into two categories: sparse rapeseed canopy siliques and rapeseed stems. The sparse canopy siliques are then segmented from the original whole rapeseed siliques point cloud using the sparse-dense point cloud mapping method, which can effectively save running time and improve efficiency. Finally, Euclidean clustering segmentation is performed on the rapeseed canopy siliques, and the RANSAC algorithm is used to perform line segmentation on the connected siliques after clustering, obtaining the three-dimensional spatial position of each silique and counting the number of siliques. The proposed method was applied to identify 1457 siliques from 12 rapeseed plants, and the experimental results showed a recognition accuracy greater than 97.80%. The proposed method achieved good results in rapeseed silique recognition and provided a useful example for the application of deep learning networks in dense 3D point cloud segmentation.
Collapse
Affiliation(s)
- Yuhui Qiao
- College of Engineering, Huazhong Agricultural University, Wuhan, China
- Key Laboratory of Agricultural Equipment for the Middle and Lower Reaches of the Yangtze River, Ministry of Agriculture, Huazhong Agricultural University, Wuhan, China
| | - Qingxi Liao
- College of Engineering, Huazhong Agricultural University, Wuhan, China
- Key Laboratory of Agricultural Equipment for the Middle and Lower Reaches of the Yangtze River, Ministry of Agriculture, Huazhong Agricultural University, Wuhan, China
| | - Moran Zhang
- College of Engineering, Huazhong Agricultural University, Wuhan, China
- Key Laboratory of Agricultural Equipment for the Middle and Lower Reaches of the Yangtze River, Ministry of Agriculture, Huazhong Agricultural University, Wuhan, China
| | - Binbin Han
- School of Mathematics and Computer Science, Wuhan Polytechnic University, Wuhan, China
| | - Chengli Peng
- School of Geosciences and Info-Physics, Central South University, Changsha, China
| | - Zhenhao Huang
- College of Engineering, Huazhong Agricultural University, Wuhan, China
- Key Laboratory of Agricultural Equipment for the Middle and Lower Reaches of the Yangtze River, Ministry of Agriculture, Huazhong Agricultural University, Wuhan, China
| | - Shaodong Wang
- College of Engineering, Huazhong Agricultural University, Wuhan, China
- Key Laboratory of Agricultural Equipment for the Middle and Lower Reaches of the Yangtze River, Ministry of Agriculture, Huazhong Agricultural University, Wuhan, China
| | - Guangsheng Zhou
- College of Plant Science & Technology, Huazhong Agricultural University, Wuhan, China
| | - Shengyong Xu
- College of Engineering, Huazhong Agricultural University, Wuhan, China
- Key Laboratory of Agricultural Equipment for the Middle and Lower Reaches of the Yangtze River, Ministry of Agriculture, Huazhong Agricultural University, Wuhan, China
| |
Collapse
|