1
|
Ang G, Zhiwei T, Wei M, Yuepeng S, Longlong R, Yuliang F, Jianping Q, Lijia X. Fruits hidden by green: an improved YOLOV8n for detection of young citrus in lush citrus trees. Front Plant Sci 2024; 15:1375118. [PMID: 38660450 PMCID: PMC11039839 DOI: 10.3389/fpls.2024.1375118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Accepted: 03/19/2024] [Indexed: 04/26/2024]
Abstract
In order to address the challenges of inefficiency and insufficient accuracy in the manual identification of young citrus fruits during thinning processes, this study proposes a detection methodology using the you only look once for complex backgrounds of young citrus fruits (YCCB-YOLO) approach. The method first constructs a dataset containing images of young citrus fruits in a real orchard environment. To improve the detection accuracy while maintaining the computational efficiency, the study reconstructs the detection head and backbone network using pointwise convolution (PWonv) lightweight network, which reduces the complexity of the model without affecting the performance. In addition, the ability of the model to accurately detect young citrus fruits in complex backgrounds is enhanced by integrating the fusion attention mechanism. Meanwhile, the simplified spatial pyramid pooling fast-large kernel separated attention (SimSPPF-LSKA) feature pyramid was introduced to further enhance the multi-feature extraction capability of the model. Finally, the Adam optimization function was used to strengthen the nonlinear representation and feature extraction ability of the model. The experimental results show that the model achieves 91.79% precision (P), 92.75% recall (R), and 97.32% mean average precision (mAP)on the test set, which were improved by 1.33%, 2.24%, and 1.73%, respectively, compared with the original model, and the size of the model is only 5.4 MB. This study could meet the performance requirements for citrus fruit identification, which provides technical support for fruit thinning.
Collapse
Affiliation(s)
- Gao Ang
- College of Mechanical and Electronic Engineering, Shandong Agricultural University, Tai’an, Shandong, China
- Institute of Urban Agriculture, Chinese Academy of Agricultural Sciences, Chengdu, China
| | - Tian Zhiwei
- Institute of Urban Agriculture, Chinese Academy of Agricultural Sciences, Chengdu, China
| | - Ma Wei
- Institute of Urban Agriculture, Chinese Academy of Agricultural Sciences, Chengdu, China
| | - Song Yuepeng
- College of Mechanical and Electronic Engineering, Shandong Agricultural University, Tai’an, Shandong, China
| | - Ren Longlong
- College of Mechanical and Electronic Engineering, Shandong Agricultural University, Tai’an, Shandong, China
| | - Feng Yuliang
- College of Engineering, China Agricultural University, Beijing, China
| | - Qian Jianping
- Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing, China
| | - Xu Lijia
- College of Mechanical and Electrical Engineering, Sichuan Agriculture University, Ya’an, China
| |
Collapse
|
2
|
Li J, Han R, Li F, Dong G, Ma Y, Yang W, Qi G, Zhang L. Apple Fruit Edge Detection Model Using a Rough Set and Convolutional Neural Network. Sensors (Basel) 2024; 24:2283. [PMID: 38610494 PMCID: PMC11014221 DOI: 10.3390/s24072283] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Revised: 03/31/2024] [Accepted: 04/01/2024] [Indexed: 04/14/2024]
Abstract
Accurately and effectively detecting the growth position and contour size of apple fruits is crucial for achieving intelligent picking and yield predictions. Thus, an effective fruit edge detection algorithm is necessary. In this study, a fusion edge detection model (RED) based on a convolutional neural network and rough sets was proposed. The Faster-RCNN was used to segment multiple apple images into a single apple image for edge detection, greatly reducing the surrounding noise of the target. Moreover, the K-means clustering algorithm was used to segment the target of a single apple image for further noise reduction. Considering the influence of illumination, complex backgrounds and dense occlusions, rough set was applied to obtain the edge image of the target for the upper and lower approximation images, and the results were compared with those of relevant algorithms in this field. The experimental results showed that the RED model in this paper had high accuracy and robustness, and its detection accuracy and stability were significantly improved compared to those of traditional operators, especially under the influence of illumination and complex backgrounds. The RED model is expected to provide a promising basis for intelligent fruit picking and yield prediction.
Collapse
Affiliation(s)
- Junqing Li
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (J.L.)
| | - Ruiyi Han
- College of Computer Science and Technology, China University of Petroleum (East China), Changjiang Road No.66, Qingdao 266580, China
| | - Fangyi Li
- College of Information and Electrical Engineering, China Agricultural University, Beijing 100083, China
| | - Guoao Dong
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (J.L.)
| | - Yu Ma
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (J.L.)
| | - Wei Yang
- National Key Laboratory of Wheat Improvement, College of Life Science, Shandong Agricultural University, Tai’an 271018, China
| | - Guanghui Qi
- College of Information Science and Engineering, Shandong Agricultural University, Tai’an 271018, China; (J.L.)
| | - Liang Zhang
- College of Computer Science and Technology, China University of Petroleum (East China), Changjiang Road No.66, Qingdao 266580, China
| |
Collapse
|
3
|
Lai Q, Wang Y, Tan Y, Sun W. Design and experiment of Panax notoginseng root orientation transplanting device based on YOLOv5s. Front Plant Sci 2024; 15:1325420. [PMID: 38525144 PMCID: PMC10957537 DOI: 10.3389/fpls.2024.1325420] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 02/21/2024] [Indexed: 03/26/2024]
Abstract
Consistent root orientation is one of the important requirements of Panax notoginseng transplanting agronomy. In this paper, a Panax notoginseng orientation transplanting method based on machine vision technology and negative pressure adsorption principle was proposed. With the cut-main root of Panax notoginseng roots as the detection object, the YOLOv5s was used to establish a root feature detection model. A Panax notoginseng root orientation transplanting device was designed. The orientation control system identifies the root posture according to the detection results and controls the orientation actuator to adjust the root posture. The detection results show that the precision rate of the model was 94.2%, the recall rate was 92.0%, and the average detection precision was 94.9%. The Box-Behnken experiments were performed to investigate the effects of suction plate rotation speed, servo rotation speed and the angle between the camera and the orientation actuator(ACOA) on the orientation qualification rate and root drop rate. Response surface method and objective optimisation algorithm were used to analyse the experimental results. The optimal working parameters were suction plate rotation speed of 5.73 r/min, servo rotation speed of 0.86 r/s and ACOA of 35°. Under this condition, the orientation qualification rate and root drop rate of the actual experiment were 89.87% and 6.57%, respectively, which met the requirements of orientation transplanting for Panax notoginseng roots. The research method of this paper is helpful to solve the problem of orientation transplanting of other root crops.
Collapse
Affiliation(s)
- Qinghui Lai
- School of Energy and Environment Science, Yunnan Provincial Rural Energy Engineering Key Laboratory, Yunnan Normal University, Kunming, China
| | - Yongjie Wang
- Faculty of Modern Agriculture Engineering, Kunming University of Science and Technology, Kunming, China
| | - Yu Tan
- Faculty of Modern Agriculture Engineering, Kunming University of Science and Technology, Kunming, China
| | - Wenqiang Sun
- Faculty of Modern Agriculture Engineering, Kunming University of Science and Technology, Kunming, China
| |
Collapse
|
4
|
Li G, Shi G, Zhu C. Dynamic Serpentine Convolution with Attention Mechanism Enhancement for Beef Cattle Behavior Recognition. Animals (Basel) 2024; 14:466. [PMID: 38338110 PMCID: PMC10854982 DOI: 10.3390/ani14030466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Revised: 01/25/2024] [Accepted: 01/26/2024] [Indexed: 02/12/2024] Open
Abstract
Behavior recognition in beef cattle is a crucial component of beef cattle behavior warning and intelligent farming. Traditional beef cattle behavior recognition faces challenges in both difficulty in identification and low accuracy. In this study, the YOLOv8n_BiF_DSC (Fusion of Dynamic Snake Convolution and BiFormer Attention) algorithm was employed for the non-intrusive recognition of beef cattle behavior. The specific steps are as follows: 45 beef cattle were observed using a fixed camera (A LINE OF DEFENSE) and a mobile phone (Huawei Mate20Pro) to collect and filter posture data, yielding usable videos ranging from 1 to 30 min in length. These videos cover nine different behaviors in various scenarios, including standing, lying, mounting, fighting, licking, eating, drinking, walking, and searching. After data augmentation, the dataset comprised 34,560 samples. The convolutional layer (CONV) was improved by introducing variable convolution and dynamic snake-like convolution modules. The dynamic snake-like convolution, which yielded the best results, expanded the model's receptive field, dynamically perceived key features of beef cattle behavior, and enhanced the algorithm's feature extraction capability. Attention mechanism modules, including SE (Squeeze-and-Excitation Networks), CBAM (Convolutional Block Attention Module), CA (Coordinate Attention), and BiFormer (Vision Transformer with Bi-Level Routing Attention), were introduced. The BiFormer attention mechanism, selected for its optimal performance, improved the algorithm's ability to capture long-distance context dependencies. The model's computational efficiency was enhanced through dynamic and query-aware perception. Experimental results indicated that YOLOv8n_BiF_DSC achieved the best results among all improved algorithms in terms of accuracy, average precision at IoU 50, and average precision at IoU 50:95. The accuracy of beef cattle behavior recognition reached 93.6%, with the average precision at IoU 50 and IoU 50:95 being 96.5% and 71.5%, respectively. This represents a 5.3%, 5.2%, and 7.1% improvement over the original YOLOv8n. Notably, the average accuracy of recognizing the lying posture of beef cattle reached 98.9%. In conclusion, the YOLOv8n_BiF_DSC algorithm demonstrates excellent performance in feature extraction and high-level data fusion, displaying high robustness and adaptability. It provides theoretical and practical support for the intelligent recognition and management of beef cattle.
Collapse
Affiliation(s)
- Guangbo Li
- College of Electronic and Information Engineering, Huaibei Institute of Technology, Huaibei 235000, China
| | - Guolong Shi
- School of Information and Artificial Intelligence, Anhui Agricultural University, Hefei 230036, China
| | - Changjie Zhu
- College of Electronic and Information Engineering, Huaibei Institute of Technology, Huaibei 235000, China
| |
Collapse
|
5
|
Chen Y, Deng C, Sun Q, Wu Z, Zou L, Zhang G, Li W. Lightweight Detection Methods for Insulator Self-Explosion Defects. Sensors (Basel) 2024; 24:290. [PMID: 38203151 PMCID: PMC10781199 DOI: 10.3390/s24010290] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Revised: 11/07/2023] [Accepted: 12/14/2023] [Indexed: 01/12/2024]
Abstract
The accurate and efficient detection of defective insulators is an essential prerequisite for ensuring the safety of the power grid in the new generation of intelligent electrical system inspections. Currently, traditional object detection algorithms for detecting defective insulators in images face issues such as excessive parameter size, low accuracy, and slow detection speed. To address the aforementioned issues, this article proposes an insulator defect detection model based on the lightweight Faster R-CNN (Faster Region-based Convolutional Network) model (Faster R-CNN-tiny). First, the Faster R-CNN model's backbone network is turned into a lightweight version of it by substituting EfficientNet for ResNet (Residual Network), greatly decreasing the model parameters while increasing its detection accuracy. The second step is to employ a feature pyramid to build feature maps with various resolutions for feature fusion, which enables the detection of objects at various scales. In addition, replacing ordinary convolutions in the network model with more efficient depth-wise separable convolutions increases detection speed while slightly reducing network detection accuracy. Transfer learning is introduced, and a training method involving freezing and unfreezing the model is employed to enhance the network's ability to detect small target defects. The proposed model is validated using the insulator self-exploding defect dataset. The experimental results show that Faster R-CNN-tiny significantly outperforms the Faster R-CNN (ResNet) model in terms of mean average precision (mAP), frames per second (FPS), and number of parameters.
Collapse
Affiliation(s)
- Yanping Chen
- School of Artificial Intelligence and Big Data, Hefei University, Hefei 230601, China
| | - Chong Deng
- School of Artificial Intelligence and Big Data, Hefei University, Hefei 230601, China
| | - Qiang Sun
- School of Artificial Intelligence and Big Data, Hefei University, Hefei 230601, China
| | - Zhize Wu
- School of Artificial Intelligence and Big Data, Hefei University, Hefei 230601, China
| | - Le Zou
- School of Artificial Intelligence and Big Data, Hefei University, Hefei 230601, China
| | - Guanhong Zhang
- School of Artificial Intelligence and Big Data, Hefei University, Hefei 230601, China
| | - Wenbo Li
- Institute of Intelligent Machines, Chinese Academy of Sciences, Hefei 230001, China
| |
Collapse
|
6
|
Yang F, Huo J, Cheng Z, Chen H, Shi Y. An Improved Mask R-CNN Micro-Crack Detection Model for the Surface of Metal Structural Parts. Sensors (Basel) 2023; 24:62. [PMID: 38202924 PMCID: PMC10780529 DOI: 10.3390/s24010062] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/14/2023] [Revised: 11/19/2023] [Accepted: 11/23/2023] [Indexed: 01/12/2024]
Abstract
Micro-crack detection is an essential task in critical equipment health monitoring. Accurate and timely detection of micro-cracks can ensure the healthy and stable service of equipment. Aiming at improving the low accuracy of the conventional target detection model during the task of detecting micro-cracks on the surface of metal structural parts, this paper built a micro-cracks dataset and explored a detection performance optimization method based on Mask R-CNN. Firstly, we improved the original FPN structure, adding a bottom-up feature fusion path to enhance the information utilization rate of the underlying feature layer. Secondly, we added the methods of deformable convolution kernel and attention mechanism to ResNet, which can improve the efficiency of feature extraction. Lastly, we modified the original loss function to optimize the network training effect and model convergence rate. The ablation comparison experiments shows that all the improvement schemes proposed in this paper have improved the performance of the original Mask R-CNN. The integration of all the improvement schemes can produce the most significant performance improvement effects in recognition, classification, and positioning simultaneously, thus proving the rationality and feasibility of the improved scheme in this paper.
Collapse
Affiliation(s)
| | - Junzhou Huo
- School of Mechanical Engineering, Dalian University of Technology, Dalian 116024, China; (F.Y.); (Z.C.); (H.C.); (Y.S.)
| | | | | | | |
Collapse
|
7
|
Xue X, Luo Q, Ji Y, Ma Z, Zhu J, Li Z, Lyu S, Sun D, Song S. Design and test of Kinect-based variable spraying control system for orchards. Front Plant Sci 2023; 14:1297879. [PMID: 38186603 PMCID: PMC10768018 DOI: 10.3389/fpls.2023.1297879] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/20/2023] [Accepted: 12/08/2023] [Indexed: 01/09/2024]
Abstract
Target detection technology and variable-rate spraying technology are key technologies for achieving precise and efficient pesticide application. To address the issues of low efficiency and high working environment requirements in detecting tree information during variable spraying in orchards, this study has designed a variable spraying control system. The system employed a Kinect sensor to real-time detect the canopy volume of citrus trees and adjusted the duty cycle of solenoid valves by pulse width modulation to control the pesticide application. A canopy volume calculation method was proposed, and precision tests for volume detection were conducted, with a maximum relative error of 10.54% compared to manual measurements. A nozzle flow model was designed to determine the spray decision coefficient. When the duty cycle ranged from 30% to 90%, the correlation coefficient of the flow model exceeded 0.95, and the actual flow rate of the system was similar to the theoretical flow rate. Field experiments were conducted to evaluate the spraying effectiveness of the variable spraying control system based on the Kinect sensor. The experimental results indicated that the variable spraying control system demonstrated good consistency between the theoretical spray volume and the actual spray volume. In deposition tests, compared to constant-rate spraying, the droplets under the variable-rate mode based on canopy volume exhibited higher deposition density. Although the amount of droplet deposit and coverage slightly decreased, they still met the requirements for spraying operation quality. Additionally, the variable-rate spray mode achieved the goal of reducing pesticide use, with a maximum pesticide saving rate of 57.14%. This study demonstrates the feasibility of the Kinect sensor in guiding spraying operations and provides a reference for their application in plant protection operations.
Collapse
Affiliation(s)
- Xiuyun Xue
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
- National Citrus Industry Technical System Machinery Research Office, Guangzhou, China
- Guangdong Provincial Agricultural Information Monitoring Engineering Technology Research Center, Guangzhou, China
- Meizhou SCAU-Zhensheng Research Institute for Modern Agricultural Equipment, Meizhou, China
| | - Qin Luo
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
| | - Yihang Ji
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
| | - Zhaoyong Ma
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
| | - Jiani Zhu
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
| | - Zhen Li
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
- National Citrus Industry Technical System Machinery Research Office, Guangzhou, China
- Guangdong Provincial Agricultural Information Monitoring Engineering Technology Research Center, Guangzhou, China
- Meizhou SCAU-Zhensheng Research Institute for Modern Agricultural Equipment, Meizhou, China
| | - Shilei Lyu
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
- National Citrus Industry Technical System Machinery Research Office, Guangzhou, China
- Guangdong Provincial Agricultural Information Monitoring Engineering Technology Research Center, Guangzhou, China
| | - Daozong Sun
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
- National Citrus Industry Technical System Machinery Research Office, Guangzhou, China
- Guangdong Provincial Agricultural Information Monitoring Engineering Technology Research Center, Guangzhou, China
| | - Shuran Song
- College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou, China
- National Citrus Industry Technical System Machinery Research Office, Guangzhou, China
- Guangdong Provincial Agricultural Information Monitoring Engineering Technology Research Center, Guangzhou, China
| |
Collapse
|
8
|
Liang B, Wang X, Zhao W, Wang X. High-Precision Carton Detection Based on Adaptive Image Augmentation for Unmanned Cargo Handling Tasks. Sensors (Basel) 2023; 24:12. [PMID: 38202874 PMCID: PMC10780547 DOI: 10.3390/s24010012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Revised: 12/04/2023] [Accepted: 12/15/2023] [Indexed: 01/12/2024]
Abstract
Unattended intelligent cargo handling is an important means to improve the efficiency and safety of port cargo trans-shipment, where high-precision carton detection is an unquestioned prerequisite. Therefore, this paper introduces an adaptive image augmentation method for high-precision carton detection. First, the imaging parameters of the images are clustered into various scenarios, and the imaging parameters and perspectives are adaptively adjusted to achieve the automatic augmenting and balancing of the carton dataset in each scenario, which reduces the interference of the scenarios on the carton detection precision. Then, the carton boundary features are extracted and stochastically sampled to synthesize new images, thus enhancing the detection performance of the trained model for dense cargo boundaries. Moreover, the weight function of the hyperparameters of the trained model is constructed to achieve their preferential crossover during genetic evolution to ensure the training efficiency of the augmented dataset. Finally, an intelligent cargo handling platform is developed and field experiments are conducted. The outcomes of the experiments reveal that the method attains a detection precision of 0.828. This technique significantly enhances the detection precision by 18.1% and 4.4% when compared to the baseline and other methods, which provides a reliable guarantee for intelligent cargo handling processes.
Collapse
Affiliation(s)
- Bing Liang
- Naval Architecture and Ocean Engineering College, Dalian Maritime University, Dalian 116026, China; (X.W.); (W.Z.); (X.W.)
| | | | | | | |
Collapse
|
9
|
Varshney N, Berweger S, Chuang J, Blandino S, Wang J, Pazare N, Gentile C, Golmie N. Adaptive Channel-State-Information Feedback in Integrated Sensing and Communication Systems. IEEE Internet Things J 2023; 10:22469-22486. [PMID: 38348220 PMCID: PMC10860372 DOI: 10.1109/jiot.2023.3304545] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/15/2024]
Abstract
Efficient design of integrated sensing and communication systems can minimize signaling overhead by reducing the size and/or rate of feedback in reporting channel state information (CSI). To minimize the signaling overhead when performing sensing operations at the transmitter, this paper proposes a procedure to reduce the feedback rate. We consider a threshold-based sensing measurement and reporting procedure, such that the CSI is transmitted only if the channel variation exceeds a threshold. However, quantifying the channel variation, determining the threshold, and recovering sensing information with a lower feedback rate are still open problems. In this paper, we first quantify the channel variation by considering several metrics including the Euclidean distance, time-reversal resonating strength, and frequency-reversal resonating strength. We then design an algorithm to adaptively select a threshold, minimizing the feedback rate, while guaranteeing sufficient sensing accuracy by reconstructing high-quality signatures of human movement. To improve sensing accuracy with irregular channel measurements, we further propose two reconstruction schemes, which can be easily employed at the transmitter in case there is no feedback available from the receiver. Finally, the sensing performance of our scheme is extensively evaluated through real and synthetic channel measurements, considering channel estimation and synchronization errors. Our results show that the amount of feedback can be reduced by 50% while maintaining good sensing performance in terms of range and velocity estimations. Moreover, in contrast to other schemes, we show that the Euclidean distance metric is better able to capture various human movements with high channel variation values.
Collapse
Affiliation(s)
- Neeraj Varshney
- Radio Access and Propagation Metrology Group, National Institute of Standards and Technology (NIST), Gaithersburg, MD 20899-6730 USA and contractor with Prometheus Computing LLC, Cullowhee, NC USA
| | - Samuel Berweger
- Communications Technology Laboratory, National Institute of Standards and Technology, Gaithersburg, MD 20899-6730 USA
| | - Jack Chuang
- Radio Access and Propagation Metrology Group, National Institute of Standards and Technology, Gaithersburg, MD 20899-6730 USA
| | - Steve Blandino
- Radio Access and Propagation Metrology Group, National Institute of Standards and Technology (NIST), Gaithersburg, MD 20899-6730 USA and contractor with Prometheus Computing LLC, Cullowhee, NC USA
- Communications Technology Laboratory, National Institute of Standards and Technology, Gaithersburg, MD 20899-6730 USA
| | - Jian Wang
- Radio Access and Propagation Metrology Group, National Institute of Standards and Technology, Gaithersburg, MD 20899-6730 USA
| | - Neha Pazare
- Department of Electrical Engineering, University of Colorado, Boulder, Colorado 80309, USA
| | - Camillo Gentile
- Radio Access and Propagation Metrology Group, National Institute of Standards and Technology, Gaithersburg, MD 20899-6730 USA
| | - Nada Golmie
- Communications Technology Laboratory, National Institute of Standards and Technology, Gaithersburg, MD 20899-6730 USA
| |
Collapse
|
10
|
Panagos II, Giotis AP, Sofianopoulos S, Nikou C. A New Benchmark for Consumer Visual Tracking and Apparent Demographic Estimation from RGB and Thermal Images. Sensors (Basel) 2023; 23:9510. [PMID: 38067883 PMCID: PMC10708599 DOI: 10.3390/s23239510] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 11/20/2023] [Accepted: 11/23/2023] [Indexed: 12/18/2023]
Abstract
Visual tracking and attribute estimation related to age or gender information of multiple person entities in a scene are mature research topics with the advent of deep learning techniques. However, when it comes to indoor images such as video sequences of retail consumers, data are not always adequate or accurate enough to essentially train effective models for consumer detection and tracking under various adverse factors. This in turn affects the quality of recognizing age or gender for those detected instances. In this work, we introduce two novel datasets: Consumers comprises 145 video sequences compliant to personal information regulations as far as facial images are concerned and BID is a set of cropped body images from each sequence that can be used for numerous computer vision tasks. We also propose an end-to-end framework which comprises CNNs as object detectors, LSTMs for motion forecasting of the tracklet association component in a sequence, along with a multi-attribute classification model for apparent demographic estimation of the detected outputs, aiming to capture useful metadata of consumer product preferences. Obtained results on tracking and age/gender prediction are promising with respect to reference systems while they indicate the proposed model's potential for practical consumer metadata extraction.
Collapse
Affiliation(s)
- Iason-Ioannis Panagos
- Department of Computer Science and Engineering (CSE), University of Ioannina, 45110 Ioannina, Greece; (I.-I.P.); (C.N.)
| | - Angelos P. Giotis
- Department of Computer Science and Engineering (CSE), University of Ioannina, 45110 Ioannina, Greece; (I.-I.P.); (C.N.)
| | - Sokratis Sofianopoulos
- Institute for Language and Speech Processing (ILSP), Athena Research and Innovation Center, 15125 Athens, Greece;
| | - Christophoros Nikou
- Department of Computer Science and Engineering (CSE), University of Ioannina, 45110 Ioannina, Greece; (I.-I.P.); (C.N.)
| |
Collapse
|
11
|
Zhang Q, Yu W, Liu W, Xu H, He Y. A Lightweight Visual Simultaneous Localization and Mapping Method with a High Precision in Dynamic Scenes. Sensors (Basel) 2023; 23:9274. [PMID: 38005660 PMCID: PMC10675022 DOI: 10.3390/s23229274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Revised: 11/07/2023] [Accepted: 11/08/2023] [Indexed: 11/26/2023]
Abstract
Currently, in most traditional VSLAM (visual SLAM) systems, static assumptions result in a low accuracy in dynamic environments, or result in a new and higher level of accuracy but at the cost of sacrificing the real-time property. In highly dynamic scenes, balancing a high accuracy and a low computational cost has become a pivotal requirement for VSLAM systems. This paper proposes a new VSLAM system, balancing the competitive demands between positioning accuracy and computational complexity and thereby further improving the overall system properties. From the perspective of accuracy, the system applies an improved lightweight target detection network to quickly detect dynamic feature points while extracting feature points at the front end of the system, and only feature points of static targets are applied for frame matching. Meanwhile, the attention mechanism is integrated into the target detection network to continuously and accurately capture dynamic factors to cope with more complex dynamic environments. From the perspective of computational expense, the lightweight network Ghostnet module is applied as the backbone network of the target detection network YOLOv5s, significantly reducing the number of model parameters and improving the overall inference speed of the algorithm. Experimental results on the TUM dynamic dataset indicate that in contrast with the ORB-SLAM3 system, the pose estimation accuracy of the system improved by 84.04%. In contrast with dynamic SLAM systems such as DS-SLAM and DVO SLAM, the system has a significantly improved positioning accuracy. In contrast with other VSLAM algorithms based on deep learning, the system has superior real-time properties while maintaining a similar accuracy index.
Collapse
Affiliation(s)
- Qi Zhang
- School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410018, China; (Q.Z.); (H.X.); (Y.H.)
| | - Wentao Yu
- School of Computer, Central South University, Changsha 410083, China;
| | - Weirong Liu
- School of Computer, Central South University, Changsha 410083, China;
| | - Hao Xu
- School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410018, China; (Q.Z.); (H.X.); (Y.H.)
| | - Yuan He
- School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410018, China; (Q.Z.); (H.X.); (Y.H.)
| |
Collapse
|
12
|
Lian J, Qiao X, Zhao Y, Li S, Wang C, Zhou J. EEG-Based Target Detection Using an RSVP Paradigm under Five Levels of Weak Hidden Conditions. Brain Sci 2023; 13:1583. [PMID: 38002543 PMCID: PMC10670035 DOI: 10.3390/brainsci13111583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 11/06/2023] [Accepted: 11/07/2023] [Indexed: 11/26/2023] Open
Abstract
Although target detection based on electroencephalogram (EEG) signals has been extensively investigated recently, EEG-based target detection under weak hidden conditions remains a problem. In this paper, we proposed a rapid serial visual presentation (RSVP) paradigm for target detection corresponding to five levels of weak hidden conditions quantitively based on the RGB color space. Eighteen subjects participated in the experiment, and the neural signatures, including P300 amplitude and latency, were investigated. Detection performance was evaluated under five levels of weak hidden conditions using the linear discrimination analysis and support vector machine classifiers on different channel sets. The experimental results showed that, compared with the benchmark condition, (1) the P300 amplitude significantly decreased (8.92 ± 1.24 μV versus 7.84 ± 1.40 μV, p = 0.021) and latency was significantly prolonged (582.39 ± 25.02 ms versus 643.83 ± 26.16 ms, p = 0.028) only under the weakest hidden condition, and (2) the detection accuracy decreased by less than 2% (75.04 ± 3.24% versus 73.35 ± 3.15%, p = 0.029) with a more than 90% reduction in channel number (62 channels versus 6 channels), determined using the proposed channel selection method under the weakest hidden condition. Our study can provide new insights into target detection under weak hidden conditions based on EEG signals with a rapid serial visual presentation paradigm. In addition, it may expand the application of brain-computer interfaces in EEG-based target detection areas.
Collapse
Affiliation(s)
- Jinling Lian
- Department of Neural Engineering and Biological Interdisciplinary Studies, Beijing Institute of Basic Medical Sciences, 27 Taiping Rd., Beijing 100850, China; (J.L.); (X.Q.); (Y.Z.); (S.L.)
| | - Xin Qiao
- Department of Neural Engineering and Biological Interdisciplinary Studies, Beijing Institute of Basic Medical Sciences, 27 Taiping Rd., Beijing 100850, China; (J.L.); (X.Q.); (Y.Z.); (S.L.)
| | - Yuwei Zhao
- Department of Neural Engineering and Biological Interdisciplinary Studies, Beijing Institute of Basic Medical Sciences, 27 Taiping Rd., Beijing 100850, China; (J.L.); (X.Q.); (Y.Z.); (S.L.)
| | - Siwei Li
- Department of Neural Engineering and Biological Interdisciplinary Studies, Beijing Institute of Basic Medical Sciences, 27 Taiping Rd., Beijing 100850, China; (J.L.); (X.Q.); (Y.Z.); (S.L.)
| | - Changyong Wang
- Department of Neural Engineering and Biological Interdisciplinary Studies, Beijing Institute of Basic Medical Sciences, 27 Taiping Rd., Beijing 100850, China; (J.L.); (X.Q.); (Y.Z.); (S.L.)
| | - Jin Zhou
- Department of Neural Engineering and Biological Interdisciplinary Studies, Beijing Institute of Basic Medical Sciences, 27 Taiping Rd., Beijing 100850, China; (J.L.); (X.Q.); (Y.Z.); (S.L.)
- Chinese Institute for Brain Research, Zhongguancun Life Science Park, Changping District, Beijing 102206, China
| |
Collapse
|
13
|
Zhan Z, Li L, Lin Y, Lv Z, Zhang H, Li X, Zhang F, Zeng Y. Rapid and accurate detection of multi-target walnut appearance quality based on the lightweight improved YOLOv5s_AMM model. Front Plant Sci 2023; 14:1247156. [PMID: 38023833 PMCID: PMC10663328 DOI: 10.3389/fpls.2023.1247156] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/25/2023] [Accepted: 10/11/2023] [Indexed: 12/01/2023]
Abstract
Introduction Nut quality detection is of paramount importance in primary nut processing. When striving to maintain the imperatives of rapid, efficient, and accurate detection, the precision of identifying small-sized nuts can be substantially compromised. Methods We introduced an optimized iteration of the YOLOv5s model designed to swiftly and precisely identify both good and bad walnut nuts across multiple targets. The M3-Net network, which is a replacement for the original C3 network in MobileNetV3's YOLOv5s, reduces the weight of the model. We explored the impact of incorporating the attention mechanism at various positions to enhance model performance. Furthermore, we introduced an attentional convolutional adaptive fusion module (Acmix) within the spatial pyramid pooling layer to improve feature extraction. In addition, we replaced the SiLU activation function in the original Conv module with MetaAconC from the CBM module to enhance feature detection in walnut images across different scales. Results In comparative trials, the YOLOv5s_AMM model surpassed the standard detection networks, exhibiting an average detection accuracy (mAP) of 80.78%, an increase of 1.81%, while reducing the model size to 20.9 MB (a compression of 22.88%) and achieving a detection speed of 40.42 frames per second. In multi-target walnut detection across various scales, the enhanced model consistently outperformed its predecessor in terms of accuracy, model size, and detection speed. It notably improves the ability to detect multi-target walnut situations, both large and small, while maintaining the accuracy and efficiency. Discussion The results underscored the superiority of the YOLOv5s_AMM model, which achieved the highest average detection accuracy (mAP) of 80.78%, while boasting the smallest model size at 20.9 MB and the highest frame rate of 40.42 FPS. Our optimized network excels in the rapid, efficient, and accurate detection of mixed multi-target dry walnut quality, accommodating lightweight edge devices. This research provides valuable insights for the detection of multi-target good and bad walnuts during the walnut processing stage.
Collapse
Affiliation(s)
- Zicheng Zhan
- Laboratory of Physical Properties of Agricultural Materials, College of Modern Agricultural Engineering, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Lixia Li
- Laboratory of Physical Properties of Agricultural Materials, College of Modern Agricultural Engineering, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Yuhao Lin
- Laboratory of Physical Properties of Agricultural Materials, College of Modern Agricultural Engineering, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Zhiyuan Lv
- Laboratory of Physical Properties of Agricultural Materials, College of Modern Agricultural Engineering, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Hao Zhang
- Laboratory of Physical Properties of Agricultural Materials, College of Modern Agricultural Engineering, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Xiaoqing Li
- 69223 Troops, People’s Liberation Army, Aksu, Xinjiang Uygur Autonomous Region, China
| | - Fujie Zhang
- Laboratory of Physical Properties of Agricultural Materials, College of Modern Agricultural Engineering, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Yumin Zeng
- Project Management Division, Yunnan Provincial Forestry and Grassland Technology Extension Station, Kunming, Yunnan, China
| |
Collapse
|
14
|
Zha W, Li H, Wu G, Zhang L, Pan W, Gu L, Jiao J, Zhang Q. Research on the Recognition and Tracking of Group-Housed Pigs' Posture Based on Edge Computing. Sensors (Basel) 2023; 23:8952. [PMID: 37960652 PMCID: PMC10649120 DOI: 10.3390/s23218952] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/21/2023] [Revised: 11/01/2023] [Accepted: 11/01/2023] [Indexed: 11/15/2023]
Abstract
The existing algorithms for identifying and tracking pigs in barns generally have a large number of parameters, relatively complex networks and a high demand for computational resources, which are not suitable for deployment in embedded-edge nodes on farms. A lightweight multi-objective identification and tracking algorithm based on improved YOLOv5s and DeepSort was developed for group-housed pigs in this study. The identification algorithm was optimized by: (i) using a dilated convolution in the YOLOv5s backbone network to reduce the number of model parameters and computational power requirements; (ii) adding a coordinate attention mechanism to improve the model precision; and (iii) pruning the BN layers to reduce the computational requirements. The optimized identification model was combined with DeepSort to form the final Tracking by Detecting algorithm and ported to a Jetson AGX Xavier edge computing node. The algorithm reduced the model size by 65.3% compared to the original YOLOv5s. The algorithm achieved a recognition precision of 96.6%; a tracking time of 46 ms; and a tracking frame rate of 21.7 FPS, and the precision of the tracking statistics was greater than 90%. The model size and performance met the requirements for stable real-time operation in embedded-edge computing nodes for monitoring group-housed pigs.
Collapse
Affiliation(s)
- Wenwen Zha
- School of Information and Computer, Anhui Agricultural University, Hefei 230036, China; (W.Z.); (G.W.); (W.P.); (L.G.)
| | - Hualong Li
- Institute of Intelligent Machines, Chinese Academy of Sciences, Hefei 230031, China;
| | - Guodong Wu
- School of Information and Computer, Anhui Agricultural University, Hefei 230036, China; (W.Z.); (G.W.); (W.P.); (L.G.)
| | - Liping Zhang
- Institute of Agricultural Economy and Information, Anhui Academy of Agricultural Sciences, Hefei 230031, China;
| | - Weihao Pan
- School of Information and Computer, Anhui Agricultural University, Hefei 230036, China; (W.Z.); (G.W.); (W.P.); (L.G.)
| | - Lichuan Gu
- School of Information and Computer, Anhui Agricultural University, Hefei 230036, China; (W.Z.); (G.W.); (W.P.); (L.G.)
| | - Jun Jiao
- School of Information and Computer, Anhui Agricultural University, Hefei 230036, China; (W.Z.); (G.W.); (W.P.); (L.G.)
| | - Qiang Zhang
- Department of Biosystems Engineering, University of Manitoba, Winnipeg, MB R3T 5V6, Canada
| |
Collapse
|
15
|
Guo Y, Wang X, Han M, Xin J, Hou Y, Gong Z, Wang L, Fan D, Feng L, Han D. Detection and Localization of Albas Velvet Goats Based on YOLOv4. Animals (Basel) 2023; 13:3242. [PMID: 37893966 PMCID: PMC10603755 DOI: 10.3390/ani13203242] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Revised: 10/14/2023] [Accepted: 10/15/2023] [Indexed: 10/29/2023] Open
Abstract
In order to achieve goat localization to help prevent goats from wandering, we proposed an efficient target localization method based on machine vision. Albas velvet goats from a farm in Ertok Banner, Ordos City, Inner Mongolia Autonomous Region, China, were the main objects of study. First, we proposed detecting the goats using a shallow convolutional neural network, ShallowSE, with the channel attention mechanism SENet, the GeLU activation function and layer normalization. Second, we designed three fully connected coordinate regression network models to predict the spatial coordinates of the goats. Finally, the target detection algorithm and the coordinate regression algorithm were combined to localize the flock. We experimentally confirmed the proposed method using our dataset. The proposed algorithm obtained a good detection accuracy and successful localization rate compared to other popular algorithms. The overall number of parameters in the target detection algorithm model was only 4.5 M. The average detection accuracy reached 95.89% and the detection time was only 8.5 ms. The average localization error of the group localization algorithm was only 0.94 m and the localization time was 0.21 s. In conclusion, the method achieved fast and accurate localization, which helped to rationalize the use of grassland resources and to promote the sustainable development of rangelands.
Collapse
Affiliation(s)
- Ying Guo
- School of Information Engineering, Inner Mongolia University of Science and Technology, Baotou 014010, China;
- College of Mechanical and Electrical Engineering, Inner Mongolia Agricultural University, Hohhot 010018, China
| | - Xihao Wang
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Mingjuan Han
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Jile Xin
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Yun Hou
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Zhuo Gong
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Liang Wang
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Daoerji Fan
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Lianjie Feng
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
| | - Ding Han
- College of Electronic Information Engineering, Inner Mongolia University, Hohhot 010021, China; (X.W.); (M.H.); (J.X.); (Y.H.); (Z.G.); (L.W.); (D.F.); (L.F.)
- Inner Mongolia State Key Laboratory of Reproductive Regulation and Breeding of Grassland Livestock, Hohhot 010020, China
| |
Collapse
|
16
|
Zhang Y, Shen S, Xu S. Strip steel surface defect detection based on lightweight YOLOv5. Front Neurorobot 2023; 17:1263739. [PMID: 37860791 PMCID: PMC10582940 DOI: 10.3389/fnbot.2023.1263739] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Accepted: 09/12/2023] [Indexed: 10/21/2023] Open
Abstract
Deep learning-based methods for detecting surface defects on strip steel have advanced detection capabilities, but there are still problems of target loss, false alarms, large computation, and imbalance between detection accuracy and detection speed. In order to achieve a good balance between detection accuracy and speed, a lightweight YOLOv5 strip steel surface defect detection algorithm based on YOLOv5s is proposed. Firstly, we introduce the efficient lightweight convolutional layer called GSConv. The Slim Neck, designed based on GSConv, replaces the original algorithm's neck, reducing the number of network parameters and improving detection speed. Secondly, we incorporate SimAM, a non-parametric attention mechanism, into the improved neck to enhance detection accuracy. Finally, we utilize the SIoU function as the regression prediction loss instead of the original CIoU to address the issue of slow convergence and improve efficiency. According to experimental findings, the YOLOv5-GSS algorithm outperforms the YOLOv5 method by 2.9% on the NEU-DET dataset and achieves an average accuracy (mAP) of 83.8% with a detection speed (FPS) of 100 Hz, which is 3.8 Hz quicker than the YOLOv5 algorithm. The proposed model outperforms existing approaches and is more useful, demonstrating the efficacy of the optimization strategy.
Collapse
Affiliation(s)
- Yongping Zhang
- School of Information Engineering, Yancheng Institute of Technology, Yancheng, China
| | | | | |
Collapse
|
17
|
Chen Y, Peng X, Cai L, Jiao M, Fu D, Xu CC, Zhang P. Research on automatic classification and detection of chicken parts based on deep learning algorithm. J Food Sci 2023; 88:4180-4193. [PMID: 37655508 DOI: 10.1111/1750-3841.16747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Revised: 07/21/2023] [Accepted: 08/07/2023] [Indexed: 09/02/2023]
Abstract
Accurate classification and identification of chicken parts are critical to improve the productivity and processing speed in poultry processing plants. However, the overlapping of chicken parts has an impact on the effectiveness of the identification process. To solve this issue, this study proposed a real-time classification and detection method for chicken parts, utilizing YOLOV4 deep learning. The method can identify segmented chicken parts on the assembly line in real time and accurately, thus improving the efficiency of poultry processing. First, 600 images containing multiple chicken part samples were collected to build a chicken part dataset after using the image broadening technique, and then the dataset was divided according to the 6:2:2 division principle, with 1200 images as the training set, 400 images as the test set, and 400 images as the validation set. Second, we utilized the single-stage target detector YOLO to predict and calculate the chicken part images, obtaining the categories and positions of the chicken leg, chicken wing, and chicken breast in the image. This allowed us to achieve real-time classification and detection of chicken parts. This approach enabled real-time and efficient classification and detection of chicken parts. Finally, the mean average precision (mAP) and the processing time per image were utilized as key metrics to evaluate the effectiveness of the model. In addition, four other target detection algorithms were introduced for comparison with YOLOV4-CSPDarknet53 in this study, which include YOLOV3-Darknet53, YOLOV3-MobileNetv3, SSD-MobileNetv3, and SSD-VGG16. A comprehensive comparison test was conducted to assess the classification and detection performance of these models for chicken parts. Finally, for the chicken part dataset, the mAP of the YOLOV4-CSPDarknet53 model was 98.86% on a single image with an inference speed of 22.2 ms, which was higher than the other four models of YOLOV3-Darknet53, YOLOV3-MobileNetv3, SSD-MobileNetv3, and SSD-VGG16 mAP by 3.27%, 3.78%, 6.91%, and 6.13%, respectively. The average detection time was reduced by 13, 1.9, 6.2, and 20.3 ms, respectively. In summary, the chicken part classification and detection method proposed in this study offers numerous benefits, including the ability to detect multiple chicken parts simultaneously, as well as delivering high levels of accuracy and speed. Furthermore, this approach effectively addresses the issue of accurately identifying individual chicken parts in the presence of occlusion, thereby reducing waste on the assembly line. PRACTICAL APPLICATION: The aim of this study is to offer visual technical assistance in minimizing wastage and resource depletion during the sorting and cutting of chicken parts in poultry production and processing facilities. Furthermore, considering the diverse demands and preferences regarding chicken parts, this research can facilitate product processing that caters specifically to consumer preferences.
Collapse
Affiliation(s)
- Yan Chen
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Xianhui Peng
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Lu Cai
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Ming Jiao
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Dandan Fu
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Chen Chen Xu
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Peng Zhang
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| |
Collapse
|
18
|
Lee VG. The attentional boost effect overcomes dual-task interference in choice-response tasks. Q J Exp Psychol (Hove) 2023; 76:2241-2255. [PMID: 36717536 DOI: 10.1177/17470218231156375] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
Dual-task interference often arises when people respond to an incoming stimulus according to an arbitrary rule, such as choosing between the gas pedal and the brake when driving. Severe interference from response selection yields a brief "Psychological Refractory Period," during which a concurrent task is put on hold. Here, we show that response selection in one task does not always hamper the processing of a secondary task. Responding to a target may paradoxically enhance the processing of secondary tasks, even when the target requires complex response selection. In three experiments, participants encoded pictures of common objects to memory while simultaneously monitoring a rapid serial visual presentation (RSVP) of characters or colours. Some of the RSVP stimuli were targets, requiring participants to press one of the two buttons to report their identity; others were distractors that participants ignored. Despite the increased response selection demands on target trials, pictures encoded with the RSVP targets were better remembered than those encoded with the RSVP distractors. Contrary to previous reports and predictions from dual-task interference, the attentional boost from target detection overcomes increased interference from response selection.
Collapse
Affiliation(s)
- Vanessa G Lee
- Department of Psychology, University of Minnesota, Minneapolis, MN, USA
| |
Collapse
|
19
|
Wei J, Tang X, Liu J, Zhang Z. Detection of Pig Movement and Aggression Using Deep Learning Approaches. Animals (Basel) 2023; 13:3074. [PMID: 37835680 PMCID: PMC10571548 DOI: 10.3390/ani13193074] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Revised: 09/25/2023] [Accepted: 09/29/2023] [Indexed: 10/15/2023] Open
Abstract
Motion and aggressive behaviors in pigs provide important information for the study of social hierarchies in pigs and can be used as a selection indicator for pig health and aggression parameters. However, relying only on visual observation or surveillance video to record the number of aggressive acts is time-consuming, labor-intensive, and lasts for only a short period of time. Manual observation is too short compared to the growth cycle of pigs, and complete recording is impractical in large farms. In addition, due to the complex process of assessing the intensity of pig aggression, manual recording is highly influenced by human subjective vision. In order to efficiently record pig motion and aggressive behaviors as parameters for breeding selection and behavioral studies, the videos and pictures were collected from typical commercial farms, with each unit including 8~20 pigs in 7~25 m2 space; they were bred in stable social groups and a video was set up to record the whole day's activities. We proposed a deep learning-based recognition method for detecting and recognizing the movement and aggressive behaviors of pigs by recording and annotating head-to-head tapping, head-to-body tapping, neck biting, body biting, and ear biting during fighting. The method uses an improved EMA-YOLOv8 model and a target tracking algorithm to assign a unique digital identity code to each pig, while efficiently recognizing and recording pig motion and aggressive behaviors and tracking them, thus providing statistics on the speed and duration of pig motion. On the test dataset, the average precision of the model was 96.4%, indicating that the model has high accuracy in detecting a pig's identity and its fighting behaviors. The model detection results were highly correlated with the manual recording results (R2 of 0.9804 and 0.9856, respectively), indicating that the method has high accuracy and effectiveness. In summary, the method realized the detection and identification of motion duration and aggressive behavior of pigs under natural conditions, and provided reliable data and technical support for the study of the social hierarchy of pigs and the selection of pig health and aggression phenotypes.
Collapse
Affiliation(s)
| | | | | | - Zhiyan Zhang
- State Key Laboratory for Pig Genetic Improvement and Production Technology, Jiangxi Agricultural University, Nanchang 330045, China
| |
Collapse
|
20
|
Zhang N, Liu W, Xia X. Video Global Motion Compensation Based on Affine Inverse Transform Model. Sensors (Basel) 2023; 23:7750. [PMID: 37765806 PMCID: PMC10534421 DOI: 10.3390/s23187750] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/21/2023] [Revised: 09/02/2023] [Accepted: 09/05/2023] [Indexed: 09/29/2023]
Abstract
Global motion greatly increases the number of false alarms for object detection in video sequences against dynamic backgrounds. Therefore, before detecting the target in the dynamic background, it is necessary to estimate and compensate the global motion to eliminate the influence of the global motion. In this paper, we use the SURF (speeded up robust features) algorithm combined with the MSAC (M-Estimate Sample Consensus) algorithm to process the video. The global motion of a video sequence is estimated according to the feature point matching pairs of adjacent frames of the video sequence and the global motion parameters of the video sequence under the dynamic background. On this basis, we propose an inverse transformation model of affine transformation, which acts on each adjacent frame of the video sequence in turn. The model compensates the global motion, and outputs a video sequence after global motion compensation from a specific view for object detection. Experimental results show that the algorithm proposed in this paper can accurately perform motion compensation on video sequences containing complex global motion, and the compensated video sequences achieve higher peak signal-to-noise ratio and better visual effects.
Collapse
Affiliation(s)
- Nan Zhang
- School of Electrical and Control Engineering, Shaanxi University of Science and Technology, Xi’an 710021, China;
| | - Weifeng Liu
- School of Electrical and Control Engineering, Shaanxi University of Science and Technology, Xi’an 710021, China;
| | - Xingyu Xia
- School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China;
| |
Collapse
|
21
|
Wang J, Alshahir A, Abbas G, Kaaniche K, Albekairi M, Alshahr S, Aljarallah W, Sahbani A, Nowakowski G, Sieja M. A Deep Recurrent Learning-Based Region-Focused Feature Detection for Enhanced Target Detection in Multi-Object Media. Sensors (Basel) 2023; 23:7556. [PMID: 37688012 PMCID: PMC10490795 DOI: 10.3390/s23177556] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/13/2023] [Revised: 08/25/2023] [Accepted: 08/29/2023] [Indexed: 09/10/2023]
Abstract
Target detection in high-contrast, multi-object images and movies is challenging. This difficulty results from different areas and objects/people having varying pixel distributions, contrast, and intensity properties. This work introduces a new region-focused feature detection (RFD) method to tackle this problem and improve target detection accuracy. The RFD method divides the input image into several smaller ones so that as much of the image as possible is processed. Each of these zones has its own contrast and intensity attributes computed. Deep recurrent learning is then used to iteratively extract these features using a similarity measure from training inputs corresponding to various regions. The target can be located by combining features from many locations that overlap. The recognized target is compared to the inputs used during training, with the help of contrast and intensity attributes, to increase accuracy. The feature distribution across regions is also used for repeated training of the learning paradigm. This method efficiently lowers false rates during region selection and pattern matching with numerous extraction instances. Therefore, the suggested method provides greater accuracy by singling out distinct regions and filtering out misleading rate-generating features. The accuracy, similarity index, false rate, extraction ratio, processing time, and others are used to assess the effectiveness of the proposed approach. The proposed RFD improves the similarity index by 10.69%, extraction ratio by 9.04%, and precision by 13.27%. The false rate and processing time are reduced by 7.78% and 9.19%, respectively.
Collapse
Affiliation(s)
- Jinming Wang
- College of Information Science & Technology, Zhejiang Shuren University, Hangzhou 310015, China
| | - Ahmed Alshahir
- Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
| | - Ghulam Abbas
- School of Electrical Engineering, Southeast University, Nanjing 210096, China
| | - Khaled Kaaniche
- Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
| | - Mohammed Albekairi
- Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
| | - Shahr Alshahr
- Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
| | - Waleed Aljarallah
- Department of Electrical Engineering, College of Engineering, Jouf University, Sakakah 72388, Saudi Arabia
| | - Anis Sahbani
- Institute for Intelligent Systems and Robotics (ISIR), CNRS, Sorbonne University, 75006 Paris, France
| | - Grzegorz Nowakowski
- Faculty of Electrical and Computer Engineering, Cracow University of Technology, Warszawska 24 Str., 31-155 Cracow, Poland
| | - Marek Sieja
- Faculty of Electrical and Computer Engineering, Cracow University of Technology, Warszawska 24 Str., 31-155 Cracow, Poland
| |
Collapse
|
22
|
Bao Z, Guo Y, Wang J, Zhu L, Huang J, Yan S. Underwater Target Detection Based on Parallel High-Resolution Networks. Sensors (Basel) 2023; 23:7337. [PMID: 37687793 PMCID: PMC10490014 DOI: 10.3390/s23177337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 08/02/2023] [Accepted: 08/20/2023] [Indexed: 09/10/2023]
Abstract
A parallel high-resolution underwater target detection network is proposed to address the problems of complex underwater scenes and limited target feature extraction capability. First, a high-resolution network (HRNet), a lighter high-resolution human posture estimation network, is used to improve the target feature representation and effectively reduce the semantic information lost in the image during sampling. Then, the attention module (A-CBAM) is improved to capture complex feature distributions by modeling the two-dimensional space in the activation function stage through the introduction of the flexible rectified linear units (FReLU) activation function to achieve pixel-level spatial information modeling capability. Feature enhancement in the spatial and channel dimensions is performed to improve understanding of fuzzy targets and small target objects and to better capture irregular and detailed object layouts. Finally, a receptive field augmentation module (RFAM) is constructed to obtain sufficient semantic information and rich detail information to further enhance the robustness and discrimination of features and improve the detection capability of the model for multi-scale underwater targets. Experimental results show that the method achieves 81.17%, 77.02%, and 82.9% mean average precision (mAP) on three publicly available datasets, specifically underwater robot professional contest (URPC2020, URPC2018) and pattern analysis, statistical modeling, and computational learning visual object classes (PASCAL VOC2007), respectively, demonstrating the effectiveness of the proposed network.
Collapse
Affiliation(s)
- Zhengwei Bao
- College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
- Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
| | - Ying Guo
- College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
- Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
| | - Jiyu Wang
- College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
- Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
| | - Linlin Zhu
- College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
- Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
| | - Jun Huang
- College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
- Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
| | - Shu Yan
- College of Automation, Nanjing University of Information Science & Technology, Nanjing 210044, China
- Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET), Nanjing 210044, China
| |
Collapse
|
23
|
Li M, Fang S, Wang X, Chen S, Cao L, Han J, Yun H. Peripheral Blood Leukocyte Detection Based on an Improved Detection Transformer Algorithm. Sensors (Basel) 2023; 23:7226. [PMID: 37631762 PMCID: PMC10459921 DOI: 10.3390/s23167226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Revised: 08/11/2023] [Accepted: 08/11/2023] [Indexed: 08/27/2023]
Abstract
The combination of a blood cell analyzer and artificial microscopy to detect white blood cells is used in hospitals. Blood cell analyzers not only have large throughput, but they also cannot detect cell morphology; although artificial microscopy has high accuracy, it is inefficient and prone to missed detections. In view of the above problems, a method based on Fourier ptychographic microscopy (FPM) and deep learning to detect peripheral blood leukocytes is proposed in this paper. Firstly, high-resolution and wide-field microscopic images of human peripheral blood cells are obtained using the FPM system, and the cell image data are enhanced with DCGANs (deep convolution generative adversarial networks) to construct datasets for performance evaluation. Then, an improved DETR (detection transformer) algorithm is proposed to improve the detection accuracy of small white blood cell targets; that is, the residual module Conv Block in the feature extraction part of the DETR network is improved to reduce the problem of information loss caused by downsampling. Finally, CIOU (complete intersection over union) is introduced as the bounding box loss function, which avoids the problem that GIOU (generalized intersection over union) is difficult to optimize when the two boxes are far away and the convergence speed is faster. The experimental results show that the mAP of the improved DETR algorithm in the detection of human peripheral white blood cells is 0.936. In addition, this algorithm is compared with other convolutional neural networks in terms of average accuracy, parameters, and number of inference frames per second, which verifies the feasibility of this method in microscopic medical image detection.
Collapse
Affiliation(s)
| | | | - Xiaoli Wang
- School of Electronic Information Engineering, Changchun University, Changchun 130000, China; (M.L.)
| | | | | | | | | |
Collapse
|
24
|
Xie L, Huang J, Li Y, Guo J. An improved model for target detection and pose estimation of a teleoperation power manipulator. Front Neurorobot 2023; 17:1193823. [PMID: 37600466 PMCID: PMC10433371 DOI: 10.3389/fnbot.2023.1193823] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2023] [Accepted: 07/17/2023] [Indexed: 08/22/2023] Open
Abstract
Introduction A hot cell is generally deployed with a teleoperation power manipulator to complete tests, operations, and maintenance. The position and pose of the manipulator are mostly acquired through radiation-resistant video cameras arranged in the hot cell. In this paper, deep learning-based target detection technology is used to establish an experimental platform to test the methods for target detection and pose estimation of teleoperation power manipulators using two cameras. Methods In view of the fact that a complex environment affects the precision of manipulator pose estimation, the dilated-fully convolutional one-stage object detection (dilated-FCOS) teleoperation power manipulator target detection algorithm is proposed based on the scale of the teleoperation power manipulator. Model pruning is used to improve the real-time performance of the dilated-FCOS teleoperation power manipulator target detection model. To improve the detection speed for the key points of the teleoperation power manipulator, the keypoint detection precision and model inference speed of different lightweight backbone networks were tested based on the SimpleBaseline algorithm. MobileNetv1 was selected as the backbone network to perform channel compression and pose distillation on the upsampling module so as to further optimize the inference speed of the model. Results and discussion Compared with the original model, the proposed model was experimentally proven to reach basically the same precision within a shorter inference time (only 58% of that of the original model). The experimental results show that the compressed model basically retains the precision of the original model and that its inference time is 48% of that of the original model.
Collapse
Affiliation(s)
- Li Xie
- School of Mechanical Engineering, Dongguan University of Technology, Dongguan, China
| | - Jiale Huang
- School of Mechanical Engineering, Dongguan University of Technology, Dongguan, China
- School of Electromechanical Engineering, Guangdong University of Technology, Guangzhou, China
| | - Yutian Li
- School of Mechanical Engineering, Dongguan University of Technology, Dongguan, China
| | - Jianwen Guo
- School of Mechanical Engineering, Dongguan University of Technology, Dongguan, China
| |
Collapse
|
25
|
Chen J, Qiu L, Zhu Z, Sun N, Huang H, Ip WH, Yung KL. An Adaptive Infrared Small-Target-Detection Fusion Algorithm Based on Multiscale Local Gradient Contrast for Remote Sensing. Micromachines (Basel) 2023; 14:1552. [PMID: 37630088 PMCID: PMC10456515 DOI: 10.3390/mi14081552] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/13/2023] [Revised: 07/26/2023] [Accepted: 07/27/2023] [Indexed: 08/27/2023]
Abstract
Space vehicles such as missiles and aircraft have relatively long tracking distances. Infrared (IR) detectors are used for small target detection. The target presents point target characteristics, which lack contour, shape, and texture information. The high-brightness cloud edge and high noise have an impact on the detection of small targets because of the complex background of the sky and ground environment. Traditional template-based filtering and local contrast-based methods do not distinguish between different complex background environments, and their strategy is to unify small-target template detection or to use absolute contrast differences; so, it is easy to have a high false alarm rate. It is necessary to study the detection and tracking methods in complex backgrounds and low signal-to-clutter ratios (SCRs). We use the complexity difference as a prior condition for detection in the background of thick clouds and ground highlight buildings. Then, we use the spatial domain filtering and improved local contrast joint algorithm to obtain a significant area. We also provide a new definition of gradient uniformity through the improvement of the local gradient method, which could further enhance the target contrast. It is important to distinguish between small targets, highlighted background edges, and noise. Furthermore, the method can be used for parallel computing. Compared with the traditional space filtering algorithm or local contrast algorithm, the flexible fusion strategy can achieve the rapid detection of small targets with a higher signal-to-clutter ratio gain (SCRG) and background suppression factor (BSF).
Collapse
Affiliation(s)
- Juan Chen
- Innovation Academy for Microsatellites of Chinese Academy of Sciences, Shanghai 200120, China; (L.Q.); (Z.Z.); (N.S.)
- University of Chinese Academy of Sciences, Beijing 100000, China
| | - Lin Qiu
- Innovation Academy for Microsatellites of Chinese Academy of Sciences, Shanghai 200120, China; (L.Q.); (Z.Z.); (N.S.)
- University of Chinese Academy of Sciences, Beijing 100000, China
| | - Zhencai Zhu
- Innovation Academy for Microsatellites of Chinese Academy of Sciences, Shanghai 200120, China; (L.Q.); (Z.Z.); (N.S.)
- University of Chinese Academy of Sciences, Beijing 100000, China
| | - Ning Sun
- Innovation Academy for Microsatellites of Chinese Academy of Sciences, Shanghai 200120, China; (L.Q.); (Z.Z.); (N.S.)
| | - Hao Huang
- Hubei Key Lab of Ferro & Piezoelectric Materials and Devices, Faculty of Physics and Electronic Science, Hubei University, Wuhan 430062, China
| | - Wai-Hung Ip
- Department of Industrial and Systems Engineering, The Hong Kong Polytechnic University, Hong Kong 100872, China; (W.-H.I.); (K.-L.Y.)
- School of Engineering, University of Saskatechewan, Saskatoon, SK S7K 0C8, Canada
| | - Kai-Leung Yung
- Department of Industrial and Systems Engineering, The Hong Kong Polytechnic University, Hong Kong 100872, China; (W.-H.I.); (K.-L.Y.)
| |
Collapse
|
26
|
Fan X, Ding W, Qin W, Xiao D, Min L, Yuan H. Fusing Self-Attention and CoordConv to Improve the YOLOv5s Algorithm for Infrared Weak Target Detection. Sensors (Basel) 2023; 23:6755. [PMID: 37571539 PMCID: PMC10422332 DOI: 10.3390/s23156755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/16/2023] [Revised: 07/21/2023] [Accepted: 07/24/2023] [Indexed: 08/13/2023]
Abstract
Convolutional neural networks have achieved good results in target detection in many application scenarios, but convolutional neural networks still face great challenges when facing scenarios with small target sizes and complex background environments. To solve the problem of low accuracy of infrared weak target detection in complex scenes, and considering the real-time requirements of the detection task, we choose the YOLOv5s target detection algorithm for improvement. We add the Bottleneck Transformer structure and CoordConv to the network to optimize the model parameters and improve the performance of the detection network. Meanwhile, a two-dimensional Gaussian distribution is used to describe the importance of pixel points in the target frame, and the normalized Guassian Wasserstein distance (NWD) is used to measure the similarity between the prediction frame and the true frame to characterize the loss function of weak targets, which will help highlight the targets with flat positional deviation transformation and improve the detection accuracy. Finally, through experimental verification, compared with other mainstream detection algorithms, the improved algorithm in this paper significantly improves the target detection accuracy, with the mAP reaching 96.7 percent, which is 2.2 percentage points higher compared with Yolov5s.
Collapse
Affiliation(s)
- Xiangsuo Fan
- School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China; (X.F.); (W.Q.); (D.X.); (H.Y.)
- Guangxi Collaborative Innovation Centre for Earthmoving Machinery, Guangxi University of Science and Technology, Liuzhou 545006, China
| | - Wentao Ding
- School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China; (X.F.); (W.Q.); (D.X.); (H.Y.)
| | - Wenlin Qin
- School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China; (X.F.); (W.Q.); (D.X.); (H.Y.)
| | - Dachuan Xiao
- School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China; (X.F.); (W.Q.); (D.X.); (H.Y.)
| | - Lei Min
- Institute of Optics and Electronics Chinese Academy of Sciences, Chengdu 610209, China;
| | - Haohao Yuan
- School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China; (X.F.); (W.Q.); (D.X.); (H.Y.)
| |
Collapse
|
27
|
Li P, Zheng J, Li P, Long H, Li M, Gao L. Tomato Maturity Detection and Counting Model Based on MHSA-YOLOv8. Sensors (Basel) 2023; 23:6701. [PMID: 37571485 PMCID: PMC10422388 DOI: 10.3390/s23156701] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/28/2023] [Revised: 07/19/2023] [Accepted: 07/25/2023] [Indexed: 08/13/2023]
Abstract
The online automated maturity grading and counting of tomato fruits has a certain promoting effect on digital supervision of fruit growth status and unmanned precision operations during the planting process. The traditional grading and counting of tomato fruit maturity is mostly done manually, which is time-consuming and laborious work, and its precision depends on the accuracy of human eye observation. The combination of artificial intelligence and machine vision has to some extent solved this problem. In this work, firstly, a digital camera is used to obtain tomato fruit image datasets, taking into account factors such as occlusion and external light interference. Secondly, based on the tomato maturity grading task requirements, the MHSA attention mechanism is adopted to improve YOLOv8's backbone to enhance the network's ability to extract diverse features. The Precision, Recall, F1-score, and mAP50 of the tomato fruit maturity grading model constructed based on MHSA-YOLOv8 were 0.806, 0.807, 0.806, and 0.864, respectively, which improved the performance of the model with a slight increase in model size. Finally, thanks to the excellent performance of MHSA-YOLOv8, the Precision, Recall, F1-score, and mAP50 of the constructed counting models were 0.990, 0.960, 0.975, and 0.916, respectively. The tomato maturity grading and counting model constructed in this study is not only suitable for online detection but also for offline detection, which greatly helps to improve the harvesting and grading efficiency of tomato growers. The main innovations of this study are summarized as follows: (1) a tomato maturity grading and counting dataset collected from actual production scenarios was constructed; (2) considering the complexity of the environment, this study proposes a new object detection method, MHSA-YOLOv8, and constructs tomato maturity grading models and counting models, respectively; (3) the models constructed in this study are not only suitable for online grading and counting but also for offline grading and counting.
Collapse
Affiliation(s)
| | | | | | | | | | - Lihong Gao
- Chongqing Academy of Agricultural Sciences, Chongqing 401329, China
| |
Collapse
|
28
|
Zhu A, Wang B, Xie J, Ma C. MFF-YOLO: An Accurate Model for Detecting Tunnel Defects Based on Multi-Scale Feature Fusion. Sensors (Basel) 2023; 23:6490. [PMID: 37514784 PMCID: PMC10383211 DOI: 10.3390/s23146490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/09/2023] [Revised: 07/12/2023] [Accepted: 07/13/2023] [Indexed: 07/30/2023]
Abstract
Tunnel linings require routine inspection as they have a big impact on a tunnel's safety and longevity. In this study, the convolutional neural network was utilized to develop the MFF-YOLO model. To improve feature learning efficiency, a multi-scale feature fusion network was constructed within the neck network. Additionally, a reweighted screening method was devised at the prediction stage to address the problem of duplicate detection frames. Moreover, the loss function was adjusted to maximize the effectiveness of model training and improve its overall performance. The results show that the model has a recall and accuracy that are 7.1% and 6.0% greater than those of the YOLOv5 model, reaching 89.5% and 89.4%, respectively, as well as the ability to reliably identify targets that the previous model error detection and miss detection. The MFF-YOLO model improves tunnel lining detection performance generally.
Collapse
Affiliation(s)
- Anfu Zhu
- School of Electronic Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
| | - Bin Wang
- School of Electronic Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
| | - Jiaxiao Xie
- School of Electronic Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
| | - Congxiao Ma
- School of Electronic Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
| |
Collapse
|
29
|
Li R, Deng L, Wang Y, Dai H, Duan R. Target Detection for Synthetic Aperture Radiometer Based on Satellite Formation Flight. Sensors (Basel) 2023; 23:6348. [PMID: 37514642 PMCID: PMC10384500 DOI: 10.3390/s23146348] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/06/2023] [Revised: 06/28/2023] [Accepted: 07/06/2023] [Indexed: 07/30/2023]
Abstract
Synthetic aperture interferometers formed by satellite formations have been adopted to improve spatial resolution. Due to the limited number of satellites and limited integrated time, the use of sparse baselines can result in distorted reconstructed images, which will generate false targets or miss true targets. When detecting a target on the Earth from a geostationary orbit, the target usually occupies only one pixel, and it is almost submerged by noise. Considering the slow-varying characteristics of the observation area, combined with historical observation data and the motion characteristics of the target itself, a target detection method based on multi-frame snapshot images is proposed. Firstly, the observation background is estimated using multi-frame historical data, and background elimination is used to suppress the background noise. Then, potential targets are selected using the local brightness temperature characteristics of the targets. Lastly, the target motion tracks are applied to erase false targets and correct the positions of missed targets. Simulation experiments have been conducted, and the false alarm rate and the missing alarm rate are counted for randomly distributed targets.
Collapse
Affiliation(s)
- Rui Li
- National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China
- School of Astronomy and Space Science, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Li Deng
- National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China
- School of Astronomy and Space Science, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Yuan Wang
- School of Astronautics, Beihang University, Beijing 100191, China
| | - Haoming Dai
- National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China
| | - Ran Duan
- National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China
| |
Collapse
|
30
|
Zhao C, Fu X, Dong J, Feng C, Chang H. LPDNet: A Lightweight Network for SAR Ship Detection Based on Multi-Level Laplacian Denoising. Sensors (Basel) 2023; 23:6084. [PMID: 37447932 DOI: 10.3390/s23136084] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/25/2023] [Revised: 06/17/2023] [Accepted: 06/20/2023] [Indexed: 07/15/2023]
Abstract
Intelligent ship detection based on synthetic aperture radar (SAR) is vital in maritime situational awareness. Deep learning methods have great advantages in SAR ship detection. However, the methods do not strike a balance between lightweight and accuracy. In this article, we propose an end-to-end lightweight SAR target detection algorithm, multi-level Laplacian pyramid denoising network (LPDNet). Firstly, an intelligent denoising method based on the multi-level Laplacian transform is proposed. Through Convolutional Neural Network (CNN)-based threshold suppression, the denoising becomes adaptive to every SAR image via back-propagation and makes the denoising processing supervised. Secondly, channel modeling is proposed to combine the spatial domain and frequency domain information. Multi-dimensional information enhances the detection effect. Thirdly, the Convolutional Block Attention Module (CBAM) is introduced into the feature fusion module of the basic framework (Yolox-tiny) so that different weights are given to each pixel of the feature map to highlight the effective features. Experiments on SSDD and AIR SARShip-1.0 demonstrate that the proposed method achieves 97.14% AP with a speed of 24.68FPS and 92.19% AP with a speed of 23.42FPS, respectively, with only 5.1 M parameters, which verifies the accuracy, efficiency, and lightweight of the proposed method.
Collapse
Affiliation(s)
- Congxia Zhao
- Beijing Institute of Technology, Beijing 100081, China
| | - Xiongjun Fu
- Beijing Institute of Technology, Beijing 100081, China
- Tangshan Research Institute of BIT, Tangshan 063000, China
| | - Jian Dong
- Beijing Institute of Technology, Beijing 100081, China
| | - Cheng Feng
- Beijing Institute of Technology, Beijing 100081, China
| | - Hao Chang
- Beijing Institute of Technology, Beijing 100081, China
| |
Collapse
|
31
|
Guo J, Liu X, Bi L, Liu H, Lou H. UN-YOLOv5s: A UAV-Based Aerial Photography Detection Algorithm. Sensors (Basel) 2023; 23:5907. [PMID: 37447757 DOI: 10.3390/s23135907] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/21/2023] [Revised: 06/19/2023] [Accepted: 06/23/2023] [Indexed: 07/15/2023]
Abstract
With the progress of science and technology, artificial intelligence is widely used in various disciplines and has produced amazing results. The research of the target detection algorithm has significantly improved the performance and role of unmanned aerial vehicles (UAVs), and plays an irreplaceable role in preventing forest fires, evacuating crowded people, surveying and rescuing explorers. At this stage, the target detection algorithm deployed in UAVs has been applied to production and life, but making the detection accuracy higher and better adaptability is still the motivation for researchers to continue to study. In aerial images, due to the high shooting height, small size, low resolution and few features, it is difficult to be detected by conventional target detection algorithms. In this paper, the UN-YOLOv5s algorithm can solve the difficult problem of small target detection excellently. The more accurate small target detection (MASD) mechanism is used to greatly improve the detection accuracy of small and medium targets, The multi-scale feature fusion (MCF) path is combined to fuse the semantic information and location information of the image to improve the expression ability of the novel model. The new convolution SimAM residual (CSR) module is introduced to make the network more stable and focused. On the VisDrone dataset, the mean average precision (mAP) of UAV necessity you only look once v5s(UN-YOLOv5s) is 8.4% higher than that of the original algorithm. Compared with the same version, YOLOv5l, the mAP is increased by 2.2%, and the Giga Floating-point Operations Per Second (GFLOPs) is reduced by 65.3%. Compared with the same series of YOLOv3, the mAP is increased by 1.8%, and GFLOPs is reduced by 75.8%. Compared with the same series of YOLOv8s, the detection accuracy of the mAP is improved by 1.1%.
Collapse
Affiliation(s)
- Junmei Guo
- The School of Information and Automation Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
| | - Xingchen Liu
- The School of Information and Automation Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
| | - Lingyun Bi
- The School of Information and Automation Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
| | - Haiying Liu
- The School of Information and Automation Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
| | - Haitong Lou
- The School of Information and Automation Engineering, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
| |
Collapse
|
32
|
Zhao Z, Wang J, Zhao H. Research on Apple Recognition Algorithm in Complex Orchard Environment Based on Deep Learning. Sensors (Basel) 2023; 23:5425. [PMID: 37420591 DOI: 10.3390/s23125425] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Revised: 05/24/2023] [Accepted: 06/02/2023] [Indexed: 07/09/2023]
Abstract
In the complex environment of orchards, in view of low fruit recognition accuracy, poor real-time and robustness of traditional recognition algorithms, this paper propose an improved fruit recognition algorithm based on deep learning. Firstly, the residual module was assembled with the cross stage parity network (CSP Net) to optimize recognition performance and reduce the computing burden of the network. Secondly, the spatial pyramid pool (SPP) module is integrated into the recognition network of the YOLOv5 to blend the local and global features of the fruit, thus improving the recall rate of the minimum fruit target. Meanwhile, the NMS algorithm was replaced by the Soft NMS algorithm to enhance the ability of identifying overlapped fruits. Finally, a joint loss function was constructed based on focal and CIoU loss to optimize the algorithm, and the recognition accuracy was significantly improved. The test results show that the MAP value of the improved model after dataset training reaches 96.3% in the test set, which is 3.8% higher than the original model. F1 value reaches 91.8%, which is 3.8% higher than the original model. The average detection speed under GPU reaches 27.8 frames/s, which is 5.6 frames/s higher than the original model. Compared with current advanced detection methods such as Faster RCNN and RetinaNet, among others, the test results show that this method has excellent detection accuracy, good robustness and real-time performance, and has important reference value for solving the problem of accurate recognition of fruit in complex environment.
Collapse
Affiliation(s)
- Zhuoqun Zhao
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
- School of Mechanical Engineering, Tianjin Sino-German University of Applied Sciences, Tianjin 300350, China
| | - Jiang Wang
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
| | - Hui Zhao
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
| |
Collapse
|
33
|
Liu P, Yin H. YOLOv7-Peach: An Algorithm for Immature Small Yellow Peaches Detection in Complex Natural Environments. Sensors (Basel) 2023; 23:s23115096. [PMID: 37299824 DOI: 10.3390/s23115096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/17/2023] [Revised: 05/15/2023] [Accepted: 05/23/2023] [Indexed: 06/12/2023]
Abstract
Using object detection techniques on immature fruits to find out their quantity and position is a crucial step for intelligent orchard management. A yellow peach target detection model (YOLOv7-Peach) based on the improved YOLOv7 was proposed to address the problem of immature yellow peach fruits in natural scenes that are similar in color to the leaves but have small sizes and are easily obscured, leading to low detection accuracy. First, the anchor frame information from the original YOLOv7 model was updated by the K-means clustering algorithm in order to generate anchor frame sizes and proportions suitable for the yellow peach dataset; second, the CA (coordinate attention) module was embedded into the backbone network of YOLOv7 so as to enhance the network's feature extraction for yellow peaches and to improve the detection accuracy; then, we accelerated the regression convergence process of the prediction box by replacing the object detection regression loss function with EIoU. Finally, the head structure of YOLOv7 added the P2 module for shallow downsampling, and the P5 module for deep downsampling was removed, effectively improving the detection of small targets. Experiments showed that the YOLOv7-Peach model had a 3.5% improvement in mAp (mean average precision) over the original one, much higher than that of SSD, Objectbox, and other target detection models in the YOLO series, and achieved better results under different weather conditions and a detection speed of up to 21 fps, suitable for real-time detection of yellow peaches. This method could provide technical support for yield estimation in the intelligent management of yellow peach orchards and also provide ideas for the real-time and accurate detection of small fruits with near background colors.
Collapse
Affiliation(s)
- Pingzhu Liu
- School of Computer and Information Engineering, Jiangxi Agricultura University, Nanchang 330045, China
| | - Hua Yin
- School of Software, Jiangxi Agricultura University, Nanchang 330045, China
| |
Collapse
|
34
|
Zhang X, Zhang Z. Research on a Traffic Sign Recognition Method under Small Sample Conditions. Sensors (Basel) 2023; 23:s23115091. [PMID: 37299816 DOI: 10.3390/s23115091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/13/2023] [Revised: 05/20/2023] [Accepted: 05/23/2023] [Indexed: 06/12/2023]
Abstract
Traffic signs are updated quickly, and there image acquisition and labeling work requires a lot of manpower and material resources, so it is difficult to provide a large number of training samples for high-precision recognition. Aiming at this problem, a traffic sign recognition method based on FSOD (few-shot object learning) is proposed. This method adjusts the backbone network of the original model and introduces dropout, which improves the detection accuracy and reduces the risk of overfitting. Secondly, an RPN (region proposal network) with improved attention mechanism is proposed to generate more accurate target candidate boxes by selectively enhancing some features. Finally, the FPN (feature pyramid network) is introduced for multi-scale feature extraction, and the feature map with higher semantic information but lower resolution is merged with the feature map with higher resolution but weaker semantic information, which further improves the detection accuracy. Compared with the baseline model, the improved algorithm improves the 5-way 3-shot and 5-way 5-shot tasks by 4.27% and 1.64%, respectively. We apply the model structure to the PASCAL VOC dataset. The results show that this method is superior to some current few-shot object detection algorithms.
Collapse
Affiliation(s)
- Xiao Zhang
- College of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
| | - Zhenyu Zhang
- Key Laboratory of Multilingual Information Technology in Xinjiang Uygur Autonomous Region, Xinjiang University, Urumqi 830017, China
| |
Collapse
|
35
|
Shen L, Su H, Mao Z, Jing X, Jia C. Signal Property Information-Based Target Detection with Dual-Output Neural Network in Complex Environments. Sensors (Basel) 2023; 23:4956. [PMID: 37430870 DOI: 10.3390/s23104956] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/03/2023] [Revised: 05/11/2023] [Accepted: 05/18/2023] [Indexed: 07/12/2023]
Abstract
The performance of traditional model-based constant false-alarm ratio (CFAR) detection algorithms can suffer in complex environments, particularly in scenarios involving multiple targets (MT) and clutter edges (CE) due to an imprecise estimation of background noise power level. Furthermore, the fixed threshold mechanism that is commonly used in the single-input single-output neural network can result in performance degradation due to changes in the scene. To overcome these challenges and limitations, this paper proposes a novel approach, a single-input dual-output network detector (SIDOND) using data-driven deep neural networks (DNN). One output is used for signal property information (SPI)-based estimation of the detection sufficient statistic, while the other is utilized to establish a dynamic-intelligent threshold mechanism based on the threshold impact factor (TIF), where the TIF is a simplified description of the target and background environment information. Experimental results demonstrate that SIDOND is more robust and performs better than model-based and single-output network detectors. Moreover, the visual explanation technique is employed to explain the working of SIDOND.
Collapse
Affiliation(s)
- Lu Shen
- National Key Laboratory of Radar Signal Processing, Xidian University, Xi'an 710071, China
| | - Hongtao Su
- National Key Laboratory of Radar Signal Processing, Xidian University, Xi'an 710071, China
| | - Zhi Mao
- National Key Laboratory of Radar Signal Processing, Xidian University, Xi'an 710071, China
| | - Xinchen Jing
- National Key Laboratory of Radar Signal Processing, Xidian University, Xi'an 710071, China
| | - Congyue Jia
- National Key Laboratory of Radar Signal Processing, Xidian University, Xi'an 710071, China
| |
Collapse
|
36
|
Saneyoshi A, Takayama R, Michimata C. Tool use moves the peri-personal space from the hand to the tip of the tool. Front Psychol 2023; 14:1142850. [PMID: 37251033 PMCID: PMC10213688 DOI: 10.3389/fpsyg.2023.1142850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2023] [Accepted: 04/13/2023] [Indexed: 05/31/2023] Open
Abstract
In this study, we used a visual target detection task to investigate three hypotheses about how the peri-personal space is extended after tool-use training: Addition, Extension, and Projection hypotheses. We compared the target detection performance before and after tool-use training. In both conditions, the participants held a hockey stick-like tool in their hands during the detection task. Furthermore, we added the no-tool-holding condition to the experimental design. In the no-tool-holding condition, a peri-hand space advantage in the visual target detection task was observed. When the participants held the tool with their hands, this peri-hand space advantage was lost. Furthermore, there was no peri-tool space advantage before tool training. After tool training, the peri-tool space advantage was observed. However, after tool training, the advantage of the peri-hand space was not observed. This result suggested that the peri-hand advantage was reduced by simply holding the tool because the participants lost the functionality of their hands. Furthermore, tool-use training improved detection performance only in the peri-tool space. Thus, these results supported the projection hypothesis that the peri-personal space advantage would move from the body to the functional part of the tool.
Collapse
Affiliation(s)
- Ayako Saneyoshi
- Department of Psychology, Teikyo University, Hachioji-shi, Tokyo, Japan
| | - Ryota Takayama
- Department of Psychology, Sophia University, Chiyoda-ku, Tokyo, Japan
| | - Chikashi Michimata
- Department of Psychology, Teikyo University, Hachioji-shi, Tokyo, Japan
- Department of Psychology, Sophia University, Chiyoda-ku, Tokyo, Japan
| |
Collapse
|
37
|
Zhao W, Wu D, Zheng X. Detection of Chrysanthemums Inflorescence Based on Improved CR-YOLOv5s Algorithm. Sensors (Basel) 2023; 23:s23094234. [PMID: 37177438 PMCID: PMC10181578 DOI: 10.3390/s23094234] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/23/2023] [Revised: 04/17/2023] [Accepted: 04/21/2023] [Indexed: 05/15/2023]
Abstract
Accurate recognition of the flowering stage is a prerequisite for flower yield estimation. In order to improve the recognition accuracy based on the complex image background, such as flowers partially covered by leaves and flowers with insignificant differences in various fluorescence, this paper proposed an improved CR-YOLOv5s to recognize flower buds and blooms for chrysanthemums by emphasizing feature representation through an attention mechanism. The coordinate attention mechanism module has been introduced to the backbone of the YOLOv5s so that the network can pay more attention to chrysanthemum flowers, thereby improving detection accuracy and robustness. Specifically, we replaced the convolution blocks in the backbone network of YOLOv5s with the convolution blocks from the RepVGG block structure to improve the feature representation ability of YOLOv5s through a multi-branch structure, further improving the accuracy and robustness of detection. The results showed that the average accuracy of the improved CR-YOLOv5s was as high as 93.9%, which is 4.5% better than that of normal YOLOv5s. This research provides the basis for the automatic picking and grading of flowers, as well as a decision-making basis for estimating flower yield.
Collapse
Affiliation(s)
- Wentao Zhao
- College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou 311300, China
- Key Laboratory of State Forestry and Grassland Administration on Forestry Sensing Technology and Intelligent Equipment, Hangzhou 311300, China
- Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Hangzhou 311300, China
| | - Dasheng Wu
- College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou 311300, China
- Key Laboratory of State Forestry and Grassland Administration on Forestry Sensing Technology and Intelligent Equipment, Hangzhou 311300, China
- Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Hangzhou 311300, China
| | - Xinyu Zheng
- College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou 311300, China
- Key Laboratory of State Forestry and Grassland Administration on Forestry Sensing Technology and Intelligent Equipment, Hangzhou 311300, China
- Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Hangzhou 311300, China
| |
Collapse
|
38
|
Fan J, Cui L, Fei S. Waste Detection System Based on Data Augmentation and YOLO_EC. Sensors (Basel) 2023; 23:s23073646. [PMID: 37050706 PMCID: PMC10098522 DOI: 10.3390/s23073646] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/27/2023] [Revised: 03/28/2023] [Accepted: 03/29/2023] [Indexed: 06/12/2023]
Abstract
The problem of waste classification has been a major concern for both the government and society, and whether waste can be effectively classified will affect the sustainable development of human society. To perform fast and efficient detection of waste targets in the sorting process, this paper proposes a data augmentation + YOLO_EC waste detection system. First of all, because of the current shortage of multi-objective waste classification datasets, the heavy workload of human data collection, and the limited improvement of data features by traditional data augmentation methods, DCGAN (deep convolution generative adversarial networks) was optimized by improving the loss function, and an image-generation model was established to realize the generation of multi-objective waste images; secondly, with YOLOv4 (You Only Look Once version 4) as the basic model, EfficientNet is used as the backbone feature extraction network to realize the light weight of the algorithm, and at the same time, the CA (coordinate attention) attention mechanism is introduced to reconstruct the MBConv module to filter out high-quality information and enhance the feature extraction ability of the model. Experimental results show that on the HPU_WASTE dataset, the proposed model outperforms other models in both data augmentation and waste detection.
Collapse
Affiliation(s)
- Jinhao Fan
- School of Electrical Engineering and Automation, Henan Polytechnic University, Jiaozuo 454000, China;
- Henan Key Laboratory of Intelligent Detection and Control of Coal Mine Equipment, Henan Polytechnic University, Jiaozuo 454000, China
| | - Lizhi Cui
- School of Electrical Engineering and Automation, Henan Polytechnic University, Jiaozuo 454000, China;
- Henan Key Laboratory of Intelligent Detection and Control of Coal Mine Equipment, Henan Polytechnic University, Jiaozuo 454000, China
| | - Shumin Fei
- School of Automation, Southeast University, Nanjing 210096, China;
| |
Collapse
|
39
|
Song C, Zhu S, Liu Y, Zhang W, Wang Z, Li W, Sun Z, Zhao P, Tian S. DCNAS-Net: deformation convolution and neural architecture search detection network for bone marrow oedema. BMC Med Imaging 2023; 23:45. [PMID: 36978011 PMCID: PMC10045610 DOI: 10.1186/s12880-023-01003-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2022] [Accepted: 03/21/2023] [Indexed: 03/30/2023] Open
Abstract
Background Lumbago is a global disease that affects more than 500 million people worldwide. Bone marrow oedema is one of the main causes of the condition and clinical diagnosis is mainly made by radiologists manually reviewing MRI images to determine whether oedema is present. However, the number of patients with Lumbago has risen dramatically in recent years, which has brought a huge workload to radiologists. In order to improve the efficiency of diagnosis, this paper is devoted to developing and evaluating a neural network for detecting bone marrow edema in MRI images. Related work Inspired by the development of deep learning and image processing techniques, we design a deep learning detection algorithm specifically for the detection of bone marrow oedema from lumbar MRI images. We introduce deformable convolution, feature pyramid networks and neural architecture search modules, and redesign the existing neural networks. We explain in detail the construction of the network and illustrate the setting of the network hyperparameters. Results and discussion The detection accuracy of our algorithm is excellent. And its accuracy of detecting bone marrow oedema reached up to 90.6\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\%$$\end{document}%, an improvement of 5.7\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\%$$\end{document}% compared to the original. The recall of our neural network is 95.1\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\%$$\end{document}%, and the F1-measure also reaches 92.8\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\%$$\end{document}%. And our algorithm is fast in detecting it, taking only 0.144 s per image. Conclusion Extensive experiments have demonstrated that deformable convolution and aggregated feature pyramid structures are conducive for the detection of bone marrow oedema. Our algorithm has better detection accuracy and good detection speed compared to other algorithms.
Collapse
Affiliation(s)
- Chengyu Song
- grid.33763.320000 0004 1761 2484Tianjin University, Tianjin, China
| | - Shan Zhu
- grid.33763.320000 0004 1761 2484Tianjin Hospital, Tianjin University, Tianjin, China
| | - Yanyan Liu
- grid.216938.70000 0000 9878 7032Nankai University, Tianjin, China
| | - Wei Zhang
- grid.33763.320000 0004 1761 2484Tianjin University, Tianjin, China
| | - Zhi Wang
- grid.33763.320000 0004 1761 2484Tianjin Hospital, Tianjin University, Tianjin, China
| | - Wangxiao Li
- grid.33763.320000 0004 1761 2484Tianjin University, Tianjin, China
| | - Zhenye Sun
- grid.33763.320000 0004 1761 2484Tianjin Hospital, Tianjin University, Tianjin, China
| | - Peng Zhao
- grid.33763.320000 0004 1761 2484Tianjin Hospital, Tianjin University, Tianjin, China
| | - Shengzhang Tian
- grid.33763.320000 0004 1761 2484Tianjin Hospital, Tianjin University, Tianjin, China
| |
Collapse
|
40
|
Ma TJ, Anderson RJ. Remote Sensing Low Signal-to-Noise-Ratio Target Detection Enhancement. Sensors (Basel) 2023; 23:3314. [PMID: 36992025 PMCID: PMC10054736 DOI: 10.3390/s23063314] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/22/2023] [Revised: 03/10/2023] [Accepted: 03/20/2023] [Indexed: 06/19/2023]
Abstract
In real-time remote sensing application, frames of data are continuously flowing into the processing system. The capability of detecting objects of interest and tracking them as they move is crucial to many critical surveillance and monitoring missions. Detecting small objects using remote sensors is an ongoing, challenging problem. Since object(s) are located far away from the sensor, the target's Signal-to-Noise-Ratio (SNR) is low. The Limit of Detection (LOD) for remote sensors is bounded by what is observable on each image frame. In this paper, we present a new method, a "Multi-frame Moving Object Detection System (MMODS)", to detect small, low SNR objects that are beyond what a human can observe in a single video frame. This is demonstrated by using simulated data where our technology-detected objects are as small as one pixel with a targeted SNR, close to 1:1. We also demonstrate a similar improvement using live data collected with a remote camera. The MMODS technology fills a major technology gap in remote sensing surveillance applications for small target detection. Our method does not require prior knowledge about the environment, pre-labeled targets, or training data to effectively detect and track slow- and fast-moving targets, regardless of the size or the distance.
Collapse
|
41
|
Chen G, Wu T, Duan J, Hu Q, Huang D, Li H. CenterPNets: A Multi-Task Shared Network for Traffic Perception. Sensors (Basel) 2023; 23:2467. [PMID: 36904671 PMCID: PMC10007440 DOI: 10.3390/s23052467] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/17/2023] [Revised: 02/20/2023] [Accepted: 02/20/2023] [Indexed: 06/18/2023]
Abstract
The importance of panoramic traffic perception tasks in autonomous driving is increasing, so shared networks with high accuracy are becoming increasingly important. In this paper, we propose a multi-task shared sensing network, called CenterPNets, that can perform the three major detection tasks of target detection, driving area segmentation, and lane detection in traffic sensing in one go and propose several key optimizations to improve the overall detection performance. First, this paper proposes an efficient detection head and segmentation head based on a shared path aggregation network to improve the overall reuse rate of CenterPNets and an efficient multi-task joint training loss function to optimize the model. Secondly, the detection head branch uses an anchor-free frame mechanism to automatically regress target location information to improve the inference speed of the model. Finally, the split-head branch fuses deep multi-scale features with shallow fine-grained features, ensuring that the extracted features are rich in detail. CenterPNets achieves an average detection accuracy of 75.8% on the publicly available large-scale Berkeley DeepDrive dataset, with an intersection ratio of 92.8% and 32.1% for driveableareas and lane areas, respectively. Therefore, CenterPNets is a precise and effective solution to the multi-tasking detection issue.
Collapse
Affiliation(s)
- Guangqiu Chen
- College of Electronic Information Engineering, Chang Chun University of Science and Technology, Changchun 130022, China
| | - Tao Wu
- College of Electronic Information Engineering, Chang Chun University of Science and Technology, Changchun 130022, China
| | - Jin Duan
- College of Electronic Information Engineering, Chang Chun University of Science and Technology, Changchun 130022, China
| | - Qi Hu
- College of Artificial Intelligence, Chang Chun University of Science and Technology, Changchun 130022, China
| | - Dandan Huang
- College of Electronic Information Engineering, Chang Chun University of Science and Technology, Changchun 130022, China
| | - Hao Li
- College of Electronic Information Engineering, Chang Chun University of Science and Technology, Changchun 130022, China
| |
Collapse
|
42
|
Huang J, Zhang G. A Study of an Online Tracking System for Spark Images of Abrasive Belt-Polishing Workpieces. Sensors (Basel) 2023; 23:2025. [PMID: 36850622 PMCID: PMC9966948 DOI: 10.3390/s23042025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/17/2022] [Revised: 02/05/2023] [Accepted: 02/06/2023] [Indexed: 06/18/2023]
Abstract
During the manual grinding of blades, the workers can estimate the material removal rate based on their experiences from observing the characteristics of the grinding sparks, leading to low grinding accuracy and low efficiency and affecting the processing quality of the blades. As an alternative to the recognition of spark images by the human eye, we used the deep learning algorithm YOLO5 to perform target detection on spark images and obtain spark image regions. First the spark images generated during one turbine blade-grinding process were collected, and some of the images were selected as training samples, with the remaining images used as test samples, which were labelled with LabelImg. Afterwards, the selected images were trained with YOLO5 to obtain an optimisation model. In the end, the trained optimisation model was used to predict the images of the test set. The proposed method was able to detect spark image regions quickly and accurately, with an average accuracy of 0.995. YOLO4 was also used to train and predict spark images, and the two methods were compared. Our findings show that YOLO5 is faster and more accurate than the YOLO4 target detection algorithm and can replace manual observation, laying a specific foundation for the automatic segmentation of spark images and the study of the relationship between the material removal rate and spark images at a later stage, which has some practical value.
Collapse
Affiliation(s)
- Jian Huang
- School of Mechanical and Precision Instrument Engineering, Xi’an University of Technology, Xi’an 710048, China
- School of Computer Science, Xijing University, Xi’an 710123, China
| | - Guangpeng Zhang
- School of Mechanical and Precision Instrument Engineering, Xi’an University of Technology, Xi’an 710048, China
| |
Collapse
|
43
|
Zou J, Zheng H, Wang F. Real-Time Target Detection System for Intelligent Vehicles Based on Multi-Source Data Fusion. Sensors (Basel) 2023; 23:1823. [PMID: 36850421 PMCID: PMC9962490 DOI: 10.3390/s23041823] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Revised: 01/28/2023] [Accepted: 02/01/2023] [Indexed: 06/18/2023]
Abstract
To improve the identification accuracy of target detection for intelligent vehicles, a real-time target detection system based on the multi-source fusion method is proposed. Based on the ROS melodic software development environment and the NVIDIA Xavier hardware development platform, this system integrates sensing devices such as millimeter-wave radar and camera, and it can realize functions such as real-time target detection and tracking. At first, the image data can be processed by the You Only Look Once v5 network, which can increase the speed and accuracy of identification; secondly, the millimeter-wave radar data are processed to provide a more accurate distance and velocity of the targets. Meanwhile, in order to improve the accuracy of the system, the sensor fusion method is used. The radar point cloud is projected onto the image, then through space-time synchronization, region of interest (ROI) identification, and data association, the target-tracking information is presented. At last, field tests of the system are conducted, the results of which indicate that the system has a more accurate recognition effect and scene adaptation ability in complex scenes.
Collapse
|
44
|
Wang L, Xiao Y, Zhang B, Liu R, Zhao B. Water Surface Targets Detection Based on the Fusion of Vision and LiDAR. Sensors (Basel) 2023; 23:1768. [PMID: 36850373 PMCID: PMC9967045 DOI: 10.3390/s23041768] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Revised: 01/28/2023] [Accepted: 02/03/2023] [Indexed: 06/18/2023]
Abstract
The use of vision for the recognition of water targets is easily influenced by reflections and ripples, resulting in misidentification. This paper proposed a detection method based on the fusion of 3D point clouds and visual information to detect and locate water surface targets. The point clouds help to reduce the impact of ripples and reflections, and the recognition accuracy is enhanced by visual information. This method consists of three steps: Firstly, the water surface target is detected using the CornerNet-Lite network, and then the candidate target box and camera detection confidence are determined. Secondly, the 3D point cloud is projected onto the two-dimensional pixel plane, and the confidence of LiDAR detection is calculated based on the ratio between the projected area of the point clouds and the pixel area of the bounding box. The target confidence is calculated with the camera detection and LiDAR detection confidence, and the water surface target is determined by combining the detection thresholds. Finally, the bounding box is used to determine the 3D point clouds of the target and estimate its 3D coordinates. The experiment results showed this method reduced the misidentification rate and had 15.5% higher accuracy compared with traditional CornerNet-Lite network. By combining the depth information from LiDAR, the position of the target relative to the detection coordinate system origin could be accurately estimated.
Collapse
Affiliation(s)
- Lin Wang
- School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
- Laboratory of Science and Technology on Marine Navigation and Control, China State Shipbuilding Corporation, Tianjin 300131, China
| | - Yufeng Xiao
- School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
- Laboratory of Science and Technology on Marine Navigation and Control, China State Shipbuilding Corporation, Tianjin 300131, China
| | - Baorui Zhang
- School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
| | - Ran Liu
- School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
- Laboratory of Science and Technology on Marine Navigation and Control, China State Shipbuilding Corporation, Tianjin 300131, China
- Engineering Product Development Pillar, Singapore University of Technology and Design, Singapore 487372, Singapore
| | - Bin Zhao
- Tianjin Navigation Instrument Research Institute, Tianjin 300131, China
| |
Collapse
|
45
|
Yu L, Guo J, Pu Y, Cen H, Li J, Liu S, Nie J, Ge J, Yang S, Zhao H, Xu Y, Wu J, Wang K. A Recognition Method of Ewe Estrus Crawling Behavior Based on Multi- Target Detection Layer Neural Network. Animals (Basel) 2023; 13:ani13030413. [PMID: 36766301 PMCID: PMC9913191 DOI: 10.3390/ani13030413] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2022] [Revised: 01/13/2023] [Accepted: 01/23/2023] [Indexed: 01/28/2023] Open
Abstract
There are some problems with estrus detection in ewes in large-scale meat sheep farming: mainly, the manual detection method is labor-intensive and the contact sensor detection method causes stress reactions in ewes. To solve the abovementioned problems, we proposed a multi-objective detection layer neural network-based method for ewe estrus crawling behavior recognition. The approach we proposed has four main parts. Firstly, to address the problem of mismatch between our constructed ewe estrus dataset and the YOLO v3 anchor box size, we propose to obtain a new anchor box size by clustering the ewe estrus dataset using the K-means++ algorithm. Secondly, to address the problem of low model recognition precision caused by small imaging of distant ewes in the dataset, we added a 104 × 104 target detection layer, making the total target detection layer reach four layers, strengthening the model's ability to learn shallow information and improving the model's ability to detect small targets. Then, we added residual units to the residual structure of the model, so that the deep feature information of the model is not easily lost and further fused with the shallow feature information to speed up the training of the model. Finally, we maintain the aspect ratio of the images in the data-loading module of the model to reduce the distortion of the image information and increase the precision of the model. The experimental results show that our proposed model has 98.56% recognition precision, while recall was 98.04%, F1 value was 98%, mAP was 99.78%, FPS was 41 f/s, and model size was 276 M, which can meet the accurate and real-time recognition of ewe estrus behavior in large-scale meat sheep farming.
Collapse
Affiliation(s)
- Longhui Yu
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
- College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, China
| | - Jianjun Guo
- College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, China
| | - Yuhai Pu
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| | - Honglei Cen
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| | - Jingbin Li
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
- Correspondence: (J.L.); (S.L.)
| | - Shuangyin Liu
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, China
- Correspondence: (J.L.); (S.L.)
| | - Jing Nie
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| | - Jianbing Ge
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| | - Shuo Yang
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| | - Hangxing Zhao
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| | - Yalei Xu
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
- College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, China
| | - Jianglin Wu
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
- College of Information Science and Technology, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, China
| | - Kang Wang
- College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China
- Xinjiang Production and Construction Corps Key Laboratory of Modern Agricultural Machinery, Shihezi 832003, China
- Industrial Technology Research Institute of Xinjiang Production and Construction Corps, Shihezi 832000, China
| |
Collapse
|
46
|
Sim Y, Heo J, Jung Y, Lee S, Jung Y. FPGA Implementation of Efficient CFAR Algorithm for Radar Systems. Sensors (Basel) 2023; 23:954. [PMID: 36679752 PMCID: PMC9861839 DOI: 10.3390/s23020954] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/30/2022] [Revised: 01/05/2023] [Accepted: 01/12/2023] [Indexed: 06/17/2023]
Abstract
The constant false-alarm rate (CFAR) algorithm is essential for detecting targets during radar signal processing. It has been improved to accurately detect targets, especially in nonhomogeneous environments, such as multitarget or clutter edge environments. For example, there are sort-based and variable index-based algorithms. However, these algorithms require large amounts of computation, making them difficult to apply in radar applications that require real-time target detection. We propose a new CFAR algorithm that determines the environment of a received signal through a new decision criterion and applies the optimal CFAR algorithms such as the modified variable index (MVI) and automatic censored cell averaging-based ordered data variability (ACCA-ODV). The Monte Carlo simulation results of the proposed CFAR algorithm showed a high detection probability of 93.8% in homogeneous and nonhomogeneous environments based on an SNR of 25 dB. In addition, this paper presents the hardware design, field-programmable gate array (FPGA)-based implementation, and verification results for the practical application of the proposed algorithm. We reduced the hardware complexity by time-sharing sum and square operations and by replacing division operations with multiplication operations when calculating decision parameters. We also developed a low-complexity and high-speed sorter architecture that performs sorting for the partial data in leading and lagging windows. As a result, the implementation used 8260 LUTs and 3823 registers and took 0.6 μs to operate. Compared with the previously proposed FPGA implementation results, it is confirmed that the complexity and operation speed of the proposed CFAR processor are very suitable for real-time implementation.
Collapse
Affiliation(s)
- Yunseong Sim
- School of Electronics and Information Engineering, Korea Aerospace University, Goyang-si 10540, Republic of Korea
| | - Jinmoo Heo
- Department of Smart Air Mobility, Korea Aerospace University, Goyang-si 10540, Republic of Korea
| | - Yongchul Jung
- Korea Electronics Technology Institute (KETI), Bundang, Seongnam 13509, Republic of Korea
| | - Seongjoo Lee
- Department of Information and Communication Engineering, Sejong University, Seoul 05006, Republic of Korea
- Department of Convergence Engineering of Intelligent Drone, Sejong University, Seoul 05006, Republic of Korea
| | - Yunho Jung
- School of Electronics and Information Engineering, Korea Aerospace University, Goyang-si 10540, Republic of Korea
- Department of Smart Air Mobility, Korea Aerospace University, Goyang-si 10540, Republic of Korea
| |
Collapse
|
47
|
Huang F, Lin G, Meng Y, Lin Y, Zheng S. The role of alerting in the attentional boost effect. Front Psychol 2023; 14:1075979. [PMID: 37089742 PMCID: PMC10117126 DOI: 10.3389/fpsyg.2023.1075979] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Accepted: 03/02/2023] [Indexed: 04/25/2023] Open
Abstract
Stimuli presented simultaneously with behaviorally relevant events (e.g., targets) are better memorized, an unusual effect defined as the attentional boost effect (ABE). We hypothesized that all types of behaviorally relevant events, including attentional cues, can promote the encoding process for the stimuli paired with them, and the attentional alerting network can amplify the ABE. The two experiments we conducted demonstrated that not all behaviorally relevant events, including alerting cues, benefit the processing of concurrently paired stimuli. We also found that the presence of a cue prior to a target can extend the memory advantage produced by target detection, but this advantage can only be observed within a limited range of time. Overall, our study provides the first evidence that the alerting network plays an important role in the ABE.
Collapse
Affiliation(s)
- Fajie Huang
- School of Health, Fujian Medical University, Fuzhou, China
- School of Psychology, Fujian Normal University, Fuzhou, China
| | - Guyang Lin
- School of Psychology, Fujian Normal University, Fuzhou, China
| | - Yingfang Meng
- School of Psychology, Fujian Normal University, Fuzhou, China
- *Correspondence: Yingfang Meng,
| | - Yuanyuan Lin
- Education Research Institution of Fujian Province, Fuzhou, China
| | - Siqi Zheng
- School of Psychology, Fujian Normal University, Fuzhou, China
| |
Collapse
|
48
|
Li H, Zheng L, Meng F. Low-Altitude Windshear Estimation Method Based on Four-Dimensional Frequency Domain Compensation for Fuselage Frustum Conformal Array. Sensors (Basel) 2022; 23:371. [PMID: 36616969 PMCID: PMC9824326 DOI: 10.3390/s23010371] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Revised: 12/05/2022] [Accepted: 12/12/2022] [Indexed: 06/17/2023]
Abstract
In this paper, a low-altitude wind speed estimation method based on the fuselage frustum conformal array system is proposed. Firstly, based on the signal model of the fuselage conformal array radar, the four-dimensional joint phase compensation of the echo data in the Doppler domain and three-dimensional space-frequency domain is performed by using the four-dimensional frequency domain compensation method. Secondly, the clutter covariance matrix is estimated by the compensated echo data, and a space-time Adaptive Processing (STAP) processor suitable for low-altitude windshear target is constructed to suppress clutter. Finally, the maximum Doppler value of each distance cell is extracted, and the wind velocity is estimated. Simulation results show that the proposed method can effectively suppress clutter and accurately estimate wind speed.
Collapse
Affiliation(s)
- Hai Li
- Tianjin Key Lab for Advanced Signal Processing, Civil Aviation University of China, Tianjin 300300, China
| | - Lei Zheng
- Tianjin Key Lab for Advanced Signal Processing, Civil Aviation University of China, Tianjin 300300, China
| | - Fanwang Meng
- AVIC Lei Hua Electronic Technology Research Institute, Wuxi 214063, China
| |
Collapse
|
49
|
Lv P, Wang B, Cheng F, Xue J. Multi-Objective Association Detection of Farmland Obstacles Based on Information Fusion of Millimeter Wave Radar and Camera. Sensors (Basel) 2022; 23:230. [PMID: 36616828 PMCID: PMC9824033 DOI: 10.3390/s23010230] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/17/2022] [Revised: 12/16/2022] [Accepted: 12/21/2022] [Indexed: 06/17/2023]
Abstract
In order to remedy the defects of single sensor in robustness, accuracy, and redundancy of target detection, this paper proposed a method for detecting obstacles in farmland based on the information fusion of a millimeter wave (mmWave) radar and a camera. Combining the advantages of the mmWave radar in range and speed measurement and the camera in type identification and lateral localization, a decision-level fusion algorithm was designed for the mmWave radar and camera information, and the global nearest neighbor method was used for data association. Then, the effective target sequences of the mmWave radar and the camera with successful data association were weighted to output, and the output included more accurate target orientation, longitudinal speed, and category. For the unassociated sequences, they were tracked as new targets by using the extended Kalman filter algorithm and were processed and output during the effective life cycle. Lastly, an experimental platform based on a tractor was built to verify the effectiveness of the proposed association detection method. The obstacle detection test was conducted under the ROS environment after solving the external parameters of the mmWave radar and the internal and external parameters of the camera. The test results show that the correct detection rate of obstacles reaches 86.18%, which is higher than that of a single camera with 62.47%. Furthermore, through the contrast experiment of the sensor fusion algorithms, the detection accuracy of the decision level fusion algorithm was 95.19%, which was higher than 4.38% and 6.63% compared with feature level and data level fusion, respectively.
Collapse
Affiliation(s)
- Pengfei Lv
- College of Engineering, Nanjing Agricultural University, Nanjing 210031, China
| | - Bingqing Wang
- Jiangsu Agricultural Machinery Information Center, Nanjing 210031, China
| | - Feng Cheng
- College of Engineering, Nanjing Agricultural University, Nanjing 210031, China
| | - Jinlin Xue
- College of Engineering, Nanjing Agricultural University, Nanjing 210031, China
| |
Collapse
|
50
|
Wang Y, Wei C, Sun H, Qu A. Design of Intelligent Detection Platform for Wine Grape Pests and Diseases in Ningxia. Plants (Basel) 2022; 12:106. [PMID: 36616237 PMCID: PMC9823901 DOI: 10.3390/plants12010106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/06/2022] [Revised: 12/22/2022] [Accepted: 12/22/2022] [Indexed: 06/17/2023]
Abstract
In order to reduce the impact of pests and diseases on the yield and quality of Ningxia wine grapes and to improve the efficiency and intelligence of detection, this paper designs an intelligent detection platform for pests and diseases. The optimal underlying network is selected by comparing the recognition accuracy of both MobileNet V2 and YOLOX_s networks trained on the Public Dataset. Based on this network, the effect of adding attention mechanism and replacing loss function on recognition effect is investigated by permutation in the Custom Dataset, resulting in the improved network YOLOX_s + CBAM. The improved network was trained on the Overall Dataset, and finally a recognition model capable of identifying nine types of pests was obtained, with a recognition accuracy of 93.35% in the validation set, an improvement of 1.35% over the original network. The recognition model is deployed on the Web side and Raspberry Pi to achieve independent detection functions; the channel between the two platforms is built through Ngrok, and remote interconnection is achieved through VNC desktop. Users can choose to upload local images on the Web side for detection, handheld Raspberry Pi for field detection, or Raspberry Pi and Web interconnection for remote detection.
Collapse
Affiliation(s)
| | | | | | - Aili Qu
- Correspondence: ; Tel.: +86-199-9538-6860 or +86-0951-2062908
| |
Collapse
|