1
|
Chen B, Zhu L, Kong C, Zhu H, Wang S, Li Z. No-Reference Image Quality Assessment by Hallucinating Pristine Features. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:6139-6151. [PMID: 36112560 DOI: 10.1109/tip.2022.3205770] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
In this paper, we propose a no-reference (NR) image quality assessment (IQA) method via feature level pseudo-reference (PR) hallucination. The proposed quality assessment framework is rooted in the view that the perceptually meaningful features could be well exploited to characterize the visual quality, and the natural image statistical behaviors are exploited in an effort to deliver the accurate predictions. Herein, the PR features from the distorted images are learned by a mutual learning scheme with the pristine reference as the supervision, and the discriminative characteristics of PR features are further ensured with the triplet constraints. Given a distorted image for quality inference, the feature level disentanglement is performed with an invertible neural layer for final quality prediction, leading to the PR and the corresponding distortion features for comparison. The effectiveness of our proposed method is demonstrated on four popular IQA databases, and superior performance on cross-database evaluation also reveals the high generalization capability of our method. The implementation of our method is publicly available on https://github.com/Baoliang93/FPR.
Collapse
|
2
|
Yin Y. Research on the Application of Animation Design Based on Machine Learning and Dynamic Image Index. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:2690415. [PMID: 35814580 PMCID: PMC9262498 DOI: 10.1155/2022/2690415] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/30/2022] [Revised: 06/05/2022] [Accepted: 06/16/2022] [Indexed: 11/18/2022]
Abstract
With the development of computer technology, animation is more and more used because of its simple, effective, and higher performance. Machine learning has become the core of artificial intelligence at present. Intelligent learning algorithms are widely used in practical problems such as evaluation. Knowledge-based automatic animation production system faces two challenges: (1) lack of learning ability and waste of data on the website; (2) the quality of animation produced that depends on the level of system designer and the inability of system users to participate in animation production.In order to solve these two problems, an active animation learning system enables the animation system to constantly learn experience and produce the most popular animation, for the first time, for animation production system design and implementation of applied research. Image retrieval technology is a research center in the field of image application. It is widely used in many fields, such as electronic commerce. Animation design will use dynamic image and machine learning to innovate.
Collapse
Affiliation(s)
- Yaodong Yin
- Department of Fine Arts, Taiyuan Normal University, Jinzhong, China
| |
Collapse
|
3
|
Tang T, Li L, Wu X, Chen R, Li H, Lu G, Cheng L. TSA-SCC: Text Semantic-Aware Screen Content Coding With Ultra Low Bitrate. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:2463-2477. [PMID: 35196232 DOI: 10.1109/tip.2022.3152003] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Due to the rapid growth of web conferences, remote screen sharing, and online games, screen content has become an important type of internet media information and over 90% of online media interactions are screen based. Meanwhile, as the main component in the screen content, textual information averagely takes up over 40% of the whole image on various commonly used screen content datasets. However, it is difficult to compress the textual information by using the traditional coding schemes as HEVC, which assumes strong spatial and temporal correlations within the image/video. State-of-the-art screen content coding (SCC) standard as HEVC-SCC still adopts a block-based coding framework and does not consider the text semantics for compression, thus inevitably blurring texts at a lower bitrate. In this paper, we propose a general text semantic-aware screen content coding scheme (TSA-SCC) for ultra low bitrate setting. This method detects the abrupt picture in a screen content video (or image), recognizes textual information (including word, position, font type, font size and font color) in the abrupt picture based on neural networks, and encodes texts with text coding tools. The other pictures as well as the background image after removing texts from the abrupt picture via inpainting, are encoded with HEVC-SCC. Compared with HEVC-SCC, the proposed method TSA-SCC reduces bitrate by up to 3× at a similar compression quality. Moreover, TSA-SCC achieves much better visual quality with less bitrate consumption when encoding the screen content video/image at ultra low bitrates.
Collapse
|
4
|
Yang J, Guo X, Li Y, Marinello F, Ercisli S, Zhang Z. A survey of few-shot learning in smart agriculture: developments, applications, and challenges. PLANT METHODS 2022; 18:28. [PMID: 35248105 PMCID: PMC8897954 DOI: 10.1186/s13007-022-00866-2] [Citation(s) in RCA: 24] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Accepted: 03/01/2022] [Indexed: 05/08/2023]
Abstract
With the rise of artificial intelligence, deep learning is gradually applied to the field of agriculture and plant science. However, the excellent performance of deep learning needs to be established on massive numbers of samples. In the field of plant science and biology, it is not easy to obtain a large amount of labeled data. The emergence of few-shot learning solves this problem. It imitates the ability of humans' rapid learning and can learn a new task with only a small number of labeled samples, which greatly reduces the time cost and financial resources. At present, the advanced few-shot learning methods are mainly divided into four categories based on: data augmentation, metric learning, external memory, and parameter optimization, solving the over-fitting problem from different viewpoints. This review comprehensively expounds on few-shot learning in smart agriculture, introduces the definition of few-shot learning, four kinds of learning methods, the publicly available datasets for few-shot learning, various applications in smart agriculture, and the challenges in smart agriculture in future development.
Collapse
Affiliation(s)
- Jiachen Yang
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Xiaolan Guo
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Yang Li
- College of Mechanical and Electrical Engineering, Shihezi University, Xinjiang, China.
| | - Francesco Marinello
- Department of Land Environment Agriculture and Forestry, University of Padova, Legnaro, Italy
| | - Sezai Ercisli
- Department of Horticulture, Faculty of Agriculture, Ataturk University, Erzurum, Turkey
| | - Zhuo Zhang
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| |
Collapse
|
5
|
GAN-Based ROI Image Translation Method for Predicting Image after Hair Transplant Surgery. ELECTRONICS 2021. [DOI: 10.3390/electronics10243066] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
In this paper, we propose a new deep learning-based image translation method to predict and generate images after hair transplant surgery from images before hair transplant surgery. Since existing image translation models use a naive strategy that trains the whole distribution of translation, the image translation models using the original image as the input data result in converting not only the hair transplant surgery region, which is the region of interest (ROI) for image translation, but also the other image regions, which are not the ROI. To solve this problem, we proposed a novel generative adversarial network (GAN)-based ROI image translation method, which converts only the ROI and retains the image for the non-ROI. Specifically, by performing image translation and image segmentation independently, the proposed method generates predictive images from the distribution of images after hair transplant surgery and specifies the ROI to be used for generated images. In addition, by applying the ensemble method to image segmentation, we propose a more robust method through complementing the shortages of various image segmentation models. From the experimental results using a real medical image dataset, e.g., 1394 images before hair transplantation and 896 images after hair transplantation, to train the GAN model, we show that the proposed GAN-based ROI image translation method performed better than the other GAN-based image translation methods, e.g., by 23% in SSIM (Structural Similarity Index Measure), 452% in IoU (Intersection over Union), and 42% in FID (Frechet Inception Distance), on average. Furthermore, the ensemble method that we propose not only improves ROI detection performance but also shows consistent performances in generating better predictive images from preoperative images taken from diverse angles.
Collapse
|
6
|
Wu L, Zhang X, Chen H, Wang D, Deng J. VP-NIQE: An opinion-unaware visual perception natural image quality evaluator. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.08.048] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
7
|
Courtney J. SEDIQA: Sound Emitting Document Image Quality Assessment in a Reading Aid for the Visually Impaired. J Imaging 2021; 7:jimaging7090168. [PMID: 34460804 PMCID: PMC8470036 DOI: 10.3390/jimaging7090168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2021] [Revised: 07/29/2021] [Accepted: 08/24/2021] [Indexed: 11/17/2022] Open
Abstract
For visually impaired people (VIPs), the ability to convert text to sound can mean a new level of independence or the simple joy of a good book. With significant advances in optical character recognition (OCR) in recent years, a number of reading aids are appearing on the market. These reading aids convert images captured by a camera to text which can then be read aloud. However, all of these reading aids suffer from a key issue—the user must be able to visually target the text and capture an image of sufficient quality for the OCR algorithm to function—no small task for VIPs. In this work, a sound-emitting document image quality assessment metric (SEDIQA) is proposed which allows the user to hear the quality of the text image and automatically captures the best image for OCR accuracy. This work also includes testing of OCR performance against image degradations, to identify the most significant contributors to accuracy reduction. The proposed no-reference image quality assessor (NR-IQA) is validated alongside established NR-IQAs and this work includes insights into the performance of these NR-IQAs on document images. SEDIQA is found to consistently select the best image for OCR accuracy. The full system includes a document image enhancement technique which introduces improvements in OCR accuracy with an average increase of 22% and a maximum increase of 68%.
Collapse
Affiliation(s)
- Jane Courtney
- School of Electrical & Electronic Engineering, Technological University Dublin, City Campus, Dublin, Ireland
| |
Collapse
|
8
|
RAVA: Region-Based Average Video Quality Assessment. SENSORS 2021; 21:s21165489. [PMID: 34450931 PMCID: PMC8401697 DOI: 10.3390/s21165489] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/28/2021] [Revised: 08/06/2021] [Accepted: 08/11/2021] [Indexed: 12/04/2022]
Abstract
Video has become the most popular medium of communication over the past decade, with nearly 90 percent of the bandwidth on the Internet being used for video transmission. Thus, evaluating the quality of an acquired or compressed video has become increasingly important. The goal of video quality assessment (VQA) is to measure the quality of a video clip as perceived by a human observer. Since manually rating every video clip to evaluate quality is infeasible, researchers have attempted to develop various quantitative metrics that estimate the perceptual quality of video. In this paper, we propose a new region-based average video quality assessment (RAVA) technique extending image quality assessment (IQA) metrics. In our experiments, we extend two full-reference (FR) image quality metrics to measure the feasibility of the proposed RAVA technique. Results on three different datasets show that our RAVA method is practical in predicting objective video scores.
Collapse
|
9
|
Yang Y, Zhang Z, Mao W, Li Y, Lv C. Radar target recognition based on few-shot learning. MULTIMEDIA SYSTEMS 2021. [PMID: 0 DOI: 10.1007/s00530-021-00832-3] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/23/2021] [Accepted: 07/09/2021] [Indexed: 05/24/2023]
|
10
|
Chao X, Zhang L. Few-shot imbalanced classification based on data augmentation. MULTIMEDIA SYSTEMS 2021. [PMID: 0 DOI: 10.1007/s00530-021-00827-0] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Accepted: 06/22/2021] [Indexed: 05/26/2023]
|
11
|
Li Y, Chao X. Semi-supervised few-shot learning approach for plant diseases recognition. PLANT METHODS 2021; 17:68. [PMID: 34176505 PMCID: PMC8237441 DOI: 10.1186/s13007-021-00770-1] [Citation(s) in RCA: 28] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/07/2021] [Accepted: 06/19/2021] [Indexed: 05/21/2023]
Abstract
BACKGROUND Learning from a few samples to automatically recognize the plant leaf diseases is an attractive and promising study to protect the agricultural yield and quality. The existing few-shot classification studies in agriculture are mainly based on supervised learning schemes, ignoring unlabeled data's helpful information. METHODS In this paper, we proposed a semi-supervised few-shot learning approach to solve the plant leaf diseases recognition. Specifically, the public PlantVillage dataset is used and split into the source domain and target domain. Extensive comparison experiments considering the domain split and few-shot parameters (N-way, k-shot) were carried out to validate the correctness and generalization of proposed semi-supervised few-shot methods. In terms of selecting pseudo-labeled samples in the semi-supervised process, we adopted the confidence interval to determine the number of unlabeled samples for pseudo-labelling adaptively. RESULTS The average improvement by the single semi-supervised method is 2.8%, and that by the iterative semi-supervised method is 4.6%. CONCLUSIONS The proposed methods can outperform other related works with fewer labeled training data.
Collapse
Affiliation(s)
- Yang Li
- College of Mechanical and Electrical Engineering, Shihezi University, Xinjiang, China
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Xuewei Chao
- College of Mechanical and Electrical Engineering, Shihezi University, Xinjiang, China.
| |
Collapse
|
12
|
Wang Z, Qu Q, Cai K, Xu T. CT Image Examination Based on Virtual Reality Analysis in Clinical Diagnosis of Gastrointestinal Stromal Tumors. JOURNAL OF HEALTHCARE ENGINEERING 2021; 2021:9996565. [PMID: 34221304 PMCID: PMC8225451 DOI: 10.1155/2021/9996565] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/20/2021] [Revised: 05/18/2021] [Accepted: 06/03/2021] [Indexed: 01/03/2023]
Abstract
With the advancement and development of medical equipment, CT images have become a common lung examination tool. This article mainly studies the application of CT imaging examination based on virtual reality analysis in the clinical diagnosis of gastrointestinal stromal tumors. Before extracting suspected lymph nodes from a CT image of the stomach, the CT image sequence is preprocessed first, which can reduce the cumbersomeness of subsequent extraction of suspected lymph nodes and speed up the subsequent processing. According to medical knowledge, CT images of the stomach show that lymph nodes mainly exist in the adipose tissue around the gastric wall, but there are no lymph nodes in the subcutaneous fat outside the chest. The most basic gray value in the image and the neighborhood average difference feature related to gray level are used as the primary features of visual attention detection. When extracting the neighborhood average difference feature, we use a 3 ∗ 3 sliding window method to traverse each point of the pixel matrix in the image, thereby calculating the feature value of each pixel in the image. After the feature extraction is completed, it is necessary to calibrate the data and make a training data set. The SP immunohistochemical staining method was used. The specimens were fixed with 10% formaldehyde, routinely embedded in paraffin, sectioned, and stained with HE. The tumor tissue was determined by immunohistochemistry, and the reagents were products of Maixin Company. All patients were followed up by regular outpatient review, letters, and visits or phone calls. The data showed that immunohistochemical tumor cells showed positive staining for CD117 (14/15, 93.3%) and CD34 (10/15, 66.7%). The results show that the application of virtual reality technology to CT imaging examination can significantly improve the diagnostic accuracy of gastrointestinal stromal tumors.
Collapse
Affiliation(s)
- Zhiying Wang
- Department of Gastroenterology, West District of Qingdao Municipal Hospital, Qingdao 266000, Shandong, China
| | - Qiaoyan Qu
- Department of Gastroenterology, West District of Qingdao Municipal Hospital, Qingdao 266000, Shandong, China
| | - Ke Cai
- Internal Medicine, Songshan Hospital of Medical College of Qingdao University, Qingdao 266000, Shandong, China
| | - Ting Xu
- Department of Gastroenterology, West District of Qingdao Municipal Hospital, Qingdao 266000, Shandong, China
| |
Collapse
|
13
|
Rai L, Li H. MyOcrTool: Visualization System for Generating Associative Images of Chinese Characters in Smart Devices. COMPLEXITY 2021; 2021:1-14. [DOI: 10.1155/2021/5583287] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/02/2023]
Abstract
Majority of Chinese characters are pictographic characters with strong associative ability and when a character appears for Chinese readers, they usually associate with the objects, or actions related to the character immediately. Having this background, we propose a system to visualize the simplified Chinese characters, so that developing any skills of either reading or writing Chinese characters is not necessary. Considering the extensive use and application of mobile devices, automatic identification of Chinese characters and display of associative images are made possible in smart devices to facilitate quick overview of a Chinese text. This work is of practical significance considering the research and development of real-time Chinese text recognition, display of associative images and for such users who would like to visualize the text with only images. The proposed Chinese character recognition system and visualization tool is named as MyOcrTool and developed for Android platform. The application recognizes the Chinese characters through OCR engine, and uses the internal voice playback interface to realize the audio functions and display the visual images of Chinese characters in real-time.
Collapse
Affiliation(s)
- Laxmisha Rai
- College of Electronic and Information Engineering, Shandong University of Science and Technology, Qingdao 266590, China
| | - Hong Li
- College of Electronic and Information Engineering, Shandong University of Science and Technology, Qingdao 266590, China
| |
Collapse
|
14
|
Tian S, Chen D, Wang H, Liu J. Deep convolution stack for waveform in underwater acoustic target recognition. Sci Rep 2021; 11:9614. [PMID: 33953232 PMCID: PMC8099869 DOI: 10.1038/s41598-021-88799-z] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2021] [Accepted: 04/13/2021] [Indexed: 11/09/2022] Open
Abstract
In underwater acoustic target recognition, deep learning methods have been proved to be effective on recognizing original signal waveform. Previous methods often utilize large convolutional kernels to extract features at the beginning of neural networks. It leads to a lack of depth and structural imbalance of networks. The power of nonlinear transformation brought by deep network has not been fully utilized. Deep convolution stack is a kind of network frame with flexible and balanced structure and it has not been explored well in underwater acoustic target recognition, even though such frame has been proven to be effective in other deep learning fields. In this paper, a multiscale residual unit (MSRU) is proposed to construct deep convolution stack network. Based on MSRU, a multiscale residual deep neural network (MSRDN) is presented to classify underwater acoustic target. Dataset acquired in a real-world scenario is used to verify the proposed unit and model. By adding MSRU into Generative Adversarial Networks, the validity of MSRU is proved. Finally, MSRDN achieves the best recognition accuracy of 83.15%, improved by 6.99% from the structure related networks which take the original signal waveform as input and 4.48% from the networks which take the time-frequency representation as input.
Collapse
Affiliation(s)
- Shengzhao Tian
- Big Data Research Center, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Duanbing Chen
- Big Data Research Center, University of Electronic Science and Technology of China, Chengdu, 611731, China.,The Research Base of Digital Culture and Media, Sichuan Provincial Key Research Base of Social Science, Chengdu, 611731, China.,Union Big Data Tech. Inc., Chengdu, 610041, China
| | - Hang Wang
- Big Data Research Center, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Jingfa Liu
- Guangzhou Key Laboratory of Multilingual Intelligent Processing, Guangdong University of Foreign Studies, Guangzhou, 510006, China. .,School of Information Science and Technology, Guangdong University of Foreign Studies, Guangzhou, 510006, China.
| |
Collapse
|
15
|
Hu B. Deep Learning Image Feature Recognition Algorithm for Judgment on the Rationality of Landscape Planning and Design. COMPLEXITY 2021; 2021:1-15. [DOI: 10.1155/2021/9921095] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/01/2023]
Abstract
This paper uses an improved deep learning algorithm to judge the rationality of the design of landscape image feature recognition. The preprocessing of the image is proposed to enhance the data. The deficiencies in landscape feature extraction are further addressed based on the new model. Then, the two-stage training method of the model is used to solve the problems of long training time and convergence difficulties in deep learning. Innovative methods for zoning and segmentation training of landscape pattern features are proposed, which makes model training faster and generates more creative landscape patterns. Because of the impact of too many types of landscape elements in landscape images, traditional convolutional neural networks can no longer effectively solve this problem. On this basis, a fully convolutional neural network model is designed to perform semantic segmentation of landscape elements in landscape images. Through the method of deconvolution, the pixel-level semantic segmentation is realized. Compared with the 65% accuracy rate of the convolutional neural network, the fully convolutional neural network has an accuracy rate of 90.3% for the recognition of landscape elements. The method is effective, accurate, and intelligent for the classification of landscape element design, which better improves the accuracy of classification, greatly reduces the cost of landscape element design classification, and ensures that the technical method is feasible. This paper classifies landscape behavior based on this model for full convolutional neural network landscape images and demonstrates the effectiveness of using the model. In terms of landscape image processing, the image evaluation provides a certain basis.
Collapse
Affiliation(s)
- Bin Hu
- Xinyang Vocational and Technical College, Xinyang 464000, Henan, China
| |
Collapse
|
16
|
Yang J, Lei Y, Tian Y, Xi M. Deep learning based six‐dimensional pose estimation in virtual reality. Comput Intell 2021. [DOI: 10.1111/coin.12453] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Affiliation(s)
- Jiachen Yang
- School of Electrical and Information Engineering Tianjin University Tianjin China
| | - Yutian Lei
- School of Electrical and Information Engineering Tianjin University Tianjin China
| | - Ying Tian
- School of Education Science and Technology Anshan Normal University Anshan China
| | - Meng Xi
- School of Electrical and Information Engineering Tianjin University Tianjin China
| |
Collapse
|
17
|
Li Y, Chao X. Distance-Entropy: An Effective Indicator for Selecting Informative Data. FRONTIERS IN PLANT SCIENCE 2021; 12:818895. [PMID: 35095987 PMCID: PMC8792929 DOI: 10.3389/fpls.2021.818895] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/20/2021] [Accepted: 12/08/2021] [Indexed: 05/11/2023]
Abstract
Smart agriculture is inseparable from data gathering, analysis, and utilization. A high-quality data improves the efficiency of intelligent algorithms and helps reduce the costs of data collection and transmission. However, the current image quality assessment research focuses on visual quality, while ignoring the crucial information aspect. In this work, taking the crop pest recognition task as an example, we proposed an effective indicator of distance-entropy to distinguish the good and bad data from the perspective of information. Many comparative experiments, considering the mapping feature dimensions and base data sizes, were conducted to testify the validity and robustness of this indicator. Both the numerical and the visual results demonstrate the effectiveness and stability of the proposed distance-entropy method. In general, this study is a relatively cutting-edge work in smart agriculture, which calls for attention to the quality assessment of the data information and provides some inspiration for the subsequent research on data mining, as well as for the dataset optimization for practical applications.
Collapse
Affiliation(s)
- Yang Li
- College of Mechanical and Electrical Engineering, Shihezi University, Xinjiang, China
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Xuewei Chao
- College of Mechanical and Electrical Engineering, Shihezi University, Xinjiang, China
- *Correspondence: Xuewei Chao
| |
Collapse
|