1
|
Hou Y, Wu Z, Cai X, Zhu T. The application of improved densenet algorithm in accurate image recognition. Sci Rep 2024; 14:8645. [PMID: 38622153 PMCID: PMC11018628 DOI: 10.1038/s41598-024-58421-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Accepted: 03/28/2024] [Indexed: 04/17/2024] Open
Abstract
Image recognition technology belongs to an important research field of artificial intelligence. In order to enhance the application value of image recognition technology in the field of computer vision and improve the technical dilemma of image recognition, the research improves the feature reuse method of dense convolutional network. Based on gradient quantization, traditional parallel algorithms have been improved. This improvement allows for independent parameter updates layer by layer, reducing communication time and data volume. The introduction of quantization error reduces the impact of gradient loss on model convergence. The test results show that the improvement strategy designed by the research improves the model parameter efficiency while ensuring the recognition effect. Narrowing the learning rate is conducive to refining the updating granularity of model parameters, and deepening the number of network layers can effectively improve the final recognition accuracy and convergence effect of the model. It is better than the existing state-of-the-art image recognition models, visual geometry group and EfficientNet. The parallel acceleration algorithm, which is improved by the gradient quantization, performs better than the traditional synchronous data parallel algorithm, and the improvement of the acceleration ratio is obvious. Compared with the traditional synchronous data parallel algorithm and stale synchronous parallel algorithm, the optimized parallel acceleration algorithm of the study ensures the image data training speed and solves the bottleneck problem of communication data. The model designed by the research improves the accuracy and training speed of image recognition technology and expands the use of image recognition technology in the field of computer vision.Please confirm the affiliation details of [1] is correct.The relevant detailed information in reference [1] has been confirmed to be correct.
Collapse
Affiliation(s)
- Yuntao Hou
- Heilongjiang Academy of Agricultural Machinery Sciences, Heilongjiang Academy of Agricultural Sciences, Harbin, 150081, China.
| | - Zequan Wu
- Heilongjiang Academy of Agricultural Machinery Sciences, Heilongjiang Academy of Agricultural Sciences, Harbin, 150081, China
| | - Xiaohua Cai
- Heilongjiang Academy of Agricultural Machinery Sciences, Heilongjiang Academy of Agricultural Sciences, Harbin, 150081, China
| | - Tianyu Zhu
- Heilongjiang Academy of Agricultural Machinery Sciences, Heilongjiang Academy of Agricultural Sciences, Harbin, 150081, China
| |
Collapse
|
2
|
Su T, Liu A, Shi Y, Zhang X. IremulbNet: Rethinking the inverted residual architecture for image recognition. Neural Netw 2024; 172:106140. [PMID: 38278090 DOI: 10.1016/j.neunet.2024.106140] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Revised: 12/23/2023] [Accepted: 01/18/2024] [Indexed: 01/28/2024]
Abstract
An increasing need of running Convolutional Neural Network (CNN) models on mobile devices encourages the studies on efficient and lightweight neural network model. In this paper, an Inverse Residual Multi-Branch Network named IremulbNet is proposed to solve the problem of insufficient classification accuracy in existing lightweight network models. The core module of this model is to reconstruct an inverse residual structure, in which a special feature fusion method, multi-branch feature extraction, and depthwise separable convolution techniques are used to improve the classification accuracy. The performance of model is tested using image databases. Experimental results show that for the fine-grained image dataset Imagenet-woof, IremulbNet achieved 10.9%, 12.2%, and 15.3% higher accuracy than that of MobileNet V3, ShuffleNet V2, and PeleeNet, respectively. Moreover, it can reduce inference time (GPU) about 42.09% and 75.56% compared to classic ResNet50 and DenseNet121.
Collapse
Affiliation(s)
- Tiantian Su
- Shaanxi Normal University, Xi'an 710119, Shaanxi, China
| | - Anan Liu
- Shaanxi Normal University, Xi'an 710119, Shaanxi, China
| | - Yongran Shi
- Shaanxi Normal University, Xi'an 710119, Shaanxi, China
| | - Xiaofeng Zhang
- Shaanxi Normal University, Xi'an 710119, Shaanxi, China.
| |
Collapse
|
3
|
Nakib MI, Mridha M. Comprehensive watermelon disease recognition dataset. Data Brief 2024; 53:110182. [PMID: 38425879 PMCID: PMC10904151 DOI: 10.1016/j.dib.2024.110182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Revised: 02/02/2024] [Accepted: 02/05/2024] [Indexed: 03/02/2024] Open
Abstract
Plant diseases pose a significant obstacle to global agricultural productivity, impacting crop quality yield and causing substantial economic losses for farmers. Watermelon, a commonly cultivated succulent vine plant, is rich in hydration and essential nutrients. However, it is susceptible to various diseases due to unfavorable environmental conditions and external factors, leading to compromised quality and substantial financial setbacks. Swift identification and management of crop diseases are imperative to minimize losses, enhance yield, reduce costs, and bolster agricultural output. Conventional disease diagnosis methods are often labor-intensive, time-consuming, ineffective, and prone to subjectivity. As a result, there is a critical need to advance research into machine-based models for disease detection in watermelons. This paper presents a large dataset of watermelons that can be used to train a machine vision-based illness detection model. Images of healthy and diseased watermelons from the Mosaic Virus, Anthracnose, and Downy Mildew Disease are included in the dataset's five separate classifications. Images were painstakingly collected on June 25, 2023, in close cooperation with agricultural experts from the highly regarded Regional Horticulture Research Station in Lebukhali, Patuakhali.
Collapse
Affiliation(s)
- Mohammad Imtiaz Nakib
- Department of Computer Science, American International University-Bangladesh, Dhaka, Bangladesh
| | - M.F. Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka, Bangladesh
| |
Collapse
|
4
|
Zhai Y, Chu L, Liu Y, Wang D, Wu Y. Using deep learning-based artificial intelligence electronic images in improving middle school teachers' literacy. PeerJ Comput Sci 2024; 10:e1844. [PMID: 38660146 PMCID: PMC11041997 DOI: 10.7717/peerj-cs.1844] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Accepted: 01/09/2024] [Indexed: 04/26/2024]
Abstract
With the rapid development of societal information, electronic educational resources have become an indispensable component of modern education. In response to the increasingly formidable challenges faced by secondary school teachers, this study endeavors to analyze and explore the application of artificial intelligence (AI) methods to enhance their cognitive literacy. Initially, this discourse delves into the application of AI-generated electronic images in the training and instruction of middle school educators, subjecting it to thorough analysis. Emphasis is placed on elucidating the pivotal role played by AI electronic images in elevating the proficiency of middle school teachers. Subsequently, an integrated intelligent device serves as the foundation for establishing a model that applies intelligent classification and algorithms based on the Structure of the Observed Learning Outcome (SOLO). This model is designed to assess the cognitive literacy and teaching efficacy of middle school educators, and its performance is juxtaposed with classification algorithms such as support vector machine (SVM) and decision trees. The findings reveal that, following 600 iterations of the model, the SVM algorithm achieves a 77% accuracy rate in recognizing teacher literacy, whereas the SOLO algorithm attains 80%. Concurrently, the spatial complexities of the SVM-based and SOLO-based intelligent literacy improvement models are determined to be 45 and 22, respectively. Notably, it is discerned that, with escalating iterations, the SOLO algorithm exhibits higher accuracy and reduced spatial complexity in evaluating teachers' pedagogical literacy. Consequently, the utilization of AI methodologies proves highly efficacious in advancing electronic imaging technology and enhancing the efficacy of image recognition in educational instruction.
Collapse
Affiliation(s)
- Yixi Zhai
- School of Foreign Studies, Tangshan Normal University, Tangshan City, China
| | - Liqing Chu
- School of Foreign Studies, Tangshan Normal University, Tangshan City, China
| | - Yanlan Liu
- School of Foreign Studies, Tangshan Normal University, Tangshan City, China
| | - Dandan Wang
- School of Foreign Studies, Tangshan Normal University, Tangshan City, China
| | - Yufei Wu
- School of Foreign Studies, Tangshan Normal University, Tangshan City, China
| |
Collapse
|
5
|
Yang C, Sun X, Wang J, Lv H, Dong P, Xi L, Shi L. YOLOv8s-CGF: a lightweight model for wheat ear Fusarium head blight detection. PeerJ Comput Sci 2024; 10:e1948. [PMID: 38660210 PMCID: PMC11041926 DOI: 10.7717/peerj-cs.1948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2023] [Accepted: 02/29/2024] [Indexed: 04/26/2024]
Abstract
Fusarium head blight (FHB) is a destructive disease that affects wheat production. Detecting FHB accurately and rapidly is crucial for improving wheat yield. Traditional models are difficult to apply to mobile devices due to large parameters, high computation, and resource requirements. Therefore, this article proposes a lightweight detection method based on an improved YOLOv8s to facilitate the rapid deployment of the model on mobile terminals and improve the detection efficiency of wheat FHB. The proposed method introduced a C-FasterNet module, which replaced the C2f module in the backbone network. It helps reduce the number of parameters and the computational volume of the model. Additionally, the Conv in the backbone network is replaced with GhostConv, further reducing parameters and computation without significantly affecting detection accuracy. Thirdly, the introduction of the Focal CIoU loss function reduces the impact of sample imbalance on the detection results and accelerates the model convergence. Lastly, the large target detection head was removed from the model for lightweight. The experimental results show that the size of the improved model (YOLOv8s-CGF) is only 11.7 M, which accounts for 52.0% of the original model (YOLOv8s). The number of parameters is only 5.7 × 106 M, equivalent to 51.4% of the original model. The computational volume is only 21.1 GFLOPs, representing 74.3% of the original model. Moreover, the mean average precision (mAP@0.5) of the model is 99.492%, which is 0.003% higher than the original model, and the mAP@0.5:0.95 is 0.269% higher than the original model. Compared to other YOLO models, the improved lightweight model not only achieved the highest detection precision but also significantly reduced the number of parameters and model size. This provides a valuable reference for FHB detection in wheat ears and deployment on mobile terminals in field environments.
Collapse
Affiliation(s)
- Chengkai Yang
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
| | - Xiaoyun Sun
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
| | - Jian Wang
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
| | - Haiyan Lv
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
| | - Ping Dong
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
| | - Lei Xi
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
| | - Lei Shi
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan, China
- Henan Grain Crop Collaborative Innovation Center, Henan Agricultural University, Zhengzhou, Henan, China
| |
Collapse
|
6
|
Zhou PK, Li Y, Zeng T, Chee MY, Huang Y, Yu Z, Yu H, Yu H, Huang W, Chen X. One-Dimensional Covalent Organic Framework-Based Multilevel Memristors for Neuromorphic Computing. Angew Chem Int Ed Engl 2024:e202402911. [PMID: 38511343 DOI: 10.1002/anie.202402911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2024] [Revised: 03/08/2024] [Accepted: 03/21/2024] [Indexed: 03/22/2024]
Abstract
Memristors are essential components of neuromorphic systems that mimic the synaptic plasticity observed in biological neurons. In this study, a novel approach employing one-dimensional covalent organic framework (1D COF) films was explored to enhance the performance of memristors. The unique structural and electronic properties of two 1D COF films (COF-4,4'-methylenedianiline (MDA) and COF-4,4'-oxydianiline (ODA)) offer advantages for multilevel resistive switching, which is a key feature in neuromorphic computing applications. By further introducing a TiO2 layer on the COF-ODA film, a built-in electric field between the COF-TiO2 interfaces could be generated, demonstrating the feasibility of utilizing COFs as a platform for constructing memristors with tunable resistive states. The 1D nanochannels of these COF structures contributed to the efficient modulation of electrical conductance, enabling precise control over synaptic weights in neuromorphic circuits. This study also investigated the potential of these COF-based memristors to achieve energy-efficient and high-density memory devices.
Collapse
Affiliation(s)
- Pan-Ke Zhou
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| | - Yiping Li
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| | - Tao Zeng
- Department of Materials Science and Engineering, National University of Singapore, Singapore, 117575, Singapore
| | - Mun Yin Chee
- School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore, 637371, Singapore
| | - Yuxing Huang
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| | - Ziyue Yu
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| | - Hongling Yu
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| | - Hong Yu
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| | - Weiguo Huang
- State Key Laboratory of Structural Chemistry, Fujian Institute of Research on the Structure of Matter, Chinese Academy of Sciences, 155 Yangqiao West Road, Fuzhou, Fujian, 350002, China
| | - Xiong Chen
- State Key Laboratory of Photocatalysis on Energy and Environment, and Key Laboratory of Molecular Synthesis and Function Discovery, College of Chemistry, Fuzhou University, Fujian, 350108, China
| |
Collapse
|
7
|
Sasaki R, Fujinami M, Nakai H. Comprehensive image dataset for enhancing object detection in chemical experiments. Data Brief 2024; 52:110054. [PMID: 38293577 PMCID: PMC10827390 DOI: 10.1016/j.dib.2024.110054] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2023] [Revised: 01/04/2024] [Accepted: 01/05/2024] [Indexed: 02/01/2024] Open
Abstract
The application of image recognition in chemical experiments has the potential to enhance experiment recording and risk management. However, the current scarcity of suitable benchmarking datasets restricts the applications of machine vision techniques in chemical experiments. This data article presents an image dataset featuring common chemical apparatuses and experimenter's hands. The images have been meticulously annotated, providing detailed information for precise object detection through deep learning methods. The images were captured from videos filmed in organic chemistry laboratories. This dataset comprises a total of 5078 images including diverse backgrounds and situations surrounding the objects. Detailed annotations are provided in accompanying text files. The dataset is organized into training, validation, and test subsets. Each subset is stored within independent folders for easy access and utilization.
Collapse
Affiliation(s)
- Ryosuke Sasaki
- Department of Chemistry and Biochemistry, School of Advanced Science and Engineering, Waseda University, 3-4-1 Okubo, Shinjuku-ku, Tokyo 169-8555, Japan
| | - Mikito Fujinami
- Waseda Research Institute for Science and Engineering, Waseda University, 3-4-1 Okubo, Shinjuku-ku, Tokyo 169-8555, Japan
| | - Hiromi Nakai
- Department of Chemistry and Biochemistry, School of Advanced Science and Engineering, Waseda University, 3-4-1 Okubo, Shinjuku-ku, Tokyo 169-8555, Japan
- Waseda Research Institute for Science and Engineering, Waseda University, 3-4-1 Okubo, Shinjuku-ku, Tokyo 169-8555, Japan
| |
Collapse
|
8
|
Khatun T, Nirob MAS, Bishshash P, Akter M, Uddin MS. A comprehensive dragon fruit image dataset for detecting the maturity and quality grading of dragon fruit. Data Brief 2024; 52:109936. [PMID: 38125368 PMCID: PMC10733099 DOI: 10.1016/j.dib.2023.109936] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2023] [Revised: 11/27/2023] [Accepted: 12/04/2023] [Indexed: 12/23/2023] Open
Abstract
Dragon fruit, often referred to as pitaya, is a tropical fruit with various types, including both white-fleshed and red-fleshed varieties. Its distinctive appearance is complemented by a range of potential health advantages. These include its abundance of nutrients and antioxidants, which contribute to a robust immune system, aid in blood sugar regulation, and support the well-being of the heart, bones, and skin. Consequently, the global desire for dragon fruit is yielding substantial economic advantages for developing nations like Bangladesh, which in turn underscores the pressing need for an automated system to identify the optimal harvest time and differentiate between fresh and defective fruits to ensure quality. To accomplish this objective, this paper introduces an extensive collection of high-resolution dragon fruits because effective detection by machine learning models necessitates a substantial amount of data. The dataset was painstakingly gathered during a span of four months from three distinct locations in Bangladesh, with the valuable assistance of domain experts. Possible application of the dataset encompasses quality evaluation, robotic harvesting, and packaging systems, ultimately boosting the effectiveness of dragon fruit production procedures. The dataset has the potential to be a valuable resource for researchers interested in dragon fruit cultivation, offering a solid foundation for the application of computer vision and deep learning methods in the agricultural industry.
Collapse
Affiliation(s)
- Tania Khatun
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Md. Asraful Sharker Nirob
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Prayma Bishshash
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Morium Akter
- Department of Computer Science and Engineering, Jahangirnagar University, Dhaka, Bangladesh
| | - Mohammad Shorif Uddin
- Department of Computer Science and Engineering, Jahangirnagar University, Dhaka, Bangladesh
| |
Collapse
|
9
|
Zhang M, Huang Y, Xie D, Huang R, Zeng G, Liu X, Deng H, Wang H, Lin Z. Machine learning constructs color features to accelerate development of long-term continuous water quality monitoring. J Hazard Mater 2024; 461:132612. [PMID: 37801971 DOI: 10.1016/j.jhazmat.2023.132612] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/28/2023] [Revised: 09/20/2023] [Accepted: 09/21/2023] [Indexed: 10/08/2023]
Abstract
Long-term continuous water quality monitoring (LTCM) is crucial to ensure the safety of water resources. However, lab-based pollutant detection via machine learning (ML) usually involves colorimetric materials or sensors, and it cannot be ignored that sensor limitations prevent their use for LTCM. To address this challenge, we propose a novel method that leverages image recognition to establish a relationship between pollutant concentration and color. By extracting efficient color variation features from raw pixel matrices using a combination of Kmeans clustering and RGB average features, the concentrations of pollutants that are difficult to distinguish by the naked eyes can be directly captured without the need for sensors and preprocessing. Four ML models (XGBoost, Linear, support vector regression (SVR), and Ridge) achieved up to a 95.9% increase in coefficient of determination (R2) compared to principal component analysis (PCA). In the prediction of the concentration of simulated pollutants such as Cu2+, Co2+, Rhodamine B, and the concentration of Cr(VI) in actual electroplating wastewater, natural resource water and drinking water, over 95% R2 was achieved. The method reported in our work can effectively capture subtle color changes that cannot be observed by the naked eyes without any preprocessing of water samples, providing a reliable method for LTCM.
Collapse
Affiliation(s)
- Mengyuan Zhang
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China
| | - Yanquan Huang
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China
| | - Dongsheng Xie
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China
| | - Renfeng Huang
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China
| | - Gongchang Zeng
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China
| | - Xueming Liu
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China.
| | - Hong Deng
- School of Environmental Science and Engineering South China University of Technology, Guangzhou 510006, China.
| | - Haiying Wang
- School of Metallurgy and Environment, Central South University, Changsha 410083, China
| | - Zhang Lin
- School of Metallurgy and Environment, Central South University, Changsha 410083, China
| |
Collapse
|
10
|
Khatun T, Razzak A, Islam MS, Uddin MS. An extensive real-world in field tomato image dataset involving maturity classification and recognition of fresh and defect tomatoes. Data Brief 2023; 51:109688. [PMID: 37920387 PMCID: PMC10618421 DOI: 10.1016/j.dib.2023.109688] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Revised: 10/02/2023] [Accepted: 10/11/2023] [Indexed: 11/04/2023] Open
Abstract
Tomato, a fruiting plant species within the Solanaceae family, is a widely used ingredient in culinary dishes due to its sweet and acidic flavor profile, as well as its rich nutritional content. Recognized for its potential health benefits, including reducing the risk of coronary artery disease and specific types of cancer, tomatoes have become a staple in global cuisine. Traditional methods for tomato maturity assessment, harvesting, quality grading, and packaging are often labor-intensive and economically inefficient. This paper introduces an extensive dataset of high-resolution tomato images collected over an eight-month period from the demonstration fields of Sher-E-Bangla Agricultural University in Dhaka, Bangladesh, in collaboration with plant breeding experts of the same university. The dataset was meticulously curated to ensure precision and consistency, encompassing various stages of tomato maturity, including images of both fresh and defective tomatoes. This dataset is a valuable resource for researchers, stakeholders, and individuals interested in tomato production in Bangladesh, providing a robust foundation for leveraging computer vision and deep learning techniques in the agriculture sector. The dataset's potential applications extend to automating tasks such as robotic harvesting, quality assessment, and packaging systems, ultimately enhancing the efficiency of tomato production processes.
Collapse
Affiliation(s)
- Tania Khatun
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Abdur Razzak
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Md. Shofiul Islam
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Mohammad Shorif Uddin
- Department of Computer Science and Engineering, Jahangirnagar University, Dhaka, Bangladesh
| |
Collapse
|
11
|
Fatima N, Rizvi SAM, Rizvi MSBA. Dermatological disease prediction and diagnosis system using deep learning. Ir J Med Sci 2023:10.1007/s11845-023-03578-1. [PMID: 38036757 DOI: 10.1007/s11845-023-03578-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Accepted: 11/20/2023] [Indexed: 12/02/2023]
Abstract
The prevalence of skin illnesses is higher than that of other diseases. Fungal infection, bacteria, allergies, viruses, genetic factors, and environmental factors are among important causative factors that have continuously escalated the degree and incidence of skin diseases. Medical technology based on lasers and photonics has made it possible to identify skin illnesses considerably more rapidly and correctly. However, the cost of such a diagnosis is currently limited and prohibitively high and restricted to developed areas. The present paper develops a holistic, critical, and important skin disease prediction system that utilizes machine learning and deep learning algorithms to accurately identify up to 20 different skin diseases with a high F1 score and efficiency. Deep learning algorithms like Xception, Inception-v3, Resnet50, DenseNet121, and Inception-ResNet-v2 were employed to accurately classify diseases based on the images. The training and testing have been performed on an enlarged dataset, and classification was performed for 20 diseases. The algorithm developed was free from any inherent bias and treated all classes equally. The present model, which was trained using the Xception algorithm, is highly efficient and accurate for 20 different skin conditions, with a dataset of over 10,000 photos. The developed system was able to classify 20 different dermatological diseases with high accuracy and precision.
Collapse
Affiliation(s)
- Neda Fatima
- Manav Rachna International Institute of Research and Studies, Faridabad, Haryana, India.
| | | | | |
Collapse
|
12
|
Zhao Z, Gao Z, Zhang K, Lun L, Xu W, Wu H, Liu B. Colon Disease Classification Method Based on Deep Learning. Stud Health Technol Inform 2023; 308:689-695. [PMID: 38007800 DOI: 10.3233/shti230901] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2023]
Abstract
Objective Colorectal cancer (CRC) is a common malignant tumor of the digestive system with a high incidence rate. It is prone to misdiagnosis or missed diagnosis in clinical practice. Therefore, researching computer-aided diagnostic methods for endoscopic colon disease image classification is of great importance. This study proposes a deep learning-based method for colon disease classification. It utilizes intestinal images or captures from an endoscope camera to achieve intelligent classification of gastrointestinal diseases, providing assistance to doctors in their decision-making process. Methods Firstly, the algorithm is used to preprocess the dataset by removing duplicates and applying enhancement techniques. Two different network architectures, namely A_Vit, MobileNet, are employed. The models are trained using the same parameters and dataset with the Adam optimizer. The training process generates loss curves, accuracy, and recall rates for each of the four network architectures. Results The results indicate that the training with A_Vit has shown better performance, achieving an accuracy rate of 95.76% and an impressive recall rate of 97.21%. Therefore, the model trained using the A_Vit network structure is ultimately selected as the preferred choice. Conclusion This method can improve the efficiency and accuracy of colon disease diagnosis.
Collapse
Affiliation(s)
- Zhihe Zhao
- Hebei University of Science and Technology
| | | | - Kun Zhang
- Hebei University of Science and Technology
| | - Lei Lun
- Hebei University of Science and Technology
| | - Weichao Xu
- Hebei Provincial Hospital of Traditional Chinese Medicine
| | - Hongxin Wu
- Hebei University of Science and Technology
| | | |
Collapse
|
13
|
Wen T, Li JH, Wang Q, Gao YY, Hao GF, Song BA. Thermal imaging: The digital eye facilitates high-throughput phenotyping traits of plant growth and stress responses. Sci Total Environ 2023; 899:165626. [PMID: 37481085 DOI: 10.1016/j.scitotenv.2023.165626] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/04/2023] [Revised: 07/13/2023] [Accepted: 07/16/2023] [Indexed: 07/24/2023]
Abstract
Plant phenotyping is important for plants to cope with environmental changes and ensure plant health. Imaging techniques are perceived as the most critical and reliable tools for studying plant phenotypes. Thermal imaging has opened up new opportunities for nondestructive imaging of plant phenotyping. However, a comprehensive summary of thermal imaging in plant phenotyping is still lacking. Here we discuss the progress and future prospects of thermal imaging for assessing plant growth and stress responses. First, we classify thermal imaging into ground-based and aerial platforms based on their adaptability to different experimental environments (including laboratory, greenhouse, and field). It is convenient to collect phenotypic information of different dimensions. Second, in order to enhance the efficiency of thermal image processing, automatic algorithms based on deep learning are employed instead of traditional manual methods, greatly reducing the time cost of experiments. Considering its ease of implementation, handling and instant response, thermal imaging has been widely used in research on environmental stress, crop yield, and seed vigor. We have found that thermal imaging can detect thermal energy dissipation caused by living organisms (e.g., pests, viruses, bacteria, fungi, and oomycetes), enabling early disease diagnosis. It also recognizes changes leaf surface temperatures resulting from reduced transpiration rates caused by nutrient deficiency, drought, salinity, or freezing. Furthermore, thermal imaging predicts crop yield under different water states and forecasts the viability of dormant seeds after water absorption by monitoring temperature changes in the seeds. This work will assist biologists and agronomists in studying plant phenotypes and serve a guide for breeders to develop high-yielding, stress-tolerant, and superior crops.
Collapse
Affiliation(s)
- Ting Wen
- National Key Laboratory of Green Pesticide, State Key Laboratory Breeding Base of Green Pesticide and Agricultural Bioengineering, Key Laboratory of Green Pesticide and Agricultural Bioengineering, Ministry of Education, Center for Research and Development of Fine Chemicals, Guizhou University, Guiyang 550025, PR China
| | - Jian-Hong Li
- National Key Laboratory of Green Pesticide, State Key Laboratory Breeding Base of Green Pesticide and Agricultural Bioengineering, Key Laboratory of Green Pesticide and Agricultural Bioengineering, Ministry of Education, Center for Research and Development of Fine Chemicals, Guizhou University, Guiyang 550025, PR China
| | - Qi Wang
- State Key Laboratory of Public Big Data, Guizhou University, Guiyang 550025, PR China.
| | - Yang-Yang Gao
- National Key Laboratory of Green Pesticide, State Key Laboratory Breeding Base of Green Pesticide and Agricultural Bioengineering, Key Laboratory of Green Pesticide and Agricultural Bioengineering, Ministry of Education, Center for Research and Development of Fine Chemicals, Guizhou University, Guiyang 550025, PR China.
| | - Ge-Fei Hao
- National Key Laboratory of Green Pesticide, State Key Laboratory Breeding Base of Green Pesticide and Agricultural Bioengineering, Key Laboratory of Green Pesticide and Agricultural Bioengineering, Ministry of Education, Center for Research and Development of Fine Chemicals, Guizhou University, Guiyang 550025, PR China; Key Laboratory of Pesticide & Chemical Biology, Ministry of Education, College of Chemistry, Central China Normal University, Wuhan 430079, China.
| | - Bao-An Song
- National Key Laboratory of Green Pesticide, State Key Laboratory Breeding Base of Green Pesticide and Agricultural Bioengineering, Key Laboratory of Green Pesticide and Agricultural Bioengineering, Ministry of Education, Center for Research and Development of Fine Chemicals, Guizhou University, Guiyang 550025, PR China
| |
Collapse
|
14
|
Fan YL, Hsu FR, Wang Y, Liao LD. Unlocking the Potential of Zebrafish Research with Artificial Intelligence: Advancements in Tracking, Processing, and Visualization. Med Biol Eng Comput 2023; 61:2797-2814. [PMID: 37558927 DOI: 10.1007/s11517-023-02903-1] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Accepted: 08/01/2023] [Indexed: 08/11/2023]
Abstract
Zebrafish have become a widely accepted model organism for biomedical research due to their strong cortisol stress response, behavioral strain differences, and sensitivity to both drug treatments and predators. However, experimental zebrafish studies generate substantial data that must be analyzed through objective, accurate, and repeatable analysis methods. Recently, advancements in artificial intelligence (AI) have enabled automated tracking, image recognition, and data analysis, leading to more efficient and insightful investigations. In this review, we examine key AI applications in zebrafish research, including behavior analysis, genomics, and neuroscience. With the development of deep learning technology, AI algorithms have been used to precisely analyze and identify images of zebrafish, enabling automated testing and analysis. By applying AI algorithms in genomics research, researchers have elucidated the relationship between genes and biology, providing a better basis for the development of disease treatments and gene therapies. Additionally, the development of more effective neuroscience tools could help researchers better understand the complex neural networks in the zebrafish brain. In the future, further advancements in AI technology are expected to enable more extensive and in-depth medical research applications in zebrafish, improving our understanding of this important animal model. This review highlights the potential of AI technology in achieving the full potential of zebrafish research by enabling researchers to efficiently track, process, and visualize the outcomes of their experiments.
Collapse
Affiliation(s)
- Yi-Ling Fan
- Institute of Biomedical Engineering and Nanomedicine, National Health Research Institutes, 35, Keyan Road, Zhunan Town, Miaoli County, 35053, Taiwan
- Department of Information Engineering and Computer Science, Feng Chia University, Taichung, 407, Taiwan
| | - Fang-Rong Hsu
- Department of Information Engineering and Computer Science, Feng Chia University, Taichung, 407, Taiwan
| | - Yuhling Wang
- Institute of Biomedical Engineering and Nanomedicine, National Health Research Institutes, 35, Keyan Road, Zhunan Town, Miaoli County, 35053, Taiwan
- Department of Electrical Engineering, National United University, 2, Lien-Da, Nan-Shih Li, Miaoli, 360302, Taiwan
| | - Lun-De Liao
- Institute of Biomedical Engineering and Nanomedicine, National Health Research Institutes, 35, Keyan Road, Zhunan Town, Miaoli County, 35053, Taiwan.
| |
Collapse
|
15
|
Okawa J, Hori K, Izuno H, Fukuda M, Ujihashi T, Kodama S, Yoshimoto T, Sato R, Ono T. Developing tongue coating status assessment using image recognition with deep learning. J Prosthodont Res 2023:JPR_D_23_00117. [PMID: 37766551 DOI: 10.2186/jpr.jpr_d_23_00117] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/29/2023]
Abstract
PURPOSE To build an image recognition network to evaluate tongue coating status. METHODS Two image recognition networks were built: one for tongue detection and another for tongue coating classification. Digital tongue photographs were used to develop both networks; images from 251 (178 women, 74.7±6.6 years) and 144 older adults (83 women, 73.8±7.3 years) who volunteered to participate were used for the tongue detection network and coating classification network, respectively. The learning objective of the tongue detection network is to extract a rectangular region that includes the tongue. You-Only-Look-Once (YOLO) v2 was used as the detection network, and transfer learning was performed using ResNet-50. The accuracy was evaluated by calculating the intersection over the union. For tongue coating classification, the rectangular area including the tongue was divided into a grid of 7×7. Five experienced panelists scored the tongue coating in each area using one of five grades, and the tongue coating index (TCI) was calculated. Transfer learning for tongue coating grades was performed using ResNet-18, and the TCI was calculated. Agreement between the panelists and network for the tongue coating grades in each area and TCI was evaluated using the kappa coefficient and intraclass correlation, respectively. RESULTS The tongue detection network recognized the tongue with a high intersection over union (0.885±0.081). The tongue coating classification network showed high agreement with tongue coating grades and TCI, with a kappa coefficient of 0.826 and an intraclass correlation coefficient of 0.807, respectively. CONCLUSIONS Image recognition enables simple and detailed assessment of tongue coating status.
Collapse
Affiliation(s)
- Jumpei Okawa
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
| | - Kazuhiro Hori
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
| | - Hiromi Izuno
- Department of Oral Health Sciences, Faculty of Nursing and Health Care, BAIKA Women's University, Ibaraki, Japan
| | - Masayo Fukuda
- Department of Oral Health Science, Faculty of Health Science, Kobe Tokiwa University, Kobe, Japan
| | - Takako Ujihashi
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
- Department of Oral Health Science, Faculty of Health Science, Kobe Tokiwa University, Kobe, Japan
| | - Shohei Kodama
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
| | - Tasuku Yoshimoto
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
| | - Rikako Sato
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
| | - Takahiro Ono
- Division of Comprehensive Prosthodontics, Faculty of Dentistry & Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Japan
- Department of Geriatric Dentistry, Osaka Dental University, Osaka, Japan
| |
Collapse
|
16
|
Sultana N, Shorif SB, Akter M, Uddin MS. A dataset for successful recognition of cucumber diseases. Data Brief 2023; 49:109320. [PMID: 37456112 PMCID: PMC10338290 DOI: 10.1016/j.dib.2023.109320] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Revised: 04/07/2023] [Accepted: 06/12/2023] [Indexed: 07/18/2023] Open
Abstract
Plant disease is a common impediment to the productivity of the world's agricultural production, which adversely affects the quality and yield of crops and causes heavy economic losses to farmers. The cucumber is a frequently grown creeping vine plant that has few calories but is high in water and several vital vitamins and minerals. Due to the unfavorable ecological environment and non-biological circumstances, cucumber diseases will adversely harm the quality of cucumber and cause heavy financial loss. Early identification and protection of crop diseases are essential for disease management, crop yield enhancement, cost reduction, and boosting agricultural production. The traditional diagnosis of crop diseases is often time-consuming, laborious, ineffective, and subjective. To cope with this scenario, the development of a machine-based model which can detect cucumber diseases is a demand of time for increasing agricultural production. This article offers a major cucumber dataset to build an effective machine vision-based model which can detect more variety of cucumber diseases. The dataset includes eight different types of classes containing disease-affected and disease-free cucumber images (Anthracnose, Bacterial Wilt, Belly Rot, Downy Mildew, Pythium Fruit Rot, Gummy Stem Blight, Fresh leaves, and Fresh cucumber) which were collected from the 6th to 30th of May 2022 from real fields with the cooperation of an expert from an agricultural institution. The dataset is hosted by the Department of Computer Science and Engineering, Jahangirnagar University, and is freely accessible at https://data.mendeley.com/datasets/y6d3z6f8z9/1.
Collapse
|
17
|
Yang L, Zhou Z. Effects of nutrition intervention on the rehabilitation level and quality of life of patients with diabetes foot: Image observation based on image recognition technology. Prev Med 2023; 173:107578. [PMID: 37343728 DOI: 10.1016/j.ypmed.2023.107578] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Revised: 06/03/2023] [Accepted: 06/17/2023] [Indexed: 06/23/2023]
Abstract
Today, the application of CT scanning technology is still limited. Therefore, an image observation method based on image recognition technology is proposed in this paper to study the impact of nutritional intervention on the rehabilitation level and quality of life of diabetic foot patients. Firstly, in view of the noise problems existing in CT images, this paper adopts the denoising technology, in which Gaussian noise and salt and pepper noise denoising algorithms are mainly used. Noise interference in the image can be removed by denoising processing, and the clarity and quality of the image can be improved. Subsequently, in order to improve the recognition and classification accuracy of the image, data enhancement and data standardization techniques are adopted to normalize the pixel values of the image, so as to make different images comparable, so as to improve the stability and classification accuracy of the model. Finally, CNN model is used to study image classification. The effects of nutritional intervention on the rehabilitation level and quality of life of patients with diabetic foot were investigated by setting up a comparative experiment. The results showed that nutrition intervention nursing can effectively improve the rehabilitation level of diabetic foot patients, improve the quality of life, improve nursing satisfaction.
Collapse
Affiliation(s)
- Lina Yang
- Department of Plastic and Burn, Hainan Western Central Hospital, Danzhou 571700, China
| | - Zhicheng Zhou
- Department of Clinical Nutrition, The Fifth Affiliated Hospital, Sun Yat-sen University, Zhuhai 519000, Guangdong Province, China.
| |
Collapse
|
18
|
Matsuoka D, Kodama C, Yamada Y, Nakano M. Tropical cyclone dataset for a high-resolution global nonhydrostatic atmospheric simulation. Data Brief 2023; 48:109135. [PMID: 37122921 PMCID: PMC10139877 DOI: 10.1016/j.dib.2023.109135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2023] [Revised: 03/29/2023] [Accepted: 04/03/2023] [Indexed: 05/02/2023] Open
Abstract
This dataset is a time series of tropical cyclones simulated using the high-resolution Nonhydrostatic Icosahedral Atmospheric Model (NICAM). By tracking tropical cyclones from 30 years of simulation data, 2,463 tracks that include the life stages of precursors (pre-TCs), tropical cyclones (TCs), and post-tropical cyclones (post-TCs), if any, were extracted. Each track data includes the time, latitude, longitude, maximum wind speed, minimum pressure, elapsed time since onset, and life-stage label of the tropical cyclone. The numbers of steps (6 h) for pre-TCs, TCs, and post-TCs were 45,288, 55,206, and 37,312, respectively. The dataset for each step also consists of atmospheric field data of multiple physical quantities, such as outgoing longwave radiation at the top-of-the-atmosphere, sea level pressure, sea surface temperature, specific humidity at 600 hPa, and zonal and meridional winds at 850 and 200 hPa over a 1000 km2 area that includes a tropical cyclone at its center. This dataset can be used to develop machine-learning models for the detection, intensity prediction, and cyclogenesis prediction of tropical cyclones.
Collapse
Affiliation(s)
- Daisuke Matsuoka
- Center for Earth Information Science and Technology (CEIST), Research Institute for Value-Added-Information Generation (VAiG), Japan Agency for Marine-Earth Science and Technology (JAMSTEC) 3173-25 Showa-machi, Kanazawa-ku, Yokohama, Kanagawa 236-0001 Japan
- Typhoon Science and Technology Research Center (TRC), Institute for Multidisciplinary Sciences (IMS), Yokohama National University 79-1 Tokiwadai, Hodogaya-ku, Yokohama, Kanagawa 240-0067 Japan
- Corresponding author.
| | - Chihiro Kodama
- Research Center for Environmental Modeling and Application (CEMA), Research Institute for Global Change (RIGC), Japan Agency for Marine-Earth Science and Technology (JAMSTEC) 3173-25 Showa-machi, Kanazawa-ku, Yokohama, Kanagawa 236-0001 Japan
- Typhoon Science and Technology Research Center (TRC), Institute for Multidisciplinary Sciences (IMS), Yokohama National University 79-1 Tokiwadai, Hodogaya-ku, Yokohama, Kanagawa 240-0067 Japan
| | - Yohei Yamada
- Research Center for Environmental Modeling and Application (CEMA), Research Institute for Global Change (RIGC), Japan Agency for Marine-Earth Science and Technology (JAMSTEC) 3173-25 Showa-machi, Kanazawa-ku, Yokohama, Kanagawa 236-0001 Japan
| | - Masuo Nakano
- Research Center for Environmental Modeling and Application (CEMA), Research Institute for Global Change (RIGC), Japan Agency for Marine-Earth Science and Technology (JAMSTEC) 3173-25 Showa-machi, Kanazawa-ku, Yokohama, Kanagawa 236-0001 Japan
- Typhoon Science and Technology Research Center (TRC), Institute for Multidisciplinary Sciences (IMS), Yokohama National University 79-1 Tokiwadai, Hodogaya-ku, Yokohama, Kanagawa 240-0067 Japan
| |
Collapse
|
19
|
Varde AS, Karthikeyan D, Wang W. Facilitating COVID recognition from X-rays with computer vision models and transfer learning. Multimed Tools Appl 2023:1-32. [PMID: 37362714 PMCID: PMC10213594 DOI: 10.1007/s11042-023-15744-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/04/2021] [Revised: 05/17/2022] [Accepted: 04/22/2023] [Indexed: 06/28/2023]
Abstract
Multimedia data plays an important role in medicine and healthcare since EHR (Electronic Health Records) entail complex images and videos for analyzing patient data. In this article, we hypothesize that transfer learning with computer vision can be adequately harnessed on such data, more specifically chest X-rays, to learn from a few images for assisting accurate, efficient recognition of COVID. While researchers have analyzed medical data (including COVID data) using computer vision models, the main contributions of our study entail the following. Firstly, we conduct transfer learning using a few images from publicly available big data on chest X-rays, suitably adapting computer vision models with data augmentation. Secondly, we aim to find the best fit models to solve this problem, adjusting the number of samples for training and validation to obtain the minimum number of samples with maximum accuracy. Thirdly, our results indicate that combining chest radiography with transfer learning has the potential to improve the accuracy and timeliness of radiological interpretations of COVID in a cost-effective manner. Finally, we outline applications of this work during COVID and its recovery phases with future issues for research and development. This research exemplifies the use of multimedia technology and machine learning in healthcare.
Collapse
Affiliation(s)
- Aparna S. Varde
- School of Computing, Montclair State University, Montclair, NJ USA
- Max Planck Institute for Informatics (Visiting Researcher), Saarbrucken, Germany
| | | | - Weitian Wang
- School of Computing, Montclair State University, Montclair, NJ USA
| |
Collapse
|
20
|
Xue J, Xia S, Li Z, Wang X, Huang L, He R, Li S. [Intelligent identification of livestock, a source of Schistosoma japonicum infection, based on deep learning of unmanned aerial vehicle images]. Zhongguo Xue Xi Chong Bing Fang Zhi Za Zhi 2023; 35:121-127. [PMID: 37253560 DOI: 10.16250/j.32.1374.2022273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
OBJECTIVE To develop an intelligent recognition model based on deep learning algorithms of unmanned aerial vehicle (UAV) images, and to preliminarily explore the value of this model for remote identification, monitoring and management of cattle, a source of Schistosoma japonicum infection. METHODS Oncomelania hupensis snail-infested marshlands around the Poyang Lake area were selected as the study area. Image datasets of the study area were captured by aerial photography with UAV and subjected to augmentation. Cattle in the sample database were annotated with the annotation software VGG Image Annotator to create the morphological recognition labels for cattle. A model was created for intelligent recognition of livestock based on deep learning-based Mask R-convolutional neural network (CNN) algorithms. The performance of the model for cattle recognition was evaluated with accuracy, precision, recall, F1 score and mean precision. RESULTS A total of 200 original UAV images were obtained, and 410 images were yielded following data augmentation. A total of 2 860 training samples of cattle recognition were labeled. The created deep learning-based Mask R-CNN model converged following 200 iterations, with an accuracy of 88.01%, precision of 92.33%, recall of 94.06%, F1 score of 93.19%, and mean precision of 92.27%, and the model was effective to detect and segment the morphological features of cattle. CONCLUSIONS The deep learning-based Mask R-CNN model is highly accurate for recognition of cattle based on UAV images, which is feasible for remote intelligent recognition, monitoring, and management of the source of S. japonicum infection.
Collapse
Affiliation(s)
- J Xue
- National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research), National Health Commission Key Laboratory of Parasite and Vector Biology, WHO Collaborating Centre for Tropical Diseases, National Center for International Research on Tropical Diseases, Shanghai 200025, China
- School of Global Health, Shanghai Jiao Tong University School of Medicine and Chinese Center for Tropical Diseases Research, Shanghai 200025, China
| | - S Xia
- National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research), National Health Commission Key Laboratory of Parasite and Vector Biology, WHO Collaborating Centre for Tropical Diseases, National Center for International Research on Tropical Diseases, Shanghai 200025, China
- School of Global Health, Shanghai Jiao Tong University School of Medicine and Chinese Center for Tropical Diseases Research, Shanghai 200025, China
| | - Z Li
- Jiangxi Provincial Institute of Parasitic Diseases Control, Jiangxi Provincial Key Laboratory of Schistosomiasis Prevention and Control, China
| | - X Wang
- National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research), National Health Commission Key Laboratory of Parasite and Vector Biology, WHO Collaborating Centre for Tropical Diseases, National Center for International Research on Tropical Diseases, Shanghai 200025, China
| | - L Huang
- National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research), National Health Commission Key Laboratory of Parasite and Vector Biology, WHO Collaborating Centre for Tropical Diseases, National Center for International Research on Tropical Diseases, Shanghai 200025, China
| | - R He
- National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research), National Health Commission Key Laboratory of Parasite and Vector Biology, WHO Collaborating Centre for Tropical Diseases, National Center for International Research on Tropical Diseases, Shanghai 200025, China
| | - S Li
- National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research), National Health Commission Key Laboratory of Parasite and Vector Biology, WHO Collaborating Centre for Tropical Diseases, National Center for International Research on Tropical Diseases, Shanghai 200025, China
- School of Global Health, Shanghai Jiao Tong University School of Medicine and Chinese Center for Tropical Diseases Research, Shanghai 200025, China
| |
Collapse
|
21
|
Sun L, Zhu J, Tan J, Li X, Li R, Deng H, Zhang X, Liu B, Zhu X. Deep learning-assisted automated sewage pipe defect detection for urban water environment management. Sci Total Environ 2023; 882:163562. [PMID: 37084915 DOI: 10.1016/j.scitotenv.2023.163562] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Revised: 04/13/2023] [Accepted: 04/13/2023] [Indexed: 05/03/2023]
Abstract
A healthy sewage pipe system plays a significant role in urban water management by collecting and transporting wastewater and stormwater, which can be assessed by hydraulic model. However, sewage pipe defects have been observed frequently in recent years during regular pipe maintenance according to the captured interior videos of underground pipes by closed-circuit television (CCTV) robots. In this case, hydraulic model constructed based on a healthy pipe would produce large deviations with that in real hydraulic performance and even be out of work, which can result in unanticipated damages such as blockage collapse or stormwater overflows. Quick defect evaluation and defect quantification are the precondition to achieve risk assessment and model calibration of urban water management, but currently pipe defects assessment still largely relies on technicians to check the CCTV videos/images. An automated sewage pipe defect detection system is necessary to timely determine pipe issues and then rehabilitate or renew sewage pipes, while the rapid development of deep learning especially in recent five years provides a fantastic opportunity to construct automated pipe defect detection system by image recognition. Given the initial success of deep learning application in CCTV interpretation, the review (i) integrated the methodological framework of automated sewage pipe defect detection, including data acquisition, image pre-processing, feature extraction, model construction and evaluation metrics, (ii) discussed the state-of-the-art performance of deep learning in pipe defects classification, location, and severity rating evaluation (e.g., up to ~96 % of accuracy and 140 FPS of processing speed), and (iii) proposed risk assessment and model calibration in urban water management by considering pipe defects. This review introduces a novel practical application-oriented methodology including defect data acquisition by CCTV, model construction by deep learning, and model application, provides references for further improving accuracy and generalization ability of urban water management models in practical application.
Collapse
Affiliation(s)
- Lianpeng Sun
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China; Guangdong Provincial Key Laboratory of Environmental Pollution Control and Remediation Technology, Sun Yat-sen University, Guangzhou 510275, China
| | - Jinjun Zhu
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China
| | - Jinxin Tan
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China
| | - Xianfeng Li
- School of Computer Science and Engineering, Macau University of Science and Technology, Macau
| | - Ruohong Li
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China; Guangdong Provincial Key Laboratory of Environmental Pollution Control and Remediation Technology, Sun Yat-sen University, Guangzhou 510275, China
| | - Huanzhong Deng
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China
| | - Xinyang Zhang
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China
| | - Bingyou Liu
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China
| | - Xinzhe Zhu
- School of Environmental Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China; Guangdong Provincial Key Laboratory of Environmental Pollution Control and Remediation Technology, Sun Yat-sen University, Guangzhou 510275, China.
| |
Collapse
|
22
|
Hofstra G, van Abeelen H, Duindam M, Houben B, Kuijpers J, Arendsen T, van der Kolk M, Rapp F, van Spaendonk J, Gonzales JL, Petie R. Automated monitoring and detection of disease using a generic facial feature scoring system - A case study on FMD infected cows. Prev Vet Med 2023; 213:105880. [PMID: 36841043 DOI: 10.1016/j.prevetmed.2023.105880] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2022] [Revised: 02/15/2023] [Accepted: 02/17/2023] [Indexed: 02/19/2023]
Abstract
Digital images are becoming more readily available and possibilities for image processing are developing rapidly. This opens the possibility to use digital images to monitor and detect diseases in animals. In this paper we present 1) a generic facial feature scoring system based on seven facial features, 2) manual scores of images of Holstein Frisian heifers during foot-and-mouth disease vaccine efficacy trials and 3) automatic disease scores of the same animals. The automatic scoring system was based on the manual version and trained on annotated images from the manual scoring system. For both systems we found an increase in disease scores three days post infection, followed by a recovery. This temporal pattern matched with observations made by animal caretakers. Importantly, the automatic system was able to discern animals that were protected by the vaccine, and did not develop blisters at the feet, and animals that were not protected. Finally, automatic scores could be used to detect healthy and sick animals with a sensitivity and specificity of 0.94 on the second and third days following infection in an experimental setting. This generic facial feature disease scoring system could be further developed and extended to lactating Holstein Frisian dairy cows, other breeds and other infectious diseases. The system could be applied during animal experiments or, after further development, in a farm setting.
Collapse
Affiliation(s)
- Gerben Hofstra
- HAS University of Applied Science, Onderwijsboulevard 221, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Hilde van Abeelen
- HAS University of Applied Science, Onderwijsboulevard 221, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Marleen Duindam
- HAS University of Applied Science, Onderwijsboulevard 221, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Bas Houben
- HAS University of Applied Science, Onderwijsboulevard 221, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Joris Kuijpers
- HAS University of Applied Science, Onderwijsboulevard 221, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Tim Arendsen
- AVANS University of Applied Science, Onderwijsboulevard 215, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Mathijs van der Kolk
- AVANS University of Applied Science, Onderwijsboulevard 215, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Felix Rapp
- AVANS University of Applied Science, Onderwijsboulevard 215, 5223 DE 's-Hertogenbosch, the Netherlands
| | - Jessy van Spaendonk
- AVANS University of Applied Science, Onderwijsboulevard 215, 5223 DE 's-Hertogenbosch, the Netherlands
| | - José L Gonzales
- Epidemiology Bioinformatics and Animal Models, Wageningen Bioveterinary Research, Houtribweg 39, 8221 RA Lelystad, the Netherlands
| | - Ronald Petie
- Epidemiology Bioinformatics and Animal Models, Wageningen Bioveterinary Research, Houtribweg 39, 8221 RA Lelystad, the Netherlands.
| |
Collapse
|
23
|
Shirabayashi JV, Braga ASM, da Silva J. Comparative approach to different convolutional neural network (CNN) architectures applied to human behavior detection. Neural Comput Appl 2023; 35:12915-12925. [PMID: 37192937 PMCID: PMC9996550 DOI: 10.1007/s00521-023-08430-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2022] [Accepted: 02/20/2023] [Indexed: 03/11/2023]
Abstract
Medical diagnostics, product classification, surveillance and detection of inappropriate behavior are becoming increasingly sophisticated due to the development of methods based on image analysis using neural networks. Considering this, in this work, we evaluate state-of-the-art convolutional neural network architectures proposed in recent years to classify the driving behavior and distractions of drivers. Our main goal is to measure the performance of such architectures using only free resources (i.e., free graphic processing unit, open source) and to evaluate how much of this technological evolution is available to regular users.
Collapse
Affiliation(s)
- Juliana Verga Shirabayashi
- Advanced Campus Jandaia do Sul, Federal University of Paraná, Rua Dr. João Maximiano, 426, Jandaia do Sul, PR 86900000 Brazil
| | | | - Jair da Silva
- Advanced Campus Jandaia do Sul, Federal University of Paraná, Rua Dr. João Maximiano, 426, Jandaia do Sul, PR 86900000 Brazil
| |
Collapse
|
24
|
Kappel S, Ramirez Montes De Oca MA, Collins S, Herborn K, Mendl M, Fureix C. Do you see what I see? Testing horses' ability to recognise real-life objects from 2D computer projections. Anim Cogn 2023:10.1007/s10071-023-01761-6. [PMID: 36864246 PMCID: PMC9980859 DOI: 10.1007/s10071-023-01761-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2022] [Revised: 02/02/2023] [Accepted: 02/20/2023] [Indexed: 03/04/2023]
Abstract
The use of 2-dimensional representations (e.g. photographs or digital images) of real-life physical objects has been an important tool in studies of animal cognition. Horses are reported to recognise objects and individuals (conspecifics and humans) from printed photographs, but it is unclear whether image recognition is also true for digital images, e.g. computer projections. We expected that horses trained to discriminate between two real-life objects would show the same learnt response to digital images of these objects indicating that the images were perceived as objects, or representations of such. Riding-school horses (N = 27) learnt to touch one of two objects (target object counterbalanced between horses) to instantly receive a food reward. After discrimination learning (three consecutive sessions of 8/10 correct trials), horses were immediately tested with on-screen images of the objects over 10 image trials interspersed with five real object trials. At first image presentation, all but two horses spontaneously responded to the images with the learnt behaviour by contacting one of the two images, but the number of horses touching the correct image was not different from chance (14/27 horses, p > 0.05). Only one horse touched the correct image above chance level across 10 image trials (9/10 correct responses, p = 0.021). Our findings thus question whether horses recognise real-life objects from digital images. We discuss how methodological factors and individual differences (i.e. age, welfare state) might have influenced animals' response to the images, and the importance of validating the suitability of stimuli of this kind for cognitive studies in horses.
Collapse
Affiliation(s)
- Sarah Kappel
- School of Biological and Marine Sciences, University of Plymouth, Portland Square, Drake Circus, Plymouth, PL4 8AA, UK.
| | | | - Sarah Collins
- School of Biological and Marine Sciences, University of Plymouth, Portland Square, Drake Circus, Plymouth, PL4 8AA, UK
| | - Katherine Herborn
- School of Biological and Marine Sciences, University of Plymouth, Portland Square, Drake Circus, Plymouth, PL4 8AA, UK
| | - Michael Mendl
- Bristol Veterinary School, University of Bristol, Langford House, Langford, BS40 5DU, UK
| | - Carole Fureix
- School of Biological and Marine Sciences, University of Plymouth, Portland Square, Drake Circus, Plymouth, PL4 8AA, UK
- Bristol Veterinary School, University of Bristol, Langford House, Langford, BS40 5DU, UK
| |
Collapse
|
25
|
Zhou H, Liu Z, Li T, Chen Y, Huang W, Zhang Z. Classification of precancerous lesions based on fusion of multiple hierarchical features. Comput Methods Programs Biomed 2023; 229:107301. [PMID: 36516661 DOI: 10.1016/j.cmpb.2022.107301] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/07/2022] [Revised: 12/01/2022] [Accepted: 12/05/2022] [Indexed: 06/17/2023]
Abstract
PURPOSE To investigate an identification method for precancerous gastric cancer based on the fusion of superficial features and deep features of gastroscopic images. The purpose of this study is to make most use of superficial features and deep features to provide clinicians with clinical decision support to assist the diagnosis of precancerous gastric diseases and reduce the workload of doctors. METHODS According to the nature of gastroscopic images, 75-dimensional shallow features were manually designed, including histogram features, texture features and high-order features of the image; then, based on the constructed convolutional neural networks such as ResNet and GoogLeNet, before the output layer. A fully connected layer is added as the deep feature of the image. In order to ensure consistent feature weights, the number of neurons in the fully connected layer is designed to be 75 dimensions. Therefore, the superficial and deep features of the image are concatenated, and a machine learning classifier is used to identify gastric polyps, there are three types of gastric precancerous diseases such as gastric polyps, gastric ulcers and gastric erosions. RESULTS A dataset with 420 images was collected for each disease, and divided into a training set and a test set with a ratio of 5:1, and then based on the dataset, three methods, such as traditional machine learning, deep learning, and feature fusion, were used respectively. For model training and testing of traditional machine learning and feature fusion, SVM, RF and BP neural network are used as the classification results of the classifier. For deep learning, the GoogLeNet, ResNet, and ResNeXt were implemented. The test results of the model on the test set show that the recognition accuracy of the proposed feature fusion method reaches (SVM: 85.18%; RF: 83.42%; BPNN: 85.18%), which is better than the traditional machine learning method (SVM: 80.17%; RF: 82.37%; BPNN: 84.12%) and the deep learning method (GoogLeNet: 82.54%; ResNet-18: 81.67%; ResNet-50: 81.67%; ResNeXt-50: 82.11%), which proves that this method has obvious advantages. CONCLUSION This study provides a new strategy for the identification of precancerous gastric cancer, improving the efficiency and accuracy of precancerous gastric cancer identification, and hopes to provide substantial practical help for the identification of gastric precancerous diseases.
Collapse
Affiliation(s)
- Huijun Zhou
- Department of Gastroenterology and Urology, Hunan Cancer Hospital/The Affiliated Cancer Hospital of Xiangya School of Medicine, Central South University, Changsha, Hunan 410013, China
| | - Zhenyang Liu
- Department of Gastroenterology and Urology, Hunan Cancer Hospital/The Affiliated Cancer Hospital of Xiangya School of Medicine, Central South University, Changsha, Hunan 410013, China
| | - Ting Li
- Department of Gastroenterology and Urology, Hunan Cancer Hospital/The Affiliated Cancer Hospital of Xiangya School of Medicine, Central South University, Changsha, Hunan 410013, China
| | - Yifei Chen
- Department of Endoscopic Diagnosis and Treatment Center, Hunan Cancer Hospital/The Affiliated Cancer Hospital of Xiangya School of Medicine, Central South University, Changsha, Hunan 410013, China
| | - Wei Huang
- Department of Radiation Oncology, Xiangya Hospital, Central South University, Changsha 410008, China; National Clinical Research Center of Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, Hunan 410008, China; Research Center of Carcinogenesis and Targeted Therapy, Xiangya Hospital, Central South University, Changsha, Hunan, China.
| | - Zijian Zhang
- Department of Radiation Oncology, Xiangya Hospital, Central South University, Changsha 410008, China; National Clinical Research Center of Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, Hunan 410008, China.
| |
Collapse
|
26
|
Das A, Khan DZ, Hanrahan JG, Marcus HJ, Stoyanov D. Automatic generation of operation notes in endoscopic pituitary surgery videos using workflow recognition. Intell Based Med 2023; 8:100107. [PMID: 38523618 PMCID: PMC10958393 DOI: 10.1016/j.ibmed.2023.100107] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Revised: 04/20/2023] [Accepted: 07/27/2023] [Indexed: 03/26/2024]
Abstract
Operation notes are a crucial component of patient care. However, writing them manually is prone to human error, particularly in high pressured clinical environments. Automatic generation of operation notes from video recordings can alleviate some of the administrative burdens, improve accuracy, and provide additional information. To achieve this for endoscopic pituitary surgery, 27-steps were identified via expert consensus. Then, for the 97-videos recorded for this study, a timestamp of each step was annotated by an expert surgeon. To automatically determine whether a step is present in a video, a three-stage architecture was created. Firstly, for each step, a convolution neural network was used for binary image classification on each frame of a video. Secondly, for each step, the binary frame classifications were passed to a discriminator for binary video classification. Thirdly, for each video, the binary video classifications were passed to an accumulator for multi-label step classification. The architecture was trained on 77-videos, and tested on 20-videos, where a 0.80 weighted-F1 score was achieved. The classifications were inputted into a clinically based predefined template, and further enriched with additional video analytics. This work therefore demonstrates automatic generation of operative notes from surgical videos is feasible, and can assist surgeons during documentation.
Collapse
Affiliation(s)
- Adrito Das
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, United Kingdom
| | - Danyal Z. Khan
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, United Kingdom
- National Hospital for Neurology and Neurosurgery, University College London, United Kingdom
| | - John G. Hanrahan
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, United Kingdom
- National Hospital for Neurology and Neurosurgery, University College London, United Kingdom
| | - Hani J. Marcus
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, United Kingdom
- National Hospital for Neurology and Neurosurgery, University College London, United Kingdom
| | - Danail Stoyanov
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, United Kingdom
| |
Collapse
|
27
|
Bao XX, Zhao C, Bao SS, Rao JS, Yang ZY, Li XG. Recognition of necrotic regions in MRI images of chronic spinal cord injury based on superpixel. Comput Methods Programs Biomed 2023; 228:107252. [PMID: 36434959 DOI: 10.1016/j.cmpb.2022.107252] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/07/2020] [Revised: 08/15/2022] [Accepted: 11/17/2022] [Indexed: 06/16/2023]
Abstract
BACKGROUND AND OBJECTIVE The cystic cavity and its surrounding dense glial scar formed in chronic spinal cord injury (SCI) hinder the regeneration of nerve axons. Accurate location of the necrotic regions formed by the scar and the cavity is conducive to eliminate the re-growth obstacles and promote SCI treatment. This work aims to realize the accurate and automatic location of necrotic regions in the chronic SCI magnetic resonance imaging (MRI). METHODS In this study, a method based on superpixel is proposed to identify the necrotic regions of spinal cord in chronic SCI MRI. Superpixels were obtained by a simple linear iterative clustering algorithm, and feature sets were constructed from intensity statistical features, gray level co-occurrence matrix features, Gabor texture features, local binary pattern features and superpixel areas. Subsequently, the recognition effects of support vector machine (SVM) and random forest (RF) classification model on necrotic regions were compared from accuracy (ACC), positive predictive value (PPV), sensitivity (SE), specificity (SP), Dice coefficient and algorithm running time. RESULTS The method is evaluated on T1- and T2-weighted MRI spinal cord images of 24 adult female Wistar rats. And an automatic recognition method for spinal cord necrosis regions was established based on the SVM classification model finally. The recognition results were 1.00±0.00 (ACC), 0.89±0.09 (PPV), 0.88±0.12 (SE), 1.00±0.00 (SP) and 0.88±0.07 (Dice), respectively. CONCLUSIONS The proposed method can accurately and noninvasively identify the necrotic regions in MRI, which is helpful for the pre-intervention assessment and post-intervention evaluation of chronic SCI research and treatments, and promoting the clinical transformation of chronic SCI research.
Collapse
Affiliation(s)
- Xing-Xing Bao
- Beijing Key Laboratory for Biomaterials and Neural Regeneration, Beijing Advanced Innovation Center for Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China
| | - Can Zhao
- Institute of Rehabilitation Engineering, China Rehabilitation Science Institute, Beijing 100068, China.
| | - Shu-Sheng Bao
- Beijing Key Laboratory for Biomaterials and Neural Regeneration, Beijing Advanced Innovation Center for Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China
| | - Jia-Sheng Rao
- Beijing Key Laboratory for Biomaterials and Neural Regeneration, Beijing Advanced Innovation Center for Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China.
| | - Zhao-Yang Yang
- Department of Neurobiology, School of Basic Medical Sciences, Capital Medical University, Beijing 100069, China
| | - Xiao-Guang Li
- Department of Neurobiology, School of Basic Medical Sciences, Capital Medical University, Beijing 100069, China
| |
Collapse
|
28
|
Chen J, Deng X, Wen Y, Chen W, Zeb A, Zhang D. Weakly-supervised learning method for the recognition of potato leaf diseases. Artif Intell Rev 2022; 56:1-18. [PMID: 36573133 PMCID: PMC9771599 DOI: 10.1007/s10462-022-10374-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/10/2022] [Indexed: 12/24/2022]
Abstract
As a crucial food crop, potatoes are highly consumed worldwide, while they are also susceptible to being infected by diverse diseases. Early detection and diagnosis can prevent the epidemic of plant diseases and raise crop yields. To this end, this study proposed a weakly-supervised learning approach for the identification of potato plant diseases. The foundation network was applied with the lightweight MobileNet V2, and to enhance the learning ability for minute lesion features, we modified the existing MobileNet-V2 architecture using the fine-tuning approach conducted by transfer learning. Then, the atrous convolution along with the SPP module was embedded into the pre-trained networks, which was followed by a hybrid attention mechanism containing channel attention and spatial attention submodules to efficiently extract high-dimensional features of plant disease images. The proposed approach outperformed other compared methods and achieved a superior performance gain. It realized an average recall rate of 91.99% for recognizing potato disease types on the publicly accessible dataset. In practical field scenarios, the proposed approach separately attained an average accuracy and specificity of 97.33% and 98.39% on the locally collected image dataset. Experimental results present a competitive performance and demonstrate the validity and feasibility of the proposed approach.
Collapse
Affiliation(s)
- Junde Chen
- Dale E. and Sarah Ann Fowler School of Engineering, Chapman University, Orange, CA 92866 USA
- School of Informatics, Xiamen University, Xiamen, 361005 China
- Department of Electronic Commerce, Xiangtan University, Xiangtan, 411105 China
| | - Xiaofang Deng
- National Academy of Forestry and Grassland Administration, Beijing, 102600 China
| | - Yuxin Wen
- Dale E. and Sarah Ann Fowler School of Engineering, Chapman University, Orange, CA 92866 USA
| | - Weirong Chen
- Department of Information and Electrical Engineering, Ningde Normal University, Ningde, 352100 China
| | - Adnan Zeb
- School of Informatics, Xiamen University, Xiamen, 361005 China
- College of Engineering, Southern University of Science and Technology, Shenzhen, 518000 China
| | - Defu Zhang
- School of Informatics, Xiamen University, Xiamen, 361005 China
| |
Collapse
|
29
|
Olorunfemi BO, Nwulu NI, Ogbolumani OA. Solar panel surface dirt detection and removal based on arduino color recognition. MethodsX 2022; 10:101967. [PMID: 36593762 PMCID: PMC9803813 DOI: 10.1016/j.mex.2022.101967] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2022] [Accepted: 12/11/2022] [Indexed: 12/14/2022] Open
Abstract
Color sensing is a technique for identifying physical changes in materials based on appearance assessment. Dirt deposition on solar panels can change their physical appearance and performance. Considering that dirt accumulation on solar panels needs monitoring to make efficient cleaning schedules, reduce unnecessary costs, and optimize solar panel output generation. Color sensing can achieve fast, accurate, and economical dirt detection, unlike the use of robotic cameras, mathematical formulae, and considering varying output current and voltage methods. Here, we introduce a method that detects and removes dirt on solar panels based on TCS3200 and Arduino Uno components. The approach targets (i.) Panel color measurement, calibration, threshold selection process, (ii.) comparison of color measurement values, and (iii.) align further calibration in response to discoloration of solar panels. This method aims to correct the dirt detection methods previously in use. Hence, a high-speed rolling brush arrangement is designed to improve the cleaning of the solar panel without using water. Further investigations of the panel's color may require some improvement in terms of increasing the sensitivity of the color sensor even with increased distance from the solar panel. Combining multiple color sensors may also be necessary.
Collapse
|
30
|
Li H, Zeng N, Wu P, Clawson K. Cov-Net: A computer-aided diagnosis method for recognizing COVID-19 from chest X-ray images via machine vision. Expert Syst Appl 2022; 207:118029. [PMID: 35812003 PMCID: PMC9252868 DOI: 10.1016/j.eswa.2022.118029] [Citation(s) in RCA: 35] [Impact Index Per Article: 17.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/06/2021] [Revised: 06/17/2022] [Accepted: 06/29/2022] [Indexed: 05/05/2023]
Abstract
In the context of global pandemic Coronavirus disease 2019 (COVID-19) that threatens life of all human beings, it is of vital importance to achieve early detection of COVID-19 among symptomatic patients. In this paper, a computer aided diagnosis (CAD) model Cov-Net is proposed for accurate recognition of COVID-19 from chest X-ray images via machine vision techniques, which mainly concentrates on powerful and robust feature learning ability. In particular, a modified residual network with asymmetric convolution and attention mechanism embedded is selected as the backbone of feature extractor, after which skip-connected dilated convolution with varying dilation rates is applied to achieve sufficient feature fusion among high-level semantic and low-level detailed information. Experimental results on two public COVID-19 radiography databases have demonstrated the practicality of proposed Cov-Net in accurate COVID-19 recognition with accuracy of 0.9966 and 0.9901, respectively. Furthermore, within same experimental conditions, proposed Cov-Net outperforms other six state-of-the-art computer vision algorithms, which validates the superiority and competitiveness of Cov-Net in building highly discriminative features from the perspective of methodology. Hence, it is deemed that proposed Cov-Net has a good generalization ability so that it can be applied to other CAD scenarios. Consequently, one can conclude that this work has both practical value in providing reliable reference to the radiologist and theoretical significance in developing methods to build robust features with strong presentation ability.
Collapse
Affiliation(s)
- Han Li
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361102, China
| | - Nianyin Zeng
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361102, China
| | - Peishu Wu
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361102, China
| | - Kathy Clawson
- School of Computer Science, University of Sunderland, Saint Peter Campus, United Kingdom
| |
Collapse
|
31
|
Lu W, Chen J. Computer vision for solid waste sorting: A critical review of academic research. Waste Manag 2022; 142:29-43. [PMID: 35172271 DOI: 10.1016/j.wasman.2022.02.009] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/12/2021] [Revised: 12/12/2021] [Accepted: 02/06/2022] [Indexed: 06/14/2023]
Abstract
Waste sorting is highly recommended for municipal solid waste (MSW) management. Increasingly, computer vision (CV), robotics, and other smart technologies are used for MSW sorting. Particularly, the field of CV-enabled waste sorting is experiencing an unprecedented explosion of academic research. However, little attention has been paid to understanding its evolvement path, status quo, and prospects and challenges ahead. To address the knowledge gap, this paper provides a critical review of academic research that focuses on CV-enabled MSW sorting. Prevalent CV algorithms, in particular their technical rationales and prediction performance, are introduced and compared. The distribution of academic research outputs is also examined from the aspects of waste sources, task objectives, application domains, and dataset accessibility. The review discovers a trend of shifting from traditional machine learning to deep learning algorithms. The robustness of CV for waste sorting is increasingly enhanced owing to the improved computation powers and algorithms. Academic studies were unevenly distributed in different sectors such as household, commerce and institution, and construction. Too often, researchers reported some preliminary studies using simplified environments and artificially collected data. Future research efforts are encouraged to consider the complexities of real-world scenarios and implement CV in industrial waste sorting practice. This paper also calls for open sharing of waste image datasets for interested researchers to train and evaluate their CV algorithms.
Collapse
Affiliation(s)
- Weisheng Lu
- Department of Real Estate and Construction, The University of Hong Kong, Pokfulam Road, Hong Kong, China
| | - Junjie Chen
- Department of Real Estate and Construction, The University of Hong Kong, Pokfulam Road, Hong Kong, China.
| |
Collapse
|
32
|
Cheng MJ, Sia WL, Liao FC, Chang LS. Adaptation capability of rainfall hotspots in water resilient cities using QGIS: a case study of Taichung City in Taiwan. Environ Monit Assess 2022; 194:219. [PMID: 35201445 DOI: 10.1007/s10661-022-09859-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/16/2021] [Accepted: 02/05/2022] [Indexed: 06/14/2023]
Abstract
In the context of extreme climate due to global climate transition, rainwater adaptation in resilient cities is a key issue for countries. The purpose of this study is to identify the rainfall hotspots in urban areas and investigate whether these hotspots have environmental conditions for rainfall adaptation. The study site is located in the Taichung area. This study collects rainfall data from rainfall stations at elevations below 500 m, employs QGIS (quantum GIS) to create an inverse distance weighted graphical distribution of rainfall to determine the hotspots where the maximum and minimum rainfalls occur, identifies the topography, green spaces, water areas, and buildings in the catchment, integrates the coverage area in the project, and estimates the amount of rainwater that could be directly absorbed by the land within the catchment. The results of this study show that, among the rainfall stations at an elevation below 100 m where most urban areas are located, the Taichung rainfall station is the area with the highest number of rainfall events from May to August. Without reliance on gully or river drainage, the natural infiltration of the land in the catchment could only adjust to 80 mm of heavy precipitation within 24 h of the rainfall warning level of the Central Weather Bureau.
Collapse
Affiliation(s)
- Ming-Jen Cheng
- School of Architecture, Feng Chia University, Taichung City 407, Taiwan
| | - Wei-Liang Sia
- Ph.D. Program for Civil Engineering, Water Resources Engineering, and Infrastructure Planning, Xitun District, Feng Chia University, No. 100 Wenhua Road, Taichung City, 407802, Taiwan.
| | - Feng-Chi Liao
- Department of Spatial Design, Kun Shan University, Tainan City 710, Taiwan
| | - Li-Shin Chang
- School of Architecture, Feng Chia University, Taichung City 407, Taiwan
| |
Collapse
|
33
|
Peng LQ, Guo YC, Wan L, Liu TA, Wang P, Zhao H, Wang YH. Forensic bone age estimation of adolescent pelvis X-rays based on two-stage convolutional neural network. Int J Legal Med 2022. [PMID: 35039894 DOI: 10.1007/s00414-021-02746-1] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2021] [Accepted: 11/10/2021] [Indexed: 12/20/2022]
Abstract
In the forensic estimation of bone age, the pelvis is important for identifying the bone age of teenagers. However, studies on this topic remain insufficient as a result of lower accuracy due to the overlapping of pelvic organs in X-ray images. Segmentation networks have been used to automate the location of key pelvic areas and minimize restrictions like doubling images of pelvic organs to increase the accuracy of estimation. This study conducted a retrospective analysis of 2164 pelvis X-ray images of Chinese Han teenagers ranging from 11 to 21 years old. Key areas of the pelvis were detected with a U-Net segmentation network, and the findings were combined with the original X-ray image for regional augmentation. Bone age estimation was conducted with the enhanced and not enhanced pelvis X-ray images by separately using three convolutional neural networks (CNNs). The root mean square errors (RMSE) of the Inception-V3, Inception-ResNet-V2, and VGG19 convolutional neural networks were 0.93 years, 1.12 years, and 1.14 years, respectively, and the mean absolute errors (MAE) of these networks were 0.67 years, 0.77 years, and 0.88 years, respectively. For comparison, a network without segmentation was employed to conduct the estimation, and it was found that the RMSE of the three CNNs above became 1.22 years, 1.25 years, and 1.63 years, respectively, and the MAE became 0.93 years, 0.96 years, and 1.23 years. Bland-Altman plots and attention maps were also generated to provide a visual comparison. The proposed segmentation network can be used to reduce the influence of restrictions like image overlapping of organs and can thus increase the accuracy of pelvic bone age estimation.
Collapse
|
34
|
Khanzhina N, Filchenkov A, Minaeva N, Novoselova L, Petukhov M, Kharisova I, Pinaeva J, Zamorin G, Putin E, Zamyatina E, Shalyto A. Combating data incompetence in pollen images detection and classification for pollinosis prevention. Comput Biol Med 2022; 140:105064. [PMID: 34861642 DOI: 10.1016/j.compbiomed.2021.105064] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2021] [Revised: 11/20/2021] [Accepted: 11/20/2021] [Indexed: 11/30/2022]
Abstract
Automatic pollen images recognition is crucial for pollinosis symptoms prevention and treatment. The problem of pollen recognition can be efficiently solved using deep learning, however neural networks require tens of thousands of images to generalize. At the same time, the existing open pollen images datasets are very small. In this paper, we present a novel open pollen dataset annotated for both detection and classification tasks. Based on our dataset we study learning from a small data using different state-of-the-art approaches. For the detection task we propose to use our new Bayesian RetinaNet network, which models aleatoric uncertainty. We compare it with the baseline RetinaNet and demonstrate that our model allows for higher detection precision. For the classification task we compare the impact of pre-training on the synthetic images from generative adversarial networks (GANs) and metric-based few-shot learning. Namely, we pre-trained our small convolutional neural network and Siamese neural network classifiers on synthetic pollen images generated by two GANs: StyleGAN and Self-attention GAN. The best classifier is the convolutional neural network pre-trained on StyleGAN images. Our best models achieved 96.3% of mean average precision for the detection task and 97.7% of F1 measure for the classification task on 13 pollen plant species. We have implemented the best models in our pollen recognition web service, which is available for palynologists on request.
Collapse
Affiliation(s)
- Natalia Khanzhina
- Machine Learning Lab, ITMO University, 49 Kronverksky Pr., Saint Petersburg, 197 101, Russia.
| | - Andrey Filchenkov
- Machine Learning Lab, ITMO University, 49 Kronverksky Pr., Saint Petersburg, 197 101, Russia
| | - Natalia Minaeva
- Perm State Medical University, 26 Petropavlovskaya St., Perm, 614 000, Russia
| | - Larisa Novoselova
- Perm State National Research University, 15 Bukireva St., Perm, 614 990, Russia
| | - Maxim Petukhov
- Machine Learning Lab, ITMO University, 49 Kronverksky Pr., Saint Petersburg, 197 101, Russia
| | - Irina Kharisova
- Perm State National Research University, 15 Bukireva St., Perm, 614 990, Russia
| | - Julia Pinaeva
- Perm State National Research University, 15 Bukireva St., Perm, 614 990, Russia
| | - Georgiy Zamorin
- Perm State National Research University, 15 Bukireva St., Perm, 614 990, Russia
| | - Evgeny Putin
- Insilico Medicine Hong Kong Ltd, Pak Shek Kok, New Territories, Hong Kong
| | - Elena Zamyatina
- Perm State National Research University, 15 Bukireva St., Perm, 614 990, Russia; National Research University "Higher School of Economics", Faculty of Economics, Management, and Business Informatics, 38 Studencheskaya St., 614 070, Perm, Russia
| | - Anatoly Shalyto
- Machine Learning Lab, ITMO University, 49 Kronverksky Pr., Saint Petersburg, 197 101, Russia
| |
Collapse
|
35
|
Zhang Q, Yang Q, Zhang X, Bao Q, Su J, Liu X. Waste image classification based on transfer learning and convolutional neural network. Waste Manag 2021; 135:150-157. [PMID: 34509053 DOI: 10.1016/j.wasman.2021.08.038] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/08/2021] [Revised: 08/24/2021] [Accepted: 08/26/2021] [Indexed: 06/13/2023]
Abstract
The rapid economic and social development has led to a rapid increase in the output of domestic waste. How to realize waste classification through intelligent methods has become a key factor for human beings to achieve sustainable development. Traditional waste classification technology has low efficiency and low accuracy. To improve the efficiency and accuracy of waste classification processing, this paper proposes a DenseNet169 waste image classification model based on transfer learning. Because of the disadvantages of the existing public waste dataset, such as uneven distribution of data, single background, obvious features, and small sample size of the waste image, the waste image dataset NWNU-TRASH is constructed. The dataset has the advantages of balanced distribution, high diversity, and rich background, which is more in line with real needs. 70% of the dataset is used as the training set and 30% as the test set. Based on the deep learning network DenseNet169 pre-trained model, we can form a DenseNet169 model suitable for this experimental dataset. The experimental results show that the accuracy of classification is over 82% in the DenseNet169 model after the transfer learning, which is better than other image classification algorithms.
Collapse
Affiliation(s)
- Qiang Zhang
- Department of Computer Science and Engineering, Northwest Normal University, Lanzhou, Gansu Province 730070, China
| | - Qifan Yang
- Department of Computer Science and Engineering, Northwest Normal University, Lanzhou, Gansu Province 730070, China
| | - Xujuan Zhang
- School of Computer Science and Artificial Intelligence, Lanzhou Institute of Technology, Lanzhou, Gansu Province 730050, China
| | - Qiang Bao
- College of Computing, Illinois Institute of Technology, Chicago, IL 60616, USA
| | - Jinqi Su
- Xi'an University of Posts&Telecommunications, Xi'an, Shanxi Province 710121, China
| | - Xueyan Liu
- Department of Computer Science and Engineering, Northwest Normal University, Lanzhou, Gansu Province 730070, China.
| |
Collapse
|
36
|
Yasue K, Fuse H, Asano Y, Kato M, Shinoda K, Ikoma H, Fujisaki T, Tamaki Y. Investigation of fiducial marker recognition possibility by water equivalent length in real-time tracking radiotherapy. Jpn J Radiol 2021; 40:318-325. [PMID: 34655387 DOI: 10.1007/s11604-021-01207-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2021] [Accepted: 09/30/2021] [Indexed: 12/28/2022]
Abstract
Real-time tumor tracking radiotherapy (RTRT) systems typically use fiducial markers implanted near the tumor to track the target using X-ray fluoroscopy. Template pattern matching, used in tracking, is often used to automatically localize the fiducial markers. In radiotherapy of the liver, the thickness of the body that can recognize the fiducial markers must be clinically assessed. The purpose of this study was to quantify the recognition of fiducial markers according to body thickness in stereotactic body radiotherapy of the liver using clinical images obtained using SyncTraX FX4. The recognition scores of fiducial markers were examined in relation to water equivalent length (WEL), tube current, and each flat panel detector. The relationship between the contrast ratio of the fiducial marker and the background and the WEL was also investigated. The average recognition score was found to be less than 20 when the WEL was greater than 25 cm. The probability of successful tracking of image recognition was mostly smaller than 0.8 when the WEL was over 30 cm. The relationship between WEL and tube current did not significantly differ between 100 and 140 mA, but there was a significant difference (p < 0.05) for all other combinations. To ensure tracking of fiducial markers during SBRT, if the WEL representing body thickness is longer than 25 cm, the X-ray fluoroscopy arrangement should be determined based on the WEL.
Collapse
Affiliation(s)
- Kenji Yasue
- Graduate School of Health Sciences, Ibaraki Prefectural University of Health Sciences, 4669-2, Ami, Inashiki, Ibaraki, 300-0394, Japan.,Department of Radiation Technology, Ibaraki Prefectural Central Hospital, 6528, Koibuchi, Kasama, Ibaraki, 309-1793, Japan
| | - Hiraku Fuse
- Department of Radiological Sciences, Ibaraki Prefectural University of Health Sciences, 4669-2, Ami, Inashiki, Ibaraki, 300-0394, Japan.
| | - Yuto Asano
- Department of Radiation Technology, Ibaraki Prefectural Central Hospital, 6528, Koibuchi, Kasama, Ibaraki, 309-1793, Japan
| | - Miho Kato
- Department of Radiation Technology, Ibaraki Prefectural Central Hospital, 6528, Koibuchi, Kasama, Ibaraki, 309-1793, Japan
| | - Kazuya Shinoda
- Department of Radiation Technology, Ibaraki Prefectural Central Hospital, 6528, Koibuchi, Kasama, Ibaraki, 309-1793, Japan
| | - Hideaki Ikoma
- Department of Radiation Technology, Ibaraki Prefectural Central Hospital, 6528, Koibuchi, Kasama, Ibaraki, 309-1793, Japan
| | - Tatsuya Fujisaki
- Department of Radiological Sciences, Ibaraki Prefectural University of Health Sciences, 4669-2, Ami, Inashiki, Ibaraki, 300-0394, Japan
| | - Yoshio Tamaki
- Department of Radiation Oncology, Ibaraki Prefectural Central Hospital, 6528, Koibuchi, Kasama, Ibaraki, 309-1793, Japan
| |
Collapse
|
37
|
Meermeier A, Jording M, Alayoubi Y, Vogel DHV, Vogeley K, Tepest R. Brief Report: Preferred Processing of Social Stimuli in Autism: A Perception Task. J Autism Dev Disord 2021; 52:3286-3293. [PMID: 34532839 PMCID: PMC9213359 DOI: 10.1007/s10803-021-05195-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/08/2021] [Indexed: 11/05/2022]
Abstract
In this study we investigate whether persons with autism spectrum disorder (ASD) perceive social images differently than control participants (CON) in a graded perception task in which stimuli emerged from noise before dissipating into noise again. We presented either social stimuli (humans) or non-social stimuli (objects or animals). ASD were slower to recognize images during their emergence, but as fast as CON when indicating the dissipation of the image irrespective of its content. Social stimuli were recognized faster and remained discernable longer in both diagnostic groups. Thus, ASD participants show a largely intact preference for the processing of social images. An exploratory analysis of response subsets reveals subtle differences between groups that could be investigated in future studies.
Collapse
Affiliation(s)
- A Meermeier
- University Hospital Cologne, NRW, Kerpener Strasse 62, Geb. 31, 50931, Cologne, Germany.
| | - M Jording
- University Hospital Cologne, NRW, Kerpener Strasse 62, Geb. 31, 50931, Cologne, Germany.,Forschungszentrum Jülich, INM3, NRW, Wilhelm-Johnen-Straße 1, 52428, Jülich, Germany
| | - Y Alayoubi
- University Hospital Cologne, NRW, Kerpener Strasse 62, Geb. 31, 50931, Cologne, Germany
| | - David H V Vogel
- University Hospital Cologne, NRW, Kerpener Strasse 62, Geb. 31, 50931, Cologne, Germany.,Forschungszentrum Jülich, INM3, NRW, Wilhelm-Johnen-Straße 1, 52428, Jülich, Germany
| | - K Vogeley
- University Hospital Cologne, NRW, Kerpener Strasse 62, Geb. 31, 50931, Cologne, Germany.,Forschungszentrum Jülich, INM3, NRW, Wilhelm-Johnen-Straße 1, 52428, Jülich, Germany
| | - R Tepest
- University Hospital Cologne, NRW, Kerpener Strasse 62, Geb. 31, 50931, Cologne, Germany
| |
Collapse
|
38
|
Yu F, Zhu Y, Qin X, Xin Y, Yang D, Xu T. A multi-class COVID-19 segmentation network with pyramid attention and edge loss in CT images. IET Image Process 2021; 15:2604-2613. [PMID: 34226836 PMCID: PMC8242907 DOI: 10.1049/ipr2.12249] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/18/2020] [Revised: 03/16/2021] [Accepted: 04/23/2021] [Indexed: 05/15/2023]
Abstract
At the end of 2019, a novel coronavirus COVID-19 broke out. Due to its high contagiousness, more than 74 million people have been infected worldwide. Automatic segmentation of the COVID-19 lesion area in CT images is an effective auxiliary medical technology which can quantitatively diagnose and judge the severity of the disease. In this paper, a multi-class COVID-19 CT image segmentation network is proposed, which includes a pyramid attention module to extract multi-scale contextual attention information, and a residual convolution module to improve the discriminative ability of the network. A wavelet edge loss function is also proposed to extract edge features of the lesion area to improve the segmentation accuracy. For the experiment, a dataset of 4369 CT slices is constructed, including three symptoms: ground glass opacities, interstitial infiltrates, and lung consolidation. The dice similarity coefficients of three symptoms of the model achieve 0.7704, 0.7900, 0.8241 respectively. The performance of the proposed network on public dataset COVID-SemiSeg is also evaluated. The results demonstrate that this model outperforms other state-of-the-art methods and can be a powerful tool to assist in the diagnosis of positive infection cases, and promote the development of intelligent technology in the medical field.
Collapse
Affiliation(s)
- Fuli Yu
- School of Information Science and EngineeringEast China University of Science and TechnologyShanghai200237People's Republic of China
| | - Yu Zhu
- School of Information Science and EngineeringEast China University of Science and TechnologyShanghai200237People's Republic of China
| | - Xiangxiang Qin
- School of Information Science and EngineeringEast China University of Science and TechnologyShanghai200237People's Republic of China
| | - Ying Xin
- Department of Endocrine and Metabolic DiseasesThe Affiliated Hospital of Qingdao UniversityQingdao266003People's Republic of China
| | - Dawei Yang
- Department of Pulmonary MedicineZhongshan HospitalFudan UniversityShanghai200032People's Republic of China
| | - Tao Xu
- Department of Pulmonary and Critical Care MedicineThe Affiliated Hospital of Qingdao UniversityQingdaoShandong266000People's Republic of China
| |
Collapse
|
39
|
Dursun G, Tandale SB, Gulakala R, Eschweiler J, Tohidnezhad M, Markert B, Stoffel M. Development of convolutional neural networks for recognition of tenogenic differentiation based on cellular morphology. Comput Methods Programs Biomed 2021; 208:106279. [PMID: 34343743 DOI: 10.1016/j.cmpb.2021.106279] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/06/2020] [Accepted: 07/06/2021] [Indexed: 06/13/2023]
Abstract
BACKGROUND AND OBJECTIVE The use of automated systems for image recognition is highly preferred for regenerative medicine applications to evaluate stem cell differentiation early in the culturing state with non-invasive methodologies instead of invasive counterparts. Bone marrow-derived mesenchymal stem cells (BMSCs) are able to differentiate into desired cell phenotypes, and thereby promise a proper cell source for tendon regeneration. The therapeutic success of stem cell therapy requires cellular characterization prior to the implantation of cells. The foremost problem is that traditional characterization techniques require cellular material which would be more useful for cell therapy, complex laboratory procedures, and human expertise. Convolutional neural networks (CNNs), a class of deep neural networks, have recently made great improvements in image-based classifications, recognition, and detection tasks. We, therefore, aim to develop a potential CNN model in order to recognize differentiated stem cells by learning features directly from image data of unlabelled cells. METHODS The differentiation of bone marrow mesenchymal stem cells (BMSCs) into tenocytes was induced with the treatment of bone morphogenetic protein-12 (BMP-12). Following the treatment and incubation step, the phase-contrast images of cells were obtained and immunofluorescence staining has been applied to characterize the differentiated state of BMSCs. CNN models were developed and trained with the phase-contrast cell images. The comparison of CNN models was performed with respect to prediction performance and training time. Moreover, we have evaluated the effect of image enhancement method, data augmentation, and fine-tuning training strategy to increase classification accuracy of CNN models. The best model was integrated into a mobile application. RESULTS All the CNN models can fit the biological data extracted from immunofluorescence characterization. CNN models enable the cell classification with satisfactory accuracies. The best result in terms of accuracy and training time is achieved by the model proposed based on Inception-ResNet V2 trained from scratch using image enhancement and data augmentation strategies (96.80%, 434.55 sec). CONCLUSION Our study reveals that the CNN models show good performance by identifying stem cell differentiation. Importantly this technique provides a faster and real-time tool in comparison to traditional methods enabling the adjustment of culture conditions during cultivation to improve the yield of therapeutic stem cells.
Collapse
Affiliation(s)
- Gözde Dursun
- Institute of General Mechanics, RWTH Aachen University, Aachen, Germany
| | | | - Rutwik Gulakala
- Institute of General Mechanics, RWTH Aachen University, Aachen, Germany
| | - Jörg Eschweiler
- Department of Orthopaedic Surgery, RWTH Aachen University, Aachen, Germany
| | | | - Bernd Markert
- Institute of General Mechanics, RWTH Aachen University, Aachen, Germany
| | - Marcus Stoffel
- Institute of General Mechanics, RWTH Aachen University, Aachen, Germany.
| |
Collapse
|
40
|
Carloto I, Johnston P, Pestana CJ, Lawton LA. Detection of morphological changes caused by chemical stress in the cyanobacterium Planktothrix agardhii using convolutional neural networks. Sci Total Environ 2021; 784:146956. [PMID: 33894604 DOI: 10.1016/j.scitotenv.2021.146956] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/23/2020] [Revised: 03/31/2021] [Accepted: 03/31/2021] [Indexed: 06/12/2023]
Abstract
The presence of harmful algal bloom in many reservoirs around the world, alongside the lack of sanitation law/ordinance regarding cyanotoxin monitoring (particularly in developing countries), create a scenario in which the local population could potentially chronically consume cyanotoxin-contaminated waters. Therefore, it is crucial to develop low cost tools to detect possible systems failures and consequent toxin release inferred by morphological changes of cyanobacteria in the raw water. This paper aimed to look for the best combination of convolutional neural network (CNN), optimizer and image segmentation technique to differentiate P. agardhii trichomes before and after chemical stress caused by the addition of hydrogen peroxide. This method takes a step towards accurate monitoring of cyanobacteria in the field without the need for a mobile lab. After testing three different network architectures (AlexNet, 3ConvLayer and 2ConvLayer), four different optimizers (Adam, Adagrad, RMSProp and SDG) and five different image segmentations methods (Canny Edge Detection, Morphological Filter, HP filter, GrabCut and Watershed), the combination 2ConvLayer with Adam optimizer and GrabCut segmentation, provided the highest median accuracy (93.33%) for identifying H2O2-induced morphological changes in P. agardhii. Our results emphasize the fact that the trichome classification problem can be adequately tackled with a limited number of learned features due to the lack of complexity in micrographs from before and after chemical stress. To the authors' knowledge, this is the first time that CNNs were applied to detect morphological changes in cyanobacteria caused by chemical stress. Thus, it is a significant step forward in developing low cost tools based on image recognition, to shield water consumers, especially in the poorest regions, against cyanotoxin-contaminated water.
Collapse
Affiliation(s)
- Ismael Carloto
- School of Pharmacy and Life Sciences, Robert Gordon University, Garthdee Road, Aberdeen AB10 7GJ, UK.
| | - Pamela Johnston
- School of Computing, Robert Gordon University, Garthdee Road, Aberdeen AB10 7GJ, UK.
| | - Carlos J Pestana
- School of Pharmacy and Life Sciences, Robert Gordon University, Garthdee Road, Aberdeen AB10 7GJ, UK.
| | - Linda A Lawton
- School of Pharmacy and Life Sciences, Robert Gordon University, Garthdee Road, Aberdeen AB10 7GJ, UK.
| |
Collapse
|
41
|
Latête T, Gauthier B, Belanger P. Towards using convolutional neural network to locate, identify and size defects in phased array ultrasonic testing. Ultrasonics 2021; 115:106436. [PMID: 33873024 DOI: 10.1016/j.ultras.2021.106436] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/20/2020] [Revised: 03/01/2021] [Accepted: 04/05/2021] [Indexed: 06/12/2023]
Abstract
Machine learning algorithms are widely used in image recognition. In Phased Array Ultrasonic Testing (PAUT), images are typically formed through constructive and destructive superpositions of signals backscattered from flaws or geometric features. However, all PAUT data acquisition schemes require several emissions and the duration of the acquisition may be too slow in high-speed manufacturing. In this study, the Faster R-CNN was used to identify, locate and size flat bottom holes (FBH) and side-drilled holes (SDH) in an immersed test specimen using a single plane wave insonification. The training was performed on segmented and classified data generated using GPU-accelerated finite element simulations. SDH and FBH of different diameters, depths and lateral positions were included in the training set. The thickness of the test specimen was also variable. An ultrasonic phased array probe of 64 elements was simulated. All elements of the phased array probe were fired at the same time and the time traces from each element were recorded. The individual time traces were concatenated to form a matrix, which was then used in the training. This inspection scenario enables fast acquisition of data at the expense of poor lateral resolution in the resulting image. The trained neural network was initially tested using finite element simulations. Results were assessed in terms of the intersection of the union (IoU) between the ground truth geometry and the predicted geometry. With the simulated cases, the thickness of the test specimen was detected in all cases. When using a 40% IoU threshold, the detection rate of the FBH was 87% while only 20% for the SDH. The smallest detected FBH had a 0.56 wavelength depth and a lateral extent of 1.04 wavelength. Drawing a box using the -6dB drop method around the FBH always led to an IoU under 15%. On average, the lateral extent of the FBH using the -6dB method was three times larger than the diameter predicted by the proposed method. Then, the training was continued with a small augmented dataset of experiments (equivalent to 3% of the simulated dataset). In experiments, the results show that the test specimen was always correctly identified. When using a 40% IoU threshold the experimental detection rate of the FBH was 70%. The smallest detected defect in experiments had a depth of 2 wavelengths.
Collapse
Affiliation(s)
- Thibault Latête
- PULÉTS, École de technologie supérieure, 1100 Notre-Dame Ouest, Montréal, Québec, Canada, H3C 1K3.
| | - Baptiste Gauthier
- PULÉTS, École de technologie supérieure, 1100 Notre-Dame Ouest, Montréal, Québec, Canada, H3C 1K3
| | - Pierre Belanger
- PULÉTS, École de technologie supérieure, 1100 Notre-Dame Ouest, Montréal, Québec, Canada, H3C 1K3
| |
Collapse
|
42
|
Del Bonifro F, Gabbrielli M, Lategano A, Zacchiroli S. Image-based many-language programming language identification. PeerJ Comput Sci 2021; 7:e631. [PMID: 34825053 PMCID: PMC8592246 DOI: 10.7717/peerj-cs.631] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2021] [Accepted: 06/18/2021] [Indexed: 06/13/2023]
Abstract
Programming language identification (PLI) is a common need in automatic program comprehension as well as a prerequisite for deeper forms of code understanding. Image-based approaches to PLI have recently emerged and are appealing due to their applicability to code screenshots and programming video tutorials. However, they remain limited to the recognition of a small amount of programming languages (up to 10 languages in the literature). We show that it is possible to perform image-based PLI on a large number of programming languages (up to 149 in our experiments) with high (92%) precision and recall, using convolutional neural networks (CNNs) and transfer learning, starting from readily-available pretrained CNNs. Results were obtained on a large real-world dataset of 300,000 code snippets extracted from popular GitHub repositories. By scrambling specific character classes and comparing identification performances we also show that the characters that contribute the most to the visual recognizability of programming languages are symbols (e.g., punctuation, mathematical operators and parentheses), followed by alphabetic characters, with digits and indentation having a negligible impact.
Collapse
|
43
|
Czyzewski A, Krawiec F, Brzezinski D, Porebski PJ, Minor W. Detecting anomalies in X-ray diffraction images using convolutional neural networks. Expert Syst Appl 2021; 174:114740. [PMID: 34366575 PMCID: PMC8341115 DOI: 10.1016/j.eswa.2021.114740] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/09/2023]
Abstract
Our understanding of life is based upon the interpretation of macromolecular structures and their dynamics. Almost 90% of currently known macromolecular models originated from electron density maps constructed using X-ray diffraction images. Even though diffraction images are critical for structure determination, due to their vast amounts and noisy, non-intuitive nature, their quality is rarely inspected. In this paper, we use recent advances in machine learning to automatically detect seven types of anomalies in X-ray diffraction images. For this purpose, we utilize a novel X-ray beam center detection algorithm, propose three different image representations, and compare the predictive performance of general-purpose classifiers and deep convolutional neural networks (CNNs). In benchmark tests on a set of 6,311 X-ray diffraction images, the proposed CNN achieved between 87% and 99% accuracy depending on the type of anomaly. Experimental results show that the proposed anomaly detection system can be considered suitable for early detection of sub-optimal data collection conditions and malfunctions at X-ray experimental stations.
Collapse
Affiliation(s)
- Adam Czyzewski
- Institute of Computing Science, Poznan University of
Technology, ul. Piotrowo 2, 60-965 Poznan, Poland
| | - Faustyna Krawiec
- Institute of Computing Science, Poznan University of
Technology, ul. Piotrowo 2, 60-965 Poznan, Poland
| | - Dariusz Brzezinski
- Institute of Computing Science, Poznan University of
Technology, ul. Piotrowo 2, 60-965 Poznan, Poland
- Center for Biocrystallographic Research, Institute of
Bioorganic Chemistry, Polish Academy of Sciences, Poznan, 61-714, Poland
- Center for Artificial Intelligence and Machine Learning,
Poznan University of Technology, ul. Piotrowo 2, 60-965 Poznan, Poland
- Department of Molecular Physiology and Biological Physics,
University of Virginia, Charlottesville, VA 22901, USA
| | - Przemyslaw Jerzy Porebski
- Department of Molecular Physiology and Biological Physics,
University of Virginia, Charlottesville, VA 22901, USA
| | - Wladek Minor
- Department of Molecular Physiology and Biological Physics,
University of Virginia, Charlottesville, VA 22901, USA
| |
Collapse
|
44
|
Chen S, Hu Z, Wang C, Pang Q, Hua L. Research on the process of small sample non-ferrous metal recognition and separation based on deep learning. Waste Manag 2021; 126:266-273. [PMID: 33789215 DOI: 10.1016/j.wasman.2021.03.019] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/23/2020] [Revised: 02/16/2021] [Accepted: 03/12/2021] [Indexed: 06/12/2023]
Abstract
Consumption of copper and aluminum has increased significantly in recent years; therefore, recycling these elements from the end-of-life vehicles (ELVs) will be of great economic value and social benefit. However, the separation of non-ferrous materials is difficult because of their different sources, various shapes and sizes, and complex surface conditions. In experimental study on the separation of these materials, few non-ferrous metal scraps can be used. To address these limitations, a traditional image recognition model and a small sample multi-target detection model (which can detect multiple targets simultaneously) based on deep learning and transfer learning were used to identify non-ferrous materials. The improved third version of You Only Look Once (YOLOv3) multi-target detection model using data augmentation, the loss function of focal loss, and a method of adjusting the threshold of Intersection over Union (IOU) between candidate bound and ground truth bound has superior target detection performance than methods. We obtained a 95.3% and 91.4% accuracy in identifying aluminum and copper scraps, respectively, and an operation speed of 18 FPS, meeting the real-time requirements of a sorting system. By using the improved YOLOv3 multi-target detection algorithm and equipment operation parameters selected, the accuracy and purity of the separation system exceeded 90%, meeting the needs of actual production.
Collapse
Affiliation(s)
- Song Chen
- Hubei Key Laboratory of Advanced Technology of Automobile Components, Wuhan University of Technology, Wuhan 430070, PR China; Hubei Collaborative Innovation Center for Automotive Components Technology, Wuhan University of Technology, Wuhan 430070, PR China
| | - Zhili Hu
- Hubei Key Laboratory of Advanced Technology of Automobile Components, Wuhan University of Technology, Wuhan 430070, PR China; Hubei Collaborative Innovation Center for Automotive Components Technology, Wuhan University of Technology, Wuhan 430070, PR China.
| | - Chao Wang
- Hubei Key Laboratory of Advanced Technology of Automobile Components, Wuhan University of Technology, Wuhan 430070, PR China; Hubei Collaborative Innovation Center for Automotive Components Technology, Wuhan University of Technology, Wuhan 430070, PR China
| | - Qiu Pang
- Key Laboratory of Metallurgical Equipment and Control Technology of Ministry of Education, Wuhan University of Science and Technology, Wuhan 430081, PR China
| | - Lin Hua
- Hubei Key Laboratory of Advanced Technology of Automobile Components, Wuhan University of Technology, Wuhan 430070, PR China; Hubei Collaborative Innovation Center for Automotive Components Technology, Wuhan University of Technology, Wuhan 430070, PR China.
| |
Collapse
|
45
|
Hwang SW, Sugiyama J. Computer vision-based wood identification and its expansion and contribution potentials in wood science: A review. Plant Methods 2021; 17:47. [PMID: 33910606 PMCID: PMC8082842 DOI: 10.1186/s13007-021-00746-1] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/16/2020] [Accepted: 04/17/2021] [Indexed: 05/23/2023]
Abstract
The remarkable developments in computer vision and machine learning have changed the methodologies of many scientific disciplines. They have also created a new research field in wood science called computer vision-based wood identification, which is making steady progress towards the goal of building automated wood identification systems to meet the needs of the wood industry and market. Nevertheless, computer vision-based wood identification is still only a small area in wood science and is still unfamiliar to many wood anatomists. To familiarize wood scientists with the artificial intelligence-assisted wood anatomy and engineering methods, we have reviewed the published mainstream studies that used or developed machine learning procedures. This review could help researchers understand computer vision and machine learning techniques for wood identification and choose appropriate techniques or strategies for their study objectives in wood science.
Collapse
Affiliation(s)
- Sung-Wook Hwang
- Graduate School of Agriculture, Kyoto University, Sakyo-ku, Kyoto, 606-8502 Japan
| | - Junji Sugiyama
- Graduate School of Agriculture, Kyoto University, Sakyo-ku, Kyoto, 606-8502 Japan
- College of Materials Science and Engineering, Nanjing Forestry University, Nanjing, 210037 China
| |
Collapse
|
46
|
Qiu P, Li Y, Liu K, Qin J, Ye K, Chen T, Lu X. Prescreening and treatment of aortic dissection through an analysis of infinite-dimension data. BioData Min 2021; 14:24. [PMID: 33794946 PMCID: PMC8015064 DOI: 10.1186/s13040-021-00249-8] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2020] [Accepted: 02/14/2021] [Indexed: 01/09/2023] Open
Abstract
Background Aortic dissection (AD) is one of the most catastrophic aortic diseases associated with a high mortality rate. In contrast to the advances in most cardiovascular diseases, both the incidence and in-hospital mortality rate of AD have experienced deviant increases over the past 20 years, highlighting the need for fresh prospects on the prescreening and in-hospital treatment strategies. Methods Through two cross-sectional studies, we adopt image recognition techniques to identify pre-disease aortic morphology for prior diagnoses; assuming that AD has occurred, we employ functional data analysis to determine the optimal timing for BP and HR interventions to offer the highest possible survival rate. Results Compared with the healthy control group, the aortic centerline is significantly more slumped for the AD group. Further, controlling patients’ blood pressure and heart rate according to the likelihood of adverse events can offer the highest possible survival probability. Conclusions The degree of slumpness is introduced to depict aortic morphological changes comprehensively. The morphology-based prediction model is associated with an improvement in the predictive accuracy of the prescreening of AD. The dynamic model reveals that blood pressure and heart rate variations have a strong predictive power for adverse events, confirming this model’s ability to improve AD management. Supplementary Information The online version contains supplementary material available at (10.1186/s13040-021-00249-8).
Collapse
Affiliation(s)
- Peng Qiu
- Department of Vascular Surgery, Shanghai Ninth People's Hospital Affiliated with Shanghai Jiao Tong University School of Medicine, Shanghai, China.,Big Data Research Lab, University of Waterloo, Waterloo, Canada
| | - Yixuan Li
- Big Data Research Lab, University of Waterloo, Waterloo, Canada.,Department of Economics, University of Waterloo, Waterloo, Canada.,Stoppingtime (Shanghai) BigData & Technology Co. Ltd., Shanghai, China
| | - Kai Liu
- Big Data Research Lab, University of Waterloo, Waterloo, Canada.,School of Mathematical and Computational Sciences, University of Prince Edward Island, Charlottetown, Canada
| | - Jinbao Qin
- Department of Vascular Surgery, Shanghai Ninth People's Hospital Affiliated with Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Kaichuang Ye
- Department of Vascular Surgery, Shanghai Ninth People's Hospital Affiliated with Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Tao Chen
- Big Data Research Lab, University of Waterloo, Waterloo, Canada. .,Department of Economics, University of Waterloo, Waterloo, Canada. .,Senior Research Fellow of Labor and Worklife Program, Harvard University, Cambridge, USA.
| | - Xinwu Lu
- Department of Vascular Surgery, Shanghai Ninth People's Hospital Affiliated with Shanghai Jiao Tong University School of Medicine, Shanghai, China.
| |
Collapse
|
47
|
Xie C, Xiang H, Zeng T, Yang Y, Yu B, Liu Q. Cross Knowledge-based Generative Zero-Shot Learning approach with Taxonomy Regularization. Neural Netw 2021; 139:168-178. [PMID: 33721699 DOI: 10.1016/j.neunet.2021.02.009] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2020] [Revised: 01/17/2021] [Accepted: 02/08/2021] [Indexed: 10/22/2022]
Abstract
Although zero-shot learning (ZSL) has an inferential capability of recognizing new classes that have never been seen before, it always faces two fundamental challenges of the cross modality and cross-domain challenges. In order to alleviate these problems, we develop a generative network-based ZSL approach equipped with the proposed Cross Knowledge Learning (CKL) scheme and Taxonomy Regularization (TR). In our approach, the semantic features are taken as inputs, and the output is the synthesized visual features generated from the corresponding semantic features. CKL enables more relevant semantic features to be trained for semantic-to-visual feature embedding in ZSL, while Taxonomy Regularization (TR) significantly improves the intersections with unseen images with more generalized visual features generated from generative network. Extensive experiments on several benchmark datasets (i.e., AwA1, AwA2, CUB, NAB and aPY) show that our approach is superior to these state-of-the-art methods in terms of ZSL image classification and retrieval.
Collapse
Affiliation(s)
- Cheng Xie
- National Pilot School of Software, Yunnan University, Kunming 650091, China
| | - Hongxin Xiang
- National Pilot School of Software, Yunnan University, Kunming 650091, China
| | - Ting Zeng
- National Pilot School of Software, Yunnan University, Kunming 650091, China
| | - Yun Yang
- National Pilot School of Software, Yunnan University, Kunming 650091, China; Kunming Key Laboratory of Data Science and Intelligent Computing, Kunming 650500, China.
| | - Beibei Yu
- National Pilot School of Software, Yunnan University, Kunming 650091, China
| | - Qing Liu
- National Pilot School of Software, Yunnan University, Kunming 650091, China
| |
Collapse
|
48
|
Morishita J, Ueda Y. New solutions for automated image recognition and identification: challenges to radiologic technology and forensic pathology. Radiol Phys Technol 2021; 14:123-133. [PMID: 33710498 DOI: 10.1007/s12194-021-00611-9] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2021] [Revised: 02/26/2021] [Accepted: 02/28/2021] [Indexed: 11/30/2022]
Abstract
This paper outlines the history of biometrics for personal identification, the current status of the initial biological fingerprint techniques for digital chest radiography, and patient verification during medical imaging, such as computed tomography and magnetic resonance imaging. Automated image recognition and identification developed for clinical images without metadata could also be applied to the identification of victims in mass disasters or other unidentified individuals. The development of methods that are adaptive to a wide range of recent imaging modalities in the fields of radiologic technology, patient safety, forensic pathology, and forensic odontology is still in its early stages. However, its importance in practice will continue to increase in the future.
Collapse
Affiliation(s)
- Junji Morishita
- Department of Health Sciences, Faculty of Medical Sciences, Kyushu University, 3-1-1 Maidashi, Higashi-ku, Fukuoka, Fukuoka, 812-8582, Japan.
| | - Yasuyuki Ueda
- Department of Medical Physics and Engineering, Area of Medical Imaging Technology and Science, Division of Health Sciences, Graduate School of Medicine, Osaka University, 1-7 Yamadaoka, Suita, Osaka, 565-0871, Japan.
| |
Collapse
|
49
|
Hosny KM, Darwish MM, Eltoukhy MM. New fractional-order shifted Gegenbauer moments for image analysis and recognition. J Adv Res 2020; 25:57-66. [PMID: 32922974 PMCID: PMC7474242 DOI: 10.1016/j.jare.2020.05.024] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2020] [Accepted: 05/23/2020] [Indexed: 11/20/2022] Open
Abstract
Orthogonal moments are used to represent digital images with minimum redundancy. Orthogonal moments with fractional-orders show better capabilities in digital image analysis than integer-order moments. In this work, the authors present new fractional-order shifted Gegenbauer polynomials. These new polynomials are used to define a novel set of orthogonal fractional-order shifted Gegenbauer moments (FrSGMs). The proposed method is applied in gray-scale image analysis and recognition. The invariances to rotation, scaling and translation (RST), are achieved using invariant fractional-order geometric moments. Experiments are conducted to evaluate the proposed FrSGMs and compare with the classical orthogonal integer-order Gegenbauer moments (GMs) and the existing orthogonal fractional-order moments. The new FrSGMs outperformed GMs and the existing orthogonal fractional-order moments in terms of image recognition and reconstruction, RST invariance, and robustness to noise.
Collapse
Affiliation(s)
- Khalid M Hosny
- Information Technology Department, Faculty of Computers and Informatics, Zagazig University, Zagazig 44519, Egypt
| | - Mohamed M Darwish
- Mathematics Department, Faculty of Science, Assiut University, Assiut 71516, Egypt
| | - Mohamed Meselhy Eltoukhy
- Computer Science Department, Faculty of Computers and Informatics, Suez Canal University, Ismailia, Egypt.,College of Computing and Information Technology, Khulais, University of Jeddah, Saudi Arabia
| |
Collapse
|
50
|
Franzoni V, Biondi G, Milani A. Emotional sounds of crowds: spectrogram-based analysis using deep learning. Multimed Tools Appl 2020; 79:36063-36075. [PMID: 32837250 PMCID: PMC7429201 DOI: 10.1007/s11042-020-09428-x] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/19/2020] [Revised: 06/10/2020] [Accepted: 07/16/2020] [Indexed: 06/11/2023]
Abstract
Crowds express emotions as a collective individual, which is evident from the sounds that a crowd produces in particular events, e.g., collective booing, laughing or cheering in sports matches, movies, theaters, concerts, political demonstrations, and riots. A critical question concerning the innovative concept of crowd emotions is whether the emotional content of crowd sounds can be characterized by frequency-amplitude features, using analysis techniques similar to those applied on individual voices, where deep learning classification is applied to spectrogram images derived by sound transformations. In this work, we present a technique based on the generation of sound spectrograms from fragments of fixed length, extracted from original audio clips recorded in high-attendance events, where the crowd acts as a collective individual. Transfer learning techniques are used on a convolutional neural network, pre-trained on low-level features using the well-known ImageNet extensive dataset of visual knowledge. The original sound clips are filtered and normalized in amplitude for a correct spectrogram generation, on which we fine-tune the domain-specific features. Experiments held on the finally trained Convolutional Neural Network show promising performances of the proposed model to classify the emotions of the crowd.
Collapse
Affiliation(s)
- Valentina Franzoni
- Department of Mathematics and Computer Science, University of Perugia, Perugia, Italy
| | - Giulio Biondi
- Department of Mathematics and Computer Science, University of Florence, Florence, Italy
| | - Alfredo Milani
- Department of Mathematics and Computer Science, University of Perugia, Perugia, Italy
| |
Collapse
|