1
|
Lindroth H, Nalaie K, Raghu R, Ayala IN, Busch C, Bhattacharyya A, Moreno Franco P, Diedrich DA, Pickering BW, Herasevich V. Applied Artificial Intelligence in Healthcare: A Review of Computer Vision Technology Application in Hospital Settings. J Imaging 2024; 10:81. [PMID: 38667979 PMCID: PMC11050909 DOI: 10.3390/jimaging10040081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 03/08/2024] [Accepted: 03/11/2024] [Indexed: 04/28/2024] Open
Abstract
Computer vision (CV), a type of artificial intelligence (AI) that uses digital videos or a sequence of images to recognize content, has been used extensively across industries in recent years. However, in the healthcare industry, its applications are limited by factors like privacy, safety, and ethical concerns. Despite this, CV has the potential to improve patient monitoring, and system efficiencies, while reducing workload. In contrast to previous reviews, we focus on the end-user applications of CV. First, we briefly review and categorize CV applications in other industries (job enhancement, surveillance and monitoring, automation, and augmented reality). We then review the developments of CV in the hospital setting, outpatient, and community settings. The recent advances in monitoring delirium, pain and sedation, patient deterioration, mechanical ventilation, mobility, patient safety, surgical applications, quantification of workload in the hospital, and monitoring for patient events outside the hospital are highlighted. To identify opportunities for future applications, we also completed journey mapping at different system levels. Lastly, we discuss the privacy, safety, and ethical considerations associated with CV and outline processes in algorithm development and testing that limit CV expansion in healthcare. This comprehensive review highlights CV applications and ideas for its expanded use in healthcare.
Collapse
Affiliation(s)
- Heidi Lindroth
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- Center for Aging Research, Regenstrief Institute, School of Medicine, Indiana University, Indianapolis, IN 46202, USA
- Center for Health Innovation and Implementation Science, School of Medicine, Indiana University, Indianapolis, IN 46202, USA
| | - Keivan Nalaie
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Roshini Raghu
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
| | - Ivan N. Ayala
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
| | - Charles Busch
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- College of Engineering, University of Wisconsin-Madison, Madison, WI 53705, USA
| | | | - Pablo Moreno Franco
- Department of Transplantation Medicine, Mayo Clinic, Jacksonville, FL 32224, USA
| | - Daniel A. Diedrich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Brian W. Pickering
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| |
Collapse
|
2
|
Li Y, Xu S, Zhu Z, Wang P, Li K, He Q, Zheng Q. EFC-YOLO: An Efficient Surface-Defect-Detection Algorithm for Steel Strips. SENSORS (BASEL, SWITZERLAND) 2023; 23:7619. [PMID: 37688077 PMCID: PMC10490735 DOI: 10.3390/s23177619] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/03/2023] [Revised: 08/28/2023] [Accepted: 08/31/2023] [Indexed: 09/10/2023]
Abstract
The pursuit of higher recognition accuracy and speed with smaller model sizes has been a major research topic in the detection of surface defects in steel. In this paper, we propose an improved high-speed and high-precision Efficient Fusion Coordination network (EFC-YOLO) without increasing the model's size. Since modifications to enhance feature extraction in shallow networks tend to affect the speed of model inference, in order to simultaneously ensure the accuracy and speed of detection, we add the improved Fusion-Faster module to the backbone network of YOLOv7. Partial Convolution (PConv) serves as the basic operator of the module, which strengthens the feature-extraction ability of shallow networks while maintaining speed. Additionally, we incorporate the Shortcut Coordinate Attention (SCA) mechanism to better capture the location information dependency, considering both lightweight design and accuracy. The de-weighted Bi-directional Feature Pyramid Network (BiFPN) structure used in the neck part of the network improves the original Path Aggregation Network (PANet)-like structure by adding step branches and reducing computations, achieving better feature fusion. In the experiments conducted on the NEU-DET dataset, the final model achieved an 85.9% mAP and decreased the GFLOPs by 60%, effectively balancing the model's size with the accuracy and speed of detection.
Collapse
Affiliation(s)
| | - Shuobo Xu
- School of Information Science and Electrical Engineering, Shandong Jiaotong University, Jinan 250357, China; (Y.L.); (Z.Z.); (P.W.); (K.L.); (Q.H.); (Q.Z.)
| | | | | | | | | | | |
Collapse
|
3
|
Avazov K, Jamil MK, Muminov B, Abdusalomov AB, Cho YI. Fire Detection and Notification Method in Ship Areas Using Deep Learning and Computer Vision Approaches. SENSORS (BASEL, SWITZERLAND) 2023; 23:7078. [PMID: 37631614 PMCID: PMC10458310 DOI: 10.3390/s23167078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/03/2023] [Revised: 08/02/2023] [Accepted: 08/07/2023] [Indexed: 08/27/2023]
Abstract
Fire incidents occurring onboard ships cause significant consequences that result in substantial effects. Fires on ships can have extensive and severe wide-ranging impacts on matters such as the safety of the crew, cargo, the environment, finances, reputation, etc. Therefore, timely detection of fires is essential for quick responses and powerful mitigation. The study in this research paper presents a fire detection technique based on YOLOv7 (You Only Look Once version 7), incorporating improved deep learning algorithms. The YOLOv7 architecture, with an improved E-ELAN (extended efficient layer aggregation network) as its backbone, serves as the basis of our fire detection system. Its enhanced feature fusion technique makes it superior to all its predecessors. To train the model, we collected 4622 images of various ship scenarios and performed data augmentation techniques such as rotation, horizontal and vertical flips, and scaling. Our model, through rigorous evaluation, showcases enhanced capabilities of fire recognition to improve maritime safety. The proposed strategy successfully achieves an accuracy of 93% in detecting fires to minimize catastrophic incidents. Objects having visual similarities to fire may lead to false prediction and detection by the model, but this can be controlled by expanding the dataset. However, our model can be utilized as a real-time fire detector in challenging environments and for small-object detection. Advancements in deep learning models hold the potential to enhance safety measures, and our proposed model in this paper exhibits this potential. Experimental results proved that the proposed method can be used successfully for the protection of ships and in monitoring fires in ship port areas. Finally, we compared the performance of our method with those of recently reported fire-detection approaches employing widely used performance matrices to test the fire classification results achieved.
Collapse
Affiliation(s)
- Kuldoshbay Avazov
- Department of Computer Engineering, Gachon University, Seongnam-si 461-701, Republic of Korea; (K.A.)
| | - Muhammad Kafeel Jamil
- Department of Computer Engineering, Gachon University, Seongnam-si 461-701, Republic of Korea; (K.A.)
| | - Bahodir Muminov
- Department of Artificial Intelligence, Tashkent State University of Economics, Tashkent 100066, Uzbekistan
| | | | - Young-Im Cho
- Department of Computer Engineering, Gachon University, Seongnam-si 461-701, Republic of Korea; (K.A.)
| |
Collapse
|
4
|
Song W, Xin L, Wang J. A grading method for Kayser Fleischer ring images based on ResNet. Heliyon 2023; 9:e16149. [PMID: 37234668 PMCID: PMC10205591 DOI: 10.1016/j.heliyon.2023.e16149] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2022] [Revised: 05/04/2023] [Accepted: 05/07/2023] [Indexed: 05/28/2023] Open
Abstract
The corneal K-F ring is the most common ophthalmic manifestation of WD patients. Early diagnosis and treatment have an important impact on the patient's condition. K-F ring is one of the gold standards for the diagnosis of WD disease. Therefore, this paper mainly focused on the detection and grading of the K-F ring. The aim of this study is three-fold. Firstly, to create a meaningful database, the K-F ring images are collected which contains 1850 images with 399 different WD patients, and then this paper uses the chi-square test and Friedman test to analyze the statistical significance. Subsequently, the all collected images were graded and labeled with an appropriate treatment approach, as a result, these images could be used to detect the corneal through the YOLO. After the detection of corneal, image segmentation was realized in batches. Finally, in this paper, different deep convolutional neural networks (VGG, ResNet, and DenseNet) were used to realize the grading of the K-F ring images in the KFID. Experimental results reveal that the entire pre-trained models obtain excellent performance. The global accuracies achieved by the six models i.e., VGG-16, VGG-19, ResNet18, ResNet34, ResNet50, and DenseNet are 89.88%, 91.89%, 94.18%, 95.31%, 93.59%, and 94.58% respectively. ResNet34 displayed the highest recall, specificity, and F1-score of 95.23%, 96.99%, and 95.23%. DenseNet showed the best precision of 95.66%. As such, the findings are encouraging, demonstrating the effectiveness of ResNet in the automatic grading of the K-F ring. Moreover, it provides effective help for the clinical diagnosis of HLD.
Collapse
Affiliation(s)
- Wei Song
- The First Affiliated Hospital of Anhui University of Chinese Medicine, Hefei, 230031, China
| | - Ling Xin
- The First Affiliated Hospital of Anhui University of Chinese Medicine, Hefei, 230031, China
| | - Jiemei Wang
- Department of Otolaryngology, The First Affiliated Hospital of Anhui University of Chinese Medicine, Hefei, 230031, China
| |
Collapse
|
5
|
Zhang Y, Sun Y, Wang Z, Jiang Y. YOLOv7-RAR for Urban Vehicle Detection. SENSORS (BASEL, SWITZERLAND) 2023; 23:1801. [PMID: 36850399 PMCID: PMC9964850 DOI: 10.3390/s23041801] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/28/2022] [Revised: 01/16/2023] [Accepted: 01/29/2023] [Indexed: 06/18/2023]
Abstract
Aiming at the problems of high missed detection rates of the YOLOv7 algorithm for vehicle detection on urban roads, weak perception of small targets in perspective, and insufficient feature extraction, the YOLOv7-RAR recognition algorithm is proposed. The algorithm is improved from the following three directions based on YOLOv7. Firstly, in view of the insufficient nonlinear feature fusion of the original backbone network, the Res3Unit structure is used to reconstruct the backbone network of YOLOv7 to improve the ability of the network model architecture to obtain more nonlinear features. Secondly, in view of the problem that there are many interference backgrounds in urban roads and that the original network is weak in positioning targets such as vehicles, a plug-and-play hybrid attention mechanism module, ACmix, is added after the SPPCSPC layer of the backbone network to enhance the network's attention to vehicles and reduce the interference of other targets. Finally, aiming at the problem that the receptive field of the original network Narrows, with the deepening of the network model, leads to a high miss rate of small targets, the Gaussian receptive field scheme used in the RFLA (Gaussian-receptive-field-based label assignment) module is used at the connection between the feature fusion area and the detection head to improve the receptive field of the network model for small objects in the image. Combining the three improvement measures, the first letter of the name of each improvement measure is selected, and the improved algorithm is named the YOLOv7-RAR algorithm. Experiments show that on urban roads with crowded vehicles and different weather patterns, the average detection accuracy of the YOLOv7-RAR algorithm reaches 95.1%, which is 2.4% higher than that of the original algorithm; the AP50:90 performance is 12.6% higher than that of the original algorithm. The running speed of the YOLOv7-RAR algorithm reaches 96 FPS, which meets the real-time requirements of vehicle detection; hence, the algorithm can be better applied to vehicle detection.
Collapse
|
6
|
Qiu Y, Lu Y, Wang Y, Jiang H. IDOD-YOLOV7: Image-Dehazing YOLOV7 for Object Detection in Low-Light Foggy Traffic Environments. SENSORS (BASEL, SWITZERLAND) 2023; 23:1347. [PMID: 36772388 PMCID: PMC9921160 DOI: 10.3390/s23031347] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Revised: 01/11/2023] [Accepted: 01/18/2023] [Indexed: 06/18/2023]
Abstract
Convolutional neural network (CNN)-based autonomous driving object detection algorithms have excellent detection results on conventional datasets, but the detector performance can be severely degraded in low-light foggy weather environments. Existing methods have difficulty in achieving a balance between low-light image enhancement and object detection. To alleviate this problem, this paper proposes a foggy traffic environment object detection framework, IDOD-YOLOV7. This network is based on joint optimal learning of image defogging module IDOD (AOD + SAIP) and YOLOV7 detection modules. Specifically, for low-light foggy images, we propose to improve the image quality by joint optimization of image defogging (AOD) and image enhancement (SAIP), where the parameters of the SAIP module are predicted by a miniature CNN network and the AOD module performs image defogging by optimizing the atmospheric scattering model. The experimental results show that the IDOD module not only improves the image defogging quality for low-light fog images but also achieves better results in objective evaluation indexes such as PSNR and SSIM. The IDOD and YOLOV7 learn jointly in an end-to-end manner so that object detection can be performed while image enhancement is executed in a weakly supervised manner. Finally, a low-light fogged traffic image dataset (FTOD) was built by physical fogging in order to solve the domain transfer problem. The training of IDOD-YOLOV7 network by a real dataset (FTOD) improves the robustness of the model. We performed various experiments to visually and quantitatively compare our method with several state-of-the-art methods to demonstrate its superiority over the others. The IDOD-YOLOV7 algorithm not only suppresses the artifacts of low-light fog images and improves the visual effect of images but also improves the perception of autonomous driving in low-light foggy environments.
Collapse
Affiliation(s)
- Yongsheng Qiu
- School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China
| | - Yuanyao Lu
- School of Information Science and Technology, North China University of Technology, Beijing 100144, China
| | - Yuantao Wang
- School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China
| | - Haiyang Jiang
- School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China
| |
Collapse
|
7
|
Hussain M, Al-Aqrabi H, Munawar M, Hill R. Feature Mapping for Rice Leaf Defect Detection Based on a Custom Convolutional Architecture. Foods 2022; 11:foods11233914. [PMID: 36496723 PMCID: PMC9738204 DOI: 10.3390/foods11233914] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2022] [Revised: 11/22/2022] [Accepted: 12/02/2022] [Indexed: 12/12/2022] Open
Abstract
Rice is a widely consumed food across the world. Whilst the world recovers from COVID-19, food manufacturers are looking to enhance their quality inspection processes for satisfying exportation requirements and providing safety assurance to their clients. Rice cultivation is a significant process, the yield of which can be significantly impacted in an adverse manner due to plant disease. Yet, a large portion of rice cultivation takes place in developing countries with less stringent quality inspection protocols due to various reasons including cost of labor. To address this, we propose the development of lightweight convolutional neural network architecture for the automated detection of rice leaf smut and rice leaf blight. In doing so, this research addresses the issue of data scarcity via a practical variance modeling mechanism (Domain Feature Mapping) and a custom filter development mechanism assisted through a reference protocol for filter suppression.
Collapse
Affiliation(s)
- Muhammad Hussain
- Department of Computer Science School of Computing and Engineering, University of Huddersfield, Queensgate, Huddersfield HD1 3DH, UK
| | - Hussain Al-Aqrabi
- Department of Computer Science School of Computing and Engineering, University of Huddersfield, Queensgate, Huddersfield HD1 3DH, UK
- Correspondence:
| | - Muhammad Munawar
- Department of Computer Science, COMSATS University of Islamabad, Park Road, Tarlai Kalan, Islamabad 45550, Pakistan
| | - Richard Hill
- Department of Computer Science School of Computing and Engineering, University of Huddersfield, Queensgate, Huddersfield HD1 3DH, UK
| |
Collapse
|
8
|
Ren J, Pan Y, Yao P, Hu Y, Gao W, Xue Z. Deep Learning-Based Intelligent Forklift Cargo Accurate Transfer System. SENSORS (BASEL, SWITZERLAND) 2022; 22:8437. [PMID: 36366133 PMCID: PMC9655510 DOI: 10.3390/s22218437] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/10/2022] [Revised: 10/26/2022] [Accepted: 11/01/2022] [Indexed: 06/16/2023]
Abstract
In this research, we present an intelligent forklift cargo precision transfer system to address the issue of poor pallet docking accuracy and low recognition rate when using current techniques. The technology is primarily used to automatically check if there is any pallet that need to be transported. The intelligent forklift is then sent to the area of the target pallet after being recognized. Images of the pallets are then collected using the forklift's camera, and a deep learning-based recognition algorithm is used to calculate the precise position of the pallets. Finally, the forklift is controlled by a high-precision control algorithm to insert the pallet in the exact location. This system creatively introduces the small target detection into the pallet target recognition system, which greatly improves the recognition rate of the system. The application of Yolov5 into the pallet positional calculation makes the coverage and recognition accuracy of the algorithm improved. In comparison with the prior approach, this system's identification rate and accuracy are substantially higher, and it requires fewer sensors and indications to help with deployment. We have collected a significant amount of real data in order to confirm the system's viability and stability. Among them, the accuracy of pallet docking is evaluated 1000 times, and the inaccuracy is kept to a maximum of 6 mm. The recognition rate of pallet recognition is above 99.5% in 7 days of continuous trials.
Collapse
Affiliation(s)
- Jie Ren
- Intelligent Perception and Control Center, Huzhou Institute of Zhejiang University, Huzhou 313098, China
| | - Yusu Pan
- Institute of Cyber-Systems and Control, Zhejiang University, Hangzhou 310027, China
| | - Pantao Yao
- Intelligent Perception and Control Center, Huzhou Institute of Zhejiang University, Huzhou 313098, China
| | - Yicheng Hu
- Intelligent Perception and Control Center, Huzhou Institute of Zhejiang University, Huzhou 313098, China
| | - Wang Gao
- Science and Technology on Complex System Control and Intelligent Agent Cooperation Laboratory, Beijing 100191, China
| | - Zhenfeng Xue
- Intelligent Perception and Control Center, Huzhou Institute of Zhejiang University, Huzhou 313098, China
- Institute of Cyber-Systems and Control, Zhejiang University, Hangzhou 310027, China
| |
Collapse
|