1
|
He S, Ye X, Xie W, Shen Y, Yang S, Zhong X, Guan H, Zhou X, Wu J, Shen L. Open ultrawidefield fundus image dataset with disease diagnosis and clinical image quality assessment. Sci Data 2024; 11:1251. [PMID: 39567563 PMCID: PMC11579006 DOI: 10.1038/s41597-024-04113-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2024] [Accepted: 11/07/2024] [Indexed: 11/22/2024] Open
Abstract
Ultrawidefield fundus (UWF) images have a wide imaging range (200° of the retinal region), which offers the opportunity to show more information for ophthalmic diseases. Image quality assessment (IQA) is a prerequisite for applying UWF and is crucial for developing artificial intelligence-driven diagnosis and screening systems. Most image quality systems have been applied to the assessments of natural images, but whether these systems are suitable for evaluating the UWF image quality remains debatable. Additionally, existing IQA datasets only provide photographs of diabetic retinopathy (DR) patients and quality evaluation results applicable for natural image, neglecting patients' clinical information. To address these issues, we established a real-world clinical practice ultra-widefield fundus images dataset, with 700 high-resolution UWF images and corresponding clinical information from six common fundus diseases and healthy volunteers. The image quality is annotated by three ophthalmologists based on the field of view, illumination, artifact, contrast, and overall quality. This dataset illustrates the distribution of UWF image quality across diseases in clinical practice, offering a foundation for developing effective IQA systems.
Collapse
Affiliation(s)
- Shucheng He
- Department of Ophthalmology, Zhejiang Provincial People's Hospital (Affiliated People's Hospital, Hangzhou Medical College), Hangzhou, Zhejiang, China
| | - Xin Ye
- Zhejiang Provincial People's Hospital Bijie Hospital, Bijie, Guizhou, China.
| | - Wenbin Xie
- Zhejiang Provincial People's Hospital Bijie Hospital, Bijie, Guizhou, China
| | - Yingjiao Shen
- Department of Ophthalmology, Zhejiang Provincial People's Hospital (Affiliated People's Hospital, Hangzhou Medical College), Hangzhou, Zhejiang, China
| | | | - Xiaxing Zhong
- Wenzhou Medical University, Wenzhou, Zhejiang, China
| | - Hanyi Guan
- Wenzhou Medical University, Wenzhou, Zhejiang, China
| | | | - Jiang Wu
- Hangzhou Medical College, Hangzhou, Zhejiang, China
| | - Lijun Shen
- Department of Ophthalmology, Zhejiang Provincial People's Hospital (Affiliated People's Hospital, Hangzhou Medical College), Hangzhou, Zhejiang, China.
- Wenzhou Medical University, Wenzhou, Zhejiang, China.
- Hangzhou Medical College, Hangzhou, Zhejiang, China.
| |
Collapse
|
2
|
Wang S, Shen W, Gao Z, Jiang X, Wang Y, Li Y, Ma X, Wang W, Xin S, Ren W, Jin K, Ye J. Enhancing the ophthalmic AI assessment with a fundus image quality classifier using local and global attention mechanisms. Front Med (Lausanne) 2024; 11:1418048. [PMID: 39175821 PMCID: PMC11339790 DOI: 10.3389/fmed.2024.1418048] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 07/23/2024] [Indexed: 08/24/2024] Open
Abstract
Background The assessment of image quality (IQA) plays a pivotal role in the realm of image-based computer-aided diagnosis techniques, with fundus imaging standing as the primary method for the screening and diagnosis of ophthalmic diseases. Conventional studies on fundus IQA tend to rely on simplistic datasets for evaluation, predominantly focusing on either local or global information, rather than a synthesis of both. Moreover, the interpretability of these studies often lacks compelling evidence. In order to address these issues, this study introduces the Local and Global Attention Aggregated Deep Neural Network (LGAANet), an innovative approach that integrates both local and global information for enhanced analysis. Methods The LGAANet was developed and validated using a Multi-Source Heterogeneous Fundus (MSHF) database, encompassing a diverse collection of images. This dataset includes 802 color fundus photography (CFP) images (302 from portable cameras), and 500 ultrawide-field (UWF) images from 904 patients with diabetic retinopathy (DR) and glaucoma, as well as healthy individuals. The assessment of image quality was meticulously carried out by a trio of ophthalmologists, leveraging the human visual system as a benchmark. Furthermore, the model employs attention mechanisms and saliency maps to bolster its interpretability. Results In testing with the CFP dataset, LGAANet demonstrated remarkable accuracy in three critical dimensions of image quality (illumination, clarity and contrast based on the characteristics of human visual system, and indicates the potential aspects to improve the image quality), recording scores of 0.947, 0.924, and 0.947, respectively. Similarly, when applied to the UWF dataset, the model achieved accuracies of 0.889, 0.913, and 0.923, respectively. These results underscore the efficacy of LGAANet in distinguishing between varying degrees of image quality with high precision. Conclusion To our knowledge, LGAANet represents the inaugural algorithm trained on an MSHF dataset specifically for fundus IQA, marking a significant milestone in the advancement of computer-aided diagnosis in ophthalmology. This research significantly contributes to the field, offering a novel methodology for the assessment and interpretation of fundus images in the detection and diagnosis of ocular diseases.
Collapse
Affiliation(s)
- Shengzhan Wang
- The Affiliated People’s Hospital of Ningbo University, Ningbo, Zhejiang, China
| | - Wenyue Shen
- Eye Center, School of Medicine, The Second Affiliated Hospital, Zhejiang University, Hangzhou, Zhejiang, China
| | - Zhiyuan Gao
- Eye Center, School of Medicine, The Second Affiliated Hospital, Zhejiang University, Hangzhou, Zhejiang, China
| | - Xiaoyu Jiang
- College of Control Science and Engineering, Zhejiang University, Hangzhou, China
| | - Yaqi Wang
- College of Media, Communication University of Zhejiang, Hangzhou, China
| | - Yunxiang Li
- College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, China
| | - Xiaoyu Ma
- Institute of Intelligent Media, Communication University of Zhejiang, Hangzhou, China
| | - Wenhao Wang
- The Affiliated People’s Hospital of Ningbo University, Ningbo, Zhejiang, China
| | - Shuanghua Xin
- The Affiliated People’s Hospital of Ningbo University, Ningbo, Zhejiang, China
| | - Weina Ren
- The Affiliated People’s Hospital of Ningbo University, Ningbo, Zhejiang, China
| | - Kai Jin
- Eye Center, School of Medicine, The Second Affiliated Hospital, Zhejiang University, Hangzhou, Zhejiang, China
| | - Juan Ye
- Eye Center, School of Medicine, The Second Affiliated Hospital, Zhejiang University, Hangzhou, Zhejiang, China
| |
Collapse
|
3
|
Zhang G, Qu Y, Zhang Y, Tang J, Wang C, Yin H, Yao X, Liang G, Shen T, Ren Q, Jia H, Sun X. Multimodal Eye Imaging, Retina Characteristics, and Psychological Assessment Dataset. Sci Data 2024; 11:836. [PMID: 39095400 PMCID: PMC11297319 DOI: 10.1038/s41597-024-03690-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2024] [Accepted: 07/26/2024] [Indexed: 08/04/2024] Open
Abstract
The eyes provide insights into psychology, potentially offering a distinctive perspective for psychological health profiles. However, there exist a notable deficiency in datasets that simultaneously encompass eye features and psychological assessments. To address this gap, our study presents a dataset that included Fundus Photography, Psychological Assessment, Retina Characteristics, and Multimodal Imaging (FPRM). FPRM dataset comprise fundus images at different wavelengths (548 nm and 605 nm), image of oxygen saturation for the retina and 8 specific retinal vessels, videos of retinal blood flow and pupillary light reflex, along with 61 items of multimodal quantitative measurement from 384 participants. Additionally, it features psychological assessments across five dimensions (geriatric depression, generalized anxiety disorder, insomnia, activities of daily living, and deterioration), accompanied by fundus photographs and 6 items of retina characteristics from 1683 participants. FPRM dataset is the first to integrate multimodal ophthalmic data and psychological assessments, not only advancing the development of machine learning applications but also facilitating in-depth research into the relationship between eye health and psychological health profiles.
Collapse
Affiliation(s)
- Guanran Zhang
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China
- National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Yanlin Qu
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China
- National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Yanping Zhang
- Jinshan District Mental Health Center, Shanghai, China
| | - Jiayi Tang
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China
- National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Chunyan Wang
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China
- National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Haogui Yin
- Jinshan District Mental Health Center, Shanghai, China
| | - Xiaoping Yao
- Jinshan District Mental Health Center, Shanghai, China
| | - Gengshi Liang
- Community Health Service Center in Lvxiang Town, Shanghai, China
| | - Ting Shen
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China
- National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Qiushi Ren
- Department of Biomedical Engineering and National Biomedical Imaging Center, College of Future Technology, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
| | - Huixun Jia
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China.
- National Clinical Research Center for Eye Diseases, Shanghai, China.
| | - Xiaodong Sun
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
- Shanghai Key Laboratory of Ocular Fundus Diseases, Shanghai, China.
- National Clinical Research Center for Eye Diseases, Shanghai, China.
| |
Collapse
|
4
|
Zhao X, Chen S, Zhang S, Liu Y, Hu Y, Yuan D, Xie L, Luo X, Zheng M, Tian R, Chen Y, Tan T, Yu Z, Sun Y, Wu Z, Zhang G. A fundus image dataset for intelligent retinopathy of prematurity system. Sci Data 2024; 11:543. [PMID: 38802420 PMCID: PMC11130119 DOI: 10.1038/s41597-024-03362-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2023] [Accepted: 05/10/2024] [Indexed: 05/29/2024] Open
Abstract
Image-based artificial intelligence (AI) systems stand as the major modality for evaluating ophthalmic conditions. However, most of the currently available AI systems are designed for experimental research using single-central datasets. Most of them fell short of application in real-world clinical settings. In this study, we collected a dataset of 1,099 fundus images in both normal and pathologic eyes from 483 premature infants for intelligent retinopathy of prematurity (ROP) system development and validation. Dataset diversity was visualized with a spatial scatter plot. Image classification was conducted by three annotators. To the best of our knowledge, this is one of the largest fundus datasets on ROP, and we believe it is conducive to the real-world application of AI systems.
Collapse
Affiliation(s)
- Xinyu Zhao
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Shaobin Chen
- Faculty of Applied Sciences, Macao Polytechnic University, Macao Special Administrative Region of China, Macao, China
| | - Sifan Zhang
- Department of Biology, New York University, New York, NY, US
| | - Yaling Liu
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Yarou Hu
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Duo Yuan
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Liqiong Xie
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China
| | - Xiayuan Luo
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Mianying Zheng
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Ruyin Tian
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Yi Chen
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Tao Tan
- Faculty of Applied Sciences, Macao Polytechnic University, Macao Special Administrative Region of China, Macao, China
| | - Zhen Yu
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Yue Sun
- Faculty of Applied Sciences, Macao Polytechnic University, Macao Special Administrative Region of China, Macao, China.
| | - Zhenquan Wu
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China.
| | - Guoming Zhang
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China.
| |
Collapse
|
5
|
Lian J, Hui G, Ma L, Zhu T, Wu X, Heidari AA, Chen Y, Chen H. Parrot optimizer: Algorithm and applications to medical problems. Comput Biol Med 2024; 172:108064. [PMID: 38452469 DOI: 10.1016/j.compbiomed.2024.108064] [Citation(s) in RCA: 54] [Impact Index Per Article: 54.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Revised: 01/09/2024] [Accepted: 01/27/2024] [Indexed: 03/09/2024]
Abstract
Stochastic optimization methods have gained significant prominence as effective techniques in contemporary research, addressing complex optimization challenges efficiently. This paper introduces the Parrot Optimizer (PO), an efficient optimization method inspired by key behaviors observed in trained Pyrrhura Molinae parrots. The study features qualitative analysis and comprehensive experiments to showcase the distinct characteristics of the Parrot Optimizer in handling various optimization problems. Performance evaluation involves benchmarking the proposed PO on 35 functions, encompassing classical cases and problems from the IEEE CEC 2022 test sets, and comparing it with eight popular algorithms. The results vividly highlight the competitive advantages of the PO in terms of its exploratory and exploitative traits. Furthermore, parameter sensitivity experiments explore the adaptability of the proposed PO under varying configurations. The developed PO demonstrates effectiveness and superiority when applied to engineering design problems. To further extend the assessment to real-world applications, we included the application of PO to disease diagnosis and medical image segmentation problems, which are highly relevant and significant in the medical field. In conclusion, the findings substantiate that the PO is a promising and competitive algorithm, surpassing some existing algorithms in the literature. The supplementary files and open source codes of the proposed Parrot Optimizer (PO) is available at https://aliasgharheidari.com/PO.html and https://github.com/junbolian/PO.
Collapse
Affiliation(s)
- Junbo Lian
- College of Mathematics and Computer Sciences, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Sensing Technology and Intelligent Equipment of Department of Forestry, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Zhejiang A & F University, Hangzhou, 311300, PR China.
| | - Guohua Hui
- College of Mathematics and Computer Sciences, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Sensing Technology and Intelligent Equipment of Department of Forestry, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Zhejiang A & F University, Hangzhou, 311300, PR China.
| | - Ling Ma
- College of Mathematics and Computer Sciences, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Sensing Technology and Intelligent Equipment of Department of Forestry, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Zhejiang A & F University, Hangzhou, 311300, PR China.
| | - Ting Zhu
- College of Mathematics and Computer Sciences, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Sensing Technology and Intelligent Equipment of Department of Forestry, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Zhejiang A & F University, Hangzhou, 311300, PR China.
| | - Xincan Wu
- College of Mathematics and Computer Sciences, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Sensing Technology and Intelligent Equipment of Department of Forestry, Zhejiang A & F University, Hangzhou, 311300, PR China; Key Laboratory of Forestry Intelligent Monitoring and Information Technology of Zhejiang Province, Zhejiang A & F University, Hangzhou, 311300, PR China.
| | - Ali Asghar Heidari
- School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran, Tehran, Iran.
| | - Yi Chen
- Key Laboratory of Intelligent Informatics for Safety & Emergency of Zhejiang Province, Wenzhou University, Wenzhou 325035, China.
| | - Huiling Chen
- Key Laboratory of Intelligent Informatics for Safety & Emergency of Zhejiang Province, Wenzhou University, Wenzhou 325035, China.
| |
Collapse
|
6
|
Deepa K, Bacanin N, Askar SS, Abouhawwash M. Elderly and visually impaired indoor activity monitoring based on Wi-Fi and Deep Hybrid convolutional neural network. Sci Rep 2023; 13:22470. [PMID: 38110422 PMCID: PMC10728209 DOI: 10.1038/s41598-023-48860-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2023] [Accepted: 11/30/2023] [Indexed: 12/20/2023] Open
Abstract
A drop in physical activity and a deterioration in the capacity to undertake daily life activities are both connected with ageing and have negative effects on physical and mental health. An Elderly and Visually Impaired Human Activity Monitoring (EV-HAM) system that keeps tabs on a person's routine and steps in if a change in behaviour or a crisis might greatly help an elderly person or a visually impaired. These individuals may find greater freedom with the help of an EVHAM system. As the backbone of human-centric applications like actively supported living and in-home monitoring for the elderly and visually impaired, an EVHAM system is essential. Big data-driven product design is flourishing in this age of 5G and the IoT. Recent advancements in processing power and software architectures have also contributed to the emergence and development of artificial intelligence (AI). In this context, the digital twin has emerged as a state-of-the-art technology that bridges the gap between the real and virtual worlds by evaluating data from several sensors using artificial intelligence algorithms. Although promising findings have been reported by Wi-Fi-based human activity identification techniques so far, their effectiveness is vulnerable to environmental variations. Using the environment-independent fingerprints generated from the Wi-Fi channel state information (CSI), we introduce Wi-Sense. This human activity identification system employs a Deep Hybrid convolutional neural network (DHCNN). The proposed system begins by collecting the CSI with a regular Wi-Fi Network Interface Controller. Wi-Sense uses the CSI ratio technique to lessen the effect of noise and the phase offset. The t- Distributed Stochastic Neighbor Embedding (t-SNE) is used to eliminate unnecessary data further. The data dimension is decreased, and the negative effects on the environment are eliminated in this process. The resulting spectrogram of the processed data exposes the activity's micro-Doppler fingerprints as a function of both time and location. These spectrograms are put to use in the training of a DHCNN. Based on our findings, EVHAM can accurately identify these actions 99% of the time.
Collapse
Affiliation(s)
- K Deepa
- Department of Computer Science and Engineering, K.Ramakrishnan College of Technology, Trichy, 621112, India
| | | | - S S Askar
- Department of Statistics and Operations Research, College of Science, King Saud University, P.O. Box 2455, 11451, Riyadh, Saudi Arabia
| | - Mohamed Abouhawwash
- Department of Computational Mathematics, Science and Engineering (CMSE), College of Engineering, Michigan State University, East Lansing, MI, 48824, USA
- Department of Mathematics, Faculty of Science, Mansoura University, Mansoura, 35516, Egypt
| |
Collapse
|
7
|
Javeed M, Abdelhaq M, Algarni A, Jalal A. Biosensor-Based Multimodal Deep Human Locomotion Decoding via Internet of Healthcare Things. MICROMACHINES 2023; 14:2204. [PMID: 38138373 PMCID: PMC10745656 DOI: 10.3390/mi14122204] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Revised: 11/28/2023] [Accepted: 11/30/2023] [Indexed: 12/24/2023]
Abstract
Multiple Internet of Healthcare Things (IoHT)-based devices have been utilized as sensing methodologies for human locomotion decoding to aid in applications related to e-healthcare. Different measurement conditions affect the daily routine monitoring, including the sensor type, wearing style, data retrieval method, and processing model. Currently, several models are present in this domain that include a variety of techniques for pre-processing, descriptor extraction, and reduction, along with the classification of data captured from multiple sensors. However, such models consisting of multiple subject-based data using different techniques may degrade the accuracy rate of locomotion decoding. Therefore, this study proposes a deep neural network model that not only applies the state-of-the-art Quaternion-based filtration technique for motion and ambient data along with background subtraction and skeleton modeling for video-based data, but also learns important descriptors from novel graph-based representations and Gaussian Markov random-field mechanisms. Due to the non-linear nature of data, these descriptors are further utilized to extract the codebook via the Gaussian mixture regression model. Furthermore, the codebook is provided to the recurrent neural network to classify the activities for the locomotion-decoding system. We show the validity of the proposed model across two publicly available data sampling strategies, namely, the HWU-USP and LARa datasets. The proposed model is significantly improved over previous systems, as it achieved 82.22% and 82.50% for the HWU-USP and LARa datasets, respectively. The proposed IoHT-based locomotion-decoding model is useful for unobtrusive human activity recognition over extended periods in e-healthcare facilities.
Collapse
Affiliation(s)
- Madiha Javeed
- Department of Computer Science, Air University, Islamabad 44000, Pakistan;
| | - Maha Abdelhaq
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Asaad Algarni
- Department of Computer Sciences, Faculty of Computing and Information Technology, Northern Border University, Rafha 91911, Saudi Arabia;
| | - Ahmad Jalal
- Department of Computer Science, Air University, Islamabad 44000, Pakistan;
| |
Collapse
|
8
|
Wang Y, Liu L, Wang C. Trends in using deep learning algorithms in biomedical prediction systems. Front Neurosci 2023; 17:1256351. [PMID: 38027475 PMCID: PMC10665494 DOI: 10.3389/fnins.2023.1256351] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2023] [Accepted: 09/25/2023] [Indexed: 12/01/2023] Open
Abstract
In the domain of using DL-based methods in medical and healthcare prediction systems, the utilization of state-of-the-art deep learning (DL) methodologies assumes paramount significance. DL has attained remarkable achievements across diverse domains, rendering its efficacy particularly noteworthy in this context. The integration of DL with health and medical prediction systems enables real-time analysis of vast and intricate datasets, yielding insights that significantly enhance healthcare outcomes and operational efficiency in the industry. This comprehensive literature review systematically investigates the latest DL solutions for the challenges encountered in medical healthcare, with a specific emphasis on DL applications in the medical domain. By categorizing cutting-edge DL approaches into distinct categories, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), generative adversarial networks (GANs), long short-term memory (LSTM) models, support vector machine (SVM), and hybrid models, this study delves into their underlying principles, merits, limitations, methodologies, simulation environments, and datasets. Notably, the majority of the scrutinized articles were published in 2022, underscoring the contemporaneous nature of the research. Moreover, this review accentuates the forefront advancements in DL techniques and their practical applications within the realm of medical prediction systems, while simultaneously addressing the challenges that hinder the widespread implementation of DL in image segmentation within the medical healthcare domains. These discerned insights serve as compelling impetuses for future studies aimed at the progressive advancement of using DL-based methods in medical and health prediction systems. The evaluation metrics employed across the reviewed articles encompass a broad spectrum of features, encompassing accuracy, precision, specificity, F-score, adoptability, adaptability, and scalability.
Collapse
Affiliation(s)
- Yanbu Wang
- School of Strength and Conditioning, Beijing Sport University, Beijing, China
| | - Linqing Liu
- Department of Physical Education, Peking University, Beijing, China
| | - Chao Wang
- Institute of Competitive Sports, Beijing Sport University, Beijing, China
| |
Collapse
|
9
|
Li M, Jiang Y, Zhang Y, Zhu H. Medical image analysis using deep learning algorithms. Front Public Health 2023; 11:1273253. [PMID: 38026291 PMCID: PMC10662291 DOI: 10.3389/fpubh.2023.1273253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Accepted: 10/05/2023] [Indexed: 12/01/2023] Open
Abstract
In the field of medical image analysis within deep learning (DL), the importance of employing advanced DL techniques cannot be overstated. DL has achieved impressive results in various areas, making it particularly noteworthy for medical image analysis in healthcare. The integration of DL with medical image analysis enables real-time analysis of vast and intricate datasets, yielding insights that significantly enhance healthcare outcomes and operational efficiency in the industry. This extensive review of existing literature conducts a thorough examination of the most recent deep learning (DL) approaches designed to address the difficulties faced in medical healthcare, particularly focusing on the use of deep learning algorithms in medical image analysis. Falling all the investigated papers into five different categories in terms of their techniques, we have assessed them according to some critical parameters. Through a systematic categorization of state-of-the-art DL techniques, such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Generative Adversarial Networks (GANs), Long Short-term Memory (LSTM) models, and hybrid models, this study explores their underlying principles, advantages, limitations, methodologies, simulation environments, and datasets. Based on our results, Python was the most frequent programming language used for implementing the proposed methods in the investigated papers. Notably, the majority of the scrutinized papers were published in 2021, underscoring the contemporaneous nature of the research. Moreover, this review accentuates the forefront advancements in DL techniques and their practical applications within the realm of medical image analysis, while simultaneously addressing the challenges that hinder the widespread implementation of DL in image analysis within the medical healthcare domains. These discerned insights serve as compelling impetuses for future studies aimed at the progressive advancement of image analysis in medical healthcare research. The evaluation metrics employed across the reviewed articles encompass a broad spectrum of features, encompassing accuracy, sensitivity, specificity, F-score, robustness, computational complexity, and generalizability.
Collapse
Affiliation(s)
- Mengfang Li
- The First Affiliated Hospital of Wenzhou Medical University, Wenzhou, China
| | - Yuanyuan Jiang
- Department of Cardiovascular Medicine, The First Affiliated Hospital of Zhengzhou University, Zhengzhou, China
| | - Yanzhou Zhang
- Department of Cardiovascular Medicine, The First Affiliated Hospital of Zhengzhou University, Zhengzhou, China
| | - Haisheng Zhu
- Department of Cardiovascular Medicine, Wencheng People’s Hospital, Wencheng, China
| |
Collapse
|
10
|
Kim S, Yoon H, Lee J, Yoo S. Facial wrinkle segmentation using weighted deep supervision and semi-automatic labeling. Artif Intell Med 2023; 145:102679. [PMID: 37925209 DOI: 10.1016/j.artmed.2023.102679] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2023] [Revised: 07/28/2023] [Accepted: 10/03/2023] [Indexed: 11/06/2023]
Abstract
Facial wrinkles are important indicators of human aging. Recently, a method using deep learning and a semi-automatic labeling was proposed to segment facial wrinkles, which showed much better performance than conventional image-processing-based methods. However, the difficulty of wrinkle segmentation remains challenging due to the thinness of wrinkles and their small proportion in the entire image. Therefore, performance improvement in wrinkle segmentation is still necessary. To address this issue, we propose a novel loss function that takes into account the thickness of wrinkles based on the semi-automatic labeling approach. First, considering the different spatial dimensions of the decoder in the U-Net architecture, we generated weighted wrinkle maps from ground truth. These weighted wrinkle maps were used to calculate the training losses more accurately than the existing deep supervision approach. This new loss computation approach is defined as weighted deep supervision in our study. The proposed method was evaluated using an image dataset obtained from a professional skin analysis device and labeled using semi-automatic labeling. In our experiment, the proposed weighted deep supervision showed higher Jaccard Similarity Index (JSI) performance for wrinkle segmentation compared to conventional deep supervision and traditional image processing methods. Additionally, we conducted experiments on the labeling using a semi-automatic labeling approach, which had not been explored in previous research, and compared it with human labeling. The semi-automatic labeling technology showed more consistent wrinkle labels than human-made labels. Furthermore, to assess the scalability of the proposed method to other domains, we applied it to retinal vessel segmentation. The results demonstrated superior performance of the proposed method compared to existing retinal vessel segmentation approaches. In conclusion, the proposed method offers high performance and can be easily applied to various biomedical domains and U-Net-based architectures. Therefore, the proposed approach will be beneficial for various biomedical imaging approaches. To facilitate this, we have made the source code of the proposed method publicly available at: https://github.com/resemin/WeightedDeepSupervision.
Collapse
Affiliation(s)
- Semin Kim
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Huisu Yoon
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Jongha Lee
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Sangwook Yoo
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| |
Collapse
|
11
|
Batool S, Gilani SO, Waris A, Iqbal KF, Khan NB, Khan MI, Eldin SM, Awwad FA. Deploying efficient net batch normalizations (BNs) for grading diabetic retinopathy severity levels from fundus images. Sci Rep 2023; 13:14462. [PMID: 37660096 PMCID: PMC10475020 DOI: 10.1038/s41598-023-41797-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2023] [Accepted: 08/31/2023] [Indexed: 09/04/2023] Open
Abstract
Diabetic retinopathy (DR) is one of the main causes of blindness in people around the world. Early diagnosis and treatment of DR can be accomplished by organizing large regular screening programs. Still, it is difficult to spot diabetic retinopathy timely because the situation might not indicate signs in the primary stages of the disease. Due to a drastic increase in diabetic patients, there is an urgent need for efficient diabetic retinopathy detecting systems. Auto-encoders, sparse coding, and limited Boltzmann machines were used as a few past deep learning (DL) techniques and features for the classification of DR. Convolutional Neural Networks (CNN) have been identified as a promising solution for detecting and classifying DR. We employ the deep learning capabilities of efficient net batch normalization (BNs) pre-trained models to automatically acquire discriminative features from fundus images. However, we successfully achieved F1 scores above 80% on all efficient net BNs in the EYE-PACS dataset (calculated F1 score for DeepDRiD another dataset) and the results are better than previous studies. In this paper, we improved the accuracy and F1 score of the efficient net BNs pre-trained models on the EYE-PACS dataset by applying a Gaussian Smooth filter and data augmentation transforms. Using our proposed technique, we have achieved F1 scores of 84% and 87% for EYE-PACS and DeepDRiD.
Collapse
Affiliation(s)
- Summiya Batool
- National University of Sciences and Technology, Islamabad, 44000, Pakistan
| | - Syed Omer Gilani
- National University of Sciences and Technology, Islamabad, 44000, Pakistan
| | - Asim Waris
- National University of Sciences and Technology, Islamabad, 44000, Pakistan
| | | | - Niaz B Khan
- National University of Sciences and Technology, Islamabad, 44000, Pakistan
- Mechanical Engineering Department, College of Engineering, University of Bahrain, Isa Town, 32038, Bahrain
| | - M Ijaz Khan
- Depaetment of Mechanical Engineering, Lebanese American University, Kraytem, Beirut, 1102-2801, Lebanon.
- Department of Mathematics and Statistics, Riphah International University I-14, Islamabad, 44000, Pakistan.
- Department of Mechanics and Engineering Science, Peking University, Beijing, 100871, China.
| | - Sayed M Eldin
- Faculty of Engineering, Center of Research, Future University in Egypt, New Cairo, 11835, Egypt
| | - Fuad A Awwad
- Department of Quantitative Analysis, College of Business Administration, King Saud University, P.O. Box 71115, 11587, Riyadh, Saudi Arabia
| |
Collapse
|
12
|
Ogundokun RO, Li A, Babatunde RS, Umezuruike C, Sadiku PO, Abdulahi AT, Babatunde AN. Enhancing Skin Cancer Detection and Classification in Dermoscopic Images through Concatenated MobileNetV2 and Xception Models. Bioengineering (Basel) 2023; 10:979. [PMID: 37627864 PMCID: PMC10451641 DOI: 10.3390/bioengineering10080979] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 08/04/2023] [Accepted: 08/17/2023] [Indexed: 08/27/2023] Open
Abstract
One of the most promising research initiatives in the healthcare field is focused on the rising incidence of skin cancer worldwide and improving early discovery methods for the disease. The most significant factor in the fatalities caused by skin cancer is the late identification of the disease. The likelihood of human survival may be significantly improved by performing an early diagnosis followed by appropriate therapy. It is not a simple process to extract the elements from the photographs of the tumors that may be used for the prospective identification of skin cancer. Several deep learning models are widely used to extract efficient features for a skin cancer diagnosis; nevertheless, the literature demonstrates that there is still room for additional improvements in various performance metrics. This study proposes a hybrid deep convolutional neural network architecture for identifying skin cancer by adding two main heuristics. These include Xception and MobileNetV2 models. Data augmentation was introduced to balance the dataset, and the transfer learning technique was utilized to resolve the challenges of the absence of labeled datasets. It has been detected that the suggested method of employing Xception in conjunction with MobileNetV2 attains the most excellent performance, particularly concerning the dataset that was evaluated: specifically, it produced 97.56% accuracy, 97.00% area under the curve, 100% sensitivity, 93.33% precision, 96.55% F1 score, and 0.0370 false favorable rates. This research has implications for clinical practice and public health, offering a valuable tool for dermatologists and healthcare professionals in their fight against skin cancer.
Collapse
Affiliation(s)
- Roseline Oluwaseun Ogundokun
- Department of Computer Science, Landmark University, Omu Aran 251103, Nigeria
- Department of Multimedia Engineering, Kaunas University of Technology, 44249 Kaunas, Lithuania
| | - Aiman Li
- School of Marxism, Guangzhou University of Chinese Medicine, Guangzhou 510006, China
| | | | | | - Peter O. Sadiku
- Department of Computer Science, University of Ilorin, Ilorin 240003, Nigeria
| | | | | |
Collapse
|