1
|
Song P, Wu Y. DR-ConvNeXt: DR classification method for reconstructing ConvNeXt model structure. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:448-460. [PMID: 39973787 DOI: 10.1177/08953996241311190] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
BackgroundDiabetic retinopathy (DR) is a major complication of diabetes and a leading cause of blindness among the working-age population. However, the complex distribution and variability of lesion characteristics within the dataset present significant challenges for achieving high-precision classification of DR images.ObjectiveWe propose an automatic classification method for DR images, named DR-ConvNeXt, which aims to achieve accurate diagnosis of lesion types.MethodsThe method involves designing a dual-branch addition convolution structure and appropriately increasing the number of stacked ConvNeXt Block convolution layers. Additionally, a unique primary-auxiliary loss function is introduced, contributing to a significant enhancement in DR classification accuracy within the DR-ConvNeXt model.ResultsThe model achieved an accuracy of 91.8%,sensitivity of 81.6%, and specificity of 97.9% on the APTOS dataset. On the Messidor-2 dataset, the model achieved an accuracy of 83.6%, sensitivity of 74.0%, and specificity of 94.6%.ConclusionsThe DR-ConvNeXt model's classification results on the two publicly available datasets illustrate the significant advantages in all evaluation indexes for DR classification.
Collapse
Affiliation(s)
- Pengfei Song
- State Key Laboratory of Public Big Data, Guizhou University, Guiyang, China
- College of Computer Science and Technology, Guizhou University, Guiyang, China
| | - Yun Wu
- State Key Laboratory of Public Big Data, Guizhou University, Guiyang, China
- College of Computer Science and Technology, Guizhou University, Guiyang, China
| |
Collapse
|
2
|
Shareef O, Soleimani M, Tu E, Jacobs DS, Ciolino JB, Rahdar A, Cheraqpour K, Ashraf M, Habib NB, Greenfield J, Yousefi S, Djalilian AR, Saeed HN. A novel artificial intelligence model for diagnosing Acanthamoeba keratitis through confocal microscopy. Ocul Surf 2024; 34:159-164. [PMID: 39084255 DOI: 10.1016/j.jtos.2024.07.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2024] [Revised: 07/26/2024] [Accepted: 07/28/2024] [Indexed: 08/02/2024]
Abstract
PURPOSE To develop an artificial intelligence (AI) model to diagnose Acanthamoeba keratitis (AK) based on in vivo confocal microscopy (IVCM) images extracted from the Heidelberg Retinal Tomograph 3 (HRT 3). METHODS This retrospective cohort study utilized HRT 3 IVCM images from patients who had received a culture-confirmed diagnosis of AK between 2013 and 2021 at Massachusetts Eye and Ear. Two cornea specialists independently labeled the images as AK or nonspecific finding (NSF) in a blind manner. Deep learning tasks were then conducted through Python and TensorFlow. Distinguishing between AK and NSF was designed as the task and completed through a devised convolutional neural network. RESULTS A dataset of 3312 confocal images from 17 patients with a culture-confirmed diagnosis of AK was used in this study. The inter-rater agreement for identifying the presence or absence of AK in IVCM images was 84 %, corresponding to a total of 2782 images on which both observers agreed and which were included in the model. 1242 and 1265 images of AK and NSF, respectively, were utilized in the training and validation sets, and 173 and 102 images of AK and NSF, respectively, were utilized in the evaluation set. Our model had an accuracy, sensitivity, and specificity of 76 % each, and a precision of 78 %. CONCLUSIONS We developed an HRT-based IVCM AI model for AK diagnosis utilizing culture-confirmed cases of AK. We achieved good accuracy in diagnosing AK and our model holds significant promise in the clinical application of AI in improving early AK diagnosis.
Collapse
Affiliation(s)
- Omar Shareef
- School of Engineering and Applied Sciences, Harvard College, Cambridge, MA, 02138, USA; Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, 02114, USA
| | - Mohammad Soleimani
- Department of Ophthalmology, Illinois Eye and Ear Infirmary, University of Illinois Chicago, Chicago, IL, 60612, USA; Department of Ophthalmology, The University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
| | - Elmer Tu
- Department of Ophthalmology, Illinois Eye and Ear Infirmary, University of Illinois Chicago, Chicago, IL, 60612, USA
| | - Deborah S Jacobs
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, 02114, USA
| | - Joseph B Ciolino
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, 02114, USA
| | - Amir Rahdar
- Department of Ophthalmology, The University of Tennessee Health Science Center, Memphis, TN, 38163, USA
| | - Kasra Cheraqpour
- Department of Ophthalmology, Illinois Eye and Ear Infirmary, University of Illinois Chicago, Chicago, IL, 60612, USA
| | - Mohammadali Ashraf
- Department of Ophthalmology, Illinois Eye and Ear Infirmary, University of Illinois Chicago, Chicago, IL, 60612, USA
| | - Nabiha B Habib
- Michigan State University College of Human Medicine, Grand Rapids, MI, 49503, USA
| | - Jason Greenfield
- Bascom Palmer Eye Institute, University of Miami Miller School of Medicine, Miami, FL, 33136, USA
| | - Siamak Yousefi
- Department of Ophthalmology, The University of Tennessee Health Science Center, Memphis, TN, 38163, USA
| | - Ali R Djalilian
- Department of Ophthalmology, Illinois Eye and Ear Infirmary, University of Illinois Chicago, Chicago, IL, 60612, USA
| | - Hajirah N Saeed
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, 02114, USA; Department of Ophthalmology, Illinois Eye and Ear Infirmary, University of Illinois Chicago, Chicago, IL, 60612, USA; Department of Ophthalmology, Loyola University Medical Center, Maywood, IL, 60153, USA.
| |
Collapse
|
3
|
Labib KM, Ghumman H, Jain S, Jarstad JS. A Review of the Utility and Limitations of Artificial Intelligence in Retinal Disorders and Pediatric Ophthalmology. Cureus 2024; 16:e71063. [PMID: 39380780 PMCID: PMC11459419 DOI: 10.7759/cureus.71063] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2024] [Accepted: 10/08/2024] [Indexed: 10/10/2024] Open
Abstract
Artificial intelligence (AI) is reshaping ophthalmology by enhancing diagnostic precision and treatment strategies, particularly in retinal disorders and pediatric ophthalmology. This review examines AI's efficacy in diagnosing conditions such as diabetic retinopathy (DR) and age-related macular degeneration (AMD) using imaging techniques, such as optical coherence tomography (OCT) and fundus photography. AI also shows promise in pediatric care, aiding in the screening of retinopathy of prematurity (ROP) and the management of conditions, including pediatric cataracts and strabismus. However, the integration of AI in ophthalmology presents challenges, including ethical concerns regarding algorithm biases, privacy issues, and limitations in data set quality. Addressing these challenges is crucial to ensure AI's responsible and effective deployment in clinical settings. This review synthesizes current research, underscoring AI's transformative potential in ophthalmology while highlighting critical considerations for its ethical use and technological advancement.
Collapse
Affiliation(s)
- Kristie M Labib
- Department of Ophthalmology, University of South Florida Health Morsani College of Medicine, Tampa, USA
| | - Haider Ghumman
- Department of Ophthalmology, University of South Florida Health Morsani College of Medicine, Tampa, USA
| | - Samyak Jain
- Department of Ophthalmology, University of South Florida Health Morsani College of Medicine, Tampa, USA
| | - John S Jarstad
- Department of Ophthalmology, University of South Florida Health Morsani College of Medicine, Tampa, USA
| |
Collapse
|
4
|
Li Z, Huang J, Chen J, Zeng J, Jiang H, Ding L, Zhang T, Sun W, Lu R, Zhang Q, Liang L. A deep-learning pipeline for the diagnosis and grading of common blinding ophthalmic diseases based on lesion-focused classification model. Front Artif Intell 2024; 7:1444136. [PMID: 39324131 PMCID: PMC11422385 DOI: 10.3389/frai.2024.1444136] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2024] [Accepted: 08/15/2024] [Indexed: 09/27/2024] Open
Abstract
Background Glaucoma (GLAU), Age-related Macular Degeneration (AMD), Retinal Vein Occlusion (RVO), and Diabetic Retinopathy (DR) are common blinding ophthalmic diseases worldwide. Purpose This approach is expected to enhance the early detection and treatment of common blinding ophthalmic diseases, contributing to the reduction of individual and economic burdens associated with these conditions. Methods We propose an effective deep-learning pipeline that combine both segmentation model and classification model for diagnosis and grading of four common blinding ophthalmic diseases and normal retinal fundus. Results In total, 102,786 fundus images of 75,682 individuals were used for training validation and external validation purposes. We test our model on internal validation data set, the micro Area Under the Receiver Operating Characteristic curve (AUROC) of which reached 0.995. Then, we fine-tuned the diagnosis model to classify each of the four disease into early and late stage, respectively, which achieved AUROCs of 0.597 (GL), 0.877 (AMD), 0.972 (RVO), and 0.961 (DR) respectively. To test the generalization of our model, we conducted two external validation experiments on Neimeng and Guangxi cohort, all of which maintained high accuracy. Conclusion Our algorithm demonstrates accurate artificial intelligence diagnosis pipeline for common blinding ophthalmic diseases based on Lesion-Focused fundus that overcomes the low-accuracy of the traditional classification method that based on raw retinal images, which has good generalization ability on diverse cases in different regions.
Collapse
Affiliation(s)
- Zhihuan Li
- Faculty of Medicine, Macau University of Science and Technology, Taipa, Macau, China
- Affiliated Hospital of Guangdong Medical University, Zhanjiang, China
- China Resources Power Intelligent Security Laboratory, Research Institute of Tsinghua University in Shenzhen, Shenzhen, China
| | - Junxiong Huang
- State Key Laboratory of Lunar and Planetary Sciences, Macau University of Science and Technology, Taipa, Macau, China
| | - Jingfang Chen
- Faculty of Medicine, Macau University of Science and Technology, Taipa, Macau, China
- Department of Research and Teaching, The Third People’s Hospital of Shenzhen, Shenzhen, China
| | - Jin Zeng
- The Institute for Sustainable Development, Macau University of Science and Technology, Taipa, Macau, China
- Guangzhou National Laboratory, Guangzhou, China
| | - Hong Jiang
- Faculty of Medicine, Macau University of Science and Technology, Taipa, Macau, China
- Statistical office, Zhuhai People’s Hospital (Zhuhai Clinical Medical College of Jinan University), Zhuhai, China
| | - Lin Ding
- Affiliated Hospital of Guangdong Medical University, Zhanjiang, China
| | - TianZi Zhang
- Department of Ophthalmology, Affiliated Hospital of Inner Mongolia University for the Nationalities, Tongliao, China
| | - Wen Sun
- Department of Ophthalmology Hainan Traditional Chinese Medicine Hospital, Haikou, China
| | - Rong Lu
- Yulin First People’s Hospital, Yulin, China
| | - Qiuli Zhang
- Affiliated Hospital of Guangdong Medical University, Zhanjiang, China
| | - Lizhong Liang
- Affiliated Hospital of Guangdong Medical University, Zhanjiang, China
- The Marine Biomedical Research Institute, School of Ocean and Tropical Medicine, Guangdong Medical University, Zhanjiang, Guangdong, China
| |
Collapse
|
5
|
Lim WX, Chen Z. Enhancing deep learning pre-trained networks on diabetic retinopathy fundus photographs with SLIC-G. Med Biol Eng Comput 2024; 62:2571-2583. [PMID: 38649629 DOI: 10.1007/s11517-024-03093-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2023] [Accepted: 04/10/2024] [Indexed: 04/25/2024]
Abstract
Diabetic retinopathy disease contains lesions (e.g., exudates, hemorrhages, and microaneurysms) that are minute to the naked eye. Determining the lesions at pixel level poses a challenge as each pixel does not reflect any semantic entities. Furthermore, the computational cost of inspecting each pixel is expensive because the number of pixels is high even at low resolution. In this work, we propose a hybrid image processing method. Simple Linear Iterative Clustering with Gaussian Filter (SLIC-G) for the purpose of overcoming pixel constraints. The SLIC-G image processing method is divided into two stages: (1) simple linear iterative clustering superpixel segmentation and (2) Gaussian smoothing operation. In such a way, a large number of new transformed datasets are generated and then used for model training. Finally, two performance evaluation metrics that are suitable for imbalanced diabetic retinopathy datasets were used to validate the effectiveness of the proposed SLIC-G. The results indicate that, in comparison to prior published works' results, the proposed SLIC-G shows better performance on image classification of class imbalanced diabetic retinopathy datasets. This research reveals the importance of image processing and how it influences the performance of deep learning networks. The proposed SLIC-G enhances pre-trained network performance by eliminating the local redundancy of an image, which preserves local structures, but avoids over-segmented, noisy clips. It closes the research gap by introducing the use of superpixel segmentation and Gaussian smoothing operation as image processing methods in diabetic retinopathy-related tasks.
Collapse
Affiliation(s)
- Wei Xiang Lim
- Faculty of Science and Engineering, School of Computer Science, University of Nottingham Malaysia, Semenyih, Malaysia
| | - Zhiyuan Chen
- Faculty of Science and Engineering, School of Computer Science, University of Nottingham Malaysia, Semenyih, Malaysia.
| |
Collapse
|
6
|
Zheng H, Liu X, Huang Z, Ren Y, Fu B, Shi T, Liu L, Guo Q, Tian C, Liang D, Wang R, Chen J, Hu Z. Deep learning for intracranial aneurysm segmentation using CT angiography. Phys Med Biol 2024; 69:155024. [PMID: 39008990 DOI: 10.1088/1361-6560/ad6372] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2024] [Accepted: 07/15/2024] [Indexed: 07/17/2024]
Abstract
Objective.This study aimed to employ a two-stage deep learning method to accurately detect small aneurysms (4-10 mm in size) in computed tomography angiography images.Approach.This study included 956 patients from 6 hospitals and a public dataset obtained with 6 CT scanners from different manufacturers. The proposed method consists of two components: a lightweight and fast head region selection (HRS) algorithm and an adaptive 3D nnU-Net network, which is used as the main architecture for segmenting aneurysms. Segments generated by the deep neural network were compared with expert-generated manual segmentation results and assessed using Dice scores.MainResults.The area under the curve (AUC) exceeded 79% across all datasets. In particular, the precision and AUC reached 85.2% and 87.6%, respectively, on certain datasets. The experimental results demonstrated the promising performance of this approach, which reduced the inference time by more than 50% compared to direct inference without HRS.Significance.Compared with a model without HRS, the deep learning approach we developed can accurately segment aneurysms by automatically localizing brain regions and can accelerate aneurysm inference by more than 50%.
Collapse
Affiliation(s)
- Huizhong Zheng
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
| | - Xinfeng Liu
- Department of Radiology, Guizhou Provincial People's Hospital, Guiyang 550002, People's Republic of China
| | - Zhenxing Huang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
| | - Yan Ren
- AI for Science (AI4S)-Preferred Program, Peking University Shenzhen Graduate School, Shenzhen 518055, People's Republic of China
- School of Electronic and Computer Engineering, Peking University Shenzhen Graduate School, Shenzhen 518005, People's Republic of China
| | - Bin Fu
- AI for Science (AI4S)-Preferred Program, Peking University Shenzhen Graduate School, Shenzhen 518055, People's Republic of China
- School of Electronic and Computer Engineering, Peking University Shenzhen Graduate School, Shenzhen 518005, People's Republic of China
| | - Tianliang Shi
- Department of Radiology, Tongren Municipal People's Hospital, Tongren, Guizhou 554300, People's Republic of China
| | - Lu Liu
- Department of Radiology, The Second People's Hospital of Guiyang, Guiyang, Guizhou 550002, People's Republic of China
| | - Qiping Guo
- Department of Radiology, Xingyi Municipal People's Hospital, Xingyi, Guizhou 562400, People's Republic of China
| | - Chong Tian
- Department of Radiology, Guizhou Provincial People's Hospital, Guiyang 550002, People's Republic of China
| | - Dong Liang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
| | - Rongpin Wang
- Department of Radiology, Guizhou Provincial People's Hospital, Guiyang 550002, People's Republic of China
| | - Jie Chen
- AI for Science (AI4S)-Preferred Program, Peking University Shenzhen Graduate School, Shenzhen 518055, People's Republic of China
- School of Electronic and Computer Engineering, Peking University Shenzhen Graduate School, Shenzhen 518005, People's Republic of China
- Peng Cheng Laboratory, Shenzhen 518005, People's Republic of China
| | - Zhanli Hu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
| |
Collapse
|
7
|
Chen T, Bai Y, Mao H, Liu S, Xu K, Xiong Z, Ma S, Yang F, Zhao Y. Cross-modality transfer learning with knowledge infusion for diabetic retinopathy grading. Front Med (Lausanne) 2024; 11:1400137. [PMID: 38808141 PMCID: PMC11130363 DOI: 10.3389/fmed.2024.1400137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2024] [Accepted: 04/15/2024] [Indexed: 05/30/2024] Open
Abstract
Background Ultra-wide-field (UWF) fundus photography represents an emerging retinal imaging technique offering a broader field of view, thus enhancing its utility in screening and diagnosing various eye diseases, notably diabetic retinopathy (DR). However, the application of computer-aided diagnosis for DR using UWF images confronts two major challenges. The first challenge arises from the limited availability of labeled UWF data, making it daunting to train diagnostic models due to the high cost associated with manual annotation of medical images. Secondly, existing models' performance requires enhancement due to the absence of prior knowledge to guide the learning process. Purpose By leveraging extensively annotated datasets within the field, which encompass large-scale, high-quality color fundus image datasets annotated at either image-level or pixel-level, our objective is to transfer knowledge from these datasets to our target domain through unsupervised domain adaptation. Methods Our approach presents a robust model for assessing the severity of diabetic retinopathy (DR) by leveraging unsupervised lesion-aware domain adaptation in ultra-wide-field (UWF) images. Furthermore, to harness the wealth of detailed annotations in publicly available color fundus image datasets, we integrate an adversarial lesion map generator. This generator supplements the grading model by incorporating auxiliary lesion information, drawing inspiration from the clinical methodology of evaluating DR severity by identifying and quantifying associated lesions. Results We conducted both quantitative and qualitative evaluations of our proposed method. In particular, among the six representative DR grading methods, our approach achieved an accuracy (ACC) of 68.18% and a precision (pre) of 67.43%. Additionally, we conducted extensive experiments in ablation studies to validate the effectiveness of each component of our proposed method. Conclusion In conclusion, our method not only improves the accuracy of DR grading, but also enhances the interpretability of the results, providing clinicians with a reliable DR grading scheme.
Collapse
Affiliation(s)
- Tao Chen
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Yanmiao Bai
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Haiting Mao
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Shouyue Liu
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Keyi Xu
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Zhouwei Xiong
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Shaodong Ma
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Fang Yang
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| | - Yitian Zhao
- Cixi Biomedical Research Institute, Wenzhou Medical University, Ningbo, China
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of SciencesNingbo, China
| |
Collapse
|
8
|
Didier AJ, Nigro A, Noori Z, Omballi MA, Pappada SM, Hamouda DM. Application of machine learning for lung cancer survival prognostication-A systematic review and meta-analysis. Front Artif Intell 2024; 7:1365777. [PMID: 38646415 PMCID: PMC11026647 DOI: 10.3389/frai.2024.1365777] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2024] [Accepted: 03/18/2024] [Indexed: 04/23/2024] Open
Abstract
Introduction Machine learning (ML) techniques have gained increasing attention in the field of healthcare, including predicting outcomes in patients with lung cancer. ML has the potential to enhance prognostication in lung cancer patients and improve clinical decision-making. In this systematic review and meta-analysis, we aimed to evaluate the performance of ML models compared to logistic regression (LR) models in predicting overall survival in patients with lung cancer. Methods We followed the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) statement. A comprehensive search was conducted in Medline, Embase, and Cochrane databases using a predefined search query. Two independent reviewers screened abstracts and conflicts were resolved by a third reviewer. Inclusion and exclusion criteria were applied to select eligible studies. Risk of bias assessment was performed using predefined criteria. Data extraction was conducted using the Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modeling Studies (CHARMS) checklist. Meta-analytic analysis was performed to compare the discriminative ability of ML and LR models. Results The literature search resulted in 3,635 studies, and 12 studies with a total of 211,068 patients were included in the analysis. Six studies reported confidence intervals and were included in the meta-analysis. The performance of ML models varied across studies, with C-statistics ranging from 0.60 to 0.85. The pooled analysis showed that ML models had higher discriminative ability compared to LR models, with a weighted average C-statistic of 0.78 for ML models compared to 0.70 for LR models. Conclusion Machine learning models show promise in predicting overall survival in patients with lung cancer, with superior discriminative ability compared to logistic regression models. However, further validation and standardization of ML models are needed before their widespread implementation in clinical practice. Future research should focus on addressing the limitations of the current literature, such as potential bias and heterogeneity among studies, to improve the accuracy and generalizability of ML models for predicting outcomes in patients with lung cancer. Further research and development of ML models in this field may lead to improved patient outcomes and personalized treatment strategies.
Collapse
Affiliation(s)
- Alexander J. Didier
- Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Anthony Nigro
- Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Zaid Noori
- Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
- Division of Pulmonary and Critical Care Medicine, Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Mohamed A. Omballi
- Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
- Division of Pulmonary and Critical Care Medicine, Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Scott M. Pappada
- Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
- Department of Anesthesiology, The University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Danae M. Hamouda
- Department of Medicine, University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
- Division of Hematology and Oncology, Department of Medicine, The University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| |
Collapse
|
9
|
Umapathy L, Brown T, Mushtaq R, Greenhill M, Lu J, Martin D, Altbach M, Bilgin A. Reducing annotation burden in MR: A novel MR-contrast guided contrastive learning approach for image segmentation. Med Phys 2024; 51:2707-2720. [PMID: 37956263 PMCID: PMC10994772 DOI: 10.1002/mp.16820] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 08/24/2023] [Accepted: 10/06/2023] [Indexed: 11/15/2023] Open
Abstract
BACKGROUND Contrastive learning, a successful form of representational learning, has shown promising results in pretraining deep learning (DL) models for downstream tasks. When working with limited annotation data, as in medical image segmentation tasks, learning domain-specific local representations can further improve the performance of DL models. PURPOSE In this work, we extend the contrastive learning framework to utilize domain-specific contrast information from unlabeled Magnetic Resonance (MR) images to improve the performance of downstream MR image segmentation tasks in the presence of limited labeled data. METHODS The contrast in MR images is controlled by underlying tissue properties (e.g., T1 or T2) and image acquisition parameters. We hypothesize that learning to discriminate local representations based on underlying tissue properties should improve subsequent segmentation tasks on MR images. We propose a novel constrained contrastive learning (CCL) strategy that uses tissue-specific information via a constraint map to define positive and negative local neighborhoods for contrastive learning, embedding this information in the representational space during pretraining. For a given MR contrast image, the proposed strategy uses local signal characteristics (constraint map) across a set of related multi-contrast MR images as a surrogate for underlying tissue information. We demonstrate the utility of the approach for downstream: (1) multi-organ segmentation tasks in T2-weighted images where a DL model learns T2 information with constraint maps from a set of 2D multi-echo T2-weighted images (n = 101) and (2) tumor segmentation tasks in multi-parametric images from the public brain tumor segmentation (BraTS) (n = 80) dataset where DL models learn T1 and T2 information from multi-parametric BraTS images. Performance is evaluated on downstream multi-label segmentation tasks with limited data in (1) T2-weighted images of the abdomen from an in-house Radial-T2 (Train/Test = 30/20), (2) public Cartesian-T2 (Train/Test = 6/12) dataset, and (3) multi-parametric MR images from the public brain tumor segmentation dataset (BraTS) (Train/Test = 40/50). The performance of the proposed CCL strategy is compared to state-of-the-art self-supervised contrastive learning techniques. In each task, a model is also trained using all available labeled data for supervised baseline performance. RESULTS The proposed CCL strategy consistently yielded improved Dice scores, Precision, and Recall metrics, and reduced HD95 values across all segmentation tasks. We also observed performance comparable to the baseline with reduced annotation effort. The t-SNE visualization of features for T2-weighted images demonstrates its ability to embed T2 information in the representational space. On the BraTS dataset, we also observed that using an appropriate multi-contrast space to learn T1+T2, T1, or T2 information during pretraining further improved the performance of tumor segmentation tasks. CONCLUSIONS Learning to embed tissue-specific information that controls MR image contrast with the proposed constrained contrastive learning improved the performance of DL models on subsequent segmentation tasks compared to conventional self-supervised contrastive learning techniques. The use of such domain-specific local representations could help understand, improve performance, and mitigate the scarcity of labeled data in MR image segmentation tasks.
Collapse
Affiliation(s)
- Lavanya Umapathy
- Department of Electrical and Computer Engineering, University of Arizona, Tucson, AZ, United States
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
- Department of Radiology, Center for Advanced Imaging Innovation and Research (CAI2R), New York University Grossman School of Medicine, New York, NY, United States
| | - Taylor Brown
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
- College of Medicine, University of Arizona, Tucson, AZ, United States
| | - Raza Mushtaq
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
- College of Medicine, University of Arizona, Tucson, AZ, United States
| | - Mark Greenhill
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
- College of Medicine, University of Arizona, Tucson, AZ, United States
| | - J’rick Lu
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
- College of Medicine, University of Arizona, Tucson, AZ, United States
| | - Diego Martin
- Department of Radiology, Houston Methodist Hospital, Houston, TX, United States
| | - Maria Altbach
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
| | - Ali Bilgin
- Department of Electrical and Computer Engineering, University of Arizona, Tucson, AZ, United States
- Department of Medical Imaging, University of Arizona, Tucson, AZ, United States
- Program in Applied Mathematics, University of Arizona, Tucson, AZ, United States
- Department of Biomedical Engineering, University of Arizona, Tucson, AZ, United States
| |
Collapse
|
10
|
Heger KA, Waldstein SM. Artificial intelligence in retinal imaging: current status and future prospects. Expert Rev Med Devices 2024; 21:73-89. [PMID: 38088362 DOI: 10.1080/17434440.2023.2294364] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2023] [Accepted: 12/09/2023] [Indexed: 12/19/2023]
Abstract
INTRODUCTION The steadily growing and aging world population, in conjunction with continuously increasing prevalences of vision-threatening retinal diseases, is placing an increasing burden on the global healthcare system. The main challenges within retinology involve identifying the comparatively few patients requiring therapy within the large mass, the assurance of comprehensive screening for retinal disease and individualized therapy planning. In order to sustain high-quality ophthalmic care in the future, the incorporation of artificial intelligence (AI) technologies into our clinical practice represents a potential solution. AREAS COVERED This review sheds light onto already realized and promising future applications of AI techniques in retinal imaging. The main attention is directed at the application in diabetic retinopathy and age-related macular degeneration. The principles of use in disease screening, grading, therapeutic planning and prediction of future developments are explained based on the currently available literature. EXPERT OPINION The recent accomplishments of AI in retinal imaging indicate that its implementation into our daily practice is likely to fundamentally change the ophthalmic healthcare system and bring us one step closer to the goal of individualized treatment. However, it must be emphasized that the aim is to optimally support clinicians by gradually incorporating AI approaches, rather than replacing ophthalmologists.
Collapse
Affiliation(s)
- Katharina A Heger
- Department of Ophthalmology, Landesklinikum Mistelbach-Gaenserndorf, Mistelbach, Austria
| | - Sebastian M Waldstein
- Department of Ophthalmology, Landesklinikum Mistelbach-Gaenserndorf, Mistelbach, Austria
| |
Collapse
|
11
|
Xia H, Long J, Song S, Tan Y. Multi-scale multi-attention network for diabetic retinopathy grading. Phys Med Biol 2023; 69:015007. [PMID: 38035368 DOI: 10.1088/1361-6560/ad111d] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Accepted: 11/30/2023] [Indexed: 12/02/2023]
Abstract
Objective.Diabetic retinopathy (DR) grading plays an important role in clinical diagnosis. However, automatic grading of DR is challenging due to the presence of intra-class variation and small lesions. On the one hand, deep features learned by convolutional neural networks often lose valid information about these small lesions. On the other hand, the great variability of lesion features, including differences in type and quantity, can exhibit considerable divergence even among fundus images of the same grade. To address these issues, we propose a novel multi-scale multi-attention network (MMNet).Approach.Firstly, to focus on different lesion features of fundus images, we propose a lesion attention module, which aims to encode multiple different lesion attention feature maps by combining channel attention and spatial attention, thus extracting global feature information and preserving diverse lesion features. Secondly, we propose a multi-scale feature fusion module to learn more feature information for small lesion regions, which combines complementary relationships between different convolutional layers to capture more detailed feature information. Furthermore, we introduce a Cross-layer Consistency Constraint Loss to overcome semantic differences between multi-scale features.Main results.The proposed MMNet obtains a high accuracy of 86.4% and a high kappa score of 88.4% for multi-class DR grading tasks on the EyePACS dataset, while 98.6% AUC, 95.3% accuracy, 92.7% recall, 95.0% precision, and 93.3% F1-score for referral and non-referral classification on the Messidor-1 dataset. Extensive experiments on two challenging benchmarks demonstrate that our MMNet achieves significant improvements and outperforms other state-of-the-art DR grading methods.Significance.MMNet has improved the diagnostic efficiency and accuracy of diabetes retinopathy and promoted the application of computer-aided medical diagnosis in DR screening.
Collapse
Affiliation(s)
- Haiying Xia
- School of Electronic and Information Engineering, Guangxi Normal University, Guilin 541004, People's Republic of China
| | - Jie Long
- School of Electronic and Information Engineering, Guangxi Normal University, Guilin 541004, People's Republic of China
| | - Shuxiang Song
- School of Electronic and Information Engineering, Guangxi Normal University, Guilin 541004, People's Republic of China
| | - Yumei Tan
- School of Computer Science and Engineering, Guangxi Normal University, Guilin 541004, People's Republic of China
| |
Collapse
|
12
|
Dao QT, Trinh HQ, Nguyen VA. An effective and comprehensible method to detect and evaluate retinal damage due to diabetes complications. PeerJ Comput Sci 2023; 9:e1585. [PMID: 37810367 PMCID: PMC10557496 DOI: 10.7717/peerj-cs.1585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Accepted: 08/20/2023] [Indexed: 10/10/2023]
Abstract
The leading cause of vision loss globally is diabetic retinopathy. Researchers are making great efforts to automatically detect and diagnose correctly diabetic retinopathy. Diabetic retinopathy includes five stages: no diabetic retinopathy, mild diabetic retinopathy, moderate diabetic retinopathy, severe diabetic retinopathy and proliferative diabetic retinopathy. Recent studies have offered several multi-tasking deep learning models to detect and assess the level of diabetic retinopathy. However, the explanation for the assessment of disease severity of these models is limited, and only stops at showing lesions through images. These studies have not explained on what basis the appraisal of disease severity is based. In this article, we present a system for assessing and interpreting the five stages of diabetic retinopathy. The proposed system is built from internal models including a deep learning model that detects lesions and an explanatory model that assesses disease stage. The deep learning model that detects lesions uses the Mask R-CNN deep learning network to specify the location and shape of the lesion and classify the lesion types. This model is a combination of two networks: one used to detect hemorrhagic and exudative lesions, and one used to detect vascular lesions like aneurysm and proliferation. The explanatory model appraises disease severity based on the severity of each type of lesion and the association between types. The severity of the disease will be decided by the model based on the number of lesions, the density and the area of the lesions. The experimental results on real-world datasets show that our proposed method achieves high accuracy of assessing five stages of diabetic retinopathy comparable to existing state-of-the-art methods and is capable of explaining the causes of disease severity.
Collapse
Affiliation(s)
- Quang Toan Dao
- Institute of Information Technology, Vietnam Academy of Science and Technology, Hanoi, Vietnam
| | - Hoang Quan Trinh
- Vietnam Space Center, Vietnam Academy of Science and Technology, Hanoi, Vietnam
| | - Viet Anh Nguyen
- Institute of Information Technology, Vietnam Academy of Science and Technology, Hanoi, Vietnam
| |
Collapse
|
13
|
Wang J, Horlacher M, Cheng L, Winther O. RNA trafficking and subcellular localization-a review of mechanisms, experimental and predictive methodologies. Brief Bioinform 2023; 24:bbad249. [PMID: 37466130 PMCID: PMC10516376 DOI: 10.1093/bib/bbad249] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Revised: 05/30/2023] [Accepted: 06/16/2023] [Indexed: 07/20/2023] Open
Abstract
RNA localization is essential for regulating spatial translation, where RNAs are trafficked to their target locations via various biological mechanisms. In this review, we discuss RNA localization in the context of molecular mechanisms, experimental techniques and machine learning-based prediction tools. Three main types of molecular mechanisms that control the localization of RNA to distinct cellular compartments are reviewed, including directed transport, protection from mRNA degradation, as well as diffusion and local entrapment. Advances in experimental methods, both image and sequence based, provide substantial data resources, which allow for the design of powerful machine learning models to predict RNA localizations. We review the publicly available predictive tools to serve as a guide for users and inspire developers to build more effective prediction models. Finally, we provide an overview of multimodal learning, which may provide a new avenue for the prediction of RNA localization.
Collapse
Affiliation(s)
- Jun Wang
- Bioinformatics Centre, Department of Biology, University of Copenhagen, København Ø 2100, Denmark
| | - Marc Horlacher
- Computational Health Center, Helmholtz Center, Munich, Germany
| | - Lixin Cheng
- Shenzhen People’s Hospital, First Affiliated Hospital of Southern University of Science and Technology, Second Clinical Medicine College of Jinan University, Shenzhen 518020, China
| | - Ole Winther
- Bioinformatics Centre, Department of Biology, University of Copenhagen, København Ø 2100, Denmark
- Center for Genomic Medicine, Rigshospitalet (Copenhagen University Hospital), Copenhagen 2100, Denmark
- Section for Cognitive Systems, Department of Applied Mathematics and Computer Science, Technical University of Denmark, Kongens Lyngby 2800, Denmark
| |
Collapse
|
14
|
Dar JA, Srivastava KK, Mishra A. Lung anomaly detection from respiratory sound database (sound signals). Comput Biol Med 2023; 164:107311. [PMID: 37552916 DOI: 10.1016/j.compbiomed.2023.107311] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Revised: 07/01/2023] [Accepted: 07/28/2023] [Indexed: 08/10/2023]
Abstract
Chest or upper body auscultation has long been considered a useful part of the physical examination going back to the time of Hippocrates. However, it did not become a prevalent practice until the invention of the stethoscope by Rene Laennec in 1816, which made the practice suitable and hygienic. Pulmonary disease is a kind of sickness that affects the lungs and various parts of the respiratory system. Lung diseases are the third largest cause of death in the world. According to the World Health Organization (WHO), the five major respiratory diseases, namely chronic obstructive pulmonary disease (COPD), tuberculosis, acute lower respiratory tract infection (LRTI), asthma, and lung cancer, cause the death of more than 3 million people each year worldwide. Respiratory sounds disclose significant information regarding the lungs of patients. Numerous methods are developed for analyzing the lung sounds. However, clinical approaches require qualified pulmonologists to diagnose such kind of signals appropriately and are also time consuming. Hence, an efficient Fractional Water Cycle Swarm Optimizer-based Deep Residual Network (Fr-WCSO-based DRN) is developed in this research for detecting the pulmonary abnormalities using respiratory sounds signals. The proposed Fr-WCSO is newly designed by the incorporation of Fractional Calculus (FC) and Water Cycle Swarm Optimizer WCSO. Meanwhile, WCSO is the combination of Water Cycle Algorithm (WCA) with Competitive Swarm Optimizer (CSO). The respiratory input sound signals are pre-processed and the important features needed for the further processing are effectively extracted. With the extracted features, data augmentation is carried out for minimizing the over fitting issues for improving the overall detection performance. Once data augmentation is done, feature selection is performed using proposed Fr-WCSO algorithm. Finally, pulmonary abnormality detection is performed using DRN where the training procedure of DRN is performed using the developed Fr-WCSO algorithm. The developed method achieved superior performance by considering the evaluation measures, namely True Positive Rate (TPR), True Negative Rate (TNR) and testing accuracy with the values of 0.963(96.3%), 0.932,(93.2%) and 0.948(94.8%), respectively.
Collapse
Affiliation(s)
- Jawad Ahmad Dar
- Department of Computer Science and Engineering, Mansarovar Global University, Madhya Pradesh, India.
| | - Kamal Kr Srivastava
- Department of Information Technology at Babu Banarasi Das Northern India Institute of Technology, Lucknow, India.
| | - Alok Mishra
- Department of Physics, Gaya College of Engineering, Gaya, India.
| |
Collapse
|
15
|
Wang JZ, Lu NH, Du WC, Liu KY, Hsu SY, Wang CY, Chen YJ, Chang LC, Twan WH, Chen TB, Huang YH. Classification of Color Fundus Photographs Using Fusion Extracted Features and Customized CNN Models. Healthcare (Basel) 2023; 11:2228. [PMID: 37570467 PMCID: PMC10418900 DOI: 10.3390/healthcare11152228] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2023] [Revised: 08/04/2023] [Accepted: 08/05/2023] [Indexed: 08/13/2023] Open
Abstract
This study focuses on overcoming challenges in classifying eye diseases using color fundus photographs by leveraging deep learning techniques, aiming to enhance early detection and diagnosis accuracy. We utilized a dataset of 6392 color fundus photographs across eight disease categories, which was later augmented to 17,766 images. Five well-known convolutional neural networks (CNNs)-efficientnetb0, mobilenetv2, shufflenet, resnet50, and resnet101-and a custom-built CNN were integrated and trained on this dataset. Image sizes were standardized, and model performance was evaluated via accuracy, Kappa coefficient, and precision metrics. Shufflenet and efficientnetb0demonstrated strong performances, while our custom 17-layer CNN outperformed all with an accuracy of 0.930 and a Kappa coefficient of 0.920. Furthermore, we found that the fusion of image features with classical machine learning classifiers increased the performance, with Logistic Regression showcasing the best results. Our study highlights the potential of AI and deep learning models in accurately classifying eye diseases and demonstrates the efficacy of custom-built models and the fusion of deep learning and classical methods. Future work should focus on validating these methods across larger datasets and assessing their real-world applicability.
Collapse
Affiliation(s)
- Jing-Zhe Wang
- Department of Information Engineering, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 84001, Taiwan
| | - Nan-Han Lu
- Department of Medical Imaging and Radiological Science, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 82445, Taiwan
- Department of Radiology, E-DA Cancer Hospital, I-Shou University, No. 21, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 82445, Taiwan
| | - Wei-Chang Du
- Department of Information Engineering, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 84001, Taiwan
| | - Kuo-Ying Liu
- Department of Radiology, E-DA Cancer Hospital, I-Shou University, No. 21, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 82445, Taiwan
| | - Shih-Yen Hsu
- Department of Information Engineering, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 84001, Taiwan
| | - Chi-Yuan Wang
- Department of Medical Imaging and Radiological Science, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 82445, Taiwan
| | - Yun-Ju Chen
- School of Medicine for International Students, I-Shu University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 84001, Taiwan
| | - Li-Ching Chang
- School of Medicine for International Students, I-Shu University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 84001, Taiwan
| | - Wen-Hung Twan
- Department of Life Sciences, National Taitung University, No. 369, Sec. 2, University Road, Taitung City 95048, Taiwan
| | - Tai-Been Chen
- Department of Medical Imaging and Radiological Science, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 82445, Taiwan
- Institute of Statistics, National Yang Ming Chiao Tung University, No. 1001, University Road, Hsinchu 30010, Taiwan
| | - Yung-Hui Huang
- Department of Medical Imaging and Radiological Science, I-Shou University, No. 8, Yida Road, Jiao-Su Village, Yan-Chao District, Kaohsiung City 82445, Taiwan
| |
Collapse
|
16
|
Tian M, Wang H, Sun Y, Wu S, Tang Q, Zhang M. Fine-grained attention & knowledge-based collaborative network for diabetic retinopathy grading. Heliyon 2023; 9:e17217. [PMID: 37449186 PMCID: PMC10336422 DOI: 10.1016/j.heliyon.2023.e17217] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 06/09/2023] [Accepted: 06/10/2023] [Indexed: 07/18/2023] Open
Abstract
Accurate diabetic retinopathy (DR) grading is crucial for making the proper treatment plan to reduce the damage caused by vision loss. This task is challenging due to the fact that the DR related lesions are often small and subtle in visual differences and intra-class variations. Moreover, relationships between the lesions and the DR levels are complicated. Although many deep learning (DL) DR grading systems have been developed with some success, there are still rooms for grading accuracy improvement. A common issue is that not much medical knowledge was used in these DL DR grading systems. As a result, the grading results are not properly interpreted by ophthalmologists, thus hinder the potential for practical applications. This paper proposes a novel fine-grained attention & knowledge-based collaborative network (FA+KC-Net) to address this concern. The fine-grained attention network dynamically divides the extracted feature maps into smaller patches and effectively captures small image features that are meaningful in the sense of its training from large amount of retinopathy fundus images. The knowledge-based collaborative network extracts a-priori medical knowledge features, i.e., lesions such as the microaneurysms (MAs), soft exudates (SEs), hard exudates (EXs), and hemorrhages (HEs). Finally, decision rules are developed to fuse the DR grading results from the fine-grained network and the knowledge-based collaborative network to make the final grading. Extensive experiments are carried out on four widely-used datasets, the DDR, Messidor, APTOS, and EyePACS to evaluate the efficacy of our method and compare with other state-of-the-art (SOTA) DL models. Simulation results show that proposed FA+KC-Net is accurate and stable, achieves the best performances on the DDR, Messidor, and APTOS datasets.
Collapse
Affiliation(s)
- Miao Tian
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Hongqiu Wang
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Yingxue Sun
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Shaozhi Wu
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Qingqing Tang
- Department of Ophthalmology, West China Hospital, Sichuan University, Chengdu, 610041, China
| | - Meixia Zhang
- Department of Ophthalmology, West China Hospital, Sichuan University, Chengdu, 610041, China
| |
Collapse
|
17
|
Hemanth SV, Alagarsamy S. Hybrid adaptive deep learning classifier for early detection of diabetic retinopathy using optimal feature extraction and classification. J Diabetes Metab Disord 2023; 22:881-895. [PMID: 37255780 PMCID: PMC10225400 DOI: 10.1007/s40200-023-01220-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/06/2022] [Accepted: 03/29/2023] [Indexed: 06/01/2023]
Abstract
Objectives Diabetic retinopathy (DR) is one of the leading causes of blindness. It is important to use a comprehensive learning method to identify the DR. However, comprehensive learning methods often rely heavily on encrypted data, which can be costly and time consuming. Also, the DR function is not displayed and is scattered in the high-definition image below. Methods Therefore, learning how to distribute such DR functions is a big challenge. In this work, we proposed a hybrid adaptive deep learning classifier for early detection of diabetic retinopathy (HADL-DR). First, we provide an improved multichannel-based generative adversarial network (MGAN) with semi-maintenance to detect blood vessels segmentation. Results By reducing the reliance on the encoded data, the following high-resolution images can be used to detect the indivisible features of some semi-observed MGAN references. Scale invariant feature transform (SIFT) function is then extracted and the best function is selected using the improved sequential approximation optimization (SAO) algorithm. After that, a hybrid recurrent neural network with long short-term memory (RNN-LSTM) is utilized for DR classification. The proposed RNN-LSTM classifier evaluated through standard benchmark Kaggle and Messidor datasets. Conclusion Finally, the simulation results are compared with the existing state-of-art classifiers in terms of accuracy, precision, recall, f-measure and area under cover (AUC), it is seen that more successful results are obtained.
Collapse
Affiliation(s)
- S V Hemanth
- Department of Computer Science and Engineering, Kalasalingam Academy of Research and Education (Deemed to Be University), Krishnankoil, TamilNadu India
| | - Saravanan Alagarsamy
- Department of Computer Science and Engineering, Kalasalingam Academy of Research and Education (Deemed to Be University), Krishnankoil, TamilNadu India
| |
Collapse
|
18
|
Chen D, Li X, Li S. A Novel Convolutional Neural Network Model Based on Beetle Antennae Search Optimization Algorithm for Computerized Tomography Diagnosis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:1418-1429. [PMID: 34460391 DOI: 10.1109/tnnls.2021.3105384] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Convolutional neural networks (CNNs) are widely used in the field of medical imaging diagnosis but have the disadvantages of slow training speed and low diagnostic accuracy due to the initialization of parameters before training. In this article, a CNN optimization method based on the beetle antennae search (BAS) optimization algorithm is proposed. The method optimizes the initial parameters of the CNN through the BAS optimization algorithm. Based on this optimization approach, a novel CNN model with a pretrained BAS optimization algorithm was developed and applied to the analysis and diagnosis of medical imaging data for intracranial hemorrhage. Experimental results on 330 test images show that the proposed method has a better diagnostic performance than the traditional CNN. The proposed method achieves a diagnostic accuracy of 93.9394% and 100% recall, and the diagnosis of 66 human head computerized tomography image data only takes 0.1596 s. Moreover, the proposed method has more advantages than the three other optimization algorithms.
Collapse
|
19
|
Detecting red-lesions from retinal fundus images using unique morphological features. Sci Rep 2023; 13:3487. [PMID: 36859429 PMCID: PMC9977778 DOI: 10.1038/s41598-023-30459-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Accepted: 02/23/2023] [Indexed: 03/03/2023] Open
Abstract
One of the most important retinal diseases is Diabetic Retinopathy (DR) which can lead to serious damage to vision if remains untreated. Red-lesions are from important demonstrations of DR helping its identification in early stages. The detection and verification of them is helpful in the evaluation of disease severity and progression. In this paper, a novel image processing method is proposed for extracting red-lesions from fundus images. The method works based on finding and extracting the unique morphological features of red-lesions. After quality improvement of images, a pixel-based verification is performed in the proposed method to find the ones which provide a significant intensity change in a curve-like neighborhood. In order to do so, a curve is considered around each pixel and the intensity changes around the curve boundary are considered. The pixels for which it is possible to find such curves in at least two directions are considered as parts of red-lesions. The simplicity of computations, the high accuracy of results, and no need to post-processing operations are the important characteristics of the proposed method endorsing its good performance.
Collapse
|
20
|
Deep learning-based hemorrhage detection for diabetic retinopathy screening. Sci Rep 2023; 13:1479. [PMID: 36707608 PMCID: PMC9883230 DOI: 10.1038/s41598-023-28680-3] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Accepted: 01/23/2023] [Indexed: 01/29/2023] Open
Abstract
Diabetic retinopathy is a retinal compilation that causes visual impairment. Hemorrhage is one of the pathological symptoms of diabetic retinopathy that emerges during disease development. Therefore, hemorrhage detection reveals the presence of diabetic retinopathy in the early phase. Diagnosing the disease in its initial stage is crucial to adopt proper treatment so the repercussions can be prevented. The automatic deep learning-based hemorrhage detection method is proposed that can be used as the second interpreter for ophthalmologists to reduce the time and complexity of conventional screening methods. The quality of the images was enhanced, and the prospective hemorrhage locations were estimated in the preprocessing stage. Modified gamma correction adaptively illuminates fundus images by using gradient information to address the nonuniform brightness levels of images. The algorithm estimated the locations of potential candidates by using a Gaussian match filter, entropy thresholding, and mathematical morphology. The required objects were segmented using the regional diversity at estimated locations. The novel hemorrhage network is propounded for hemorrhage classification and compared with the renowned deep models. Two datasets benchmarked the model's performance using sensitivity, specificity, precision, and accuracy metrics. Despite being the shallowest network, the proposed network marked competitive results than LeNet-5, AlexNet, ResNet50, and VGG-16. The hemorrhage network was assessed using training time and classification accuracy through synthetic experimentation. Results showed promising accuracy in the classification stage while significantly reducing training time. The research concluded that increasing deep network layers does not guarantee good results but rather increases training time. The suitable architecture of a deep model and its appropriate parameters are critical for obtaining excellent outcomes.
Collapse
|
21
|
Luo X, Wang W, Xu Y, Lai Z, Jin X, Zhang B, Zhang D. A deep convolutional neural network for diabetic retinopathy detection via mining local and long‐range dependence. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY 2023. [DOI: 10.1049/cit2.12155] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/26/2023] Open
Affiliation(s)
- Xiaoling Luo
- Shenzhen Key Laboratory of Visual Object Detection and Recognition Harbin Institute of Technology Shenzhen China
| | - Wei Wang
- Shenzhen Key Laboratory of Visual Object Detection and Recognition Harbin Institute of Technology Shenzhen China
| | - Yong Xu
- Shenzhen Key Laboratory of Visual Object Detection and Recognition Harbin Institute of Technology Shenzhen China
- Peng Cheng Laboratory Shenzhen China
| | - Zhihui Lai
- Shenzhen Institute of Artificial Intelligence and Robotics for Society Shenzhen China
| | - Xiaopeng Jin
- College of Big Data and Internet Shenzhen Technology University Shenzhen China
| | - Bob Zhang
- The Department of Computer and Information Science University of Macau Macao Macau
| | - David Zhang
- The Chinese University of Hong Kong (Shenzhen) Shenzhen China
| |
Collapse
|
22
|
Zhao S, Wu Y, Tong M, Yao Y, Qian W, Qi S. CoT-XNet: contextual transformer with Xception network for diabetic retinopathy grading. Phys Med Biol 2022; 67. [PMID: 36322995 DOI: 10.1088/1361-6560/ac9fa0] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Accepted: 11/01/2022] [Indexed: 11/07/2022]
Abstract
Objective.Diabetic retinopathy (DR) grading is primarily performed by assessing fundus images. Many types of lesions, such as microaneurysms, hemorrhages, and soft exudates, are available simultaneously in a single image. However, their sizes may be small, making it difficult to differentiate adjacent DR grades even using deep convolutional neural networks (CNNs). Recently, a vision transformer has shown comparable or even superior performance to CNNs, and it also learns different visual representations from CNNs. Inspired by this finding, we propose a two-path contextual transformer with Xception network (CoT-XNet) to improve the accuracy of DR grading.Approach.The representations learned by CoT through one path and those by the Xception network through another path are concatenated before the fully connected layer. Meanwhile, the dedicated pre-processing, data resampling, and test time augmentation strategies are implemented. The performance of CoT-XNet is evaluated in the publicly available datasets of DDR, APTOS2019, and EyePACS, which include over 50 000 images. Ablation experiments and comprehensive comparisons with various state-of-the-art (SOTA) models have also been performed.Main results.Our proposed CoT-XNet shows better performance than available SOTA models, and the accuracy and Kappa are 83.10% and 0.8496, 84.18% and 0.9000 and 84.10% and 0.7684 respectively, in the three datasets (listed above). Class activation maps of CoT and Xception networks are different and complementary in most images.Significance.By concatenating the different visual representations learned by CoT and Xception networks, CoT-XNet can accurately grade DR from fundus images and present good generalizability. CoT-XNet will promote the application of artificial intelligence-based systems in the DR screening of large-scale populations.
Collapse
Affiliation(s)
- Shuiqing Zhao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, People's Republic of China.,Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, People's Republic of China
| | - Yanan Wu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, People's Republic of China
| | - Mengmeng Tong
- Ningbo Blue Illumination Tech Co., Ltd, Ningbo, People's Republic of China
| | - Yudong Yao
- Department of Electrical and Computer Engineering, Stevens Institute of Technology, Hoboken, NJ, United States of America
| | - Wei Qian
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, People's Republic of China
| | - Shouliang Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, People's Republic of China.,Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, People's Republic of China
| |
Collapse
|
23
|
Lin J, Pan Y, Xu J, Bao Y, Zhuo H. A meta-fusion RCNN network for endoscopic visual bladder lesions intelligent detection. Comput Med Imaging Graph 2022; 102:102138. [PMID: 36444783 DOI: 10.1016/j.compmedimag.2022.102138] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2021] [Revised: 09/26/2022] [Accepted: 10/24/2022] [Indexed: 11/11/2022]
Abstract
This study investigates a visual object detection technology in order to help doctors diagnose bladder lesions with endoscopy. A new object detection approach based on deep learning is presented, which derived from the cascade R-CNN and extended the ability of network for adapting insufficient endoscopic lesions samples when training a deep neural network. We propose a feature adaptive fusion model to increase the network's mobility and reduce the possibility of overfitting problems, and use task adaptation meta-learning approach to train the feature fusion process of the entire model and the target network update process in order to complete the task-adaptive classification and detection. The new model has been evaluated on the challenging object detection data set Pascal VOC and its converted format of Microsoft COCO, and the results show that the performance of our proposed method is superior to the original method. Therefore, we apply the proposed method to a custom bladder lesions data set to solve the auxiliary detection problem in the intelligent diagnosis of bladder lesions and demonstrated the effectiveness.
Collapse
Affiliation(s)
- Jie Lin
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 611731, China.
| | - Yulong Pan
- Department of Urology, The Third People's Hospital of Chengdu/The Affiliated Hospital of Southwest Jiaotong University, Chengdu, Sichuan, 610014, China.
| | - Jiajun Xu
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 611731, China.
| | - Yige Bao
- West China Hospital, Sichuan University, China.
| | - Hui Zhuo
- Department of Urology, The Third People's Hospital of Chengdu/The Affiliated Hospital of Southwest Jiaotong University, Chengdu, Sichuan, 610014, China.
| |
Collapse
|
24
|
Liu F, Demosthenes P. Real-world data: a brief review of the methods, applications, challenges and opportunities. BMC Med Res Methodol 2022; 22:287. [PMID: 36335315 PMCID: PMC9636688 DOI: 10.1186/s12874-022-01768-6] [Citation(s) in RCA: 141] [Impact Index Per Article: 47.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2022] [Accepted: 10/22/2022] [Indexed: 11/07/2022] Open
Abstract
Abstract
Background
The increased adoption of the internet, social media, wearable devices, e-health services, and other technology-driven services in medicine and healthcare has led to the rapid generation of various types of digital data, providing a valuable data source beyond the confines of traditional clinical trials, epidemiological studies, and lab-based experiments.
Methods
We provide a brief overview on the type and sources of real-world data and the common models and approaches to utilize and analyze real-world data. We discuss the challenges and opportunities of using real-world data for evidence-based decision making This review does not aim to be comprehensive or cover all aspects of the intriguing topic on RWD (from both the research and practical perspectives) but serves as a primer and provides useful sources for readers who interested in this topic.
Results and Conclusions
Real-world hold great potential for generating real-world evidence for designing and conducting confirmatory trials and answering questions that may not be addressed otherwise. The voluminosity and complexity of real-world data also call for development of more appropriate, sophisticated, and innovative data processing and analysis techniques while maintaining scientific rigor in research findings, and attentions to data ethics to harness the power of real-world data.
Collapse
|
25
|
Yang Y, Shang F, Wu B, Yang D, Wang L, Xu Y, Zhang W, Zhang T. Robust Collaborative Learning of Patch-Level and Image-Level Annotations for Diabetic Retinopathy Grading From Fundus Image. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:11407-11417. [PMID: 33961571 DOI: 10.1109/tcyb.2021.3062638] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Diabetic retinopathy (DR) grading from fundus images has attracted increasing interest in both academic and industrial communities. Most convolutional neural network-based algorithms treat DR grading as a classification task via image-level annotations. However, these algorithms have not fully explored the valuable information in the DR-related lesions. In this article, we present a robust framework, which collaboratively utilizes patch-level and image-level annotations, for DR severity grading. By an end-to-end optimization, this framework can bidirectionally exchange the fine-grained lesion and image-level grade information. As a result, it exploits more discriminative features for DR grading. The proposed framework shows better performance than the recent state-of-the-art algorithms and three clinical ophthalmologists with over nine years of experience. By testing on datasets of different distributions (such as label and camera), we prove that our algorithm is robust when facing image quality and distribution variations that commonly exist in real-world practice. We inspect the proposed framework through extensive ablation studies to indicate the effectiveness and necessity of each motivation. The code and some valuable annotations are now publicly available.
Collapse
|
26
|
Deng Y, Chen Y, Xie L, Wang L, Zhan J. The investigation of construction and clinical application of image recognition technology assisted bronchoscopy diagnostic model of lung cancer. Front Oncol 2022; 12:1001840. [PMID: 36387178 PMCID: PMC9647035 DOI: 10.3389/fonc.2022.1001840] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2022] [Accepted: 10/07/2022] [Indexed: 12/02/2022] Open
Abstract
Background The incidence and mortality of lung cancer ranks first in China. Bronchoscopy is one of the most common diagnostic methods for lung cancer. In recent years, image recognition technology(IRT) has been more and more widely studied and applied in the medical field. We developed a diagnostic model of lung cancer under bronchoscopy based on deep learning method and tried to classify pathological types. Methods A total of 2238 lesion images were collected retrospectively from 666 cases of lung cancer diagnosed by pathology in the bronchoscopy center of the Third Xiangya Hospital from Oct.01 2017 to Dec.31 2020 and 152 benign cases from Jun.01 2015 to Dec.31 2020. The benign and malignant images were divided into training, verification and test set according to 7:1:2 respectively. The model was trained and tested based on deep learning method. We also tried to classify different pathological types of lung cancer using the model. Furthermore, 9 clinicians with different experience were invited to diagnose the same test images and the results were compared with the model. Results The diagnostic model took a total of 30s to diagnose 467 test images. The overall accuracy, sensitivity, specificity and area under curve (AUC) of the model to differentiate benign and malignant lesions were 0.951, 0.978, 0.833 and 0.940, which were equivalent to the judgment results of 2 doctors in the senior group and higher than those of other doctors. In the classification of squamous cell carcinoma (SCC) and adenocarcinoma (AC), the overall accuracy was 0.745, including 0.790 for SCC, 0.667 for AC and AUC was 0.728. Conclusion The performance of our diagnostic model to distinguish benign and malignant lesions in bronchoscopy is roughly the same as that of experienced clinicians and the efficiency is much higher than manually. Our study verifies the possibility of applying IRT in diagnosis of lung cancer during white light bronchoscopy.
Collapse
Affiliation(s)
- Yihong Deng
- Department of Pulmonary and Critical Care Medicine, the Third Xiangya Hospital, Central South University, Changsha, Hunan, China
| | - Yuan Chen
- Department of Computer Science, School of Informatics, Xiamen University, Xiamen, Fujian, China
| | - Lihua Xie
- Department of Pulmonary and Critical Care Medicine, the Third Xiangya Hospital, Central South University, Changsha, Hunan, China
- *Correspondence: Lihua Xie, ; Liansheng Wang, ; Juan Zhan,
| | - Liansheng Wang
- Department of Computer Science, School of Informatics, Xiamen University, Xiamen, Fujian, China
- *Correspondence: Lihua Xie, ; Liansheng Wang, ; Juan Zhan,
| | - Juan Zhan
- Department of Oncology, Zhongshan Hospital affiliated to Xiamen University, Xiamen, Fujian, China
- *Correspondence: Lihua Xie, ; Liansheng Wang, ; Juan Zhan,
| |
Collapse
|
27
|
Fang H, Li F, Fu H, Sun X, Cao X, Lin F, Son J, Kim S, Quellec G, Matta S, Shankaranarayana SM, Chen YT, Wang CH, Shah NA, Lee CY, Hsu CC, Xie H, Lei B, Baid U, Innani S, Dang K, Shi W, Kamble R, Singhal N, Wang CW, Lo SC, Orlando JI, Bogunovic H, Zhang X, Xu Y. ADAM Challenge: Detecting Age-Related Macular Degeneration From Fundus Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:2828-2847. [PMID: 35507621 DOI: 10.1109/tmi.2022.3172773] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Age-related macular degeneration (AMD) is the leading cause of visual impairment among elderly in the world. Early detection of AMD is of great importance, as the vision loss caused by this disease is irreversible and permanent. Color fundus photography is the most cost-effective imaging modality to screen for retinal disorders. Cutting edge deep learning based algorithms have been recently developed for automatically detecting AMD from fundus images. However, there are still lack of a comprehensive annotated dataset and standard evaluation benchmarks. To deal with this issue, we set up the Automatic Detection challenge on Age-related Macular degeneration (ADAM), which was held as a satellite event of the ISBI 2020 conference. The ADAM challenge consisted of four tasks which cover the main aspects of detecting and characterizing AMD from fundus images, including detection of AMD, detection and segmentation of optic disc, localization of fovea, and detection and segmentation of lesions. As part of the ADAM challenge, we have released a comprehensive dataset of 1200 fundus images with AMD diagnostic labels, pixel-wise segmentation masks for both optic disc and AMD-related lesions (drusen, exudates, hemorrhages and scars, among others), as well as the coordinates corresponding to the location of the macular fovea. A uniform evaluation framework has been built to make a fair comparison of different models using this dataset. During the ADAM challenge, 610 results were submitted for online evaluation, with 11 teams finally participating in the onsite challenge. This paper introduces the challenge, the dataset and the evaluation methods, as well as summarizes the participating methods and analyzes their results for each task. In particular, we observed that the ensembling strategy and the incorporation of clinical domain knowledge were the key to improve the performance of the deep learning models.
Collapse
|
28
|
Kundu S, Karale V, Ghorai G, Sarkar G, Ghosh S, Dhara AK. Nested U-Net for Segmentation of Red Lesions in Retinal Fundus Images and Sub-image Classification for Removal of False Positives. J Digit Imaging 2022; 35:1111-1119. [PMID: 35474556 PMCID: PMC9582103 DOI: 10.1007/s10278-022-00629-4] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2020] [Revised: 03/21/2022] [Accepted: 04/03/2022] [Indexed: 11/29/2022] Open
Abstract
Diabetic retinopathy is a pathological change of the retina that occurs for long-term diabetes. The patients become symptomatic in advanced stages of diabetic retinopathy resulting in severe non-proliferative diabetic retinopathy or proliferative diabetic retinopathy stages. There is a need of an automated screening tool for the early detection and treatment of patients with diabetic retinopathy. This paper focuses on the segmentation of red lesions using nested U-Net Zhou et al. (Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Springer, 2018) followed by removal of false positives based on the sub-image classification method. Different sizes of sub-images were studied for the reduction in false positives in the sub-image classification method. The network could capture semantic features and fine details due to dense convolutional blocks connected via skip connections in between down sampling and up sampling paths. False-negative candidates were very few and the sub-image classification network effectively reduced the falsely detected candidates. The proposed framework achieves a sensitivity of [Formula: see text], precision of [Formula: see text], and F1-Score of [Formula: see text] for the DIARETDB1 data set Kalviainen and Uusutalo (Medical Image Understanding and Analysis, Citeseer, 2007). It outperforms the state-of-the-art networks such as U-Net Ronneberger et al. (International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2015) and attention U-Net Oktay et al. (Attention u-net: Learning where to look for the pancreas, 2018).
Collapse
Affiliation(s)
- Swagata Kundu
- Electrical Engineering Department, National Institute of Technology Durgapur, Durgapur, 713209 India
| | - Vikrant Karale
- Department of Electronics and Electrical Communication Engineering, Indian Institute of Technology Kharagpur, Kharagpur, 721302 India
| | - Goutam Ghorai
- Department of Electrical Engineering, Jadavpur University, Kolkata, 700032 India
| | - Gautam Sarkar
- Department of Electrical Engineering, Jadavpur University, Kolkata, 700032 India
| | - Sambuddha Ghosh
- Department of Ophthalmology, Calcutta National Medical College and Hospital, Kolkata, 700014 India
| | - Ashis Kumar Dhara
- Electrical Engineering Department, National Institute of Technology Durgapur, Durgapur, 713209 India
| |
Collapse
|
29
|
Dubey S, Dixit M. Recent developments on computer aided systems for diagnosis of diabetic retinopathy: a review. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:14471-14525. [PMID: 36185322 PMCID: PMC9510498 DOI: 10.1007/s11042-022-13841-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Revised: 04/27/2022] [Accepted: 09/06/2022] [Indexed: 06/16/2023]
Abstract
Diabetes is a long-term condition in which the pancreas quits producing insulin or the body's insulin isn't utilised properly. One of the signs of diabetes is Diabetic Retinopathy. Diabetic retinopathy is the most prevalent type of diabetes, if remains unaddressed, diabetic retinopathy can affect all diabetics and become very serious, raising the chances of blindness. It is a chronic systemic condition that affects up to 80% of patients for more than ten years. Many researchers believe that if diabetes individuals are diagnosed early enough, they can be rescued from the condition in 90% of cases. Diabetes damages the capillaries, which are microscopic blood vessels in the retina. On images, blood vessel damage is usually noticeable. Therefore, in this study, several traditional, as well as deep learning-based approaches, are reviewed for the classification and detection of this particular diabetic-based eye disease known as diabetic retinopathy, and also the advantage of one approach over the other is also described. Along with the approaches, the dataset and the evaluation metrics useful for DR detection and classification are also discussed. The main finding of this study is to aware researchers about the different challenges occurs while detecting diabetic retinopathy using computer vision, deep learning techniques. Therefore, a purpose of this review paper is to sum up all the major aspects while detecting DR like lesion identification, classification and segmentation, security attacks on the deep learning models, proper categorization of datasets and evaluation metrics. As deep learning models are quite expensive and more prone to security attacks thus, in future it is advisable to develop a refined, reliable and robust model which overcomes all these aspects which are commonly found while designing deep learning models.
Collapse
Affiliation(s)
- Shradha Dubey
- Madhav Institute of Technology & Science (Department of Computer Science and Engineering), Gwalior, M.P. India
| | - Manish Dixit
- Madhav Institute of Technology & Science (Department of Computer Science and Engineering), Gwalior, M.P. India
| |
Collapse
|
30
|
Spectral features and optimal Hierarchical attention networks for pulmonary abnormality detection from the respiratory sound signals. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103905] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
31
|
Zhang X, Peng Z, Meng M, Wu J, Han Y, Zhang Y, Yang J, Zhao Q. ID-NET: Inception deconvolutional neural network for multi-class classification in retinal fundus image. J MECH MED BIOL 2022. [DOI: 10.1142/s0219519422400292] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
32
|
Li H, Bhatt M, Qu Z, Zhang S, Hartel MC, Khademhosseini A, Cloutier G. Deep learning in ultrasound elastography imaging: A review. Med Phys 2022; 49:5993-6018. [PMID: 35842833 DOI: 10.1002/mp.15856] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2021] [Revised: 02/04/2022] [Accepted: 07/06/2022] [Indexed: 11/11/2022] Open
Abstract
It is known that changes in the mechanical properties of tissues are associated with the onset and progression of certain diseases. Ultrasound elastography is a technique to characterize tissue stiffness using ultrasound imaging either by measuring tissue strain using quasi-static elastography or natural organ pulsation elastography, or by tracing a propagated shear wave induced by a source or a natural vibration using dynamic elastography. In recent years, deep learning has begun to emerge in ultrasound elastography research. In this review, several common deep learning frameworks in the computer vision community, such as multilayer perceptron, convolutional neural network, and recurrent neural network are described. Then, recent advances in ultrasound elastography using such deep learning techniques are revisited in terms of algorithm development and clinical diagnosis. Finally, the current challenges and future developments of deep learning in ultrasound elastography are prospected. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Hongliang Li
- Laboratory of Biorheology and Medical Ultrasonics, University of Montreal Hospital Research Center, Montréal, Québec, Canada.,Institute of Biomedical Engineering, University of Montreal, Montréal, Québec, Canada
| | - Manish Bhatt
- Laboratory of Biorheology and Medical Ultrasonics, University of Montreal Hospital Research Center, Montréal, Québec, Canada
| | - Zhen Qu
- Laboratory of Biorheology and Medical Ultrasonics, University of Montreal Hospital Research Center, Montréal, Québec, Canada
| | - Shiming Zhang
- California Nanosystems Institute, University of California, Los Angeles, California, USA
| | - Martin C Hartel
- California Nanosystems Institute, University of California, Los Angeles, California, USA
| | - Ali Khademhosseini
- California Nanosystems Institute, University of California, Los Angeles, California, USA
| | - Guy Cloutier
- Laboratory of Biorheology and Medical Ultrasonics, University of Montreal Hospital Research Center, Montréal, Québec, Canada.,Institute of Biomedical Engineering, University of Montreal, Montréal, Québec, Canada.,Department of Radiology, Radio-Oncology and Nuclear Medicine, University of Montreal, Montréal, Québec, Canada
| |
Collapse
|
33
|
Gudhe NR, Behravan H, Sudah M, Okuma H, Vanninen R, Kosma VM, Mannermaa A. Area-based breast percentage density estimation in mammograms using weight-adaptive multitask learning. Sci Rep 2022; 12:12060. [PMID: 35835933 PMCID: PMC9283472 DOI: 10.1038/s41598-022-16141-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2022] [Accepted: 07/05/2022] [Indexed: 12/02/2022] Open
Abstract
Breast density, which is a measure of the relative amount of fibroglandular tissue within the breast area, is one of the most important breast cancer risk factors. Accurate segmentation of fibroglandular tissues and breast area is crucial for computing the breast density. Semiautomatic and fully automatic computer-aided design tools have been developed to estimate the percentage of breast density in mammograms. However, the available approaches are usually limited to specific mammogram views and are inadequate for complete delineation of the pectoral muscle. These tools also perform poorly in cases of data variability and often require an experienced radiologist to adjust the segmentation threshold for fibroglandular tissue within the breast area. This study proposes a new deep learning architecture that automatically estimates the area-based breast percentage density from mammograms using a weight-adaptive multitask learning approach. The proposed approach simultaneously segments the breast and dense tissues and further estimates the breast percentage density. We evaluate the performance of the proposed model in both segmentation and density estimation on an independent evaluation set of 7500 craniocaudal and mediolateral oblique-view mammograms from Kuopio University Hospital, Finland. The proposed multitask segmentation approach outperforms and achieves average relative improvements of 2.88% and 9.78% in terms of F-score compared to the multitask U-net and a fully convolutional neural network, respectively. The estimated breast density values using our approach strongly correlate with radiologists' assessments with a Pearson's correlation of [Formula: see text] (95% confidence interval [0.89, 0.91]). We conclude that our approach greatly improves the segmentation accuracy of the breast area and dense tissues; thus, it can play a vital role in accurately computing the breast density. Our density estimation model considerably reduces the time and effort needed to estimate density values from mammograms by radiologists and therefore, decreases inter- and intra-reader variability.
Collapse
Affiliation(s)
- Naga Raju Gudhe
- Institute of Clinical Medicine, Pathology and Forensic Medicine, Multidisciplinary Cancer Research community, University of Eastern Finland, P.O. Box 1627, 70211, Kuopio, Finland.
| | - Hamid Behravan
- Institute of Clinical Medicine, Pathology and Forensic Medicine, Multidisciplinary Cancer Research community, University of Eastern Finland, P.O. Box 1627, 70211, Kuopio, Finland.
| | - Mazen Sudah
- Department of Clinical Radiology, Kuopio University Hospital, P.O. Box 100, 70029, Kuopio, Finland
| | - Hidemi Okuma
- Department of Clinical Radiology, Kuopio University Hospital, P.O. Box 100, 70029, Kuopio, Finland
| | - Ritva Vanninen
- Department of Clinical Radiology, Kuopio University Hospital, P.O. Box 100, 70029, Kuopio, Finland
- Institute of Clinical Medicine, Radiology, Translational Cancer Research Area, University of Eastern Finland, P.O. Box 1627, 70211, Kuopio, Finland
| | - Veli-Matti Kosma
- Institute of Clinical Medicine, Pathology and Forensic Medicine, Multidisciplinary Cancer Research community, University of Eastern Finland, P.O. Box 1627, 70211, Kuopio, Finland
- Biobank of Eastern Finland, Kuopio University Hospital, Kuopio, Finland
| | - Arto Mannermaa
- Institute of Clinical Medicine, Pathology and Forensic Medicine, Multidisciplinary Cancer Research community, University of Eastern Finland, P.O. Box 1627, 70211, Kuopio, Finland
- Biobank of Eastern Finland, Kuopio University Hospital, Kuopio, Finland
| |
Collapse
|
34
|
Wong NCK, Meshkinfamfard S, Turbé V, Whitaker M, Moshe M, Bardanzellu A, Dai T, Pignatelli E, Barclay W, Darzi A, Elliott P, Ward H, Tanaka RJ, Cooke GS, McKendry RA, Atchison CJ, Bharath AA. Machine learning to support visual auditing of home-based lateral flow immunoassay self-test results for SARS-CoV-2 antibodies. COMMUNICATIONS MEDICINE 2022; 2:78. [PMID: 35814295 PMCID: PMC9259560 DOI: 10.1038/s43856-022-00146-z] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2022] [Accepted: 06/15/2022] [Indexed: 12/24/2022] Open
Abstract
Background Lateral flow immunoassays (LFIAs) are being used worldwide for COVID-19 mass testing and antibody prevalence studies. Relatively simple to use and low cost, these tests can be self-administered at home, but rely on subjective interpretation of a test line by eye, risking false positives and false negatives. Here, we report on the development of ALFA (Automated Lateral Flow Analysis) to improve reported sensitivity and specificity. Methods Our computational pipeline uses machine learning, computer vision techniques and signal processing algorithms to analyse images of the Fortress LFIA SARS-CoV-2 antibody self-test, and subsequently classify results as invalid, IgG negative and IgG positive. A large image library of 595,339 participant-submitted test photographs was created as part of the REACT-2 community SARS-CoV-2 antibody prevalence study in England, UK. Alongside ALFA, we developed an analysis toolkit which could also detect device blood leakage issues. Results Automated analysis showed substantial agreement with human experts (Cohen's kappa 0.90-0.97) and performed consistently better than study participants, particularly for weak positive IgG results. Specificity (98.7-99.4%) and sensitivity (90.1-97.1%) were high compared with visual interpretation by human experts (ranges due to the varying prevalence of weak positive IgG tests in datasets). Conclusions Given the potential for LFIAs to be used at scale in the COVID-19 response (for both antibody and antigen testing), even a small improvement in the accuracy of the algorithms could impact the lives of millions of people by reducing the risk of false-positive and false-negative result read-outs by members of the public. Our findings support the use of machine learning-enabled automated reading of at-home antibody lateral flow tests as a tool for improved accuracy for population-level community surveillance.
Collapse
Affiliation(s)
| | | | - Valérian Turbé
- London Centre for Nanotechnology, University College London, London, UK
| | | | - Maya Moshe
- Department of Infectious Disease, Imperial College London, London, UK
| | | | - Tianhong Dai
- Department of Bioengineering, Imperial College London, London, UK
| | | | - Wendy Barclay
- Department of Infectious Disease, Imperial College London, London, UK
- Imperial College Healthcare NHS Trust, London, UK
- National Institute for Health Research Imperial Biomedical Research Centre, London, UK
| | - Ara Darzi
- Imperial College Healthcare NHS Trust, London, UK
- National Institute for Health Research Imperial Biomedical Research Centre, London, UK
- Institute of Global Health Innovation, Imperial College London, London, UK
| | - Paul Elliott
- School of Public Health, Imperial College London, London, UK
- Imperial College Healthcare NHS Trust, London, UK
- National Institute for Health Research Imperial Biomedical Research Centre, London, UK
- MRC Centre for Environment and Health, School of Public Health, Imperial College London, London, UK
| | - Helen Ward
- School of Public Health, Imperial College London, London, UK
- Imperial College Healthcare NHS Trust, London, UK
- National Institute for Health Research Imperial Biomedical Research Centre, London, UK
| | - Reiko J. Tanaka
- Department of Bioengineering, Imperial College London, London, UK
| | - Graham S. Cooke
- Department of Infectious Disease, Imperial College London, London, UK
- National Institute for Health Research Imperial Biomedical Research Centre, London, UK
| | - Rachel A. McKendry
- London Centre for Nanotechnology, University College London, London, UK
- Division of Medicine, University College London, London, UK
| | - Christina J. Atchison
- School of Public Health, Imperial College London, London, UK
- Imperial College Healthcare NHS Trust, London, UK
| | - Anil A. Bharath
- Department of Bioengineering, Imperial College London, London, UK
| |
Collapse
|
35
|
Xia H, Rao Z, Zhou Z. A multi-scale gated network for retinal hemorrhage detection. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03476-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
36
|
Investigations of CNN for Medical Image Analysis for Illness Prediction. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:7968200. [PMID: 35676956 PMCID: PMC9168160 DOI: 10.1155/2022/7968200] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Revised: 01/31/2022] [Accepted: 02/11/2022] [Indexed: 11/17/2022]
Abstract
When it comes to diabetic retinopathy, exudates are the most common sign; alarms for early screening and diagnosis are suggested. The images taken by cameras and high-definition ophthalmoscopes are riddled with flaws and noise. Overcoming noise difficulties and pursuing automated/computer-aided diagnosis is always a challenge. The major objective of this approach is to obtain a better prediction rate of diabetic retinopathy analysis. The accuracy, sensitivity, specificity, and prediction rate improvement are focused on the objective view. The images are separated into relevant patches of various sizes and stacked for use as inputs to CNN, which is then trained, tested, and validated. The article presents a mathematical approach to determine the prevalence, shape in precise, color, and density in the populations among image patches to operate and discover the fact the image collection consists of symptoms of exudates and methods to comprehend the diagnosis and suggest risks of early hospital treatment. The experimental result analysis of malignant quality shows the accuracy, sensitivity, specificity, and predictive value. Here, 78% of accuracy, 78.8% of sensitivity, and 78.3% of specificity are obtained, and both positive and negative predictive values are obtained.
Collapse
|
37
|
Cao P, Hou Q, Song R, Wang H, Zaiane O. Collaborative learning of weakly-supervised domain adaptation for diabetic retinopathy grading on retinal images. Comput Biol Med 2022; 144:105341. [PMID: 35279423 DOI: 10.1016/j.compbiomed.2022.105341] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2021] [Revised: 02/20/2022] [Accepted: 02/20/2022] [Indexed: 11/25/2022]
Abstract
Early detection and treatment of diabetic retinopathy (DR) can significantly reduce the risk of vision loss in patients. In essence, we are faced with two challenges: (i) how to simultaneously achieve domain adaptation from the different domains and (ii) how to build an interpretable multi-instance learning (MIL) on the target domain in an end-to-end framework. In this paper, we address these issues and propose a unified weakly-supervised domain adaptation framework, which consists of three components: domain adaptation, instance progressive discriminator and multi-instance learning with attention. The method models the relationship between the patches and images in the target domain with a multi-instance learning scheme and an attention mechanism. Meanwhile, it incorporates all available information from both source and target domains for a jointly learning strategy. We validate the performance of the proposed framework for DR grading on the Messidor dataset and the large-scale Eyepacs dataset. The experimental results demonstrate that it achieves an average accuracy of 0.949 (95% CI 0.931-0.958)/0.764 (95% CI 0.755-0.772) and an average AUC value of 0.958 (95% CI 0.945-0.962)/0.749 (95% CI 0.732-0.761) for binary-class/multi-class classification tasks on the Messidor dataset. Moreover, the proposed method achieves an accuracy of 0.887 and a quadratic weighted kappa score value of 0.860 on the Eyepacs dataset, outperforming the state-of-the-art approaches. Comprehensive experiments confirm the effectiveness of the approach in terms of both grading performance and interpretability. The source code is available at https://github.com/HouQingshan/WAD-Net.
Collapse
Affiliation(s)
- Peng Cao
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Shenyang, China.
| | - Qingshan Hou
- Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Ruoxian Song
- Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Haonan Wang
- Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Osmar Zaiane
- Alberta Machine Intelligence Institute, University of Alberta, Edmonton, Alberta, Canada
| |
Collapse
|
38
|
Chandrasekaran R, Loganathan B. Retinopathy grading with deep learning and wavelet hyper-analytic activations. THE VISUAL COMPUTER 2022; 39:1-16. [PMID: 35493724 PMCID: PMC9035984 DOI: 10.1007/s00371-022-02489-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Accepted: 03/12/2022] [Indexed: 06/14/2023]
Abstract
Recent developments reveal the prominence of Diabetic Retinopathy (DR) grading. In the past few decades, Wavelet-based DR classification has shown successful impacts and the Deep Learning models, like Convolutional Neural Networks (CNN's), have evolved in offering the highest prediction accuracy. In this work, the features of the input image are enhanced with the integration of Multi-Resolution Analysis (MRA) and a CNN framework without costing more convolution filters. The bottleneck with conventional activation functions, used in CNN's, is the nullification of the feature maps that are negative in value. In this work, a novel Hyper-analytic Wavelet (HW) phase activation function is formulated with unique characteristics for the wavelet sub-bands. Instead of dismissal, the function transforms these negative coefficients that correspond to significant edge feature maps. The hyper-analytic wavelet phase forms the imaginary part of the complex activation. And the hyper-parameter of the activation function is selected such that the corresponding magnitude spectrum produces monotonic and effective activations. The performance of 3 CNN models (1 custom, shallow CNN, ResNet with Soft attention, Alex Net for DR) with spatial-Wavelet quilts is better. With the spatial-Wavelet quilts, the Alex Net for DR has an improvement with an 11% of accuracy level (from 87 to 98%). The highest accuracy level of 98% and the highest Sensitivity of 99% are attained through Modified Alex Net for DR. The proposal also illustrates the visualization of the negative edge preservation with assumed image patches. From this study, the researcher infers that models with spatial-Wavelet quilts, with the hyper-analytic activations, have better generalization ability. And the visualization of heat maps provides evidence of better learning of the feature maps from the wavelet sub-bands.
Collapse
Affiliation(s)
| | - Balaji Loganathan
- Department of ECE, Vel Tech Rangarajan Dr. Sagunthala R & D Institute of Science and Technology, Chennai, Tamil Nadu 600062 India
| |
Collapse
|
39
|
Das D, Biswas SK, Bandyopadhyay S. A critical review on diagnosis of diabetic retinopathy using machine learning and deep learning. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 81:25613-25655. [PMID: 35342328 PMCID: PMC8940593 DOI: 10.1007/s11042-022-12642-4] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/25/2020] [Revised: 06/29/2021] [Accepted: 02/09/2022] [Indexed: 06/12/2023]
Abstract
Diabetic Retinopathy (DR) is a health condition caused due to Diabetes Mellitus (DM). It causes vision problems and blindness due to disfigurement of human retina. According to statistics, 80% of diabetes patients battling from long diabetic period of 15 to 20 years, suffer from DR. Hence, it has become a dangerous threat to the health and life of people. To overcome DR, manual diagnosis of the disease is feasible but overwhelming and cumbersome at the same time and hence requires a revolutionary method. Thus, such a health condition necessitates primary recognition and diagnosis to prevent DR from developing into severe stages and prevent blindness. Innumerable Machine Learning (ML) models are proposed by researchers across the globe, to achieve this purpose. Various feature extraction techniques are proposed for extraction of DR features for early detection. However, traditional ML models have shown either meagre generalization throughout feature extraction and classification for deploying smaller datasets or consumes more of training time causing inefficiency in prediction while using larger datasets. Hence Deep Learning (DL), a new domain of ML, is introduced. DL models can handle a smaller dataset with help of efficient data processing techniques. However, they generally incorporate larger datasets for their deep architectures to enhance performance in feature extraction and image classification. This paper gives a detailed review on DR, its features, causes, ML models, state-of-the-art DL models, challenges, comparisons and future directions, for early detection of DR.
Collapse
Affiliation(s)
- Dolly Das
- National Institute of Technology Silchar, Cachar, Assam India
| | | | | |
Collapse
|
40
|
Joint DR-DME classification using deep learning-CNN based modified grey-wolf optimizer with variable weights. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2021.103439] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
41
|
Deep Red Lesion Classification for Early Screening of Diabetic Retinopathy. MATHEMATICS 2022. [DOI: 10.3390/math10050686] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Diabetic retinopathy (DR) is an asymptotic and vision-threatening complication among working-age adults. To prevent blindness, a deep convolutional neural network (CNN) based diagnosis can help to classify less-discriminative and small-sized red lesions in early screening of DR patients. However, training deep models with minimal data is a challenging task. Fine-tuning through transfer learning is a useful alternative, but performance degradation, overfitting, and domain adaptation issues further demand architectural amendments to effectively train deep models. Various pre-trained CNNs are fine-tuned on an augmented set of image patches. The best-performing ResNet50 model is modified by introducing reinforced skip connections, a global max-pooling layer, and the sum-of-squared-error loss function. The performance of the modified model (DR-ResNet50) on five public datasets is found to be better than state-of-the-art methods in terms of well-known metrics. The highest scores (0.9851, 0.991, 0.991, 0.991, 0.991, 0.9939, 0.0029, 0.9879, and 0.9879) for sensitivity, specificity, AUC, accuracy, precision, F1-score, false-positive rate, Matthews’s correlation coefficient, and kappa coefficient are obtained within a 95% confidence interval for unseen test instances from e-Ophtha_MA. This high sensitivity and low false-positive rate demonstrate the worth of a proposed framework. It is suitable for early screening due to its performance, simplicity, and robustness.
Collapse
|
42
|
Automatic Left Ventricle Segmentation from Short-Axis Cardiac MRI Images Based on Fully Convolutional Neural Network. Diagnostics (Basel) 2022; 12:diagnostics12020414. [PMID: 35204504 PMCID: PMC8871002 DOI: 10.3390/diagnostics12020414] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2021] [Revised: 01/06/2022] [Accepted: 01/16/2022] [Indexed: 11/26/2022] Open
Abstract
Background: Left ventricle (LV) segmentation using a cardiac magnetic resonance imaging (MRI) dataset is critical for evaluating global and regional cardiac functions and diagnosing cardiovascular diseases. LV clinical metrics such as LV volume, LV mass and ejection fraction (EF) are frequently extracted based on the LV segmentation from short-axis MRI images. Manual segmentation to assess such functions is tedious and time-consuming for medical experts to diagnose cardiac pathologies. Therefore, a fully automated LV segmentation technique is required to assist medical experts in working more efficiently. Method: This paper proposes a fully convolutional network (FCN) architecture for automatic LV segmentation from short-axis MRI images. Several experiments were conducted in the training phase to compare the performance of the network and the U-Net model with various hyper-parameters, including optimization algorithms, epochs, learning rate, and mini-batch size. In addition, a class weighting method was introduced to avoid having a high imbalance of pixels in the classes of image’s labels since the number of background pixels was significantly higher than the number of LV and myocardium pixels. Furthermore, effective image conversion with pixel normalization was applied to obtain exact features representing target organs (LV and myocardium). The segmentation models were trained and tested on a public dataset, namely the evaluation of myocardial infarction from the delayed-enhancement cardiac MRI (EMIDEC) dataset. Results: The dice metric, Jaccard index, sensitivity, and specificity were used to evaluate the network’s performance, with values of 0.93, 0.87, 0.98, and 0.94, respectively. Based on the experimental results, the proposed network outperforms the standard U-Net model and is an advanced fully automated method in terms of segmentation performance. Conclusion: This proposed method is applicable in clinical practice for doctors to diagnose cardiac diseases from short-axis MRI images.
Collapse
|
43
|
Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination. Biomedicines 2022; 10:biomedicines10010124. [PMID: 35052803 PMCID: PMC8773350 DOI: 10.3390/biomedicines10010124] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2021] [Revised: 12/31/2021] [Accepted: 01/03/2022] [Indexed: 01/02/2023] Open
Abstract
Retinal microaneurysm (MA) is the initial symptom of diabetic retinopathy (DR). The automatic detection of MA is helpful to assist doctors in diagnosis and treatment. Previous algorithms focused on the features of the target itself; however, the local structural features of the target and background are also worth exploring. To achieve MA detection, an efficient local structure awareness-based retinal MA detection with the multi-feature combination (LSAMFC) is proposed in this paper. We propose a novel local structure feature called a ring gradient descriptor (RGD) to describe the structural differences between an object and its surrounding area. Then, a combination of RGD with the salience and texture features is used by a Gradient Boosting Decision Tree (GBDT) for candidate classification. We evaluate our algorithm on two public datasets, i.e., the e-ophtha MA dataset and retinopathy online challenge (ROC) dataset. The experimental results show that the performance of the trained model significantly improved after combining traditional features with RGD, and the area under the receiver operating characteristic curve (AUC) values in the test results of the datasets e-ophtha MA and ROC increased from 0.9615 to 0.9751 and from 0.9066 to 0.9409, respectively.
Collapse
|
44
|
Review of Machine Learning Applications Using Retinal Fundus Images. Diagnostics (Basel) 2022; 12:diagnostics12010134. [PMID: 35054301 PMCID: PMC8774893 DOI: 10.3390/diagnostics12010134] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Revised: 01/03/2022] [Accepted: 01/03/2022] [Indexed: 02/04/2023] Open
Abstract
Automating screening and diagnosis in the medical field saves time and reduces the chances of misdiagnosis while saving on labor and cost for physicians. With the feasibility and development of deep learning methods, machines are now able to interpret complex features in medical data, which leads to rapid advancements in automation. Such efforts have been made in ophthalmology to analyze retinal images and build frameworks based on analysis for the identification of retinopathy and the assessment of its severity. This paper reviews recent state-of-the-art works utilizing the color fundus image taken from one of the imaging modalities used in ophthalmology. Specifically, the deep learning methods of automated screening and diagnosis for diabetic retinopathy (DR), age-related macular degeneration (AMD), and glaucoma are investigated. In addition, the machine learning techniques applied to the retinal vasculature extraction from the fundus image are covered. The challenges in developing these systems are also discussed.
Collapse
|
45
|
Sharma Y, Singh BK. One-dimensional convolutional neural network and hybrid deep-learning paradigm for classification of specific language impaired children using their speech. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 213:106487. [PMID: 34763173 DOI: 10.1016/j.cmpb.2021.106487] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/30/2021] [Accepted: 10/15/2021] [Indexed: 06/13/2023]
Abstract
BACKGROUND AND OBJECTIVE Screening children for communicational disorders such as specific language impairment (SLI) is always challenging as it requires clinicians to follow a series of steps to evaluate the subjects. Artificial intelligence and computer-aided diagnosis have supported health professionals in making swift and error-free decisions about the neurodevelopmental state of children vis-à-vis language comprehension and production. Past studies have claimed that typical developing (TD) and SLI children show distinct vocal characteristics that can serve as discriminating facets between them. The objective of this study is to group children in SLI or TD categories by processing their raw speech signals using two proposed approaches: a customized convolutional neural network (CNN) model and a hybrid deep-learning framework where CNN is combined with long-short-term-memory (LSTM). METHOD We considered a publicly available speech database of SLI and typical children of Czech accents for this study. The convolution filters in both the proposed CNN and hybrid models (CNN-LSTM) estimated fuzzy-automated features from the speech utterance. We performed the experiments in five separate sessions. Data augmentations were performed in each of those sessions to enhance the training strength. RESULTS Our hybrid model exhibited a perfect 100% accuracy and F-measure for almost all the session-trials compared to CNN alone which achieved an average accuracy close to 90% and F-measure ≥ 92%. The models have further illustrated their robust classification essences by securing values of reliability indexes over 90%. CONCLUSION The results confirm the effectiveness of proposed approaches for the detection of SLI in children using their raw speech signals. Both the models do not require any dedicated feature extraction unit for their operations. The models may also be suitable for screening SLI and other neurodevelopmental disorders in children of different linguistic accents.
Collapse
Affiliation(s)
- Yogesh Sharma
- Department of Biomedical Engineering, National Institute of Technology Raipur, Chhattisgarh, 492010, India.
| | - Bikesh Kumar Singh
- Department of Biomedical Engineering, National Institute of Technology Raipur, Chhattisgarh, 492010, India.
| |
Collapse
|
46
|
Luca AR, Ursuleanu TF, Gheorghe L, Grigorovici R, Iancu S, Hlusneac M, Grigorovici A. Impact of quality, type and volume of data used by deep learning models in the analysis of medical images. INFORMATICS IN MEDICINE UNLOCKED 2022. [DOI: 10.1016/j.imu.2022.100911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022] Open
|
47
|
Shekar S, Satpute N, Gupta A. Review on diabetic retinopathy with deep learning methods. JOURNAL OF MEDICAL IMAGING (BELLINGHAM, WASH.) 2021; 8:060901. [PMID: 34859116 DOI: 10.1117/1.jmi.8.6.060901] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Received: 05/13/2021] [Accepted: 10/27/2021] [Indexed: 11/14/2022]
Abstract
Purpose: The purpose of our review paper is to examine many existing works of literature presenting the different methods utilized for diabetic retinopathy (DR) recognition employing deep learning (DL) and machine learning (ML) techniques, and also to address the difficulties faced in various datasets used by DR. Approach: DR is a progressive illness and may become a reason for vision loss. Early identification of DR lesions is, therefore, helpful and prevents damage to the retina. However, it is a complex job in view of the fact that it is symptomless earlier, and also ophthalmologists have been needed in traditional approaches. Recently, automated identification of DR-based studies has been stated based on image processing, ML, and DL. We analyze the recent literature and provide a comparative study that also includes the limitations of the literature and future work directions. Results: A relative analysis among the databases used, performance metrics employed, and ML and DL techniques adopted recently in DR detection based on various DR features is presented. Conclusion: Our review paper discusses the methods employed in DR detection along with the technical and clinical challenges that are encountered, which is missing in existing reviews, as well as future scopes to assist researchers in the field of retinal imaging.
Collapse
Affiliation(s)
- Shreya Shekar
- College of Engineering Pune, Department of Electronics and Telecommunication Engineering, Pune, Maharashtra, India
| | - Nitin Satpute
- Aarhus University, Department of Electrical and Computer Engineering, Aarhus, Denmark
| | - Aditya Gupta
- College of Engineering Pune, Department of Electronics and Telecommunication Engineering, Pune, Maharashtra, India
| |
Collapse
|
48
|
Melissourgos D, Gao H, Ma C, Chen S, Wu SS. On Outsourcing Artificial Neural Network Learning of Privacy-Sensitive Medical Data to the Cloud. INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE : [PROCEEDINGS]. INTERNATIONAL CONFERENCE ON TOOLS FOR ARTIFICIAL INTELLIGENCE 2021; 2021:381-385. [PMID: 35095256 PMCID: PMC8796752 DOI: 10.1109/ictai52525.2021.00062] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
Machine learning and artificial neural networks (ANNs) have been at the forefront of medical research in the last few years. It is well known that ANNs benefit from big data and the collection of the data is often decentralized, meaning that it is stored in different computer systems. There is a practical need to bring the distributed data together with the purpose of training a more accurate ANN. However, the privacy concern prevents medical institutes from sharing patient data freely. Federated learning and multi-party computation have been proposed to address this concern. However, they require the medical data collectors to participate in the deep-learning computations of the data users, which is inconvenient or even infeasible in practice. In this paper, we propose to use matrix masking for privacy protection of patient data. It allows the data collectors to outsource privacy-sensitive medical data to the cloud in a masked form, and allows the data users to outsource deep learning to the cloud as well, where the ANN models can be trained directly from the masked data. Our experimental results on deep-learning models for diagnosis of Alzheimer's disease and Parkinson's disease show that the diagnosis accuracy of the models trained from the masked data is similar to that of the models from the original patient data.
Collapse
Affiliation(s)
| | - Hanzhi Gao
- University of Florida, Gainesville, FL, USA
| | - Chaoyi Ma
- University of Florida, Gainesville, FL, USA
| | | | | |
Collapse
|
49
|
Wu J, Hu R, Xiao Z, Chen J, Liu J. Vision Transformer-based recognition of diabetic retinopathy grade. Med Phys 2021; 48:7850-7863. [PMID: 34693536 DOI: 10.1002/mp.15312] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2021] [Revised: 10/14/2021] [Accepted: 10/15/2021] [Indexed: 11/06/2022] Open
Abstract
BACKGROUND In the domain of natural language processing, Transformers are recognized as state-of-the-art models, which opposing to typical convolutional neural networks (CNNs) do not rely on convolution layers. Instead, Transformers employ multi-head attention mechanisms as the main building block to capture long-range contextual relations between image pixels. Recently, CNNs dominated the deep learning solutions for diabetic retinopathy grade recognition. However, spurred by the advantages of Transformers, we propose a Transformer-based method that is appropriate for recognizing the grade of diabetic retinopathy. PURPOSE The purposes of this work are to demonstrate that (i) the pure attention mechanism is suitable for diabetic retinopathy grade recognition and (ii) Transformers can replace traditional CNNs for diabetic retinopathy grade recognition. METHODS This paper proposes a Vision Transformer-based method to recognize the grade of diabetic retinopathy. Fundus images are subdivided into non-overlapping patches, which are then converted into sequences by flattening, and undergo a linear and positional embedding process to preserve positional information. Then, the generated sequence is input into several multi-head attention layers to generate the final representation. The first token sequence is input to a softmax classification layer to produce the recognition output in the classification stage. RESULTS The dataset for training and testing employs fundus images of different resolutions, subdivided into patches. We challenge our method against current CNNs and extreme learning machines and achieve an appealing performance. Specifically, the suggested deep learning architecture attains an accuracy of 91.4%, specificity = 0.977 (95% confidence interval (CI) (0.951-1)), precision = 0.928 (95% CI (0.852-1)), sensitivity = 0.926 (95% CI (0.863-0.989)), quadratic weighted kappa score = 0.935, and area under curve (AUC) = 0.986. CONCLUSION Our comparative experiments against current methods conclude that our model is competitive and highlight that an attention mechanism based on a Vision Transformer model is promising for the diabetic retinopathy grade recognition task.
Collapse
Affiliation(s)
- Jianfang Wu
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Ruo Hu
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Zhenghong Xiao
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Jiaxu Chen
- School of Traditional Chinese Medicine, Jinan University, Guangzhou, China
| | | |
Collapse
|
50
|
Ajitha S, Akkara JD, Judy MV. Identification of glaucoma from fundus images using deep learning techniques. Indian J Ophthalmol 2021; 69:2702-2709. [PMID: 34571619 PMCID: PMC8597466 DOI: 10.4103/ijo.ijo_92_21] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023] Open
Abstract
Purpose Glaucoma is one of the preeminent causes of incurable visual disability and blindness across the world due to elevated intraocular pressure within the eyes. Accurate and timely diagnosis is essential for preventing visual disability. Manual detection of glaucoma is a challenging task that needs expertise and years of experience. Methods In this paper, we suggest a powerful and accurate algorithm using a convolutional neural network (CNN) for the automatic diagnosis of glaucoma. In this work, 1113 fundus images consisting of 660 normal and 453 glaucomatous images from four databases have been used for the diagnosis of glaucoma. A 13-layer CNN is potently trained from this dataset to mine vital features, and these features are classified into either glaucomatous or normal class during testing. The proposed algorithm is implemented in Google Colab, which made the task straightforward without spending hours installing the environment and supporting libraries. To evaluate the effectiveness of our algorithm, the dataset is divided into 70% for training, 20% for validation, and the remaining 10% utilized for testing. The training images are augmented to 12012 fundus images. Results Our model with SoftMax classifier achieved an accuracy of 93.86%, sensitivity of 85.42%, specificity of 100%, and precision of 100%. In contrast, the model with the SVM classifier achieved accuracy, sensitivity, specificity, and precision of 95.61, 89.58, 100, and 100%, respectively. Conclusion These results demonstrate the ability of the deep learning model to identify glaucoma from fundus images and suggest that the proposed system can help ophthalmologists in a fast, accurate, and reliable diagnosis of glaucoma.
Collapse
Affiliation(s)
- S Ajitha
- Department of Computer Applications, Cochin University of Science and Technology, Kochi, Kerala, India
| | - John D Akkara
- Ophthalmology Department, Sri Ramachandra Institute of Higher Education and Research, Chennai, Tamil Nadu, India
| | - M V Judy
- Department of Computer Applications, Cochin University of Science and Technology, Kochi, Kerala, India
| |
Collapse
|