1
|
Chu B, Zhao J, Zheng W, Xu Z. (DA-U) 2Net: double attention U 2Net for retinal vessel segmentation. BMC Ophthalmol 2025; 25:86. [PMID: 39984892 PMCID: PMC11844045 DOI: 10.1186/s12886-025-03908-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Accepted: 02/10/2025] [Indexed: 02/23/2025] Open
Abstract
BACKGROUND Morphological changes in the retina are crucial and serve as valuable references in the clinical diagnosis of ophthalmic and cardiovascular diseases. However, the retinal vascular structure is complex, making manual segmentation time-consuming and labor-intensive. METHODS This paper proposes a retinal segmentation network that integrates feature channel attention and the Convolutional Block Attention Module (CBAM) attention within the U2Net model. First, a feature channel attention module is introduced into the RSU (Residual Spatial Unit) block of U2Net, forming an Attention-RSU block, which focuses more on significant areas during feature extraction and suppresses the influence of noise; Second, a Spatial Attention Module (SAM) is introduced into the high-resolution module of Attention-RSU to enrich feature extraction from both spatial and channel dimensions, and a Channel Attention Module (CAM) is integrated into the lowresolution module of Attention-RSU, which uses dual channel attention to reduce detail loss.Finally, dilated convolution is applied during the upscaling and downscaling processes to expand the receptive field in low-resolution states, allowing the model to better integrate contextual information. RESULTS The evaluation across multiple clinical datasets demonstrated excellent performance on various metrics, with an accuracy (ACC) of 98.71%. CONCLUSION The proposed Network is general enough and we believe it can be easily extended to other medical image segmentation tasks where large scale variation and complicated features are the main challenges.
Collapse
Affiliation(s)
- Bing Chu
- Department of Medical Engineering, Wannan Medical College, WuHu, AnHui, 241002, China
| | - Jinsong Zhao
- School of Medical Imageology, Wannan Medical College, WuHu, AnHui, 241002, China
| | - Wenqiang Zheng
- Department of Nuclear Medicine, First Affiliated Hospital of Wannan Medical College, Wuhu, AnHui, 241001, China
| | - Zhengyuan Xu
- Department of Medical Engineering, Wannan Medical College, WuHu, AnHui, 241002, China.
| |
Collapse
|
2
|
Kande GB, Nalluri MR, Manikandan R, Cho J, Veerappampalayam Easwaramoorthy S. Multi scale multi attention network for blood vessel segmentation in fundus images. Sci Rep 2025; 15:3438. [PMID: 39870673 PMCID: PMC11772654 DOI: 10.1038/s41598-024-84255-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Accepted: 12/20/2024] [Indexed: 01/29/2025] Open
Abstract
Precise segmentation of retinal vasculature is crucial for the early detection, diagnosis, and treatment of vision-threatening ailments. However, this task is challenging due to limited contextual information, variations in vessel thicknesses, the complexity of vessel structures, and the potential for confusion with lesions. In this paper, we introduce a novel approach, the MSMA Net model, which overcomes these challenges by replacing traditional convolution blocks and skip connections with an improved multi-scale squeeze and excitation block (MSSE Block) and Bottleneck residual paths (B-Res paths) with spatial attention blocks (SAB). Our experimental findings on publicly available datasets of fundus images, specifically DRIVE, STARE, CHASE_DB1, HRF and DR HAGIS consistently demonstrate that our approach outperforms other segmentation techniques, achieving higher accuracy, sensitivity, Dice score, and area under the receiver operator characteristic (AUC) in the segmentation of blood vessels with different thicknesses, even in situations involving diverse contextual information, the presence of coexisting lesions, and intricate vessel morphologies.
Collapse
Affiliation(s)
- Giri Babu Kande
- Vasireddy Venkatadri Institute of Technology, Nambur, 522508, India
| | - Madhusudana Rao Nalluri
- School of Computing, Amrita Vishwa Vidyapeetham, Amaravati, 522503, India.
- Department of Computer Science & Engineering, Faculty of Science and Technology (IcfaiTech), ICFAI Foundation for Higher Education, Hyderabad, India.
| | - R Manikandan
- School of Computing, SASTRA Deemed University, Thanjavur, 613401, India
| | - Jaehyuk Cho
- Department of Software Engineering & Division of Electronics and Information Engineering, Jeonbuk National University, Jeonju-si, 54896, Republic of Korea.
| | | |
Collapse
|
3
|
Ye RZ, Iezzi R. Intraoperative Augmented Reality for Vitreoretinal Surgery Using Edge Computing. J Pers Med 2025; 15:20. [PMID: 39852212 PMCID: PMC11766602 DOI: 10.3390/jpm15010020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2024] [Revised: 12/25/2024] [Accepted: 12/27/2024] [Indexed: 01/26/2025] Open
Abstract
Purpose: Augmented reality (AR) may allow vitreoretinal surgeons to leverage microscope-integrated digital imaging systems to analyze and highlight key retinal anatomic features in real time, possibly improving safety and precision during surgery. By employing convolutional neural networks (CNNs) for retina vessel segmentation, a retinal coordinate system can be created that allows pre-operative images of capillary non-perfusion or retinal breaks to be digitally aligned and overlayed upon the surgical field in real time. Such technology may be useful in assuring thorough laser treatment of capillary non-perfusion or in using pre-operative optical coherence tomography (OCT) to guide macular surgery when microscope-integrated OCT (MIOCT) is not available. Methods: This study is a retrospective analysis involving the development and testing of a novel image-registration algorithm for vitreoretinal surgery. Fifteen anonymized cases of pars plana vitrectomy with epiretinal membrane peeling, along with corresponding preoperative fundus photographs and optical coherence tomography (OCT) images, were retrospectively collected from the Mayo Clinic database. We developed a TPU (Tensor-Processing Unit)-accelerated CNN for semantic segmentation of retinal vessels from fundus photographs and subsequent real-time image registration in surgical video streams. An iterative patch-wise cross-correlation (IPCC) algorithm was developed for image registration, with a focus on optimizing processing speeds and maintaining high spatial accuracy. The primary outcomes measured were processing speed in frames per second (FPS) and the spatial accuracy of image registration, quantified by the Dice coefficient between registered and manually aligned images. Results: When deployed on an Edge TPU, the CNN model combined with our image-registration algorithm processed video streams at a rate of 14 FPS, which is superior to processing rates achieved on other standard hardware configurations. The IPCC algorithm efficiently aligned pre-operative and intraoperative images, showing high accuracy in comparison to manual registration. Conclusions: This study demonstrates the feasibility of using TPU-accelerated CNNs for enhanced AR in vitreoretinal surgery.
Collapse
Affiliation(s)
| | - Raymond Iezzi
- Department of Ophthalmology, Mayo Clinic, Rochester, MN 55905, USA;
| |
Collapse
|
4
|
Zhang Y, Huang Y, Hu K. Multi-scale object equalization learning network for intracerebral hemorrhage region segmentation. Neural Netw 2024; 179:106507. [PMID: 39003984 DOI: 10.1016/j.neunet.2024.106507] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2023] [Revised: 05/31/2024] [Accepted: 07/02/2024] [Indexed: 07/16/2024]
Abstract
Segmentation and the subsequent quantitative assessment of the target object in computed tomography (CT) images provide valuable information for the analysis of intracerebral hemorrhage (ICH) pathology. However, most existing methods lack a reasonable strategy to explore the discriminative semantics of multi-scale ICH regions, making it difficult to address the challenge of complex morphology in clinical data. In this paper, we propose a novel multi-scale object equalization learning network (MOEL-Net) for accurate ICH region segmentation. Specifically, we first introduce a shallow feature extraction module (SFEM) for obtaining shallow semantic representations to maintain sufficient and effective detailed location information. Then, a deep feature extraction module (DFEM) is leveraged to extract the deep semantic information of the ICH region from the combination of SFEM and original image features. To further achieve equalization learning in different scales of ICH regions, we introduce a multi-level semantic feature equalization fusion module (MSFEFM), which explores the equalized fusion features of the described objects with the assistance of shallow and deep semantic information provided by SFEM and DFEM. Driven by the above three designs, MOEL-Net shows a solid capacity to capture more discriminative features in various ICH region segmentation. To promote the research of clinical automatic ICH region segmentation, we collect two datasets, VMICH and FRICH (divided into Test A and Test B) for evaluation. Experimental results show that the proposed model achieves the Dice scores of 88.28%, 90.92%, and 90.95% on the VMICH, FRICH Test A, and Test B, respectively, which outperform fourteen competing methods.
Collapse
Affiliation(s)
- Yuan Zhang
- Key Laboratory of Intelligent Computing and Information Processing of Ministry of Education, Xiangtan University, Xiangtan 411105, China
| | - Yanglin Huang
- Key Laboratory of Intelligent Computing and Information Processing of Ministry of Education, Xiangtan University, Xiangtan 411105, China.
| | - Kai Hu
- Key Laboratory of Intelligent Computing and Information Processing of Ministry of Education, Xiangtan University, Xiangtan 411105, China; Key Laboratory of Medical Imaging and Artificial Intelligence of Hunan Province, Xiangnan University, Chenzhou 423000, China.
| |
Collapse
|
5
|
Ghislain F, Beaudelaire ST, Daniel T. An improved semi-supervised segmentation of the retinal vasculature using curvelet-based contrast adjustment and generalized linear model. Heliyon 2024; 10:e38027. [PMID: 39347436 PMCID: PMC11437861 DOI: 10.1016/j.heliyon.2024.e38027] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2024] [Revised: 08/12/2024] [Accepted: 09/16/2024] [Indexed: 10/01/2024] Open
Abstract
Diagnosis of most ophthalmic conditions, such as diabetic retinopathy, generally relies on an effective analysis of retinal blood vessels. Techniques that depend solely on the visual observation of clinicians can be tedious and prone to numerous errors. In this article, we propose a semi-supervised automated approach for segmenting blood vessels in retinal color images. Our method effectively combines some classical filters with a Generalized Linear Model (GLM). We first apply the Curvelet Transform along with the Contrast-Limited Histogram Adaptive Equalization (CLAHE) technique to significantly enhance the contrast of vessels in the retinal image during the preprocessing phase. We then use Gabor transform to extract features from the enhanced image. For retinal vasculature identification, we use a GLM learning model with a simple link identity function. Binarization is then performed using an automatic optimal threshold based on the maximum Youden index. A morphological cleaning operation is applied to remove isolated or unwanted segments from the final segmented image. The proposed model is evaluated using statistical parameters on images from three publicly available databases. We achieve average accuracies of 0.9593, 0.9553 and 0.9643, with Receiver Operating Characteristic (ROC) analysis yielding Area Under Curve (AUC) values of 0.9722, 0.9682 and 0.9767 for the CHASE_DB1, STARE and DRIVE databases, respectively. Compared to some of the best results from similar approaches published recently, our results exceed their performance on several datasets.
Collapse
Affiliation(s)
- Feudjio Ghislain
- Research Unit of Condensed Matter, Electronics and Signal Processing (UR-MACETS). Department of Physics, Faculty of Sciences, University of Dschang, P.O. Box 67, Dschang, Cameroon
- Research Unit of Automation and Applied Computer (UR-AIA), Electrical Engineering Department of IUT-FV, University of Dschang, P.O. Box: 134, Bandjoun, Cameroon
| | - Saha Tchinda Beaudelaire
- Research Unit of Automation and Applied Computer (UR-AIA), Electrical Engineering Department of IUT-FV, University of Dschang, P.O. Box: 134, Bandjoun, Cameroon
| | - Tchiotsop Daniel
- Research Unit of Automation and Applied Computer (UR-AIA), Electrical Engineering Department of IUT-FV, University of Dschang, P.O. Box: 134, Bandjoun, Cameroon
| |
Collapse
|
6
|
Ghislain F, Beaudelaire ST, Daniel T. An accurate unsupervised extraction of retinal vasculature using curvelet transform and classical morphological operators. Comput Biol Med 2024; 178:108801. [PMID: 38917533 DOI: 10.1016/j.compbiomed.2024.108801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2024] [Revised: 06/18/2024] [Accepted: 06/21/2024] [Indexed: 06/27/2024]
Abstract
BACKGROUND Many ophthalmic disorders such as diabetic retinopathy and hypertension can be early diagnosed by analyzing changes related to the vascular structure of the retina. Accuracy and efficiency of the segmentation of retinal blood vessels are important parameters that can help the ophthalmologist to better characterize the targeted anomalies. METHOD In this work, we propose a new method for accurate unsupervised automatic segmentation of retinal blood vessels based on a simple and adequate combination of classical filters. Initially, contrast of vessels in retinal image is significantly improved by adding the Curvelet Transform to commonly used Contrast-Limited Adaptive Histogram Equalization technique. Afterwards, a morphological operator using Top Hat is applied to highlight vascular network. Then, a global threshold-based Otsu technique using minimum of intra-class variance is applied for vessel detection. Finally, a cleanup operation based on Match Filter and First Derivative Order Gaussian with fixed parameters is used to remove unwanted or isolated segments. We test the proposed method on images from two publicly available STARE and DRIVE databases. RESULTS We achieve in terms of sensitivity, specificity and accuracy the respective average performances of 0.7407, 0.9878 and 0.9667 on the DRIVE database, then 0.7028, 0.9755 and 0.9507 on the STARE database. CONCLUSIONS Compared to some recent similar work, the obtained results are quite promising and can thus contribute to the optimization of automatic tools to aid in the diagnosis of eye disorders.
Collapse
Affiliation(s)
- Feudjio Ghislain
- Unité de Recherche de Matière Condensée, d'Electronique et de Traitements du Signal (URMACETS), Department of Physics, Faculty of Science, University of Dschang, P.O.Box 67, Dschang, Cameroon; Unité de Recherche d'Automatique et d'Informatique Appliquée (URAIA), IUT-FV de Bandjoun, Université de Dschang-Cameroun, B.P. 134, Bandjoun, Cameroon.
| | - Saha Tchinda Beaudelaire
- Unité de Recherche d'Automatique et d'Informatique Appliquée (URAIA), IUT-FV de Bandjoun, Université de Dschang-Cameroun, B.P. 134, Bandjoun, Cameroon.
| | - Tchiotsop Daniel
- Unité de Recherche d'Automatique et d'Informatique Appliquée (URAIA), IUT-FV de Bandjoun, Université de Dschang-Cameroun, B.P. 134, Bandjoun, Cameroon.
| |
Collapse
|
7
|
Faierstein K, Fiman M, Loutati R, Rubin N, Manor U, Am-Shalom A, Cohen-Shelly M, Blank N, Lotan D, Zhao Q, Schwammenthal E, Klempfner R, Zimlichman E, Raanani E, Maor E. Artificial Intelligence Assessment of Biological Age From Transthoracic Echocardiography: Discrepancies with Chronologic Age Predict Significant Excess Mortality. J Am Soc Echocardiogr 2024; 37:725-735. [PMID: 38740271 DOI: 10.1016/j.echo.2024.04.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Revised: 04/29/2024] [Accepted: 04/30/2024] [Indexed: 05/16/2024]
Abstract
BACKGROUND Age and sex can be estimated using artificial intelligence on the basis of various sources. The aims of this study were to test whether convolutional neural networks could be trained to estimate age and predict sex using standard transthoracic echocardiography and to evaluate the prognostic implications. METHODS The algorithm was trained on 76,342 patients, validated in 22,825 patients, and tested in 20,960 patients. It was then externally validated using data from a different hospital (n = 556). Finally, a prospective cohort of handheld point-of-care ultrasound devices (n = 319; ClinicalTrials.gov identifier NCT05455541) was used to confirm the findings. A multivariate Cox regression model was used to investigate the association between age estimation and chronologic age with overall survival. RESULTS The mean absolute error in age estimation was 4.9 years, with a Pearson correlation coefficient of 0.922. The probabilistic value of sex had an overall accuracy of 96.1% and an area under the curve of 0.993. External validation and prospective study cohorts yielded consistent results. Finally, survival analysis demonstrated that age prediction ≥5 years vs chronologic age was associated with an independent 34% increased risk for death during follow-up (P < .001). CONCLUSIONS Applying artificial intelligence to standard transthoracic echocardiography allows the prediction of sex and the estimation of age. Machine-based estimation is an independent predictor of overall survival and, with further evaluation, can be used for risk stratification and estimation of biological age.
Collapse
Affiliation(s)
- Kobi Faierstein
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel; Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel.
| | | | - Ranel Loutati
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel
| | | | - Uri Manor
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel; Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel
| | | | | | - Nimrod Blank
- Echocardiography Unit, Division of Cardiovascular Medicine, Baruch-Padeh Medical Center, Poria, Israel
| | - Dor Lotan
- Division of Cardiology, Department of Medicine, NewYork-Presbyterian Hospital, Columbia University Irving Medical Center, New York, New York
| | - Qiong Zhao
- Inova Heart and Vascular Institute, Inova Fairfax Hospital, Falls Church, Virginia
| | - Ehud Schwammenthal
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel; Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel; Aisap.ai, Ramat Gan, Israel
| | - Robert Klempfner
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel; Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel; Aisap.ai, Ramat Gan, Israel
| | - Eyal Zimlichman
- Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel
| | - Ehud Raanani
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel; Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel; Aisap.ai, Ramat Gan, Israel
| | - Elad Maor
- Leviev Cardiovascular Institute, Sheba Medical Center, Ramat Gan, Israel; Faculty of Medical and Health Sciences, Tel Aviv University, Tel Aviv, Israel; Aisap.ai, Ramat Gan, Israel
| |
Collapse
|
8
|
Zhang L, Wu X, Zhang J, Liu Z, Fan Y, Zheng L, Liu P, Song H, Lyu G. SEG-LUS: A novel ultrasound segmentation method for liver and its accessory structures based on muti-head self-attention. Comput Med Imaging Graph 2024; 113:102338. [PMID: 38290353 DOI: 10.1016/j.compmedimag.2024.102338] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2023] [Revised: 12/13/2023] [Accepted: 01/14/2024] [Indexed: 02/01/2024]
Abstract
Although liver ultrasound (US) is quick and convenient, it presents challenges due to patient variations. Previous research has predominantly focused on computer-aided diagnosis (CAD), particularly for disease analysis. However, characterizing liver US images is complex due to structural diversity and a limited number of samples. Normal liver US images are crucial, especially for standard section diagnosis. This study explicitly addresses Liver US standard sections (LUSS) and involves detailed labeling of eight anatomical structures. We propose SEG-LUS, a US image segmentation model for the liver and its accessory structures. In SEG-LUS, we have adopted the shifted windows feature encoder combined with the cross-attention mechanism to adapt to capturing image information at different scales and resolutions and address context mismatch and sample imbalance in the segmentation task. By introducing the UUF module, we achieve the perfect fusion of shallow and deep information, making the information retained by the network in the feature extraction process more comprehensive. We have improved the Focal Loss to tackle the imbalance of pixel-level distribution. The results show that the SEG-LUS model exhibits significant performance improvement, with mPA, mDice, mIOU, and mASD reaching 85.05%, 82.60%, 74.92%, and 0.31, respectively. Compared with seven state-of-the-art semantic segmentation methods, the mPA improves by 5.32%. SEG-LUS is positioned to serve as a crucial reference for research in computer-aided modeling using liver US images, thereby advancing the field of US medicine research.
Collapse
Affiliation(s)
- Lei Zhang
- College of Physics and Electronic Engineering, Northwest Normal University, Lanzhou 730070, China
| | - Xiuming Wu
- Department of Ultrasound, Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou 362000, China
| | - Jiansong Zhang
- College of Medicine, Huaqiao University, Quanzhou 362021, China
| | - Zhonghua Liu
- Department of Ultrasound, Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou 362000, China
| | - Yuling Fan
- College of Engineering, Huaqiao University, Quanzhou 362021, China
| | - Lan Zheng
- College of Engineering, Huaqiao University, Quanzhou 362021, China
| | - Peizhong Liu
- College of Medicine, Huaqiao University, Quanzhou 362021, China; College of Engineering, Huaqiao University, Quanzhou 362021, China; Collaborative Innovation Center for Maternal and Infant Health Service Application Technology, Quanzhou Medical College, Quanzhou 362011, China.
| | - Haisheng Song
- College of Physics and Electronic Engineering, Northwest Normal University, Lanzhou 730070, China.
| | - Guorong Lyu
- Collaborative Innovation Center for Maternal and Infant Health Service Application Technology, Quanzhou Medical College, Quanzhou 362011, China; Department of Ultrasound, The Second Hospital of Quanzhou Affiliated to Fujian Medical University, Quanzhou, China.
| |
Collapse
|
9
|
Yang X, He D, Li Y, Li C, Wang X, Zhu X, Sun H, Xu Y. Deep learning-based vessel extraction in 3D confocal microscope images of cleared human glioma tissues. BIOMEDICAL OPTICS EXPRESS 2024; 15:2498-2516. [PMID: 38633068 PMCID: PMC11019690 DOI: 10.1364/boe.516541] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/20/2023] [Revised: 02/25/2024] [Accepted: 02/26/2024] [Indexed: 04/19/2024]
Abstract
Comprehensive visualization and accurate extraction of tumor vasculature are essential to study the nature of glioma. Nowadays, tissue clearing technology enables 3D visualization of human glioma vasculature at micron resolution, but current vessel extraction schemes cannot well cope with the extraction of complex tumor vessels with high disruption and irregularity under realistic conditions. Here, we developed a framework, FineVess, based on deep learning to automatically extract glioma vessels in confocal microscope images of cleared human tumor tissues. In the framework, a customized deep learning network, named 3D ResCBAM nnU-Net, was designed to segment the vessels, and a novel pipeline based on preprocessing and post-processing was developed to refine the segmentation results automatically. On the basis of its application to a practical dataset, we showed that the FineVess enabled extraction of variable and incomplete vessels with high accuracy in challenging 3D images, better than other traditional and state-of-the-art schemes. For the extracted vessels, we calculated vascular morphological features including fractal dimension and vascular wall integrity of different tumor grades, and verified the vascular heterogeneity through quantitative analysis.
Collapse
Affiliation(s)
- Xiaodu Yang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, China
- Guangdong Provincial Key Laboratory of Medical Imaging Processing, Southern Medical University, Guangzhou, China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, China
| | - Dian He
- Clinical Biobank Center, Microbiome Medicine Center, Department of Laboratory Medicine, Zhujiang Hospital, Southern Medical University, Guangzhou, China
- Neurosurgery Center, The National Key Clinical Specialty, The Engineering Technology Research Center of Education Ministry of China on Diagnosis and Treatment of Cerebrovascular Disease, Guangdong Provincial Key Laboratory on Brain Function Repair and Regeneration, The Neurosurgery Institute of Guangdong Province Zhujiang Hospital, Southern Medical University, Guangzhou, China
| | - Yu Li
- School of Biomedical Engineering, Southern Medical University, Guangzhou, China
- Guangdong Provincial Key Laboratory of Medical Imaging Processing, Southern Medical University, Guangzhou, China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, China
| | - Chenyang Li
- Clinical Biobank Center, Microbiome Medicine Center, Department of Laboratory Medicine, Zhujiang Hospital, Southern Medical University, Guangzhou, China
- Neurosurgery Center, The National Key Clinical Specialty, The Engineering Technology Research Center of Education Ministry of China on Diagnosis and Treatment of Cerebrovascular Disease, Guangdong Provincial Key Laboratory on Brain Function Repair and Regeneration, The Neurosurgery Institute of Guangdong Province Zhujiang Hospital, Southern Medical University, Guangzhou, China
| | - Xinyue Wang
- Clinical Biobank Center, Microbiome Medicine Center, Department of Laboratory Medicine, Zhujiang Hospital, Southern Medical University, Guangzhou, China
- Neurosurgery Center, The National Key Clinical Specialty, The Engineering Technology Research Center of Education Ministry of China on Diagnosis and Treatment of Cerebrovascular Disease, Guangdong Provincial Key Laboratory on Brain Function Repair and Regeneration, The Neurosurgery Institute of Guangdong Province Zhujiang Hospital, Southern Medical University, Guangzhou, China
| | - Xingzheng Zhu
- Institute of Applied Artificial Intelligence of the Guangdong-Hong Kong-Macao Greater Bay Area, Shenzhen Polytechnic University, Shenzhen, China
| | - Haitao Sun
- Clinical Biobank Center, Microbiome Medicine Center, Department of Laboratory Medicine, Zhujiang Hospital, Southern Medical University, Guangzhou, China
- Neurosurgery Center, The National Key Clinical Specialty, The Engineering Technology Research Center of Education Ministry of China on Diagnosis and Treatment of Cerebrovascular Disease, Guangdong Provincial Key Laboratory on Brain Function Repair and Regeneration, The Neurosurgery Institute of Guangdong Province Zhujiang Hospital, Southern Medical University, Guangzhou, China
- Key Laboratory of Mental Health of the Ministry of Education, Guangdong-Hong Kong-Macao Greater Bay Area Center for Brain Science and Brain-Inspired Intelligence, Southern Medical University, Guangzhou, China
| | - Yingying Xu
- School of Biomedical Engineering, Southern Medical University, Guangzhou, China
- Guangdong Provincial Key Laboratory of Medical Imaging Processing, Southern Medical University, Guangzhou, China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, China
| |
Collapse
|
10
|
Zheng C, Li H, Ge Y, He Y, Yi Y, Zhu M, Sun H, Kong J. Retinal vessel segmentation based on multi-scale feature and style transfer. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:49-74. [PMID: 38303413 DOI: 10.3934/mbe.2024003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
Retinal vessel segmentation is very important for diagnosing and treating certain eye diseases. Recently, many deep learning-based retinal vessel segmentation methods have been proposed; however, there are still many shortcomings (e.g., they cannot obtain satisfactory results when dealing with cross-domain data or segmenting small blood vessels). To alleviate these problems and avoid overly complex models, we propose a novel network based on a multi-scale feature and style transfer (MSFST-NET) for retinal vessel segmentation. Specifically, we first construct a lightweight segmentation module named MSF-Net, which introduces the selective kernel (SK) module to increase the multi-scale feature extraction ability of the model to achieve improved small blood vessel segmentation. Then, to alleviate the problem of model performance degradation when segmenting cross-domain datasets, we propose a style transfer module and a pseudo-label learning strategy. The style transfer module is used to reduce the style difference between the source domain image and the target domain image to improve the segmentation performance for the target domain image. The pseudo-label learning strategy is designed to be combined with the style transfer module to further boost the generalization ability of the model. Moreover, we trained and tested our proposed MSFST-NET in experiments on the DRIVE and CHASE_DB1 datasets. The experimental results demonstrate that MSFST-NET can effectively improve the generalization ability of the model on cross-domain datasets and achieve improved retinal vessel segmentation results than other state-of-the-art methods.
Collapse
Affiliation(s)
- Caixia Zheng
- Jilin Animation Institute, Changchun 130013, China
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| | - Huican Li
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| | - Yingying Ge
- Jilin Animation Institute, Changchun 130013, China
| | - Yanlin He
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| | - Yugen Yi
- School of Software, Jiangxi Normal University, Nanchang 330022, China
| | - Meili Zhu
- Jilin Animation Institute, Changchun 130013, China
| | - Hui Sun
- School of Science and Technology, Changchun Humanities and Sciences College, Changchun 130117, China
| | - Jun Kong
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| |
Collapse
|
11
|
Ma Z, Li X. An improved supervised and attention mechanism-based U-Net algorithm for retinal vessel segmentation. Comput Biol Med 2024; 168:107770. [PMID: 38056215 DOI: 10.1016/j.compbiomed.2023.107770] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Revised: 11/08/2023] [Accepted: 11/26/2023] [Indexed: 12/08/2023]
Abstract
The segmentation results of retinal blood vessels are crucial for automatically diagnosing ophthalmic diseases such as diabetic retinopathy, hypertension, cardiovascular and cerebrovascular diseases. To improve the accuracy of vessel segmentation and better extract information about small vessels and edges, we introduce the U-Net algorithm with a supervised attention mechanism for retinal vessel segmentation. We achieve this by introducing a decoder fusion module (DFM) in the encoding part, effectively combining different convolutional blocks to extract features comprehensively. Additionally, in the decoding part of U-Net, we propose the context squeeze and excitation (CSE) decoding module to enhance important contextual feature information and the detection of tiny blood vessels. For the final output, we introduce the supervised fusion mechanism (SFM), which combines multiple branches from shallow to deep layers, effectively fusing multi-scale features and capturing information from different levels, fully integrating low-level and high-level features to improve segmentation performance. Our experimental results on the public datasets of DRIVE, STARE, and CHASED_B1 demonstrate the excellent performance of our proposed network.
Collapse
Affiliation(s)
- Zhendi Ma
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China
| | - Xiaobo Li
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China.
| |
Collapse
|
12
|
Yang S, Liang Y, Wu S, Sun P, Chen Z. SADSNet: A robust 3D synchronous segmentation network for liver and liver tumors based on spatial attention mechanism and deep supervision. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:707-723. [PMID: 38552134 DOI: 10.3233/xst-230312] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2024]
Abstract
Highlights • Introduce a data augmentation strategy to expand the required different morphological data during the training and learning phase, and improve the algorithm's feature learning ability for complex and diverse tumor morphology CT images.• Design attention mechanisms for encoding and decoding paths to extract fine pixel level features, improve feature extraction capabilities, and achieve efficient spatial channel feature fusion.• The deep supervision layer is used to correct and decode the final image data to provide high accuracy of results.• The effectiveness of this method has been affirmed through validation on the LITS, 3DIRCADb, and SLIVER datasets. BACKGROUND Accurately extracting liver and liver tumors from medical images is an important step in lesion localization and diagnosis, surgical planning, and postoperative monitoring. However, the limited number of radiation therapists and a great number of images make this work time-consuming. OBJECTIVE This study designs a spatial attention deep supervised network (SADSNet) for simultaneous automatic segmentation of liver and tumors. METHOD Firstly, self-designed spatial attention modules are introduced at each layer of the encoder and decoder to extract image features at different scales and resolutions, helping the model better capture liver tumors and fine structures. The designed spatial attention module is implemented through two gate signals related to liver and tumors, as well as changing the size of convolutional kernels; Secondly, deep supervision is added behind the three layers of the decoder to assist the backbone network in feature learning and improve gradient propagation, enhancing robustness. RESULTS The method was testing on LITS, 3DIRCADb, and SLIVER datasets. For the liver, it obtained dice similarity coefficients of 97.03%, 96.11%, and 97.40%, surface dice of 81.98%, 82.53%, and 86.29%, 95% hausdorff distances of 8.96 mm, 8.26 mm, and 3.79 mm, and average surface distances of 1.54 mm, 1.19 mm, and 0.81 mm. Additionally, it also achieved precise tumor segmentation, which with dice scores of 87.81% and 87.50%, surface dice of 89.63% and 84.26%, 95% hausdorff distance of 12.96 mm and 16.55 mm, and average surface distances of 1.11 mm and 3.04 mm on LITS and 3DIRCADb, respectively. CONCLUSION The experimental results show that the proposed method is effective and superior to some other methods. Therefore, this method can provide technical support for liver and liver tumor segmentation in clinical practice.
Collapse
Affiliation(s)
- Sijing Yang
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
| | - Yongbo Liang
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
| | - Shang Wu
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
| | - Peng Sun
- School of Electronic Engineering and Automation, Guilin University of Electronic Technology, Guilin, China
| | - Zhencheng Chen
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
- School of Electronic Engineering and Automation, Guilin University of Electronic Technology, Guilin, China
- Guangxi Colleges and Universities Key Laboratory of Biomedical Sensors and Intelligent Instruments, Guilin, China
- Guangxi Engineering Technology Research Center of Human Physiological Information Noninvasive Detection, Guilin, China
| |
Collapse
|
13
|
Tomić M, Vrabec R, Hendelja Đ, Kolarić V, Bulum T, Rahelić D. Diagnostic Accuracy of Hand-Held Fundus Camera and Artificial Intelligence in Diabetic Retinopathy Screening. Biomedicines 2023; 12:34. [PMID: 38255141 PMCID: PMC10813433 DOI: 10.3390/biomedicines12010034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2023] [Revised: 12/19/2023] [Accepted: 12/20/2023] [Indexed: 01/24/2024] Open
Abstract
Our study aimed to assess the role of a hand-held fundus camera and artificial intelligence (AI)-based grading system in diabetic retinopathy (DR) screening and determine its diagnostic accuracy in detecting DR compared with clinical examination and a standard fundus camera. This cross-sectional instrument validation study, as a part of the International Diabetes Federation (IDF) Diabetic Retinopathy Screening Project, included 160 patients (320 eyes) with type 2 diabetes (T2DM). After the standard indirect slit-lamp fundoscopy, each patient first underwent fundus photography with a standard 45° camera VISUCAM Zeiss and then with a hand-held camera TANG (Shanghai Zhi Tang Health Technology Co., Ltd.). Two retina specialists independently graded the images taken with the standard camera, while the images taken with the hand-held camera were graded using the DeepDR system and an independent IDF ophthalmologist. The three screening methods did not differ in detecting moderate/severe nonproliferative and proliferative DR. The area under the curve, sensitivity, specificity, positive predictive value, negative predictive value, positive likelihood ratio, negative likelihood ratio, kappa (ĸ) agreement, diagnostic odds ratio, and diagnostic effectiveness for a hand-held camera compared to clinical examination were 0.921, 89.1%, 100%, 100%, 91.4%, infinity, 0.11, 0.86, 936.48, and 94.9%, while compared to the standard fundus camera were 0.883, 83.2%, 100%, 100%, 87.3%, infinity, 0.17, 0.78, 574.6, and 92.2%. The results of our study suggest that fundus photography with a hand-held camera and AI-based grading system is a short, simple, and accurate method for the screening and early detection of DR, comparable to clinical examination and fundus photography with a standard camera.
Collapse
Affiliation(s)
- Martina Tomić
- Department of Ophthalmology, Vuk Vrhovac University Clinic for Diabetes, Endocrinology and Metabolic Diseases, Merkur University Hospital, Dugi dol 4a, 10000 Zagreb, Croatia
| | - Romano Vrabec
- Department of Ophthalmology, Vuk Vrhovac University Clinic for Diabetes, Endocrinology and Metabolic Diseases, Merkur University Hospital, Dugi dol 4a, 10000 Zagreb, Croatia
| | - Đurđica Hendelja
- Department of Ophthalmology, Vuk Vrhovac University Clinic for Diabetes, Endocrinology and Metabolic Diseases, Merkur University Hospital, Dugi dol 4a, 10000 Zagreb, Croatia
| | - Vilma Kolarić
- Department of Diabetes and Endocrinology, Vuk Vrhovac University Clinic for Diabetes, Endocrinology and Metabolic Diseases, Merkur University Hospital, Dugi dol 4a, 10000 Zagreb, Croatia
| | - Tomislav Bulum
- Department of Diabetes and Endocrinology, Vuk Vrhovac University Clinic for Diabetes, Endocrinology and Metabolic Diseases, Merkur University Hospital, Dugi dol 4a, 10000 Zagreb, Croatia
- School of Medicine, University of Zagreb, Šalata 3, 10000 Zagreb, Croatia
| | - Dario Rahelić
- Department of Diabetes and Endocrinology, Vuk Vrhovac University Clinic for Diabetes, Endocrinology and Metabolic Diseases, Merkur University Hospital, Dugi dol 4a, 10000 Zagreb, Croatia
- School of Medicine, Catholic University of Croatia, Ilica 242, 10000 Zagreb, Croatia
- School of Medicine, Josip Juraj Strossmayer University, Josipa Huttlera 4, 31000 Osijek, Croatia
| |
Collapse
|
14
|
Li C, Li Z, Liu W. TDCAU-Net: retinal vessel segmentation using transformer dilated convolutional attention-based U-Net method. Phys Med Biol 2023; 69:015003. [PMID: 38052089 DOI: 10.1088/1361-6560/ad1273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Accepted: 12/05/2023] [Indexed: 12/07/2023]
Abstract
Retinal vessel segmentation plays a vital role in the medical field, facilitating the identification of numerous chronic conditions based on retinal vessel images. These conditions include diabetic retinopathy, hypertensive retinopathy, glaucoma, and others. Although the U-Net model has shown promising results in retinal vessel segmentation, it tends to struggle with fine branching and dense vessel segmentation. To further enhance the precision of retinal vessel segmentation, we propose a novel approach called transformer dilated convolution attention U-Net (TDCAU-Net), which builds upon the U-Net architecture with improved Transformer-based dilated convolution attention mechanisms. The proposed model retains the three-layer architecture of the U-Net network. The Transformer component enables the learning of contextual information for each pixel in the image, while the dilated convolution attention prevents information loss. The algorithm efficiently addresses several challenges to optimize blood vessel detection. The process starts with five-step preprocessing of the images, followed by chunking them into segments. Subsequently, the retinal images are fed into the modified U-Net network introduced in this paper for segmentation. The study employs eye fundus images from the DRIVE and CHASEDB1 databases for both training and testing purposes. Evaluation metrics are utilized to compare the algorithm's results with state-of-the-art methods. The experimental analysis on both databases demonstrates that the algorithm achieves high values of sensitivity, specificity, accuracy, and AUC. Specifically, for the first database, the achieved values are 0.8187, 0.9756, 0.9556, and 0.9795, respectively. For the second database, the corresponding values are 0.8243, 0.9836, 0.9738, and 0.9878, respectively. These results demonstrate that the proposed approach outperforms state-of-the-art methods, achieving higher performance on both datasets. The TDCAU-Net model presented in this study exhibits substantial capabilities in accurately segmenting fine branching and dense vessels. The segmentation performance of the network surpasses that of the U-Net algorithm and several mainstream methods.
Collapse
Affiliation(s)
- Chunyang Li
- School of Electronics and Information Engineering, University of Science and Technology Liaoning, Anshan, People's Republic of China
| | - Zhigang Li
- School of Electronics and Information Engineering, University of Science and Technology Liaoning, Anshan, People's Republic of China
| | - Weikang Liu
- School of Electronics and Information Engineering, University of Science and Technology Liaoning, Anshan, People's Republic of China
| |
Collapse
|
15
|
Feng M, Xu J. Detection of ASD Children through Deep-Learning Application of fMRI. CHILDREN (BASEL, SWITZERLAND) 2023; 10:1654. [PMID: 37892317 PMCID: PMC10605350 DOI: 10.3390/children10101654] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Revised: 10/01/2023] [Accepted: 10/04/2023] [Indexed: 10/29/2023]
Abstract
Autism spectrum disorder (ASD) necessitates prompt diagnostic scrutiny to enable immediate, targeted interventions. This study unveils an advanced convolutional-neural-network (CNN) algorithm that was meticulously engineered to examine resting-state functional magnetic resonance imaging (fMRI) for early ASD detection in pediatric cohorts. The CNN architecture amalgamates convolutional, pooling, batch-normalization, dropout, and fully connected layers, optimized for high-dimensional data interpretation. Rigorous preprocessing yielded 22,176 two-dimensional echo planar samples from 126 subjects (56 ASD, 70 controls) who were sourced from the Autism Brain Imaging Data Exchange (ABIDE I) repository. The model, trained on 17,740 samples across 50 epochs, demonstrated unparalleled diagnostic metrics-accuracy of 99.39%, recall of 98.80%, precision of 99.85%, and an F1 score of 99.32%-and thereby eclipsed extant computational methodologies. Feature map analyses substantiated the model's hierarchical feature extraction capabilities. This research elucidates a deep learning framework for computer-assisted ASD screening via fMRI, with transformative implications for early diagnosis and intervention.
Collapse
Affiliation(s)
- Min Feng
- Nanjing Rehabilitation Medical Center, The Affiliated Brain Hospital, Nanjing Medical University, Nanjing 210029, China
- School of Chinese Language and Literature, Nanjing Normal University, Nanjing 210024, China
| | - Juncai Xu
- School of Engineering, Case Western Reserve University, Cleveland, OH 44106, USA;
| |
Collapse
|
16
|
Mojaver KHR, Zhao B, Leung E, Safaee SMR, Liboiron-Ladouceur O. Addressing the programming challenges of practical interferometric mesh based optical processors. OPTICS EXPRESS 2023; 31:23851-23866. [PMID: 37475226 DOI: 10.1364/oe.489493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Accepted: 06/08/2023] [Indexed: 07/22/2023]
Abstract
We demonstrate a novel mesh of Mach-Zehnder interferometers (MZIs) for programmable optical processors. We thoroughly analyze the benefits and drawbacks of previously known meshes and compare our newly proposed mesh with these prior architectures, highlighting its unique features and advantages. The proposed mesh, referred to as Bokun mesh, is an architecture that merges the attributes of the prior topologies Diamond and Clements. Similar to Diamond, Bokun provides diagonal paths passing through every individual MZI enabling direct phase monitoring. However, unlike Diamond and similar to Clements, Bokun maintains a minimum optical depth leading to better scalability. Providing the monitoring option, Bokun's programming is faster improving the total energy efficiency of the processor. The performance of Bokun mesh enabled by an optimal optical depth is also more resilient to the loss and fabrication imperfections compared to architectures with longer depth such as Reck and Diamond. Employing an efficient programming scheme, the proposed architecture improves energy efficiency by 83% maintaining the same computation accuracy for weight matrix changes at 2 kHz.
Collapse
|
17
|
Yan S, Xu W, Liu W, Yang H, Wang L, Deng Y, Gao F. TBENet:A two-branch boundary enhancement Network for cerebrovascular segmentation. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2023; 2023:1-7. [PMID: 38083508 DOI: 10.1109/embc40787.2023.10340540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2023]
Abstract
Cerebrovascular segmentation in digital subtraction angiography (DSA) images is the gold standard for clinical diagnosis. However, owing to the complexity of cerebrovascular, automatic cerebrovascular segmentation in DSA images is a challenging task. In this paper, we propose a CNN-based Two-branch Boundary Enhancement Network (TBENet) for automatic segmentation of cerebrovascular in DSA images. The TBENet is inspired by U-Net and designed as an encoder-decoder architecture. We propose an additional boundary branch to segment the boundary of cerebrovascular and a Main and Boundary branches Fusion Module (MBFM) to integrate the boundary branch outcome with the main branch outcome to achieve better segmentation performance. The TBENet was evaluated on HMCDSA (an in-house DSA cerebrovascular dataset), and reaches 0.9611, 0.7486, 0.7152, 0.9860 and 0.9556 in Accuracy, F1 score, Sensitivity, Specificity, and AUC, respectively. Meanwhile, we tested our TBENet on the public vessel segmentation benchmark DRIVE, and the results show that our TBENet can be extended to diverse vessel segmentation tasks.
Collapse
|
18
|
Shen J, Zhang Y, Liang H, Zhao Z, Zhu K, Qian K, Dong Q, Zhang X, Hu B. Depression Recognition From EEG Signals Using an Adaptive Channel Fusion Method via Improved Focal Loss. IEEE J Biomed Health Inform 2023; 27:3234-3245. [PMID: 37037251 DOI: 10.1109/jbhi.2023.3265805] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/12/2023]
Abstract
Depression is a serious and common psychiatric disease characterized by emotional and cognitive dysfunction. In addition, the rates of clinical diagnosis and treatment for depression are low. Therefore, the accurate recognition of depression is important for its effective treatment. Electroencephalogram (EEG) signals, which can objectively reflect the inner states of human brains, are regarded as promising physiological tools that can enable effective and efficient clinical depression diagnosis and recognition. However, one of the challenges regarding EEG-based depression recognition involves sufficiently optimizing the spatial information derived from the multichannel space of EEG signals. Consequently, we propose an adaptive channel fusion method via improved focal loss (FL) functions for depression recognition based on EEG signals to effectively address this challenge. In this method, we propose two improved FL functions that can enhance the separability of hard examples by upweighting their losses as optimization objectives and can optimize the channel weights by a proposed adaptive channel fusion framework. The experimental results obtained on two EEG datasets show that the developed channel fusion method can achieve improved classification performance. The learned channel weights include the individual characteristics of each EEG epoch, which can effectively optimize the spatial information of each EEG epoch via the channel fusion method. In addition, the proposed method performs better than the state-of-the-art channel fusion methods.
Collapse
|
19
|
Shamsan A, Senan EM, Shatnawi HSA. Automatic Classification of Colour Fundus Images for Prediction Eye Disease Types Based on Hybrid Features. Diagnostics (Basel) 2023; 13:diagnostics13101706. [PMID: 37238190 DOI: 10.3390/diagnostics13101706] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Revised: 05/06/2023] [Accepted: 05/08/2023] [Indexed: 05/28/2023] Open
Abstract
Early detection of eye diseases is the only solution to receive timely treatment and prevent blindness. Colour fundus photography (CFP) is an effective fundus examination technique. Because of the similarity in the symptoms of eye diseases in the early stages and the difficulty in distinguishing between the type of disease, there is a need for computer-assisted automated diagnostic techniques. This study focuses on classifying an eye disease dataset using hybrid techniques based on feature extraction with fusion methods. Three strategies were designed to classify CFP images for the diagnosis of eye disease. The first method is to classify an eye disease dataset using an Artificial Neural Network (ANN) with features from the MobileNet and DenseNet121 models separately after reducing the high dimensionality and repetitive features using Principal Component Analysis (PCA). The second method is to classify the eye disease dataset using an ANN on the basis of fused features from the MobileNet and DenseNet121 models before and after reducing features. The third method is to classify the eye disease dataset using ANN based on the fused features from the MobileNet and DenseNet121 models separately with handcrafted features. Based on the fused MobileNet and handcrafted features, the ANN attained an AUC of 99.23%, an accuracy of 98.5%, a precision of 98.45%, a specificity of 99.4%, and a sensitivity of 98.75%.
Collapse
Affiliation(s)
- Ahlam Shamsan
- Computer Department, Applied College, Najran University, Najran 66462, Saudi Arabia
| | - Ebrahim Mohammed Senan
- Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, Alrazi University, Sana'a, Yemen
| | | |
Collapse
|
20
|
Wang J, Cheng S, Tian J, Gao Y. A 2D CNN-LSTM hybrid algorithm using time series segments of EEG data for motor imagery classification. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/09/2023]
|
21
|
Sule O, Viriri S. Contrast Enhancement of RGB Retinal Fundus Images for Improved Segmentation of Blood Vessels Using Convolutional Neural Networks. J Digit Imaging 2023; 36:414-432. [PMID: 36456839 PMCID: PMC10039198 DOI: 10.1007/s10278-022-00738-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2020] [Revised: 09/27/2021] [Accepted: 11/03/2021] [Indexed: 12/03/2022] Open
Abstract
Retinal fundus images are non-invasively acquired and faced with low contrast, noise, and uneven illumination. The low-contrast problem makes objects in the retinal fundus image indistinguishable and the segmentation of blood vessels very challenging. Retinal blood vessels are significant because of their diagnostic importance in ophthalmologic diseases. This paper proposes improved retinal fundus images for optimal segmentation of blood vessels using convolutional neural networks (CNNs). This study explores some robust contrast enhancement tools on the RGB and the green channel of the retinal fundus images. The improved images undergo quality evaluation using mean square error (MSE), peak signal to noise ratio (PSNR), Similar Structure Index Matrix (SSIM), histogram, correlation, and intersection distance measures for histogram comparison before segmentation in the CNN-based model. The simulation results analysis reveals that the improved RGB quality outperforms the improved green channel. This revelation implies that the choice of RGB to the green channel for contrast enhancement is adequate and effectively improves the quality of the fundus images. This improved contrast will, in turn, boost the predictive accuracy of the CNN-based model during the segmentation process. The evaluation of the proposed method on the DRIVE dataset achieves an accuracy of 94.47, sensitivity of 70.92, specificity of 98.20, and AUC (ROC) of 97.56.
Collapse
Affiliation(s)
- Olubunmi Sule
- School of Mathematics, Statistics and Computer Science, University of KwaZulu-Natal, Durban, South Africa
| | - Serestina Viriri
- School of Mathematics, Statistics and Computer Science, University of KwaZulu-Natal, Durban, South Africa.
| |
Collapse
|
22
|
He S, Li Q, Li X, Zhang M. An optimized segmentation convolutional neural network with dynamic energy loss function for 3D reconstruction of lumbar spine MR images. Comput Biol Med 2023; 160:106839. [PMID: 37187132 DOI: 10.1016/j.compbiomed.2023.106839] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Revised: 03/21/2023] [Accepted: 03/26/2023] [Indexed: 05/17/2023]
Abstract
3D reconstruction for lumbar spine based on segmentation of Magnetic Resonance (MR) images is meaningful for diagnosis of degenerative lumbar spine diseases. However, spine MR images with unbalanced pixel distribution often cause the segmentation performance of Convolutional Neural Network (CNN) reduced. Designing a composite loss function for CNN is an effective way to enhance the segmentation capacity, yet composition loss values with fixed weight may still cause underfitting in CNN training. In this study, we designed a composite loss function with a dynamic weight, called Dynamic Energy Loss, for spine MR images segmentation. In our loss function, the weight percentage of different loss values could be dynamically adjusted during training, thus CNN could fast converge in earlier training stage and focus on detail learning in the later stage. Two datasets were used in control experiments, and the U-net CNN model with our proposed loss function achieved superior performance with Dice similarity coefficient values of 0.9484 and 0.8284 respectively, which were also verified by the Pearson correlation, Bland-Altman, and intra-class correlation coefficient analysis. Furthermore, to improve the 3D reconstruction based on the segmentation results, we proposed a filling algorithm to generate contextually related slices by computing the pixel-level difference between adjacent slices of segmented images, which could enhance the structural information of tissues between slices, and improve the performance of 3D lumbar spine model rendering. Our methods could help radiologists to build a 3D lumbar spine graphical model accurately for diagnosis while reducing burden of manual image reading.
Collapse
Affiliation(s)
- Siyuan He
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, Jilin, China
| | - Qi Li
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, Jilin, China; Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, Guangdong, China.
| | - Xianda Li
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, Jilin, China
| | - Mengchao Zhang
- Division of Radiology, China-Japan Union Hospital of Jilin University, Changchun, Jilin, China.
| |
Collapse
|
23
|
Wang G, Huang Y, Ma K, Duan Z, Luo Z, Xiao P, Yuan J. Automatic vessel crossing and bifurcation detection based on multi-attention network vessel segmentation and directed graph search. Comput Biol Med 2023; 155:106647. [PMID: 36848799 DOI: 10.1016/j.compbiomed.2023.106647] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2022] [Revised: 01/04/2023] [Accepted: 02/07/2023] [Indexed: 02/17/2023]
Abstract
Analysis of the vascular tree is the basic premise to automatically diagnose retinal biomarkers associated with ophthalmic and systemic diseases, among which accurate identification of intersection and bifurcation points is quite challenging but important for disentangling complex vascular network and tracking vessel morphology. In this paper, we present a novel directed graph search-based multi-attentive neural network approach to automatically segment the vascular network and separate intersections and bifurcations from color fundus images. Our approach uses multi-dimensional attention to adaptively integrate local features and their global dependencies while learning to focus on target structures at different scales to generate binary vascular maps. A directed graphical representation of the vascular network is constructed to represent the topology and spatial connectivity of the vascular structures. Using local geometric information including color difference, diameter, and angle, the complex vascular tree is decomposed into multiple sub-trees to finally classify and label vascular feature points. The proposed method has been tested on the DRIVE dataset and the IOSTAR dataset containing 40 images and 30 images, respectively, with 0.863 and 0.764 F1-score of detection points and average accuracy of 0.914 and 0.854 for classification points. These results demonstrate the superiority of our proposed method outperforming state-of-the-art methods in feature point detection and classification.
Collapse
Affiliation(s)
- Gengyuan Wang
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China; School of Life Sciences, South China University of Technology, Guangzhou, 510006, Guangdong, China
| | - Yuancong Huang
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China
| | - Ke Ma
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China
| | - Zhengyu Duan
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China
| | - Zhongzhou Luo
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China
| | - Peng Xiao
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China.
| | - Jin Yuan
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangzhou, 510060, China.
| |
Collapse
|
24
|
GDF-Net: A multi-task symmetrical network for retinal vessel segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104426] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022]
|
25
|
Expectation-maximization algorithm leads to domain adaptation for a perineural invasion and nerve extraction task in whole slide digital pathology images. Med Biol Eng Comput 2023; 61:457-473. [PMID: 36496513 DOI: 10.1007/s11517-022-02711-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2022] [Accepted: 10/22/2022] [Indexed: 12/14/2022]
Abstract
In addition to lymphatic and vascular channels, tumor cells can also spread via nerves, i.e., perineural invasion (PNI). PNI serves as an independent prognostic indicator in many malignancies. As a result, identifying and determining the extent of PNI is an important yet extremely tedious task in surgical pathology. In this work, we present a computational approach to extract nerves and PNI from whole slide histopathology images. We make manual annotations on selected prostate cancer slides once but then apply the trained model for nerve segmentation to both prostate cancer slides and head and neck cancer slides. For the purpose of multi-domain learning/prediction and investigation on the generalization capability of deep neural network, an expectation-maximization (EM)-based domain adaptation approach is proposed to improve the segmentation performance, in particular for the head and neck cancer slides. Experiments are conducted to demonstrate the segmentation performances. The average Dice coefficient for prostate cancer slides is 0.82 and 0.79 for head and neck cancer slides. Comparisons are then made for segmentations with and without the proposed EM-based domain adaptation on prostate cancer and head and neck cancer whole slide histopathology images from The Cancer Genome Atlas (TCGA) database and significant improvements are observed.
Collapse
|
26
|
Morano J, Hervella ÁS, Rouco J, Novo J, Fernández-Vigo JI, Ortega M. Weakly-supervised detection of AMD-related lesions in color fundus images using explainable deep learning. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 229:107296. [PMID: 36481530 DOI: 10.1016/j.cmpb.2022.107296] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/24/2022] [Revised: 11/16/2022] [Accepted: 11/29/2022] [Indexed: 06/17/2023]
Abstract
BACKGROUND AND OBJECTIVES Age-related macular degeneration (AMD) is a degenerative disorder affecting the macula, a key area of the retina for visual acuity. Nowadays, AMD is the most frequent cause of blindness in developed countries. Although some promising treatments have been proposed that effectively slow down its development, their effectiveness significantly diminishes in the advanced stages. This emphasizes the importance of large-scale screening programs for early detection. Nevertheless, implementing such programs for a disease like AMD is usually unfeasible, since the population at risk is large and the diagnosis is challenging. For the characterization of the disease, clinicians have to identify and localize certain retinal lesions. All this motivates the development of automatic diagnostic methods. In this sense, several works have achieved highly positive results for AMD detection using convolutional neural networks (CNNs). However, none of them incorporates explainability mechanisms linking the diagnosis to its related lesions to help clinicians to better understand the decisions of the models. This is specially relevant, since the absence of such mechanisms limits the application of automatic methods in the clinical practice. In that regard, we propose an explainable deep learning approach for the diagnosis of AMD via the joint identification of its associated retinal lesions. METHODS In our proposal, a CNN with a custom architectural setting is trained end-to-end for the joint identification of AMD and its associated retinal lesions. With the proposed setting, the lesion identification is directly derived from independent lesion activation maps; then, the diagnosis is obtained from the identified lesions. The training is performed end-to-end using image-level labels. Thus, lesion-specific activation maps are learned in a weakly-supervised manner. The provided lesion information is of high clinical interest, as it allows clinicians to assess the developmental stage of the disease. Additionally, the proposed approach allows to explain the diagnosis obtained by the models directly from the identified lesions and their corresponding activation maps. The training data necessary for the approach can be obtained without much extra work on the part of clinicians, since the lesion information is habitually present in medical records. This is an important advantage over other methods, including fully-supervised lesion segmentation methods, which require pixel-level labels whose acquisition is arduous. RESULTS The experiments conducted in 4 different datasets demonstrate that the proposed approach is able to identify AMD and its associated lesions with satisfactory performance. Moreover, the evaluation of the lesion activation maps shows that the models trained using the proposed approach are able to identify the pathological areas within the image and, in most cases, to correctly determine to which lesion they correspond. CONCLUSIONS The proposed approach provides meaningful information-lesion identification and lesion activation maps-that conveniently explains and complements the diagnosis, and is of particular interest to clinicians for the diagnostic process. Moreover, the data needed to train the networks using the proposed approach is commonly easy to obtain, what represents an important advantage in fields with particularly scarce data, such as medical imaging.
Collapse
Affiliation(s)
- José Morano
- Centro de Investigación CITIC, Universidade da Coruña, A Coruña, Spain; VARPA Research Group, Instituto de Investigación Biomédica de A Coruńa (INIBIC), Universidade da Coruña, A Coruña, Spain.
| | - Álvaro S Hervella
- Centro de Investigación CITIC, Universidade da Coruña, A Coruña, Spain; VARPA Research Group, Instituto de Investigación Biomédica de A Coruńa (INIBIC), Universidade da Coruña, A Coruña, Spain.
| | - José Rouco
- Centro de Investigación CITIC, Universidade da Coruña, A Coruña, Spain; VARPA Research Group, Instituto de Investigación Biomédica de A Coruńa (INIBIC), Universidade da Coruña, A Coruña, Spain.
| | - Jorge Novo
- Centro de Investigación CITIC, Universidade da Coruña, A Coruña, Spain; VARPA Research Group, Instituto de Investigación Biomédica de A Coruńa (INIBIC), Universidade da Coruña, A Coruña, Spain.
| | - José I Fernández-Vigo
- Department of Ophthalmology, Hospital Clínico San Carlos, Instituto de Investigación Sanitaria (IdISSC), Madrid, Spain; Department of Ophthalmology, Centro Internacional de Oftalmología Avanzada, Madrid, Spain.
| | - Marcos Ortega
- Centro de Investigación CITIC, Universidade da Coruña, A Coruña, Spain; VARPA Research Group, Instituto de Investigación Biomédica de A Coruńa (INIBIC), Universidade da Coruña, A Coruña, Spain.
| |
Collapse
|
27
|
Retinal blood vessel segmentation by using the MS-LSDNet network and geometric skeleton reconnection method. Comput Biol Med 2023; 153:106416. [PMID: 36586230 DOI: 10.1016/j.compbiomed.2022.106416] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Revised: 11/21/2022] [Accepted: 12/04/2022] [Indexed: 12/29/2022]
Abstract
Automatic retinal blood vessel segmentation is a key link in the diagnosis of ophthalmic diseases. Recent deep learning methods have achieved high accuracy in vessel segmentation but still face challenges in maintaining vascular structural connectivity. Therefore, this paper proposes a novel retinal blood vessel segmentation strategy that includes three stages: vessel structure detection, vessel branch extraction and broken vessel segment reconnection. First, we propose a multiscale linear structure detection network (MS-LSDNet), which improves the detection ability of fine blood vessels by learning the types of rich hierarchical features. In addition, to maintain the connectivity of the vascular structure in the process of binarization of the vascular probability map, an adaptive hysteresis threshold method for vascular extraction is proposed. Finally, we propose a vascular tree structure reconstruction algorithm based on a geometric skeleton to connect the broken vessel segments. Experimental results on three publicly available datasets show that compared with current state-of-the-art algorithms, our strategy effectively maintains the connectivity of retinal vascular tree structure.
Collapse
|
28
|
Gobinath C, Gopinath M. Attention aware fully convolutional deep learning model for retinal blood vessel segmentation. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2023. [DOI: 10.3233/jifs-224229] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
Recent reports indicate a rise in retinal issues, and automatic artery vein categorization offers data that is particularly instructive for the medical evaluation of serious retinal disorders including glaucoma and diabetic retinopathy. This work presents a competent and precise deep-learning model designed for vessel segmentation in retinal fundus imaging. This article aims to segment the retinal images using an attention-based dense fully convolutional neural network (A-DFCNN) after removing uncertainty. The artery extraction layers encompass vessel-specific convolutional blocks to focus the tiny blood vessels and dense layers with skip connections for feature propagation. Segmentation is associated with artery extraction layers via individual loss function. Blood vessel maps produced from individual loss functions are authenticated for performance. The proposed technique attains improved outcomes in terms of Accuracy (0.9834), Sensitivity (0.8553), and Specificity (0.9835) from DRIVE, STARE, and CHASE-DB1 datasets. The result demonstrates that the proposed A-DFCNN is capable of segmenting minute vessel bifurcation breakdowns during the training and testing phases.
Collapse
Affiliation(s)
- C. Gobinath
- School of Computer Science and Engineering, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| | - M.P. Gopinath
- School of Computer Science and Engineering, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| |
Collapse
|
29
|
Sengar N, Joshi RC, Dutta MK, Burget R. EyeDeep-Net: a multi-class diagnosis of retinal diseases using deep neural network. Neural Comput Appl 2023. [DOI: 10.1007/s00521-023-08249-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/22/2023]
|
30
|
Imran SMA, Saleem MW, Hameed MT, Hussain A, Naqvi RA, Lee SW. Feature preserving mesh network for semantic segmentation of retinal vasculature to support ophthalmic disease analysis. Front Med (Lausanne) 2023; 9:1040562. [PMID: 36714120 PMCID: PMC9880050 DOI: 10.3389/fmed.2022.1040562] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2022] [Accepted: 12/20/2022] [Indexed: 01/14/2023] Open
Abstract
Introduction Ophthalmic diseases are approaching an alarming count across the globe. Typically, ophthalmologists depend on manual methods for the analysis of different ophthalmic diseases such as glaucoma, Sickle cell retinopathy (SCR), diabetic retinopathy, and hypertensive retinopathy. All these manual assessments are not reliable, time-consuming, tedious, and prone to error. Therefore, automatic methods are desirable to replace conventional approaches. The accuracy of this segmentation of these vessels using automated approaches directly depends on the quality of fundus images. Retinal vessels are assumed as a potential biomarker for the diagnosis of many ophthalmic diseases. Mostly newly developed ophthalmic diseases contain minor changes in vasculature which is a critical job for the early detection and analysis of disease. Method Several artificial intelligence-based methods suggested intelligent solutions for automated retinal vessel detection. However, existing methods exhibited significant limitations in segmentation performance, complexity, and computational efficiency. Specifically, most of the existing methods failed in detecting small vessels owing to vanishing gradient problems. To overcome the stated problems, an intelligence-based automated shallow network with high performance and low cost is designed named Feature Preserving Mesh Network (FPM-Net) for the accurate segmentation of retinal vessels. FPM-Net employs a feature-preserving block that preserves the spatial features and helps in maintaining a better segmentation performance. Similarly, FPM-Net architecture uses a series of feature concatenation that also boosts the overall segmentation performance. Finally, preserved features, low-level input image information, and up-sampled spatial features are aggregated at the final concatenation stage for improved pixel prediction accuracy. The technique is reliable since it performs better on the DRIVE database, CHASE-DB1 database, and STARE dataset. Results and discussion Experimental outcomes confirm that FPM-Net outperforms state-of-the-art techniques with superior computational efficiency. In addition, presented results are achieved without using any preprocessing or postprocessing scheme. Our proposed method FPM-Net gives improvement results which can be observed with DRIVE datasets, it gives Se, Sp, and Acc as 0.8285, 0.98270, 0.92920, for CHASE-DB1 dataset 0.8219, 0.9840, 0.9728 and STARE datasets it produces 0.8618, 0.9819 and 0.9727 respectively. Which is a remarkable difference and enhancement as compared to the conventional methods using only 2.45 million trainable parameters.
Collapse
Affiliation(s)
| | | | | | - Abida Hussain
- Faculty of CS and IT, Superior University, Lahore, Pakistan
| | - Rizwan Ali Naqvi
- Department of Unmanned Vehicle Engineering, Sejong University, Seoul, Republic of Korea,*Correspondence: Rizwan Ali Naqvi ✉
| | - Seung Won Lee
- School of Medicine, Sungkyunkwan University, Suwon, Republic of Korea,Seung Won Lee ✉
| |
Collapse
|
31
|
Liu Y, Shen J, Yang L, Bian G, Yu H. ResDO-UNet: A deep residual network for accurate retinal vessel segmentation from fundus images. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104087] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
32
|
Liu Y, Shen J, Yang L, Yu H, Bian G. Wave-Net: A lightweight deep network for retinal vessel segmentation from fundus images. Comput Biol Med 2023; 152:106341. [PMID: 36463794 DOI: 10.1016/j.compbiomed.2022.106341] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2022] [Revised: 10/25/2022] [Accepted: 11/16/2022] [Indexed: 11/26/2022]
Abstract
Accurate segmentation of retinal vessels from fundus images is fundamental for the diagnosis of numerous diseases of eye, and an automated vessel segmentation method can effectively help clinicians to make accurate diagnosis for the patients and provide the appropriate treatment schemes. It is important to note that both thick and thin vessels play the key role for disease judgements. Because of complex factors, the precise segmentation of thin vessels is still a great challenge, such as the presence of various lesions, image noise, complex backgrounds and poor contrast in the fundus images. Recently, because of the advantage of context feature representation learning capabilities, deep learning has shown a remarkable segmentation performance on retinal vessels. However, it still has some shortcomings on high-precision retinal vessel extraction due to some factors, such as semantic information loss caused by pooling operations, limited receptive field, etc. To address these problems, this paper proposes a new lightweight segmentation network for precise retinal vessel segmentation, which is called as Wave-Net model on account of the whole shape. To alleviate the influence of semantic information loss problem to thin vessels, to acquire more contexts about micro structures and details, a detail enhancement and denoising block (DED) is proposed to improve the segmentation precision on thin vessels, which replaces the simple skip connections of original U-Net. On the other hand, it could well alleviate the influence of the semantic gap problem. Further, faced with limited receptive field, for multi-scale vessel detection, a multi-scale feature fusion block (MFF) is proposed to fuse cross-scale contexts to achieve higher segmentation accuracy and realize effective processing of local feature maps. Experiments indicate that proposed Wave-Net achieves an excellent performance on retinal vessel segmentation while maintaining a lightweight network design compared to other advanced segmentation methods, and it also has shown a better segmentation ability to thin vessels.
Collapse
Affiliation(s)
- Yanhong Liu
- School of Electrical and Information Engineering, Zhengzhou University, 450001, China; Robot Perception and Control Engineering Laboratory, Henan Province, 450001, China
| | - Ji Shen
- School of Electrical and Information Engineering, Zhengzhou University, 450001, China; Robot Perception and Control Engineering Laboratory, Henan Province, 450001, China
| | - Lei Yang
- School of Electrical and Information Engineering, Zhengzhou University, 450001, China; Robot Perception and Control Engineering Laboratory, Henan Province, 450001, China.
| | - Hongnian Yu
- School of Electrical and Information Engineering, Zhengzhou University, 450001, China; The Built Environment, Edinburgh Napier University, Edinburgh EH10 5DT, UK
| | - Guibin Bian
- School of Electrical and Information Engineering, Zhengzhou University, 450001, China; The State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| |
Collapse
|
33
|
Liu M, Liang H, Hou M. Research on cassava disease classification using the multi-scale fusion model based on EfficientNet and attention mechanism. FRONTIERS IN PLANT SCIENCE 2022; 13:1088531. [PMID: 36618625 PMCID: PMC9815107 DOI: 10.3389/fpls.2022.1088531] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/03/2022] [Accepted: 11/21/2022] [Indexed: 06/17/2023]
Abstract
Cassava disease is one of the leading causes to the serious decline of cassava yield. Because it is difficult to identify the characteristics of cassava disease, if not professional cassava growers, it will be prone to misjudgment. In order to strengthen the judgment of cassava diseases, the identification characteristics of cassava diseases such as different color of cassava leaf disease spots, abnormal leaf shape and disease spot area were studied. In this paper, deep convolutional neural network was used to classify cassava leaf diseases, and image classification technology was used to recognize and classify cassava leaf diseases. A lightweight module Multi-scale fusion model (MSFM) based on attention mechanism was proposed to extract disease features of cassava leaves to enhance the classification of disease features. The resulting feature map contained key disease identification information. The study used 22,000 cassava disease leaf images as a data set, including four different cassava leaf disease categories and healthy cassava leaves. The experimental results show that the cassava leaf disease classification model based on multi-scale fusion Convolutional Neural Network (CNN) improves EfficientNet compared with the original model, with the average recognition rate increased by nearly 4% and the average recognition rate up to 88.1%. It provides theoretical support and practical tools for the recognition and early diagnosis of plant disease leaves.
Collapse
Affiliation(s)
- Mingxin Liu
- School of Electronic and Information, Guangdong Ocean University, Zhanjiang, China
| | - Haofeng Liang
- School of Electronic and Information, Guangdong Ocean University, Zhanjiang, China
| | - Mingxin Hou
- School of Mechanical and Power Engineering, Guangdong Ocean University, Zhanjiang, China
| |
Collapse
|
34
|
Xu GX, Ren CX. SPNet: A novel deep neural network for retinal vessel segmentation based on shared decoder and pyramid-like loss. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.12.039] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
|
35
|
Kumar KS, Singh NP. An efficient registration-based approach for retinal blood vessel segmentation using generalized Pareto and fatigue pdf. Med Eng Phys 2022; 110:103936. [PMID: 36529622 DOI: 10.1016/j.medengphy.2022.103936] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 11/05/2022] [Accepted: 12/01/2022] [Indexed: 12/12/2022]
Abstract
Segmentation of Retinal Blood Vessel (RBV) extraction in the retina images and Registration of segmented RBV structure is implemented to identify changes in vessel structure by ophthalmologists in diagnosis of various illnesses like Glaucoma, Diabetes, and Hypertension's. The Retinal Blood Vessel provides blood to the inner retinal neurons, RBV are located mainly in internal retina but it may partly in the ganglion cell layer, following network failure haven't been identified with past methods. Classifications of accurate RBV and Registration of segmented blood vessels are challenging tasks in the low intensity background of Retinal Image. So, we projected a novel approach of segmentation of RBV extraction used matched filter of Generalized Pareto Probability Distribution Function (pdf) and Registration approach on feature-based segmented retinal blood vessel of Binary Robust Invariant Scalable Key point (BRISK). The BRISK provides the predefined sampling pattern as compared to Pdf. The BRISK feature is implemented for attention point recognition & matching approach for change in vessel structure. The proposed approaches contain 3 levels: pre-processing, matched filter-based Generalized Pareto pdf as a source along with the novel approach of fatigue pdf as a target, and BRISK framework is used for Registration on segmented retinal images of supply & intention images. This implemented system's performance is estimated in experimental analysis by the Average accuracy, Normalized Cross-Correlation (NCC), and computation time process of the segmented retinal source and target image. The NCC is main element to give more statistical information about retinal image segmentation. The proposed approach of Generalized Pareto value pdf has Average Accuracy of 95.21%, NCC of both image pairs is 93%, and Average accuracy of Registration of segmented source images and the target image is 98.51% respectively. The proposed approach of average computational time taken is around 1.4 s, which has been identified on boundary condition of Pdf function.
Collapse
Affiliation(s)
- K Susheel Kumar
- GITAM University, Bengaluru, 561203, India; National Institute of Technology Hamirpur, Himachal Pradesh 177005, India.
| | | |
Collapse
|
36
|
Iqbal S, Khan TM, Naveed K, Naqvi SS, Nawaz SJ. Recent trends and advances in fundus image analysis: A review. Comput Biol Med 2022; 151:106277. [PMID: 36370579 DOI: 10.1016/j.compbiomed.2022.106277] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2022] [Revised: 10/19/2022] [Accepted: 10/30/2022] [Indexed: 11/05/2022]
Abstract
Automated retinal image analysis holds prime significance in the accurate diagnosis of various critical eye diseases that include diabetic retinopathy (DR), age-related macular degeneration (AMD), atherosclerosis, and glaucoma. Manual diagnosis of retinal diseases by ophthalmologists takes time, effort, and financial resources, and is prone to error, in comparison to computer-aided diagnosis systems. In this context, robust classification and segmentation of retinal images are primary operations that aid clinicians in the early screening of patients to ensure the prevention and/or treatment of these diseases. This paper conducts an extensive review of the state-of-the-art methods for the detection and segmentation of retinal image features. Existing notable techniques for the detection of retinal features are categorized into essential groups and compared in depth. Additionally, a summary of quantifiable performance measures for various important stages of retinal image analysis, such as image acquisition and preprocessing, is provided. Finally, the widely used in the literature datasets for analyzing retinal images are described and their significance is emphasized.
Collapse
Affiliation(s)
- Shahzaib Iqbal
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan
| | - Tariq M Khan
- School of Computer Science and Engineering, University of New South Wales, Sydney, NSW, Australia.
| | - Khuram Naveed
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan; Department of Electrical and Computer Engineering, Aarhus University, Aarhus, Denmark
| | - Syed S Naqvi
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan
| | - Syed Junaid Nawaz
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan
| |
Collapse
|
37
|
Aswini A, Sivarani T. Modified capsule network for diabetic retinopathy detection and classification using fundus images. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2022. [DOI: 10.3233/jifs-221112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Diabetic retinopathy becomes an increasingly popular cause of vision loss in diabetic patients. Deep learning has recently received attention as one of the most popular methods for boosting performance in a range of sectors, including medical image analysis and classification. The proposed system comprises three steps; they are image preprocessing, image segmentation, and classification. In preprocessing, the image will be resized, denoising the image and enhancing the contrast of the image which is used for further processing. The lesion region of diabetic retinopathy fundus image is segmented by using Feature Fusion-based U-Net architecture. A blood vessel of a retinal image is extracted by using the spatial fuzzy c means clustering (SFCM) algorithm. Finally, the diabetic retinopathy images are classified using a modified capsule network. The convolution and primary capsule layers collect features from fundus images, while the class capsule and softmax layers decide whether the image belongs to a certain class. Using the Messidor dataset, the proposed system’s network efficiency is evaluated in terms of four performance indicators. The modified contrast limited adaptive histogram equalization technique enhanced the Peak Signal to Noise Ratio (PSNR), mean square error, and Structural Similarity Index Measure (SSIM) have average values of 36.18, 6.15, and 0.95, respectively. After enhancing the image, segmentation is performed to segment the vessel and lesion region. The segmentation accuracy is measured for the proposed segmentation algorithm by using two metrics namely intersection over union (IoU) and Dice similarity coefficient. Then modified capsule network is constructed for classifying the stages of diabetic retinopathy. The experimental result shows that the proposed modified capsule network got 98.57% of classification accuracy.
Collapse
Affiliation(s)
| | - T.S. Sivarani
- Department of EEE, Arunachala College of Engineering for Women, Vellichanthai, Tamil Nadu, India
| |
Collapse
|
38
|
Panda NR, Sahoo AK. A Detailed Systematic Review on Retinal Image Segmentation Methods. J Digit Imaging 2022; 35:1250-1270. [PMID: 35508746 PMCID: PMC9582172 DOI: 10.1007/s10278-022-00640-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2021] [Revised: 04/12/2022] [Accepted: 04/14/2022] [Indexed: 11/27/2022] Open
Abstract
The separation of blood vessels in the retina is a major aspect in detecting ailment and is carried out by segregating the retinal blood vessels from the fundus images. Moreover, it helps to provide earlier therapy for deadly diseases and prevent further impacts due to diabetes and hypertension. Many reviews already exist for this problem, but those reviews have presented the analysis of a single framework. Hence, this article on retinal segmentation review has revealed distinct methodologies with diverse frameworks that are utilized for blood vessel separation. The novelty of this review research lies in finding the best neural network model by comparing its efficiency. For that, machine learning (ML) and deep learning (DL) were compared and have been reported as the best model. Moreover, different datasets were used to segment the retinal blood vessels. The execution of each approach is compared based on the performance metrics such as sensitivity, specificity, and accuracy using publically accessible datasets like STARE, DRIVE, ROSE, REFUGE, and CHASE. This article discloses the implementation capacity of distinct techniques implemented for each segmentation method. Finally, the finest accuracy of 98% and sensitivity of 96% were achieved for the technique of Convolution Neural Network with Ranking Support Vector Machine (CNN-rSVM). Moreover, this technique has utilized public datasets to verify efficiency. Hence, the overall review of this article has revealed a method for earlier diagnosis of diseases to deliver earlier therapy.
Collapse
Affiliation(s)
- Nihar Ranjan Panda
- Department of Electronics and Communication Engineering, Silicon Institute of Technology, Bhubaneswar, Orissa, 751024, India.
| | - Ajit Kumar Sahoo
- Department of Electronics and Communication Engineering, National Institute of Technology, Rourkela, Odisha, 769008, India
| |
Collapse
|
39
|
Segmentation-Based Classification Deep Learning Model Embedded with Explainable AI for COVID-19 Detection in Chest X-ray Scans. Diagnostics (Basel) 2022; 12:diagnostics12092132. [PMID: 36140533 PMCID: PMC9497601 DOI: 10.3390/diagnostics12092132] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2022] [Revised: 08/26/2022] [Accepted: 08/30/2022] [Indexed: 12/16/2022] Open
Abstract
Background and Motivation: COVID-19 has resulted in a massive loss of life during the last two years. The current imaging-based diagnostic methods for COVID-19 detection in multiclass pneumonia-type chest X-rays are not so successful in clinical practice due to high error rates. Our hypothesis states that if we can have a segmentation-based classification error rate <5%, typically adopted for 510 (K) regulatory purposes, the diagnostic system can be adapted in clinical settings. Method: This study proposes 16 types of segmentation-based classification deep learning-based systems for automatic, rapid, and precise detection of COVID-19. The two deep learning-based segmentation networks, namely UNet and UNet+, along with eight classification models, namely VGG16, VGG19, Xception, InceptionV3, Densenet201, NASNetMobile, Resnet50, and MobileNet, were applied to select the best-suited combination of networks. Using the cross-entropy loss function, the system performance was evaluated by Dice, Jaccard, area-under-the-curve (AUC), and receiver operating characteristics (ROC) and validated using Grad-CAM in explainable AI framework. Results: The best performing segmentation model was UNet, which exhibited the accuracy, loss, Dice, Jaccard, and AUC of 96.35%, 0.15%, 94.88%, 90.38%, and 0.99 (p-value <0.0001), respectively. The best performing segmentation-based classification model was UNet+Xception, which exhibited the accuracy, precision, recall, F1-score, and AUC of 97.45%, 97.46%, 97.45%, 97.43%, and 0.998 (p-value <0.0001), respectively. Our system outperformed existing methods for segmentation-based classification models. The mean improvement of the UNet+Xception system over all the remaining studies was 8.27%. Conclusion: The segmentation-based classification is a viable option as the hypothesis (error rate <5%) holds true and is thus adaptable in clinical practice.
Collapse
|
40
|
Zhang Y, Ye F, Gao X. MCA-Net: Multi-Feature Coding and Attention Convolutional Neural Network for Predicting lncRNA-Disease Association. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2022; 19:2907-2919. [PMID: 34283719 DOI: 10.1109/tcbb.2021.3098126] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
With the advent of the era of big data, it is troublesome to accurately predict the associations between lncRNAs and diseases based on traditional biological experiments due to its time-consuming and subjective. In this paper, we propose a novel deep learning method for predicting lncRNA-disease associations using multi-feature coding and attention convolutional neural network (MCA-Net). We first calculate six similarity features to extract different types of lncRNA and disease feature information. Second, a multi-feature coding method is proposed to construct the feature vectors of lncRNA-disease association samples by integrating the six similarity features. Furthermore, an attention convolutional neural network is developed to identify lncRNA-disease associations under 10-fold cross-validation. Finally, we evaluate the performance of MCA-Net from different perspectives including the effects of the model parameters, distinct deep learning models, and the necessity of attention mechanism. We also compare MCA-Net with several state-of-the-art methods on three publicly available datasets, i.e., LncRNADisease, Lnc2Cancer, and LncRNADisease2.0. The results show that our MCA-Net outperforms the state-of-the-art methods on all three dataset. Besides, case studies on breast cancer and lung cancer further verify that MCA-Net is effective and accurate for the lncRNA-disease association prediction.
Collapse
|
41
|
Analysis Model of Image Colour Data Elements Based on Deep Neural Network. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:7631788. [PMID: 35898791 PMCID: PMC9313933 DOI: 10.1155/2022/7631788] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 06/08/2022] [Accepted: 06/14/2022] [Indexed: 11/26/2022]
Abstract
At present, the classification method used in image colour element analysis in China is still based on subjective visual evaluation. Because the evaluation process will inevitably be disturbed by human factors, it will not only have low efficiency but also produce large errors. To solve the above problems, this paper proposes an image colour data element analysis model based on depth neural network. Firstly, intelligent analysis of image colour data elements based on tensorflow is constructed, and the isomerized tensorflow framework is designed with the idea of Docker cluster to improve the efficiency of image element analysis. Secondly, considering the time error and spatial error diffusion model in the process of image analysis, the quantization modified error diffusion model is replaced by the original model for more accurate colour management. Image colour management is an important link in the process of image reproduction; the rotating principal component analysis method is used to correct and analyze the image colour error. Finally, using the properties of transfer learning and convolution neural network, an image colour element analysis model based on depth neural network is established. Large-scale image data is collected, and the effectiveness and reliability of the algorithm are verified from different angles. The results show that the new image colour analysis method can not only reveal the true colour components of the target image; furthermore, the real colour component of the target image also has high spectral data reconstruction accuracy, and the analysis results have strong adaptability.
Collapse
|
42
|
Biswas S, Khan MIA, Hossain MT, Biswas A, Nakai T, Rohdin J. Which Color Channel Is Better for Diagnosing Retinal Diseases Automatically in Color Fundus Photographs? LIFE (BASEL, SWITZERLAND) 2022; 12:life12070973. [PMID: 35888063 PMCID: PMC9321111 DOI: 10.3390/life12070973] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/27/2022] [Revised: 05/25/2022] [Accepted: 06/01/2022] [Indexed: 11/22/2022]
Abstract
Color fundus photographs are the most common type of image used for automatic diagnosis of retinal diseases and abnormalities. As all color photographs, these images contain information about three primary colors, i.e., red, green, and blue, in three separate color channels. This work aims to understand the impact of each channel in the automatic diagnosis of retinal diseases and abnormalities. To this end, the existing works are surveyed extensively to explore which color channel is used most commonly for automatically detecting four leading causes of blindness and one retinal abnormality along with segmenting three retinal landmarks. From this survey, it is clear that all channels together are typically used for neural network-based systems, whereas for non-neural network-based systems, the green channel is most commonly used. However, from the previous works, no conclusion can be drawn regarding the importance of the different channels. Therefore, systematic experiments are conducted to analyse this. A well-known U-shaped deep neural network (U-Net) is used to investigate which color channel is best for segmenting one retinal abnormality and three retinal landmarks.
Collapse
Affiliation(s)
- Sangeeta Biswas
- Faculty of Engineering, University of Rajshahi, Rajshahi 6205, Bangladesh; (M.I.A.K.); (M.T.H.)
- Correspondence: or
| | - Md. Iqbal Aziz Khan
- Faculty of Engineering, University of Rajshahi, Rajshahi 6205, Bangladesh; (M.I.A.K.); (M.T.H.)
| | - Md. Tanvir Hossain
- Faculty of Engineering, University of Rajshahi, Rajshahi 6205, Bangladesh; (M.I.A.K.); (M.T.H.)
| | - Angkan Biswas
- CAPM Company Limited, Bonani, Dhaka 1213, Bangladesh;
| | - Takayoshi Nakai
- Faculty of Engineering, Shizuoka University, Hamamatsu 432-8561, Japan;
| | - Johan Rohdin
- Faculty of Information Technology, Brno University of Technology, 61200 Brno, Czech Republic;
| |
Collapse
|
43
|
Asymmetry between right and left optical coherence tomography images identified using convolutional neural networks. Sci Rep 2022; 12:9925. [PMID: 35705663 PMCID: PMC9200978 DOI: 10.1038/s41598-022-14140-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2022] [Accepted: 05/17/2022] [Indexed: 11/08/2022] Open
Abstract
In a previous study, we identified biocular asymmetries in fundus photographs, and macula was discriminative area to distinguish left and right fundus images with > 99.9% accuracy. The purposes of this study were to investigate whether optical coherence tomography (OCT) images of the left and right eyes could be discriminated by convolutional neural networks (CNNs) and to support the previous result. We used a total of 129,546 OCT images. CNNs identified right and left horizontal images with high accuracy (99.50%). Even after flipping the left images, all of the CNNs were capable of discriminating them (DenseNet121: 90.33%, ResNet50: 88.20%, VGG19: 92.68%). The classification accuracy results were similar for the right and left flipped images (90.24% vs. 90.33%, respectively; p = 0.756). The CNNs also differentiated right and left vertical images (86.57%). In all cases, the discriminatory ability of the CNNs yielded a significant p value (< 0.001). However, the CNNs could not well-discriminate right horizontal images (50.82%, p = 0.548). There was a significant difference in identification accuracy between right and left horizontal and vertical OCT images and between flipped and non-flipped images. As this could result in bias in machine learning, care should be taken when flipping images.
Collapse
|
44
|
Deep Feature Extraction for Cymbidium Species Classification Using Global–Local CNN. HORTICULTURAE 2022. [DOI: 10.3390/horticulturae8060470] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/05/2023]
Abstract
Cymbidium is the most famous and widely distributed type of plant in the Orchidaceae family. It has extremely high ornamental and economic value. With the continuous development of the Cymbidium industry in recent years, it has become increasingly difficult to classify, identify, develop, and utilize orchids. In this study, a classification model GL-CNN based on a convolutional neural network was proposed to solve the problem of Cymbidium classification. First, the image set was expanded by four methods (mirror rotation, salt-and-pepper noise, image sharpening, and random angle flip), and then a cascade fusion strategy was used to fit the multiscale features obtained from the two branches. Comparing the performance of GL-CNN with other four classic models (AlexNet, ResNet50, GoogleNet, and VGG16), the results showed that GL-CNN achieves the highest classification prediction accuracy with a value of 94.13%. This model can effectively detect different species of Cymbidium and provide a reference for the identification of Cymbidium germplasm resources.
Collapse
|
45
|
Segmenting Retinal Vessels Using a Shallow Segmentation Network to Aid Ophthalmic Analysis. MATHEMATICS 2022. [DOI: 10.3390/math10091536] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
Abstract
Retinal blood vessels possess a complex structure in the retina and are considered an important biomarker for several retinal diseases. Ophthalmic diseases result in specific changes in the retinal vasculature; for example, diabetic retinopathy causes the retinal vessels to swell, and depending upon disease severity, fluid or blood can leak. Similarly, hypertensive retinopathy causes a change in the retinal vasculature due to the thinning of these vessels. Central retinal vein occlusion (CRVO) is a phenomenon in which the main vein causes drainage of the blood from the retina and this main vein can close completely or partially with symptoms of blurred vision and similar eye problems. Considering the importance of the retinal vasculature as an ophthalmic disease biomarker, ophthalmologists manually analyze retinal vascular changes. Manual analysis is a tedious task that requires constant observation to detect changes. The deep learning-based methods can ease the problem by learning from the annotations provided by an expert ophthalmologist. However, current deep learning-based methods are relatively inaccurate, computationally expensive, complex, and require image preprocessing for final detection. Moreover, existing methods are unable to provide a better true positive rate (sensitivity), which shows that the model can predict most of the vessel pixels. Therefore, this study presents the so-called vessel segmentation ultra-lite network (VSUL-Net) to accurately extract the retinal vasculature from the background. The proposed VSUL-Net comprises only 0.37 million trainable parameters and uses an original image as input without preprocessing. The VSUL-Net uses a retention block that specifically maintains the larger feature map size and low-level spatial information transfer. This retention block results in better sensitivity of the proposed VSUL-Net without using expensive preprocessing schemes. The proposed method was tested on three publicly available datasets: digital retinal images for vessel extraction (DRIVE), structured analysis of retina (STARE), and children’s heart health study in England database (CHASE-DB1) for retinal vasculature segmentation. The experimental results demonstrated that VSUL-Net provides robust segmentation of retinal vasculature with sensitivity (Sen), specificity (Spe), accuracy (Acc), and area under the curve (AUC) values of 83.80%, 98.21%, 96.95%, and 98.54%, respectively, for DRIVE, 81.73%, 98.35%, 97.17%, and 98.69%, respectively, for CHASE-DB1, and 86.64%, 98.13%, 97.27%, and 99.01%, respectively, for STARE datasets. The proposed method provides an accurate segmentation mask for deep ophthalmic analysis.
Collapse
|
46
|
CSAUNet: A cascade self-attention u-shaped network for precise fundus vessel segmentation. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103613] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
47
|
Hussain S, Guo F, Li W, Shen Z. DilUnet: A U-net based architecture for blood vessels segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 218:106732. [PMID: 35279601 DOI: 10.1016/j.cmpb.2022.106732] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/30/2021] [Revised: 02/24/2022] [Accepted: 03/03/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND AND OBJECTIVE Retinal image segmentation can help clinicians detect pathological disorders by studying changes in retinal blood vessels. This early detection can help prevent blindness and many other vision impairments. So far, several supervised and unsupervised methods have been proposed for the task of automatic blood vessel segmentation. However, the sensitivity and the robustness of these methods can be improved by correctly classifying more vessel pixels. METHOD We proposed an automatic, retinal blood vessel segmentation method based on the U-net architecture. This end-to-end framework utilizes preprocessing and a data augmentation pipeline for training. The architecture utilizes multiscale input and multioutput modules with improved skip connections and the correct use of dilated convolutions for effective feature extraction. In multiscale input, the input image is scaled down and concatenated with the output of convolutional blocks at different points in the encoder path to ensure the feature transfer of the original image. The multioutput module obtains upsampled outputs from each decoder block that are combined to obtain the final output. Skip paths connect each encoder block with the corresponding decoder block, and the whole architecture utilizes different dilation rates to improve the overall feature extraction. RESULTS The proposed method achieved an accuracy: of 0.9680, 0.9694, and 0.9701; sensitivity of 0.8837, 0.8263, and 0.8713; and Intersection Over Union (IOU) of 0.8698, 0.7951, and 0.8184 with three publicly available datasets: DRIVE, STARE, and CHASE, respectively. An ablation study is performed to show the contribution of each proposed module and technique. CONCLUSION The evaluation metrics revealed that the performance of the proposed method is higher than that of the original U-net and other U-net-based architectures, as well as many other state-of-the-art segmentation techniques, and that the proposed method is robust to noise.
Collapse
Affiliation(s)
- Snawar Hussain
- School of Automation, Central South University, Changsha, Hunan 410083, China
| | - Fan Guo
- School of Automation, Central South University, Changsha, Hunan 410083, China.
| | - Weiqing Li
- School of Automation, Central South University, Changsha, Hunan 410083, China
| | - Ziqi Shen
- School of Automation, Central South University, Changsha, Hunan 410083, China
| |
Collapse
|
48
|
Khan MS, Tafshir N, Alam KN, Dhruba AR, Khan MM, Albraikan AA, Almalki FA. Deep Learning for Ocular Disease Recognition: An Inner-Class Balance. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:5007111. [PMID: 35528343 PMCID: PMC9071974 DOI: 10.1155/2022/5007111] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/24/2022] [Revised: 03/18/2022] [Accepted: 04/12/2022] [Indexed: 12/25/2022]
Abstract
It can be challenging for doctors to identify eye disorders early enough using fundus pictures. Diagnosing ocular illnesses by hand is time-consuming, error-prone, and complicated. Therefore, an automated ocular disease detection system with computer-aided tools is necessary to detect various eye disorders using fundus pictures. Such a system is now possible as a consequence of deep learning algorithms that have improved image classification capabilities. A deep-learning-based approach to targeted ocular detection is presented in this study. For this study, we used state-of-the-art image classification algorithms, such as VGG-19, to classify the ODIR dataset, which contains 5000 images of eight different classes of the fundus. These classes represent different ocular diseases. However, the dataset within these classes is highly unbalanced. To resolve this issue, the work suggested converting this multiclass classification problem into a binary classification problem and taking the same number of images for both classifications. Then, the binary classifications were trained with VGG-19. The accuracy of the VGG-19 model was 98.13% for the normal (N) versus pathological myopia (M) class; the model reached an accuracy of 94.03% for normal (N) versus cataract (C), and the model provided an accuracy of 90.94% for normal (N) versus glaucoma (G). All of the other models also improve the accuracy when the data is balanced.
Collapse
Affiliation(s)
- Md Shakib Khan
- Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka 1229, Bangladesh
| | - Nafisa Tafshir
- Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka 1229, Bangladesh
| | - Kazi Nabiul Alam
- Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka 1229, Bangladesh
| | - Abdur Rab Dhruba
- Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka 1229, Bangladesh
| | - Mohammad Monirujjaman Khan
- Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka 1229, Bangladesh
| | - Amani Abdulrahman Albraikan
- Department of Computer Science, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Faris A. Almalki
- Department of Computer Engineering, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
| |
Collapse
|
49
|
Yang R, Yu J, Yin J, Liu K, Xu S. An FA-SegNet Image Segmentation Model Based on Fuzzy Attention and Its Application in Cardiac MRI Segmentation. INT J COMPUT INT SYS 2022. [DOI: 10.1007/s44196-022-00080-x] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/25/2022] Open
Abstract
AbstractAiming at the medical images segmentation with low-recognition and high background noise, a deep convolution neural network image segmentation model based on fuzzy attention mechanism is proposed, which is called FA-SegNet. It takes SegNet as the basic framework. In the down-sampling module for image feature extraction, a fuzzy channel-attention module is added to strengthen the discrimination of different target regions. In the up-sampling module for image size restoration and multi-scale feature fusion, a fuzzy spatial-attention module is added to reduce the loss of image details and expand the receptive field. In this paper, fuzzy cognition is introduced into the feature fusion of CNNs. Based on the attention mechanism, fuzzy membership is used to re-calibrate the importance of the pixel value in local regions. It can strengthen the distinguishing ability of image features, and the fusion ability of the contextual information, which improves the segmentation accuracy of the target regions. Taking MRI segmentation as an experimental example, multiple targets such as the left ventricles, right ventricles, and left ventricular myocardium are selected as the segmentation targets. The pixels accuracy is 92.47%, the mean intersection to union is 86.18%, and the Dice coefficient is 92.44%, which are improved compared with other methods. It verifies the accuracy and applicability of the proposed method for the medical images segmentation, especially the targets with low-recognition and serious occlusion.
Collapse
|
50
|
Xu X, Wang Y, Liang Y, Luo S, Wang J, Jiang W, Lai X. Retinal Vessel Automatic Segmentation Using SegNet. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2022; 2022:3117455. [PMID: 35378728 PMCID: PMC8976667 DOI: 10.1155/2022/3117455] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Revised: 03/10/2022] [Accepted: 03/12/2022] [Indexed: 12/22/2022]
Abstract
Extracting retinal vessels accurately is very important for diagnosing some diseases such as diabetes retinopathy, hypertension, and cardiovascular. Clinically, experienced ophthalmologists diagnose these diseases through segmenting retinal vessels manually and analysing its structural feature, such as tortuosity and diameter. However, manual segmentation of retinal vessels is a time-consuming and laborious task with strong subjectivity. The automatic segmentation technology of retinal vessels can not only reduce the burden of ophthalmologists but also effectively solve the problem that is a lack of experienced ophthalmologists in remote areas. Therefore, the automatic segmentation technology of retinal vessels is of great significance for clinical auxiliary diagnosis and treatment of ophthalmic diseases. A method using SegNet is proposed in this paper to improve the accuracy of the retinal vessel segmentation. The performance of the retinal vessel segmentation model with SegNet is evaluated on the three public datasets (DRIVE, STARE, and HRF) and achieved accuracy of 0.9518, 0.9683, and 0.9653, sensitivity of 0.7580, 0.7747, and 0.7070, specificity of 0.9804, 0.9910, and 0.9885, F 1 score of 0.7992, 0.8369, and 0.7918, MCC of 0.7749, 0.8227, and 0.7643, and AUC of 0.9750, 0.9893, and 0.9740, respectively. The experimental results showed that the method proposed in this research presented better results than many classical methods studied and may be expected to have clinical application prospects.
Collapse
Affiliation(s)
- Xiaomei Xu
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| | - Yixin Wang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| | - Yu Liang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| | - Siyuan Luo
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| | - Jianqing Wang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| | - Weiwei Jiang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| | - Xiaobo Lai
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou 310053, China
| |
Collapse
|