1
|
Ding L, Zhang C, Lyu X, Cheng D, Xu S. Unified Framework for Enhancement of Low-Quality Fundus Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01509-3. [PMID: 40301293 DOI: 10.1007/s10278-025-01509-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/24/2025] [Revised: 04/14/2025] [Accepted: 04/15/2025] [Indexed: 05/01/2025]
Abstract
Compared to desktop fundus cameras, handheld ones offer portability and affordability, although they often produce lower-quality images. This paper primarily addresses the issue of reduced image quality commonly associated with images captured by handheld fundus cameras. We first collected 538 fundus images obtained from handheld devices to form a dataset called Mule. A unified framework that consists of three main modules is then proposed to enhance the quality of fundus images. The Light Balance Module is employed first to suppress overexposure and underexposure. This is followed by the Super Resolution Module to enhance vascular details. Finally, the Vessel Enhancement Module is applied to improve image contrast. And a special preservation strategy is additionally applied to retain mocular features in the final fundus image. Objective evaluations demonstrate that the proposed framework yields the most promising results. Further experiments also suggest that it improves accuracy in downstream tasks, such as vessel segmentation, optic disc/optic cup detection, macula detection, and fundus image quality assessment. Our code is available at: https://github.com/Alen880/UFELQ.
Collapse
Affiliation(s)
- Lihua Ding
- School of Information Science and Technology, HangZhou Normal University, Hangzhou, 311100, Zhejiang, China
| | - Chengyi Zhang
- School of Information Science and Technology, HangZhou Normal University, Hangzhou, 311100, Zhejiang, China
| | - Xingzheng Lyu
- Hangzhou Mocular Medical Technology Inc., Lin'an District Future Eye Valley, Hangzhou, 311100, Zhejiang, China
| | - Deji Cheng
- Hangzhou Mocular Medical Technology Inc., Lin'an District Future Eye Valley, Hangzhou, 311100, Zhejiang, China
| | - Shuchang Xu
- School of Information Science and Technology, HangZhou Normal University, Hangzhou, 311100, Zhejiang, China.
| |
Collapse
|
2
|
Yang B, Han H, Zhang W, Li H. General retinal image enhancement via reconstruction: Bridging distribution shifts using latent diffusion adaptors. Med Image Anal 2025; 103:103603. [PMID: 40300379 DOI: 10.1016/j.media.2025.103603] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Revised: 01/21/2025] [Accepted: 04/12/2025] [Indexed: 05/01/2025]
Abstract
Deep learning-based fundus image enhancement has attracted extensive research attention recently, which has shown remarkable effectiveness in improving the visibility of low-quality images. However, these methods are often constrained to specific datasets and degradations, leading to poor generalization capabilities and having challenges in the fine-tuning process. Therefore, a general method for fundus image enhancement is proposed for improved generalizability and flexibility, which decomposes the enhancement task into reconstruction and adaptation phases. In the reconstruction phase, self-supervised training with unpaired data is employed, allowing the utilization of extensive public datasets to improve the generalizability of the model. During the adaptation phase, the model is fine-tuned according to the target datasets and their degradations, utilizing the pre-trained weights from the reconstruction. The proposed method improves the feasibility of latent diffusion models for retinal image enhancement. Adaptation loss and enhancement adaptor are proposed in autoencoders and diffusion networks for fewer paired training data, fewer trainable parameters, and faster convergence compared with training from scratch. The proposed method can be easily fine-tuned and experiments demonstrate the adaptability for different datasets and degradations. Additionally, the reconstruction-adaptation framework can be utilized in different backbones and other modalities, which shows its generality.
Collapse
Affiliation(s)
- Bingyu Yang
- Beijing Institute of Technology, Beijing, 100081, China
| | - Haonan Han
- Beijing Institute of Technology, Beijing, 100081, China
| | - Weihang Zhang
- Beijing Institute of Technology, Beijing, 100081, China
| | - Huiqi Li
- Beijing Institute of Technology, Beijing, 100081, China.
| |
Collapse
|
3
|
Wang X, Gong D, Chen Y, Zong Z, Li M, Fan K, Jia L, Cao Q, Liu Q, Yang Q. Hybrid CNN-Mamba model for multi-scale fundus image enhancement. BIOMEDICAL OPTICS EXPRESS 2025; 16:1104-1117. [PMID: 40109520 PMCID: PMC11919352 DOI: 10.1364/boe.542471] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Revised: 11/26/2024] [Accepted: 02/06/2025] [Indexed: 03/22/2025]
Abstract
This study proposes a multi-scale fundus image enhancement approach that combines CNN with Mamba, demonstrating clear superiority across multiple benchmarks. The model consistently achieves top performance on public datasets, with the lowest FID and KID scores, and the highest PSNR and SSIM values, particularly excelling at larger image resolutions. Notably, its performance improves as the image size increases, with several metrics reaching optimal values at 1024 × 1024 resolution. Scale generalizability further highlights the model's exceptional structural preservation capability. Additionally, its high VSD and IOU scores in segmentation tasks further validate its practical effectiveness, making it a valuable tool for enhancing fundus images and improving diagnostic accuracy.
Collapse
Affiliation(s)
- Xiaopeng Wang
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| | - Di Gong
- China-Japan Friendship Hospital, Beijing 100029, China
| | - Yi Chen
- China-Japan Friendship Hospital, Beijing 100029, China
| | - Zheng Zong
- Beijing Information Science and Technology University, Beijing 100096, China
| | - Meng Li
- Tianjin University of Technology, Tianjin 300384, China
| | - Kun Fan
- Taihe Intelligent Technology Group Co., Ltd., Anhui 230000, China
| | - Lina Jia
- Beijing Etop Smartinfo Technology Co., Ltd., Beijing 102617, China
| | - Qiyuan Cao
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| | - Qiang Liu
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| | - Qiang Yang
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| |
Collapse
|
4
|
Liu Q, Han Y, Shen L, Du J, Tania MH. Adaptive enhancement of shoulder x-ray images using tissue attenuation and type-II fuzzy sets. PLoS One 2025; 20:e0316585. [PMID: 39913419 PMCID: PMC11801559 DOI: 10.1371/journal.pone.0316585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 12/12/2024] [Indexed: 02/09/2025] Open
Abstract
Shoulder X-ray images typically have low contrast and high noise levels, making it challenging to distinguish and identify subtle anatomical structures. While existing image enhancement techniques are effective in improving contrast, they often overlook the enhancement of sharpness, especially when amplifying blurring and noise. These techniques may improve detail contrast but fail to maintain overall image clarity and the distinction between the target and the background. To address these issues, we propose a novel image enhancement method aimed at simultaneously improving both the contrast and sharpness of shoulder X-ray images. The method integrates automatic tissue attenuation techniques, which enhance the image contrast by removing non-essential tissue components while preserving important tissues and bones. Additionally, we apply an improved Type-II fuzzy set algorithm to further optimize image sharpness. By simultaneously enhancing contrast and sharpness, the method significantly improves image quality and detail distinguishability. When tested on certain images from the MURA dataset, the proposed method achieved the best or second-best results, outperforming five no-reference image quality assessment metrics. In comparative studies, the method demonstrated significant performance advantages over 10 contemporary X-ray image enhancement algorithms and was validated through ablation experiments to confirm the effectiveness of each module.
Collapse
Affiliation(s)
- Qifeng Liu
- Centre for Big Data Research in Health, University of New South Wales, Sydney, Australia
| | - Yong Han
- School of Software Engineering, Xi’an Jiaotong University, Xi’an, China
| | - Lu Shen
- International Digital Economy College, Minjiang University, Fuzhou, China
| | - Jialei Du
- Business School, University of New South Wales, Sydney, Australia
| | - Marzia Hoque Tania
- Centre for Big Data Research in Health, University of New South Wales, Sydney, Australia
| |
Collapse
|
5
|
Yu Z, Zhao B, Zhang S, Chen X, Yan F, Feng J, Peng T, Zhang XY. HiFi-Syn: Hierarchical granularity discrimination for high-fidelity synthesis of MR images with structure preservation. Med Image Anal 2025; 100:103390. [PMID: 39602984 DOI: 10.1016/j.media.2024.103390] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2023] [Revised: 11/01/2024] [Accepted: 11/08/2024] [Indexed: 11/29/2024]
Abstract
Synthesizing medical images while preserving their structural information is crucial in medical research. In such scenarios, the preservation of anatomical content becomes especially important. Although recent advances have been made by incorporating instance-level information to guide translation, these methods overlook the spatial coherence of structural-level representation and the anatomical invariance of content during translation. To address these issues, we introduce hierarchical granularity discrimination, which exploits various levels of semantic information present in medical images. Our strategy utilizes three levels of discrimination granularity: pixel-level discrimination using a Brain Memory Bank, structure-level discrimination on each brain structure with a re-weighting strategy to focus on hard samples, and global-level discrimination to ensure anatomical consistency during translation. The image translation performance of our strategy has been evaluated on three independent datasets (UK Biobank, IXI, and BraTS 2018), and it has outperformed state-of-the-art algorithms. Particularly, our model excels not only in synthesizing normal structures but also in handling abnormal (pathological) structures, such as brain tumors, despite the variations in contrast observed across different imaging modalities due to their pathological characteristics. The diagnostic value of synthesized MR images containing brain tumors has been evaluated by radiologists. This indicates that our model may offer an alternative solution in scenarios where specific MR modalities of patients are unavailable. Extensive experiments further demonstrate the versatility of our method, providing unique insights into medical image translation.
Collapse
Affiliation(s)
- Ziqi Yu
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Faculty of Medical Imaging Technology, College of Health Science and Technology, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Clinical Neuroscience Center, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China; National Engineering Research Center of Advanced Magnetic Resonance Technologies for Diagnosis and Therapy, Shanghai Jiao Tong University, Shanghai, China
| | - Botao Zhao
- Ping An Technology (Shenzhen) Co., Ltd., China
| | - Shengjie Zhang
- Institute of Science and Technology for Brain-Inspired Intelligence, MOE Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Fudan University, Shanghai, China
| | - Xiang Chen
- Institute of Science and Technology for Brain-Inspired Intelligence, MOE Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Fudan University, Shanghai, China
| | - Fuhua Yan
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Faculty of Medical Imaging Technology, College of Health Science and Technology, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Clinical Neuroscience Center, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Jianfeng Feng
- Institute of Science and Technology for Brain-Inspired Intelligence, MOE Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Fudan University, Shanghai, China
| | - Tingying Peng
- Helmholtz AI, Helmholtz zentrum Muenchen, Munich, Germany.
| | - Xiao-Yong Zhang
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Faculty of Medical Imaging Technology, College of Health Science and Technology, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Clinical Neuroscience Center, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China; National Engineering Research Center of Advanced Magnetic Resonance Technologies for Diagnosis and Therapy, Shanghai Jiao Tong University, Shanghai, China.
| |
Collapse
|
6
|
Liu B, Du Y. MVE-Net: A label-free microscopic image visual enhancement network via mRetinex and nonreference loss guidance. Comput Biol Med 2025; 184:109456. [PMID: 39581123 DOI: 10.1016/j.compbiomed.2024.109456] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2024] [Revised: 11/07/2024] [Accepted: 11/19/2024] [Indexed: 11/26/2024]
Abstract
Label-free microscopic cell image analysis (segmentation, detection, counting, e.g.) is elementary for unravelling the biological functions of cells and their organelles. However, low contrast, darker brightness, background inhomogeneous, and weak edges of cells cause challenges in subsequent cell image analysis processes. To address these challenges, a Microscopic Visual Enhancement Network (MVE-Net) is proposed to improve microscopic visual effects through pre-enhancement and enhancement processes. In the pre-enhancement stage, to overcome the difficulty of acquiring paired or unpaired images for training, the mRetinex block is proposed to guide the pre-enhancement network to image contrast, cell details, and structural features. Furthermore, a multi-scale extraction module is employed to extract and fuse cell texture and structural features from the pre-enhanced images at various scales, guiding the generator training. In the enhancement stage, a nonreference loss block is designed, incorporating spatial consistency, uneven illumination smoothness, and exposure adjustment loss terms, to further enhance the contrast between cells and the background, smooth the inhomogeneous background, and adjust overall image brightness, thereby guiding the generator's enhancement process and improving the visual effect of microscopic images. Experiments on the LIVECell and PNT1A datasets demonstrate that MVE-Net outperforms state-of-the-art image enhancement methods, significantly improving image contrast, brightness, cell detail, and structural features without the need for paired or unpaired reference standard images for training.
Collapse
Affiliation(s)
- Bo Liu
- School of Biomedical Science, Huaqiao University, Quanzhou, Fujian, 362000, China
| | - Yongzhao Du
- School of Biomedical Science, Huaqiao University, Quanzhou, Fujian, 362000, China; College of Engineering, Huaqiao University, Quanzhou, Fujian, 362000, China; College of Internet of Things Industry, Huaqiao University, Quanzhou, Fujian, 362021, China.
| |
Collapse
|
7
|
He C, Li K, Xu G, Yan J, Tang L, Zhang Y, Wang Y, Li X. HQG-Net: Unpaired Medical Image Enhancement With High-Quality Guidance. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:18404-18418. [PMID: 37796672 DOI: 10.1109/tnnls.2023.3315307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/07/2023]
Abstract
Unpaired medical image enhancement (UMIE) aims to transform a low-quality (LQ) medical image into a high-quality (HQ) one without relying on paired images for training. While most existing approaches are based on Pix2Pix/CycleGAN and are effective to some extent, they fail to explicitly use HQ information to guide the enhancement process, which can lead to undesired artifacts and structural distortions. In this article, we propose a novel UMIE approach that avoids the above limitation of existing methods by directly encoding HQ cues into the LQ enhancement process in a variational fashion and thus model the UMIE task under the joint distribution between the LQ and HQ domains. Specifically, we extract features from an HQ image and explicitly insert the features, which are expected to encode HQ cues, into the enhancement network to guide the LQ enhancement with the variational normalization module. We train the enhancement network adversarially with a discriminator to ensure the generated HQ image falls into the HQ domain. We further propose a content-aware loss to guide the enhancement process with wavelet-based pixel-level and multiencoder-based feature-level constraints. Additionally, as a key motivation for performing image enhancement is to make the enhanced images serve better for downstream tasks, we propose a bi-level learning scheme to optimize the UMIE task and downstream tasks cooperatively, helping generate HQ images both visually appealing and favorable for downstream tasks. Experiments on three medical datasets verify that our method outperforms existing techniques in terms of both enhancement quality and downstream task performance. The code and the newly collected datasets are publicly available at https://github.com/ChunmingHe/HQG-Net.
Collapse
|
8
|
Kim S, Chung H, Park SH, Chung ES, Yi K, Ye JC. Fundus Image Enhancement Through Direct Diffusion Bridges. IEEE J Biomed Health Inform 2024; 28:7275-7286. [PMID: 39167517 DOI: 10.1109/jbhi.2024.3446866] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/23/2024]
Abstract
We propose FD3, a fundus image enhancement method based on direct diffusion bridges, which can cope with a wide range of complex degradations, including haze, blur, noise, and shadow. We first propose a synthetic forward model through a human feedback loop with board-certified ophthalmologists for maximal quality improvement of low-quality in-vivo images. Using the proposed forward model, we train a robust and flexible diffusion-based image enhancement network that is highly effective as a stand-alone method, unlike previous diffusion model-based approaches which act only as a refiner on top of pre-trained models. Through extensive experiments, we show that FD3 establishes superior quality not only on synthetic degradations but also on in vivo studies with low-quality fundus photos taken from patients with cataracts or small pupils.
Collapse
|
9
|
Almarri B, Naveen Kumar B, Aditya Pai H, Bhatia Khan S, Asiri F, Mahesh TR. Redefining retinal vessel segmentation: empowering advanced fundus image analysis with the potential of GANs. Front Med (Lausanne) 2024; 11:1470941. [PMID: 39497847 PMCID: PMC11532151 DOI: 10.3389/fmed.2024.1470941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2024] [Accepted: 09/13/2024] [Indexed: 11/07/2024] Open
Abstract
Retinal vessel segmentation is a critical task in fundus image analysis, providing essential insights for diagnosing various retinal diseases. In recent years, deep learning (DL) techniques, particularly Generative Adversarial Networks (GANs), have garnered significant attention for their potential to enhance medical image analysis. This paper presents a novel approach for retinal vessel segmentation by harnessing the capabilities of GANs. Our method, termed GANVesselNet, employs a specialized GAN architecture tailored to the intricacies of retinal vessel structures. In GANVesselNet, a dual-path network architecture is employed, featuring an Auto Encoder-Decoder (AED) pathway and a UNet-inspired pathway. This unique combination enables the network to efficiently capture multi-scale contextual information, improving the accuracy of vessel segmentation. Through extensive experimentation on publicly available retinal datasets, including STARE and DRIVE, GANVesselNet demonstrates remarkable performance compared to traditional methods and state-of-the-art deep learning approaches. The proposed GANVesselNet exhibits superior sensitivity (0.8174), specificity (0.9862), and accuracy (0.9827) in segmenting retinal vessels on the STARE dataset, and achieves commendable results on the DRIVE dataset with sensitivity (0.7834), specificity (0.9846), and accuracy (0.9709). Notably, GANVesselNet achieves remarkable performance on previously unseen data, underscoring its potential for real-world clinical applications. Furthermore, we present qualitative visualizations of the generated vessel segmentations, illustrating the network's proficiency in accurately delineating retinal vessels. In summary, this paper introduces GANVesselNet, a novel and powerful approach for retinal vessel segmentation. By capitalizing on the advanced capabilities of GANs and incorporating a tailored network architecture, GANVesselNet offers a quantum leap in retinal vessel segmentation accuracy, opening new avenues for enhanced fundus image analysis and improved clinical decision-making.
Collapse
Affiliation(s)
- Badar Almarri
- Department of Computer Science, College of Computer Sciences and Information Technology, King Faisal University, Alhasa, Saudi Arabia
| | - Baskaran Naveen Kumar
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-be University), Bengaluru, India
| | - Haradi Aditya Pai
- Department of Computer Science and Engineering, MIT School of Computing, MIT Art, Design and Technology University, Pune, India
| | - Surbhi Bhatia Khan
- School of Science, Engineering and Environment, University of Salford, Manchester, United Kingdom
- Adjunct Research Faculty at the Centre for Research Impact & Outcome, Chitkara University, Punjab, India
| | - Fatima Asiri
- College of Computer Science, Informatics and Computer Systems Department, King Khalid University, Abha, Saudi Arabia
| | - Thyluru Ramakrishna Mahesh
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-be University), Bengaluru, India
| |
Collapse
|
10
|
Tahir AM, Guo L, Ward RK, Yu X, Rideout A, Hore M, Wang ZJ. Explainable machine learning for assessing upper respiratory tract of racehorses from endoscopy videos. Comput Biol Med 2024; 181:109030. [PMID: 39173488 DOI: 10.1016/j.compbiomed.2024.109030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2023] [Revised: 06/20/2024] [Accepted: 08/13/2024] [Indexed: 08/24/2024]
Abstract
Laryngeal hemiplegia (LH) is a major upper respiratory tract (URT) complication in racehorses. Endoscopy imaging of horse throat is a gold standard for URT assessment. However, current manual assessment faces several challenges, stemming from the poor quality of endoscopy videos and subjectivity of manual grading. To overcome such limitations, we propose an explainable machine learning (ML)-based solution for efficient URT assessment. Specifically, a cascaded YOLOv8 architecture is utilized to segment the key semantic regions and landmarks per frame. Several spatiotemporal features are then extracted from key landmarks points and fed to a decision tree (DT) model to classify LH as Grade 1,2,3 or 4 denoting absence of LH, mild, moderate, and severe LH, respectively. The proposed method, validated through 5-fold cross-validation on 107 videos, showed promising performance in classifying different LH grades with 100%, 91.18%, 94.74% and 100% sensitivity values for Grade 1 to 4, respectively. Further validation on an external dataset of 72 cases confirmed its generalization capability with 90%, 80.95%, 100%, and 100% sensitivity values for Grade 1 to 4, respectively. We introduced several explainability related assessment functions, including: (i) visualization of YOLOv8 output to detect landmark estimation errors which can affect the final classification, (ii) time-series visualization to assess video quality, and (iii) backtracking of the DT output to identify borderline cases. We incorporated domain knowledge (e.g., veterinarian diagnostic procedures) into the proposed ML framework. This provides an assistive tool with clinical-relevance and explainability that can ease and speed up the URT assessment by veterinarians.
Collapse
Affiliation(s)
- Anas Mohammed Tahir
- Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada.
| | - Li Guo
- Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada.
| | - Rabab K Ward
- Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada.
| | - Xinhui Yu
- Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada.
| | - Andrew Rideout
- Point To Point Research & Development, Vancouver, BC, Canada.
| | - Michael Hore
- Hagyard Equine Medical Institute, Lexington, KY, USA.
| | - Z Jane Wang
- Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada.
| |
Collapse
|
11
|
Zhong Y, Liu Z, Zhang X, Liang Z, Chen W, Dai C, Qi L. Unsupervised adversarial neural network for enhancing vasculature in photoacoustic tomography images using optical coherence tomography angiography. Comput Med Imaging Graph 2024; 117:102425. [PMID: 39216343 DOI: 10.1016/j.compmedimag.2024.102425] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2024] [Revised: 08/23/2024] [Accepted: 08/23/2024] [Indexed: 09/04/2024]
Abstract
Photoacoustic tomography (PAT) is a powerful imaging modality for visualizing tissue physiology and exogenous contrast agents. However, PAT faces challenges in visualizing deep-seated vascular structures due to light scattering, absorption, and reduced signal intensity with depth. Optical coherence tomography angiography (OCTA) offers high-contrast visualization of vasculature networks, yet its imaging depth is limited to a millimeter scale. Herein, we propose OCPA-Net, a novel unsupervised deep learning method that utilizes the rich vascular feature of OCTA to enhance PAT images. Trained on unpaired OCTA and PAT images, OCPA-Net incorporates a vessel-aware attention module to enhance deep-seated vessel details captured from OCTA. It leverages a domain-adversarial loss function to enforce structural consistency and a novel identity invariant loss to mitigate excessive image content generation. We validate the structural fidelity of OCPA-Net on simulation experiments, and then demonstrate its vascular enhancement performance on in vivo imaging experiments of tumor-bearing mice and contrast-enhanced pregnant mice. The results show the promise of our method for comprehensive vessel-related image analysis in preclinical research applications.
Collapse
Affiliation(s)
- Yutian Zhong
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Zhenyang Liu
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China; Department of Radiotherapy, The Second Hospital of Nanjing, Nanjing University of Chinese Medicine, Nanjing, 210003, China
| | - Xiaoming Zhang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Zhaoyong Liang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Wufan Chen
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Cuixia Dai
- College of Science, Shanghai Institute of Technology, Shanghai, 201418, China
| | - Li Qi
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| |
Collapse
|
12
|
Guo R, Xu Y, Tompkins A, Pagnucco M, Song Y. Multi-degradation-adaptation network for fundus image enhancement with degradation representation learning. Med Image Anal 2024; 97:103273. [PMID: 39029157 DOI: 10.1016/j.media.2024.103273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2024] [Revised: 05/16/2024] [Accepted: 07/09/2024] [Indexed: 07/21/2024]
Abstract
Fundus image quality serves a crucial asset for medical diagnosis and applications. However, such images often suffer degradation during image acquisition where multiple types of degradation can occur in each image. Although recent deep learning based methods have shown promising results in image enhancement, they tend to focus on restoring one aspect of degradation and lack generalisability to multiple modes of degradation. We propose an adaptive image enhancement network that can simultaneously handle a mixture of different degradations. The main contribution of this work is to introduce our Multi-Degradation-Adaptive module which dynamically generates filters for different types of degradation. Moreover, we explore degradation representation learning and propose the degradation representation network and Multi-Degradation-Adaptive discriminator for our accompanying image enhancement network. Experimental results demonstrate that our method outperforms several existing state-of-the-art methods in fundus image enhancement. Code will be available at https://github.com/RuoyuGuo/MDA-Net.
Collapse
Affiliation(s)
- Ruoyu Guo
- School of Computer Science and Engineering, University of New South Wales, Australia
| | - Yiwen Xu
- School of Computer Science and Engineering, University of New South Wales, Australia
| | - Anthony Tompkins
- School of Computer Science and Engineering, University of New South Wales, Australia
| | - Maurice Pagnucco
- School of Computer Science and Engineering, University of New South Wales, Australia
| | - Yang Song
- School of Computer Science and Engineering, University of New South Wales, Australia.
| |
Collapse
|
13
|
Liu L, Hong J, Wu Y, Liu S, Wang K, Li M, Zhao L, Liu Z, Li L, Cui T, Tsui CK, Xu F, Hu W, Yun D, Chen X, Shang Y, Bi S, Wei X, Lai Y, Lin D, Fu Z, Deng Y, Cai K, Xie Y, Cao Z, Wang D, Zhang X, Dongye M, Lin H, Wu X. Digital ray: enhancing cataractous fundus images using style transfer generative adversarial networks to improve retinopathy detection. Br J Ophthalmol 2024; 108:1423-1429. [PMID: 38839251 PMCID: PMC11503040 DOI: 10.1136/bjo-2024-325403] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2024] [Accepted: 05/15/2024] [Indexed: 06/07/2024]
Abstract
BACKGROUND/AIMS The aim of this study was to develop and evaluate digital ray, based on preoperative and postoperative image pairs using style transfer generative adversarial networks (GANs), to enhance cataractous fundus images for improved retinopathy detection. METHODS For eligible cataract patients, preoperative and postoperative colour fundus photographs (CFP) and ultra-wide field (UWF) images were captured. Then, both the original CycleGAN and a modified CycleGAN (C2ycleGAN) framework were adopted for image generation and quantitatively compared using Frechet Inception Distance (FID) and Kernel Inception Distance (KID). Additionally, CFP and UWF images from another cataract cohort were used to test model performances. Different panels of ophthalmologists evaluated the quality, authenticity and diagnostic efficacy of the generated images. RESULTS A total of 959 CFP and 1009 UWF image pairs were included in model development. FID and KID indicated that images generated by C2ycleGAN presented significantly improved quality. Based on ophthalmologists' average ratings, the percentages of inadequate-quality images decreased from 32% to 18.8% for CFP, and from 18.7% to 14.7% for UWF. Only 24.8% and 13.8% of generated CFP and UWF images could be recognised as synthetic. The accuracy of retinopathy detection significantly increased from 78% to 91% for CFP and from 91% to 93% for UWF. For retinopathy subtype diagnosis, the accuracies also increased from 87%-94% to 91%-100% for CFP and from 87%-95% to 93%-97% for UWF. CONCLUSION Digital ray could generate realistic postoperative CFP and UWF images with enhanced quality and accuracy for overall detection and subtype diagnosis of retinopathies, especially for CFP.\ TRIAL REGISTRATION NUMBER: This study was registered with ClinicalTrials.gov (NCT05491798).
Collapse
Affiliation(s)
- Lixue Liu
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Jiaming Hong
- School of Medical Information Engineering, Guangzhou University of Chinese Medicine, Guangzhou, China
| | - Yuxuan Wu
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Shaopeng Liu
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Kai Wang
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Mingyuan Li
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Lanqin Zhao
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Zhenzhen Liu
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Longhui Li
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Tingxin Cui
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Ching-Kit Tsui
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Fabao Xu
- Qilu Hospital of Shandong University, Jinan, Shandong, China
| | - Weiling Hu
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Dongyuan Yun
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Xi Chen
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Yuanjun Shang
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Shaowei Bi
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Xiaoyue Wei
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Yunxi Lai
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Duoru Lin
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Zhe Fu
- Sun Yat-sen University Zhongshan School of Medicine, Guangzhou, Guangdong, China
| | - Yaru Deng
- Sun Yat-sen University Zhongshan School of Medicine, Guangzhou, Guangdong, China
| | - Kaimin Cai
- Sun Yat-sen University Zhongshan School of Medicine, Guangzhou, Guangdong, China
| | - Yi Xie
- Sun Yat-sen University Zhongshan School of Medicine, Guangzhou, Guangdong, China
| | - Zizheng Cao
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Dongni Wang
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Xulin Zhang
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Meimei Dongye
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Haotian Lin
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Xiaohang Wu
- Zhongshan Ophthalmic Center State Key Laboratory of Ophthalmology, Sun Yat-Sen University, Guangzhou, Guangdong, China
| |
Collapse
|
14
|
Song K, Zhu W, Zhang Z, Liu B, Zhang M, Tang T, Liang J, Wu W. Synthetic lumbar MRI can aid in diagnosis and treatment strategies based on self-pix networks. Sci Rep 2024; 14:20382. [PMID: 39223186 PMCID: PMC11368963 DOI: 10.1038/s41598-024-71288-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Accepted: 08/27/2024] [Indexed: 09/04/2024] Open
Abstract
CT and MR tools are commonly used to diagnose lumbar fractures (LF). However, numerous limitations have been found in practice. The aims of this study were to innovate and develop a spinal disease-specific neural network and to evaluate whether synthetic MRI of the LF affected clinical diagnosis and treatment strategies. A total of 675 LF patients who met the inclusion and exclusion criteria were included in the study. For each participant, two mid-sagittal CT and T2-weighted MR images were selected; 1350 pairs of LF images were also included. A new Self-pix based on Pix2pix and Self-Attention was constructed. A total of 1350 pairs of CT and MR images, which were randomly divided into a training group (1147 pairs) and a test group (203 pairs), were fed into Pix2pix and Self-pix. The quantitative evaluation included PSNR and SSIM (PSNR1 and SSIM1: real MR images and Pix2pix-generated MR images; PSNR2 and SSIM2: real MR images and Self-pix-generated MR images). The qualitative evaluation, including accurate diagnosis of acute fractures and accurate selection of treatment strategies based on Self-pix-generated MRI, was performed by three spine surgeons. In the LF group, PSNR1 and PSNR2 were 10.884 and 11.021 (p < 0.001), and SSIM1 and SSIM2 were 0.766 and 0.771 (p < 0.001), respectively. In the ROI group, PSNR1 and PSNR2 were 12.350 and 12.670 (p = 0.004), and SSIM1 and SSIM2 were 0.816 and 0.832 (p = 0.005), respectively. According to the qualitative evaluation, Self-pix-generated MRI showed no significant difference from real MRI in identifying acute fractures (p = 0.689), with a good sensitivity of 84.36% and specificity of 96.65%. No difference in treatment strategy was found between the Self-pix-generated MRI group and the real MRI group (p = 0.135). In this study, a disease-specific GAN named Self-pix was developed, which demonstrated better image generation performance compared to traditional GAN. The spine surgeon could accurately diagnose LF and select treatment strategies based on Self-pix-generated T2 MR images.
Collapse
Affiliation(s)
- Ke Song
- The First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China
- Yichang Central People's Hospital, Yichang, 443000, China
| | - Wendong Zhu
- College of Computer and Information Technology, China Three Gorges University, Yichang, 430002, China
| | - Zhenxi Zhang
- School of Biomedical Engineering, Sun Yat-Sen University, Shenzhen, 518107, China
| | - Bin Liu
- Wendeng Orthopaedic and Traumatologic Hospital of Shandong Province, Weihai, 264400, China
| | - Meiling Zhang
- The First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China
- Yichang Central People's Hospital, Yichang, 443000, China
| | - Tinglong Tang
- College of Computer and Information Technology, China Three Gorges University, Yichang, 430002, China
| | - Jie Liang
- The First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China
- Yichang Central People's Hospital, Yichang, 443000, China
| | - Weifei Wu
- The First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China.
- Yichang Central People's Hospital, Yichang, 443000, China.
| |
Collapse
|
15
|
Bardozzo F, Fiore P, Valentino M, Bianco V, Memmolo P, Miccio L, Brancato V, Smaldone G, Gambacorta M, Salvatore M, Ferraro P, Tagliaferri R. Enhanced tissue slide imaging in the complex domain via cross-explainable GAN for Fourier ptychographic microscopy. Comput Biol Med 2024; 179:108861. [PMID: 39018884 DOI: 10.1016/j.compbiomed.2024.108861] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Revised: 05/28/2024] [Accepted: 07/06/2024] [Indexed: 07/19/2024]
Abstract
Achieving microscopy with large space-bandwidth products plays a key role in diagnostic imaging and is widely significant in the overall field of clinical practice. Among quantitative microscopy techniques, Fourier Ptychography (FP) provides a wide field of view and high-resolution images, suitable to the histopathological field, but onerous in computational terms. Artificial intelligence can be a solution in this sense. In particular, this research delves into the application of Generative Adversarial Networks (GAN) for the dual-channel complex FP image enhancement of human kidney samples. The study underscores the GANs' efficacy in promoting biological architectures in FP domain, thereby still guaranteeing high resolution and visibility of detailed microscopic structures. We demonstrate successful GAN-based enhanced reconstruction through two strategies: cross-explainability and expert survey. The cross-explainability is evaluated through the comparison of explanation maps for both real and imaginary components underlining its robustness. This comparison further shows that their interplay is pivotal for accurate reconstruction without hallucinations. Secondly, the enhanced reconstruction accuracy and effectiveness in a clinical workflow are confirmed through a two-step survey conducted with nephrologists.
Collapse
Affiliation(s)
- Francesco Bardozzo
- NeuroneLab - Department of Management and Innovation Systems (DISA-MIS), University of Salerno, Via Giovanni Paolo II, 132, Fisciano (SA), 84084, Italy; CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy
| | - Pierpaolo Fiore
- NeuroneLab - Department of Management and Innovation Systems (DISA-MIS), University of Salerno, Via Giovanni Paolo II, 132, Fisciano (SA), 84084, Italy
| | - Marika Valentino
- CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy; DIETI, Department of Electrical Engineering and Information Technologies, University of Naples "Federico II", via Claudio 21, Napoli, 80125, Italy
| | - Vittorio Bianco
- CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy.
| | - Pasquale Memmolo
- CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy
| | - Lisa Miccio
- CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy
| | | | | | | | - Marco Salvatore
- IRCCS SYNLAB SDN, Via E. Gianturco 113, Napoli, 80143, Italy
| | - Pietro Ferraro
- CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy.
| | - Roberto Tagliaferri
- NeuroneLab - Department of Management and Innovation Systems (DISA-MIS), University of Salerno, Via Giovanni Paolo II, 132, Fisciano (SA), 84084, Italy; CNR-ISASI, Institute of Applied Sciences and Intelligent Systems "E. Caianiello", Via Campi Flegrei 34, Pozzuoli (NA), 80078, Italy
| |
Collapse
|
16
|
Ye Z, Luo S, Wang L. Deep Learning Based Cystoscopy Image Enhancement. J Endourol 2024; 38:962-968. [PMID: 38753720 DOI: 10.1089/end.2023.0751] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/18/2024] Open
Abstract
Background: Endoscopy image enhancement technology provides doctors with clearer and more detailed images for observation and diagnosis, allowing doctors to assess lesions more accurately. Unlike most other endoscopy images, cystoscopy images face more complex and diverse image degradation because of their underwater imaging characteristics. Among the various causes of image degradation, the blood haze resulting from bladder mucosal bleeding make the background blurry and unclear, severely affecting diagnostic efficiency, even leading to misjudgment. Materials and Methods: We propose a deep learning-based approach to mitigate the impact of blood haze on cystoscopy images. The approach consists of two parts as follows: a blood haze removal network and a contrast enhancement algorithm. First, we adopt Feature Fusion Attention Network (FFA-Net) and transfer learning in the field of deep learning to remove blood haze from cystoscopy images and introduce perceptual loss to constrain the network for better visual results. Second, we enhance the image contrast by remapping the gray scale of the blood haze-free image and performing weighted fusion of the processed image and the original image. Results: In the blood haze removal stage, the algorithm proposed in this article achieves an average peak signal-to-noise ratio of 29.44 decibels, which is 15% higher than state-of-the-art traditional methods. The average structural similarity and perceptual image patch similarity reach 0.9269 and 0.1146, respectively, both superior to state-of-the-art traditional methods. Besides, our method is the best in keeping color balance after removing the blood haze. In the image enhancement stage, our algorithm enhances the contrast of vessels and tissues while preserving the original colors, expanding the dynamic range of the image. Conclusion: The deep learning-based cystoscopy image enhancement method is significantly better than other traditional methods in both qualitative and quantitative evaluation. The application of artificial intelligence will provide clearer, higher contrast cystoscopy images for medical diagnosis.
Collapse
Affiliation(s)
- Zixing Ye
- Department of Urology, Peking Union Medical College Hospital, Beijing, China
| | - Shun Luo
- School of Software, Northwestern Polytechnical University, Xi' an, China
| | - Lianpo Wang
- School of Software, Northwestern Polytechnical University, Xi' an, China
- Research & Development Institute of Northwestern, Polytechnical University in Shenzhen, Shenzhen, China
| |
Collapse
|
17
|
Wasilewski T, Kamysz W, Gębicki J. AI-Assisted Detection of Biomarkers by Sensors and Biosensors for Early Diagnosis and Monitoring. BIOSENSORS 2024; 14:356. [PMID: 39056632 PMCID: PMC11274923 DOI: 10.3390/bios14070356] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/09/2024] [Revised: 06/25/2024] [Accepted: 06/28/2024] [Indexed: 07/28/2024]
Abstract
The steady progress in consumer electronics, together with improvement in microflow techniques, nanotechnology, and data processing, has led to implementation of cost-effective, user-friendly portable devices, which play the role of not only gadgets but also diagnostic tools. Moreover, numerous smart devices monitor patients' health, and some of them are applied in point-of-care (PoC) tests as a reliable source of evaluation of a patient's condition. Current diagnostic practices are still based on laboratory tests, preceded by the collection of biological samples, which are then tested in clinical conditions by trained personnel with specialistic equipment. In practice, collecting passive/active physiological and behavioral data from patients in real time and feeding them to artificial intelligence (AI) models can significantly improve the decision process regarding diagnosis and treatment procedures via the omission of conventional sampling and diagnostic procedures while also excluding the role of pathologists. A combination of conventional and novel methods of digital and traditional biomarker detection with portable, autonomous, and miniaturized devices can revolutionize medical diagnostics in the coming years. This article focuses on a comparison of traditional clinical practices with modern diagnostic techniques based on AI and machine learning (ML). The presented technologies will bypass laboratories and start being commercialized, which should lead to improvement or substitution of current diagnostic tools. Their application in PoC settings or as a consumer technology accessible to every patient appears to be a real possibility. Research in this field is expected to intensify in the coming years. Technological advancements in sensors and biosensors are anticipated to enable the continuous real-time analysis of various omics fields, fostering early disease detection and intervention strategies. The integration of AI with digital health platforms would enable predictive analysis and personalized healthcare, emphasizing the importance of interdisciplinary collaboration in related scientific fields.
Collapse
Affiliation(s)
- Tomasz Wasilewski
- Department of Inorganic Chemistry, Faculty of Pharmacy, Medical University of Gdansk, Hallera 107, 80-416 Gdansk, Poland
| | - Wojciech Kamysz
- Department of Inorganic Chemistry, Faculty of Pharmacy, Medical University of Gdansk, Hallera 107, 80-416 Gdansk, Poland
| | - Jacek Gębicki
- Department of Process Engineering and Chemical Technology, Faculty of Chemistry, Gdansk University of Technology, Narutowicza 11/12, 80-233 Gdansk, Poland;
| |
Collapse
|
18
|
Zhang S, Webers CAB, Berendschot TTJM. Computational single fundus image restoration techniques: a review. FRONTIERS IN OPHTHALMOLOGY 2024; 4:1332197. [PMID: 38984141 PMCID: PMC11199880 DOI: 10.3389/fopht.2024.1332197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Accepted: 04/19/2024] [Indexed: 07/11/2024]
Abstract
Fundus cameras are widely used by ophthalmologists for monitoring and diagnosing retinal pathologies. Unfortunately, no optical system is perfect, and the visibility of retinal images can be greatly degraded due to the presence of problematic illumination, intraocular scattering, or blurriness caused by sudden movements. To improve image quality, different retinal image restoration/enhancement techniques have been developed, which play an important role in improving the performance of various clinical and computer-assisted applications. This paper gives a comprehensive review of these restoration/enhancement techniques, discusses their underlying mathematical models, and shows how they may be effectively applied in real-life practice to increase the visual quality of retinal images for potential clinical applications including diagnosis and retinal structure recognition. All three main topics of retinal image restoration/enhancement techniques, i.e., illumination correction, dehazing, and deblurring, are addressed. Finally, some considerations about challenges and the future scope of retinal image restoration/enhancement techniques will be discussed.
Collapse
Affiliation(s)
- Shuhe Zhang
- University Eye Clinic Maastricht, Maastricht University Medical Center, Maastricht, Netherlands
| | - Carroll A B Webers
- University Eye Clinic Maastricht, Maastricht University Medical Center, Maastricht, Netherlands
| | - Tos T J M Berendschot
- University Eye Clinic Maastricht, Maastricht University Medical Center, Maastricht, Netherlands
| |
Collapse
|
19
|
Li H, Lin Z, Qiu Z, Li Z, Niu K, Guo N, Fu H, Hu Y, Liu J. Enhancing and Adapting in the Clinic: Source-Free Unsupervised Domain Adaptation for Medical Image Enhancement. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1323-1336. [PMID: 38015687 DOI: 10.1109/tmi.2023.3335651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/30/2023]
Abstract
Medical imaging provides many valuable clues involving anatomical structure and pathological characteristics. However, image degradation is a common issue in clinical practice, which can adversely impact the observation and diagnosis by physicians and algorithms. Although extensive enhancement models have been developed, these models require a well pre-training before deployment, while failing to take advantage of the potential value of inference data after deployment. In this paper, we raise an algorithm for source-free unsupervised domain adaptive medical image enhancement (SAME), which adapts and optimizes enhancement models using test data in the inference phase. A structure-preserving enhancement network is first constructed to learn a robust source model from synthesized training data. Then a teacher-student model is initialized with the source model and conducts source-free unsupervised domain adaptation (SFUDA) by knowledge distillation with the test data. Additionally, a pseudo-label picker is developed to boost the knowledge distillation of enhancement tasks. Experiments were implemented on ten datasets from three medical image modalities to validate the advantage of the proposed algorithm, and setting analysis and ablation studies were also carried out to interpret the effectiveness of SAME. The remarkable enhancement performance and benefits for downstream tasks demonstrate the potential and generalizability of SAME. The code is available at https://github.com/liamheng/Annotation-free-Medical-Image-Enhancement.
Collapse
|
20
|
Chen J, Lu R, Ye S, Guang M, Tassew TM, Jing B, Zhang G, Chen G, Shen D. Image Recovery Matters: A Recovery-Extraction Framework for Robust Fetal Brain Extraction From MR Images. IEEE J Biomed Health Inform 2024; 28:823-834. [PMID: 37995170 DOI: 10.1109/jbhi.2023.3333953] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2023]
Abstract
The extraction of the fetal brain from magnetic resonance (MR) images is a challenging task. In particular, fetal MR images suffer from different kinds of artifacts introduced during the image acquisition. Among those artifacts, intensity inhomogeneity is a common one affecting brain extraction. In this work, we propose a deep learning-based recovery-extraction framework for fetal brain extraction, which is particularly effective in handling fetal MR images with intensity inhomogeneity. Our framework involves two stages. First, the artifact-corrupted images are recovered with the proposed generative adversarial learning-based image recovery network with a novel region-of-darkness discriminator that enforces the network focusing on artifacts of the images. Second, we propose a brain extraction network for more effective fetal brain segmentation by strengthening the association between lower- and higher-level features as well as suppressing task-irrelevant features. Thanks to the proposed recovery-extraction strategy, our framework is able to accurately segment fetal brains from artifact-corrupted MR images. The experiments show that our framework achieves promising performance in both quantitative and qualitative evaluations, and outperforms state-of-the-art methods in both image recovery and fetal brain extraction.
Collapse
|
21
|
Guo S, Wang H, Agaian S, Han L, Song X. LRENet: a location-related enhancement network for liver lesions in CT images. Phys Med Biol 2024; 69:035019. [PMID: 38211307 DOI: 10.1088/1361-6560/ad1d6b] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2023] [Accepted: 01/11/2024] [Indexed: 01/13/2024]
Abstract
Objective. Liver cancer is a major global health problem expected to increase by more than 55% by 2040. Accurate segmentation of liver tumors from computed tomography (CT) images is essential for diagnosis and treatment planning. However, this task is challenging due to the variations in liver size, the low contrast between tumor and normal tissue, and the noise in the images. APPROACH In this study, we propose a novel method called location-related enhancement network (LRENet) which can enhance the contrast of liver lesions in CT images and facilitate their segmentation. LRENet consists of two steps: (1) locating the lesions and the surrounding tissues using a morphological approach and (2) enhancing the lesions and smoothing the other regions using a new loss function. MAIN RESULTS We evaluated LRENet on two public datasets (LiTS and 3Dircadb01) and one dataset collected from a collaborative hospital (Liver cancer dateset), and compared it with state-of-the-art methods regarding several metrics. The results of the experiments showed that our proposed method outperformed the compared methods on three datasets in several metrics. We also trained the Swin-Transformer network on the enhanced datasets and showed that our method could improve the segmentation performance of both liver and lesions. SIGNIFICANCE Our method has potential applications in clinical diagnosis and treatment planning, as it can provide more reliable and informative CT images of liver tumors.
Collapse
Affiliation(s)
- Shuli Guo
- State Key Laboratory of Intelligent Control and Decision of Complex Systems, School of Automation, Beijing Institute of Technology, Beijing, 100081, People's Republic of China
| | - Hui Wang
- State Key Laboratory of Intelligent Control and Decision of Complex Systems, School of Automation, Beijing Institute of Technology, Beijing, 100081, People's Republic of China
| | - Sos Agaian
- Computer Science Departments, College of Staten Island and the Graduate Center, City University of New York, 2800 Victory Boulevard, Staten Island, NY,10314, United States of America
| | - Lina Han
- Department of Cardiology, The Second Medical Center, National Clinical Research Center for Geriatric Diseases, Chinese PLA General Hospital, Beijing, 100853, People's Republic of China
| | - Xiaowei Song
- State Key Laboratory of Intelligent Control and Decision of Complex Systems, School of Automation, Beijing Institute of Technology, Beijing, 100081, People's Republic of China
| |
Collapse
|
22
|
Zhang J, Huang X, Liu Y, Han Y, Xiang Z. GAN-based medical image small region forgery detection via a two-stage cascade framework. PLoS One 2024; 19:e0290303. [PMID: 38166011 PMCID: PMC10760893 DOI: 10.1371/journal.pone.0290303] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Accepted: 08/06/2023] [Indexed: 01/04/2024] Open
Abstract
Using generative adversarial network (GAN) Goodfellow et al. (2014) for data enhancement of medical images is significantly helpful for many computer-aided diagnosis (CAD) tasks. A new GAN-based automated tampering attack, like CT-GAN Mirsky et al. (2019), has emerged. It can inject or remove lung cancer lesions to CT scans. Because the tampering region may even account for less than 1% of the original image, even state-of-the-art methods are challenging to detect the traces of such tampering. This paper proposes a two-stage cascade framework to detect GAN-based medical image small region forgery like CT-GAN. In the local detection stage, we train the detector network with small sub-images so that interference information in authentic regions will not affect the detector. We use depthwise separable convolution and residual networks to prevent the detector from over-fitting and enhance the ability to find forged regions through the attention mechanism. The detection results of all sub-images in the same image will be combined into a heatmap. In the global classification stage, using gray-level co-occurrence matrix (GLCM) can better extract features of the heatmap. Because the shape and size of the tampered region are uncertain, we use hyperplanes in an infinite-dimensional space for classification. Our method can classify whether a CT image has been tampered and locate the tampered position. Sufficient experiments show that our method can achieve excellent performance than the state-of-the-art detection methods.
Collapse
Affiliation(s)
- Jianyi Zhang
- Beijing Electronic Science and Technology Institute, Beijing, China
- University of Louisiana at Lafayette, Lafayette, Louisiana, United States of America
| | - Xuanxi Huang
- Beijing Electronic Science and Technology Institute, Beijing, China
| | - Yaqi Liu
- Beijing Electronic Science and Technology Institute, Beijing, China
| | - Yuyang Han
- Beijing Electronic Science and Technology Institute, Beijing, China
| | - Zixiao Xiang
- Beijing Electronic Science and Technology Institute, Beijing, China
| |
Collapse
|
23
|
Li H, Liu H, Fu H, Xu Y, Shu H, Niu K, Hu Y, Liu J. A generic fundus image enhancement network boosted by frequency self-supervised representation learning. Med Image Anal 2023; 90:102945. [PMID: 37703674 DOI: 10.1016/j.media.2023.102945] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2022] [Revised: 06/12/2023] [Accepted: 08/29/2023] [Indexed: 09/15/2023]
Abstract
Fundus photography is prone to suffer from image quality degradation that impacts clinical examination performed by ophthalmologists or intelligent systems. Though enhancement algorithms have been developed to promote fundus observation on degraded images, high data demands and limited applicability hinder their clinical deployment. To circumvent this bottleneck, a generic fundus image enhancement network (GFE-Net) is developed in this study to robustly correct unknown fundus images without supervised or extra data. Levering image frequency information, self-supervised representation learning is conducted to learn robust structure-aware representations from degraded images. Then with a seamless architecture that couples representation learning and image enhancement, GFE-Net can accurately correct fundus images and meanwhile preserve retinal structures. Comprehensive experiments are implemented to demonstrate the effectiveness and advantages of GFE-Net. Compared with state-of-the-art algorithms, GFE-Net achieves superior performance in data dependency, enhancement performance, deployment efficiency, and scale generalizability. Follow-up fundus image analysis is also facilitated by GFE-Net, whose modules are respectively verified to be effective for image enhancement.
Collapse
Affiliation(s)
- Heng Li
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China.
| | - Haofeng Liu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China.
| | - Huazhu Fu
- Institute of High Performance Computing (IHPC), Agency for Science, Technology and Research (A*STAR), Singapore
| | - Yanwu Xu
- School of Future Technology, South China University of Technology, Guangzhou, China; Pazhou Lab, Guangzhou, China
| | - Hai Shu
- Department of Biostatistics, School of Global Public Health, New York University, NY, USA
| | - Ke Niu
- Computer School, Beijing Information Science and Technology University, Beijing, China
| | - Yan Hu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China.
| | - Jiang Liu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China; Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China; Guangdong Provincial Key Laboratory of Brain-inspired Intelligent Computation, Southern University of Science and Technology, Shenzhen, China.
| |
Collapse
|
24
|
Guo Y, Hu M, Min X, Wang Y, Dai M, Zhai G, Zhang XP, Yang X. Blind Image Quality Assessment for Pathological Microscopic Image Under Screen and Immersion Scenarios. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3295-3306. [PMID: 37267133 DOI: 10.1109/tmi.2023.3282387] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
The high-quality pathological microscopic images are essential for physicians or pathologists to make a correct diagnosis. Image quality assessment (IQA) can quantify the visual distortion degree of images and guide the imaging system to improve image quality, thus raising the quality of pathological microscopic images. Current IQA methods are not ideal for pathological microscopy images due to their specificity. In this paper, we present deep learning-based blind image quality assessment model with saliency block and patch block for pathological microscopic images. The saliency block and patch block can handle the local and global distortions, respectively. To better capture the area of interest of pathologists when viewing pathological images, the saliency block is fine-tuned by eye movement data of pathologists. The patch block can capture lots of global information strongly related to image quality via the interaction between different image patches from different positions. The performance of the developed model is validated by the home-made Pathological Microscopic Image Quality Database under Screen and Immersion Scenarios (PMIQD-SIS) and cross-validated by the five public datasets. The results of ablation experiments demonstrate the contribution of the added blocks. The dataset and the corresponding code are publicly available at: https://github.com/mikugyf/PMIQD-SIS.
Collapse
|
25
|
Huang ZH, Liu YY, Wu WJ, Huang KW. Design and Validation of a Deep Learning Model for Renal Stone Detection and Segmentation on Kidney-Ureter-Bladder Images. Bioengineering (Basel) 2023; 10:970. [PMID: 37627855 PMCID: PMC10452034 DOI: 10.3390/bioengineering10080970] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 08/11/2023] [Accepted: 08/13/2023] [Indexed: 08/27/2023] Open
Abstract
Kidney-ureter-bladder (KUB) imaging is used as a frontline investigation for patients with suspected renal stones. In this study, we designed a computer-aided diagnostic system for KUB imaging to assist clinicians in accurately diagnosing urinary tract stones. The image dataset used for training and testing the model comprised 485 images provided by Kaohsiung Chang Gung Memorial Hospital. The proposed system was divided into two subsystems, 1 and 2. Subsystem 1 used Inception-ResNetV2 to train a deep learning model on preprocessed KUB images to verify the improvement in diagnostic accuracy with image preprocessing. Subsystem 2 trained an image segmentation model using the ResNet hybrid, U-net, to accurately identify the contours of renal stones. The performance was evaluated using a confusion matrix for the classification model. We conclude that the model can assist clinicians in accurately diagnosing renal stones via KUB imaging. Therefore, the proposed system can assist doctors in diagnosis, reduce patients' waiting time for CT scans, and minimize the radiation dose absorbed by the body.
Collapse
Affiliation(s)
- Zih-Hao Huang
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 807618, Taiwan; (Z.-H.H.); (Y.-Y.L.); (W.-J.W.)
| | - Yi-Yang Liu
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 807618, Taiwan; (Z.-H.H.); (Y.-Y.L.); (W.-J.W.)
- Department of Urology, Kaohsiung Chang Gung Memorial Hospital and Chang Gung University College of Medicine, Kaohsiung City 83301, Taiwan
| | - Wei-Juei Wu
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 807618, Taiwan; (Z.-H.H.); (Y.-Y.L.); (W.-J.W.)
| | - Ko-Wei Huang
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 807618, Taiwan; (Z.-H.H.); (Y.-Y.L.); (W.-J.W.)
| |
Collapse
|
26
|
Li T, Roberts R, Liu Z, Tong W. TransOrGAN: An Artificial Intelligence Mapping of Rat Transcriptomic Profiles between Organs, Ages, and Sexes. Chem Res Toxicol 2023; 36:916-925. [PMID: 37200521 PMCID: PMC10433534 DOI: 10.1021/acs.chemrestox.3c00037] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Indexed: 05/20/2023]
Abstract
Animal studies are required for the evaluation of candidate drugs to ensure patient and volunteer safety. Toxicogenomics is often applied in these studies to gain understanding of the underlying mechanisms of toxicity, which is usually focused on critical organs such as the liver or kidney in young male rats. There is a strong ethical reason to reduce, refine and replace animal use (the 3Rs), where the mapping of data between organs, sexes and ages could reduce the cost and time of drug development. Herein, we proposed a generative adversarial network (GAN)-based framework entitled TransOrGAN that allowed the molecular mapping of gene expression profiles in different rodent organ systems and across sex and age groups. We carried out a proof-of-concept study based on rat RNA-seq data from 288 samples in 9 different organs of both sexes and 4 developmental stages. First, we demonstrated that TransOrGAN could infer transcriptomic profiles between any 2 of the 9 organs studied, yielding an average cosine similarity of 0.984 between synthetic transcriptomic profiles and their corresponding real profiles. Second, we found that TransOrGAN could infer transcriptomic profiles observed in females from males, with an average cosine similarity of 0.984. Third, we found that TransOrGAN could infer transcriptomic profiles in juvenile, adult, and aged animals from adolescent animals with an average cosine similarity of 0.981, 0.983, and 0.989, respectively. Altogether, TransOrGAN is an innovative approach to infer transcriptomic profiles between ages, sexes, and organ systems, offering the opportunity to reduce animal usage and to provide an integrated assessment of toxicity in the whole organism irrespective of sex or age.
Collapse
Affiliation(s)
- Ting Li
- National
Center for Toxicological Research, Food
and Drug Administration, Jefferson, Arkansas 72079, United States
| | - Ruth Roberts
- ApconiX Ltd, Alderley Park, Alderley Edge SK10 4TG, United Kingdom
- University
of Birmingham, Edgbaston, Birmingham B15 2TT, United
Kingdom
| | - Zhichao Liu
- Integrative
Toxicology, Nonclinical Drug Safety, Boehringer
Ingelheim Pharmaceuticals, Inc., Ridgefield, Connecticut 06877, United States
| | - Weida Tong
- National
Center for Toxicological Research, Food
and Drug Administration, Jefferson, Arkansas 72079, United States
| |
Collapse
|
27
|
Fu L, Lu B, Tian J, Hu Z. PSSGAN: Towards spectrum shift based perceptual quality enhancement for fluorescence imaging. Comput Med Imaging Graph 2023; 107:102216. [DOI: 10.1016/j.compmedimag.2023.102216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Revised: 03/08/2023] [Accepted: 03/09/2023] [Indexed: 03/19/2023]
|
28
|
Zheng Q, Yang R, Ni X, Yang S, Jiang Z, Wang L, Chen Z, Liu X. Development and validation of a deep learning-based laparoscopic system for improving video quality. Int J Comput Assist Radiol Surg 2023; 18:257-268. [PMID: 36243805 DOI: 10.1007/s11548-022-02777-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2022] [Accepted: 10/05/2022] [Indexed: 02/03/2023]
Abstract
PURPOSE A clear surgical field of view is a prerequisite for successful laparoscopic surgery. Surgical smoke, image blur, and lens fogging can affect the clarity of laparoscopic imaging. We aimed to develop a real-time assistance system (namely LVQIS) for removing these interfering factors during laparoscopic surgery, thereby improving laparoscopic video quality. METHODS LVQIS was developed with generative adversarial networks (GAN) and transfer learning, which included two classification models (ResNet-50), a motion blur removal model (MPRNet), and a smoke/fog removal model (GAN). 136 laparoscopic surgery videos were retrospectively collected in a tripartite dataset for training and validation. A synthetic dataset was simulated using the image enhancement library Albumentations and the 3D rendering software Blender. The objective evaluation results were through PSNR, SSIM and FID, and the subjective evaluation includes the operation pause time and the degree of anxiety of surgeons. RESULTS The synthesized dataset contained 19,245 clear images, 19,245 motion blur images, and 19,245 smoke/fog images. The ResNet-50 CNN model identified whether a single laparoscopic image had motion blur and smoke/fog with an accuracy of over 0.99. The PSNR, SSIM and FID of the de-smoke model were 29.67, 0.9551 and 74.72, respectively, and the PSNR, SSIM and FID of the de-blurring model were 26.78, 0.9020 and 80.10, respectively, which were better than other advanced de-blurring and de-smoke/fog models. In a comparative study of 100 laparoscopic surgeries, the use of LVQIS significantly reduced the operation pause time (P < 0.001) and the anxiety of surgeons (P = 0.004). CONCLUSIONS In this study, LVQIS is an efficient and robust system that can improve the quality of laparoscopic video, reduce surgical pause time and the anxiety of surgeons, and has the potential for real-time application in real clinical settings.
Collapse
Affiliation(s)
- Qingyuan Zheng
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China
| | - Rui Yang
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China
| | - Xinmiao Ni
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China
| | - Song Yang
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China
| | - Zhengyu Jiang
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China
| | - Lei Wang
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China
| | - Zhiyuan Chen
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China.
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China.
| | - Xiuheng Liu
- Department of Urology, Renmin Hospital of Wuhan University, 99 Zhang Zhi-dong Road, Wuhan, Hubei, 430060, People's Republic of China.
- Institute of Urologic Disease, Renmin Hospital of Wuhan University, Wuhan, 430060, Hubei, China.
| |
Collapse
|
29
|
PISDGAN: Perceive image structure and details for laryngeal image enhancement. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
30
|
Poonkodi S, Kanchana M. 3D-MedTranCSGAN: 3D Medical Image Transformation using CSGAN. Comput Biol Med 2023; 153:106541. [PMID: 36652868 DOI: 10.1016/j.compbiomed.2023.106541] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2022] [Revised: 11/30/2022] [Accepted: 01/10/2023] [Indexed: 01/15/2023]
Abstract
Computer vision techniques are a rapidly growing area of transforming medical images for various specific medical applications. In an end-to-end application, this paper proposes a 3D Medical Image Transformation Using a CSGAN model named a 3D-MedTranCSGAN. The 3D-MedTranCSGAN model is an integration of non-adversarial loss components and the Cyclic Synthesized Generative Adversarial Networks. The proposed model utilizes PatchGAN's discriminator network, to penalize the difference between the synthesized image and the original image. The model also computes the non-adversary loss functions such as content, perception, and style transfer losses. 3DCascadeNet is a new generator architecture introduced in the paper, which is used to enhance the perceptiveness of the transformed medical image by encoding-decoding pairs. We use the 3D-MedTranCSGAN model to do various tasks without modifying specific applications: PET to CT image transformation; reconstruction of CT to PET; modification of movement artefacts in MR images; and removing noise in PET images. We found that 3D-MedTranCSGAN outperformed other transformation methods in our experiments. For the first task, the proposed model yields SSIM is 0.914, PSNR is 26.12, MSE is 255.5, VIF is 0.4862, UQI is 0.9067 and LPIPs is 0.2284. For the second task, the model yields 0.9197, 25.7, 257.56, 0.4962, 0.9027, 0.2262. For the third task, the model yields 0.8862, 24.94, 0.4071, 0.6410, 0.2196. For the final task, the model yields 0.9521, 33.67, 33.57, 0.6091, 0.9255, 0.0244. Based on the result analysis, the proposed model outperforms the other techniques.
Collapse
Affiliation(s)
- S Poonkodi
- Department of Computing Technologies, School of Computing, SRM Institute of Science and Technology, Kattankulathur, India
| | - M Kanchana
- Department of Computing Technologies, School of Computing, SRM Institute of Science and Technology, Kattankulathur, India.
| |
Collapse
|
31
|
Zhang S, Webers CAB, Berendschot TTJM. Luminosity rectified blind Richardson-Lucy deconvolution for single retinal image restoration. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 229:107297. [PMID: 36563648 DOI: 10.1016/j.cmpb.2022.107297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/11/2022] [Revised: 11/14/2022] [Accepted: 11/29/2022] [Indexed: 06/17/2023]
Abstract
BACKGROUND AND OBJECTIVE Due to imperfect imaging conditions, retinal images can be degraded by uneven/insufficient illumination, blurriness caused by optical aberrations and unintentional motions. Degraded images reduce the effectiveness of diagnosis by an ophthalmologist. To restore the image quality, in this research we propose the luminosity rectified Richardson-Lucy (LRRL) blind deconvolution framework for single retinal image restoration. METHODS We established an image formation model based on the double-pass fundus reflection feature and developed a differentiable non-convex cost function that jointly achieves illumination correction and blind deconvolution. To solve this non-convex optimization problem, we derived the closed-form expression of the gradients and used gradient descent with Nesterov-accelerated adaptive momentum estimation to accelerate the optimization, which is more efficient than the traditional half quadratic splitting method. RESULTS The LRRL was tested on 1719 images from three public databases. Four image quality matrixes including image definition, image sharpness, image entropy, and image multiscale contrast were used for objective assessments. The LRRL was compared against the state-of-the-art retinal image blind deconvolution methods. CONCLUSIONS Our LRRL corrects the problematic illumination and improves the clarity of the retinal image simultaneously, showing its superiority in terms of restoration quality and implementation efficiency. The MATLAB code is available on Github.
Collapse
Affiliation(s)
- Shuhe Zhang
- University Eye Clinic Maastricht, Maastricht University Medical Center +, P.O. Box 5800, Maastricht, AZ 6202, the Netherlands.
| | - Carroll A B Webers
- University Eye Clinic Maastricht, Maastricht University Medical Center +, P.O. Box 5800, Maastricht, AZ 6202, the Netherlands
| | - Tos T J M Berendschot
- University Eye Clinic Maastricht, Maastricht University Medical Center +, P.O. Box 5800, Maastricht, AZ 6202, the Netherlands
| |
Collapse
|
32
|
Cao J, Xu Z, Xu M, Ma Y, Zhao Y. A two-stage framework for optical coherence tomography angiography image quality improvement. Front Med (Lausanne) 2023; 10:1061357. [PMID: 36756179 PMCID: PMC9899819 DOI: 10.3389/fmed.2023.1061357] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Accepted: 01/02/2023] [Indexed: 01/24/2023] Open
Abstract
Introduction Optical Coherence Tomography Angiography (OCTA) is a new non-invasive imaging modality that gains increasing popularity for the observation of the microvasculatures in the retina and the conjunctiva, assisting clinical diagnosis and treatment planning. However, poor imaging quality, such as stripe artifacts and low contrast, is common in the acquired OCTA and in particular Anterior Segment OCTA (AS-OCTA) due to eye microtremor and poor illumination conditions. These issues lead to incomplete vasculature maps that in turn makes it hard to make accurate interpretation and subsequent diagnosis. Methods In this work, we propose a two-stage framework that comprises a de-striping stage and a re-enhancing stage, with aims to remove stripe noise and to enhance blood vessel structure from the background. We introduce a new de-striping objective function in a Stripe Removal Net (SR-Net) to suppress the stripe noise in the original image. The vasculatures in acquired AS-OCTA images usually exhibit poor contrast, so we use a Perceptual Structure Generative Adversarial Network (PS-GAN) to enhance the de-striped AS-OCTA image in the re-enhancing stage, which combined cyclic perceptual loss with structure loss to achieve further image quality improvement. Results and discussion To evaluate the effectiveness of the proposed method, we apply the proposed framework to two synthetic OCTA datasets and a real AS-OCTA dataset. Our results show that the proposed framework yields a promising enhancement performance, which enables both conventional and deep learning-based vessel segmentation methods to produce improved results after enhancement of both retina and AS-OCTA modalities.
Collapse
Affiliation(s)
- Juan Cao
- School of Information Science and Engineering, Chongqing Jiaotong University, Chongqing, China
| | - Zihao Xu
- School of Information Science and Engineering, Chongqing Jiaotong University, Chongqing, China,Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Mengjia Xu
- Affiliated Cixi Hospital, Wenzhou Medical University, Ningbo, China,*Correspondence: Mengjia Xu ✉
| | - Yuhui Ma
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China,Yuhui Ma ✉
| | - Yitian Zhao
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China
| |
Collapse
|
33
|
Han R, Tang C, Xu M, Lei Z. A Retinex-based variational model for noise suppression and nonuniform illumination correction in corneal confocal microscopy images. Phys Med Biol 2023; 68. [PMID: 36577141 DOI: 10.1088/1361-6560/acaeef] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2022] [Accepted: 12/28/2022] [Indexed: 12/29/2022]
Abstract
Objective.Corneal confocal microscopy (CCM) image analysis is a non-invasivein vivoclinical technique that can quantify corneal nerve fiber damage. However, the acquired CCM images are often accompanied by speckle noise and nonuniform illumination, which seriously affects the analysis and diagnosis of the diseases.Approach.In this paper, first we propose a variational Retinex model for the inhomogeneity correction and noise removal of CCM images. In this model, the Beppo Levi space is introduced to constrain the smoothness of the illumination layer for the first time, and the fractional order differential is adopted as the regularization term to constrain reflectance layer. Then, a denoising regularization term is also constructed with Block Matching 3D (BM3D) to suppress noise. Finally, by adjusting the uneven illumination layer, we obtain the final results. Second, an image quality evaluation metric is proposed to evaluate the illumination uniformity of images objectively.Main results.To demonstrate the effectiveness of our method, the proposed method is tested on 628 low-quality CCM images from the CORN-2 dataset. Extensive experiments show the proposed method outperforms the other four related methods in terms of noise removal and uneven illumination suppression.SignificanceThis demonstrates that the proposed method may be helpful for the diagnostics and analysis of eye diseases.
Collapse
Affiliation(s)
- Rui Han
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Chen Tang
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Min Xu
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Zhenkun Lei
- State Key Laboratory of Structural Analysis for Industrial Equipment, Dalian University of Technology, Dalian 116024, People's Republic of China
| |
Collapse
|
34
|
Guo T, Liang Z, Gu Y, Liu K, Xu X, Yang J, Yu Q. Learning for retinal image quality assessment with label regularization. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 228:107238. [PMID: 36423485 DOI: 10.1016/j.cmpb.2022.107238] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/26/2022] [Revised: 10/03/2022] [Accepted: 11/08/2022] [Indexed: 06/16/2023]
Abstract
BACKGROUND AND OBJECTIVE The assessment of the image quality is crucial before the computer-aided diagnosis of fundus images. This task is very challenging. Firstly, the subjective judgments of graders on image quality lead to ambiguous labels. Secondly, despite being treated as classification in existing works, grading has regression properties that cannot be ignored. Solving the ambiguity problem and regression problem in the label space, and extracting discriminative features, have become the keys to quality assessment. METHODS In this paper, we proposed a framework that can assess the quality of fundus images accurately and reasonably based on deep convolutional neural networks. Drawing on the experience of human graders, a dual-path convolutional neural network with attention blocks is designed to better extract discriminative features and present the bases of decision. Label smoothing and cost-sensitive regularization are designed to solve the label ambiguity problem and the potential regression problem respectively. Besides, a large number of images are annotated by us to further improve the results. RESULTS We conducted our experiments on the largest retinal image quality assessment dataset with 28,792 retinal images. Our approach achieves 0.8868 precision, 0.8786 recall, 0.8820 F1, and 0.9138 Kappa score. Results show that our approach outperforms state-of-the-art methods. CONCLUSIONS The promising performances reveal that our methods are beneficial to retinal image quality assessment and have potential in other grading tasks.
Collapse
Affiliation(s)
- Tianjiao Guo
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China; Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, China; School of Biomedical Engineering, Shanghai Jiao Tong Univeristy, Shanghai, China.
| | - Ziyun Liang
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China; Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, China
| | - Yun Gu
- Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China; Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, China
| | - Kun Liu
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine,Shanghai, China; National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Xun Xu
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine,Shanghai, China; National Clinical Research Center for Eye Diseases, Shanghai, China
| | - Jie Yang
- Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, Shanghai, China.
| | - Qi Yu
- Department of Ophthalmology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine,Shanghai, China; National Clinical Research Center for Eye Diseases, Shanghai, China.
| |
Collapse
|
35
|
Liu YY, Huang ZH, Huang KW. Deep Learning Model for Computer-Aided Diagnosis of Urolithiasis Detection from Kidney-Ureter-Bladder Images. Bioengineering (Basel) 2022; 9:811. [PMID: 36551017 PMCID: PMC9774756 DOI: 10.3390/bioengineering9120811] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2022] [Revised: 12/08/2022] [Accepted: 12/12/2022] [Indexed: 12/24/2022] Open
Abstract
Kidney-ureter-bladder (KUB) imaging is a radiological examination with a low cost, low radiation, and convenience. Although emergency room clinicians can arrange KUB images easily as a first-line examination for patients with suspicious urolithiasis, interpreting the KUB images correctly is difficult for inexperienced clinicians. Obtaining a formal radiology report immediately after a KUB imaging examination can also be challenging. Recently, artificial-intelligence-based computer-aided diagnosis (CAD) systems have been developed to help clinicians who are not experts make correct diagnoses for further treatment more effectively. Therefore, in this study, we proposed a CAD system for KUB imaging based on a deep learning model designed to help first-line emergency room clinicians diagnose urolithiasis accurately. A total of 355 KUB images were retrospectively collected from 104 patients who were diagnosed with urolithiasis at Kaohsiung Chang Gung Memorial Hospital. Then, we trained a deep learning model with a ResNet architecture to classify KUB images in terms of the presence or absence of kidney stones with this dataset of pre-processed images. Finally, we tuned the parameters and tested the model experimentally. The results show that the accuracy, sensitivity, specificity, and F1-measure of the model were 0.977, 0.953, 1, and 0.976 on the validation set and 0.982, 0.964, 1, and 0.982 on the testing set, respectively. Moreover, the results demonstrate that the proposed model performed well compared to the existing CNN-based methods and was able to detect urolithiasis in KUB images successfully. We expect the proposed approach to help emergency room clinicians make accurate diagnoses and reduce unnecessary radiation exposure from computed tomography (CT) scans, along with the associated medical costs.
Collapse
Affiliation(s)
- Yi-Yang Liu
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 80778, Taiwan
- Department of Urology, Kaohsiung Chang Gung Memorial Hospital and Chang Gung University College of Medicine, Kaohsiung City 83301, Taiwan
| | - Zih-Hao Huang
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 80778, Taiwan
| | - Ko-Wei Huang
- Department of Electrical Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 80778, Taiwan
| |
Collapse
|
36
|
Bhatia S, Alojail M, Sengan S, Dadheech P. An efficient modular framework for automatic LIONC classification of MedIMG using unified medical language. Front Public Health 2022; 10:926229. [PMID: 36033768 PMCID: PMC9399779 DOI: 10.3389/fpubh.2022.926229] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2022] [Accepted: 06/27/2022] [Indexed: 01/24/2023] Open
Abstract
Handwritten prescriptions and radiological reports: doctors use handwritten prescriptions and radiological reports to give drugs to patients who have illnesses, injuries, or other problems. Clinical text data, like physician prescription visuals and radiology reports, should be labelled with specific information such as disease type, features, and anatomical location for more effective use. The semantic annotation of vast collections of biological and biomedical texts, like scientific papers, medical reports, and general practitioner observations, has lately been examined by doctors and scientists. By identifying and disambiguating references to biomedical concepts in texts, medical semantics annotators could generate such annotations automatically. For Medical Images (MedIMG), we provide a methodology for learning an effective holistic representation (handwritten word pictures as well as radiology reports). Deep Learning (DL) methods have recently gained much interest for their capacity to achieve expert-level accuracy in automated MedIMG analysis. We discovered that tasks requiring significant responsive fields are ideal for downscaled input images that are qualitatively verified by examining functional, responsive areas and class activating maps for training models. This article focuses on the following contributions: (a) Information Extraction from Narrative MedImages, (b) Automatic categorisation on image resolution with an impact on MedIMG, and (c) Hybrid Model to Predictions of Named Entity Recognition utilising RNN + LSTM + GRM that perform admirably in every trainee for every input purpose. At the same time, supplying understandable scale weight implies that such multi-scale structures are also crucial for extracting information from high-resolution MedIMG. A portion of the reports (30%) are manually evaluated by trained physicians, while the rest were automatically categorised using deep supervised training models based on attention mechanisms and supplied with test reports. MetaMapLite proved recall and precision, but also an F1-score equivalent for primary biomedicine text search techniques and medical text examination on many databases of MedIMG. In addition to implementing as well as getting the requirements for MedIMG, the article explores the quality of medical data by using DL techniques for reaching large-scale labelled clinical data and also the significance of their real-time efforts in the biomedical study that have played an instrumental role in its extramural diffusion and global appeal.
Collapse
Affiliation(s)
- Surbhi Bhatia
- Department of Information Systems, College of Computer Science and Information Technology, King Faisal University, Al Hasa, Saudi Arabia,*Correspondence: Surbhi Bhatia
| | - Mohammed Alojail
- Department of Information Systems, College of Computer Science and Information Technology, King Faisal University, Al Hasa, Saudi Arabia
| | - Sudhakar Sengan
- Department of Computer Science and Engineering, PSN College of Engineering and Technology, Tirunelveli, India
| | - Pankaj Dadheech
- Department of Computer Science and Engineering, Swami Keshvanand Institute of Technology, Management & Gramothan (SKIT), Jaipur, India
| |
Collapse
|
37
|
GAN Training Acceleration Using Fréchet Descriptor-Based Coreset. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12157599] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Generative Adversarial Networks (GANs) are a class of deep learning models being applied to image processing. GANs have demonstrated state-of-the-art performance in applications such as image generation and image-to-image translation, just to name a few. However, with this success comes the realization that the training of GANs takes a long time and is often limited by available computing resources. In this research, we propose to construct a Coreset using Fréchet Descriptor Distances (FDD-Coreset) to accelerate the training of GAN for blob identification. We first propose a Fréchet Descriptor Distance (FDD) to measure the difference between each pair of blob images based on the statistics derived from blob distribution. The Coreset is then employed using our proposed FDD metric to select samples from the entire dataset for GAN training. A 3D-simulated dataset of blobs and a 3D MRI dataset of human kidneys are studied. Using computation time and eight performance metrics, the GAN trained on the FDD-Coreset is compared against the model trained on the entire dataset and an Inception and Euclidean Distance-based Coreset (IED-Coreset). We conclude that the FDD-Coreset not only significantly reduces the training time, but also achieves higher denoising performance and maintains approximate performance of blob identification compared with training on the entire dataset.
Collapse
|
38
|
|
39
|
Deng Z, Cai Y, Chen L, Gong Z, Bao Q, Yao X, Fang D, Yang W, Zhang S, Ma L. RFormer: Transformer-Based Generative Adversarial Network for Real Fundus Image Restoration on a New Clinical Benchmark. IEEE J Biomed Health Inform 2022; 26:4645-4655. [PMID: 35767498 DOI: 10.1109/jbhi.2022.3187103] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Ophthalmologists have used fundus images to screen and diagnose eye diseases. However, different equipments and ophthalmologists pose large variations to the quality of fundus images. Low-quality (LQ) degraded fundus images easily lead to uncertainty in clinical screening and generally increase the risk of misdiagnosis. Thus, real fundus image restoration is worth studying. Unfortunately, real clinical benchmark has not been explored for this task so far. In this paper, we investigate the real clinical fundus image restoration problem. Firstly, We establish a clinical dataset, Real Fundus (RF), including 120 low- and high-quality (HQ) image pairs. Then we propose a novel Transformer-based Generative Adversarial Network (RFormer) to restore the real degradation of clinical fundus images. The key component in our network is the Window-based Self-Attention Block (WSAB) which captures non-local self-similarity and long-range dependencies. To produce more visually pleasant results, a Transformer-based discriminator is introduced. Extensive experiments on our clinical benchmark show that the proposed RFormer significantly outperforms the state-of-the-art (SOTA) methods. In addition, experiments of downstream tasks such as vessel segmentation and optic disc/cup detection demonstrate that our proposed RFormer benefits clinical fundus image analysis and applications. The dataset, code, and models will be made publicly available at https://github.com/dengzhuo-AI/Real-Fundus.
Collapse
|
40
|
A Novel Un-Supervised GAN for Fundus Image Enhancement with Classification Prior Loss. ELECTRONICS 2022. [DOI: 10.3390/electronics11071000] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Fundus images captured for clinical diagnosis usually suffer from degradation factors due to variation in equipment, operators, or environment. These degraded fundus images need to be enhanced to achieve better diagnosis and improve the results of downstream tasks. As there is no paired low- and high-quality fundus image, existing methods mainly focus on supervised or semi-supervised learning methods for color fundus image enhancement (CFIE) tasks by utilizing synthetic image pairs. Consequently, domain gaps between real images and synthetic images arise. With respect to existing unsupervised methods, the most important low scale pathological features and structural information in degraded fundus images are prone to be erased after enhancement. To solve these problems, an unsupervised GAN is proposed for CFIE tasks utilizing adversarial training to enhance low quality fundus images. Synthetic image pairs are no longer required during the training. A specially designed U-Net with skip connection in our enhancement network can effectively remove degradation factors while preserving pathological features and structural information. Global and local discriminators adopted in the GAN lead to better illumination uniformity in the enhanced fundus image. To better improve the visual quality of enhanced fundus images, a novel non-reference loss function based on a pretrained fundus image quality classification network was designed to guide the enhancement network to produce high quality images. Experiments demonstrated that our method could effectively remove degradation factors in low-quality fundus images and produce a competitive result compared with previous methods in both quantitative and qualitative metrics.
Collapse
|
41
|
Huang J, Ding W, Lv J, Yang J, Dong H, Del Ser J, Xia J, Ren T, Wong ST, Yang G. Edge-enhanced dual discriminator generative adversarial network for fast MRI with parallel imaging using multi-view information. APPL INTELL 2022; 52:14693-14710. [PMID: 36199853 PMCID: PMC9526695 DOI: 10.1007/s10489-021-03092-w] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/09/2021] [Indexed: 12/24/2022]
Abstract
In clinical medicine, magnetic resonance imaging (MRI) is one of the most important tools for diagnosis, triage, prognosis, and treatment planning. However, MRI suffers from an inherent slow data acquisition process because data is collected sequentially in k-space. In recent years, most MRI reconstruction methods proposed in the literature focus on holistic image reconstruction rather than enhancing the edge information. This work steps aside this general trend by elaborating on the enhancement of edge information. Specifically, we introduce a novel parallel imaging coupled dual discriminator generative adversarial network (PIDD-GAN) for fast multi-channel MRI reconstruction by incorporating multi-view information. The dual discriminator design aims to improve the edge information in MRI reconstruction. One discriminator is used for holistic image reconstruction, whereas the other one is responsible for enhancing edge information. An improved U-Net with local and global residual learning is proposed for the generator. Frequency channel attention blocks (FCA Blocks) are embedded in the generator for incorporating attention mechanisms. Content loss is introduced to train the generator for better reconstruction quality. We performed comprehensive experiments on Calgary-Campinas public brain MR dataset and compared our method with state-of-the-art MRI reconstruction methods. Ablation studies of residual learning were conducted on the MICCAI13 dataset to validate the proposed modules. Results show that our PIDD-GAN provides high-quality reconstructed MR images, with well-preserved edge information. The time of single-image reconstruction is below 5ms, which meets the demand of faster processing.
Collapse
Affiliation(s)
- Jiahao Huang
- College of Information Science and Technology, Zhejiang Shuren University, 310015 Hangzhou, China
- National Heart and Lung Institute, Imperial College London, London, UK
| | - Weiping Ding
- School of Information Science and Technology, Nantong University, 226019 Nantong, China
| | - Jun Lv
- School of Computer and Control Engineering, Yantai University, 264005 Yantai, China
| | - Jingwen Yang
- Department of Prosthodontics, Peking University School and Hospital of Stomatology, Beijing, China
| | - Hao Dong
- Center on Frontiers of Computing Studies, Peking University, Beijing, China
| | - Javier Del Ser
- TECNALIA, Basque Research and Technology Alliance (BRTA), 48160 Derio, Spain
- University of the Basque Country (UPV/EHU), 48013 Bilbao, Spain
| | - Jun Xia
- Department of Radiology, Shenzhen Second People’s Hospital, The First Afliated Hospital of Shenzhen University Health Science Center, Shenzhen, China
| | - Tiaojuan Ren
- College of Information Science and Technology, Zhejiang Shuren University, 310015 Hangzhou, China
| | - Stephen T. Wong
- Systems Medicine and Bioengineering Department, Departments of Radiology and Pathology, Houston Methodist Cancer Center, Houston Methodist Hospital, Weill Cornell Medicine, 77030 Houston, TX USA
| | - Guang Yang
- National Heart and Lung Institute, Imperial College London, London, UK
- Cardiovascular Research Centre, Royal Brompton Hospital, London, UK
| |
Collapse
|
42
|
Wan C, Zhou X, You Q, Sun J, Shen J, Zhu S, Jiang Q, Yang W. Retinal Image Enhancement Using Cycle-Constraint Adversarial Network. Front Med (Lausanne) 2022; 8:793726. [PMID: 35096883 PMCID: PMC8789669 DOI: 10.3389/fmed.2021.793726] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Accepted: 12/14/2021] [Indexed: 11/25/2022] Open
Abstract
Retinal images are the most intuitive medical images for the diagnosis of fundus diseases. Low-quality retinal images cause difficulties in computer-aided diagnosis systems and the clinical diagnosis of ophthalmologists. The high quality of retinal images is an important basis of precision medicine in ophthalmology. In this study, we propose a retinal image enhancement method based on deep learning to enhance multiple low-quality retinal images. A generative adversarial network is employed to build a symmetrical network, and a convolutional block attention module is introduced to improve the feature extraction capability. The retinal images in our dataset are sorted into two sets according to their quality: low and high quality. Generators and discriminators alternately learn the features of low/high-quality retinal images without the need for paired images. We analyze the proposed method both qualitatively and quantitatively on public datasets and a private dataset. The study results demonstrate that the proposed method is superior to other advanced algorithms, especially in enhancing color-distorted retinal images. It also performs well in the task of retinal vessel segmentation. The proposed network effectively enhances low-quality retinal images, aiding ophthalmologists and enabling computer-aided diagnosis in pathological analysis. Our method enhances multiple types of low-quality retinal images using a deep learning network.
Collapse
Affiliation(s)
- Cheng Wan
- College of Electronic Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Xueting Zhou
- College of Electronic Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Qijing You
- College of Electronic Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Jing Sun
- College of Electronic Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Jianxin Shen
- College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Shaojun Zhu
- School of Information Engineering, Huzhou University, Huzhou, China
| | - Qin Jiang
- The Laboratory of Artificial Intelligence and Bigdata in Ophthalmology, The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| | - Weihua Yang
- The Laboratory of Artificial Intelligence and Bigdata in Ophthalmology, The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| |
Collapse
|