1
|
Jamrasnarodom J, Rajborirug P, Pisespongsa P, Pasupa K. Optimizing colorectal polyp detection and localization: Impact of RGB color adjustment on CNN performance. MethodsX 2025; 14:103187. [PMID: 39975856 PMCID: PMC11836512 DOI: 10.1016/j.mex.2025.103187] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2024] [Accepted: 01/26/2025] [Indexed: 02/21/2025] Open
Abstract
Colorectal cancer, arising from adenomatous polyps, is a leading cause of cancer-related mortality, making early detection and removal crucial for preventing cancer progression. Machine learning is increasingly used to enhance polyp detection during colonoscopy, the gold standard for colorectal cancer screening, despite its operator-dependent miss rates. This study explores the impact of RGB color adjustment on Convolutional Neural Network (CNN) models for improving polyp detection and localization in colonoscopic images. Using datasets from Harvard Dataverse for training and internal validation, and LDPolypVideo-Benchmark for external validation, RGB color adjustments were applied, and YOLOv8s was used to develop models. Bayesian optimization identified the best RGB adjustments, with performance assessed using mean average precision (mAP) and F1-scores. Results showed that RGB adjustment with 1.0 R-1.0 G-0.8 B improved polyp detection, achieving an mAP of 0.777 and an F1-score of 0.720 on internal test sets, and localization performance with an F1-score of 0.883 on adjusted images. External validation showed improvement but with a lower F1-score of 0.556. While RGB adjustments improved performance in our study, their generalizability to diverse datasets and clinical settings has yet to be validated. Thus, although RGB color adjustment enhances CNN model performance for detecting and localizing colorectal polyps, further research is needed to verify these improvements across diverse datasets and clinical settings.•RGB Color Adjustment: Applied RGB color adjustments to colonoscopic images to enhance the performance of Convolutional Neural Network (CNN) models.•Model Development: Used YOLOv8s for polyp detection and localization, with Bayesian optimization to identify the best RGB adjustments.•Performance Evaluation: Assessed model performance using mAP and F1-scores on both internal and external validation datasets.
Collapse
Affiliation(s)
- Jirakorn Jamrasnarodom
- Faculty of Medicine, King Mongkut's Institute of Technology Ladkrabang, Bangkok, Thailand
| | - Pharuj Rajborirug
- Faculty of Medicine, King Mongkut's Institute of Technology Ladkrabang, Bangkok, Thailand
| | - Pises Pisespongsa
- Faculty of Medicine, King Mongkut's Institute of Technology Ladkrabang, Bangkok, Thailand
| | - Kitsuchart Pasupa
- School of Information Technology, King Mongkut's Institute of Technology Ladkrabang, Bangkok, Thailand
| |
Collapse
|
2
|
Song Y, Du S, Wang R, Liu F, Lin X, Chen J, Li Z, Li Z, Yang L, Zhang Z, Yan H, Zhang Q, Qian D, Li X. Polyp-Size: A Precise Endoscopic Dataset for AI-Driven Polyp Sizing. Sci Data 2025; 12:918. [PMID: 40450075 DOI: 10.1038/s41597-025-05251-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2024] [Accepted: 05/21/2025] [Indexed: 06/03/2025] Open
Abstract
Colorectal cancer often arises from precancerous polyps, where accurate size assessment is vital for clinical decisions but challenged by subjective methods. While artificial intelligence (AI) has shown promise in improving the accuracy of polyp size estimation, its development depends on large, meticulously annotated datasets. We present Polyp-Size, a dataset of 42 high-resolution white-light colonoscopy videos with polyp sizes precisely measured post-resection using vernier calipers to submillimeter precision. Unlike existing datasets primarily focused on polyp detection or segmentation, Polyp-Size offers validated size annotations, diverse polyp features (Paris classification, anatomical location and histological type), and standardized video formats, enabling robust AI models for size estimation. By making this resource publicly available, we aim to foster research collaboration and innovation in automated polyp measurement to ultimately improve clinical practice.
Collapse
Affiliation(s)
- Yiming Song
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Sijia Du
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Ruilan Wang
- Department of Gastroenterology, Armed Police Forces Hospital of Sichuan, Leshan, Sichuan Province, China
| | - Fei Liu
- Departmant of Gastroenterology, Nine Division Hospital of Xinjiang Production and Construction Corps, Tacheng Xinjiang Uygur Autonomous Region, Tacheng, China
| | - Xiaolu Lin
- Department of Digestive Endoscopy Center, Fujian Provincial Hospital, Shengli Clinical Medical College of Fujian Medical University, Fuzhou, Fujian, China
| | - Jinnan Chen
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Zeyu Li
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Zhao Li
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Liuyi Yang
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Zhengjie Zhang
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Hao Yan
- The Second Clinical Medical College, Harbin Medical University, Harbin, 150081, China
| | - Qingwei Zhang
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
| | - Dahong Qian
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China.
| | - Xiaobo Li
- Division of Gastroenterology and Hepatology, Shanghai Institute of Digestive Disease, NHC Key Laboratory of Digestive Diseases, Renji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
| |
Collapse
|
3
|
Selvaraj J, Sadaf K, Aslam SM, Umapathy S. Multiclassification of Colorectal Polyps from Colonoscopy Images Using AI for Early Diagnosis. Diagnostics (Basel) 2025; 15:1285. [PMID: 40428278 PMCID: PMC12109892 DOI: 10.3390/diagnostics15101285] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2025] [Revised: 05/10/2025] [Accepted: 05/12/2025] [Indexed: 05/29/2025] Open
Abstract
Background/Objectives: Colorectal cancer (CRC) remains one of the leading causes of cancer-related mortality worldwide, emphasizing the critical need for the accurate classification of precancerous polyps. This research presents an extensive analysis of the multiclassification framework leveraging various deep learning (DL) architectures for the automated classification of colorectal polyps from colonoscopy images. Methods: The proposed methodology integrates real-time data for training and utilizes a publicly available dataset for testing, ensuring generalizability. The real-time images were cautiously annotated and verified by a panel of experts, including post-graduate medical doctors and gastroenterology specialists. The DL models were designed to categorize the preprocessed colonoscopy images into four clinically significant classes: hyperplastic, serrated, adenoma, and normal. A suite of state-of-the-art models, including VGG16, VGG19, ResNet50, DenseNet121, EfficientNetV2, InceptionNetV3, Vision Transformer (ViT), and the custom-developed CRP-ViT, were trained and rigorously evaluated for this task. Results: Notably, the CRP-ViT model exhibited superior capability in capturing intricate features, achieving an impressive accuracy of 97.28% during training and 96.02% during validation with real-time images. Furthermore, the model demonstrated remarkable performance during testing on the public dataset, attaining an accuracy of 95.69%. To facilitate real-time interaction and clinical applicability, a user-friendly interface was developed using Gradio, allowing healthcare professionals to upload colonoscopy images and receive instant classification results. Conclusions: The CRP-ViT model effectively predicts and categorizes colonoscopy images into clinically relevant classes, aiding gastroenterologists in decision-making. This study highlights the potential of integrating AI-driven models into routine clinical practice to improve colorectal cancer screening outcomes and reduce diagnostic variability.
Collapse
Affiliation(s)
- Jothiraj Selvaraj
- Department of Biomedical Engineering, College of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu 603203, India;
| | - Kishwar Sadaf
- Department of Computer Science, College of Computer and Information Sciences, Majmaah University, Al Majmaah 11952, Saudi Arabia
| | - Shabnam Mohamed Aslam
- Department of Information Technology, College of Computer and Information Sciences, Majmaah University, Al Majmaah 11952, Saudi Arabia;
| | - Snekhalatha Umapathy
- Department of Biomedical Engineering, College of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu 603203, India;
| |
Collapse
|
4
|
Wang Z, Liu C, Zhu L, Wang T, Zhang S, Dou Q. Improving Foundation Model for Endoscopy Video Analysis via Representation Learning on Long Sequences. IEEE J Biomed Health Inform 2025; 29:3526-3536. [PMID: 40031835 DOI: 10.1109/jbhi.2025.3532311] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Recent advancements in endoscopy video analysis have relied on the utilization of relatively short video clips extracted from longer videos or millions of individual frames. However, these approaches tend to neglect the domain-specific characteristics of endoscopy data, which is typically presented as a long stream containing valuable semantic spatial and temporal information. To address this limitation, we propose EndoFM-LV, a foundation model developed under a minute-level pre-training framework upon long endoscopy video sequences. To be specific, we propose a novel masked token modeling scheme within a teacher-student framework for self-supervised video pre-training, which is tailored for learning representations from long video sequences. For pre-training, we construct a large-scale long endoscopy video dataset comprising 6,469 long endoscopic video samples, each longer than 1 minute and totaling over 13 million frames. Our EndoFM-LV is evaluated on four types of endoscopy tasks, namely classification, segmentation, detection, and workflow recognition, serving as the backbone or temporal module. Extensive experimental results demonstrate that our framework outperforms previous state-of-the-art video-based and frame-based approaches by a significant margin, surpassing Endo-FM (5.6% F1, 9.3% Dice, 8.4% F1, and 3.3% accuracy for classification, segmentation, detection, and workflow recognition) and EndoSSL (5.0% F1, 8.1% Dice, 9.3% F1 and 3.1% accuracy for classification, segmentation, detection, and workflow recognition).
Collapse
|
5
|
Kim Y, Keum JS, Kim JH, Chun J, Oh SI, Kim KN, Yoon YH, Park H. Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor. Diagnostics (Basel) 2025; 15:901. [PMID: 40218251 PMCID: PMC11988911 DOI: 10.3390/diagnostics15070901] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2025] [Revised: 03/25/2025] [Accepted: 03/31/2025] [Indexed: 04/14/2025] Open
Abstract
Background/Objectives: Artificial intelligence (AI) integration in colon polyp detection often exhibits high sensitivity but notably low specificity in real-world settings, primarily due to reliance on publicly available datasets alone. To address this limitation, we proposed a semi-automatic annotation method using real colonoscopy videos to enhance AI model performance and reduce manual labeling labor. Methods: An integrated AI model was trained and validated on 86,258 training images and 17,616 validation images. Model 1 utilized only publicly available datasets, while Model 2 additionally incorporated images obtained from real colonoscopy videos of patients through a semi-automatic annotation process, significantly reducing the labeling burden on expert endoscopists. Results: The integrated AI model (Model 2) significantly outperformed the public-dataset-only model (Model 1). At epoch 35, Model 2 achieved a sensitivity of 90.6%, a specificity of 96.0%, an overall accuracy of 94.5%, and an F1 score of 89.9%. All polyps in the test videos were successfully detected, demonstrating considerable enhancement in detection performance compared to the public-dataset-only model. Conclusions: Integrating real-world colonoscopy video data using semi-automatic annotation markedly improved diagnostic accuracy while potentially reducing the need for extensive manual annotation typically performed by expert endoscopists. However, the findings need validation through multicenter external datasets to ensure generalizability.
Collapse
Affiliation(s)
- Yuna Kim
- Department of Internal Medicine, Division of Gastroenterology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul 06273, Republic of Korea; (Y.K.)
| | - Ji-Soo Keum
- Waycen Inc., Seoul 06167, Republic of Korea; (J.-S.K.)
| | - Jie-Hyun Kim
- Department of Internal Medicine, Division of Gastroenterology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul 06273, Republic of Korea; (Y.K.)
| | - Jaeyoung Chun
- Department of Internal Medicine, Division of Gastroenterology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul 06273, Republic of Korea; (Y.K.)
| | - Sang-Il Oh
- Waycen Inc., Seoul 06167, Republic of Korea; (J.-S.K.)
| | - Kyung-Nam Kim
- Waycen Inc., Seoul 06167, Republic of Korea; (J.-S.K.)
| | - Young-Hoon Yoon
- Department of Internal Medicine, Division of Gastroenterology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul 06273, Republic of Korea; (Y.K.)
| | - Hyojin Park
- Department of Internal Medicine, Division of Gastroenterology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul 06273, Republic of Korea; (Y.K.)
| |
Collapse
|
6
|
Elamin S, Johri S, Rajpurkar P, Geisler E, Berzin TM. From data to artificial intelligence: evaluating the readiness of gastrointestinal endoscopy datasets. J Can Assoc Gastroenterol 2025; 8:S81-S86. [PMID: 39990508 PMCID: PMC11842897 DOI: 10.1093/jcag/gwae041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 02/25/2025] Open
Abstract
The incorporation of artificial intelligence (AI) into gastrointestinal (GI) endoscopy represents a promising advancement in gastroenterology. With over 40 published randomized controlled trials and numerous ongoing clinical trials, gastroenterology leads other medical disciplines in AI research. Computer-aided detection algorithms for identifying colorectal polyps have achieved regulatory approval and are in routine clinical use, while other AI applications for GI endoscopy are in advanced development stages. Near-term opportunities include the potential for computer-aided diagnosis to replace conventional histopathology for diagnosing small colon polyps and increased AI automation in capsule endoscopy. Despite significant development in research settings, the generalizability and robustness of AI models in real clinical practice remain inconsistent. The GI field lags behind other medical disciplines in the breadth of novel AI algorithms, with only 13 out of 882 Food and Drug Administration (FDA)-approved AI models focussed on GI endoscopy as of June 2024. Additionally, existing GI endoscopy image databases are disproportionately focussed on colon polyps, lacking representation of the diversity of other endoscopic findings. High-quality datasets, encompassing a wide range of patient demographics, endoscopic equipment types, and disease states, are crucial for developing effective AI models for GI endoscopy. This article reviews the current state of GI endoscopy datasets, barriers to progress, including dataset size, data diversity, annotation quality, and ethical issues in data collection and usage, and future needs for advancing AI in GI endoscopy.
Collapse
Affiliation(s)
- Sami Elamin
- Department of Biomedical Informatics, Harvard Medical School, Boston, MA 02115, USA
| | - Shreya Johri
- Department of Biomedical Informatics, Harvard Medical School, Boston, MA 02115, USA
| | - Pranav Rajpurkar
- Department of Biomedical Informatics, Harvard Medical School, Boston, MA 02115, USA
| | - Enrik Geisler
- Center for Advanced Endoscopy, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, MA 02115, USA
| | - Tyler M Berzin
- Center for Advanced Endoscopy, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, MA 02115, USA
| |
Collapse
|
7
|
He D, Liu Z, Yin X, Liu H, Gao W, Fu Y. Synthesized colonoscopy dataset from high-fidelity virtual colon with abnormal simulation. Comput Biol Med 2025; 186:109672. [PMID: 39826299 DOI: 10.1016/j.compbiomed.2025.109672] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2024] [Revised: 01/06/2025] [Accepted: 01/08/2025] [Indexed: 01/22/2025]
Abstract
With the advent of the deep learning-based colonoscopy system, the need for a vast amount of high-quality colonoscopy image datasets for training is crucial. However, the generalization ability of deep learning models is challenged by the limited availability of colonoscopy images due to regulatory restrictions and privacy concerns. In this paper, we propose a method for rendering high-fidelity 3D colon models and synthesizing diversified colonoscopy images with abnormalities such as polyps, bleeding, and ulcers, which can be used to train deep learning models. The geometric model of the colon is derived from CT images. We employed dedicated surface mesh deformation to mimic the shapes of polyps and ulcers and applied texture mapping techniques to generate realistic, lifelike appearances. The generated polyp models were then attached to the inner surface of the colon model, while the ulcers were created directly on the inner surface of the colon model. To realistically model blood behavior, we developed a simulation of the blood diffusion process on the colon's inner surface and colored vertices in the traversed region to reflect blood flow. Ultimately, we generated a comprehensive dataset comprising high-fidelity rendered colonoscopy images with the abnormalities. To validate the effectiveness of the synthesized colonoscopy dataset, we trained state-of-the-art deep learning models on it and other publicly available datasets and assessed the performance of these models in abnormal classification, detection, and segmentation. Notably, the models trained on the synthesized dataset exhibit an enhanced performance in the aforementioned tasks, as evident from the results.
Collapse
Affiliation(s)
- Dongdong He
- School of Life Science and Technology, Harbin Institute of Technology, Harbin, 150080, China
| | - Ziteng Liu
- School of Life Science and Technology, Harbin Institute of Technology, Harbin, 150080, China
| | - Xunhai Yin
- Department of Gastroenterology, The First Affiliated Hospital of Harbin Medical University, Harbin, 150001, China
| | - Hao Liu
- State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang, 110016, China
| | - Wenpeng Gao
- School of Life Science and Technology, Harbin Institute of Technology, Harbin, 150080, China; State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, 150080, China.
| | - Yili Fu
- State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, 150080, China.
| |
Collapse
|
8
|
Fu J, Chen K, Dou Q, Gao Y, He Y, Zhou P, Lin S, Wang Y, Guo Y. IPNet: An Interpretable Network With Progressive Loss for Whole-Stage Colorectal Disease Diagnosis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:789-800. [PMID: 39298304 DOI: 10.1109/tmi.2024.3459910] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/21/2024]
Abstract
Colorectal cancer plays a dominant role in cancer-related deaths, primarily due to the absence of obvious early-stage symptoms. Whole-stage colorectal disease diagnosis is crucial for assessing lesion evolution and determining treatment plans. However, locality difference and disease progression lead to intra-class disparities and inter-class similarities for colorectal lesion representation. In addition, interpretable algorithms explaining the lesion progression are still lacking, making the prediction process a "black box". In this paper, we propose IPNet, a dual-branch interpretable network with progressive loss for whole-stage colorectal disease diagnosis. The dual-branch architecture captures unbiased features representing diverse localities to suppress intra-class variation. The progressive loss function considers inter-class relationship, using prior knowledge of disease evolution to guide classification. Furthermore, a novel Grain-CAM is designed to interpret IPNet by visualizing pixel-wise attention maps from shallow to deep layers, providing regions semantically related to IPNet's progressive classification. We conducted whole-stage diagnosis on two image modalities, i.e., colorectal lesion classification on 129,893 endoscopic optical images and rectal tumor T-staging on 11,072 endoscopic ultrasound images. IPNet is shown to surpass other state-of-the-art algorithms, accordingly achieving an accuracy of 93.15% and 89.62%. Especially, it establishes effective decision boundaries for challenges like polyp vs. adenoma and T2 vs. T3. The results demonstrate an explainable attempt for colorectal lesion classification at a whole-stage level, and rectal tumor T-staging by endoscopic ultrasound is also unprecedentedly explored. IPNet is expected to be further applied, assisting physicians in whole-stage disease diagnosis and enhancing diagnostic interpretability.
Collapse
|
9
|
Raju ASN, Venkatesh K, Gatla RK, Konakalla EP, Eid MM, Titova N, Ghoneim SSM, Ghaly RNR. Colorectal cancer detection with enhanced precision using a hybrid supervised and unsupervised learning approach. Sci Rep 2025; 15:3180. [PMID: 39863646 PMCID: PMC11763007 DOI: 10.1038/s41598-025-86590-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2024] [Accepted: 01/13/2025] [Indexed: 01/27/2025] Open
Abstract
The current work introduces the hybrid ensemble framework for the detection and segmentation of colorectal cancer. This framework will incorporate both supervised classification and unsupervised clustering methods to present more understandable and accurate diagnostic results. The method entails several steps with CNN models: ADa-22 and AD-22, transformer networks, and an SVM classifier, all inbuilt. The CVC ClinicDB dataset supports this process, containing 1650 colonoscopy images classified as polyps or non-polyps. The best performance in the ensembles was done by the AD-22 + Transformer + SVM model, with an AUC of 0.99, a training accuracy of 99.50%, and a testing accuracy of 99.00%. This group also saw a high accuracy of 97.50% for Polyps and 99.30% for Non-Polyps, together with a recall of 97.80% for Polyps and 98.90% for Non-Polyps, hence performing very well in identifying both cancerous and healthy regions. The framework proposed here uses K-means clustering in combination with the visualisation of bounding boxes, thereby improving segmentation and yielding a silhouette score of 0.73 with the best cluster configuration. It discusses how to combine feature interpretation challenges into medical imaging for accurate localization and precise segmentation of malignant regions. A good balance between performance and generalization shall be done by hyperparameter optimization-heavy learning rates; dropout rates and overfitting shall be suppressed effectively. The hybrid schema of this work treats the deficiencies of the previous approaches, such as incorporating CNN-based effective feature extraction, Transformer networks for developing attention mechanisms, and finally the fine decision boundary of the support vector machine. Further, we refine this process via unsupervised clustering for the purpose of enhancing the visualisation of such a procedure. Such a holistic framework, hence, further boosts classification and segmentation results by generating understandable outcomes for more rigorous benchmarking of detecting colorectal cancer and with higher reality towards clinical application feasibility.
Collapse
Affiliation(s)
- Akella S Narasimha Raju
- Department of Computer Science and Engineering (Data Science), Institute of Aeronautical Engineering, Dundigul, Hyderabad, Telangana, 500043, India.
| | - K Venkatesh
- Department of Networking and Communications, School of Computing, SRM Institute of Science and Technology, Kattankulathur, Chennai, Tamilnadu, 603203, India.
| | - Ranjith Kumar Gatla
- Department of Computer Science and Engineering (Data Science), Institute of Aeronautical Engineering, Dundigul, Hyderabad, Telangana, 500043, India
| | - Eswara Prasad Konakalla
- Department of Physics and Electronics, B.V.Raju College, Bhimavaram, Garagaparru Road, Kovvada, Andhra Pradesh, 534202, India
| | - Marwa M Eid
- College of Applied Medical Science, Taif University, 21944, Taif, Saudi Arabia
| | - Nataliia Titova
- Biomedical Engineering Department, National University Odesa Polytechnic, Odesa, 65044, Ukraine.
| | - Sherif S M Ghoneim
- Department of Electrical Engineering, College of Engineering, Taif University, 21944, Taif, Saudi Arabia
| | - Ramy N R Ghaly
- Ministry of Higher Education, Mataria Technical College, Cairo, 11718, Egypt
- Chitkara Centre for Research and Development, Chitkara University, Solan, Himachal Pradesh, 174103, India
| |
Collapse
|
10
|
Albuquerque C, Henriques R, Castelli M. Deep learning-based object detection algorithms in medical imaging: Systematic review. Heliyon 2025; 11:e41137. [PMID: 39758372 PMCID: PMC11699422 DOI: 10.1016/j.heliyon.2024.e41137] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Revised: 12/04/2024] [Accepted: 12/10/2024] [Indexed: 01/06/2025] Open
Abstract
Over the past decade, Deep Learning (DL) techniques have demonstrated remarkable advancements across various domains, driving their widespread adoption. Particularly in medical image analysis, DL received greater attention for tasks like image segmentation, object detection, and classification. This paper provides an overview of DL-based object recognition in medical images, exploring recent methods and emphasizing different imaging techniques and anatomical applications. Utilizing a meticulous quantitative and qualitative analysis following PRISMA guidelines, we examined publications based on citation rates to explore into the utilization of DL-based object detectors across imaging modalities and anatomical domains. Our findings reveal a consistent rise in the utilization of DL-based object detection models, indicating unexploited potential in medical image analysis. Predominantly within Medicine and Computer Science domains, research in this area is most active in the US, China, and Japan. Notably, DL-based object detection methods have gotten significant interest across diverse medical imaging modalities and anatomical domains. These methods have been applied to a range of techniques including CR scans, pathology images, and endoscopic imaging, showcasing their adaptability. Moreover, diverse anatomical applications, particularly in digital pathology and microscopy, have been explored. The analysis underscores the presence of varied datasets, often with significant discrepancies in size, with a notable percentage being labeled as private or internal, and with prospective studies in this field remaining scarce. Our review of existing trends in DL-based object detection in medical images offers insights for future research directions. The continuous evolution of DL algorithms highlighted in the literature underscores the dynamic nature of this field, emphasizing the need for ongoing research and fitted optimization for specific applications.
Collapse
|
11
|
Braverman-Jaiven D, Manfredi L. Advancements in the use of AI in the diagnosis and management of inflammatory bowel disease. Front Robot AI 2024; 11:1453194. [PMID: 39498116 PMCID: PMC11532194 DOI: 10.3389/frobt.2024.1453194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2024] [Accepted: 10/07/2024] [Indexed: 11/07/2024] Open
Abstract
Inflammatory bowel disease (IBD) causes chronic inflammation of the colon and digestive tract, and it can be classified as Crohn's disease (CD) and Ulcerative colitis (UC). IBD is more prevalent in Europe and North America, however, since the beginning of the 21st century it has been increasing in South America, Asia, and Africa, leading to its consideration as a worldwide problem. Optical colonoscopy is one of the crucial tests in diagnosing and assessing the progression and prognosis of IBD, as it allows a real-time optical visualization of the colonic wall and ileum and allows for the collection of tissue samples. The accuracy of colonoscopy procedures depends on the expertise and ability of the endoscopists. Therefore, algorithms based on Deep Learning (DL) and Convolutional Neural Networks (CNN) for colonoscopy images and videos are growing in popularity, especially for the detection and classification of colorectal polyps. The performance of this system is dependent on the quality and quantity of the data used for training. There are several datasets publicly available for endoscopy images and videos, but most of them are solely specialized in polyps. The use of DL algorithms to detect IBD is still in its inception, most studies are based on assessing the severity of UC. As artificial intelligence (AI) grows in popularity there is a growing interest in the use of these algorithms for diagnosing and classifying the IBDs and managing their progression. To tackle this, more annotated colonoscopy images and videos will be required for the training of new and more reliable AI algorithms. This article discusses the current challenges in the early detection of IBD, focusing on the available AI algorithms, and databases, and the challenges ahead to improve the detection rate.
Collapse
Affiliation(s)
| | - Luigi Manfredi
- Division of Imaging Science and Technology, School of Medicine, University of Dundee, Dundee, United Kingdom
| |
Collapse
|
12
|
Tan J, Yuan J, Fu X, Bai Y. Colonoscopy polyp classification via enhanced scattering wavelet Convolutional Neural Network. PLoS One 2024; 19:e0302800. [PMID: 39392783 PMCID: PMC11469526 DOI: 10.1371/journal.pone.0302800] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 08/26/2024] [Indexed: 10/13/2024] Open
Abstract
Among the most common cancers, colorectal cancer (CRC) has a high death rate. The best way to screen for colorectal cancer (CRC) is with a colonoscopy, which has been shown to lower the risk of the disease. As a result, Computer-aided polyp classification technique is applied to identify colorectal cancer. But visually categorizing polyps is difficult since different polyps have different lighting conditions. Different from previous works, this article presents Enhanced Scattering Wavelet Convolutional Neural Network (ESWCNN), a polyp classification technique that combines Convolutional Neural Network (CNN) and Scattering Wavelet Transform (SWT) to improve polyp classification performance. This method concatenates simultaneously learnable image filters and wavelet filters on each input channel. The scattering wavelet filters can extract common spectral features with various scales and orientations, while the learnable filters can capture image spatial features that wavelet filters may miss. A network architecture for ESWCNN is designed based on these principles and trained and tested using colonoscopy datasets (two public datasets and one private dataset). An n-fold cross-validation experiment was conducted for three classes (adenoma, hyperplastic, serrated) achieving a classification accuracy of 96.4%, and 94.8% accuracy in two-class polyp classification (positive and negative). In the three-class classification, correct classification rates of 96.2% for adenomas, 98.71% for hyperplastic polyps, and 97.9% for serrated polyps were achieved. The proposed method in the two-class experiment reached an average sensitivity of 96.7% with 93.1% specificity. Furthermore, we compare the performance of our model with the state-of-the-art general classification models and commonly used CNNs. Six end-to-end models based on CNNs were trained using 2 dataset of video sequences. The experimental results demonstrate that the proposed ESWCNN method can effectively classify polyps with higher accuracy and efficacy compared to the state-of-the-art CNN models. These findings can provide guidance for future research in polyp classification.
Collapse
Affiliation(s)
- Jun Tan
- School of Mathematics, Sun Yat-Sen University, Guangzhou, Guangdong, China
- Guangdong Province Key Laboratory of Computational Science, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Jiamin Yuan
- Health construction administration center, Guangdong Provincial Hospital of Chinese Medicine, Guangzhou, Guangdong, China
- The Second Affiliated Hospital of Guangzhou University of Traditional Chinese Medicine(TCM), Guangzhou, Guangdong, China
| | - Xiaoyong Fu
- School of Mathematics, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Yilin Bai
- School of Mathematics, Sun Yat-Sen University, Guangzhou, Guangdong, China
- China Southern Airlines, Guangzhou, Guangdong, China
| |
Collapse
|
13
|
Wang Y, Ni H, Zhou J, Liu L, Lin J, Yin M, Gao J, Zhu S, Yin Q, Zhu J, Li R. A Semi-Supervised Learning Framework for Classifying Colorectal Neoplasia Based on the NICE Classification. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2342-2353. [PMID: 38653910 PMCID: PMC11522217 DOI: 10.1007/s10278-024-01123-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/21/2024] [Revised: 04/02/2024] [Accepted: 04/12/2024] [Indexed: 04/25/2024]
Abstract
Labelling medical images is an arduous and costly task that necessitates clinical expertise and large numbers of qualified images. Insufficient samples can lead to underfitting during training and poor performance of supervised learning models. In this study, we aim to develop a SimCLR-based semi-supervised learning framework to classify colorectal neoplasia based on the NICE classification. First, the proposed framework was trained under self-supervised learning using a large unlabelled dataset; subsequently, it was fine-tuned on a limited labelled dataset based on the NICE classification. The model was evaluated on an independent dataset and compared with models based on supervised transfer learning and endoscopists using accuracy, Matthew's correlation coefficient (MCC), and Cohen's kappa. Finally, Grad-CAM and t-SNE were applied to visualize the models' interpretations. A ResNet-backboned SimCLR model (accuracy of 0.908, MCC of 0.862, and Cohen's kappa of 0.896) outperformed supervised transfer learning-based models (means: 0.803, 0.698, and 0.742) and junior endoscopists (0.816, 0.724, and 0.863), while performing only slightly worse than senior endoscopists (0.916, 0.875, and 0.944). Moreover, t-SNE showed a better clustering of ternary samples through self-supervised learning in SimCLR than through supervised transfer learning. Compared with traditional supervised learning, semi-supervised learning enables deep learning models to achieve improved performance with limited labelled endoscopic images.
Collapse
Affiliation(s)
- Yu Wang
- Department of Hepatobiliary Surgery, Jintan Affiliated Hospital of Jiangsu University, Changzhou, Jiangsu, 213200, China
| | - Haoxiang Ni
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China
| | - Jielu Zhou
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China
- Department of Geriatrics, Kowloon Affiliated Hospital of Shanghai Jiao Tong University, Suzhou, Jiangsu, 215006, China
| | - Lihe Liu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China
| | - Jiaxi Lin
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China
| | - Minyue Yin
- Department of Gastroenterology, Beijing Friendship Hospital, Capital Medical University, Beijing, 100050, China
- National Clinical Research Center for Digestive Disease, Beijing Digestive Disease Center, State Key Laboratory of Digestive Health, Beijing, 100050, China
| | - Jingwen Gao
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China
| | - Shiqi Zhu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China
| | - Qi Yin
- Department of Anesthesiology, Jintan Affiliated Hospital of Jiangsu University, Changzhou, Jiangsu, 213200, China
| | - Jinzhou Zhu
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China.
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China.
- Key Laboratory of Hepatosplenic Surgery, Ministry of Education, The First Affiliated Hospital of Harbin Medical University, Harbin, 150001, China.
| | - Rui Li
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, # 899 Pinghai St., Suzhou, Jiangsu, 215006, China.
- Suzhou Clinical Center of Digestive Disease, Suzhou, Jiangsu, 215006, China.
| |
Collapse
|
14
|
Tudela Y, Majó M, de la Fuente N, Galdran A, Krenzer A, Puppe F, Yamlahi A, Tran TN, Matuszewski BJ, Fitzgerald K, Bian C, Pan J, Liu S, Fernández-Esparrach G, Histace A, Bernal J. A complete benchmark for polyp detection, segmentation and classification in colonoscopy images. Front Oncol 2024; 14:1417862. [PMID: 39381041 PMCID: PMC11458519 DOI: 10.3389/fonc.2024.1417862] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 07/11/2024] [Indexed: 10/10/2024] Open
Abstract
Introduction Colorectal cancer (CRC) is one of the main causes of deaths worldwide. Early detection and diagnosis of its precursor lesion, the polyp, is key to reduce its mortality and to improve procedure efficiency. During the last two decades, several computational methods have been proposed to assist clinicians in detection, segmentation and classification tasks but the lack of a common public validation framework makes it difficult to determine which of them is ready to be deployed in the exploration room. Methods This study presents a complete validation framework and we compare several methodologies for each of the polyp characterization tasks. Results Results show that the majority of the approaches are able to provide good performance for the detection and segmentation task, but that there is room for improvement regarding polyp classification. Discussion While studied show promising results in the assistance of polyp detection and segmentation tasks, further research should be done in classification task to obtain reliable results to assist the clinicians during the procedure. The presented framework provides a standarized method for evaluating and comparing different approaches, which could facilitate the identification of clinically prepared assisting methods.
Collapse
Affiliation(s)
- Yael Tudela
- Computer Vision Center and Computer Science Department, Universitat Autònoma de Cerdanyola del Valles, Barcelona, Spain
| | - Mireia Majó
- Computer Vision Center and Computer Science Department, Universitat Autònoma de Cerdanyola del Valles, Barcelona, Spain
| | - Neil de la Fuente
- Computer Vision Center and Computer Science Department, Universitat Autònoma de Cerdanyola del Valles, Barcelona, Spain
| | - Adrian Galdran
- Department of Information and Communication Technologies, SymBioSys Research Group, BCNMedTech, Barcelona, Spain
| | - Adrian Krenzer
- Artificial Intelligence and Knowledge Systems, Institute for Computer Science, Julius-Maximilians University of Würzburg, Würzburg, Germany
| | - Frank Puppe
- Artificial Intelligence and Knowledge Systems, Institute for Computer Science, Julius-Maximilians University of Würzburg, Würzburg, Germany
| | - Amine Yamlahi
- Division of Intelligent Medical Systems, German Cancer Research Center (DKFZ), Heidelberg, Germany
| | - Thuy Nuong Tran
- Division of Intelligent Medical Systems, German Cancer Research Center (DKFZ), Heidelberg, Germany
| | - Bogdan J. Matuszewski
- Computer Vision and Machine Learning (CVML) Research Group, University of Central Lancashir (UCLan), Preston, United Kingdom
| | - Kerr Fitzgerald
- Computer Vision and Machine Learning (CVML) Research Group, University of Central Lancashir (UCLan), Preston, United Kingdom
| | - Cheng Bian
- Hebei University of Technology, Baoding, China
| | | | - Shijle Liu
- Hebei University of Technology, Baoding, China
| | | | - Aymeric Histace
- ETIS UMR 8051, École Nationale Supérieure de l'Électronique et de ses Applications (ENSEA), Centre national de la recherche scientifique (CNRS), CY Paris Cergy University, Cergy, France
| | - Jorge Bernal
- Computer Vision Center and Computer Science Department, Universitat Autònoma de Cerdanyola del Valles, Barcelona, Spain
| |
Collapse
|
15
|
Lin Q, Tan W, Cai S, Yan B, Li J, Zhong Y. Lesion-Decoupling-Based Segmentation With Large-Scale Colon and Esophageal Datasets for Early Cancer Diagnosis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:11142-11156. [PMID: 37028330 DOI: 10.1109/tnnls.2023.3248804] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Lesions of early cancers often show flat, small, and isochromatic characteristics in medical endoscopy images, which are difficult to be captured. By analyzing the differences between the internal and external features of the lesion area, we propose a lesion-decoupling-based segmentation (LDS) network for assisting early cancer diagnosis. We introduce a plug-and-play module called self-sampling similar feature disentangling module (FDM) to obtain accurate lesion boundaries. Then, we propose a feature separation loss (FSL) function to separate pathological features from normal ones. Moreover, since physicians make diagnoses with multimodal data, we propose a multimodal cooperative segmentation network with two different modal images as input: white-light images (WLIs) and narrowband images (NBIs). Our FDM and FSL show a good performance for both single-modal and multimodal segmentations. Extensive experiments on five backbones prove that our FDM and FSL can be easily applied to different backbones for a significant lesion segmentation accuracy improvement, and the maximum increase of mean Intersection over Union (mIoU) is 4.58. For colonoscopy, we can achieve up to mIoU of 91.49 on our Dataset A and 84.41 on the three public datasets. For esophagoscopy, mIoU of 64.32 is best achieved on the WLI dataset and 66.31 on the NBI dataset.
Collapse
|
16
|
Marchese Aizenman G, Salvagnini P, Cherubini A, Biffi C. Assessing clinical efficacy of polyp detection models using open-access datasets. Front Oncol 2024; 14:1422942. [PMID: 39148908 PMCID: PMC11324571 DOI: 10.3389/fonc.2024.1422942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2024] [Accepted: 07/08/2024] [Indexed: 08/17/2024] Open
Abstract
Background Ensuring accurate polyp detection during colonoscopy is essential for preventing colorectal cancer (CRC). Recent advances in deep learning-based computer-aided detection (CADe) systems have shown promise in enhancing endoscopists' performances. Effective CADe systems must achieve high polyp detection rates from the initial seconds of polyp appearance while maintaining low false positive (FP) detection rates throughout the procedure. Method We integrated four open-access datasets into a unified platform containing over 340,000 images from various centers, including 380 annotated polyps, with distinct data splits for comprehensive model development and benchmarking. The REAL-Colon dataset, comprising 60 full-procedure colonoscopy videos from six centers, is used as the fifth dataset of the platform to simulate clinical conditions for model evaluation on unseen center data. Performance assessment includes traditional object detection metrics and new metrics that better meet clinical needs. Specifically, by defining detection events as sequences of consecutive detections, we compute per-polyp recall at early detection stages and average per-patient FPs, enabling the generation of Free-Response Receiver Operating Characteristic (FROC) curves. Results Using YOLOv7, we trained and tested several models across the proposed data splits, showcasing the robustness of our open-access platform for CADe system development and benchmarking. The introduction of new metrics allows for the optimization of CADe operational parameters based on clinically relevant criteria, such as per-patient FPs and early polyp detection. Our findings also reveal that omitting full-procedure videos leads to non-realistic assessments and that detecting small polyp bounding boxes poses the greatest challenge. Conclusion This study demonstrates how newly available open-access data supports ongoing research progress in environments that closely mimic clinical settings. The introduced metrics and FROC curves illustrate CADe clinical efficacy and can aid in tuning CADe hyperparameters.
Collapse
Affiliation(s)
| | | | - Andrea Cherubini
- Cosmo Intelligent Medical Devices, Dublin, Ireland
- Milan Center for Neuroscience, University of Milano-Bicocca, Milano, Italy
| | - Carlo Biffi
- Cosmo Intelligent Medical Devices, Dublin, Ireland
| |
Collapse
|
17
|
Guan H, Yap PT, Bozoki A, Liu M. Federated learning for medical image analysis: A survey. PATTERN RECOGNITION 2024; 151:110424. [PMID: 38559674 PMCID: PMC10976951 DOI: 10.1016/j.patcog.2024.110424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Machine learning in medical imaging often faces a fundamental dilemma, namely, the small sample size problem. Many recent studies suggest using multi-domain data pooled from different acquisition sites/centers to improve statistical power. However, medical images from different sites cannot be easily shared to build large datasets for model training due to privacy protection reasons. As a promising solution, federated learning, which enables collaborative training of machine learning models based on data from different sites without cross-site data sharing, has attracted considerable attention recently. In this paper, we conduct a comprehensive survey of the recent development of federated learning methods in medical image analysis. We have systematically gathered research papers on federated learning and its applications in medical image analysis published between 2017 and 2023. Our search and compilation were conducted using databases from IEEE Xplore, ACM Digital Library, Science Direct, Springer Link, Web of Science, Google Scholar, and PubMed. In this survey, we first introduce the background of federated learning for dealing with privacy protection and collaborative learning issues. We then present a comprehensive review of recent advances in federated learning methods for medical image analysis. Specifically, existing methods are categorized based on three critical aspects of a federated learning system, including client end, server end, and communication techniques. In each category, we summarize the existing federated learning methods according to specific research problems in medical image analysis and also provide insights into the motivations of different approaches. In addition, we provide a review of existing benchmark medical imaging datasets and software platforms for current federated learning research. We also conduct an experimental study to empirically evaluate typical federated learning methods for medical image analysis. This survey can help to better understand the current research status, challenges, and potential research opportunities in this promising research field.
Collapse
Affiliation(s)
- Hao Guan
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Pew-Thian Yap
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Andrea Bozoki
- Department of Neurology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Mingxia Liu
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| |
Collapse
|
18
|
Kim ES, Lee KS. Artificial intelligence in colonoscopy: from detection to diagnosis. Korean J Intern Med 2024; 39:555-562. [PMID: 38695105 PMCID: PMC11236815 DOI: 10.3904/kjim.2023.332] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Revised: 10/30/2023] [Accepted: 11/13/2023] [Indexed: 07/12/2024] Open
Abstract
This study reviews the recent progress of artificial intelligence for colonoscopy from detection to diagnosis. The source of data was 27 original studies in PubMed. The search terms were "colonoscopy" (title) and "deep learning" (abstract). The eligibility criteria were: (1) the dependent variable of gastrointestinal disease; (2) the interventions of deep learning for classification, detection and/or segmentation for colonoscopy; (3) the outcomes of accuracy, sensitivity, specificity, area under the curve (AUC), precision, F1, intersection of union (IOU), Dice and/or inference frames per second (FPS); (3) the publication year of 2021 or later; (4) the publication language of English. Based on the results of this study, different deep learning methods would be appropriate for different tasks for colonoscopy, e.g., Efficientnet with neural architecture search (AUC 99.8%) in the case of classification, You Only Look Once with the instance tracking head (F1 96.3%) in the case of detection, and Unet with dense-dilation-residual blocks (Dice 97.3%) in the case of segmentation. Their performance measures reported varied within 74.0-95.0% for accuracy, 60.0-93.0% for sensitivity, 60.0-100.0% for specificity, 71.0-99.8% for the AUC, 70.1-93.3% for precision, 81.0-96.3% for F1, 57.2-89.5% for the IOU, 75.1-97.3% for Dice and 66-182 for FPS. In conclusion, artificial intelligence provides an effective, non-invasive decision support system for colonoscopy from detection to diagnosis.
Collapse
Affiliation(s)
- Eun Sun Kim
- Department of Gastroenterology, Korea University Anam Hospital, Seoul, Korea
| | - Kwang-Sig Lee
- AI Center, Korea University Anam Hospital, Seoul, Korea
| |
Collapse
|
19
|
Biffi C, Antonelli G, Bernhofer S, Hassan C, Hirata D, Iwatate M, Maieron A, Salvagnini P, Cherubini A. REAL-Colon: A dataset for developing real-world AI applications in colonoscopy. Sci Data 2024; 11:539. [PMID: 38796533 PMCID: PMC11127922 DOI: 10.1038/s41597-024-03359-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Accepted: 05/10/2024] [Indexed: 05/28/2024] Open
Abstract
Detection and diagnosis of colon polyps are key to preventing colorectal cancer. Recent evidence suggests that AI-based computer-aided detection (CADe) and computer-aided diagnosis (CADx) systems can enhance endoscopists' performance and boost colonoscopy effectiveness. However, most available public datasets primarily consist of still images or video clips, often at a down-sampled resolution, and do not accurately represent real-world colonoscopy procedures. We introduce the REAL-Colon (Real-world multi-center Endoscopy Annotated video Library) dataset: a compilation of 2.7 M native video frames from sixty full-resolution, real-world colonoscopy recordings across multiple centers. The dataset contains 350k bounding-box annotations, each created under the supervision of expert gastroenterologists. Comprehensive patient clinical data, colonoscopy acquisition information, and polyp histopathological information are also included in each video. With its unprecedented size, quality, and heterogeneity, the REAL-Colon dataset is a unique resource for researchers and developers aiming to advance AI research in colonoscopy. Its openness and transparency facilitate rigorous and reproducible research, fostering the development and benchmarking of more accurate and reliable colonoscopy-related algorithms and models.
Collapse
Affiliation(s)
- Carlo Biffi
- Cosmo Intelligent Medical Devices, Dublin, Ireland.
| | - Giulio Antonelli
- Gastroenterology and Digestive Endoscopy Unit, Ospedale dei Castelli (N.O.C.), Rome, Italy
| | - Sebastian Bernhofer
- Karl Landsteiner University of Health Sciences, Krems, Austria
- Department of Internal Medicine 2, University Hospital St. Pölten, St. Pölten, Austria
| | - Cesare Hassan
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy
- Endoscopy Unit, Humanitas Clinical and Research Center IRCCS, Rozzano, Italy
| | - Daizen Hirata
- Gastrointestinal Center, Sano Hospital, Hyogo, Japan
| | - Mineo Iwatate
- Gastrointestinal Center, Sano Hospital, Hyogo, Japan
| | - Andreas Maieron
- Karl Landsteiner University of Health Sciences, Krems, Austria
- Department of Internal Medicine 2, University Hospital St. Pölten, St. Pölten, Austria
| | | | - Andrea Cherubini
- Cosmo Intelligent Medical Devices, Dublin, Ireland.
- Milan Center for Neuroscience, University of Milano-Bicocca, Milano, Italy.
| |
Collapse
|
20
|
Rahman R, Indris C, Bramesfeld G, Zhang T, Li K, Chen X, Grijalva I, McCornack B, Flippo D, Sharda A, Wang G. A New Dataset and Comparative Study for Aphid Cluster Detection and Segmentation in Sorghum Fields. J Imaging 2024; 10:114. [PMID: 38786568 PMCID: PMC11122648 DOI: 10.3390/jimaging10050114] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2024] [Revised: 04/24/2024] [Accepted: 05/04/2024] [Indexed: 05/25/2024] Open
Abstract
Aphid infestations are one of the primary causes of extensive damage to wheat and sorghum fields and are one of the most common vectors for plant viruses, resulting in significant agricultural yield losses. To address this problem, farmers often employ the inefficient use of harmful chemical pesticides that have negative health and environmental impacts. As a result, a large amount of pesticide is wasted on areas without significant pest infestation. This brings to attention the urgent need for an intelligent autonomous system that can locate and spray sufficiently large infestations selectively within the complex crop canopies. We have developed a large multi-scale dataset for aphid cluster detection and segmentation, collected from actual sorghum fields and meticulously annotated to include clusters of aphids. Our dataset comprises a total of 54,742 image patches, showcasing a variety of viewpoints, diverse lighting conditions, and multiple scales, highlighting its effectiveness for real-world applications. In this study, we trained and evaluated four real-time semantic segmentation models and three object detection models specifically for aphid cluster segmentation and detection. Considering the balance between accuracy and efficiency, Fast-SCNN delivered the most effective segmentation results, achieving 80.46% mean precision, 81.21% mean recall, and 91.66 frames per second (FPS). For object detection, RT-DETR exhibited the best overall performance with a 61.63% mean average precision (mAP), 92.6% mean recall, and 72.55 on an NVIDIA V100 GPU. Our experiments further indicate that aphid cluster segmentation is more suitable for assessing aphid infestations than using detection models.
Collapse
Affiliation(s)
- Raiyan Rahman
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| | - Christopher Indris
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| | - Goetz Bramesfeld
- Department of Aerospace Engineering, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| | - Tianxiao Zhang
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Kaidong Li
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Xiangyu Chen
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Ivan Grijalva
- Department of Entomology, Kansas State University, Manhattan, KS 66506-4004, USA
| | - Brian McCornack
- Department of Entomology, Kansas State University, Manhattan, KS 66506-4004, USA
| | - Daniel Flippo
- Department of Biological and Agricultural Engineering, Kansas State University, Manhattan, KS 66506, USA
| | - Ajay Sharda
- Department of Biological and Agricultural Engineering, Kansas State University, Manhattan, KS 66506, USA
| | - Guanghui Wang
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| |
Collapse
|
21
|
Hsu CM, Chen TH, Hsu CC, Wu CH, Lin CJ, Le PH, Lin CY, Kuo T. Two-stage deep-learning-based colonoscopy polyp detection incorporating fisheye and reflection correction. J Gastroenterol Hepatol 2024; 39:733-739. [PMID: 38225761 DOI: 10.1111/jgh.16470] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Revised: 11/28/2023] [Accepted: 12/14/2023] [Indexed: 01/17/2024]
Abstract
BACKGROUND AND AIM Colonoscopy is a useful method for the diagnosis and management of colorectal diseases. Many computer-aided systems have been developed to assist clinicians in detecting colorectal lesions by analyzing colonoscopy images. However, fisheye-lens distortion and light reflection in colonoscopy images can substantially affect the clarity of these images and their utility in detecting polyps. This study proposed a two-stage deep-learning model to correct distortion and reflections in colonoscopy images and thus facilitate polyp detection. METHODS Images were collected from the PolypSet dataset, the Kvasir-SEG dataset, and one medical center's patient archiving and communication system. The training, validation, and testing datasets comprised 808, 202, and 1100 images, respectively. The first stage involved the correction of fisheye-related distortion in colonoscopy images and polyp detection, which was performed using a convolutional neural network. The second stage involved the use of generative and adversarial networks for correcting reflective colonoscopy images before the convolutional neural network was used for polyp detection. RESULTS The model had higher accuracy when it was validated using corrected images than when it was validated using uncorrected images (96.8% vs 90.8%, P < 0.001). The model's accuracy in detecting polyps in the Kvasir-SEG dataset reached 96%, and the area under the receiver operating characteristic curve was 0.94. CONCLUSION The proposed model can facilitate the clinical diagnosis of colorectal polyps and improve the quality of colonoscopy.
Collapse
Affiliation(s)
- Chen-Ming Hsu
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Taoyuan Branch, Taoyuan, Taiwan
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Linkou Main Branch, Taoyuan, Taiwan
- Chang Gung University College of Medicine, Taoyuan, Taiwan
| | - Tsung-Hsing Chen
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Linkou Main Branch, Taoyuan, Taiwan
- Chang Gung University College of Medicine, Taoyuan, Taiwan
| | - Chien-Chang Hsu
- Department of Computer Science and Information Engineering, Fu Jen Catholic University, Taipei, Taiwan
| | - Che-Hao Wu
- Department of Computer Science and Information Engineering, Fu Jen Catholic University, Taipei, Taiwan
| | - Chun-Jung Lin
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Linkou Main Branch, Taoyuan, Taiwan
- Chang Gung University College of Medicine, Taoyuan, Taiwan
| | - Puo-Hsien Le
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Linkou Main Branch, Taoyuan, Taiwan
- Chang Gung University College of Medicine, Taoyuan, Taiwan
| | - Cheng-Yu Lin
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Linkou Main Branch, Taoyuan, Taiwan
| | - Tony Kuo
- Department of Gastroenterology and Hepatology, Chang Gung Memorial Hospital Linkou Main Branch, Taoyuan, Taiwan
| |
Collapse
|
22
|
Pan X, Mu Y, Ma C, He Q. TFCNet: A texture-aware and fine-grained feature compensated polyp detection network. Comput Biol Med 2024; 171:108144. [PMID: 38382386 DOI: 10.1016/j.compbiomed.2024.108144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Revised: 01/14/2024] [Accepted: 02/12/2024] [Indexed: 02/23/2024]
Abstract
PURPOSE Abnormal tissue detection is a prerequisite for medical image analysis and computer-aided diagnosis and treatment. The use of neural networks (CNN) to achieve accurate detection of intestinal polyps is beneficial to the early diagnosis and treatment of colorectal cancer. Currently, image detection models using multi-scale feature processing perform well in polyp detection. However, these methods do not fully consider the misalignment of information in the process of feature scale change, resulting in the loss of fine-grained features, and eventually cause the missed and false detection of targets. METHOD To solve this problem, a texture-aware and fine-grained feature compensated polyp detection network (TFCNet) is proposed in this paper. Firstly, design Texture Awareness Module (TAM) to excavate the rich texture information from the low-level layers and utilize high-level semantic information for background suppression, thereby capturing purer fine-grained features. Secondly, the Texture Feature Enhancement Module (TFEM) is designed to enhance the low-level texture information in TAM, and the enhanced texture features were fused with the high-level features. By making full use of the low-level texture features and multi-scale context information, the semantic consistency and integrity of the features were ensured. Finally, the Residual Pyramid Splittable Attention Module (RPSA) is designed to balance the loss of channel information caused by skip connections, and further improve the detection performance of the network. RESULTS Experimental results on 4 datasets demonstrate that the TFCNet network outperforms existing methods. Particularly, on the large dataset PolypSets, the mAP@0.5-0.95 has been improved to 88.9%. On the small datasets CVC-ClinicDB and Kvasir, the mAP@0.5-0.95 is increased by 2% and 1.6%, respectively, compared to the baseline, showcasing a significant superiority over competing methods.
Collapse
Affiliation(s)
- Xiaoying Pan
- Shanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi'an, 710121, China; School of Computer Science & Technology, Xi'an University of Post & Telecommunications, Xi'an, 710121, China.
| | - Yaya Mu
- Shanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi'an, 710121, China; School of Computer Science & Technology, Xi'an University of Post & Telecommunications, Xi'an, 710121, China
| | - Chenyang Ma
- Shanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi'an, 710121, China; School of Computer Science & Technology, Xi'an University of Post & Telecommunications, Xi'an, 710121, China
| | - Qiqi He
- Shanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi'an, 710121, China; School of Computer Science & Technology, Xi'an University of Post & Telecommunications, Xi'an, 710121, China
| |
Collapse
|
23
|
Selvaraj J, Umapathy S. CRPU-NET: a deep learning model based semantic segmentation for the detection of colorectal polyp in lower gastrointestinal tract. Biomed Phys Eng Express 2023; 10:015018. [PMID: 38100789 DOI: 10.1088/2057-1976/ad160f] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2023] [Accepted: 12/15/2023] [Indexed: 12/17/2023]
Abstract
Purpose. The objectives of the proposed work are twofold. Firstly, to develop a specialized light weight CRPU-Net for the segmentation of polyps in colonoscopy images. Secondly, to conduct a comparative analysis of the performance of CRPU-Net with implemented state-of-the-art models.Methods. We have utilized two distinct colonoscopy image datasets such as CVC-ColonDB and CVC-ClinicDB. This paper introduces the CRPU-Net, a novel approach for the automated segmentation of polyps in colorectal regions. A comprehensive series of experiments was conducted using the CRPU-Net, and its performance was compared with that of state-of-the-art models such as VGG16, VGG19, U-Net and ResUnet++. Additional analysis such as ablation study, generalizability test and 5-fold cross validation were performed.Results. The CRPU-Net achieved the segmentation accuracy of 96.42% compared to state-of-the-art model like ResUnet++ (90.91%). The Jaccard coefficient of 93.96% and Dice coefficient of 95.77% was obtained by comparing the segmentation performance of the CRPU-Net with ground truth.Conclusion. The CRPU-Net exhibits outstanding performance in Segmentation of polyp and holds promise for integration into colonoscopy devices enabling efficient operation.
Collapse
Affiliation(s)
- Jothiraj Selvaraj
- Department of Biomedical Engineering, College of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu-603203, Tamil Nadu, India
| | - Snekhalatha Umapathy
- Department of Biomedical Engineering, College of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu-603203, Tamil Nadu, India
| |
Collapse
|
24
|
Wei MYK, Zhang J, Schmidt R, Miller AS, Yeung JMC. Artificial intelligence (AI) in the management of colorectal cancer: on the horizon? ANZ J Surg 2023; 93:2052-2053. [PMID: 37489622 DOI: 10.1111/ans.18504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Accepted: 04/26/2023] [Indexed: 07/26/2023]
Affiliation(s)
- Matthew Y K Wei
- Department of Surgery, Western Precinct, University of Melbourne, Melbourne, Victoria, Australia
- Department of Colorectal Surgery, Western Health, Melbourne, Victoria, Australia
| | - Junyao Zhang
- Department of Surgery, Western Precinct, University of Melbourne, Melbourne, Victoria, Australia
| | - Reuben Schmidt
- Department of Radiology, Western Health, Melbourne, Victoria, Australia
| | - Andrew S Miller
- Department of Colorectal Surgery, Whangarei Hospital, Whangarei, New Zealand
| | - Justin M C Yeung
- Department of Surgery, Western Precinct, University of Melbourne, Melbourne, Victoria, Australia
- Department of Colorectal Surgery, Western Health, Melbourne, Victoria, Australia
| |
Collapse
|
25
|
Zhang T, Li K, Chen X, Zhong C, Luo B, Grijalva I, McCornack B, Flippo D, Sharda A, Wang G. Aphid cluster recognition and detection in the wild using deep learning models. Sci Rep 2023; 13:13410. [PMID: 37591898 PMCID: PMC10435548 DOI: 10.1038/s41598-023-38633-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2023] [Accepted: 07/12/2023] [Indexed: 08/19/2023] Open
Abstract
Aphid infestation poses a significant threat to crop production, rural communities, and global food security. While chemical pest control is crucial for maximizing yields, applying chemicals across entire fields is both environmentally unsustainable and costly. Hence, precise localization and management of aphids are essential for targeted pesticide application. The paper primarily focuses on using deep learning models for detecting aphid clusters. We propose a novel approach for estimating infection levels by detecting aphid clusters. To facilitate this research, we have captured a large-scale dataset from sorghum fields, manually selected 5447 images containing aphids, and annotated each individual aphid cluster within these images. To facilitate the use of machine learning models, we further process the images by cropping them into patches, resulting in a labeled dataset comprising 151,380 image patches. Then, we implemented and compared the performance of four state-of-the-art object detection models (VFNet, GFLV2, PAA, and ATSS) on the aphid dataset. Extensive experimental results show that all models yield stable similar performance in terms of average precision and recall. We then propose to merge close neighboring clusters and remove tiny clusters caused by cropping, and the performance is further boosted by around 17%. The study demonstrates the feasibility of automatically detecting and managing insects using machine learning models. The labeled dataset will be made openly available to the research community.
Collapse
Affiliation(s)
- Tianxiao Zhang
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS, 66045, USA
| | - Kaidong Li
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS, 66045, USA
| | - Xiangyu Chen
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS, 66045, USA
| | - Cuncong Zhong
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS, 66045, USA
| | - Bo Luo
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS, 66045, USA
| | - Ivan Grijalva
- Department of Entomology, Kansas State University, Manhattan, KS, 66506, USA
| | - Brian McCornack
- Department of Entomology, Kansas State University, Manhattan, KS, 66506, USA
| | - Daniel Flippo
- Department of Biological and Agricultural Engineering, Kansas State University, Manhattan, KS, 66506, USA
| | - Ajay Sharda
- Department of Biological and Agricultural Engineering, Kansas State University, Manhattan, KS, 66506, USA
| | - Guanghui Wang
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON, M5B 2K3, Canada.
| |
Collapse
|
26
|
Zhang T, Bur AM, Kraft S, Kavookjian H, Renslo B, Chen X, Luo B, Wang G. Gender, Smoking History, and Age Prediction from Laryngeal Images. J Imaging 2023; 9:109. [PMID: 37367457 DOI: 10.3390/jimaging9060109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2023] [Revised: 05/22/2023] [Accepted: 05/25/2023] [Indexed: 06/28/2023] Open
Abstract
Flexible laryngoscopy is commonly performed by otolaryngologists to detect laryngeal diseases and to recognize potentially malignant lesions. Recently, researchers have introduced machine learning techniques to facilitate automated diagnosis using laryngeal images and achieved promising results. The diagnostic performance can be improved when patients' demographic information is incorporated into models. However, the manual entry of patient data is time-consuming for clinicians. In this study, we made the first endeavor to employ deep learning models to predict patient demographic information to improve the detector model's performance. The overall accuracy for gender, smoking history, and age was 85.5%, 65.2%, and 75.9%, respectively. We also created a new laryngoscopic image set for the machine learning study and benchmarked the performance of eight classical deep learning models based on CNNs and Transformers. The results can be integrated into current learning models to improve their performance by incorporating the patient's demographic information.
Collapse
Affiliation(s)
- Tianxiao Zhang
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Andrés M Bur
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Shannon Kraft
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Hannah Kavookjian
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Bryan Renslo
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Xiangyu Chen
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Bo Luo
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Guanghui Wang
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| |
Collapse
|
27
|
Lo CM, Yang YW, Lin JK, Lin TC, Chen WS, Yang SH, Chang SC, Wang HS, Lan YT, Lin HH, Huang SC, Cheng HH, Jiang JK, Lin CC. Modeling the survival of colorectal cancer patients based on colonoscopic features in a feature ensemble vision transformer. Comput Med Imaging Graph 2023; 107:102242. [PMID: 37172354 DOI: 10.1016/j.compmedimag.2023.102242] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2023] [Revised: 05/05/2023] [Accepted: 05/07/2023] [Indexed: 05/14/2023]
Abstract
The prognosis of patients with colorectal cancer (CRC) mostly relies on the classic tumor node metastasis (TNM) staging classification. A more accurate and convenient prediction model would provide a better prognosis and assist in treatment. From May 2014 to December 2017, patients who underwent an operation for CRC were enrolled. The proposed feature ensemble vision transformer (FEViT) used ensemble classifiers to benefit the combinations of relevant colonoscopy features from the pretrained vision transformer and clinical features, including sex, age, family history of CRC, and tumor location, to establish the prognostic model. A total of 1729 colonoscopy images were enrolled in the current retrospective study. For the prediction of patient survival, FEViT achieved an accuracy of 94 % with an area under the receiver operating characteristic curve of 0.93, which was better than the TNM staging classification (90 %, 0.83) in the experiment. FEViT reduced the limited receptive field and gradient disappearance in the conventional convolutional neural network and was a relatively effective and efficient procedure. The promising accuracy of FEViT in modeling survival makes the prognosis of CRC patients more predictable and practical.
Collapse
Affiliation(s)
- Chung-Ming Lo
- Graduate Institute of Library, Information and Archival Studies, National Chengchi University, Taipei, Taiwan
| | - Yi-Wen Yang
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Jen-Kou Lin
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Tzu-Chen Lin
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Wei-Shone Chen
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Shung-Haur Yang
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Department of Surgery, National Yang Ming Chiao Tung University Hospital, Yilan, Taiwan
| | - Shih-Ching Chang
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Huann-Sheng Wang
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Yuan-Tzu Lan
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Hung-Hsin Lin
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Sheng-Chieh Huang
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Hou-Hsuan Cheng
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Jeng-Kai Jiang
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Chun-Chi Lin
- Division of Colon and Rectal Surgery, Department of Surgery, Taipei Veterans General Hospital, Taipei, Taiwan; Department of Surgery, School of Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan.
| |
Collapse
|
28
|
Nogueira-Rodríguez A, Glez-Peña D, Reboiro-Jato M, López-Fernández H. Negative Samples for Improving Object Detection-A Case Study in AI-Assisted Colonoscopy for Polyp Detection. Diagnostics (Basel) 2023; 13:diagnostics13050966. [PMID: 36900110 PMCID: PMC10001273 DOI: 10.3390/diagnostics13050966] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2023] [Accepted: 03/01/2023] [Indexed: 03/08/2023] Open
Abstract
Deep learning object-detection models are being successfully applied to develop computer-aided diagnosis systems for aiding polyp detection during colonoscopies. Here, we evidence the need to include negative samples for both (i) reducing false positives during the polyp-finding phase, by including images with artifacts that may confuse the detection models (e.g., medical instruments, water jets, feces, blood, excessive proximity of the camera to the colon wall, blurred images, etc.) that are usually not included in model development datasets, and (ii) correctly estimating a more realistic performance of the models. By retraining our previously developed YOLOv3-based detection model with a dataset that includes 15% of additional not-polyp images with a variety of artifacts, we were able to generally improve its F1 performance in our internal test datasets (from an average F1 of 0.869 to 0.893), which now include such type of images, as well as in four public datasets that include not-polyp images (from an average F1 of 0.695 to 0.722).
Collapse
Affiliation(s)
- Alba Nogueira-Rodríguez
- CINBIO, Department of Computer Science, ESEI—Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
- Correspondence:
| | - Daniel Glez-Peña
- CINBIO, Department of Computer Science, ESEI—Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| | - Miguel Reboiro-Jato
- CINBIO, Department of Computer Science, ESEI—Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| | - Hugo López-Fernández
- CINBIO, Department of Computer Science, ESEI—Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| |
Collapse
|
29
|
Mansur A, Saleem Z, Elhakim T, Daye D. Role of artificial intelligence in risk prediction, prognostication, and therapy response assessment in colorectal cancer: current state and future directions. Front Oncol 2023; 13:1065402. [PMID: 36761957 PMCID: PMC9905815 DOI: 10.3389/fonc.2023.1065402] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2022] [Accepted: 01/09/2023] [Indexed: 01/26/2023] Open
Abstract
Artificial Intelligence (AI) is a branch of computer science that utilizes optimization, probabilistic and statistical approaches to analyze and make predictions based on a vast amount of data. In recent years, AI has revolutionized the field of oncology and spearheaded novel approaches in the management of various cancers, including colorectal cancer (CRC). Notably, the applications of AI to diagnose, prognosticate, and predict response to therapy in CRC, is gaining traction and proving to be promising. There have also been several advancements in AI technologies to help predict metastases in CRC and in Computer-Aided Detection (CAD) Systems to improve miss rates for colorectal neoplasia. This article provides a comprehensive review of the role of AI in predicting risk, prognosis, and response to therapies among patients with CRC.
Collapse
Affiliation(s)
- Arian Mansur
- Harvard Medical School, Boston, MA, United States
| | | | - Tarig Elhakim
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
| | - Dania Daye
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
| |
Collapse
|
30
|
Hsu CM, Hsu CC, Hsu ZM, Chen TH, Kuo T. Intraprocedure Artificial Intelligence Alert System for Colonoscopy Examination. SENSORS (BASEL, SWITZERLAND) 2023; 23:1211. [PMID: 36772251 PMCID: PMC9921893 DOI: 10.3390/s23031211] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/11/2022] [Revised: 01/13/2023] [Accepted: 01/18/2023] [Indexed: 06/18/2023]
Abstract
Colonoscopy is a valuable tool for preventing and reducing the incidence and mortality of colorectal cancer. Although several computer-aided colorectal polyp detection and diagnosis systems have been proposed for clinical application, many remain susceptible to interference problems, including low image clarity, unevenness, and low accuracy for the analysis of dynamic images; these drawbacks affect the robustness and practicality of these systems. This study proposed an intraprocedure alert system for colonoscopy examination developed on the basis of deep learning. The proposed system features blurred image detection, foreign body detection, and polyp detection modules facilitated by convolutional neural networks. The training and validation datasets included high-quality images and low-quality images, including blurred images and those containing folds, fecal matter, and opaque water. For the detection of blurred images and images containing folds, fecal matter, and opaque water, the accuracy rate was 96.2%. Furthermore, the study results indicated a per-polyp detection accuracy of 100% when the system was applied to video images. The recall rates for high-quality image frames and polyp image frames were 95.7% and 92%, respectively. The overall alert accuracy rate and the false-positive rate of low quality for video images obtained through per-frame analysis were 95.3% and 0.18%, respectively. The proposed system can be used to alert colonoscopists to the need to slow their procedural speed or to perform flush or lumen inflation in cases where the colonoscope is being moved too rapidly, where fecal residue is present in the intestinal tract, or where the colon has been inadequately distended.
Collapse
Affiliation(s)
- Chen-Ming Hsu
- Department of Gastroenterology and Hepatology, Taoyuan Chang Gung Memorial Hospital, Taoyuan 333, Taiwan
- Department of Gastroenterology and Hepatology, Linkou Chang Gung Memorial Hospital, Taoyuan 333, Taiwan
- College of Medicine, Chang Gung University, Taoyuan 333, Taiwan
| | - Chien-Chang Hsu
- Department of Computer Science and Information Engineering, Fu-Jen Catholic University, New Taipei 242, Taiwan
| | - Zhe-Ming Hsu
- Department of Computer Science and Information Engineering, Fu-Jen Catholic University, New Taipei 242, Taiwan
| | - Tsung-Hsing Chen
- Department of Gastroenterology and Hepatology, Linkou Chang Gung Memorial Hospital, Taoyuan 333, Taiwan
- College of Medicine, Chang Gung University, Taoyuan 333, Taiwan
| | - Tony Kuo
- Department of Gastroenterology and Hepatology, Linkou Chang Gung Memorial Hospital, Taoyuan 333, Taiwan
| |
Collapse
|
31
|
Cui R, Yang R, Liu F, Cai C. N-Net: Lesion region segmentations using the generalized hybrid dilated convolutions for polyps in colonoscopy images. Front Bioeng Biotechnol 2022; 10:963590. [DOI: 10.3389/fbioe.2022.963590] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2022] [Accepted: 08/12/2022] [Indexed: 11/13/2022] Open
Abstract
Colorectal cancer is the cancer with the second highest and the third highest incidence rates for the female and the male, respectively. Colorectal polyps are potential prognostic indicators of colorectal cancer, and colonoscopy is the gold standard for the biopsy and the removal of colorectal polyps. In this scenario, one of the main concerns is to ensure the accuracy of lesion region identifications. However, the missing rate of polyps through manual observations in colonoscopy can reach 14%–30%. In this paper, we focus on the identifications of polyps in clinical colonoscopy images and propose a new N-shaped deep neural network (N-Net) structure to conduct the lesion region segmentations. The encoder-decoder framework is adopted in the N-Net structure and the DenseNet modules are implemented in the encoding path of the network. Moreover, we innovatively propose the strategy to design the generalized hybrid dilated convolution (GHDC), which enables flexible dilated rates and convolutional kernel sizes, to facilitate the transmission of the multi-scale information with the respective fields expanded. Based on the strategy of GHDC designing, we design four GHDC blocks to connect the encoding and the decoding paths. Through the experiments on two publicly available datasets on polyp segmentations of colonoscopy images: the Kvasir-SEG dataset and the CVC-ClinicDB dataset, the rationality and superiority of the proposed GHDC blocks and the proposed N-Net are verified. Through the comparative studies with the state-of-the-art methods, such as TransU-Net, DeepLabV3+ and CA-Net, we show that even with a small amount of network parameters, the N-Net outperforms with the Dice of 94.45%, the average symmetric surface distance (ASSD) of 0.38 pix and the mean intersection-over-union (mIoU) of 89.80% on the Kvasir-SEG dataset, and with the Dice of 97.03%, the ASSD of 0.16 pix and the mIoU of 94.35% on the CVC-ClinicDB dataset.
Collapse
|
32
|
Zhang T, Luo B, Sharda A, Wang G. Dynamic Label Assignment for Object Detection by Combining Predicted IoUs and Anchor IoUs. J Imaging 2022; 8:193. [PMID: 35877638 PMCID: PMC9322857 DOI: 10.3390/jimaging8070193] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2022] [Revised: 07/06/2022] [Accepted: 07/07/2022] [Indexed: 02/01/2023] Open
Abstract
Label assignment plays a significant role in modern object detection models. Detection models may yield totally different performances with different label assignment strategies. For anchor-based detection models, the IoU (Intersection over Union) threshold between the anchors and their corresponding ground truth bounding boxes is the key element since the positive samples and negative samples are divided by the IoU threshold. Early object detectors simply utilize the fixed threshold for all training samples, while recent detection algorithms focus on adaptive thresholds based on the distribution of the IoUs to the ground truth boxes. In this paper, we introduce a simple while effective approach to perform label assignment dynamically based on the training status with predictions. By introducing the predictions in label assignment, more high-quality samples with higher IoUs to the ground truth objects are selected as the positive samples, which could reduce the discrepancy between the classification scores and the IoU scores, and generate more high-quality boundary boxes. Our approach shows improvements in the performance of the detection models with the adaptive label assignment algorithm and lower bounding box losses for those positive samples, indicating more samples with higher-quality predicted boxes are selected as positives.
Collapse
Affiliation(s)
- Tianxiao Zhang
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA; (T.Z.); (B.L.)
| | - Bo Luo
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA; (T.Z.); (B.L.)
| | - Ajay Sharda
- Department of Biological and Agricultural Engineering, Kansas State University, Manhattan, KS 66506, USA;
| | - Guanghui Wang
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| |
Collapse
|
33
|
Cheng YW, Li YC. Examining the Factors That Affect the Diagnosis of Patients with Positive Fecal Occult Blood Test Results. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:ijerph19137569. [PMID: 35805251 PMCID: PMC9265584 DOI: 10.3390/ijerph19137569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/21/2022] [Revised: 06/02/2022] [Accepted: 06/20/2022] [Indexed: 11/17/2022]
Abstract
Due to the threat of colorectal cancer (CRC) to health, Taiwan included the fecal occult blood test (FOBT) under preventive health services in 2010. We examined the factors that affect the diagnosis of people with positive FOBT results. Data were retrospectively collected from the CRC screening database. In the model predicting factors that affect the diagnosis of 89,046 people with positive FOBT results, the risks of disease in the CRC group were lower in medical institutions that conducted follow-up examinations in regions such as Northern Taiwan compared to that in Eastern Taiwan (p = 0.013); they were lower in the age group of 50 to 65 years than those in the age group of 71 to 75 years (p < 0.001, p = 0.016), and lower in the outpatient medical units that conducted follow-up examinations than those in the inpatient medical units by 0.565 times (p < 0.001, 95% CI: 0.493−0.647). Factors affecting the diagnosis of patients with positive FOBT results were gender, the region of the medical institution, medical unit for follow-up examinations, age, screening site, family history, type of follow-up examinations, and follow-up time. Therefore, the identification of characteristics of patients with positive FOBT results and the promotion of follow-up examination are important prevention strategies for CRC.
Collapse
Affiliation(s)
- Yin-Wen Cheng
- Department of Business Management, College of Management, National Sun Yat-Sen University, No. 70, Lien-Hai Rd., Gushan Dist., Kaohsiung 80424, Taiwan;
| | - Ying-Chun Li
- Institute of Health Care Management, National Sun Yat-Sen University, No. 70, Lien-Hai Rd., Gushan Dist., Kaohsiung 80424, Taiwan
- Correspondence: ; Tel.: +886-7-5252000 (ext. 4875)
| |
Collapse
|
34
|
Nogueira-Rodríguez A, Reboiro-Jato M, Glez-Peña D, López-Fernández H. Performance of Convolutional Neural Networks for Polyp Localization on Public Colonoscopy Image Datasets. Diagnostics (Basel) 2022; 12:898. [PMID: 35453946 PMCID: PMC9027927 DOI: 10.3390/diagnostics12040898] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2022] [Revised: 03/31/2022] [Accepted: 04/01/2022] [Indexed: 01/10/2023] Open
Abstract
Colorectal cancer is one of the most frequent malignancies. Colonoscopy is the de facto standard for precancerous lesion detection in the colon, i.e., polyps, during screening studies or after facultative recommendation. In recent years, artificial intelligence, and especially deep learning techniques such as convolutional neural networks, have been applied to polyp detection and localization in order to develop real-time CADe systems. However, the performance of machine learning models is very sensitive to changes in the nature of the testing instances, especially when trying to reproduce results for totally different datasets to those used for model development, i.e., inter-dataset testing. Here, we report the results of testing of our previously published polyp detection model using ten public colonoscopy image datasets and analyze them in the context of the results of other 20 state-of-the-art publications using the same datasets. The F1-score of our recently published model was 0.88 when evaluated on a private test partition, i.e., intra-dataset testing, but it decayed, on average, by 13.65% when tested on ten public datasets. In the published research, the average intra-dataset F1-score is 0.91, and we observed that it also decays in the inter-dataset setting to an average F1-score of 0.83.
Collapse
Affiliation(s)
- Alba Nogueira-Rodríguez
- CINBIO, Department of Computer Science, ESEI-Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain; (A.N.-R.); (M.R.-J.); (D.G.-P.)
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| | - Miguel Reboiro-Jato
- CINBIO, Department of Computer Science, ESEI-Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain; (A.N.-R.); (M.R.-J.); (D.G.-P.)
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| | - Daniel Glez-Peña
- CINBIO, Department of Computer Science, ESEI-Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain; (A.N.-R.); (M.R.-J.); (D.G.-P.)
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| | - Hugo López-Fernández
- CINBIO, Department of Computer Science, ESEI-Escuela Superior de Ingeniería Informática, Universidade de Vigo, 32004 Ourense, Spain; (A.N.-R.); (M.R.-J.); (D.G.-P.)
- SING Research Group, Galicia Sur Health Research Institute (IIS Galicia Sur), SERGAS-UVIGO, 36213 Vigo, Spain
| |
Collapse
|
35
|
Yang Y, Zhang T, Li G, Kim T, Wang G. An unsupervised domain adaptation model based on dual-module adversarial training. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.12.060] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
36
|
Xu W, Wang G. A Domain Gap Aware Generative Adversarial Network for Multi-Domain Image Translation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 31:72-84. [PMID: 34762587 DOI: 10.1109/tip.2021.3125266] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Recent image-to-image translation models have shown great success in mapping local textures between two domains. Existing approaches rely on a cycle-consistency constraint that supervises the generators to learn an inverse mapping. However, learning the inverse mapping introduces extra trainable parameters and it is unable to learn the inverse mapping for some domains. As a result, they are ineffective in the scenarios where (i) multiple visual image domains are involved; (ii) both structure and texture transformations are required; and (iii) semantic consistency is preserved. To solve these challenges, the paper proposes a unified model to translate images across multiple domains with significant domain gaps. Unlike previous models that constrain the generators with the ubiquitous cycle-consistency constraint to achieve the content similarity, the proposed model employs a perceptual self-regularization constraint. With a single unified generator, the model can maintain consistency over the global shapes as well as the local texture information across multiple domains. Extensive qualitative and quantitative evaluations demonstrate the effectiveness and superior performance over state-of-the-art models. It is more effective in representing shape deformation in challenging mappings with significant dataset variation across multiple domains.
Collapse
|
37
|
Deep Learning Approaches to Colorectal Cancer Diagnosis: A Review. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app112210982] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Abstract
Unprecedented breakthroughs in the development of graphical processing systems have led to great potential for deep learning (DL) algorithms in analyzing visual anatomy from high-resolution medical images. Recently, in digital pathology, the use of DL technologies has drawn a substantial amount of attention for use in the effective diagnosis of various cancer types, especially colorectal cancer (CRC), which is regarded as one of the dominant causes of cancer-related deaths worldwide. This review provides an in-depth perspective on recently published research articles on DL-based CRC diagnosis and prognosis. Overall, we provide a retrospective synopsis of simple image-processing-based and machine learning (ML)-based computer-aided diagnosis (CAD) systems, followed by a comprehensive appraisal of use cases with different types of state-of-the-art DL algorithms for detecting malignancies. We first list multiple standardized and publicly available CRC datasets from two imaging types: colonoscopy and histopathology. Secondly, we categorize the studies based on the different types of CRC detected (tumor tissue, microsatellite instability, and polyps), and we assess the data preprocessing steps and the adopted DL architectures before presenting the optimum diagnostic results. CRC diagnosis with DL algorithms is still in the preclinical phase, and therefore, we point out some open issues and provide some insights into the practicability and development of robust diagnostic systems in future health care and oncology.
Collapse
|
38
|
Patel K, Bur AM, Wang G. Enhanced U-Net: A Feature Enhancement Network for Polyp Segmentation. PROCEEDINGS OF THE INTERNATIONAL ROBOTS & VISION CONFERENCE. INTERNATIONAL ROBOTS & VISION CONFERENCE 2021; 2021:181-188. [PMID: 34368816 PMCID: PMC8341462 DOI: 10.1109/crv52889.2021.00032] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Colonoscopy is a procedure to detect colorectal polyps which are the primary cause for developing colorectal cancer. However, polyp segmentation is a challenging task due to the diverse shape, size, color, and texture of polyps, shuttle difference between polyp and its background, as well as low contrast of the colonoscopic images. To address these challenges, we propose a feature enhancement network for accurate polyp segmentation in colonoscopy images. Specifically, the proposed network enhances the semantic information using the novel Semantic Feature Enhance Module (SFEM). Furthermore, instead of directly adding encoder features to the respective decoder layer, we introduce an Adaptive Global Context Module (AGCM), which focuses only on the encoder's significant and hard fine-grained features. The integration of these two modules improves the quality of features layer by layer, which in turn enhances the final feature representation. The proposed approach is evaluated on five colonoscopy datasets and demonstrates superior performance compared to other state-of-the-art models.
Collapse
Affiliation(s)
- Krushi Patel
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence KS, USA, 66045
| | - Andrés M Bur
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas, Kansas City, Kansas, USA, 66160
| | - Guanghui Wang
- Department of Computer Science, Ryerson University, Toronto ON, Canada, M5B 2K3
| |
Collapse
|