1
|
Hossain T, Shamrat FMJM, Zhou X, Mahmud I, Mazumder MSA, Sharmin S, Gururajan R. Development of a multi-fusion convolutional neural network (MF-CNN) for enhanced gastrointestinal disease diagnosis in endoscopy image analysis. PeerJ Comput Sci 2024; 10:e1950. [PMID: 38660192 PMCID: PMC11041948 DOI: 10.7717/peerj-cs.1950] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Accepted: 02/29/2024] [Indexed: 04/26/2024]
Abstract
Gastrointestinal (GI) diseases are prevalent medical conditions that require accurate and timely diagnosis for effective treatment. To address this, we developed the Multi-Fusion Convolutional Neural Network (MF-CNN), a deep learning framework that strategically integrates and adapts elements from six deep learning models, enhancing feature extraction and classification of GI diseases from endoscopic images. The MF-CNN architecture leverages truncated and partially frozen layers from existing models, augmented with novel components such as Auxiliary Fusing Layers (AuxFL), Fusion Residual Block (FuRB), and Alpha Dropouts (αDO) to improve precision and robustness. This design facilitates the precise identification of conditions such as ulcerative colitis, polyps, esophagitis, and healthy colons. Our methodology involved preprocessing endoscopic images sourced from open databases, including KVASIR and ETIS-Larib Polyp DB, using adaptive histogram equalization (AHE) to enhance their quality. The MF-CNN framework supports detailed feature mapping for improved interpretability of the model's internal workings. An ablation study was conducted to validate the contribution of each component, demonstrating that the integration of AuxFL, αDO, and FuRB played a crucial part in reducing overfitting and efficiency saturation and enhancing overall model performance. The MF-CNN demonstrated outstanding performance in terms of efficacy, achieving an accuracy rate of 99.25%. It also excelled in other key performance metrics with a precision of 99.27%, a recall of 99.25%, and an F1-score of 99.25%. These metrics confirmed the model's proficiency in accurate classification and its capability to minimize false positives and negatives across all tested GI disease categories. Furthermore, the AUC values were exceptional, averaging 1.00 for both test and validation sets, indicating perfect discriminative ability. The findings of the P-R curve analysis and confusion matrix further confirmed the robust classification performance of the MF-CNN. This research introduces a technique for medical imaging that can potentially transform diagnostics in gastrointestinal healthcare facilities worldwide.
Collapse
Affiliation(s)
- Tanzim Hossain
- Department of Software Engineering, Daffodil International University, Dhaka, Bangladesh
| | | | - Xujuan Zhou
- School of Business, University of Southern Queensland, Springfield, Australia
| | - Imran Mahmud
- Department of Software Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Md. Sakib Ali Mazumder
- Department of Software Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Sharmin Sharmin
- Department of Computer System and Technology, University of Malaya, Kuala Lumpur, Malaysia
| | - Raj Gururajan
- School of Business, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
2
|
Massironi S. Advancements in Barrett's esophagus detection: The role of artificial intelligence and its implications. World J Gastroenterol 2024; 30:1494-1496. [PMID: 38617459 PMCID: PMC11008413 DOI: 10.3748/wjg.v30.i11.1494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/07/2024] [Revised: 01/27/2024] [Accepted: 02/27/2024] [Indexed: 03/21/2024] Open
Abstract
Artificial intelligence (AI) is making significant strides in revolutionizing the detection of Barrett's esophagus (BE), a precursor to esophageal adenocarcinoma. In the research article by Tsai et al, researchers utilized endoscopic images to train an AI model, challenging the traditional distinction between endoscopic and histological BE. This approach yielded remarkable results, with the AI system achieving an accuracy of 94.37%, sensitivity of 94.29%, and specificity of 94.44%. The study's extensive dataset enhances the AI model's practicality, offering valuable support to endoscopists by minimizing unnecessary biopsies. However, questions about the applicability to different endoscopic systems remain. The study underscores the potential of AI in BE detection while highlighting the need for further research to assess its adaptability to diverse clinical settings.
Collapse
Affiliation(s)
- Sara Massironi
- Division of Gastroenterology and Center for Autoimmune Liver Diseases, Department of Medicine and Surgery, Fondazione IRCCS San Gerardo dei Tintori, University of Milano-Bicocca, Monza 20900, Italy
| |
Collapse
|
3
|
Kiziloluk S, Yildirim M, Bingol H, Alatas B. Multi-feature fusion and dandelion optimizer based model for automatically diagnosing the gastrointestinal diseases. PeerJ Comput Sci 2024; 10:e1919. [PMID: 38435605 PMCID: PMC10909187 DOI: 10.7717/peerj-cs.1919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2023] [Accepted: 02/12/2024] [Indexed: 03/05/2024]
Abstract
It is a known fact that gastrointestinal diseases are extremely common among the public. The most common of these diseases are gastritis, reflux, and dyspepsia. Since the symptoms of these diseases are similar, diagnosis can often be confused. Therefore, it is of great importance to make these diagnoses faster and more accurate by using computer-aided systems. Therefore, in this article, a new artificial intelligence-based hybrid method was developed to classify images with high accuracy of anatomical landmarks that cause gastrointestinal diseases, pathological findings and polyps removed during endoscopy, which usually cause cancer. In the proposed method, firstly trained InceptionV3 and MobileNetV2 architectures are used and feature extraction is performed with these two architectures. Then, the features obtained from InceptionV3 and MobileNetV2 architectures are merged. Thanks to this merging process, different features belonging to the same images were brought together. However, these features contain irrelevant and redundant features that may have a negative impact on classification performance. Therefore, Dandelion Optimizer (DO), one of the most recent metaheuristic optimization algorithms, was used as a feature selector to select the appropriate features to improve the classification performance and support vector machine (SVM) was used as a classifier. In the experimental study, the proposed method was also compared with different convolutional neural network (CNN) models and it was found that the proposed method achieved better results. The accuracy value obtained in the proposed model is 93.88%.
Collapse
Affiliation(s)
- Soner Kiziloluk
- Computer Engineering, Malatya Turgut Ozal University, Malatya, Turkey
| | - Muhammed Yildirim
- Computer Engineering, Malatya Turgut Ozal University, Malatya, Turkey
| | - Harun Bingol
- Software Engineering, Malatya Turgut Ozal University, Malatya, Turkey
| | - Bilal Alatas
- Software Engineering, Firat (Euphrates) University, Elazig, Turkey
| |
Collapse
|
4
|
Yin M, Zhang R, Lin J, Zhu S, Liu L, Liu X, Lu J, Xu C, Zhu J. Identification of gastric signet ring cell carcinoma based on endoscopic images using few-shot learning. Dig Liver Dis 2023; 55:1725-1734. [PMID: 37455154 DOI: 10.1016/j.dld.2023.07.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/06/2023] [Revised: 07/05/2023] [Accepted: 07/07/2023] [Indexed: 07/18/2023]
Abstract
BACKGROUND Deep learning (DL) models perform poorly when there are limited gastric signet ring cell carcinoma (SRCC) samples. Few-shot learning (FSL) can address the small sample problem. METHODS EfficientNetV2-S was first pretrained on ImageNet and then pretrained on esophageal endoscopic images (i.e., base classes: normal vs. early cancer vs. advanced cancer) using transfer learning. Second, images of gastric benign ulcers, adenocarcinoma and SRCC, i.e., novel classes (n = 50 per class), were included. Image features were extracted as vectors using the dual pretrained EfficientNetV2-S. Finally, a k-nearest neighbor classifier was used to identify SRCC. The above proposed 3-way 3-shot FSL framework was conducted in three rounds. RESULTS Dual pretrained FSL performed better than single pretrained FSL, endoscopists and traditional EfficientNetV2-S models. Dual pretrained FSL obtained the highest accuracy (79.4%), sensitivity (68.8%), recall (68.8%), precision (69.3%) and F1-score (0.691), and the senior endoscopist achieved the highest specificity of 93.6% when identifying SRCC. The macro-AUC and F1-score for dual pretraining (0.763 and 0.703, respectively) were higher than those for single pretraining (0.656 and 0.537, respectively), along with endoscopists and traditional EfficientNetV2-S models. The 2-way 3-shot FSL also performed better. CONCLUSION The proposed FSL framework showed practical performance in the differentiation of SRCC on endoscopic images, suggesting the potential of FSL in the computer-aided diagnosis for rare diseases.
Collapse
Affiliation(s)
- Minyue Yin
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Rufa Zhang
- Department of Gastroenterology, Changshu Hospital Affiliated to Soochow University, Changshu No.1 People's Hospital, Suzhou, Jiangsu, 215500, China
| | - Jiaxi Lin
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Shiqi Zhu
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Lu Liu
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Xiaolin Liu
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Jianying Lu
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Chunfang Xu
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China
| | - Jinzhou Zhu
- Department of Gastroenterology, the First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215006, China; Suzhou Clinical Center of Digestive Diseases, Suzhou, Jiangsu, 215006, China.
| |
Collapse
|
5
|
Wang KN, Zhuang S, Ran QY, Zhou P, Hua J, Zhou GQ, He X. DLGNet: A dual-branch lesion-aware network with the supervised Gaussian Mixture model for colon lesions classification in colonoscopy images. Med Image Anal 2023; 87:102832. [PMID: 37148864 DOI: 10.1016/j.media.2023.102832] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Revised: 01/20/2023] [Accepted: 04/20/2023] [Indexed: 05/08/2023]
Abstract
Colorectal cancer is one of the malignant tumors with the highest mortality due to the lack of obvious early symptoms. It is usually in the advanced stage when it is discovered. Thus the automatic and accurate classification of early colon lesions is of great significance for clinically estimating the status of colon lesions and formulating appropriate diagnostic programs. However, it is challenging to classify full-stage colon lesions due to the large inter-class similarities and intra-class differences of the images. In this work, we propose a novel dual-branch lesion-aware neural network (DLGNet) to classify intestinal lesions by exploring the intrinsic relationship between diseases, composed of four modules: lesion location module, dual-branch classification module, attention guidance module, and inter-class Gaussian loss function. Specifically, the elaborate dual-branch module integrates the original image and the lesion patch obtained by the lesion localization module to explore and interact with lesion-specific features from a global and local perspective. Also, the feature-guided module guides the model to pay attention to the disease-specific features by learning remote dependencies through spatial and channel attention after network feature learning. Finally, the inter-class Gaussian loss function is proposed, which assumes that each feature extracted by the network is an independent Gaussian distribution, and the inter-class clustering is more compact, thereby improving the discriminative ability of the network. The extensive experiments on the collected 2568 colonoscopy images have an average accuracy of 91.50%, and the proposed method surpasses the state-of-the-art methods. This study is the first time that colon lesions are classified at each stage and achieves promising colon disease classification performance. To motivate the community, we have made our code publicly available via https://github.com/soleilssss/DLGNet.
Collapse
Affiliation(s)
- Kai-Ni Wang
- School of Biological Science and Medical Engineering, Southeast University, Nanjing, China; State Key Laboratory of Digital Medical Engineering, Southeast University, Nanjing, China; Jiangsu Key Laboratory of Biomaterials and Devices, Southeast University, Nanjing, China
| | - Shuaishuai Zhuang
- The First Affiliated Hospital of Nanjing Medical University, Nanjing, China
| | - Qi-Yong Ran
- School of Biological Science and Medical Engineering, Southeast University, Nanjing, China; State Key Laboratory of Digital Medical Engineering, Southeast University, Nanjing, China; Jiangsu Key Laboratory of Biomaterials and Devices, Southeast University, Nanjing, China
| | - Ping Zhou
- School of Biological Science and Medical Engineering, Southeast University, Nanjing, China; State Key Laboratory of Digital Medical Engineering, Southeast University, Nanjing, China; Jiangsu Key Laboratory of Biomaterials and Devices, Southeast University, Nanjing, China
| | - Jie Hua
- The First Affiliated Hospital of Nanjing Medical University, Nanjing, China; Liyang People's Hospital, Liyang Branch Hospital of Jiangsu Province Hospital, Liyang, China
| | - Guang-Quan Zhou
- School of Biological Science and Medical Engineering, Southeast University, Nanjing, China; State Key Laboratory of Digital Medical Engineering, Southeast University, Nanjing, China; Jiangsu Key Laboratory of Biomaterials and Devices, Southeast University, Nanjing, China.
| | - Xiaopu He
- The First Affiliated Hospital of Nanjing Medical University, Nanjing, China.
| |
Collapse
|
6
|
Chu Y, Li H, Li X, Ding Y, Yang X, Ai D, Chen X, Wang Y, Yang J. Endoscopic image feature matching via motion consensus and global bilateral regression. Comput Methods Programs Biomed 2020; 190:105370. [PMID: 32036206 DOI: 10.1016/j.cmpb.2020.105370] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/24/2019] [Revised: 12/17/2019] [Accepted: 01/26/2020] [Indexed: 06/10/2023]
Abstract
BACKGROUND AND OBJECTIVE Feature matching of endoscopic images is of crucial importance in many clinical applications, such as object tracking and surface reconstruction. However, with the presence of low texture, specular reflections and deformations, the feature matching methods of natural scene are facing great challenges in minimally invasive surgery (MIS) scenarios. We propose a novel motion consensus-based method for endoscopic image feature matching to address these problems. METHODS Our method starts by correcting the radial distortion with the spherical projection model and removing the specular reflection regions with an adaptive detection method, which helps to eliminate the image distortion and to reduce the quantity of outliers. We solve the matching problem with a two-stage strategy that progressively estimates a consensus of inliers; the result is a precisely smoothed motion field. First, we construct a spatial motion field from candidate feature matches and estimate its maximum posterior with expectation maximization algorithm, which is computationally efficient and able to obtain smoothed motion field quickly. Second, we extend the smoothed motion field to the affine domain and refine it with bilateral regression to preserve locally subtle motions. The true matches can be identified by checking the difference of feature motion against the estimated field. RESULTS Evaluations are implemented on two simulation datasets of deformation (218 images) and four different types of endoscopic datasets (1032 images). Our method is compared with three other state-of-the-art methods and achieves the best performance on affine transformation and nonrigid deformation simulations, with inlier ratio of 86.7% and 94.3%, sensitivity of 90.0% and 96.2%, precision of 88.2% and 93.9%, and F1-Score of 89.1% and 95.0%, respectively. On clinical datasets evaluations, the proposed method achieves an average reprojection error of 3.7 pixels and a consistent performance in multi-image correspondence of sequential images. Furthermore, we also present a surface reconstruction result from rhinoscopic images to validate the reliability of our method, which shows high-quality feature matching results. CONCLUSIONS The proposed motion consensus-based feature matching method is proved effective and robust for endoscopic images correspondence. This demonstrates its capability to generate reliable feature matches for surface reconstruction and other meaningful applications in MIS scenarios.
Collapse
Affiliation(s)
- Yakui Chu
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
| | - Heng Li
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China.
| | - Xu Li
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
| | - Yuan Ding
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
| | - Xilin Yang
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
| | - Danni Ai
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
| | - Xiaohong Chen
- Department of Otolaryngology, Head and Neck Surgery, Beijing Tongren Hospital, Beijing 100730, China
| | - Yongtian Wang
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
| | - Jian Yang
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China.
| |
Collapse
|
7
|
Alsaleh SM, Aviles-Rivero AI, Hahn JK. ReTouchImg: Fusioning from-local-to-global context detection and graph data structures for fully-automatic specular reflection removal for endoscopic images. Comput Med Imaging Graph 2019; 73:39-48. [PMID: 30877992 DOI: 10.1016/j.compmedimag.2019.02.002] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2018] [Revised: 12/31/2018] [Accepted: 02/18/2019] [Indexed: 11/20/2022]
Abstract
Minimally invasive surgical and diagnostic systems are commonly used in clinical practices. However, the accuracy and robustness of these systems depend heavily on computer based processes such as tracking, detecting or segmenting clinically meaningful regions of interest, which are significantly affected by the inherent specular reflections that appear on the organs' surfaces. Restoration of the acquired data for clinical purposes still presents challenges because of the high texture and color variations across the image. In this work, we propose a novel fully-automated solution for endoscopic image restoration, which we call ReTouchImg. Our approach is designed as a two-step scheme. The first is a detection step that is based on the synergy of a set of color variations and gradient information conditions. For the second step, we introduce an inpainting process which is based on graph data structures for recovering the missing information. We exhaustively evaluate our approach on real endoscopic datasets and compare it against some works from the body of literature. We also demonstrate that our solution deals with complex cases such as strong illumination variation and large affected areas through a careful quantitative evaluation of a range of numerical results.
Collapse
|
8
|
Khan TH, Mohammed SK, Imtiaz MS, Wahid KA. Color reproduction and processing algorithm based on real-time mapping for endoscopic images. Springerplus 2016; 5:17. [PMID: 26759756 PMCID: PMC4703600 DOI: 10.1186/s40064-015-1612-4] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/27/2015] [Accepted: 12/13/2015] [Indexed: 01/03/2023]
Abstract
In this paper, we present a real-time preprocessing algorithm for image enhancement for endoscopic images. A novel dictionary based color mapping algorithm is used for reproducing the color information from a theme image. The theme image is selected from a nearby anatomical location. A database of color endoscopy image for different location is prepared for this purpose. The color map is dynamic as its contents change with the change of the theme image. This method is used on low contrast grayscale white light images and raw narrow band images to highlight the vascular and mucosa structures and to colorize the images. It can also be applied to enhance the tone of color images. The statistic visual representation and universal image quality measures show that the proposed method can highlight the mucosa structure compared to other methods. The color similarity has been verified using Delta E color difference, structure similarity index, mean structure similarity index and structure and hue similarity. The color enhancement was measured using color enhancement factor that shows considerable improvements. The proposed algorithm has low and linear time complexity, which results in higher execution speed than other related works.
Collapse
Affiliation(s)
- Tareq H. Khan
- Department of Electrical and Computer Engineering, University of Saskatchewan, Saskatoon, SK S7N5A9 Canada
| | - Shahed K. Mohammed
- Department of Electrical and Computer Engineering, University of Saskatchewan, Saskatoon, SK S7N5A9 Canada
| | - Mohammad S. Imtiaz
- Department of Electrical and Computer Engineering, University of Saskatchewan, Saskatoon, SK S7N5A9 Canada
| | - Khan A. Wahid
- Department of Electrical and Computer Engineering, University of Saskatchewan, Saskatoon, SK S7N5A9 Canada
| |
Collapse
|