151
|
Nam Y, Kim SY, Kim KA, Kwon E, Lee YH, Jang J, Lee MK, Kim J, Choi Y. Development and Validation of Deep Learning-Based Automated Detection of Cervical Lymphadenopathy in Patients with Lymphoma for Treatment Response Assessment: A Bi-institutional Feasibility Study. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:734-743. [PMID: 38316667 PMCID: PMC11031526 DOI: 10.1007/s10278-024-00966-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/03/2023] [Revised: 11/11/2023] [Accepted: 11/13/2023] [Indexed: 02/07/2024]
Abstract
The purpose is to train and evaluate a deep learning (DL) model for the accurate detection and segmentation of abnormal cervical lymph nodes (LN) on head and neck contrast-enhanced CT scans in patients diagnosed with lymphoma and evaluate the clinical utility of the DL model in response assessment. This retrospective study included patients who underwent CT for abnormal cervical LN and lymphoma assessment between January 2021 and July 2022. Patients were grouped into the development (n = 76), internal test 1 (n = 27), internal test 2 (n = 87), and external test (n = 26) cohorts. A 3D SegResNet model was used to train the CT images. The volume change rates of cervical LN across longitudinal CT scans were compared among patients with different treatment outcomes (stable, response, and progression). Dice similarity coefficient (DSC) and the Bland-Altman plot were used to assess the model's segmentation performance and reliability, respectively. No significant differences in baseline clinical characteristics were found across cohorts (age, P = 0.55; sex, P = 0.13; diagnoses, P = 0.06). The mean DSC was 0.39 ± 0.2 with a precision and recall of 60.9% and 57.0%, respectively. Most LN volumes were within the limits of agreement on the Bland-Altman plot. The volume change rates among the three groups differed significantly (progression (n = 74), 342.2%; response (n = 8), - 79.2%; stable (n = 5), - 8.1%; all P < 0.01). Our proposed DL segmentation model showed modest performance in quantifying the cervical LN burden on CT in patients with lymphoma. Longitudinal changes in cervical LN volume, as predicted by the DL model, were useful for treatment response assessment.
Collapse
Affiliation(s)
- Yoonho Nam
- Division of Biomedical Engineering, Hankuk University of Foreign Studies, Yongin-Si, Gyeonggi-do, Republic of Korea
| | - Su-Youn Kim
- Division of Biomedical Engineering, Hankuk University of Foreign Studies, Yongin-Si, Gyeonggi-do, Republic of Korea
| | - Kyu-Ah Kim
- Division of Biomedical Engineering, Hankuk University of Foreign Studies, Yongin-Si, Gyeonggi-do, Republic of Korea
| | - Euna Kwon
- Division of Biomedical Engineering, Hankuk University of Foreign Studies, Yongin-Si, Gyeonggi-do, Republic of Korea
| | - Yoo Hyun Lee
- College of Medicine, The Catholic University of Korea, Seoul, Republic of Korea
| | - Jinhee Jang
- Department of Radiology, College of Medicine, Seoul St. Mary's Hospital, The Catholic University of Korea, Seoul, Republic of Korea
| | - Min Kyoung Lee
- Department of Radiology, College of Medicine, Yeouido St. Mary's Hospital, The Catholic University of Korea, Seoul, Republic of Korea
| | - Jiwoong Kim
- Department of Mathematics and Statistics, University of South Florida, Tampa, FL, USA
| | - Yangsean Choi
- Department of Radiology, College of Medicine, Seoul St. Mary's Hospital, The Catholic University of Korea, Seoul, Republic of Korea.
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine, Asan Medical Centre, 43 Olympic-Ro 88, Songpa-Gu, Seoul, 05505, Republic of Korea.
| |
Collapse
|
152
|
Yi X, He Y, Gao S, Li M. A review of the application of deep learning in obesity: From early prediction aid to advanced management assistance. Diabetes Metab Syndr 2024; 18:103000. [PMID: 38604060 DOI: 10.1016/j.dsx.2024.103000] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/08/2022] [Revised: 01/23/2024] [Accepted: 03/29/2024] [Indexed: 04/13/2024]
Abstract
BACKGROUND AND AIMS Obesity is a chronic disease which can cause severe metabolic disorders. Machine learning (ML) techniques, especially deep learning (DL), have proven to be useful in obesity research. However, there is a dearth of systematic reviews of DL applications in obesity. This article aims to summarize the current trend of DL usage in obesity research. METHODS An extensive literature review was carried out across multiple databases, including PubMed, Embase, Web of Science, Scopus, and Medline, to collate relevant studies published from January 2018 to September 2023. The focus was on research detailing the application of DL in the context of obesity. We have distilled critical insights pertaining to the utilized learning models, encompassing aspects of their development, principal results, and foundational methodologies. RESULTS Our analysis culminated in the synthesis of new knowledge regarding the application of DL in the context of obesity. Finally, 40 research articles were included. The final collection of these research can be divided into three categories: obesity prediction (n = 16); obesity management (n = 13); and body fat estimation (n = 11). CONCLUSIONS This is the first review to examine DL applications in obesity. It reveals DL's superiority in obesity prediction over traditional ML methods, showing promise for multi-omics research. DL also innovates in obesity management through diet, fitness, and environmental analyses. Additionally, DL improves body fat estimation, offering affordable and precise monitoring tools. The study is registered with PROSPERO (ID: CRD42023475159).
Collapse
Affiliation(s)
- Xinghao Yi
- Department of Endocrinology, NHC Key Laboratory of Endocrinology, Peking Union Medical College Hospital, Peking Union Medical College and Chinese Academy of Medical Sciences, Beijing 100730, China
| | - Yangzhige He
- Department of Medical Research Center, Peking Union Medical College Hospital, Chinese Academy of Medical Science & Peking Union Medical College, Beijing 100730, China
| | - Shan Gao
- Department of Endocrinology, Xuan Wu Hospital, Capital Medical University, Beijing 10053, China
| | - Ming Li
- Department of Endocrinology, NHC Key Laboratory of Endocrinology, Peking Union Medical College Hospital, Peking Union Medical College and Chinese Academy of Medical Sciences, Beijing 100730, China.
| |
Collapse
|
153
|
Zhang Z, Hwang M, Kilbaugh TJ, Katz J. Improving sub-pixel accuracy in ultrasound localization microscopy using supervised and self-supervised deep learning. MEASUREMENT SCIENCE & TECHNOLOGY 2024; 35:045701. [PMID: 38205381 PMCID: PMC10774911 DOI: 10.1088/1361-6501/ad1671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Revised: 11/30/2023] [Accepted: 12/17/2023] [Indexed: 01/12/2024]
Abstract
With a spatial resolution of tens of microns, ultrasound localization microscopy (ULM) reconstructs microvascular structures and measures intravascular flows by tracking microbubbles (1-5 μm) in contrast enhanced ultrasound (CEUS) images. Since the size of CEUS bubble traces, e.g. 0.5-1 mm for ultrasound with a wavelength λ = 280 μm, is typically two orders of magnitude larger than the bubble diameter, accurately localizing microbubbles in noisy CEUS data is vital to the fidelity of the ULM results. In this paper, we introduce a residual learning based supervised super-resolution blind deconvolution network (SupBD-net), and a new loss function for a self-supervised blind deconvolution network (SelfBD-net), for detecting bubble centers at a spatial resolution finer than λ/10. Our ultimate purpose is to improve the ability to distinguish closely located microvessels and the accuracy of the velocity profile measurements in macrovessels. Using realistic synthetic data, the performance of these methods is calibrated and compared against several recently introduced deep learning and blind deconvolution techniques. For bubble detection, errors in bubble center location increase with the trace size, noise level, and bubble concentration. For all cases, SupBD-net yields the least error, keeping it below 0.1 λ. For unknown bubble trace morphology, where all the supervised learning methods fail, SelfBD-net can still maintain an error of less than 0.15 λ. SupBD-net also outperforms the other methods in separating closely located bubbles and parallel microvessels. In macrovessels, SupBD-net maintains the least errors in the vessel radius and velocity profile after introducing a procedure that corrects for terminated tracks caused by overlapping traces. Application of these methods is demonstrated by mapping the cerebral microvasculature of a neonatal pig, where neighboring microvessels separated by 0.15 λ can be readily distinguished by SupBD-net and SelfBD-net, but not by the other techniques. Hence, the newly proposed residual learning based methods improve the spatial resolution and accuracy of ULM in micro- and macro-vessels.
Collapse
Affiliation(s)
- Zeng Zhang
- Department of Mechanical Engineering, Johns Hopkins University, Baltimore, MD, United States of America
| | - Misun Hwang
- Departments of Radiology, Children’s Hospital of Philadelphia, Philadelphia, PA, United States of America
- Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, United States of America
| | - Todd J Kilbaugh
- Department of Anesthesiology and Critical Care Medicine, Children’s Hospital of Philadelphia, Philadelphia, PA, United States of America
| | - Joseph Katz
- Department of Mechanical Engineering, Johns Hopkins University, Baltimore, MD, United States of America
| |
Collapse
|
154
|
Wu R, Qin K, Fang Y, Xu Y, Zhang H, Li W, Luo X, Han Z, Liu S, Li Q. Application of the convolution neural network in determining the depth of invasion of gastrointestinal cancer: a systematic review and meta-analysis. J Gastrointest Surg 2024; 28:538-547. [PMID: 38583908 DOI: 10.1016/j.gassur.2023.12.029] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/22/2023] [Revised: 12/16/2023] [Accepted: 12/30/2023] [Indexed: 04/09/2024]
Abstract
BACKGROUND With the development of endoscopic technology, endoscopic submucosal dissection (ESD) has been widely used in the treatment of gastrointestinal tumors. It is necessary to evaluate the depth of tumor invasion before the application of ESD. The convolution neural network (CNN) is a type of artificial intelligence that has the potential to assist in the classification of the depth of invasion in endoscopic images. This meta-analysis aimed to evaluate the performance of CNN in determining the depth of invasion of gastrointestinal tumors. METHODS A search on PubMed, Web of Science, and SinoMed was performed to collect the original publications about the use of CNN in determining the depth of invasion of gastrointestinal neoplasms. Pooled sensitivity and specificity were calculated using an exact binominal rendition of the bivariate mixed-effects regression model. I2 was used for the evaluation of heterogeneity. RESULTS A total of 17 articles were included; the pooled sensitivity was 84% (95% CI, 0.81-0.88), specificity was 91% (95% CI, 0.85-0.94), and the area under the curve (AUC) was 0.93 (95% CI, 0.90-0.95). The performance of CNN was significantly better than that of endoscopists (AUC: 0.93 vs 0.83, respectively; P = .0005). CONCLUSION Our review revealed that CNN is one of the most effective methods of endoscopy to evaluate the depth of invasion of early gastrointestinal tumors, which has the potential to work as a remarkable tool for clinical endoscopists to make decisions on whether the lesion is feasible for endoscopic treatment.
Collapse
Affiliation(s)
- Ruo Wu
- Nanfang Hospital (The First School of Clinical Medicine), Southern Medical University, Guangzhou, Guangdong, China
| | - Kaiwen Qin
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Yuxin Fang
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Yuyuan Xu
- Department of Hepatology Unit and Infectious Diseases, State Key Laboratory of Organ Failure Research, Guangdong Provincial Key Laboratory of Viral Hepatitis Research, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Haonan Zhang
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Wenhua Li
- Nanfang Hospital (The First School of Clinical Medicine), Southern Medical University, Guangzhou, Guangdong, China
| | - Xiaobei Luo
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Zelong Han
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Side Liu
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China; Pazhou Lab, Guangzhou, Guangdong, China
| | - Qingyuan Li
- Department of Gastroenterology, Guangdong Provincial Key Laboratory of Gastroenterology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China.
| |
Collapse
|
155
|
Zhang Y, Xie J, Fu E, Cai W, Xu W. Artificial intelligence in cardiology: a bibliometric study. Am J Transl Res 2024; 16:1029-1035. [PMID: 38586089 PMCID: PMC10994793 DOI: 10.62347/hsfe6936] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2023] [Accepted: 11/28/2023] [Indexed: 04/09/2024]
Abstract
OBJECTIVES To perform a comprehensive bibliometric analysis of global publications on the applications of artificial intelligence (AI) in cardiology. METHODS Documents related to AI in cardiology published between 2002 and 2022 were retrieved from Web of Science Core Collection. R package "bibliometrix", VOSviewers and Microsoft Excel were applied to perform the bibliometric analysis. RESULTS A total of 4332 articles were included. United States topped the list of countries publishing articles, followed by China and United Kingdom. The Harvard University was the institution that contributed the most to this field, followed by University of California System and University of London. Disease risk prediction, diagnosis, treatment, disease detection, and prognosis assessment were the research hotspots for AI in cardiology. CONCLUSIONS Enhancing cooperation between different countries and institutions is a critical step in leading to breakthroughs in the application of AI in cardiology. It is foreseeable that the application of machine learning and deep learning in various areas of cardiology will be a research priority in the coming years.
Collapse
Affiliation(s)
- Yalan Zhang
- Department of Pharmacy, The Second Affiliated Hospital of Fujian Medical UniversityQuanzhou, Fujian, China
| | - Jingwen Xie
- Guangzhou University of Chinese MedicineGuangzhou, Guangdong, China
| | - Enlong Fu
- Guangzhou University of Chinese MedicineGuangzhou, Guangdong, China
| | - Wan Cai
- Shanghai University of Traditional Chinese MedicineShanghai, China
| | - Wentan Xu
- Department of Pharmacy, Jinjiang Municipal HospitalJinjiang, Fujian, China
| |
Collapse
|
156
|
Hernández HO, Montoya F, Hernández-Herrera P, Díaz-Guerrero DS, Olveres J, Bloomfield-Gadêlha H, Darszon A, Escalante-Ramírez B, Corkidi G. Feature-based 3D+t descriptors of hyperactivated human sperm beat patterns. Heliyon 2024; 10:e26645. [PMID: 38444471 PMCID: PMC10912238 DOI: 10.1016/j.heliyon.2024.e26645] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Revised: 11/23/2023] [Accepted: 02/16/2024] [Indexed: 03/07/2024] Open
Abstract
The flagellar movement of the mammalian sperm plays a crucial role in fertilization. In the female reproductive tract, human spermatozoa undergo a process called capacitation which promotes changes in their motility. Only capacitated spermatozoa may be hyperactivated and only those that transition to hyperactivated motility are capable of fertilizing the egg. Hyperactivated motility is characterized by asymmetric flagellar bends of greater amplitude and lower frequency. Historically, clinical fertilization studies have used two-dimensional analysis to classify sperm motility, despite the inherently three-dimensional (3D) nature of sperm motion. Recent research has described several 3D beating features of sperm flagella. However, the 3D motility pattern of hyperactivated spermatozoa has not yet been characterized. One of the main challenges in classifying these patterns in 3D is the lack of a ground-truth reference, as it can be difficult to visually assess differences in flagellar beat patterns. Additionally, it is worth noting that only a relatively small proportion, approximately 10-20% of sperm incubated under capacitating conditions exhibit hyperactivated motility. In this work, we used a multifocal image acquisition system that can acquire, segment, and track sperm flagella in 3D+t. We developed a feature-based vector that describes the spatio-temporal flagellar sperm motility patterns by an envelope of ellipses. The classification results obtained using our 3D feature-based descriptors can serve as potential label for future work involving deep neural networks. By using the classification results as labels, it will be possible to train a deep neural network to automatically classify spermatozoa based on their 3D flagellar beating patterns. We demonstrated the effectiveness of the descriptors by applying them to a dataset of human sperm cells and showing that they can accurately differentiate between non-hyperactivated and hyperactivated 3D motility patterns of the sperm cells. This work contributes to the understanding of 3D flagellar hyperactive motility patterns and provides a framework for research in the fields of human and animal fertility.
Collapse
Affiliation(s)
- Haydee O. Hernández
- Posgrado en Ciencia e Ingeniería de la Computación, Universidad Nacional Autónoma de México, UNAM, Ciudad de México, Mexico
- Laboratorio de Imágenes y Visión por Computadora, Instituto de Biotecnología, UNAM, Cuernavaca, Mexico
| | - Fernando Montoya
- Laboratorio de Imágenes y Visión por Computadora, Instituto de Biotecnología, UNAM, Cuernavaca, Mexico
| | - Paul Hernández-Herrera
- Laboratorio de Imágenes y Visión por Computadora, Instituto de Biotecnología, UNAM, Cuernavaca, Mexico
- Facultad de Ciencias, Universidad Autónoma de San Luis Potosí, San Luis Potosí, Mexico
| | - Dan S. Díaz-Guerrero
- Laboratorio de Imágenes y Visión por Computadora, Instituto de Biotecnología, UNAM, Cuernavaca, Mexico
| | - Jimena Olveres
- Departamento de Procesamiento de Señales, Facultad de Ingeniería, UNAM, Ciudad de México, Mexico
| | - Hermes Bloomfield-Gadêlha
- Department of Engineering Mathematics and Technology, Bristol Robotics Laboratory, University of Bristol, Bristol, United Kingdom
| | - Alberto Darszon
- Departamento de Genética del Desarrollo y Fisiología Molecular, Instituto de Biotecnología, UNAM, Ciudad de México, Mexico
| | - Boris Escalante-Ramírez
- Departamento de Procesamiento de Señales, Facultad de Ingeniería, UNAM, Ciudad de México, Mexico
| | - Gabriel Corkidi
- Laboratorio de Imágenes y Visión por Computadora, Instituto de Biotecnología, UNAM, Cuernavaca, Mexico
| |
Collapse
|
157
|
Jung J, Han J, Han JM, Ko J, Yoon J, Hwang JS, Park JI, Hwang G, Jung JH, Hwang DDJ. Prediction of neovascular age-related macular degeneration recurrence using optical coherence tomography images with a deep neural network. Sci Rep 2024; 14:5854. [PMID: 38462646 PMCID: PMC10925587 DOI: 10.1038/s41598-024-56309-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2023] [Accepted: 03/05/2024] [Indexed: 03/12/2024] Open
Abstract
Neovascular age-related macular degeneration (nAMD) can result in blindness if left untreated, and patients often require repeated anti-vascular endothelial growth factor injections. Although, the treat-and-extend method is becoming popular to reduce vision loss attributed to recurrence, it may pose a risk of overtreatment. This study aimed to develop a deep learning model based on DenseNet201 to predict nAMD recurrence within 3 months after confirming dry-up 1 month following three loading injections in treatment-naïve patients. A dataset of 1076 spectral domain optical coherence tomography (OCT) images from 269 patients diagnosed with nAMD was used. The performance of the model was compared with that of 6 ophthalmologists, using 100 randomly selected samples. The DenseNet201-based model achieved 53.0% accuracy in predicting nAMD recurrence using a single pre-injection image and 60.2% accuracy after viewing all the images immediately after the 1st, 2nd, and 3rd injections. The model outperformed experienced ophthalmologists, with an average accuracy of 52.17% using a single pre-injection image and 53.3% after examining four images before and after three loading injections. In conclusion, the artificial intelligence model demonstrated a promising ability to predict nAMD recurrence using OCT images and outperformed experienced ophthalmologists. These findings suggest that deep learning models can assist in nAMD recurrence prediction, thus improving patient outcomes and optimizing treatment strategies.
Collapse
Affiliation(s)
- Juho Jung
- Department of Applied Artificial Intelligence, Sungkyunkwan University, Seoul, Korea
| | - Jinyoung Han
- Department of Applied Artificial Intelligence, Sungkyunkwan University, Seoul, Korea
- Department of Human-Artificial Intelligence Interaction, Sungkyunkwan University, Seoul, Korea
| | - Jeong Mo Han
- Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Korea
- Kong Eye Center, Seoul, Korea
| | | | | | | | - Ji In Park
- Department of Medicine, Kangwon National University Hospital, Kangwon National University School of Medicine, Chuncheon, Gangwon-do, Korea
| | - Gyudeok Hwang
- Department of Ophthalmology, Hangil Eye Hospital, 35 Bupyeong-Daero, Bupyeong-gu, Incheon, 21388, Korea
| | - Jae Ho Jung
- Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Korea
| | - Daniel Duck-Jin Hwang
- Department of Ophthalmology, Hangil Eye Hospital, 35 Bupyeong-Daero, Bupyeong-gu, Incheon, 21388, Korea.
- Department of Ophthalmology, Catholic Kwandong University College of Medicine, Incheon, Korea.
- Lux Mind, Incheon, Korea.
| |
Collapse
|
158
|
Zubair M. Clinical applications of artificial intelligence in identification and management of bacterial infection: Systematic review and meta-analysis. Saudi J Biol Sci 2024; 31:103934. [PMID: 38304541 PMCID: PMC10831261 DOI: 10.1016/j.sjbs.2024.103934] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2023] [Revised: 12/28/2023] [Accepted: 01/11/2024] [Indexed: 02/03/2024] Open
Abstract
Pneumonia is declared a global emergency public health crisis in children less than five age and the geriatric population. Recent advancements in deep learning models could be utilized effectively for the timely and early diagnosis of pneumonia in immune-compromised patients to avoid complications. This systematic review and meta-analysis utilized PRISMA guidelines for the selection of ten articles included in this study. The literature search was done through electronic databases including PubMed, Scopus, and Google Scholar from 1st January 2016 till 1 July 2023. Overall studies included a total of 126,610 images and 1706 patients in this meta-analysis. At a 95% confidence interval, for pooled sensitivity was 0.90 (0.85-0.94) and I2 statistics 90.20 (88.56 - 91.92). The pooled specificity for deep learning models' diagnostic accuracy was 0.89 (0.86---0.92) and I2 statistics 92.72 (91.50 - 94.83). I2 statistics showed low heterogeneity across studies highlighting consistent and reliable estimates, and instilling confidence in these findings for researchers and healthcare practitioners. The study highlighted the recent deep learning models single or in combination with high accuracy, sensitivity, and specificity to ensure reliable use for bacterial pneumonia identification and differentiate from other viral, fungal pneumonia in children and adults through chest x-rays and radiographs.
Collapse
Affiliation(s)
- Mohammad Zubair
- Department of Medical Microbiology, Faculty of Medicine, University of Tabuk, Tabuk 71491, Kingdom of Saudi Arabia
| |
Collapse
|
159
|
Li Z, Zhou L, Bin X, Tan S, Tan Z, Tang A. Utility of deep learning for the diagnosis of cochlear malformation on temporal bone CT. Jpn J Radiol 2024; 42:261-267. [PMID: 37812304 DOI: 10.1007/s11604-023-01494-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Accepted: 09/19/2023] [Indexed: 10/10/2023]
Abstract
OBJECTIVE Diagnosis of cochlear malformation on temporal bone CT images is often difficult. Our aim was to assess the utility of deep learning analysis in diagnosing cochlear malformation on temporal bone CT images. METHODS A total of 654 images from 165 temporal bone CTs were divided into the training set (n = 534) and the testing set (n = 120). A target region that includes the area of the cochlear was extracted to create a diagnostic model. 4 models were used: ResNet10, ResNet50, SE-ResNet50, and DenseNet121. The testing data set was subsequently analyzed using these models and by 4 doctors. RESULTS The areas under the curve was 0.91, 0.94, 0.93, and 0.73 in ResNet10, ResNet50, SE-ResNet50, and DenseNet121. The accuracy of ResNet10, ResNet50, and SE-ResNet50 is better than chief physician. CONCLUSIONS Deep learning technique implied a promising prospect for clinical application of artificial intelligence in the diagnosis of cochlear malformation based on CT images.
Collapse
Affiliation(s)
- Zhenhua Li
- Department of Otorhinolaryngology-Head and Neck Surgery, Hunan Provincial People's Hospital, The First Affiliated Hospital of Hunan Normal University, Changsha, Hunan, People's Republic of China
| | - Langtao Zhou
- School of Computer Science and Cyber Engineering, Guangzhou University, Guangzhou, People's Republic of China
| | - Xiang Bin
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Guangxi Medical University, Nanning, Guangxi, People's Republic of China
| | - Songhua Tan
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Guangxi Medical University, Nanning, Guangxi, People's Republic of China
| | - Zhiqiang Tan
- Department of Otorhinolaryngology-Head and Neck Surgery, Hunan Provincial People's Hospital, The First Affiliated Hospital of Hunan Normal University, Changsha, Hunan, People's Republic of China
| | - Anzhou Tang
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Guangxi Medical University, Nanning, Guangxi, People's Republic of China.
- Clinical Teaching Building, Guangxi Medical University, Nanning, 530000, People's Republic of China.
| |
Collapse
|
160
|
Matsumoto S, Nakahara Y, Yonezawa T, Nakamura Y, Tanabe M, Higashi M, Shiraishi J. Development of an individual display optimization system based on deep convolutional neural network transition learning for somatostatin receptor scintigraphy. Radiol Phys Technol 2024; 17:195-206. [PMID: 38165579 DOI: 10.1007/s12194-023-00766-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Revised: 11/13/2023] [Accepted: 11/26/2023] [Indexed: 01/04/2024]
Abstract
Somatostatin receptor scintigraphy (SRS) is an essential examination for the diagnosis of neuroendocrine tumors (NETs). This study developed a method to individually optimize the display of whole-body SRS images using a deep convolutional neural network (DCNN) reconstructed by transfer learning of a DCNN constructed using Gallium-67 (67Ga) images. The initial DCNN was constructed using U-Net to optimize the display of 67Ga images (493 cases/986 images), and a DCNN with transposed weight coefficients was reconstructed for the optimization of whole-body SRS images (133 cases/266 images). A DCNN was constructed for each observer using reference display conditions estimated in advance. Furthermore, to eliminate information loss in the original image, a grayscale linear process is performed based on the DCNN output image to obtain the final linearly corrected DCNN (LcDCNN) image. To verify the usefulness of the proposed method, an observer study using a paired-comparison method was conducted on the original, reference, and LcDCNN images of 15 cases with 30 images. The paired comparison method showed that in most cases (29/30), the LcDCNN images were significantly superior to the original images in terms of display conditions. When comparing the LcDCNN and reference images, the number of LcDCNN and reference images that were superior to each other in the display condition was 17 and 13, respectively, and in both cases, 6 of these images showed statistically significant differences. The optimized SRS images obtained using the proposed method, while reflecting the observer's preference, were superior to the conventional manually adjusted images.
Collapse
Affiliation(s)
- Shun Matsumoto
- Division of Radiology, Yamaguchi University Hospital, 1-1-1 Minamikogushi, Yamaguchi, 755-8505, Japan.
- Graduate School of Health Sciences, Kumamoto University, 4-24-1 Kuhonji, Kumamoto, 862-0976, Japan.
| | - Yuki Nakahara
- Division of Radiology, Yamaguchi University Hospital, 1-1-1 Minamikogushi, Yamaguchi, 755-8505, Japan
| | - Teppei Yonezawa
- Division of Radiology, Yamaguchi University Hospital, 1-1-1 Minamikogushi, Yamaguchi, 755-8505, Japan
| | - Yuto Nakamura
- Division of Radiology, Yamaguchi University Hospital, 1-1-1 Minamikogushi, Yamaguchi, 755-8505, Japan
| | - Masahiro Tanabe
- Department of Radiology, Yamaguchi University School of Medicine, 1-1-1 Minamikogushi, Yamaguchi, 755-8505, Japan
| | - Mayumi Higashi
- Department of Radiology, Yamaguchi University School of Medicine, 1-1-1 Minamikogushi, Yamaguchi, 755-8505, Japan
| | - Junji Shiraishi
- Faculty of Life Sciences, Kumamoto University, 4-24-1 Kuhonji, Kumamoto, 862-0976, Japan
| |
Collapse
|
161
|
Murray J, Heng D, Lygate A, Porto L, Abade A, Manica S, Franco A. Applying artificial intelligence to determination of legal age of majority from radiographic data. Morphologie 2024; 108:100723. [PMID: 37897941 DOI: 10.1016/j.morpho.2023.100723] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Accepted: 08/24/2023] [Indexed: 10/30/2023]
Abstract
Forensic odontologists use biological patterns to estimate chronological age for the judicial system. The age of majority is a legally significant period with a limited set of reliable oral landmarks. Currently, experts rely on the questionable development of third molars to assess whether litigants can be prosecuted as legal adults. Identification of new and novel patterns may illuminate features more dependably indicative of chronological age, which have, until now, remained unseen. Unfortunately, biased perceptions and limited cognitive capacity compromise the ability of researchers to notice new patterns. The present study demonstrates how artificial intelligence can break through identification barriers and generate new estimation modalities. A convolutional neural network was trained with 4003 panoramic-radiographs to sort subjects into 'under-18' and 'over-18' age categories. The resultant architecture identified legal adults with a high predictive accuracy equally balanced between precision, specificity and recall. Moving forward, AI-based methods could improve courtroom efficiency, stand as automated assessment methods and contribute to our understanding of biological ageing.
Collapse
Affiliation(s)
- J Murray
- Department of Forensic Odontology, University of Dundee, Nethergate, Dundee DD1 4HN, UK.
| | - D Heng
- Department of Forensic Odontology, University of Dundee, Nethergate, Dundee DD1 4HN, UK
| | - A Lygate
- Department of Forensic Odontology, University of Dundee, Nethergate, Dundee DD1 4HN, UK
| | - L Porto
- Department of Mechanical Engineering, University of Brasilia, Federal District 70910-900, Brazil
| | - A Abade
- Departmento de Computacao, Instituto Federal de Educacao, Ciencie e Tecnologia de Mato Grosso, Cuiaba, Mato Grosso, Brazil
| | - S Manica
- Department of Forensic Odontology, University of Dundee, Nethergate, Dundee DD1 4HN, UK
| | - A Franco
- Department of Forensic Odontology, University of Dundee, Nethergate, Dundee DD1 4HN, UK; Division of Forensic Dentistry, Faculdade São Leopoldo Mandic, Campinas, Brazil
| |
Collapse
|
162
|
Al Muhaisen S, Safi O, Ulayan A, Aljawamis S, Fakhoury M, Baydoun H, Abuquteish D. Artificial Intelligence-Powered Mammography: Navigating the Landscape of Deep Learning for Breast Cancer Detection. Cureus 2024; 16:e56945. [PMID: 38665752 PMCID: PMC11044525 DOI: 10.7759/cureus.56945] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/26/2024] [Indexed: 04/28/2024] Open
Abstract
Worldwide, breast cancer (BC) is one of the most commonly diagnosed malignancies in women. Early detection is key to improving survival rates and health outcomes. This literature review focuses on how artificial intelligence (AI), especially deep learning (DL), can enhance the ability of mammography, a key tool in BC detection, to yield more accurate results. Artificial intelligence has shown promise in reducing diagnostic errors and increasing early cancer detection chances. Nevertheless, significant challenges exist, including the requirement for large amounts of high-quality data and concerns over data privacy. Despite these hurdles, AI and DL are advancing the field of radiology, offering better ways to diagnose, detect, and treat diseases. The U.S. Food and Drug Administration (FDA) has approved several AI diagnostic tools. Yet, the full potential of these technologies, especially for more advanced screening methods like digital breast tomosynthesis (DBT), depends on further clinical studies and the development of larger databases. In summary, this review highlights the exciting potential of AI in BC screening. It calls for more research and validation to fully employ the power of AI in clinical practice, ensuring that these technologies can help save lives by improving diagnosis accuracy and efficiency.
Collapse
Affiliation(s)
| | - Omar Safi
- Medicine, Faculty of Medicine, The Hashemite University, Zarqa, JOR
| | - Ahmad Ulayan
- Medicine, Faculty of Medicine, The Hashemite University, Zarqa, JOR
| | - Sara Aljawamis
- Medicine, Faculty of Medicine, The Hashemite University, Zarqa, JOR
| | - Maryam Fakhoury
- Medicine, Faculty of Medicine, The Hashemite University, Zarqa, JOR
| | - Haneen Baydoun
- Diagnostic Radiology, King Hussein Cancer Center, Amman, JOR
| | - Dua Abuquteish
- Microbiology, Pathology and Forensic Medicine, Faculty of Medicine, The Hashemite University, Zarqa, JOR
- Pathology and Laboratory Medicine, King Hussein Cancer Center, Amman, JOR
| |
Collapse
|
163
|
Haq AU, Li JP, Khan I, Agbley BLY, Ahmad S, Uddin MI, Zhou W, Khan S, Alam I. DEBCM: Deep Learning-Based Enhanced Breast Invasive Ductal Carcinoma Classification Model in IoMT Healthcare Systems. IEEE J Biomed Health Inform 2024; 28:1207-1217. [PMID: 37015704 DOI: 10.1109/jbhi.2022.3228577] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Abstract
Accurate breast cancer (BC) diagnosis is a difficult task that is critical for the proper treatment of BC in IoMT (Internet of Medical Things) healthcare systems. This paper proposes a convolutional neural network (CNN)-based diagnosis method for detecting early-stage breast cancer. In developing the proposed method, we incorporated the CNN model for the invasive ductal carcinoma (IDC) classification using breast histology image data. We have incorporated transfer learning (TL) and data augmentation (DA) mechanisms to improve the CNN model's predictive outcomes. For the fine-tuning process, the CNN model was trained with breast histology image data. Furthermore, the held-out cross-validation method for best model selection and hyper-parameter tuning was incorporated. In addition, various performance evaluation metrics for model performance assessment were computed. The experimental results confirmed that the proposed model outperformed the baseline models across all evaluation metrics, achieving 99.04% accuracy. We recommend the proposed method for early recognition of BC in IoMT healthcare systems due to its high performance.
Collapse
|
164
|
Mo W, Lee J, Abdel-Aty M, Mao S, Jiang Q. Dynamic short-term crash analysis and prediction at toll plazas for proactive safety management. ACCIDENT; ANALYSIS AND PREVENTION 2024; 197:107456. [PMID: 38184886 DOI: 10.1016/j.aap.2024.107456] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/04/2023] [Revised: 12/29/2023] [Accepted: 12/31/2023] [Indexed: 01/09/2024]
Abstract
Toll plazas are commonly recognized as bottlenecks on toll roads, where vehicles are prone to crashes. However, there has been a lack of research analyzing and predicting dynamic short-term crash risk specifically at toll plazas. This study utilizes traffic, geometric, and weather data to analyze and predict dynamic short-term collision occurrence probability at mainline toll plazas. A random-effects logit regression model is employed to identify crash precursors and assess their impacts on the probability of crash occurrence at toll plazas. Meanwhile, a Long Short-Term Memory Convolutional Neural Network (LSTM-CNN) network is applied for crash prediction. The results of random-effects logit regression model indicate that the flow standard deviation of downstream, upstream occupancy, speed difference and occupancy difference between upstream and downstream positively influence the probability of crash occurrence. Conversely, an increase in the proportion of ETC lanes negatively impacts the probability of crash occurrence. Additionally, there appears a higher likelihood of crashes occurring during summer at toll plaza area. Furthermore, to address the issue of data imbalance, Synthetic Minority Oversampling Techniques (SMOTE) and class weight methods were employed. Stacked Sparse AutoEncoder-Long Short-Term Memory (SSAE-LSTM) and CatBoost were developed and their performance was compared with the proposed model. The results demonstrated that the LSTM-CNN model outperformed the other models in terms of the Area Under the Curve (AUC) values and the true positive rate. The findings of this study can assist engineers in selecting suitable traffic control strategies to improve traffic safety in toll plaza areas. Moreover, the developed collision prediction model can be incorporated into a real-time safety management system to proactively prevent traffic crash.
Collapse
Affiliation(s)
- Weiwei Mo
- School of Traffic and Transportation Engineering, Central South University, Changsha, Hunan 410075, China
| | - Jaeyoung Lee
- School of Traffic and Transportation Engineering, Central South University, Changsha, Hunan 410075, China; Department of Civil, Environmental & Construction Engineering, University of Central Florida, Orlando, FL 32816, United States.
| | - Mohamed Abdel-Aty
- Department of Civil, Environmental & Construction Engineering, University of Central Florida, Orlando, FL 32816, United States
| | - Suyi Mao
- School of Traffic and Transportation Engineering, Central South University, Changsha, Hunan 410075, China; Department of Environment, Land and Infrastructure Engineering (DIATI), Politecnico di Torino, Torino, Piemonte 10129, Italy
| | - Qianshan Jiang
- School of Traffic and Transportation Engineering, Central South University, Changsha, Hunan 410075, China; Department of Environment, Land and Infrastructure Engineering (DIATI), Politecnico di Torino, Torino, Piemonte 10129, Italy
| |
Collapse
|
165
|
Charters JA, Luximon D, Petragallo R, Neylon J, Low DA, Lamb JM. Automated detection of vertebral body misalignments in orthogonal kV and MV guided radiotherapy: application to a comprehensive retrospective dataset. Biomed Phys Eng Express 2024; 10:025039. [PMID: 38382110 DOI: 10.1088/2057-1976/ad2baa] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Accepted: 02/21/2024] [Indexed: 02/23/2024]
Abstract
Objective. In image-guided radiotherapy (IGRT), off-by-one vertebral body misalignments are rare but potentially catastrophic. In this study, a novel detection method for such misalignments in IGRT was investigated using densely-connected convolutional networks (DenseNets) for applications towards real-time error prevention and retrospective error auditing.Approach. A total of 4213 images acquired from 527 radiotherapy patients aligned with planar kV or MV radiographs were used to develop and test error-detection software modules. Digitally reconstructed radiographs (DRRs) and setup images were retrieved and co-registered according to the clinically applied alignment contained in the DICOM REG files. A semi-automated algorithm was developed to simulate patient positioning errors on the anterior-posterior (AP) and lateral (LAT) images shifted by one vertebral body. A DenseNet architecture was designed to classify either AP images individually or AP and LAT image pairs. Receiver-operator characteristic curves (ROC) and areas under the curves (AUC) were computed to evaluate the classifiers on test subsets. Subsequently, the algorithm was applied to the entire dataset in order to retrospectively determine the absolute off-by-one vertebral body error rate for planar radiograph guided RT at our institution from 2011-2021.Main results. The AUCs for the kV models were 0.98 for unpaired AP and 0.99 for paired AP-LAT. The AUC for the MV AP model was 0.92. For a specificity of 95%, the paired kV model achieved a sensitivity of 99%. Application of the model to the entire dataset yielded a per-fraction off-by-one vertebral body error rate of 0.044% [0.0022%, 0.21%] for paired kV IGRT including one previously unreported error.Significance. Our error detection algorithm was successful in classifying vertebral body positioning errors with sufficient accuracy for retrospective quality control and real-time error prevention. The reported positioning error rate for planar radiograph IGRT is unique in being determined independently of an error reporting system.
Collapse
Affiliation(s)
- John A Charters
- Department of Radiation Oncology, University of California, Los Angeles, CA 90095, United States of America
| | - Dishane Luximon
- Department of Radiation Oncology, University of California, Los Angeles, CA 90095, United States of America
| | - Rachel Petragallo
- Department of Radiation Oncology, University of California, Los Angeles, CA 90095, United States of America
| | - Jack Neylon
- Department of Radiation Oncology, University of California, Los Angeles, CA 90095, United States of America
| | - Daniel A Low
- Department of Radiation Oncology, University of California, Los Angeles, CA 90095, United States of America
| | - James M Lamb
- Department of Radiation Oncology, University of California, Los Angeles, CA 90095, United States of America
| |
Collapse
|
166
|
Guo X, Xu L, Liu Z, Hao Y, Wang P, Zhu H, Du Y. Automated classification of ulcerative lesions in small intestine using densenet with channel attention and residual dilated blocks. Phys Med Biol 2024; 69:055017. [PMID: 38316034 DOI: 10.1088/1361-6560/ad2637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Accepted: 02/05/2024] [Indexed: 02/07/2024]
Abstract
Objective. Ulceration of the small intestine, which has a high incidence, includes Crohn's disease (CD), intestinal tuberculosis (ITB), primary small intestinal lymphoma (PSIL), cryptogenic multifocal ulcerous stenosing enteritis (CMUSE), and non-specific ulcer (NSU). However, the ulceration morphology can easily be misdiagnosed through enteroscopy.Approach. In this study, DRCA-DenseNet169, which is based on DenseNet169, with residual dilated blocks and a channel attention block, is proposed to identify CD, ITB, PSIL, CMUSE, and NSU intelligently. In addition, a novel loss function that incorporates dynamic weights is designed to enhance the precision of imbalanced datasets with limited samples. DRCA-Densenet169 was evaluated using 10883 enteroscopy images, including 5375 ulcer images and 5508 normal images, which were obtained from the Shanghai Changhai Hospital.Main results. DRCA-Densenet169 achieved an overall accuracy of 85.27% ± 0.32%, a weighted-precision of 83.99% ± 2.47%, a weighted-recall of 84.36% ± 0.88% and a weighted-F1-score of 84.07% ± 2.14%.Significance. The results demonstrate that DRCA-Densenet169 has high recognition accuracy and strong robustness in identifying different types of ulcers when obtaining immediate and preliminary diagnoses.
Collapse
Affiliation(s)
- Xudong Guo
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Lei Xu
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Zhang Liu
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Youguo Hao
- Department of Rehabilitation, Shanghai Putuo People's Hospital, Shanghai 200060, People's Republic of China
| | - Peng Wang
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, People's Republic of China
| | - Huiyun Zhu
- Department of Gastroenterology, Changhai Hospital, Naval Medical University, Shanghai 200433, People's Republic of China
| | - Yiqi Du
- Department of Gastroenterology, Changhai Hospital, Naval Medical University, Shanghai 200433, People's Republic of China
| |
Collapse
|
167
|
Darvish M, Kist AM. A Generative Method for a Laryngeal Biosignal. J Voice 2024:S0892-1997(24)00019-5. [PMID: 38395653 DOI: 10.1016/j.jvoice.2024.01.016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2023] [Revised: 01/26/2024] [Accepted: 01/26/2024] [Indexed: 02/25/2024]
Abstract
The Glottal Area Waveform (GAW) is an important component in quantitative clinical voice assessment, providing valuable insights into vocal fold function. In this study, we introduce a novel method employing Variational Autoencoders (VAEs) to generate synthetic GAWs. Our approach enables the creation of synthetic GAWs that closely replicate real-world data, offering a versatile tool for researchers and clinicians. We elucidate the process of manipulating the VAE latent space using the Glottal Opening Vector (GlOVe). The GlOVe allows precise control over the synthetic closure and opening of the vocal folds. By utilizing the GlOVe, we generate synthetic laryngeal biosignals. These biosignals accurately reflect vocal fold behavior, allowing for the emulation of realistic glottal opening changes. This manipulation extends to the introduction of arbitrary oscillations in the vocal folds, closely resembling real vocal fold oscillations. The range of factor coefficient values enables the generation of diverse biosignals with varying frequencies and amplitudes. Our results demonstrate that this approach yields highly accurate laryngeal biosignals, with the Normalized Mean Absolute Error values for various frequencies ranging from 9.6 ⋅ 10-3 to 1.20 ⋅ 10-2 for different experimented frequencies, alongside a remarkable training effectiveness, reflected in reductions of up to approximately 89.52% in key loss components. This proposed method may have implications for downstream speech synthesis and phonetics research, offering the potential for advanced and natural-sounding speech technologies.
Collapse
Affiliation(s)
- Mahdi Darvish
- Department Artificial Intelligence in Biomedical Engineering, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
| | - Andreas M Kist
- Department Artificial Intelligence in Biomedical Engineering, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany.
| |
Collapse
|
168
|
Zhao G, Chen X, Zhu M, Liu Y, Wang Y. Exploring the application and future outlook of Artificial intelligence in pancreatic cancer. Front Oncol 2024; 14:1345810. [PMID: 38450187 PMCID: PMC10915754 DOI: 10.3389/fonc.2024.1345810] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Accepted: 01/29/2024] [Indexed: 03/08/2024] Open
Abstract
Pancreatic cancer, an exceptionally malignant tumor of the digestive system, presents a challenge due to its lack of typical early symptoms and highly invasive nature. The majority of pancreatic cancer patients are diagnosed when curative surgical resection is no longer possible, resulting in a poor overall prognosis. In recent years, the rapid progress of Artificial intelligence (AI) in the medical field has led to the extensive utilization of machine learning and deep learning as the prevailing approaches. Various models based on AI technology have been employed in the early screening, diagnosis, treatment, and prognostic prediction of pancreatic cancer patients. Furthermore, the development and application of three-dimensional visualization and augmented reality navigation techniques have also found their way into pancreatic cancer surgery. This article provides a concise summary of the current state of AI technology in pancreatic cancer and offers a promising outlook for its future applications.
Collapse
Affiliation(s)
- Guohua Zhao
- Department of General Surgery, Cancer Hospital of China Medical University, Liaoning Cancer Hospital & Institute, Liaoning, China
| | - Xi Chen
- Department of General Surgery, Cancer Hospital of China Medical University, Liaoning Cancer Hospital & Institute, Liaoning, China
- Department of Clinical integration of traditional Chinese and Western medicine, Liaoning University of Traditional Chinese Medicine, Liaoning, China
| | - Mengying Zhu
- Department of General Surgery, Cancer Hospital of China Medical University, Liaoning Cancer Hospital & Institute, Liaoning, China
- Department of Clinical integration of traditional Chinese and Western medicine, Liaoning University of Traditional Chinese Medicine, Liaoning, China
| | - Yang Liu
- Department of Ophthalmology, First Hospital of China Medical University, Liaoning, China
| | - Yue Wang
- Department of General Surgery, Cancer Hospital of China Medical University, Liaoning Cancer Hospital & Institute, Liaoning, China
| |
Collapse
|
169
|
Yang Y, Long Z, Lei B, Liu W, Ye J. Clinical decision support system based on deep learning for evaluating implantable collamer lens size and vault after implantable collamer lens surgery: a retrospective study. BMJ Open 2024; 14:e081050. [PMID: 38365302 PMCID: PMC10875548 DOI: 10.1136/bmjopen-2023-081050] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Accepted: 01/23/2024] [Indexed: 02/18/2024] Open
Abstract
OBJECTIVES To aid doctors in selecting the optimal preoperative implantable collamer lens (ICL) size and to enhance the safety and surgical outcomes of ICL procedures, a clinical decision support system (CDSS) is proposed in our study. DESIGN A retrospective study of patients after ICL surgery. SETTING China Tertiary Myopia Prevention and Control Center. PARTICIPANTS 2772 eyes belonging to 1512 patients after ICL surgery. Data were collected between 2018 and 2022. OUTCOME MEASURES A CDSS is constructed and used to predict vault at 1 month postoperatively and preoperative ICL dimensions using various artificial intelligence methods. Accuracy metrics as well as area under curve (AUC) parameters are used to determine the CDSS prediction methods. RESULTS Among the ICL size prediction models, conventional neural networks (CNNs) achieve the best prediction accuracy at 91.37% and exhibit the highest AUC of 0.842. Regarding the prediction model for vault values 1 month after surgery, CNN surpasses the other methods with an accuracy of 85.27%, which has the uppermost AUC of 0.815. Thus, we select CNN as the prediction algorithm for the CDSS. CONCLUSIONS This study introduces a CDSS to assist doctors in selecting the optimal ICL size for patients while improving the safety and postoperative outcomes of ICL surgery.
Collapse
Affiliation(s)
- Yixuan Yang
- Department of Ophthalmology, The Third Hospital Affiliated to the Third Military Medical University Department of Ophthalmology, Chongqing, China
| | - Zhengqin Long
- Chongqing University Qianjiang Hospital, Chongqing, China
| | - Bo Lei
- Department of Ophthalmology, The Third Hospital Affiliated to the Third Military Medical University Department of Ophthalmology, Chongqing, China
| | - Wei Liu
- Department of Ophthalmology, The Third Hospital Affiliated to the Third Military Medical University Department of Ophthalmology, Chongqing, China
| | - Jian Ye
- Department of Ophthalmology, The Third Hospital Affiliated to the Third Military Medical University Department of Ophthalmology, Chongqing, China
| |
Collapse
|
170
|
Ramli AA, Liu X, Berndt K, Goude E, Hou J, Kaethler LB, Liu R, Lopez A, Nicorici A, Owens C, Rodriguez D, Wang J, Zhang H, Aranki D, McDonald CM, Henricson EK. Gait Characterization in Duchenne Muscular Dystrophy (DMD) Using a Single-Sensor Accelerometer: Classical Machine Learning and Deep Learning Approaches. SENSORS (BASEL, SWITZERLAND) 2024; 24:1123. [PMID: 38400281 PMCID: PMC10892016 DOI: 10.3390/s24041123] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/30/2023] [Revised: 02/02/2024] [Accepted: 02/05/2024] [Indexed: 02/25/2024]
Abstract
Differences in gait patterns of children with Duchenne muscular dystrophy (DMD) and typically developing (TD) peers are visible to the eye, but quantifications of those differences outside of the gait laboratory have been elusive. In this work, we measured vertical, mediolateral, and anteroposterior acceleration using a waist-worn iPhone accelerometer during ambulation across a typical range of velocities. Fifteen TD and fifteen DMD children from 3 to 16 years of age underwent eight walking/running activities, including five 25 m walk/run speed-calibration tests at a slow walk to running speeds (SC-L1 to SC-L5), a 6-min walk test (6MWT), a 100 m fast walk/jog/run (100MRW), and a free walk (FW). For clinical anchoring purposes, participants completed a Northstar Ambulatory Assessment (NSAA). We extracted temporospatial gait clinical features (CFs) and applied multiple machine learning (ML) approaches to differentiate between DMD and TD children using extracted temporospatial gait CFs and raw data. Extracted temporospatial gait CFs showed reduced step length and a greater mediolateral component of total power (TP) consistent with shorter strides and Trendelenberg-like gait commonly observed in DMD. ML approaches using temporospatial gait CFs and raw data varied in effectiveness at differentiating between DMD and TD controls at different speeds, with an accuracy of up to 100%. We demonstrate that by using ML with accelerometer data from a consumer-grade smartphone, we can capture DMD-associated gait characteristics in toddlers to teens.
Collapse
Affiliation(s)
- Albara Ah Ramli
- Department of Computer Science, School of Engineering, University of California, Davis, CA 95616, USA; (A.A.R.); (X.L.); (R.L.)
| | - Xin Liu
- Department of Computer Science, School of Engineering, University of California, Davis, CA 95616, USA; (A.A.R.); (X.L.); (R.L.)
| | - Kelly Berndt
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Erica Goude
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Jiahui Hou
- Department of Electrical and Computer Engineering, School of Engineering, University of Waterloo, Waterloo, ON N2L 3G1, Canada;
| | - Lynea B. Kaethler
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Rex Liu
- Department of Computer Science, School of Engineering, University of California, Davis, CA 95616, USA; (A.A.R.); (X.L.); (R.L.)
| | - Amanda Lopez
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Alina Nicorici
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Corey Owens
- UC Davis Center for Health and Technology, University of California, Davis, CA 95616, USA;
| | - David Rodriguez
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Jane Wang
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Huanle Zhang
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Daniel Aranki
- Berkeley School of Information, University of California Berkeley, Berkeley, CA 94720, USA;
| | - Craig M. McDonald
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
| | - Erik K. Henricson
- Department of Physical Medicine and Rehabilitation, School of Medicine, University of California, Davis, CA 95616, USA; (K.B.); (E.G.); (L.B.K.); (A.L.); (A.N.); (D.R.); (J.W.); (H.Z.); (C.M.M.)
- Graduate Group in Computer Science (GGCS), University of California, Davis, CA 95616, USA
| |
Collapse
|
171
|
Bekollari M, Dettoraki M, Stavrou V, Glotsos D, Liaparinos P. Computer-Aided Discrimination of Glaucoma Patients from Healthy Subjects Using the RETeval Portable Device. Diagnostics (Basel) 2024; 14:349. [PMID: 38396388 PMCID: PMC10888400 DOI: 10.3390/diagnostics14040349] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Revised: 01/29/2024] [Accepted: 02/01/2024] [Indexed: 02/25/2024] Open
Abstract
Glaucoma is a chronic, progressive eye disease affecting the optic nerve, which may cause visual damage and blindness. In this study, we present a machine-learning investigation to classify patients with glaucoma (case group) with respect to normal participants (control group). We examined 172 eyes at the Ophthalmology Clinic of the "Elpis" General Hospital of Athens between October 2022 and September 2023. In addition, we investigated the glaucoma classification in terms of the following: (a) eye selection and (b) gender. Our methodology was based on the features extracted via two diagnostic optical systems: (i) conventional optical coherence tomography (OCT) and (ii) a modern RETeval portable device. The machine-learning approach comprised three different classifiers: the Bayesian, the Probabilistic Neural Network (PNN), and Support Vectors Machines (SVMs). For all cases examined, classification accuracy was found to be significantly higher when using the RETeval device with respect to the OCT system, as follows: 14.7% for all participants, 13.4% and 29.3% for eye selection (right and left, respectively), and 25.6% and 22.6% for gender (male and female, respectively). The most efficient classifier was found to be the SVM compared to the PNN and Bayesian classifiers. In summary, all aforementioned comparisons demonstrate that the RETeval device has the advantage over the OCT system for the classification of glaucoma patients by using the machine-learning approach.
Collapse
Affiliation(s)
- Marsida Bekollari
- Department of Biomedical Engineering, University of West Attica, Ag. Spyridonos, 12243 Athens, Greece; (M.B.); (D.G.)
| | - Maria Dettoraki
- Department of Ophthalmology, “Elpis” General Hospital, 11522 Athens, Greece
| | - Valentina Stavrou
- Department of Ophthalmology, “Elpis” General Hospital, 11522 Athens, Greece
| | - Dimitris Glotsos
- Department of Biomedical Engineering, University of West Attica, Ag. Spyridonos, 12243 Athens, Greece; (M.B.); (D.G.)
| | - Panagiotis Liaparinos
- Department of Biomedical Engineering, University of West Attica, Ag. Spyridonos, 12243 Athens, Greece; (M.B.); (D.G.)
| |
Collapse
|
172
|
Pei Z, Zhang Q, Qi Y, Wen Z, Zhang Z. Identification of the normative use of medical protective equipment by fusion of object detection and keypoints detection. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 244:107972. [PMID: 38070391 DOI: 10.1016/j.cmpb.2023.107972] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/29/2023] [Revised: 11/24/2023] [Accepted: 12/05/2023] [Indexed: 01/26/2024]
Abstract
BACKGROUND AND OBJECTIVE Irregular use of personal protective equipment (PPE) seriously affects the occupational health and safety of healthcare workers, especially in large public health emergencies such as COVID-19. The danger often occurs in some complex scenarios where the use of certain equipment does not comply with the spatial rules it specifies. The single object detection-based method is difficult to effectively verify whether the worker's use of PPE is normative. Also detecting the use of six classes of PPE brings a large computational load to the task. METHODS In this paper, we proposed an identification approach that combined human keypoints detection with deep learning object detection to help facilitate the monitoring of healthcare workers' standard PPE use. We used YOLOv4 as the baseline model for PPE detection and MobileNetv3 as the backbone of the detector to reduce the computational effort. In addition, High-Resolution Net (HRNet) was the benchmark for keypoints detection, characterizing the coordinates of 25 key points in the human body. Generalized Intersection over Union (GIoU) was used to establish the association between PPEs and key points, and by calculating the matching score between a PPE and the corresponding body bounding boxes, the authenticity of the PPE specification could be effectively inferred. RESULTS The proposed detector is able to identify whether a healthcare worker is normatively using multiple equipment with a higher precision (95.81 %), recall (96.38 %), and F1-score (96.09 %). Meanwhile, the number of parameters (2.87 M) and the size of the model (6.4 MB) are also more lightweight than other comparative detectors. CONCLUSIONS Our approach is more reliable for reasoning about the normality of personal protection for healthcare workers in some complex scenarios than a single object detection-based approach. The developed identification framework provides a new automated monitoring solution for protection management in healthcare, and the modular design brings more flexible applications for different medical operation scenarios.
Collapse
Affiliation(s)
- Ziyu Pei
- Department of Computer Science and Engineering, Northwest Normal University, #967 Anning east RD, Lanzhou, Gansu Province 730070, China
| | - Qiang Zhang
- Department of Computer Science and Engineering, Northwest Normal University, #967 Anning east RD, Lanzhou, Gansu Province 730070, China.
| | - Ying Qi
- Department of Computer Science and Engineering, Northwest Normal University, #967 Anning east RD, Lanzhou, Gansu Province 730070, China
| | - Zexin Wen
- Department of Computer Science and Engineering, Northwest Normal University, #967 Anning east RD, Lanzhou, Gansu Province 730070, China
| | - Zheng Zhang
- Department of Computer Science and Engineering, Northwest Normal University, #967 Anning east RD, Lanzhou, Gansu Province 730070, China
| |
Collapse
|
173
|
Zheng L, Shi S, Lu M, Fang P, Pan Z, Zhang H, Zhou Z, Zhang H, Mou M, Huang S, Tao L, Xia W, Li H, Zeng Z, Zhang S, Chen Y, Li Z, Zhu F. AnnoPRO: a strategy for protein function annotation based on multi-scale protein representation and a hybrid deep learning of dual-path encoding. Genome Biol 2024; 25:41. [PMID: 38303023 PMCID: PMC10832132 DOI: 10.1186/s13059-024-03166-1] [Citation(s) in RCA: 32] [Impact Index Per Article: 32.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2023] [Accepted: 01/05/2024] [Indexed: 02/03/2024] Open
Abstract
Protein function annotation has been one of the longstanding issues in biological sciences, and various computational methods have been developed. However, the existing methods suffer from a serious long-tail problem, with a large number of GO families containing few annotated proteins. Herein, an innovative strategy named AnnoPRO was therefore constructed by enabling sequence-based multi-scale protein representation, dual-path protein encoding using pre-training, and function annotation by long short-term memory-based decoding. A variety of case studies based on different benchmarks were conducted, which confirmed the superior performance of AnnoPRO among available methods. Source code and models have been made freely available at: https://github.com/idrblab/AnnoPRO and https://zenodo.org/records/10012272.
Collapse
Affiliation(s)
- Lingyan Zheng
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
- Industry Solutions Research and Development, Alibaba Cloud Computing, Hangzhou, 330110, China
| | - Shuiyang Shi
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Mingkun Lu
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Pan Fang
- Industry Solutions Research and Development, Alibaba Cloud Computing, Hangzhou, 330110, China
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou, 330110, China
| | - Ziqi Pan
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Hongning Zhang
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Zhimeng Zhou
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Hanyu Zhang
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Minjie Mou
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Shijie Huang
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China
| | - Lin Tao
- Key Laboratory of Elemene Class Anti-Cancer Chinese Medicines, Engineering Laboratory of Development and Application of Traditional Chinese Medicines, Collaborative Innovation Center of Traditional Chinese Medicines of Zhejiang Province, School of Pharmacy, Hangzhou Normal University, Hangzhou, 311121, China
| | - Weiqi Xia
- Pharmaceutical Department, Zhejiang Provincial People's Hospital, Hangzhou, 310014, China
| | - Honglin Li
- School of Pharmacy, East China University of Science and Technology, Shanghai, 200237, China
| | - Zhenyu Zeng
- Industry Solutions Research and Development, Alibaba Cloud Computing, Hangzhou, 330110, China
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou, 330110, China
| | - Shun Zhang
- Industry Solutions Research and Development, Alibaba Cloud Computing, Hangzhou, 330110, China
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou, 330110, China
| | - Yuzong Chen
- State Key Laboratory of Chemical Oncogenomics, Key Laboratory of Chemical Biology, The Graduate School at Shenzhen, Tsinghua University, Shenzhen, 518055, China
| | - Zhaorong Li
- Industry Solutions Research and Development, Alibaba Cloud Computing, Hangzhou, 330110, China.
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou, 330110, China.
| | - Feng Zhu
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou, 310058, China.
- Industry Solutions Research and Development, Alibaba Cloud Computing, Hangzhou, 330110, China.
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou, 330110, China.
| |
Collapse
|
174
|
Shoaib MR, Emara HM, Zhao J, El-Shafai W, Soliman NF, Mubarak AS, Omer OA, El-Samie FEA, Esmaiel H. Deep learning innovations in diagnosing diabetic retinopathy: The potential of transfer learning and the DiaCNN model. Comput Biol Med 2024; 169:107834. [PMID: 38159396 DOI: 10.1016/j.compbiomed.2023.107834] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2023] [Revised: 10/31/2023] [Accepted: 12/05/2023] [Indexed: 01/03/2024]
Abstract
Diabetic retinopathy (DR) is a significant cause of vision impairment, emphasizing the critical need for early detection and timely intervention to avert visual deterioration. Diagnosing DR is inherently complex, as it necessitates the meticulous examination of intricate retinal images by experienced specialists. This makes the early diagnosis of DR essential for effective treatment and prevention of eventual blindness. Traditional diagnostic methods, relying on human interpretation of medical images, face challenges in terms of accuracy and efficiency. In the present research, we introduce a novel method that offers superior precision in DR diagnosis, compared to traditional methods, by employing advanced deep learning techniques. Central to this approach is the concept of transfer learning. This entails the utilization of pre-existing, well-established models, specifically InceptionResNetv2 and Inceptionv3, to extract features and fine-tune selected layers to cater to the unique requirements of this specific diagnostic task. Concurrently, we also present a newly devised model, DiaCNN, which is tailored for the classification of eye diseases. To prove the efficacy of the proposed methodology, we leveraged the Ocular Disease Intelligent Recognition (ODIR) dataset, which comprises eight different eye disease categories. The results are promising. The InceptionResNetv2 model, incorporating transfer learning, registered an impressive 97.5% accuracy in both the training and testing phases. Its counterpart, the Inceptionv3 model, achieved an even more commendable 99.7% accuracy during training, and 97.5% during testing. Remarkably, the DiaCNN model showcased unparalleled precision, achieving 100% accuracy in training and 98.3% in testing. These figures represent a significant leap in classification accuracy when juxtaposed with existing state-of-the-art diagnostic methods. Such advancements hold immense promise for the future, emphasizing the potential of our proposed technique to revolutionize the accuracy of DR and other eye disease diagnoses. By facilitating earlier detection and more timely interventions, this approach stands poised to significantly reduce the incidence of blindness associated with DR, thus heralding a new era of improved patient outcomes. Therefore, this work, through its novel approach and stellar results, not only pushes the boundaries of DR diagnostic accuracy but also promises a transformative impact in early detection and intervention, aiming to substantially diminish DR-induced blindness and champion enhanced patient care.
Collapse
Affiliation(s)
- Mohamed R Shoaib
- School of Computer Science and Engineering, Nanyang Technological University, Singapore, 639798, Singapore
| | - Heba M Emara
- Department of Electronics and Electrical Communications Engineering, Ministry of Higher Education Pyramids Higher Institute (PHI) for Engineering and Technology, 6th of October, 12566, Egypt
| | - Jun Zhao
- School of Computer Science and Engineering, Nanyang Technological University, Singapore, 639798, Singapore.
| | - Walid El-Shafai
- Security Engineering Lab, Computer Science Department, Prince Sultan University, Riyadh, 11586, Saudi Arabia; Department of Electronics and Electrical Communications Engineering, Faculty of Electronic Engineering, Menoufia University, Menouf, 32952, Egypt
| | - Naglaa F Soliman
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, 11671, Saudi Arabia
| | - Ahmed S Mubarak
- Electrical Engineering Department, Aswan Faculty of Engineering, Aswan University, Aswan, 81542, Egypt
| | - Osama A Omer
- Electrical Engineering Department, Aswan Faculty of Engineering, Aswan University, Aswan, 81542, Egypt
| | - Fathi E Abd El-Samie
- Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, 11671, Saudi Arabia
| | - Hamada Esmaiel
- Electrical Engineering Department, Aswan Faculty of Engineering, Aswan University, Aswan, 81542, Egypt
| |
Collapse
|
175
|
Liu C, Hu Y, Chen Y, Fang J, Liu R, Bi L, Tan X, Sheng B, Wu Q. Improvements to a GLCM-based machine-learning approach for quantifying posterior capsule opacification. J Appl Clin Med Phys 2024; 25:e14268. [PMID: 38259111 PMCID: PMC10860560 DOI: 10.1002/acm2.14268] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Revised: 12/26/2023] [Accepted: 01/02/2024] [Indexed: 01/24/2024] Open
Abstract
BACKGROUND Posterior capsular opacification (PCO) is a common complication following cataract surgery that leads to visual disturbances and decreased quality of vision. The aim of our study was to employ a machine-learning methodology to characterize and validate enhancements applied to the grey-level co-occurrence matrix (GLCM) while assessing its validity in comparison to clinical evaluations for evaluating PCO. METHODS One hundred patients diagnosed with age-related cataracts who were scheduled for phacoemulsification surgery were included in the study. Following mydriasis, anterior segment photographs were captured using a high-resolution photographic system. The GLCM was utilized as the feature extractor, and a supported vector machine as the regressor. Three variations, namely, GLCM, GLCM+C (+axial information), and GLCM+V (+regional voting), were analyzed. The reference value for regression was determined by averaging clinical scores obtained through subjective analysis. The relationships between the predicted PCO outcome scores and the ground truth were assessed using Pearson correlation analysis and a Bland-Altman plot, while agreement between them was assessed through the Bland-Altman plot. RESULTS Relative to the ground truth, the GLCM, GLCM+C, and GLCM+V methods exhibited correlation coefficients of 0.706, 0.768, and 0.829, respectively. The relationship between the PCO score predicted by the GLCM+V method and the ground truth was statistically significant (p < 0.001). Furthermore, the GLCM+V method demonstrated competitive performance comparable to that of two experienced clinicians (r = 0.825, 0.843) and superior to that of two junior clinicians (r = 0.786, 0.756). Notably, a high level of agreement was observed between predictions and the ground truth, without significant evidence of proportional bias (p > 0.05). CONCLUSIONS Overall, our findings suggest that a machine-learning approach incorporating the GLCM, specifically the GLCM+V method, holds promise as an objective and reliable tool for assessing PCO progression. Further studies in larger patient cohorts are warranted to validate these findings and explore their potential clinical applications.
Collapse
Affiliation(s)
- Chang Liu
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Ying Hu
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Yan Chen
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Jian Fang
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Ruhan Liu
- Furong Laboratory, Central South University, Changsha, Hunan, China
| | - Lei Bi
- Institute of Translational Medicine, Shanghai Jiao Tong University, Shanghai, China
| | - Xunan Tan
- Shanghai University of Sport School of Exercise and Health, Shanghai, China
| | - Bin Sheng
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Qiang Wu
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| |
Collapse
|
176
|
Zheng D. Transfer learning-based English translation text classification in a multimedia network environment. PeerJ Comput Sci 2024; 10:e1842. [PMID: 38435557 PMCID: PMC10909173 DOI: 10.7717/peerj-cs.1842] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Accepted: 01/09/2024] [Indexed: 03/05/2024]
Abstract
In recent years, with the rapid development of the Internet and multimedia technology, English translation text classification has played an important role in various industries. However, English translation remains a complex and difficult problem. Seeking an efficient and accurate English translation method has become an urgent problem to be solved. The study first elucidated the possibility of the development of transfer learning technology in multimedia environments, which was recognized. Then, previous research on this issue, as well as the Bidirectional Encoder Representations from Transformers (BERT) model, the attention mechanism and bidirectional long short-term memory (Att-BILSTM) model, and the transfer learning based cross domain model (TLCM) and their theoretical foundations, were comprehensively explained. Through the application of transfer learning in multimedia network technology, we deconstructed and integrated these methods. A new text classification technology fusion model, the BATCL transfer learning model, has been established. We analyzed its requirements and label classification methods, proposed a data preprocessing method, and completed experiments to analyze different influencing factors. The research results indicate that the classification system obtained from the study has a similar trend to the BERT model at the macro level, and the classification method proposed in this study can surpass the BERT model by up to 28%. The classification accuracy of the Att-BILSTM model improves over time, but it does not exceed the classification accuracy of the method proposed in this study. This study not only helps to improve the accuracy of English translation, but also enhances the efficiency of machine learning algorithms, providing a new approach for solving English translation problems.
Collapse
Affiliation(s)
- Danyang Zheng
- School of Foreign Languages, Tianjin University, Tianjin, China
| |
Collapse
|
177
|
Thanchomnang T, Chaibutr N, Maleewong W, Janwan P. Automatic detection of Opisthorchis viverrini egg in stool examination using convolutional-based neural networks. PeerJ 2024; 12:e16773. [PMID: 38313031 PMCID: PMC10836206 DOI: 10.7717/peerj.16773] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Accepted: 12/18/2023] [Indexed: 02/06/2024] Open
Abstract
Background Human opisthorchiasis is a dangerous infectious chronic disease distributed in many Asian areas in the water-basins of large rivers, Siberia, and Europe. The gold standard for human opisthorchiasis laboratory diagnosis is the routine examination of Opisthorchis spp. eggs under a microscope. Manual detection is laborious, time-consuming, and dependent on the microscopist's abilities and expertise. Automatic screening of Opisthorchis spp. eggs with deep learning techniques is a useful diagnostic aid. Methods Herein, we propose a convolutional neural network (CNN) for classifying and automatically detecting O. viverrini eggs from digitized images. The image data acquisition was acquired from infected human feces and was processed using the gold standard formalin ethyl acetate concentration technique, and then captured under the microscope digital camera at 400x. Microscopic images containing artifacts and O.viverrini egg were augmented using image rotation, filtering, noising, and sharpening techniques. This augmentation increased the image dataset from 1 time to 36 times in preparation for the training and validation step. Furthermore, the overall dataset was subdivided into a training-validation and test set at an 80:20 ratio, trained with a five-fold cross-validation to test model stability. For model training, we customized a CNN for image classification. An object detection method was proposed using a patch search algorithm to detect eggs and their locations. A performance matrix was used to evaluate model efficiency after training and IoU analysis for object detection. Results The proposed model, initially trained on non-augmented data of artifacts (class 0) and O. viverrini eggs (class 1), showed limited performance with 50.0% accuracy, 25.0% precision, 50.0% recall, and a 33.0% F1-score. After implementing data augmentation, the model significantly improved, reaching 100% accuracy, precision, recall, and F1-score. Stability assessments using 5-fold cross-validation indicated better stability with augmented data, evidenced by an ROC-AUC metric improvement from 0.5 to 1.00. Compared to other models such as ResNet50, InceptionV3, VGG16, DenseNet121, and Xception, the proposed model, with a smaller file size of 2.7 MB, showed comparable perfect performance. In object detection, the augmented data-trained model achieved an IoU score over 0.5 in 139 out of 148 images, with an average IoU of 0.6947. Conclusion This study demonstrated the successful application of CNN in classifying and automating the detection of O. viverrini eggs in human stool samples. Our CNN model's performance metrics and true positive detection rates were outstanding. This innovative application of deep learning can automate and improve diagnostic precision, speed, and efficiency, particularly in regions where O. viverrini infections are prevalent, thereby possibly improving infection sustainable control and treatment program.
Collapse
Affiliation(s)
| | - Natthanai Chaibutr
- Medical Innovation and Technology Program, School of Allied Health Sciences, Walailak University, Nakhon Si Thammarat, Thailand
- Hematology and Transfusion Science Research Center, Walailak University, Nakhon Si Thammarat, Thailand
| | - Wanchai Maleewong
- Department of Parasitology, Faculty of Medicine, Khon Kaen University, Khon Kaen, Thailand
- Mekong Health Science Research Institute, Khon Kaen University, Khon Kaen, Thailand
| | - Penchom Janwan
- Medical Innovation and Technology Program, School of Allied Health Sciences, Walailak University, Nakhon Si Thammarat, Thailand
- Hematology and Transfusion Science Research Center, Walailak University, Nakhon Si Thammarat, Thailand
- Department of Medical Technology, School of Allied Health Sciences, Walailak University, Nakhon Si Thammarat, Thailand
| |
Collapse
|
178
|
Ocampo J, Heyes G, Dehghani H, Scanlon T, Jolly S, Gibson A. Determination of output factor for CyberKnife using scintillation dosimetry and deep learning. Phys Med Biol 2024; 69:025024. [PMID: 38181420 DOI: 10.1088/1361-6560/ad1b69] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 01/05/2024] [Indexed: 01/07/2024]
Abstract
Objective. Small-field dosimetry is an ongoing challenge in radiotherapy quality assurance (QA) especially for radiosurgery systems such as CyberKnifeTM. The objective of this work is to demonstrate the use of a plastic scintillator imaged with a commercial camera to measure the output factor of a CyberKnife system. The output factor describes the dose on the central axis as a function of collimator size, and is a fundamental part of CyberKnife QA and integral to the data used in the treatment planning system.Approach. A self-contained device consisting of a solid plastic scintillator and a camera was build in a portable Pelicase. Photographs were analysed using classical methods and with convolutional neural networks (CNN) to predict beam parameters which were then compared to measurements.Main results. Initial results using classical image processing to determine standard QA parameters such as percentage depth dose (PDD) were unsuccessful, with 34% of points failing to meet the Gamma criterion (which measures the distance between corresponding points and the relative difference in dose) of 2 mm/2%. However, when images were processed using a CNN trained on simulated data and a green scintillator sheet, 92% of PDD curves agreed with measurements with a microdiamond detector to within 2 mm/2% and 78% to 1%/1 mm. The mean difference between the output factors measured using this system and a microdiamond detector was 1.1%. Confidence in the results was enhanced by using the algorithm to predict the known collimator sizes from the photographs which it was able to do with an accuracy of less than 1 mm.Significance. With refinement, a full output factor curve could be measured in less than an hour, offering a new approach for rapid, convenient small-field dosimetry.
Collapse
Affiliation(s)
- Jeremy Ocampo
- UCL Physics and Astronomy, London, WC1E 6BT, United Kingdom
| | - Geoff Heyes
- Radiotherapy Physics, University Hospitals Birmingham NHS Foundation Trust, Birmingham B15 2TH, United Kingdom
| | - Hamid Dehghani
- School of Computer Science, University of Birmingham, Birmingham B15 2TT, United Kingdom
| | - Tim Scanlon
- UCL Physics and Astronomy, London, WC1E 6BT, United Kingdom
| | - Simon Jolly
- UCL Physics and Astronomy, London, WC1E 6BT, United Kingdom
| | - Adam Gibson
- UCL Medical Physics & Biomedical Engineering, London, WC1E 6BT, United Kingdom
| |
Collapse
|
179
|
Zeng J, Gao X, Gao L, Yu Y, Shen L, Pan X. Recognition of rare antinuclear antibody patterns based on a novel attention-based enhancement framework. Brief Bioinform 2024; 25:bbad531. [PMID: 38279651 PMCID: PMC10818137 DOI: 10.1093/bib/bbad531] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 12/17/2023] [Accepted: 12/19/2023] [Indexed: 01/28/2024] Open
Abstract
Rare antinuclear antibody (ANA) pattern recognition has been a widely applied technology for routine ANA screening in clinical laboratories. In recent years, the application of deep learning methods in recognizing ANA patterns has witnessed remarkable advancements. However, the majority of studies in this field have primarily focused on the classification of the most common ANA patterns, while another subset has concentrated on the detection of mitotic metaphase cells. To date, no prior research has been specifically dedicated to the identification of rare ANA patterns. In the present paper, we introduce a novel attention-based enhancement framework, which was designed for the recognition of rare ANA patterns in ANA-indirect immunofluorescence images. More specifically, we selected the algorithm with the best performance as our target detection network by conducting comparative experiments. We then further developed and enhanced the chosen algorithm through a series of optimizations. Then, attention mechanism was introduced to facilitate neural networks in expediting the learning process, extracting more essential and distinctive features for the target features that belong to the specific patterns. The proposed approach has helped to obtained high precision rate of 86.40%, 82.75% recall, 84.24% F1 score and 84.64% mean average precision for a 9-category rare ANA pattern detection task on our dataset. Finally, we evaluated the potential of the model as medical technologist assistant and observed that the technologist's performance improved after referring to the results of the model prediction. These promising results highlighted its potential as an efficient and reliable tool to assist medical technologists in their clinical practice.
Collapse
Affiliation(s)
- Junxiang Zeng
- Department of Clinical Laboratory, Xinhua Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
- Faculty of Medical Laboratory Science, College of Health Science and Technology, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Institute of Artificial Intelligence Medicine, Shanghai Academy of Experimental Medicine, Shanghai, China
| | - Xiupan Gao
- Department of Clinical Laboratory, Xinhua Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
| | - Limei Gao
- Department of Immunology and Rheumatology, Xinhua Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
| | - Youyou Yu
- Department of Clinical Laboratory, Xinhua Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
| | - Lisong Shen
- Department of Clinical Laboratory, Xinhua Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
- Faculty of Medical Laboratory Science, College of Health Science and Technology, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Institute of Artificial Intelligence Medicine, Shanghai Academy of Experimental Medicine, Shanghai, China
| | - Xiujun Pan
- Department of Clinical Laboratory, Xinhua Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
| |
Collapse
|
180
|
Niu ZB, Jia SY, Xu HH. Automated graptolite identification at high taxonomic resolution using residual networks. iScience 2024; 27:108549. [PMID: 38213629 PMCID: PMC10783601 DOI: 10.1016/j.isci.2023.108549] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Revised: 08/23/2023] [Accepted: 11/20/2023] [Indexed: 01/13/2024] Open
Abstract
Graptolites, fossils significant for evolutionary studies and shale gas exploration, are traditionally identified visually by taxonomists due to their intricate morphologies and preservation challenges. Artificial intelligence (AI) holds great promise for transforming such meticulous tasks. In this paper, we demonstrate that graptolites can be identified with taxonomist accuracy using a deep learning model. We construct the most sophisticated and largest professional single organisms image dataset to date, which is composed of >34,000 images of 113 graptolite species annotated at pixel-level resolution to train the model, develop, and evaluate deep learning networks to classify graptolites. The model's performance surpassed taxonomists in accuracy, time, and generalization, achieving 86% and 81% accuracy in identifying graptolite genus and species, respectively. This AI-based method, capable of recognizing minute morphological details better than taxonomists, can be integrated into web and mobile apps, extending graptolite identification beyond research institutes and enhancing shale gas exploration efficiency.
Collapse
Affiliation(s)
- Zhi-Bin Niu
- College of Intelligence and Computing, Tianjin University, Tianjin 300354, China
- State Key Laboratory of Palaeobiology and Stratigraphy, Nanjing Institute of Geology and Palaeontology and Centre for Excellence in Life and Paleoenvironment, Chinese Academy of Sciences, Nanjing 210008, China
| | - Si-Yuan Jia
- College of Intelligence and Computing, Tianjin University, Tianjin 300354, China
| | - Hong-He Xu
- State Key Laboratory of Palaeobiology and Stratigraphy, Nanjing Institute of Geology and Palaeontology and Centre for Excellence in Life and Paleoenvironment, Chinese Academy of Sciences, Nanjing 210008, China
| |
Collapse
|
181
|
Raut P, Baldini G, Schöneck M, Caldeira L. Using a generative adversarial network to generate synthetic MRI images for multi-class automatic segmentation of brain tumors. FRONTIERS IN RADIOLOGY 2024; 3:1336902. [PMID: 38304344 PMCID: PMC10830800 DOI: 10.3389/fradi.2023.1336902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/11/2023] [Accepted: 12/28/2023] [Indexed: 02/03/2024]
Abstract
Challenging tasks such as lesion segmentation, classification, and analysis for the assessment of disease progression can be automatically achieved using deep learning (DL)-based algorithms. DL techniques such as 3D convolutional neural networks are trained using heterogeneous volumetric imaging data such as MRI, CT, and PET, among others. However, DL-based methods are usually only applicable in the presence of the desired number of inputs. In the absence of one of the required inputs, the method cannot be used. By implementing a generative adversarial network (GAN), we aim to apply multi-label automatic segmentation of brain tumors to synthetic images when not all inputs are present. The implemented GAN is based on the Pix2Pix architecture and has been extended to a 3D framework named Pix2PixNIfTI. For this study, 1,251 patients of the BraTS2021 dataset comprising sequences such as T1w, T2w, T1CE, and FLAIR images equipped with respective multi-label segmentation were used. This dataset was used for training the Pix2PixNIfTI model for generating synthetic MRI images of all the image contrasts. The segmentation model, namely DeepMedic, was trained in a five-fold cross-validation manner for brain tumor segmentation and tested using the original inputs as the gold standard. The inference of trained segmentation models was later applied to synthetic images replacing missing input, in combination with other original images to identify the efficacy of generated images in achieving multi-class segmentation. For the multi-class segmentation using synthetic data or lesser inputs, the dice scores were observed to be significantly reduced but remained similar in range for the whole tumor when compared with evaluated original image segmentation (e.g. mean dice of synthetic T2w prediction NC, 0.74 ± 0.30; ED, 0.81 ± 0.15; CET, 0.84 ± 0.21; WT, 0.90 ± 0.08). A standard paired t-tests with multiple comparison correction were performed to assess the difference between all regions (p < 0.05). The study concludes that the use of Pix2PixNIfTI allows us to segment brain tumors when one input image is missing.
Collapse
Affiliation(s)
- P. Raut
- Department of Pediatric Pulmonology, Erasmus Medical Center, Rotterdam, Netherlands
- Department of Radiology & Nuclear Medicine, Erasmus Medical Center, Rotterdam, Netherlands
- Institute for Diagnostic and Interventional Radiology, University Hospital Cologne, Cologne, Germany
| | - G. Baldini
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
| | - M. Schöneck
- Institute for Diagnostic and Interventional Radiology, University Hospital Cologne, Cologne, Germany
| | - L. Caldeira
- Institute for Diagnostic and Interventional Radiology, University Hospital Cologne, Cologne, Germany
| |
Collapse
|
182
|
Tawhid MNA, Siuly S, Kabir E, Li Y. Exploring Frequency Band-Based Biomarkers of EEG Signals for Mild Cognitive Impairment Detection. IEEE Trans Neural Syst Rehabil Eng 2024; 32:189-199. [PMID: 38145525 DOI: 10.1109/tnsre.2023.3347032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Mild Cognitive Impairment (MCI) is often considered a precursor to Alzheimer's disease (AD), with a high likelihood of progression. Accurate and timely diagnosis of MCI is essential for halting the progression of AD and other forms of dementia. Electroencephalography (EEG) is the prevalent method for identifying MCI biomarkers. Frequency band-based EEG biomarkers are crucial for identifying MCI as they capture neuronal activities and connectivity patterns linked to cognitive functions. However, traditional approaches struggle to identify precise frequency band-based biomarkers for MCI diagnosis. To address this challenge, a novel framework has been developed for identifying important frequency sub-bands within EEG signals for MCI detection. In the proposed scheme, the signals are first denoised using a stationary wavelet transformation and segmented into small time frames. Then, four frequency sub-bands are extracted from each segment, and spectrogram images are generated for each sub-band as well as for the full filtered frequency band signal segments. This process produces five different sets of images for five separate frequency bands. Afterwards, a convolutional neural network is used individually on those image sets to perform the classification task. Finally, the obtained results for the tested four sub-bands are compared with the results obtained using the full bandwidth. Our proposed framework was tested on two MCI datasets, and the results indicate that the 16-32 Hz sub-band range has the greatest impact on MCI detection, followed by 4-8 Hz. Furthermore, our framework, utilizing the full frequency band, outperformed existing state-of-the-art methods, indicating its potential for developing diagnostic tools for MCI detection.
Collapse
|
183
|
Tripathi S, Tabari A, Mansur A, Dabbara H, Bridge CP, Daye D. From Machine Learning to Patient Outcomes: A Comprehensive Review of AI in Pancreatic Cancer. Diagnostics (Basel) 2024; 14:174. [PMID: 38248051 PMCID: PMC10814554 DOI: 10.3390/diagnostics14020174] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2023] [Revised: 12/28/2023] [Accepted: 12/29/2023] [Indexed: 01/23/2024] Open
Abstract
Pancreatic cancer is a highly aggressive and difficult-to-detect cancer with a poor prognosis. Late diagnosis is common due to a lack of early symptoms, specific markers, and the challenging location of the pancreas. Imaging technologies have improved diagnosis, but there is still room for improvement in standardizing guidelines. Biopsies and histopathological analysis are challenging due to tumor heterogeneity. Artificial Intelligence (AI) revolutionizes healthcare by improving diagnosis, treatment, and patient care. AI algorithms can analyze medical images with precision, aiding in early disease detection. AI also plays a role in personalized medicine by analyzing patient data to tailor treatment plans. It streamlines administrative tasks, such as medical coding and documentation, and provides patient assistance through AI chatbots. However, challenges include data privacy, security, and ethical considerations. This review article focuses on the potential of AI in transforming pancreatic cancer care, offering improved diagnostics, personalized treatments, and operational efficiency, leading to better patient outcomes.
Collapse
Affiliation(s)
- Satvik Tripathi
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA; (S.T.); (A.T.); (A.M.); (C.P.B.)
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA 02129, USA
- Harvard Medical School, Boston, MA 02115, USA
| | - Azadeh Tabari
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA; (S.T.); (A.T.); (A.M.); (C.P.B.)
- Harvard Medical School, Boston, MA 02115, USA
| | - Arian Mansur
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA; (S.T.); (A.T.); (A.M.); (C.P.B.)
- Harvard Medical School, Boston, MA 02115, USA
| | - Harika Dabbara
- Boston University Chobanian & Avedisian School of Medicine, Boston, MA 02118, USA;
| | - Christopher P. Bridge
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA; (S.T.); (A.T.); (A.M.); (C.P.B.)
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA 02129, USA
- Harvard Medical School, Boston, MA 02115, USA
| | - Dania Daye
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA; (S.T.); (A.T.); (A.M.); (C.P.B.)
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA 02129, USA
- Harvard Medical School, Boston, MA 02115, USA
| |
Collapse
|
184
|
Mulugeta AK, Sharma DP, Mesfin AH. Deep learning for medicinal plant species classification and recognition: a systematic review. FRONTIERS IN PLANT SCIENCE 2024; 14:1286088. [PMID: 38250440 PMCID: PMC10796487 DOI: 10.3389/fpls.2023.1286088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 12/12/2023] [Indexed: 01/23/2024]
Abstract
Knowledge of medicinal plant species is necessary to preserve medicinal plants and safeguard biodiversity. The classification and identification of these plants by botanist experts are complex and time-consuming activities. This systematic review's main objective is to systematically assess the prior research efforts on the applications and usage of deep learning approaches in classifying and recognizing medicinal plant species. Our objective was to pinpoint systematic reviews following the PRISMA guidelines related to the classification and recognition of medicinal plant species through the utilization of deep learning techniques. This review encompassed studies published between January 2018 and December 2022. Initially, we identified 1644 studies through title, keyword, and abstract screening. After applying our eligibility criteria, we selected 31 studies for a thorough and critical review. The main findings of this reviews are (1) the selected studies were carried out in 16 different countries, and India leads in paper contributions with 29%, followed by Indonesia and Sri Lanka. (2) A private dataset has been used in 67.7% of the studies subjected to image augmentation and preprocessing techniques. (3) In 96.7% of the studies, researchers have employed plant leaf organs, with 74% of them utilizing leaf shapes for the classification and recognition of medicinal plant species. (4) Transfer learning with the pre-trained model was used in 83.8% of the studies as a future extraction technique. (5) Convolutional Neural Network (CNN) is used by 64.5% of the paper as a deep learning classifier. (6) The lack of a globally available and public dataset need for medicinal plants indigenous to a specific country and the trustworthiness of the deep learning approach for the classification and recognition of medicinal plants is an observable research gap in this literature review. Therefore, further investigations and collaboration between different stakeholders are required to fulfilling the aforementioned research gaps.
Collapse
Affiliation(s)
- Adibaru Kiflie Mulugeta
- Department of Computer Science and Engineering, School of Electrical Engineering and Computing, Adama Science and Technology University, Adama, Ethiopia
| | | | - Abebe Haile Mesfin
- Department of Computer Science and Engineering, School of Electrical Engineering and Computing, Adama Science and Technology University, Adama, Ethiopia
| |
Collapse
|
185
|
Van Der Aar JF, Van Den Ende DA, Fonseca P, Van Meulen FB, Overeem S, Van Gilst MM, Peri E. Deep transfer learning for automated single-lead EEG sleep staging with channel and population mismatches. Front Physiol 2024; 14:1287342. [PMID: 38250654 PMCID: PMC10796543 DOI: 10.3389/fphys.2023.1287342] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Accepted: 12/08/2023] [Indexed: 01/23/2024] Open
Abstract
Introduction: Automated sleep staging using deep learning models typically requires training on hundreds of sleep recordings, and pre-training on public databases is therefore common practice. However, suboptimal sleep stage performance may occur from mismatches between source and target datasets, such as differences in population characteristics (e.g., an unrepresented sleep disorder) or sensors (e.g., alternative channel locations for wearable EEG). Methods: We investigated three strategies for training an automated single-channel EEG sleep stager: pre-training (i.e., training on the original source dataset), training-from-scratch (i.e., training on the new target dataset), and fine-tuning (i.e., training on the original source dataset, fine-tuning on the new target dataset). As source dataset, we used the F3-M2 channel of healthy subjects (N = 94). Performance of the different training strategies was evaluated using Cohen's Kappa (κ) in eight smaller target datasets consisting of healthy subjects (N = 60), patients with obstructive sleep apnea (OSA, N = 60), insomnia (N = 60), and REM sleep behavioral disorder (RBD, N = 22), combined with two EEG channels, F3-M2 and F3-F4. Results: No differences in performance between the training strategies was observed in the age-matched F3-M2 datasets, with an average performance across strategies of κ = .83 in healthy, κ = .77 in insomnia, and κ = .74 in OSA subjects. However, in the RBD set, where data availability was limited, fine-tuning was the preferred method (κ = .67), with an average increase in κ of .15 to pre-training and training-from-scratch. In the presence of channel mismatches, targeted training is required, either through training-from-scratch or fine-tuning, increasing performance with κ = .17 on average. Discussion: We found that, when channel and/or population mismatches cause suboptimal sleep staging performance, a fine-tuning approach can yield similar to superior performance compared to building a model from scratch, while requiring a smaller sample size. In contrast to insomnia and OSA, RBD data contains characteristics, either inherent to the pathology or age-related, which apparently demand targeted training.
Collapse
Affiliation(s)
- Jaap F. Van Der Aar
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Philips Research, Eindhoven, Netherlands
| | | | - Pedro Fonseca
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Philips Research, Eindhoven, Netherlands
| | - Fokke B. Van Meulen
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Kempenhaeghe Center for Sleep Medicine, Heeze, Netherlands
| | - Sebastiaan Overeem
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Kempenhaeghe Center for Sleep Medicine, Heeze, Netherlands
| | - Merel M. Van Gilst
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
- Kempenhaeghe Center for Sleep Medicine, Heeze, Netherlands
| | - Elisabetta Peri
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
| |
Collapse
|
186
|
Liu R, Wang Q, Zhang X. Identification of prognostic coagulation-related signatures in clear cell renal cell carcinoma through integrated multi-omics analysis and machine learning. Comput Biol Med 2024; 168:107779. [PMID: 38061153 DOI: 10.1016/j.compbiomed.2023.107779] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 10/30/2023] [Accepted: 11/28/2023] [Indexed: 01/10/2024]
Abstract
Clear cell renal cell carcinoma is a threat to public health with high morbidity and mortality. Clinical evidence has shown that cancer-associated thrombosis poses significant challenges to treatments, including drug resistance and difficulties in surgical decision-making in ccRCC. However, the coagulation pathway, one of the core mechanisms of cancer-associated thrombosis, recently found closely related to the tumor microenvironment and immune-related pathway, is rarely researched in ccRCC. Therefore, we integrated bulk RNA-seq data, DNA mutation and methylation data, single-cell data, and proteomic data to perform a comprehensive analysis of coagulation-related genes in ccRCC. First, we demonstrated the importance of the coagulation-related gene set by consensus clustering. Based on machine learning, we identified 5 coagulation signature genes and verified their clinical value in TCGA, ICGC, and E-MTAB-1980 databases. It's also demonstrated that the specific expression patterns of coagulation signature genes driven by CNV and methylation were closely correlated with pathways including apoptosis, immune infiltration, angiogenesis, and the construction of extracellular matrix. Moreover, we identified two types of tumor cells in single-cell data by machine learning, and the coagulation signature genes were differentially expressed in two types of tumor cells. Besides, the signature genes were proven to influence immune cells especially the differentiation of T cells. And their protein level was also validated.
Collapse
Affiliation(s)
- Ruijie Liu
- Department of Urology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, China.
| | - Qi Wang
- Department of Urology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, China.
| | - Xiaoping Zhang
- Department of Urology, Union Hospital, Tongji Medical College, Huazhong University of Science and Technology, China.
| |
Collapse
|
187
|
Silveira A, Greving I, Longo E, Scheel M, Weitkamp T, Fleck C, Shahar R, Zaslansky P. Deep learning to overcome Zernike phase-contrast nanoCT artifacts for automated micro-nano porosity segmentation in bone. JOURNAL OF SYNCHROTRON RADIATION 2024; 31:136-149. [PMID: 38095668 PMCID: PMC10833422 DOI: 10.1107/s1600577523009852] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 11/13/2023] [Indexed: 01/09/2024]
Abstract
Bone material contains a hierarchical network of micro- and nano-cavities and channels, known as the lacuna-canalicular network (LCN), that is thought to play an important role in mechanobiology and turnover. The LCN comprises micrometer-sized lacunae, voids that house osteocytes, and submicrometer-sized canaliculi that connect bone cells. Characterization of this network in three dimensions is crucial for many bone studies. To quantify X-ray Zernike phase-contrast nanotomography data, deep learning is used to isolate and assess porosity in artifact-laden tomographies of zebrafish bones. A technical solution is proposed to overcome the halo and shade-off domains in order to reliably obtain the distribution and morphology of the LCN in the tomographic data. Convolutional neural network (CNN) models are utilized with increasing numbers of images, repeatedly validated by `error loss' and `accuracy' metrics. U-Net and Sensor3D CNN models were trained on data obtained from two different synchrotron Zernike phase-contrast transmission X-ray microscopes, the ANATOMIX beamline at SOLEIL (Paris, France) and the P05 beamline at PETRA III (Hamburg, Germany). The Sensor3D CNN model with a smaller batch size of 32 and a training data size of 70 images showed the best performance (accuracy 0.983 and error loss 0.032). The analysis procedures, validated by comparison with human-identified ground-truth images, correctly identified the voids within the bone matrix. This proposed approach may have further application to classify structures in volumetric images that contain non-linear artifacts that degrade image quality and hinder feature identification.
Collapse
Affiliation(s)
- Andreia Silveira
- Department for Restorative, Preventive and Pediatric Dentistry, Charité-Universitaetsmedizin, Berlin, Germany
| | - Imke Greving
- Institute of Materials Physics, Helmholtz-Zentrum Hereon, Geesthacht, Germany
| | - Elena Longo
- Elettra – Sincrotrone Trieste SCpA, Basovizza, Trieste, Italy
| | | | | | - Claudia Fleck
- Fachgebiet Werkstofftechnik / Chair of Materials Science and Engineering, Institute of Materials Science and Technology, Faculty III Process Sciences, Technische Universität Berlin, Berlin, Germany
| | - Ron Shahar
- Koret School of Veterinary Medicine, The Robert H. Smith Faculty of Agriculture, Food and Environmental Sciences, Hebrew University of Jerusalem, Rehovot, Israel
| | - Paul Zaslansky
- Department for Restorative, Preventive and Pediatric Dentistry, Charité-Universitaetsmedizin, Berlin, Germany
| |
Collapse
|
188
|
Su X, Liu W, Jiang S, Gao X, Chu Y, Ma L. Deep learning-based anatomical position recognition for gastroscopic examination. Technol Health Care 2024; 32:39-48. [PMID: 38669495 PMCID: PMC11191429 DOI: 10.3233/thc-248004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/28/2024]
Abstract
BACKGROUND The gastroscopic examination is a preferred method for the detection of upper gastrointestinal lesions. However, gastroscopic examination has high requirements for doctors, especially for the strict position and quantity of the archived images. These requirements are challenging for the education and training of junior doctors. OBJECTIVE The purpose of this study is to use deep learning to develop automatic position recognition technology for gastroscopic examination. METHODS A total of 17182 gastroscopic images in eight anatomical position categories are collected. Convolutional neural network model MogaNet is used to identify all the anatomical positions of the stomach for gastroscopic examination The performance of four models is evaluated by sensitivity, precision, and F1 score. RESULTS The average sensitivity of the method proposed is 0.963, which is 0.074, 0.066 and 0.065 higher than ResNet, GoogleNet and SqueezeNet, respectively. The average precision of the method proposed is 0.964, which is 0.072, 0.067 and 0.068 higher than ResNet, GoogleNet, and SqueezeNet, respectively. And the average F1-Score of the method proposed is 0.964, which is 0.074, 0.067 and 0.067 higher than ResNet, GoogleNet, and SqueezeNet, respectively. The results of the t-test show that the method proposed is significantly different from other methods (p< 0.05). CONCLUSION The method proposed exhibits the best performance for anatomical positions recognition. And the method proposed can help junior doctors meet the requirements of completeness of gastroscopic examination and the number and position of archived images quickly.
Collapse
Affiliation(s)
- Xiufeng Su
- Weihai Municipal Hospital, Cheeloo College of Medicine, Shandong University, Weihai, Shandong, China
| | - Weiyu Liu
- Weihai Municipal Hospital, Cheeloo College of Medicine, Shandong University, Weihai, Shandong, China
| | - Suyi Jiang
- Weihai Municipal Hospital, Cheeloo College of Medicine, Shandong University, Weihai, Shandong, China
| | - Xiaozhong Gao
- Weihai Municipal Hospital, Cheeloo College of Medicine, Shandong University, Weihai, Shandong, China
| | - Yanliu Chu
- Weihai Municipal Hospital, Cheeloo College of Medicine, Shandong University, Weihai, Shandong, China
| | - Liyong Ma
- School of Information Science and Engineering, Harbin Institute of Technology, Weihai, Shandong, China
| |
Collapse
|
189
|
Fang C, Song C, Wen Z, Liu G, Wang X, Li S, Shang Y, Tao H, Lyu L, Song K. A novel chlorophyll-a retrieval model based on suspended particulate matter classification and different machine learning. ENVIRONMENTAL RESEARCH 2024; 240:117430. [PMID: 37866530 DOI: 10.1016/j.envres.2023.117430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/03/2023] [Revised: 10/05/2023] [Accepted: 10/15/2023] [Indexed: 10/24/2023]
Abstract
Chlorophyll-a (Chla) in inland waters is one of the most significant optical parameters of aquatic ecosystem assessment, and long-term and daily Chla concentration monitoring has the potential to facilitate in early warning of algal blooms. MOD09 products have multiple observation advantages (higher temporal, spatial resolution and signal-to-noise ratio), and play an extremely important role in the remote sensing of water color. For developing a high accuracy machine learning model of remotely estimating Chla concentration in inland waters based on MOD09 products, this study proposed an assumption that the accuracy of Chla concentration retrieval will be improved after classifying water bodies into three groups by suspended particulate matter (SPM) concentration. A total of 10 commonly used machine learning models were compared and evaluated in this study, including random forest regressor (RFR), deep neural networks (DNN), extreme gradient boosting (XGBoost), and convolutional neural network (CNN). Altogether, 41 basic bands and 820 band ratios between the 41 bands were filtered by measuring their correlation with Ln(Chla) and several bands brought into different machine learning models. Results demonstrated that the construction of Chla concentration remote estimation model based on SPM classification could significantly improve the correlation between Ln(Chla) and 41 basic spectral band combinations, the correlation between Ln(Chla) and 820 band ratios, and the model verification R2 from 0.41 to 0.83. Furthermore, B3, B20, and B32 were finally selected based on correlation with SPM to classify SPM and the classification accuracy could reach 0.9. Finally, we concluded that RFR model performed best via comparing the R2, RMSE, and MAPE. By comparing the relative contribution of input bands in different groups, B3 contributed most to three groups. The model constructed in this study has promising prospects for promotion and application in other inland waters, and could provide systematic research reference for subsequent research.
Collapse
Affiliation(s)
- Chong Fang
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China
| | - Changchun Song
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China; Faculty of Infrastructure Engineering, Dalian University of Technology, Dalian, 116024, China
| | - Zhidan Wen
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China
| | - Ge Liu
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China
| | - Xiaodi Wang
- School of Geography and Tourism, Harbin University, Harbin, 150086, China
| | - Sijia Li
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China
| | - Yingxin Shang
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China
| | - Hui Tao
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China; University of Chinese Academy of Sciences, Beijing, 100049, China
| | - Lili Lyu
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China; University of Chinese Academy of Sciences, Beijing, 100049, China
| | - Kaishan Song
- Northeast Institute of Geography and Agroecology, Chinese Academy of Sciences, Changchun, 130102, China; School of Environment and Planning, Liaocheng University, Liaocheng, 252000, China.
| |
Collapse
|
190
|
Sharma P, Nayak DR, Balabantaray BK, Tanveer M, Nayak R. A survey on cancer detection via convolutional neural networks: Current challenges and future directions. Neural Netw 2024; 169:637-659. [PMID: 37972509 DOI: 10.1016/j.neunet.2023.11.006] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 10/21/2023] [Accepted: 11/04/2023] [Indexed: 11/19/2023]
Abstract
Cancer is a condition in which abnormal cells uncontrollably split and damage the body tissues. Hence, detecting cancer at an early stage is highly essential. Currently, medical images play an indispensable role in detecting various cancers; however, manual interpretation of these images by radiologists is observer-dependent, time-consuming, and tedious. An automatic decision-making process is thus an essential need for cancer detection and diagnosis. This paper presents a comprehensive survey on automated cancer detection in various human body organs, namely, the breast, lung, liver, prostate, brain, skin, and colon, using convolutional neural networks (CNN) and medical imaging techniques. It also includes a brief discussion about deep learning based on state-of-the-art cancer detection methods, their outcomes, and the possible medical imaging data used. Eventually, the description of the dataset used for cancer detection, the limitations of the existing solutions, future trends, and challenges in this domain are discussed. The utmost goal of this paper is to provide a piece of comprehensive and insightful information to researchers who have a keen interest in developing CNN-based models for cancer detection.
Collapse
Affiliation(s)
- Pallabi Sharma
- School of Computer Science, UPES, Dehradun, 248007, Uttarakhand, India.
| | - Deepak Ranjan Nayak
- Department of Computer Science and Engineering, Malaviya National Institute of Technology, Jaipur, 302017, Rajasthan, India.
| | - Bunil Kumar Balabantaray
- Computer Science and Engineering, National Institute of Technology Meghalaya, Shillong, 793003, Meghalaya, India.
| | - M Tanveer
- Department of Mathematics, Indian Institute of Technology Indore, Simrol, 453552, Indore, India.
| | - Rajashree Nayak
- School of Applied Sciences, Birla Global University, Bhubaneswar, 751029, Odisha, India.
| |
Collapse
|
191
|
Wang SX, Li Y, Zhu JQ, Wang ML, Zhang W, Tie CW, Wang GQ, Ni XG. The Detection of Nasopharyngeal Carcinomas Using a Neural Network Based on Nasopharyngoscopic Images. Laryngoscope 2024; 134:127-135. [PMID: 37254946 DOI: 10.1002/lary.30781] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2022] [Revised: 05/14/2023] [Accepted: 05/15/2023] [Indexed: 06/01/2023]
Abstract
OBJECTIVE To construct and validate a deep convolutional neural network (DCNN)-based artificial intelligence (AI) system for the detection of nasopharyngeal carcinoma (NPC) using archived nasopharyngoscopic images. METHODS We retrospectively collected 14107 nasopharyngoscopic images (7108 NPCs and 6999 noncancers) to construct a DCNN model and prepared a validation dataset containing 3501 images (1744 NPCs and 1757 noncancers) from a single center between January 2009 and December 2020. The DCNN model was established using the You Only Look Once (YOLOv5) architecture. Four otolaryngologists were asked to review the images of the validation set to benchmark the DCNN model performance. RESULTS The DCNN model analyzed the 3501 images in 69.35 s. For the validation dataset, the precision, recall, accuracy, and F1 score of the DCNN model in the detection of NPCs on white light imaging (WLI) and narrow band imaging (NBI) were 0.845 ± 0.038, 0.942 ± 0.021, 0.920 ± 0.024, and 0.890 ± 0.045, and 0.895 ± 0.045, 0.941 ± 0.018, and 0.975 ± 0.013, 0.918 ± 0.036, respectively. The diagnostic outcome of the DCNN model on WLI and NBI images was significantly higher than that of two junior otolaryngologists (p < 0.05). CONCLUSION The DCNN model showed better diagnostic outcomes for NPCs than those of junior otolaryngologists. Therefore, it could assist them in improving their diagnostic level and reducing missed diagnoses. LEVEL OF EVIDENCE 3 Laryngoscope, 134:127-135, 2024.
Collapse
Affiliation(s)
- Shi-Xu Wang
- Department of Head and Neck Surgery, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Ying Li
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Ji-Qing Zhu
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Mei-Ling Wang
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Wei Zhang
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Cheng-Wei Tie
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Gui-Qi Wang
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Xiao-Guang Ni
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| |
Collapse
|
192
|
Li K, Zhu Q, Wu J, Ding J, Liu B, Zhu X, Lin S, Yan W, Li W. DCT-Net: An effective method to diagnose retinal tears from B-scan ultrasound images. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:1110-1124. [PMID: 38303456 DOI: 10.3934/mbe.2024046] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
Retinal tears (RTs) are usually detected by B-scan ultrasound images, particularly for individuals with complex eye conditions. However, traditional manual techniques for reading ultrasound images have the potential to overlook or inaccurately diagnose conditions. Thus, the development of rapid and accurate approaches for the diagnosis of an RT is highly important and urgent. The present study introduces a novel hybrid deep-learning model called DCT-Net to enable the automatic and precise diagnosis of RTs. The implemented model utilizes a vision transformer as the backbone and feature extractor. Additionally, in order to accommodate the edge characteristics of the lesion areas, a novel module called the residual deformable convolution has been incorporated. Furthermore, normalization is employed to mitigate the issue of overfitting and, a Softmax layer has been included to achieve the final classification following the acquisition of the global and local representations. The study was conducted by using both our proprietary dataset and a publicly available dataset. In addition, interpretability of the trained model was assessed by generating attention maps using the attention rollout approach. On the private dataset, the model demonstrated a high level of performance, with an accuracy of 97.78%, precision of 97.34%, recall rate of 97.13%, and an F1 score of 0.9682. On the other hand, the model developed by using the public funds image dataset demonstrated an accuracy of 83.82%, a sensitivity of 82.69% and a specificity of 82.40%. The findings, therefore present a novel framework for the diagnosis of RTs that is characterized by a high degree of efficiency, accuracy and interpretability. Accordingly, the technology exhibits considerable promise and has the potential to serve as a reliable tool for ophthalmologists.
Collapse
Affiliation(s)
- Ke Li
- The First Affiliated Hospital of Wenzhou Medical University, Wenzhou 325035, China
| | - Qiaolin Zhu
- The Eye Hospital, School of Ophthalmology & Optometry, Wenzhou Medical University, Wenzhou 325027, China
| | - Jianzhang Wu
- Oujiang Laboratory (Zhejiang Lab for Regenerative Medicine, Vision and Brain Health), Wenzhou 325000, China
- The Eye Hospital, School of Ophthalmology & Optometry, Wenzhou Medical University, Wenzhou 325027, China
| | - Juntao Ding
- The First Affiliated Hospital of Wenzhou Medical University, Wenzhou 325035, China
| | - Bo Liu
- The First Affiliated Hospital of Wenzhou Medical University, Wenzhou 325035, China
| | - Xixi Zhu
- The Eye Hospital, School of Ophthalmology & Optometry, Wenzhou Medical University, Wenzhou 325027, China
| | - Shishi Lin
- The Eye Hospital, School of Ophthalmology & Optometry, Wenzhou Medical University, Wenzhou 325027, China
| | - Wentao Yan
- The Eye Hospital, School of Ophthalmology & Optometry, Wenzhou Medical University, Wenzhou 325027, China
| | - Wulan Li
- The First Affiliated Hospital of Wenzhou Medical University, Wenzhou 325035, China
| |
Collapse
|
193
|
Watanabe H, Ezawa Y, Matsuyama E, Kondo Y, Hayashi N, Maruyama S, Ogura T, Shimosegawa M. Auto-evaluation of skull radiograph accuracy using unsupervised anomaly detection. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:1151-1162. [PMID: 38943422 DOI: 10.3233/xst-230431] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/01/2024]
Abstract
BACKGROUND Radiography plays an important role in medical care, and accurate positioning is essential for providing optimal quality images. Radiographs with insufficient diagnostic value are rejected, and retakes are required. However, determining the suitability of retaking radiographs is a qualitative evaluation. OBJECTIVE To evaluate skull radiograph accuracy automatically using an unsupervised learning-based autoencoder (AE) and a variational autoencoder (VAE). In this study, we eliminated visual qualitative evaluation and used unsupervised learning to identify skull radiography retakes from the quantitative evaluation. METHODS Five skull phantoms were imaged on radiographs, and 1,680 images were acquired. These images correspond to two categories: normal images captured at appropriate positions and images captured at inappropriate positions. This study verified the discriminatory ability of skull radiographs using anomaly detection methods. RESULTS The areas under the curves for AE and VAE were 0.7060 and 0.6707, respectively, in receiver operating characteristic analysis. Our proposed method showed a higher discrimination ability than those of previous studies which had an accuracy of 52%. CONCLUSIONS Our findings suggest that the proposed method has high classification accuracy in determining the suitability of retaking skull radiographs. Automation of optimal image consideration, whether or not to retake radiographs, contributes to improving operational efficiency in busy X-ray imaging operations.
Collapse
Affiliation(s)
- Haruyuki Watanabe
- School of Radiological Technology, Gunma Prefectural College of Health Sciences, Maebashi, Japan
| | - Yuina Ezawa
- School of Radiological Technology, Gunma Prefectural College of Health Sciences, Maebashi, Japan
| | - Eri Matsuyama
- Faculty of Informatics, The University of Fukuchiyama, Fukuchiyama, Japan
| | - Yohan Kondo
- Graduate School of Health Sciences, Niigata University, Niigata, Japan
| | - Norio Hayashi
- School of Radiological Technology, Gunma Prefectural College of Health Sciences, Maebashi, Japan
| | - Sho Maruyama
- School of Radiological Technology, Gunma Prefectural College of Health Sciences, Maebashi, Japan
| | - Toshihiro Ogura
- School of Radiological Technology, Gunma Prefectural College of Health Sciences, Maebashi, Japan
| | - Masayuki Shimosegawa
- School of Radiological Technology, Gunma Prefectural College of Health Sciences, Maebashi, Japan
| |
Collapse
|
194
|
Mohseninia N, Zamani-Siahkali N, Harsini S, Divband G, Pirich C, Beheshti M. Bone Metastasis in Prostate Cancer: Bone Scan Versus PET Imaging. Semin Nucl Med 2024; 54:97-118. [PMID: 37596138 DOI: 10.1053/j.semnuclmed.2023.07.004] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Accepted: 07/11/2023] [Indexed: 08/20/2023]
Abstract
Prostate cancer is the second most common cause of malignancy among men, with bone metastasis being a significant source of morbidity and mortality in advanced cases. Detecting and treating bone metastasis at an early stage is crucial to improve the quality of life and survival of prostate cancer patients. This objective strongly relies on imaging studies. While CT and MRI have their specific utilities, they also possess certain drawbacks. Bone scintigraphy, although cost-effective and widely available, presents high false-positive rates. The emergence of PET/CT and PET/MRI, with their ability to overcome the limitations of standard imaging methods, offers promising alternatives for the detection of bone metastasis. Various radiotracers targeting cell division activity or cancer-specific membrane proteins, as well as bone seeking agents, have been developed and tested. The use of positron-emitting isotopes such as fluorine-18 and gallium-68 for labeling allows for a reduced radiation dose and unaffected biological properties. Furthermore, the integration of artificial intelligence (AI) and radiomics techniques in medical imaging has shown significant advancements in reducing interobserver variability, improving accuracy, and saving time. This article provides an overview of the advantages and limitations of bone scan using SPECT and SPECT/CT and PET imaging methods with different radiopharmaceuticals and highlights recent developments in hybrid scanners, AI, and radiomics for the identification of prostate cancer bone metastasis using molecular imaging.
Collapse
Affiliation(s)
- Nasibeh Mohseninia
- Division of Molecular Imaging and Theranostics, Department of Nuclear Medicine, University Hospital, Paracelsus Medical University, Salzburg, Austria
| | - Nazanin Zamani-Siahkali
- Division of Molecular Imaging and Theranostics, Department of Nuclear Medicine, University Hospital, Paracelsus Medical University, Salzburg, Austria; Department of Nuclear Medicine, Research center for Nuclear Medicine and Molecular Imaging, Shariati Hospital, Tehran University of Medical Sciences, Tehran, Iran
| | - Sara Harsini
- Department of Molecular Oncology, BC Cancer Research Institute, Vancouver, BC, Canada
| | | | - Christian Pirich
- Division of Molecular Imaging and Theranostics, Department of Nuclear Medicine, University Hospital, Paracelsus Medical University, Salzburg, Austria
| | - Mohsen Beheshti
- Division of Molecular Imaging and Theranostics, Department of Nuclear Medicine, University Hospital, Paracelsus Medical University, Salzburg, Austria.
| |
Collapse
|
195
|
Jia W, Zhao G. Automated recognition of the major muscle injury in athletes on X-ray CT images1. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:107-121. [PMID: 37483059 DOI: 10.3233/xst-230135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/25/2023]
Abstract
BACKGROUND In this research, imaging techniques such as CT and X-ray are used to locate important muscles in the shoulders and legs. Athletes who participate in sports that require running, jumping, or throwing are more likely to get injuries such as sprains, strains, tendinitis, fractures, and dislocations. One proposed automated technique has the overarching goal of enhancing recognition. OBJECTIVE This study aims to determine how to recognize the major muscles in the shoulder and leg utilizing X-ray CT images as its primary diagnostic tool. METHODS Using a shape model, discovering landmarks, and generating a form model are the steps necessary to identify injuries in key shoulder and leg muscles. The method also involves identifying injuries in significant abdominal muscles. The use of adversarial deep learning, and more specifically Deep-Injury Region Identification, can improve the ability to identify damaged muscle in X-ray and CT images. RESULTS Applying the proposed diagnostic model to 150 sets of CT images, the study results show that Jaccard similarity coefficient (JSC) rate for the procedure is 0.724, the repeatability is 0.678, and the accuracy is 94.9% respectively. CONCLUSION The study results demonstrate feasibility of using adversarial deep learning and deep-injury region identification to automatically detect severe muscle injuries in the shoulder and leg, which can enhance the identification and diagnosis of injuries in athletes, especially for those who compete in sports that include running, jumping, and throwing.
Collapse
Affiliation(s)
- Wanping Jia
- Center for International Education, Philippine Christian University, Manila, Philippines
| | - Guangyong Zhao
- Department of Sports and Health, Linyi University, Shandong, Linyi, China
| |
Collapse
|
196
|
Wu D, Ni J, Fan W, Jiang Q, Wang L, Sun L, Cai Z. Opportunities and challenges of computer aided diagnosis in new millennium: A bibliometric analysis from 2000 to 2023. Medicine (Baltimore) 2023; 102:e36703. [PMID: 38134105 PMCID: PMC10735127 DOI: 10.1097/md.0000000000036703] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 11/27/2023] [Indexed: 12/24/2023] Open
Abstract
BACKGROUND After entering the new millennium, computer-aided diagnosis (CAD) is rapidly developing as an emerging technology worldwide. Expanding the spectrum of CAD-related diseases is a possible future research trend. Nevertheless, bibliometric studies in this area have not yet been reported. This study aimed to explore the hotspots and frontiers of research on CAD from 2000 to 2023, which may provide a reference for researchers in this field. METHODS In this paper, we use bibliometrics to analyze CAD-related literature in the Web of Science database between 2000 and 2023. The scientometric softwares VOSviewer and CiteSpace were used to visually analyze the countries, institutions, authors, journals, references and keywords involved in the literature. Keywords burst analysis were utilized to further explore the current state and development trends of research on CAD. RESULTS A total of 13,970 publications were included in this study, with a noticeably rising annual publication trend. China and the United States are major contributors to the publication, with the United States being the dominant position in CAD research. The American research institutions, lead by the University of Chicago, are pioneers of CAD. Acharya UR, Zheng B and Chan HP are the most prolific authors. Institute of Electrical and Electronics Engineers Transactions on Medical Imaging focuses on CAD and publishes the most articles. New computer technologies related to CAD are in the forefront of attention. Currently, CAD is used extensively in breast diseases, pulmonary diseases and brain diseases. CONCLUSION Expanding the spectrum of CAD-related diseases is a possible future research trend. How to overcome the lack of large sample datasets and establish a universally accepted standard for the evaluation of CAD system performance are urgent issues for CAD development and validation. In conclusion, this paper provides valuable information on the current state of CAD research and future developments.
Collapse
Affiliation(s)
- Di Wu
- Department of Proctology, Yongchuan Hospital of Traditional Chinese Medicine, Chongqing Medical University, Chongqing, China
- Department of Proctology, Bishan Hospital of Traditional Chinese Medicine, Chongqing, China
- Chongqing College of Traditional Chinese Medicine, Chongqing, China
| | - Jiachun Ni
- Department of Coloproctology, Yueyang Hospital of Integrated Traditional Chinese and Western Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai, China
| | - Wenbin Fan
- Department of Proctology, Bishan Hospital of Traditional Chinese Medicine, Chongqing, China
- Chongqing College of Traditional Chinese Medicine, Chongqing, China
| | - Qiong Jiang
- Chongqing College of Traditional Chinese Medicine, Chongqing, China
| | - Ling Wang
- Department of Proctology, Yongchuan Hospital of Traditional Chinese Medicine, Chongqing Medical University, Chongqing, China
| | - Li Sun
- Department of Proctology, Yongchuan Hospital of Traditional Chinese Medicine, Chongqing Medical University, Chongqing, China
| | - Zengjin Cai
- Department of Proctology, Yongchuan Hospital of Traditional Chinese Medicine, Chongqing Medical University, Chongqing, China
| |
Collapse
|
197
|
Xin Y, Zhang Q, Liu X, Li B, Mao T, Li X. Application of artificial intelligence in endoscopic gastrointestinal tumors. Front Oncol 2023; 13:1239788. [PMID: 38144533 PMCID: PMC10747923 DOI: 10.3389/fonc.2023.1239788] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 11/17/2023] [Indexed: 12/26/2023] Open
Abstract
With an increasing number of patients with gastrointestinal cancer, effective and accurate early diagnostic clinical tools are required provide better health care for patients with gastrointestinal cancer. Recent studies have shown that artificial intelligence (AI) plays an important role in the diagnosis and treatment of patients with gastrointestinal tumors, which not only improves the efficiency of early tumor screening, but also significantly improves the survival rate of patients after treatment. With the aid of efficient learning and judgment abilities of AI, endoscopists can improve the accuracy of diagnosis and treatment through endoscopy and avoid incorrect descriptions or judgments of gastrointestinal lesions. The present article provides an overview of the application status of various artificial intelligence in gastric and colorectal cancers in recent years, and the direction of future research and clinical practice is clarified from a clinical perspective to provide a comprehensive theoretical basis for AI as a promising diagnostic and therapeutic tool for gastrointestinal cancer.
Collapse
Affiliation(s)
| | | | | | | | | | - Xiaoyu Li
- Department of Gastroenterology, The Affiliated Hospital of Qingdao University, Qingdao, China
| |
Collapse
|
198
|
Kavitha P, Ayyappan G, Jayagopal P, Mathivanan SK, Mallik S, Al-Rasheed A, Alqahtani MS, Soufiene BO. Detection for melanoma skin cancer through ACCF, BPPF, and CLF techniques with machine learning approach. BMC Bioinformatics 2023; 24:458. [PMID: 38053030 DOI: 10.1186/s12859-023-05584-7] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 11/27/2023] [Indexed: 12/07/2023] Open
Abstract
Intense sun exposure is a major risk factor for the development of melanoma, an abnormal proliferation of skin cells. Yet, this more prevalent type of skin cancer can also develop in less-exposed areas, such as those that are shaded. Melanoma is the sixth most common type of skin cancer. In recent years, computer-based methods for imaging and analyzing biological systems have made considerable strides. This work investigates the use of advanced machine learning methods, specifically ensemble models with Auto Correlogram Methods, Binary Pyramid Pattern Filter, and Color Layout Filter, to enhance the detection accuracy of Melanoma skin cancer. These results suggest that the Color Layout Filter model of the Attribute Selection Classifier provides the best overall performance. Statistics for ROC, PRC, Kappa, F-Measure, and Matthews Correlation Coefficient were as follows: 90.96% accuracy, 0.91 precision, 0.91 recall, 0.95 ROC, 0.87 PRC, 0.87 Kappa, 0.91 F-Measure, and 0.82 Matthews Correlation Coefficient. In addition, its margins of error are the smallest. The research found that the Attribute Selection Classifier performed well when used in conjunction with the Color Layout Filter to improve image quality.
Collapse
Affiliation(s)
- P Kavitha
- Department of Artificial Intelligence and Data Science, Panimalar Engineering College, Chennai, India
| | - G Ayyappan
- Department of Information Technology, Prince Shri Venkateshwara Padmavathy Engineering College, Chennai, India
| | - Prabhu Jayagopal
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, 632014, India
| | - Sandeep Kumar Mathivanan
- School of Computing Science and Engineering, Galgotias University, Greater Noida, Uttar Pradesh, 203201, India
| | - Saurav Mallik
- Department of Environmental Health, Harvard T H Chan School of Public Health, Boston, MA, 02115, USA
- Department of Pharmacology and Toxicology, The University of Arizona, Tucson, AZ, 85721, USA
| | - Amal Al-Rasheed
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, 11671, Riyadh, Saudi Arabia
| | - Mohammed S Alqahtani
- Radiological Sciences Department, College of Applied Medical Sciences, King Khalid University, 61421, Abha, Saudi Arabia
- BioImaging Unit, Space Research Centre, University of Leicester, Michael Atiyah Building, Leicester, LE1 7RH, UK
| | - Ben Othman Soufiene
- PRINCE Laboratory Research, ISITcom, Hammam Sousse, University of Sousse, Sousse, Tunisia.
| |
Collapse
|
199
|
Than J, Sim PY, Muttuvelu D, Ferraz D, Koh V, Kang S, Huemer J. Teleophthalmology and retina: a review of current tools, pathways and services. Int J Retina Vitreous 2023; 9:76. [PMID: 38053188 PMCID: PMC10699065 DOI: 10.1186/s40942-023-00502-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2023] [Accepted: 10/02/2023] [Indexed: 12/07/2023] Open
Abstract
Telemedicine, the use of telecommunication and information technology to deliver healthcare remotely, has evolved beyond recognition since its inception in the 1970s. Advances in telecommunication infrastructure, the advent of the Internet, exponential growth in computing power and associated computer-aided diagnosis, and medical imaging developments have created an environment where telemedicine is more accessible and capable than ever before, particularly in the field of ophthalmology. Ever-increasing global demand for ophthalmic services due to population growth and ageing together with insufficient supply of ophthalmologists requires new models of healthcare provision integrating telemedicine to meet present day challenges, with the recent COVID-19 pandemic providing the catalyst for the widespread adoption and acceptance of teleophthalmology. In this review we discuss the history, present and future application of telemedicine within the field of ophthalmology, and specifically retinal disease. We consider the strengths and limitations of teleophthalmology, its role in screening, community and hospital management of retinal disease, patient and clinician attitudes, and barriers to its adoption.
Collapse
Affiliation(s)
- Jonathan Than
- Moorfields Eye Hospital NHS Foundation Trust, 162 City Road, London, UK
| | - Peng Y Sim
- Moorfields Eye Hospital NHS Foundation Trust, 162 City Road, London, UK
| | - Danson Muttuvelu
- Department of Ophthalmology, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- MitØje ApS/Danske Speciallaeger Aps, Aarhus, Denmark
| | - Daniel Ferraz
- D'Or Institute for Research and Education (IDOR), São Paulo, Brazil
- Institute of Ophthalmology, University College London, London, UK
| | - Victor Koh
- Department of Ophthalmology, National University Hospital, Singapore, Singapore
| | - Swan Kang
- Moorfields Eye Hospital NHS Foundation Trust, 162 City Road, London, UK
| | - Josef Huemer
- Moorfields Eye Hospital NHS Foundation Trust, 162 City Road, London, UK.
- Department of Ophthalmology and Optometry, Kepler University Hospital, Johannes Kepler University, Linz, Austria.
| |
Collapse
|
200
|
Asif S, Zhao M, Chen X, Zhu Y. StoneNet: An Efficient Lightweight Model Based on Depthwise Separable Convolutions for Kidney Stone Detection from CT Images. Interdiscip Sci 2023; 15:633-652. [PMID: 37452930 DOI: 10.1007/s12539-023-00578-8] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Revised: 06/27/2023] [Accepted: 06/30/2023] [Indexed: 07/18/2023]
Abstract
Kidney stone disease is one of the most common and serious health problems in much of the world, leading to many hospitalizations with severe pain. Detecting small stones is difficult and time-consuming, so an early diagnosis of kidney disease is needed to prevent the loss of kidney failure. Recent advances in artificial intelligence (AI) found to be very successful in the diagnosis of various diseases in the biomedical field. However, existing models using deep networks have several problems, such as high computational cost, long training time, and huge parameters. Providing a low-cost solution for diagnosing kidney stones in a medical decision support system is of paramount importance. Therefore, in this study, we propose "StoneNet", a lightweight and high-performance model for the detection of kidney stones based on MobileNet using depthwise separable convolution. The proposed model includes a combination of global average pooling (GAP), batch normalization, dropout layer, and dense layers. Our study shows that using GAP instead of flattening layers greatly improves the robustness of the model by significantly reducing the parameters. The developed model is benchmarked against four pre-trained models as well as the state-of-the-art heavy model. The results show that the proposed model can achieve the highest accuracy of 97.98%, and only requires training and testing time of 996.88 s and 14.62 s. Several parameters, such as different batch sizes and optimizers, were considered to validate the proposed model. The proposed model is computationally faster and provides optimal performance than other considered models. Experiments on a large kidney dataset of 1799 CT images show that StoneNet has superior performance in terms of higher accuracy and lower complexity. The proposed model can assist the radiologist in faster diagnosis of kidney stones and has great potential for deployment in real-time applications.
Collapse
Affiliation(s)
- Sohaib Asif
- School of Computer Science and Engineering, Central South University, Changsha, China
| | - Ming Zhao
- School of Computer Science and Engineering, Central South University, Changsha, China.
| | - Xuehan Chen
- School of Computer Science and Engineering, Central South University, Changsha, China.
| | - Yusen Zhu
- School of Mathematics, Hunan University, Changsha, China
| |
Collapse
|