1
|
Liu G, Huang W, Li Y, Zhang Q, Fu J, Tang H, Huang J, Zhang Z, Zhang L, Wang Y, Hu J. A weakly-supervised follicle segmentation method in ultrasound images. Sci Rep 2025; 15:13771. [PMID: 40258856 PMCID: PMC12012036 DOI: 10.1038/s41598-025-95957-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 03/25/2025] [Indexed: 04/23/2025] Open
Abstract
Accurate follicle segmentation in ultrasound images is crucial for monitoring follicle development, a key factor in fertility treatments. However, obtaining pixel-level annotations for fully supervised instance segmentation is often impractical due to time and workload constraints. This paper presents a weakly supervised instance segmentation method that leverages bounding boxes as approximate annotations, aiming to assist clinicians with automated tools for follicle development monitoring. We propose the Weakly Supervised Follicle Segmentation (WSFS) method, a novel one-stage weakly supervised segmentation technique model designed to enhance the ultrasound images of follicles, which incorporates a Convolutional Neural Network (CNN) backbone augmented with a Feature Pyramid Network (FPN) module for multi-scale feature representation, critical for capturing the diverse sizes and shapes of follicles. By leveraging Multiple Instance Learning (MIL), we formulated the learning process in a weakly supervised manner and developed an end-to-end trainable model that efficiently addresses the issue of annotation scarcity. Furthermore, the WSFS can be used as a prompt proposal to enhance the performance of the Segmentation Anything Model (SAM), a well-known pre-trained segmentation model utilizing few-shot learning strategies. In addition, this study introduces the Follicle Ultrasound Image Dataset (FUID), addressing the scarcity in reproductive health data and aiding future research in computer-aided diagnosis. The experimental results on both the public dataset USOVA3D and private dataset FUID showed that our method performs competitively with fully supervised methods. Our approach achieves performance with mAP of 0.957, IOU of 0.714 and Dice Score of 0.83, competitive to fully supervised methods that rely on pixel-level labeled masks, despite operating with less detailed annotations.
Collapse
Affiliation(s)
- Guanyu Liu
- Big Data Institute, Central South University, Changsha, 410083, China
| | - Weihong Huang
- Big Data Institute, Central South University, Changsha, 410083, China
- Mobile Health Ministry of Education - China Mobile Joint Laboratory, Xiangya Hospital, Central South University, Changsha, 410000, China
- Xiangjiang Laboratory, Changsha, 410205, China
| | - Yanping Li
- Department of Reproductive Medicine, Xiangya Hospital, Central South University, Changsha, Hunan, 410000, China
- Clinical Research Center for Women's Reproductive Health in Hunan Province, Changsha, Hunan, 410000, China
| | - Qiong Zhang
- Department of Reproductive Medicine, Xiangya Hospital, Central South University, Changsha, Hunan, 410000, China
- Clinical Research Center for Women's Reproductive Health in Hunan Province, Changsha, Hunan, 410000, China
| | - Jing Fu
- Department of Reproductive Medicine, Xiangya Hospital, Central South University, Changsha, Hunan, 410000, China
- Clinical Research Center for Women's Reproductive Health in Hunan Province, Changsha, Hunan, 410000, China
| | - Hongying Tang
- Department of Reproductive Medicine, Xiangya Hospital, Central South University, Changsha, Hunan, 410000, China
- Clinical Research Center for Women's Reproductive Health in Hunan Province, Changsha, Hunan, 410000, China
| | - Jia Huang
- School of Life Science, Central South University, Changsha, 410083, China
| | - Zhongteng Zhang
- School of Computer Sciences and Engineering, Central South University, Changsha, 410083, China
| | - Lei Zhang
- Laboratory of Vision Engineering (LoVE), School of computer science, University of Lincoln, Lincoln, LN6 7TS, UK
| | - Yu Wang
- Department of Reproductive Medicine, Xiangya Hospital, Central South University, Changsha, Hunan, 410000, China.
- Clinical Research Center for Women's Reproductive Health in Hunan Province, Changsha, Hunan, 410000, China.
| | - Jianzhong Hu
- Big Data Institute, Central South University, Changsha, 410083, China.
- Mobile Health Ministry of Education - China Mobile Joint Laboratory, Xiangya Hospital, Central South University, Changsha, 410000, China.
| |
Collapse
|
2
|
Agyekum EA, Wang YG, Issaka E, Ren YZ, Tan G, Shen X, Qian XQ. Predicting the efficacy of microwave ablation of benign thyroid nodules from ultrasound images using deep convolutional neural networks. BMC Med Inform Decis Mak 2025; 25:161. [PMID: 40217199 PMCID: PMC11987319 DOI: 10.1186/s12911-025-02989-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2024] [Accepted: 03/26/2025] [Indexed: 04/15/2025] Open
Abstract
BACKGROUND Thyroid nodules are frequent in clinical settings, and their diagnosis in adults is growing, with some persons experiencing symptoms. Ultrasound-guided thermal ablation can shrink nodules and alleviate discomfort. Because the degree and rate of lesion absorption vary greatly between individuals, there is no reliable model for predicting the therapeutic efficacy of thermal ablation. METHODS Five convolutional neural network models including VGG19, Resnet 50, EfficientNetB1, EfficientNetB0, and InceptionV3, pre-trained with ImageNet, were compared for predicting the efficacy of ultrasound-guided microwave ablation (MWA) for benign thyroid nodules using ultrasound data. The patients were randomly assigned to one of two data sets: training (70%) or validation (30%). Accuracy, sensitivity, specificity, positive predictive value, negative predictive value, and area under the curve (AUC) were all used to assess predictive performance. RESULTS In the validation set, fine-tuned EfficientNetB1 performed best, with an AUC of 0.85 and an ACC of 0.79. CONCLUSIONS The study found that our deep learning model accurately predicts nodules with VRR < 50% after a single MWA session. Indeed, when thermal therapies compete with surgery, anticipating which nodules will be poor responders provides useful information that may assist physicians and patients determine whether thermal ablation or surgery is the preferable option. This was a preliminary study of deep learning, with a gap in actual clinical applications. As a result, more in-depth study should be undertaken to develop deep-learning models that can better help clinics. Prospective studies are expected to generate high-quality evidence and improve clinical performance in subsequent research.
Collapse
Affiliation(s)
- Enock Adjei Agyekum
- Department of Ultrasound, Affiliated People's Hospital of Jiangsu University, Zhenjiang, 212002, China
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, Jiangsu Province, China
| | - Yu-Guo Wang
- Department of Ultrasound, Jiangsu Hospital of Integrated Traditional Chinese and Western Medicine, Nanjing, China
| | - Eliasu Issaka
- College of Engineering, Birmingham City University, Birmingham, B4 7XG, UK
| | - Yong-Zhen Ren
- Department of Ultrasound, Affiliated People's Hospital of Jiangsu University, Zhenjiang, 212002, China
| | - Gongxun Tan
- Department of Ultrasound, Affiliated People's Hospital of Jiangsu University, Zhenjiang, 212002, China
| | - Xiangjun Shen
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, Jiangsu Province, China.
| | - Xiao-Qin Qian
- Northern Jiangsu People's Hospital Affiliated to Yangzhou University, Yangzhou, China.
- Northern Jiangsu People's Hospital, Yangzhou, Jiangsu Province, China.
- The Yangzhou Clinical Medical College of Xuzhou Medical University, Yangzhou, Jiangsu Province, China.
| |
Collapse
|
3
|
Wu M, Yan C, Sen G. Computer-aided diagnosis of hepatic cystic echinococcosis based on deep transfer learning features from ultrasound images. Sci Rep 2025; 15:607. [PMID: 39753933 PMCID: PMC11698856 DOI: 10.1038/s41598-024-85004-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2024] [Accepted: 12/30/2024] [Indexed: 01/06/2025] Open
Abstract
Hepatic cystic echinococcosis (HCE), a life-threatening liver disease, has 5 subtypes, i.e., single-cystic, polycystic, internal capsule collapse, solid mass, and calcified subtypes. And each subtype has different treatment methods. An accurate diagnosis is the prerequisite for effective HCE treatment. However, clinicians with less diagnostic experience often make misdiagnoses of HCE and confuse its 5 subtypes in clinical practice. Computer-aided diagnosis (CAD) techniques can help clinicians to improve their diagnostic performance. This paper aims to propose an efficient CAD system that automatically differentiates 5 subtypes of HCE from the ultrasound images. The proposed CAD system adopts the concept of deep transfer learning and uses a pre-trained convolutional neural network (CNN) named VGG19 to extract deep CNN features from the ultrasound images. The proven classifier models, k - nearest neighbor (KNN) and support vecter machine (SVM) models, are integrated to classify the extracted deep CNN features. 3 distinct experiments with the same deep CNN features but different classifier models (softmax, KNN, SVM) are performed. The experiments followed 10 runs of the five-fold cross-validation process on a total of 1820 ultrasound images and the results were compared using Wilcoxon signed-rank test. The overall classification accuracy from low to high was 90.46 ± 1.59% for KNN classifier, 90.92 ± 2.49% for transfer learned VGG19, and 92.01 ± 1.48% for SVM, indicating SVM classifiers with deep CNN features achieved the best performance (P < 0.05). Other performance measures used in the study are specificity, sensitivity, precision, F1-score, and area under the curve (AUC). In addition, the paper addresses a practical aspect by evaluating the system with smaller training data to demonstrate the capability of the proposed classification system. The observations of the study imply that transfer learning is a useful technique when the availability of medical images is limited. The proposed classification system by using deep CNN features and SVM classifier is potentially helpful for clinicians to improve their HCE diagnostic performance in clinical practice.
Collapse
Affiliation(s)
- Miao Wu
- College of Medical Engineering and Technology, Xinjiang Medical University, Urumqi, 830017, China.
| | - Chuanbo Yan
- College of Medical Engineering and Technology, Xinjiang Medical University, Urumqi, 830017, China
| | - Gan Sen
- College of Medical Engineering and Technology, Xinjiang Medical University, Urumqi, 830017, China
| |
Collapse
|
4
|
Yang X, Geng H, Wang X, Li L, An X, Cong Z. Identification of lesion location and discrimination between benign and malignant findings in thyroid ultrasound imaging. Sci Rep 2024; 14:32118. [PMID: 39738724 PMCID: PMC11685495 DOI: 10.1038/s41598-024-83888-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2024] [Accepted: 12/18/2024] [Indexed: 01/02/2025] Open
Abstract
Thyroid nodules are a common thyroid disorder, and ultrasound imaging, as the primary diagnostic tool, is susceptible to variations based on the physician's experience, leading to misdiagnosis. This paper constructs an end-to-end thyroid nodule detection framework based on YOLOv8, enabling automatic detection and classification of nodules by extracting grayscale and elastic features from ultrasound images. First, an attention-weighted DCN is introduced to enhance superficial feature extraction and capture local information. Next, the CPCA mechanism is employed to reduce the interference of redundant information. Finally, a feature fusion network based on an aggregation-distribution mechanism is utilized to improve the learning capability of fine-grained features, enhancing the performance of early nodule detection. Experimental results demonstrate that our method is accurate and effective for thyroid nodule detection, achieving diagnostic rates of 89.3% for benign and 90.4% for malignant nodules based on tests conducted on 611 clinical ultrasound images, with a mean Average Precision at IoU = 0.5 (mAP@50) of 95.5%, representing a 6.6% improvement over baseline models.
Collapse
Affiliation(s)
- Xu Yang
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130022, China
| | - Hongliang Geng
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, 130022, China
| | - Xue Wang
- Department of Electrodiagnosis, The Affiliated Hospital to Changchun University of Traditional Chinese Medicine, Changchun, 130021, China
| | - Lingxiao Li
- Human Resources Department, The Third Affiliated Hospital of C.C.U.C.M, Changchun, 130117, China
| | - Xiaofeng An
- Education Quality Monitoring Center, Jilin Engineering Normal University, Changchun, 130052, China.
| | - Zhibin Cong
- Department of Electrodiagnosis, The Affiliated Hospital to Changchun University of Traditional Chinese Medicine, Changchun, 130021, China.
| |
Collapse
|
5
|
Leoncini A, Curti M, Ruinelli L, Gamarra E, Trimboli P. Performance of ACR-TIRADS in assessing thyroid nodules does not vary according to patient age. Hormones (Athens) 2024; 23:667-674. [PMID: 39028415 PMCID: PMC11519249 DOI: 10.1007/s42000-024-00585-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/19/2024] [Accepted: 07/09/2024] [Indexed: 07/20/2024]
Abstract
AIMS A few studies have evaluated the performance of the American College of Radiology Thyroid Imaging Reporting And Data System (ACR-TIRADS) in pediatric and elderly patients and found differences between the latter two age groups and middle adulthood. Thus, the present study was undertaken to explore the possible variation of ACR-TIRADS performance across different ages of patients. METHODS A retrospective population undergoing thyroidectomy was selected to use histology as the reference standard. Ultrasound images were reviewed, and alignment of ACR-TIRADS with the corresponding histological diagnosis was made afterwards. Results of the age groups were compared. The ACR-TIRADS diagnostic performance was calculated considering the assessment of nodules across risk categories (i.e., from TR1 to TR5), rate of unnecessary FNAC (UN-FNAC), and rate of necessary but non-performed FNAC (NNP-FNAC). RESULTS Overall, 114 patients with a total of 220 nodules (46 carcinomas) were included. The rate of UN-FNAC was 66.3%, being 93.1% in TR3, 82.1% in TR4, and 31.4% in TR5. There were 15 NNP-FNACs. No significant difference was observed between age groups in terms of sample size, nodule, cancer, and FNAC. The nodule assessment according to ACR-TIRADS categories did not vary across ages. Sensitivity and specificity recorded in three age tertiles were not significantly different. CONCLUSIONS The present study shows that the performance of ACR-TIRADS is not significantly influenced by patient age.
Collapse
Affiliation(s)
- Andrea Leoncini
- Servizio Di Radiologia E Radiologia Interventistica, Istituto Di Imaging Della Svizzera Italiana (IIMSI), Ente Ospedaliero Cantonale (EOC), 6900, Lugano, Switzerland
| | - Marco Curti
- Servizio Di Radiologia E Radiologia Interventistica, Istituto Di Imaging Della Svizzera Italiana (IIMSI), Ente Ospedaliero Cantonale (EOC), 6900, Lugano, Switzerland
| | - Lorenzo Ruinelli
- Servizio Di Endocrinologia E Diabetologia, Ospedale Regionale Di Lugano, Ente Ospedaliero Cantonale (EOC), 6900, Lugano, Switzerland
- Team Data Science & Research, Ente Ospedaliero Cantonale, Area ICT, 6500, Bellinzona, Switzerland
- Clinical Trial Unit, Ente Ospedaliero Cantonale (EOC), Bellinzona, Switzerland
| | - Elena Gamarra
- Servizio Di Endocrinologia E Diabetologia, Ospedale Regionale Di Lugano, Ente Ospedaliero Cantonale (EOC), 6900, Lugano, Switzerland
| | - Pierpaolo Trimboli
- Servizio Di Endocrinologia E Diabetologia, Ospedale Regionale Di Lugano, Ente Ospedaliero Cantonale (EOC), 6900, Lugano, Switzerland.
- Facoltà Di Scienze Biomediche, Università Della Svizzera Italiana (USI), 6900, Lugano, Switzerland.
| |
Collapse
|
6
|
Saproo D, Mahajan AN, Narwal S. Deep learning based binary classification of diabetic retinopathy images using transfer learning approach. J Diabetes Metab Disord 2024; 23:2289-2314. [PMID: 39610484 PMCID: PMC11599653 DOI: 10.1007/s40200-024-01497-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/17/2024] [Accepted: 08/17/2024] [Indexed: 11/30/2024]
Abstract
Objective Diabetic retinopathy (DR) is a common problem of diabetes, and it is the cause of blindness worldwide. Detection of diabetic radiology disease in the early detection stage is crucial for preventing vision loss. In this work, a deep learning-based binary classification of DR images has been proposed to classify DR images into healthy and unhealthy. Transfer learning-based 20 pre-trained networks have been fine-tuned using a robust dataset of diabetic radiology images. The combined dataset has been collected from three robust databases of diabetic patients annotated by experienced ophthalmologists indicating healthy or non-healthy diabetic retina images. Method This work has improved robust models by pre-processing the DR images by applying a denoising algorithm, normalization, and data augmentation. In this work, three rubout datasets of diabetic retinopathy images have been selected, named DRD- EyePACS, IDRiD, and APTOS-2019, for the extensive experiments, and a combined diabetic retinopathy image dataset has been generated for the exhaustive experiments. The datasets have been divided into training, testing, and validation sets, and the models use classification accuracy, sensitivity, specificity, precision, F1-score, and ROC-AUC to assess the model's efficiency for evaluating network performance. The present work has selected 20 different pre-trained networks based on three categories: Series, DAG, and lightweight. Results This study uses pre-processed data augmentation and normalization of data to solve overfitting problems. From the exhaustive experiments, the three best pre-trained have been selected based on the best classification accuracy from each category. It is concluded that the trained model ResNet101 based on the DAG category effectively identifies diabetic retinopathy disease accurately from radiological images from all cases. It is noted that 97.33% accuracy has been achieved using ResNet101 in the category of DAG network. Conclusion Based on the experiment results, the proposed model ResNet101 helps healthcare professionals detect retina diseases early and provides practical solutions to diabetes patients. It also gives patients and experts a second opinion for early detection of diabetic retinopathy.
Collapse
Affiliation(s)
- Dimple Saproo
- Maharaja Agrasen University Baddi, Baddi, Himachal Pradesh 173205 India
| | - Aparna N. Mahajan
- Maharaja Agrasen Institute of Technology (MAIT), Maharaja Agrasen University Baddi, Baddi, Himachal Pradesh 173205 India
| | - Seema Narwal
- Dronacharya College of Engineering, Gurugram, Haryana 122001 India
| |
Collapse
|
7
|
Moral P, Mustafi D, Mustafi A, Sahana SK. CystNet: An AI driven model for PCOS detection using multilevel thresholding of ultrasound images. Sci Rep 2024; 14:25012. [PMID: 39443622 PMCID: PMC11499604 DOI: 10.1038/s41598-024-75964-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2024] [Accepted: 10/09/2024] [Indexed: 10/25/2024] Open
Abstract
Polycystic Ovary Syndrome (PCOS) is a widespread endocrinological dysfunction impacting women of reproductive age, categorized by excess androgens and a variety of associated syndromes, consisting of acne, alopecia, and hirsutism. It involves the presence of multiple immature follicles in the ovaries, which can disrupt normal ovulation and lead to hormonal imbalances and associated health complications. Routine diagnostic methods rely on manual interpretation of ultrasound (US) images and clinical assessments, which are time-consuming and prone to errors. Therefore, implementing an automated system is essential for streamlining the diagnostic process and enhancing accuracy. By automatically analyzing follicle characteristics and other relevant features, this research aims to facilitate timely intervention and reduce the burden on healthcare professionals. The present study proposes an advanced automated system for detecting and classifying PCOS from ultrasound images. Leveraging Artificial Intelligence (AI) based techniques, the system examines affected and unaffected cases to enhance diagnostic accuracy. The pre-processing of input images incorporates techniques such as image resizing, normalization, augmentation, Watershed technique, multilevel thresholding, etc. approaches for precise image segmentation. Feature extraction is facilitated by the proposed CystNet technique, followed by PCOS classification utilizing both fully connected layers with 5-fold cross-validation and traditional machine learning classifiers. The performance of the model is rigorously evaluated using a comprehensive range of metrics, incorporating AUC score, accuracy, specificity, precision, F1-score, recall, and loss, along with a detailed confusion matrix analysis. The model demonstrated a commendable accuracy of [Formula: see text] when utilizing a fully connected classification layer, as determined by a thorough 5-fold cross-validation process. Additionally, it has achieved an accuracy of [Formula: see text] when employing an ensemble ML classifier. This proposed approach could be suggested for predicting PCOS or similar diseases using datasets that exhibit multimodal characteristics.
Collapse
Affiliation(s)
- Poonam Moral
- Department of Computer Science and Engineering, Birla Institute of Technology, Mesra, Ranchi, 835215, India.
| | - Debjani Mustafi
- Department of Computer Science and Engineering, Birla Institute of Technology, Mesra, Ranchi, 835215, India
| | - Abhijit Mustafi
- Department of Computer Science and Engineering, Birla Institute of Technology, Mesra, Ranchi, 835215, India
| | - Sudip Kumar Sahana
- Department of Computer Science and Engineering, Birla Institute of Technology, Mesra, Ranchi, 835215, India
| |
Collapse
|
8
|
Nastase INA, Moldovanu S, Biswas KC, Moraru L. Role of inter- and extra-lesion tissue, transfer learning, and fine-tuning in the robust classification of breast lesions. Sci Rep 2024; 14:22754. [PMID: 39354128 PMCID: PMC11448494 DOI: 10.1038/s41598-024-74316-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Accepted: 09/25/2024] [Indexed: 10/03/2024] Open
Abstract
Accurate and unbiased classification of breast lesions is pivotal for early diagnosis and treatment, and a deep learning approach can effectively represent and utilize the digital content of images for more precise medical image analysis. Breast ultrasound imaging is useful for detecting and distinguishing benign masses from malignant masses. Based on the different ways in which benign and malignant tumors affect neighboring tissues, i.e., the pattern of growth and border irregularities, the penetration degree of the adjacent tissue, and tissue-level changes, we investigated the relationship between breast cancer imaging features and the roles of inter- and extra-lesional tissues and their impact on refining the performance of deep learning classification. The novelty of the proposed approach lies in considering the features extracted from the tissue inside the tumor (by performing an erosion operation) and from the lesion and surrounding tissue (by performing a dilation operation) for classification. This study uses these new features and three pre-trained deep neuronal networks to address the challenge of breast lesion classification in ultrasound images. To improve the classification accuracy and interpretability of the model, the proposed model leverages transfer learning to accelerate the training process. Three modern pre-trained CNN architectures (MobileNetV2, VGG16, and EfficientNetB7) are used for transfer learning and fine-tuning for optimization. There are concerns related to the neuronal networks producing erroneous outputs in the presence of noisy images, variations in input data, or adversarial attacks; thus, the proposed system uses the BUS-BRA database (two classes/benign and malignant) for training and testing and the unseen BUSI database (two classes/benign and malignant) for testing. Extensive experiments have recorded accuracy and AUC as performance parameters. The results indicate that the proposed system outperforms the existing breast cancer detection algorithms reported in the literature. AUC values of 1.00 are calculated for VGG16 and EfficientNet-B7 in the dilation cases. The proposed approach will facilitate this challenging and time-consuming classification task.
Collapse
Affiliation(s)
- Iulia-Nela Anghelache Nastase
- The Modeling & Simulation Laboratory, Dunarea de Jos University of Galati, 47 Domneasca Street, Galati, 800008, Romania
- Emil Racovita Theoretical Highschool, 12-14, Regiment 11 Siret Street, Galati, 800332, Romania
| | - Simona Moldovanu
- The Modeling & Simulation Laboratory, Dunarea de Jos University of Galati, 47 Domneasca Street, Galati, 800008, Romania.
- Department of Computer Science and Information Technology, Faculty of Automation, Computers, Electrical Engineering and Electronics, Dunarea de Jos University of Galati, 47 Domneasca Street, Galati, 800008, Romania.
| | - Keka C Biswas
- Department of Biological Sciences, University of Alabama at Huntsville, Huntsville, AL, 35899, USA
| | - Luminita Moraru
- The Modeling & Simulation Laboratory, Dunarea de Jos University of Galati, 47 Domneasca Street, Galati, 800008, Romania.
- Department of Chemistry, Physics & Environment, Faculty of Sciences and Environment, Dunarea de Jos University of Galati, 47 Domneasca Street, Galati, 800008, Romania.
- Department of Physics, School of Science and Technology, Sefako Makgatho Health Sciences University, Medunsa-0204, Pretoria, South Africa.
| |
Collapse
|
9
|
Latha M, Kumar PS, Chandrika RR, Mahesh TR, Kumar VV, Guluwadi S. Revolutionizing breast ultrasound diagnostics with EfficientNet-B7 and Explainable AI. BMC Med Imaging 2024; 24:230. [PMID: 39223507 PMCID: PMC11367906 DOI: 10.1186/s12880-024-01404-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2024] [Accepted: 08/19/2024] [Indexed: 09/04/2024] Open
Abstract
Breast cancer is a leading cause of mortality among women globally, necessitating precise classification of breast ultrasound images for early diagnosis and treatment. Traditional methods using CNN architectures such as VGG, ResNet, and DenseNet, though somewhat effective, often struggle with class imbalances and subtle texture variations, leading to reduced accuracy for minority classes such as malignant tumors. To address these issues, we propose a methodology that leverages EfficientNet-B7, a scalable CNN architecture, combined with advanced data augmentation techniques to enhance minority class representation and improve model robustness. Our approach involves fine-tuning EfficientNet-B7 on the BUSI dataset, implementing RandomHorizontalFlip, RandomRotation, and ColorJitter to balance the dataset and improve model robustness. The training process includes early stopping to prevent overfitting and optimize performance metrics. Additionally, we integrate Explainable AI (XAI) techniques, such as Grad-CAM, to enhance the interpretability and transparency of the model's predictions, providing visual and quantitative insights into the features and regions of ultrasound images influencing classification outcomes. Our model achieves a classification accuracy of 99.14%, significantly outperforming existing CNN-based approaches in breast ultrasound image classification. The incorporation of XAI techniques enhances our understanding of the model's decision-making process, thereby increasing its reliability and facilitating clinical adoption. This comprehensive framework offers a robust and interpretable tool for the early detection and diagnosis of breast cancer, advancing the capabilities of automated diagnostic systems and supporting clinical decision-making processes.
Collapse
Affiliation(s)
- M Latha
- Department of Information Technology, SRM Institute of Science and Technology, Ramapuram, Chennai, India
| | - P Santhosh Kumar
- Department of Information Technology, SRM Institute of Science and Technology, Ramapuram, Chennai, India
| | - R Roopa Chandrika
- Department of Computer Science and Engineering, Faculty of Engineering, Karpagam Academy of Higher Education(Deemed to be University, Coimbatore, India
| | - T R Mahesh
- Department of Computer Science & Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-be University), Bengaluru, 562112, India
| | - V Vinoth Kumar
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology University, Vellore, 632014, India
| | - Suresh Guluwadi
- Adama Science and Technology University, Adama, 302120, Ethiopia.
| |
Collapse
|
10
|
Zheng Y, Zhang Y, Lu K, Wang J, Li L, Xu D, Liu J, Lou J. Diagnostic value of an interpretable machine learning model based on clinical ultrasound features for follicular thyroid carcinoma. Quant Imaging Med Surg 2024; 14:6311-6324. [PMID: 39281129 PMCID: PMC11400673 DOI: 10.21037/qims-24-601] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Accepted: 07/11/2024] [Indexed: 09/18/2024]
Abstract
Background Follicular thyroid carcinoma (FTC) and follicular thyroid adenoma (FTA) present diagnostic challenges due to overlapping clinical and ultrasound features. Improving the diagnosis of FTC can enhance patient prognosis and effectiveness in clinical management. This study seeks to develop a predictive model for FTC based on ultrasound features using machine learning (ML) algorithms and assess its diagnostic effectiveness. Methods Patients diagnosed with FTA or FTC based on surgical pathology between January 2009 and February 2023 at Zhejiang Provincial Cancer Hospital and Zhejiang Provincial People's Hospital were retrospectively included. A total of 562 patients from Zhejiang Provincial Cancer Hospital comprised the training set, and 218 patients from Zhejiang Provincial People's Hospital constituted the validation set. Subsequently, clinical parameters and ultrasound characteristics of the patients were collected. The diagnostic parameters were analyzed using the least absolute shrinkage and selection operator and multivariate logistic regression screening methods. Next, a comparative analysis was performed using seven ML models. The area under the receiver operating characteristic (ROC) curve (AUC), accuracy, sensitivity, specificity, positive predicted value (PPV), negative predicted value (NPV), precision, recall, and comprehensive evaluation index (F-score) were calculated to compare the diagnostic efficacy among the seven models and determine the optimal model. Further, the optimal model was validated, and the SHapley Additive ExPlanations (SHAP) approach was applied to explain the significance of the model variables. Finally, an individualized risk assessment was conducted. Results Age, echogenicity, thyroglobulin antibody (TGAb), echotexture, composition, triiodothyronine (T3), thyroglobulin (TG), margin, thyroid-stimulating hormone (TSH), calcification, and halo thickness >2 mm were influential factors for diagnosing FTC. The XGBoost model was identified as the optimal model after a comprehensive evaluation. The AUC of this model in the validation set was 0.969 [95% confidence interval (CI), 0.946-0.992], while its precision sensitivity, specificity, and accuracy were 0.791, 0.930, 0.913 and 0.917, respectively. Conclusions XGBoost model based on ultrasound features was constructed and interpreted using the SHAP method, providing evidence for the diagnosis of FTC and guidance for the personalized treatment of patients.
Collapse
Affiliation(s)
- Yuxin Zheng
- Second Clinical College, Zhejiang University of Traditional Chinese Medicine, Hangzhou, China
- Department of Diagnostic Ultrasound Imaging & Interventional Therapy, Zhejiang Cancer Hospital, Hangzhou, China
- Key Laboratory of Head & Neck Cancer Translational Research of Zhejiang Province, Hangzhou, China
| | - Yajiao Zhang
- Second Clinical College, Zhejiang University of Traditional Chinese Medicine, Hangzhou, China
- Department of Diagnostic Ultrasound Imaging & Interventional Therapy, Zhejiang Cancer Hospital, Hangzhou, China
- Key Laboratory of Head & Neck Cancer Translational Research of Zhejiang Province, Hangzhou, China
| | - Kefeng Lu
- Department of Ultrasound, Zhejiang Provincial People's Hospital, Affiliated People's Hospital, Hangzhou Medical College, Hangzhou, China
| | - Jiafeng Wang
- Otolaryngology & Head and Neck Center, Cancer Center, Department of Head and Neck Surgery, Zhejiang Provincial People's Hospital (Affiliated People's Hospital), Hangzhou Medical College, Hangzhou, China
- Department of Thyroid and Breast Surgery, Zhejiang Provincial People's Hospital Bijie Hospital, Bijie, China
| | - Linlin Li
- Otolaryngology & Head and Neck Center, Cancer Center, Department of Head and Neck Surgery, Zhejiang Provincial People's Hospital (Affiliated People's Hospital), Hangzhou Medical College, Hangzhou, China
| | - Dong Xu
- Department of Diagnostic Ultrasound Imaging & Interventional Therapy, Zhejiang Cancer Hospital, Hangzhou, China
- Hangzhou Institute of Medicine (HIM), Chinese Academy of Sciences, Hangzhou, China
| | - Junping Liu
- Department of Diagnostic Ultrasound Imaging & Interventional Therapy, Zhejiang Cancer Hospital, Hangzhou, China
- Hangzhou Institute of Medicine (HIM), Chinese Academy of Sciences, Hangzhou, China
| | - Jiangyan Lou
- Department of Pediatrics, Zhejiang Provincial People's Hospital, Affiliated People's Hospital, Hangzhou Medical College, Hangzhou, China
| |
Collapse
|
11
|
Ru J, Zhu Z, Shi J. Spatial and geometric learning for classification of breast tumors from multi-center ultrasound images: a hybrid learning approach. BMC Med Imaging 2024; 24:133. [PMID: 38840240 PMCID: PMC11155188 DOI: 10.1186/s12880-024-01307-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2024] [Accepted: 05/27/2024] [Indexed: 06/07/2024] Open
Abstract
BACKGROUND Breast cancer is the most common cancer among women, and ultrasound is a usual tool for early screening. Nowadays, deep learning technique is applied as an auxiliary tool to provide the predictive results for doctors to decide whether to make further examinations or treatments. This study aimed to develop a hybrid learning approach for breast ultrasound classification by extracting more potential features from local and multi-center ultrasound data. METHODS We proposed a hybrid learning approach to classify the breast tumors into benign and malignant. Three multi-center datasets (BUSI, BUS, OASBUD) were used to pretrain a model by federated learning, then every dataset was fine-tuned at local. The proposed model consisted of a convolutional neural network (CNN) and a graph neural network (GNN), aiming to extract features from images at a spatial level and from graphs at a geometric level. The input images are small-sized and free from pixel-level labels, and the input graphs are generated automatically in an unsupervised manner, which saves the costs of labor and memory space. RESULTS The classification AUCROC of our proposed method is 0.911, 0.871 and 0.767 for BUSI, BUS and OASBUD. The balanced accuracy is 87.6%, 85.2% and 61.4% respectively. The results show that our method outperforms conventional methods. CONCLUSIONS Our hybrid approach can learn the inter-feature among multi-center data and the intra-feature of local data. It shows potential in aiding doctors for breast tumor classification in ultrasound at an early stage.
Collapse
Affiliation(s)
- Jintao Ru
- Department of Medical Engineering, Shaoxing Hospital of Traditional Chinese Medicine, Shaoxing, Zhejiang, People's Republic of China.
| | - Zili Zhu
- Department of Radiology, The First Affiliated Hospital of Ningbo University, Ningbo, Zhejiang, People's Republic of China
| | - Jialin Shi
- Rehabilitation Medicine Institute, Zhejiang Rehabilitation Medical Center, Hangzhou, Zhejiang, People's Republic of China
| |
Collapse
|
12
|
Liang B, Peng F, Luo D, Zeng Q, Wen H, Zheng B, Zou Z, An L, Wen H, Wen X, Liao Y, Yuan Y, Li S. Automatic segmentation of 15 critical anatomical labels and measurements of cardiac axis and cardiothoracic ratio in fetal four chambers using nnU-NetV2. BMC Med Inform Decis Mak 2024; 24:128. [PMID: 38773456 PMCID: PMC11106923 DOI: 10.1186/s12911-024-02527-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2024] [Accepted: 05/02/2024] [Indexed: 05/23/2024] Open
Abstract
BACKGROUND Accurate segmentation of critical anatomical structures in fetal four-chamber view images is essential for the early detection of congenital heart defects. Current prenatal screening methods rely on manual measurements, which are time-consuming and prone to inter-observer variability. This study develops an AI-based model using the state-of-the-art nnU-NetV2 architecture for automatic segmentation and measurement of key anatomical structures in fetal four-chamber view images. METHODS A dataset, consisting of 1,083 high-quality fetal four-chamber view images, was annotated with 15 critical anatomical labels and divided into training/validation (867 images) and test (216 images) sets. An AI-based model using the nnU-NetV2 architecture was trained on the annotated images and evaluated using the mean Dice coefficient (mDice) and mean intersection over union (mIoU) metrics. The model's performance in automatically computing the cardiac axis (CAx) and cardiothoracic ratio (CTR) was compared with measurements from sonographers with varying levels of experience. RESULTS The AI-based model achieved a mDice coefficient of 87.11% and an mIoU of 77.68% for the segmentation of critical anatomical structures. The model's automated CAx and CTR measurements showed strong agreement with those of experienced sonographers, with respective intraclass correlation coefficients (ICCs) of 0.83 and 0.81. Bland-Altman analysis further confirmed the high agreement between the model and experienced sonographers. CONCLUSION We developed an AI-based model using the nnU-NetV2 architecture for accurate segmentation and automated measurement of critical anatomical structures in fetal four-chamber view images. Our model demonstrated high segmentation accuracy and strong agreement with experienced sonographers in computing clinically relevant parameters. This approach has the potential to improve the efficiency and reliability of prenatal cardiac screening, ultimately contributing to the early detection of congenital heart defects.
Collapse
Affiliation(s)
- Bocheng Liang
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Fengfeng Peng
- Department of Computer Science and Electronic Engineering, Hunan University, Changsha, 410082, China
| | - Dandan Luo
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Qing Zeng
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Huaxuan Wen
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Bowen Zheng
- Department of Computer Science and Electronic Engineering, Hunan University, Changsha, 410082, China
| | - Zhiying Zou
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Liting An
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Huiying Wen
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Xin Wen
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Yimei Liao
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Ying Yuan
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China
| | - Shengli Li
- Department of Ultrasound, Shenzhen Maternity&Child Healthcare Hospital, Shenzhen, 518028, China.
| |
Collapse
|