1
|
Ahluwalia VS, Doiphode N, Mankowski WC, Cohen EA, Pati S, Pantalone L, Bakas S, Brooks A, Vachon CM, Conant EF, Gastounioti A, Kontos D. Volumetric Breast Density Estimation From Three-Dimensional Reconstructed Digital Breast Tomosynthesis Images Using Deep Learning. JCO Clin Cancer Inform 2024; 8:e2400103. [PMID: 39652797 PMCID: PMC11643139 DOI: 10.1200/cci.24.00103] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2024] [Revised: 08/06/2024] [Accepted: 10/14/2024] [Indexed: 12/15/2024] Open
Abstract
PURPOSE Breast density is a widely established independent breast cancer risk factor. With the increasing utilization of digital breast tomosynthesis (DBT) in breast cancer screening, there is an opportunity to estimate volumetric breast density (VBD) routinely. However, current available methods extrapolate VBD from two-dimensional (2D) images acquired using DBT and/or depend on the existence of raw DBT data, which is rarely archived by clinical centers because of storage constraints. METHODS We retrospectively analyzed 1,080 nonactionable three-dimensional (3D) reconstructed DBT screening examinations acquired between 2011 and 2016. Reference tissue segmentations were generated using previously validated software that uses 3D reconstructed slices and raw 2D DBT data. We developed a deep learning (DL) model that segments dense and fatty breast tissue from background. We then applied this model to estimate %VBD and absolute dense volume (ADV) in cm3 in a separate case-control sample (180 cases and 654 controls). We created two conditional logistic regression models, relating each model-derived density measurement to likelihood of contralateral breast cancer diagnosis, adjusted for age, BMI, family history, and menopausal status. RESULTS The DL model achieved unweighted and weighted Dice scores of 0.88 (standard deviation [SD] = 0.08) and 0.76 (SD = 0.15), respectively, on the held-out test set, demonstrating good agreement between the model and 3D reference segmentations. There was a significant association between the odds of breast cancer diagnosis and model-derived VBD (odds ratio [OR], 1.41 [95 % CI, 1.13 to 1.77]; P = .002), with an AUC of 0.65 (95% CI, 0.60 to 0.69). ADV was also significantly associated with breast cancer diagnosis (OR, 1.45 [95% CI, 1.22 to 1.73]; P < .001) with an AUC of 0.67 (95% CI, 0.62 to 0.71). CONCLUSION DL-derived density measures derived from 3D reconstructed DBT images are associated with breast cancer diagnosis.
Collapse
Affiliation(s)
- Vinayak S. Ahluwalia
- Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA
- Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA
- Leonard Davis Institute of Health Economics, University of Pennsylvania, Philadelphia, PA
| | - Nehal Doiphode
- Department of Radiology, Columbia University Irving Medical Center, New York, NY
| | - Walter C. Mankowski
- Department of Radiology, Columbia University Irving Medical Center, New York, NY
| | - Eric A. Cohen
- Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA
| | - Sarthak Pati
- Department of Pathology and Laboratory Medicine, Indiana University School of Medicine, Indianapolis, IN
| | - Lauren Pantalone
- Center for Biomedical Image Computing and Analytics (CBICA), University of Pennsylvania, Philadelphia, PA
| | - Spyridon Bakas
- Department of Surgery, Division of Endocrine and Oncologic Surgery, Hospital of the University of Pennsylvania, Philadelphia, PA
| | - Ari Brooks
- Department of Surgery, Division of Endocrine and Oncologic Surgery, Hospital of the University of Pennsylvania, Philadelphia, PA
| | - Celine M. Vachon
- Department of Quantitative Health Sciences, Division of Epidemiology, Mayo Clinic, Rochester, MN
| | - Emily F. Conant
- Department of Radiology, Hospital of the University of Pennsylvania, Philadelphia, PA
| | - Aimilia Gastounioti
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, MO
| | - Despina Kontos
- Department of Radiology, Columbia University Irving Medical Center, New York, NY
| |
Collapse
|
2
|
Al-Battal AF, Duong STM, Nguyen CDT, Truong SQH, Phan C, Nguyen TQ, An C. Efficient In-Training Adaptive Compound Loss Function Contribution Control for Medical Image Segmentation. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-6. [PMID: 40040122 DOI: 10.1109/embc53108.2024.10781657] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
Image segmentation plays a crucial role in many clinical applications, including disease diagnosis and monitoring. Current state-of-the-art segmentation approaches use deep neural networks that are trained on their target tasks by minimizing a loss function. Class imbalance is one of the major challenges that these networks face, where the target object is significantly underrepresented. Compound loss functions that incorporate the binary cross-entropy (BCE) and Dice loss are among the most prominent approaches to address this issue. However, determining the contribution of each individual loss to the overall compound loss function is a tedious process. It requires hyperparameter fine-tuning and multiple iterations of training, which is highly inefficient in terms of time and energy consumption. To address this issue, we propose an approach that adaptively controls the contribution of each of these individual loss functions during training. This eliminates the need for multiple fine-tuning iterations to achieve the desired precision and recall for segmentation models.
Collapse
|
3
|
Fu G, El Jurdi R, Chougar L, Dormont D, Valabregue R, Lehéricy S, Colliot O. Projected pooling loss for red nucleus segmentation with soft topology constraints. J Med Imaging (Bellingham) 2024; 11:044002. [PMID: 38988992 PMCID: PMC11232703 DOI: 10.1117/1.jmi.11.4.044002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Revised: 05/28/2024] [Accepted: 06/11/2024] [Indexed: 07/12/2024] Open
Abstract
Purpose Deep learning is the standard for medical image segmentation. However, it may encounter difficulties when the training set is small. Also, it may generate anatomically aberrant segmentations. Anatomical knowledge can be potentially useful as a constraint in deep learning segmentation methods. We propose a loss function based on projected pooling to introduce soft topological constraints. Our main application is the segmentation of the red nucleus from quantitative susceptibility mapping (QSM) which is of interest in parkinsonian syndromes. Approach This new loss function introduces soft constraints on the topology by magnifying small parts of the structure to segment to avoid that they are discarded in the segmentation process. To that purpose, we use projection of the structure onto the three planes and then use a series of MaxPooling operations with increasing kernel sizes. These operations are performed both for the ground truth and the prediction and the difference is computed to obtain the loss function. As a result, it can reduce topological errors as well as defects in the structure boundary. The approach is easy to implement and computationally efficient. Results When applied to the segmentation of the red nucleus from QSM data, the approach led to a very high accuracy (Dice 89.9%) and no topological errors. Moreover, the proposed loss function improved the Dice accuracy over the baseline when the training set was small. We also studied three tasks from the medical segmentation decathlon challenge (MSD) (heart, spleen, and hippocampus). For the MSD tasks, the Dice accuracies were similar for both approaches but the topological errors were reduced. Conclusions We propose an effective method to automatically segment the red nucleus which is based on a new loss for introducing topology constraints in deep learning segmentation.
Collapse
Affiliation(s)
- Guanghui Fu
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
| | - Rosana El Jurdi
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
| | - Lydia Chougar
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
- ICM, Centre de NeuroImagerie de Recherche-CENIR, Paris, France
- Hôpital de la Pitié Salpêtrière, AP-HP, DMU DIAMENT, Department of Neuroradiology, Paris, France
- McGill University, The Neuro (Montreal Neurological Institute – MNI), Montreal, Canada
| | - Didier Dormont
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
- Hôpital de la Pitié Salpêtrière, AP-HP, DMU DIAMENT, Department of Neuroradiology, Paris, France
| | - Romain Valabregue
- ICM, Centre de NeuroImagerie de Recherche-CENIR, Paris, France
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
| | - Stéphane Lehéricy
- ICM, Centre de NeuroImagerie de Recherche-CENIR, Paris, France
- Hôpital de la Pitié Salpêtrière, AP-HP, DMU DIAMENT, Department of Neuroradiology, Paris, France
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
| | - Olivier Colliot
- Sorbonne Université, Institut du Cerveau – Paris Brain Institute – ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié Salpêtrière, Paris, France
| | | |
Collapse
|
4
|
Ellison J, Caliva F, Damasceno P, Luks TL, LaFontaine M, Cluceru J, Kemisetti A, Li Y, Molinaro AM, Pedoia V, Villanueva-Meyer JE, Lupo JM. Improving the Generalizability of Deep Learning for T2-Lesion Segmentation of Gliomas in the Post-Treatment Setting. Bioengineering (Basel) 2024; 11:497. [PMID: 38790363 PMCID: PMC11117752 DOI: 10.3390/bioengineering11050497] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2024] [Revised: 04/24/2024] [Accepted: 05/07/2024] [Indexed: 05/26/2024] Open
Abstract
Although fully automated volumetric approaches for monitoring brain tumor response have many advantages, most available deep learning models are optimized for highly curated, multi-contrast MRI from newly diagnosed gliomas, which are not representative of post-treatment cases in the clinic. Improving segmentation for treated patients is critical to accurately tracking changes in response to therapy. We investigated mixing data from newly diagnosed (n = 208) and treated (n = 221) gliomas in training, applying transfer learning (TL) from pre- to post-treatment imaging domains, and incorporating spatial regularization for T2-lesion segmentation using only T2 FLAIR images as input to improve generalization post-treatment. These approaches were evaluated on 24 patients suspected of progression who had received prior treatment. Including 26% of treated patients in training improved performance by 13.9%, and including more treated and untreated patients resulted in minimal changes. Fine-tuning with treated glioma improved sensitivity compared to data mixing by 2.5% (p < 0.05), and spatial regularization further improved performance when used with TL by 95th HD, Dice, and sensitivity (6.8%, 0.8%, 2.2%; p < 0.05). While training with ≥60 treated patients yielded the majority of performance gain, TL and spatial regularization further improved T2-lesion segmentation to treated gliomas using a single MR contrast and minimal processing, demonstrating clinical utility in response assessment.
Collapse
Affiliation(s)
- Jacob Ellison
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
- UCSF/UC Berkeley Graduate Program in Bioengineering, San Francisco, CA 94143, USA
| | - Francesco Caliva
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
| | - Pablo Damasceno
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
| | - Tracy L. Luks
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
| | - Marisa LaFontaine
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
| | - Julia Cluceru
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
| | - Anil Kemisetti
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
| | - Yan Li
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
| | | | - Valentina Pedoia
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
- UCSF/UC Berkeley Graduate Program in Bioengineering, San Francisco, CA 94143, USA
| | - Javier E. Villanueva-Meyer
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
| | - Janine M. Lupo
- Department of Radiology and Biomedical Imaging, UCSF, San Francisco, CA 94143, USA; (J.E.); (F.C.); (P.D.); (T.L.L.); (M.L.); (J.C.); (A.K.); (Y.L.); (V.P.); (J.E.V.-M.)
- Center for Intelligent Imaging, UCSF, San Francisco, CA 94143, USA
- UCSF/UC Berkeley Graduate Program in Bioengineering, San Francisco, CA 94143, USA
| |
Collapse
|
5
|
Sugino T, Kin T, Saito N, Nakajima Y. Improved segmentation of basal ganglia from MR images using convolutional neural network with crossover-typed skip connection. Int J Comput Assist Radiol Surg 2024; 19:433-442. [PMID: 37982960 DOI: 10.1007/s11548-023-03015-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2023] [Accepted: 08/29/2023] [Indexed: 11/21/2023]
Abstract
PURPOSE Accurate and automatic segmentation of basal ganglia from magnetic resonance (MR) images is important for diagnosis and treatment of various brain disorders. However, the basal ganglia segmentation is a challenging task because of the class imbalance and the unclear boundaries among basal ganglia anatomical structures. Thus, we aim to present an encoder-decoder convolutional neural network (CNN)-based method for improved segmentation of basal ganglia by focusing on skip connections that determine the segmentation performance of encoder-decoder CNNs. We also aim to reveal the effect of skip connections on the segmentation of basal ganglia with unclear boundaries. METHODS We used the encoder-decoder CNNs with the following five patterns of skip connections: without skip connection, with full-resolution horizontal skip connection, with horizontal skip connections, with vertical skip connections, and with crossover-typed skip connections (the proposed method). We compared and evaluated the performance of the CNNs in the experiment of basal ganglia segmentation using T1-weighted MR brain images of 79 patients. RESULTS The experimental results showed that the skip connections at each scale level help CNNs to acquire multi-scale image features, the vertical skip connections contribute on acquiring finer image features for segmentation of smaller anatomical structures with more blurred boundaries, and the crossover-typed skip connections, a combination of horizontal and vertical skip connections, provided better segmentation accuracy. CONCLUSION This paper investigated the effect of skip connections on the basal ganglia segmentation and revealed the crossover-typed skip connections might be effective for improving the segmentation of basal ganglia with the class imbalance and the unclear boundaries.
Collapse
Affiliation(s)
- Takaaki Sugino
- Department of Biomedical Informatics, Institute of Biomaterials and Bioengineering, Tokyo Medical and Dental University, Tokyo, Japan.
| | - Taichi Kin
- Department of Neurosurgery, Graduate School of Medicine, The University of Tokyo, Tokyo, Japan
| | - Nobuhito Saito
- Department of Neurosurgery, Graduate School of Medicine, The University of Tokyo, Tokyo, Japan
| | - Yoshikazu Nakajima
- Department of Biomedical Informatics, Institute of Biomaterials and Bioengineering, Tokyo Medical and Dental University, Tokyo, Japan
| |
Collapse
|
6
|
Chaki J, Woźniak M. A deep learning based four-fold approach to classify brain MRI: BTSCNet. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/03/2023]
|
7
|
Rezk E, Eltorki M, El-Dakhakhni W. Interpretable Skin Cancer Classification based on Incremental Domain Knowledge Learning. JOURNAL OF HEALTHCARE INFORMATICS RESEARCH 2023; 7:59-83. [PMID: 36910915 PMCID: PMC9995827 DOI: 10.1007/s41666-023-00127-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2022] [Revised: 01/02/2023] [Accepted: 02/03/2023] [Indexed: 02/17/2023]
Abstract
The recent advances in artificial intelligence have led to the rapid development of computer-aided skin cancer diagnosis applications that perform on par with dermatologists. However, the black-box nature of such applications makes it difficult for physicians to trust the predicted decisions, subsequently preventing the proliferation of such applications in the clinical workflow. In this work, we aim to address this challenge by developing an interpretable skin cancer diagnosis approach using clinical images. Accordingly, a skin cancer diagnosis model consolidated with two interpretability methods is developed. The first interpretability method integrates skin cancer diagnosis domain knowledge, characterized by a skin lesion taxonomy, into model development, whereas the other method focuses on visualizing the decision-making process by highlighting the dominant of interest regions of skin lesion images. The proposed model is trained and validated on clinical images since the latter are easily obtainable by non-specialist healthcare providers. The results demonstrate the effectiveness of incorporating lesion taxonomy in improving model classification accuracy, where our model can predict the skin lesion origin as melanocytic or non-melanocytic with an accuracy of 87%, predict lesion malignancy with 77% accuracy, and provide disease diagnosis with an accuracy of 71%. In addition, the implemented interpretability methods assist understand the model's decision-making process and detecting misdiagnoses. This work is a step toward achieving interpretability in skin cancer diagnosis using clinical images. The developed approach can assist general practitioners to make an early diagnosis, thus reducing the redundant referrals that expert dermatologists receive for further investigations.
Collapse
Affiliation(s)
- Eman Rezk
- School of Computational Science and Engineering, McMaster University, Hamilton, ON Canada
| | - Mohamed Eltorki
- Faculty of Health Sciences, McMaster University, Hamilton, ON Canada
| | - Wael El-Dakhakhni
- School of Computational Science and Engineering, McMaster University, Hamilton, ON Canada
| |
Collapse
|
8
|
Tchetchenian A, Zhu Y, Zhang F, O'Donnell LJ, Song Y, Meijering E. A comparison of manual and automated neural architecture search for white matter tract segmentation. Sci Rep 2023; 13:1617. [PMID: 36709392 PMCID: PMC9884270 DOI: 10.1038/s41598-023-28210-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2022] [Accepted: 01/16/2023] [Indexed: 01/30/2023] Open
Abstract
Segmentation of white matter tracts in diffusion magnetic resonance images is an important first step in many imaging studies of the brain in health and disease. Similar to medical image segmentation in general, a popular approach to white matter tract segmentation is to use U-Net based artificial neural network architectures. Despite many suggested improvements to the U-Net architecture in recent years, there is a lack of systematic comparison of architectural variants for white matter tract segmentation. In this paper, we evaluate multiple U-Net based architectures specifically for this purpose. We compare the results of these networks to those achieved by our own various architecture changes, as well as to new U-Net architectures designed automatically via neural architecture search (NAS). To the best of our knowledge, this is the first study to systematically compare multiple U-Net based architectures for white matter tract segmentation, and the first to use NAS. We find that the recently proposed medical imaging segmentation network UNet3+ slightly outperforms the current state of the art for white matter tract segmentation, and achieves a notably better mean Dice score for segmentation of the fornix (+ 0.01 and + 0.006 mean Dice increase for left and right fornix respectively), a tract that the current state of the art model struggles to segment. UNet3+ also outperforms the current state of the art when little training data is available. Additionally, manual architecture search found that a minor segmentation improvement is observed when an additional, deeper layer is added to the U-shape of UNet3+. However, all networks, including those designed via NAS, achieve similar results, suggesting that there may be benefit in exploring networks that deviate from the general U-Net paradigm.
Collapse
Affiliation(s)
- Ari Tchetchenian
- Biomedical Image Computing Group, School of Computer Science and Engineering, University of New South Wales (UNSW), Sydney, NSW, Australia.
| | - Yanming Zhu
- Biomedical Image Computing Group, School of Computer Science and Engineering, University of New South Wales (UNSW), Sydney, NSW, Australia
| | - Fan Zhang
- Brigham and Women's Hospital, Harvard Medical School, Boston, USA
| | | | - Yang Song
- Biomedical Image Computing Group, School of Computer Science and Engineering, University of New South Wales (UNSW), Sydney, NSW, Australia
| | - Erik Meijering
- Biomedical Image Computing Group, School of Computer Science and Engineering, University of New South Wales (UNSW), Sydney, NSW, Australia
| |
Collapse
|
9
|
Genc A, Kovarik L, Fraser HL. A deep learning approach for semantic segmentation of unbalanced data in electron tomography of catalytic materials. Sci Rep 2022; 12:16267. [PMID: 36171204 PMCID: PMC9519981 DOI: 10.1038/s41598-022-16429-3] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2022] [Accepted: 07/11/2022] [Indexed: 11/09/2022] Open
Abstract
In computed TEM tomography, image segmentation represents one of the most basic tasks with implications not only for 3D volume visualization, but more importantly for quantitative 3D analysis. In case of large and complex 3D data sets, segmentation can be an extremely difficult and laborious task, and thus has been one of the biggest hurdles for comprehensive 3D analysis. Heterogeneous catalysts have complex surface and bulk structures, and often sparse distribution of catalytic particles with relatively poor intrinsic contrast, which possess a unique challenge for image segmentation, including the current state-of-the-art deep learning methods. To tackle this problem, we apply a deep learning-based approach for the multi-class semantic segmentation of a γ-Alumina/Pt catalytic material in a class imbalance situation. Specifically, we used the weighted focal loss as a loss function and attached it to the U-Net's fully convolutional network architecture. We assessed the accuracy of our results using Dice similarity coefficient (DSC), recall, precision, and Hausdorff distance (HD) metrics on the overlap between the ground-truth and predicted segmentations. Our adopted U-Net model with the weighted focal loss function achieved an average DSC score of 0.96 ± 0.003 in the γ-Alumina support material and 0.84 ± 0.03 in the Pt NPs segmentation tasks. We report an average boundary-overlap error of less than 2 nm at the 90th percentile of HD for γ-Alumina and Pt NPs segmentations. The complex surface morphology of γ-Alumina and its relation to the Pt NPs were visualized in 3D by the deep learning-assisted automatic segmentation of a large data set of high-angle annular dark-field (HAADF) scanning transmission electron microscopy (STEM) tomography reconstructions.
Collapse
Affiliation(s)
- Arda Genc
- Center for the Accelerated Maturation of Materials, Department of Materials Science and Engineering, The Ohio State University, Columbus, OH, USA
- Materials Department, University of California Santa Barbara, Santa Barbara, CA, USA
| | - Libor Kovarik
- Institute for Integrated Catalysis, Pacific Northwest National Laboratory, Richland, WA, USA.
| | - Hamish L Fraser
- Center for the Accelerated Maturation of Materials, Department of Materials Science and Engineering, The Ohio State University, Columbus, OH, USA
| |
Collapse
|
10
|
Estrade V, Daudon M, Richard E, Bernhard JC, Bladou F, Robert G, Facq L, Denis de Senneville B. Deep morphological recognition of kidney stones using intra-operative endoscopic digital videos. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac8592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Accepted: 07/29/2022] [Indexed: 11/11/2022]
Abstract
Abstract
Objective. To assess the performance and added value of processing complete digital endoscopic video sequences for the automatic recognition of stone morphological features during a standard-of-care intra-operative session.Approach. A computer-aided video classifier was developed to predict in-situ the morphology of stone using an intra-operative digital endoscopic video acquired in a clinical setting. Using dedicated artificial intelligence (AI) networks, the proposed pipeline selects adequate frames in steady sequences of the video, ensures the presence of (potentially fragmented) stones and predicts the stone morphologies on a frame-by-frame basis. The automatic endoscopic stone recognition (A-ESR) is subsequently carried out by mixing all collected morphological observations.Main results. The proposed technique was evaluated on pure (i.e. include one morphology) and mixed (i.e. include at least two morphologies) stones involving ‘Ia/Calcium Oxalate Monohydrate’ (COM), ‘IIb/Calcium Oxalate Dihydrate’ (COD) and ‘IIIb/Uric Acid’ (UA) morphologies. The gold standard ESR was provided by a trained endo-urologist and confirmed by microscopy and infra-red spectroscopy. For the AI-training, 585 static images were collected (349 and 236 observations of stone surface and section, respectively) and used. Using the proposed video classifier, 71 digital endoscopic videos were analyzed: 50 exhibited only one morphological type and 21 displayed two. Taken together, both pure and mixed stone types yielded a mean diagnostic performances as follows: balanced accuracy = [88 ± 6] (min = 81)%, sensitivity = [80 ± 13] (min = 69)%, specificity = [95 ± 2] (min = 92)%, precision = [78 ± 12] (min = 62)% and F1-score = [78 ± 7] (min = 69)%.Significance. These results demonstrate that AI applied on digital endoscopic video sequences is a promising tool for collecting morphological information during the time-course of the stone fragmentation process without resorting to any human intervention for stone delineation or the selection of adequate steady frames.
Collapse
|
11
|
Deep 3D Neural Network for Brain Structures Segmentation Using Self-Attention Modules in MRI Images. SENSORS 2022; 22:s22072559. [PMID: 35408173 PMCID: PMC9002763 DOI: 10.3390/s22072559] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 03/15/2022] [Accepted: 03/21/2022] [Indexed: 01/03/2023]
Abstract
In recent years, the use of deep learning-based models for developing advanced healthcare systems has been growing due to the results they can achieve. However, the majority of the proposed deep learning-models largely use convolutional and pooling operations, causing a loss in valuable data and focusing on local information. In this paper, we propose a deep learning-based approach that uses global and local features which are of importance in the medical image segmentation process. In order to train the architecture, we used extracted three-dimensional (3D) blocks from the full magnetic resonance image resolution, which were sent through a set of successive convolutional neural network (CNN) layers free of pooling operations to extract local information. Later, we sent the resulting feature maps to successive layers of self-attention modules to obtain the global context, whose output was later dispatched to the decoder pipeline composed mostly of upsampling layers. The model was trained using the Mindboggle-101 dataset. The experimental results showed that the self-attention modules allow segmentation with a higher Mean Dice Score of 0.90 ± 0.036 compared with other UNet-based approaches. The average segmentation time was approximately 0.038 s per brain structure. The proposed model allows tackling the brain structure segmentation task properly. Exploiting the global context that the self-attention modules incorporate allows for more precise and faster segmentation. We segmented 37 brain structures and, to the best of our knowledge, it is the largest number of structures under a 3D approach using attention mechanisms.
Collapse
|
12
|
Mahmood U, Bates DDB, Erdi YE, Mannelli L, Corrias G, Kanan C. Deep Learning and Domain-Specific Knowledge to Segment the Liver from Synthetic Dual Energy CT Iodine Scans. Diagnostics (Basel) 2022; 12:672. [PMID: 35328225 PMCID: PMC8947702 DOI: 10.3390/diagnostics12030672] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2022] [Revised: 02/24/2022] [Accepted: 03/03/2022] [Indexed: 11/23/2022] Open
Abstract
We map single energy CT (SECT) scans to synthetic dual-energy CT (synth-DECT) material density iodine (MDI) scans using deep learning (DL) and demonstrate their value for liver segmentation. A 2D pix2pix (P2P) network was trained on 100 abdominal DECT scans to infer synth-DECT MDI scans from SECT scans. The source and target domain were paired with DECT monochromatic 70 keV and MDI scans. The trained P2P algorithm then transformed 140 public SECT scans to synth-DECT scans. We split 131 scans into 60% train, 20% tune, and 20% held-out test to train four existing liver segmentation frameworks. The remaining nine low-dose SECT scans tested system generalization. Segmentation accuracy was measured with the dice coefficient (DSC). The DSC per slice was computed to identify sources of error. With synth-DECT (and SECT) scans, an average DSC score of 0.93±0.06 (0.89±0.01) and 0.89±0.01 (0.81±0.02) was achieved on the held-out and generalization test sets. Synth-DECT-trained systems required less data to perform as well as SECT-trained systems. Low DSC scores were primarily observed around the scan margin or due to non-liver tissue or distortions within ground-truth annotations. In general, training with synth-DECT scans resulted in improved segmentation performance with less data.
Collapse
Affiliation(s)
- Usman Mahmood
- Department of Medical Physics, Memorial Sloan Kettering Cancer Center, New York, NY 10065, USA;
| | - David D. B. Bates
- Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY 10065, USA;
| | - Yusuf E. Erdi
- Department of Medical Physics, Memorial Sloan Kettering Cancer Center, New York, NY 10065, USA;
| | | | - Giuseppe Corrias
- Department of Radiology, University of Cagliari, 09124 Cagliari, Italy;
| | - Christopher Kanan
- Chester F. Carlson Center for Imaging Science, Rochester Institute of Technology, Rochester, NY 14623, USA;
| |
Collapse
|
13
|
Chanchal AK, Lal S, Kini J. Deep structured residual encoder-decoder network with a novel loss function for nuclei segmentation of kidney and breast histopathology images. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 81:9201-9224. [PMID: 35125928 PMCID: PMC8809220 DOI: 10.1007/s11042-021-11873-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2020] [Revised: 11/10/2021] [Accepted: 12/23/2021] [Indexed: 06/14/2023]
Abstract
To improve the process of diagnosis and treatment of cancer disease, automatic segmentation of haematoxylin and eosin (H & E) stained cell nuclei from histopathology images is the first step in digital pathology. The proposed deep structured residual encoder-decoder network (DSREDN) focuses on two aspects: first, it effectively utilized residual connections throughout the network and provides a wide and deep encoder-decoder path, which results to capture relevant context and more localized features. Second, vanished boundary of detected nuclei is addressed by proposing an efficient loss function that better train our proposed model and reduces the false prediction which is undesirable especially in healthcare applications. The proposed architecture experimented on three different publicly available H&E stained histopathological datasets namely: (I) Kidney (RCC) (II) Triple Negative Breast Cancer (TNBC) (III) MoNuSeg-2018. We have considered F1-score, Aggregated Jaccard Index (AJI), the total number of parameters, and FLOPs (Floating point operations), which are mostly preferred performance measure metrics for comparison of nuclei segmentation. The evaluated score of nuclei segmentation indicated that the proposed architecture achieved a considerable margin over five state-of-the-art deep learning models on three different histopathology datasets. Visual segmentation results show that the proposed DSREDN model accurately segment the nuclear regions than those of the state-of-the-art methods.
Collapse
Affiliation(s)
- Amit Kumar Chanchal
- Department of Electronics and Communication Engineering, National Institute of Technology Karnataka, Surathkal, Mangaluru-575025 Karnataka India
| | - Shyam Lal
- Department of Electronics and Communication Engineering, National Institute of Technology Karnataka, Surathkal, Mangaluru-575025 Karnataka India
| | - Jyoti Kini
- Department of Pathology, Kasturba Medical College Mangalore, Manipal Academy of Higher Education, Manipal, India
| |
Collapse
|
14
|
Benlala I, De Senneville BD, Dournes G, Menant M, Gramond C, Thaon I, Clin B, Brochard P, Gislard A, Andujar P, Chammings S, Gallet J, Lacourt A, Delva F, Paris C, Ferretti G, Pairon JC, Laurent F. Deep Learning for the Automatic Quantification of Pleural Plaques in Asbestos-Exposed Subjects. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:ijerph19031417. [PMID: 35162440 PMCID: PMC8835296 DOI: 10.3390/ijerph19031417] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/16/2021] [Revised: 01/22/2022] [Accepted: 01/23/2022] [Indexed: 12/10/2022]
Abstract
Objective: This study aimed to develop and validate an automated artificial intelligence (AI)-driven quantification of pleural plaques in a population of retired workers previously occupationally exposed to asbestos. Methods: CT scans of former workers previously occupationally exposed to asbestos who participated in the multicenter APEXS (Asbestos PostExposure Survey) study were collected retrospectively between 2010 and 2017 during the second and the third rounds of the survey. A hundred and forty-one participants with pleural plaques identified by expert radiologists at the 2nd and the 3rd CT screenings were included. Maximum Intensity Projection (MIP) with 5 mm thickness was used to reduce the number of CT slices for manual delineation. A Deep Learning AI algorithm using 2D-convolutional neural networks was trained with 8280 images from 138 CT scans of 69 participants for the semantic labeling of Pleural Plaques (PP). In all, 2160 CT images from 36 CT scans of 18 participants were used for AI testing versus ground-truth labels (GT). The clinical validity of the method was evaluated longitudinally in 54 participants with pleural plaques. Results: The concordance correlation coefficient (CCC) between AI-driven and GT was almost perfect (>0.98) for the volume extent of both PP and calcified PP. The 2D pixel similarity overlap of AI versus GT was good (DICE = 0.63) for PP, whether they were calcified or not, and very good (DICE = 0.82) for calcified PP. A longitudinal comparison of the volumetric extent of PP showed a significant increase in PP volumes (p < 0.001) between the 2nd and the 3rd CT screenings with an average delay of 5 years. Conclusions: AI allows a fully automated volumetric quantification of pleural plaques showing volumetric progression of PP over a five-year period. The reproducible PP volume evaluation may enable further investigations for the comprehension of the unclear relationships between pleural plaques and both respiratory function and occurrence of thoracic malignancy.
Collapse
Affiliation(s)
- Ilyes Benlala
- Faculté de Médecine, Université de Bordeaux, 33000 Bordeaux, France; (G.D.); (P.B.); (F.L.)
- Service d’Imagerie Médicale Radiologie Diagnostique et Thérapeutique, CHU de Bordeaux, 33000 Bordeaux, France
- Centre de Recherche Cardio-Thoracique de Bordeaux, INSERM U1045, Université de Bordeaux, 33000 Bordeaux, France
- Correspondence:
| | - Baudouin Denis De Senneville
- Mathematical Institute of Bordeaux (IMB), CNRS, INRIA, Bordeaux INP, UMR 5251, Université de Bordeaux, 33400 Talence, France;
| | - Gael Dournes
- Faculté de Médecine, Université de Bordeaux, 33000 Bordeaux, France; (G.D.); (P.B.); (F.L.)
- Service d’Imagerie Médicale Radiologie Diagnostique et Thérapeutique, CHU de Bordeaux, 33000 Bordeaux, France
- Centre de Recherche Cardio-Thoracique de Bordeaux, INSERM U1045, Université de Bordeaux, 33000 Bordeaux, France
| | - Morgane Menant
- Epicene Team, Bordeaux Population Health Research Center, INSERM UMR 1219, Université de Bordeaux, 33000 Bordeaux, France; (M.M.); (C.G.); (J.G.); (A.L.); (F.D.)
| | - Celine Gramond
- Epicene Team, Bordeaux Population Health Research Center, INSERM UMR 1219, Université de Bordeaux, 33000 Bordeaux, France; (M.M.); (C.G.); (J.G.); (A.L.); (F.D.)
| | - Isabelle Thaon
- Centre de Consultation de Pathologies Professionnelles, CHRU de Nancy, Université de Lorraine, 54000 Nancy, France;
| | - Bénédicte Clin
- Service de Santé au Travail et Pathologie Professionnelle, CHU Caen, 14000 Caen, France;
- Faculté de Médecine, Université de Caen, 14000 Caen, France
| | - Patrick Brochard
- Faculté de Médecine, Université de Bordeaux, 33000 Bordeaux, France; (G.D.); (P.B.); (F.L.)
- Service de Médecine du Travail et de Pathologies Professionnelles, CHU de Bordeaux, 33000 Bordeaux, France
| | - Antoine Gislard
- Faculté de Médecine, Normandie Université, UNIROUEN, UNICAEN, ABTE, 76000 Rouen, France;
- Centre de Consultations de Pathologie Professionnelle, CHU de Rouen, CEDEX, 76031 Rouen, France
| | - Pascal Andujar
- Equipe GEIC20, INSERM U955, 94000 Créteil, France; (P.A.); (J.-C.P.)
- Faculté de Santé, Université Paris-Est Créteil, 94000 Créteil, France
- Service de Pathologies Professionnelles et de l’Environnement, Centre Hospitalier Intercommunal Créteil, Institut Santé-Travail Paris-Est, 94000 Créteil, France
- Institut Interuniversitaire de Médecine du Travail de Paris-Ile de France, 94000 Créteil, France;
| | - Soizick Chammings
- Institut Interuniversitaire de Médecine du Travail de Paris-Ile de France, 94000 Créteil, France;
| | - Justine Gallet
- Epicene Team, Bordeaux Population Health Research Center, INSERM UMR 1219, Université de Bordeaux, 33000 Bordeaux, France; (M.M.); (C.G.); (J.G.); (A.L.); (F.D.)
| | - Aude Lacourt
- Epicene Team, Bordeaux Population Health Research Center, INSERM UMR 1219, Université de Bordeaux, 33000 Bordeaux, France; (M.M.); (C.G.); (J.G.); (A.L.); (F.D.)
| | - Fleur Delva
- Epicene Team, Bordeaux Population Health Research Center, INSERM UMR 1219, Université de Bordeaux, 33000 Bordeaux, France; (M.M.); (C.G.); (J.G.); (A.L.); (F.D.)
| | - Christophe Paris
- Service de Santé au Travail et Pathologie Professionnelle, CHU Rennes, 35000 Rennes, France;
- Institut de Recherche en Santé, Environnement et Travail, INSERM U1085, 35000 Rennes, France
| | - Gilbert Ferretti
- INSERM U 1209 IAB, 38700 La Tronche, France;
- Domaine de la Merci, Faculté de Médecine, Université Grenoble Alpes, 38706 La Tronche, France
- Service de Radiologie Diagnostique et Interventionnelle Nord, CHU Grenoble Alpes, CS 10217, 38043 Grenoble, France
| | - Jean-Claude Pairon
- Equipe GEIC20, INSERM U955, 94000 Créteil, France; (P.A.); (J.-C.P.)
- Faculté de Santé, Université Paris-Est Créteil, 94000 Créteil, France
- Service de Pathologies Professionnelles et de l’Environnement, Centre Hospitalier Intercommunal Créteil, Institut Santé-Travail Paris-Est, 94000 Créteil, France
- Institut Interuniversitaire de Médecine du Travail de Paris-Ile de France, 94000 Créteil, France;
| | - François Laurent
- Faculté de Médecine, Université de Bordeaux, 33000 Bordeaux, France; (G.D.); (P.B.); (F.L.)
- Service d’Imagerie Médicale Radiologie Diagnostique et Thérapeutique, CHU de Bordeaux, 33000 Bordeaux, France
- Centre de Recherche Cardio-Thoracique de Bordeaux, INSERM U1045, Université de Bordeaux, 33000 Bordeaux, France
| |
Collapse
|
15
|
Deciphering tumour tissue organization by 3D electron microscopy and machine learning. Commun Biol 2021; 4:1390. [PMID: 34903822 PMCID: PMC8668903 DOI: 10.1038/s42003-021-02919-z] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2021] [Accepted: 11/12/2021] [Indexed: 12/02/2022] Open
Abstract
Despite recent progress in the characterization of tumour components, the tri-dimensional (3D) organization of this pathological tissue and the parameters determining its internal architecture remain elusive. Here, we analysed the spatial organization of patient-derived xenograft tissues generated from hepatoblastoma, the most frequent childhood liver tumour, by serial block-face scanning electron microscopy using an integrated workflow combining 3D imaging, manual and machine learning-based semi-automatic segmentations, mathematics and infographics. By digitally reconstituting an entire hepatoblastoma sample with a blood capillary, a bile canaliculus-like structure, hundreds of tumour cells and their main organelles (e.g. cytoplasm, nucleus, mitochondria), we report unique 3D ultrastructural data about the organization of tumour tissue. We found that the size of hepatoblastoma cells correlates with the size of their nucleus, cytoplasm and mitochondrial mass. We also found anatomical connections between the blood capillary and the planar alignment and size of tumour cells in their 3D milieu. Finally, a set of tumour cells polarized in the direction of a hot spot corresponding to a bile canaliculus-like structure. In conclusion, this pilot study allowed the identification of bioarchitectural parameters that shape the internal and spatial organization of tumours, thus paving the way for future investigations in the emerging onconanotomy field. de Senneville et al. demonstrate an integrated workflow combining 3D imaging, manual and machine learning-based semi-automatic segmentation, mathematics and infographics to study the spatial organization of patient-derived hepatoblastoma xenograft tissues. Their approach potentially assists investigations of this childhood liver tumour and other types of tumour tissues.
Collapse
|