201
|
Going to Extremes: Weakly Supervised Medical Image Segmentation. MACHINE LEARNING AND KNOWLEDGE EXTRACTION 2021. [DOI: 10.3390/make3020026] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Medical image annotation is a major hurdle for developing precise and robust machine-learning models. Annotation is expensive, time-consuming, and often requires expert knowledge, particularly in the medical field. Here, we suggest using minimal user interaction in the form of extreme point clicks to train a segmentation model which, in effect, can be used to speed up medical image annotation. An initial segmentation is generated based on the extreme points using the random walker algorithm. This initial segmentation is then used as a noisy supervision signal to train a fully convolutional network that can segment the organ of interest, based on the provided user clicks. Through experimentation on several medical imaging datasets, we show that the predictions of the network can be refined using several rounds of training with the prediction from the same weakly annotated data. Further improvements are shown using the clicked points within a custom-designed loss and attention mechanism. Our approach has the potential to speed up the process of generating new training datasets for the development of new machine-learning and deep-learning-based models for, but not exclusively, medical image analysis.
Collapse
|
202
|
Kart T, Fischer M, Küstner T, Hepp T, Bamberg F, Winzeck S, Glocker B, Rueckert D, Gatidis S. Deep Learning-Based Automated Abdominal Organ Segmentation in the UK Biobank and German National Cohort Magnetic Resonance Imaging Studies. Invest Radiol 2021; 56:401-408. [PMID: 33930003 DOI: 10.1097/rli.0000000000000755] [Citation(s) in RCA: 23] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/15/2022]
Abstract
PURPOSE The aims of this study were to train and evaluate deep learning models for automated segmentation of abdominal organs in whole-body magnetic resonance (MR) images from the UK Biobank (UKBB) and German National Cohort (GNC) MR imaging studies and to make these models available to the scientific community for analysis of these data sets. METHODS A total of 200 T1-weighted MR image data sets of healthy volunteers each from UKBB and GNC (400 data sets in total) were available in this study. Liver, spleen, left and right kidney, and pancreas were segmented manually on all 400 data sets, providing labeled ground truth data for training of a previously described U-Net-based deep learning framework for automated medical image segmentation (nnU-Net). The trained models were tested on all data sets using a 4-fold cross-validation scheme. Qualitative analysis of automated segmentation results was performed visually; performance metrics between automated and manual segmentation results were computed for quantitative analysis. In addition, interobserver segmentation variability between 2 human readers was assessed on a subset of the data. RESULTS Automated abdominal organ segmentation was performed with high qualitative and quantitative accuracy on UKBB and GNC data. In more than 90% of data sets, no or only minor visually detectable qualitative segmentation errors occurred. Mean Dice scores of automated segmentations compared with manual reference segmentations were well higher than 0.9 for the liver, spleen, and kidneys on UKBB and GNC data and around 0.82 and 0.89 for the pancreas on UKBB and GNC data, respectively. Mean average symmetric surface distance was between 0.3 and 1.5 mm for the liver, spleen, and kidneys and between 2 and 2.2 mm for pancreas segmentation. The quantitative accuracy of automated segmentation was comparable with the agreement between 2 human readers for all organs on UKBB and GNC data. CONCLUSION Automated segmentation of abdominal organs is possible with high qualitative and quantitative accuracy on whole-body MR imaging data acquired as part of UKBB and GNC. The results obtained and deep learning models trained in this study can be used as a foundation for automated analysis of thousands of MR data sets of UKBB and GNC and thus contribute to tackling topical and original scientific questions.
Collapse
Affiliation(s)
- Turkay Kart
- From the Biomedical Image Analysis Group, Department of Computing, Imperial College London, London, UK
| | - Marc Fischer
- Medical Image and Data Analysis Lab, Department of Radiology, University Hospital Tübingen, Tübingen, Germany
| | - Thomas Küstner
- Medical Image and Data Analysis Lab, Department of Radiology, University Hospital Tübingen, Tübingen, Germany
| | | | - Fabian Bamberg
- Department of Radiology, University Hospital Freiburg, Freiburg, Germany
| | - Stefan Winzeck
- From the Biomedical Image Analysis Group, Department of Computing, Imperial College London, London, UK
| | - Ben Glocker
- From the Biomedical Image Analysis Group, Department of Computing, Imperial College London, London, UK
| | | | | |
Collapse
|
203
|
Hussain MA, Hamarneh G, Garbi R. Cascaded Regression Neural Nets for Kidney Localization and Segmentation-free Volume Estimation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1555-1567. [PMID: 33606626 DOI: 10.1109/tmi.2021.3060465] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Kidney volume is an essential biomarker for a number of kidney disease diagnoses, for example, chronic kidney disease. Existing total kidney volume estimation methods often rely on an intermediate kidney segmentation step. On the other hand, automatic kidney localization in volumetric medical images is a critical step that often precedes subsequent data processing and analysis. Most current approaches perform kidney localization via an intermediate classification or regression step. This paper proposes an integrated deep learning approach for (i) kidney localization in computed tomography scans and (ii) segmentation-free renal volume estimation. Our localization method uses a selection-convolutional neural network that approximates the kidney inferior-superior span along the axial direction. Cross-sectional (2D) slices from the estimated span are subsequently used in a combined sagittal-axial Mask-RCNN that detects the organ bounding boxes on the axial and sagittal slices, the combination of which produces a final 3D organ bounding box. Furthermore, we use a fully convolutional network to estimate the kidney volume that skips the segmentation procedure. We also present a mathematical expression to approximate the 'volume error' metric from the 'Sørensen-Dice coefficient.' We accessed 100 patients' CT scans from the Vancouver General Hospital records and obtained 210 patients' CT scans from the 2019 Kidney Tumor Segmentation Challenge database to validate our method. Our method produces a kidney boundary wall localization error of ~2.4mm and a mean volume estimation error of ~5%.
Collapse
|
204
|
Lei W, Mei H, Sun Z, Ye S, Gu R, Wang H, Huang R, Zhang S, Zhang S, Wang G. Automatic segmentation of organs-at-risk from head-and-neck CT using separable convolutional neural network with hard-region-weighted loss. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.01.135] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
205
|
Villarini B, Asaturyan H, Kurugol S, Afacan O, Bell JD, Thomas EL. 3D Deep Learning for Anatomical Structure Segmentation in Multiple Imaging Modalities. PROCEEDINGS. IEEE INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS 2021; 2021:166-171. [PMID: 35224185 PMCID: PMC8867534 DOI: 10.1109/cbms52027.2021.00066] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Accurate, quantitative segmentation of anatomical structures in radiological scans, such as Magnetic Resonance Imaging (MRI) and Computer Tomography (CT), can produce significant biomarkers and can be integrated into computer-aided assisted diagnosis (CADx) systems to support the interpretation of medical images from multi-protocol scanners. However, there are serious challenges towards developing robust automated segmentation techniques, including high variations in anatomical structure and size, the presence of edge-based artefacts, and heavy un-controlled breathing that can produce blurred motion-based artefacts. This paper presents a novel computing approach for automatic organ and muscle segmentation in medical images from multiple modalities by harnessing the advantages of deep learning techniques in a two-part process. (1) a 3D encoder-decoder, Rb-UNet, builds a localisation model and a 3D Tiramisu network generates a boundary-preserving segmentation model for each target structure; (2) the fully trained Rb-UNet predicts a 3D bounding box encapsulating the target structure of interest, after which the fully trained Tiramisu model performs segmentation to reveal detailed organ or muscle boundaries. The proposed approach is evaluated on six different datasets, including MRI, Dynamic Contrast Enhanced (DCE) MRI and CT scans targeting the pancreas, liver, kidneys and psoas-muscle and achieves quantitative measures of mean Dice similarity coefficient (DSC) that surpass or are comparable with the state-of-the-art. A qualitative evaluation performed by two independent radiologists verified the preservation of detailed organ and muscle boundaries.
Collapse
Affiliation(s)
| | - Hykoush Asaturyan
- School of Computer Science, University of Westminster, London, United Kingdom
| | - Sila Kurugol
- Department of Radiology, Boston Children’s Hospital & Harvard Medical School, Boston, Massachusetts, USA
| | - Onur Afacan
- Department of Radiology Boston Children’s Hospital & Harvard Medical School, Boston, Massachusetts, USA
| | - Jimmy D. Bell
- School of Life Sciences, University of Westminster, London, United Kingdom
| | - E. Louise Thomas
- School of Life Sciences, University of Westminster, London, United Kingdom
| |
Collapse
|
206
|
Barat M, Chassagnon G, Dohan A, Gaujoux S, Coriat R, Hoeffel C, Cassinotto C, Soyer P. Artificial intelligence: a critical review of current applications in pancreatic imaging. Jpn J Radiol 2021; 39:514-523. [PMID: 33550513 DOI: 10.1007/s11604-021-01098-5] [Citation(s) in RCA: 43] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2021] [Accepted: 01/25/2021] [Indexed: 12/11/2022]
Abstract
The applications of artificial intelligence (AI), including machine learning and deep learning, in the field of pancreatic disease imaging are rapidly expanding. AI can be used for the detection of pancreatic ductal adenocarcinoma and other pancreatic tumors but also for pancreatic lesion characterization. In this review, the basic of radiomics, recent developments and current results of AI in the field of pancreatic tumors are presented. Limitations and future perspectives of AI are discussed.
Collapse
Affiliation(s)
- Maxime Barat
- Department of Radiology, Hopital Cochin, Assistance Publique-Hopitaux de Paris, 27 Rue du Faubourg Saint-Jacques, Paris, France
- Université de Paris, Descartes-Paris 5, 75006, Paris, France
| | - Guillaume Chassagnon
- Department of Radiology, Hopital Cochin, Assistance Publique-Hopitaux de Paris, 27 Rue du Faubourg Saint-Jacques, Paris, France
- Université de Paris, Descartes-Paris 5, 75006, Paris, France
| | - Anthony Dohan
- Department of Radiology, Hopital Cochin, Assistance Publique-Hopitaux de Paris, 27 Rue du Faubourg Saint-Jacques, Paris, France
- Université de Paris, Descartes-Paris 5, 75006, Paris, France
| | - Sébastien Gaujoux
- Université de Paris, Descartes-Paris 5, 75006, Paris, France
- Department of Abdominal Surgery, Hopital Cochin, Assistance Publique-Hopitaux de Paris, 75014, Paris, France
| | - Romain Coriat
- Université de Paris, Descartes-Paris 5, 75006, Paris, France
- Department of Gastroenterology, Hopital Cochin, Assistance Publique-Hopitaux de Paris, 75014, Paris, France
| | - Christine Hoeffel
- Department of Radiology, Robert Debré Hospital, 51092, Reims, France
| | - Christophe Cassinotto
- Department of Radiology, CHU Montpellier, University of Montpellier, Saint-Éloi Hospital, 34000, Montpellier, France
| | - Philippe Soyer
- Department of Radiology, Hopital Cochin, Assistance Publique-Hopitaux de Paris, 27 Rue du Faubourg Saint-Jacques, Paris, France.
- Université de Paris, Descartes-Paris 5, 75006, Paris, France.
| |
Collapse
|
207
|
Li H, Liu B, Zhang Y, Fu C, Han X, Du L, Gao W, Chen Y, Liu X, Wang Y, Wang T, Ma G, Lei B. 3D IFPN: Improved Feature Pyramid Network for Automatic Segmentation of Gastric Tumor. Front Oncol 2021; 11:618496. [PMID: 34094903 PMCID: PMC8173118 DOI: 10.3389/fonc.2021.618496] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2020] [Accepted: 04/21/2021] [Indexed: 11/24/2022] Open
Abstract
Automatic segmentation of gastric tumor not only provides image-guided clinical diagnosis but also assists radiologists to read images and improve the diagnostic accuracy. However, due to the inhomogeneous intensity distribution of gastric tumors in CT scans, the ambiguous/missing boundaries, and the highly variable shapes of gastric tumors, it is quite challenging to develop an automatic solution. This study designs a novel 3D improved feature pyramidal network (3D IFPN) to automatically segment gastric tumors in computed tomography (CT) images. To meet the challenges of this extremely difficult task, the proposed 3D IFPN makes full use of the complementary information within the low and high layers of deep convolutional neural networks, which is equipped with three types of feature enhancement modules: 3D adaptive spatial feature fusion (ASFF) module, single-level feature refinement (SLFR) module, and multi-level feature refinement (MLFR) module. The 3D ASFF module adaptively suppresses the feature inconsistency in different levels and hence obtains the multi-level features with high feature invariance. Then, the SLFR module combines the adaptive features and previous multi-level features at each level to generate the multi-level refined features by skip connection and attention mechanism. The MLFR module adaptively recalibrates the channel-wise and spatial-wise responses by adding the attention operation, which improves the prediction capability of the network. Furthermore, a stage-wise deep supervision (SDS) mechanism and a hybrid loss function are also embedded to enhance the feature learning ability of the network. CT volumes dataset collected in three Chinese medical centers was used to evaluate the segmentation performance of the proposed 3D IFPN model. Experimental results indicate that our method outperforms state-of-the-art segmentation networks in gastric tumor segmentation. Moreover, to explore the generalization for other segmentation tasks, we also extend the proposed network to liver tumor segmentation in CT images of the MICCAI 2017 Liver Tumor Segmentation Challenge.
Collapse
Affiliation(s)
- Haimei Li
- Department of Radiology, Fuxing Hospital, Capital Medical University, Beijing, China
| | - Bing Liu
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
- Graduate School of Peking Union Medical College, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Yongtao Zhang
- School of Biomedical Engineering, Shenzhen University, National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, Shenzhen, China
| | - Chao Fu
- Department of Radiology, Dongzhimen Hospital, Beijing University of Chinese Medicine, Beijing, China
| | - Xiaowei Han
- Department of Radiology, The Affiliated Drum Tower Hospital of Nanjing University Medical School, Nanjing, China
| | - Lei Du
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Wenwen Gao
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Yue Chen
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Xiuxiu Liu
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Yige Wang
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Tianfu Wang
- School of Biomedical Engineering, Shenzhen University, National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, Shenzhen, China
| | - Guolin Ma
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Baiying Lei
- School of Biomedical Engineering, Shenzhen University, National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, Shenzhen, China
| |
Collapse
|
208
|
Petit O, Thome N, Soler L. Iterative confidence relabeling with deep ConvNets for organ segmentation with partial labels. Comput Med Imaging Graph 2021; 91:101938. [PMID: 34153879 DOI: 10.1016/j.compmedimag.2021.101938] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2020] [Revised: 03/22/2021] [Accepted: 04/27/2021] [Indexed: 11/16/2022]
Abstract
Training deep ConvNets requires large labeled datasets. However, collecting pixel-level labels for medical image segmentation is very expensive and requires a high level of expertise. In addition, most existing segmentation masks provided by clinical experts focus on specific anatomical structures. In this paper, we propose a method dedicated to handle such partially labeled medical image datasets. We propose a strategy to identify pixels for which labels are correct, and to train Fully Convolutional Neural Networks with a multi-label loss adapted to this context. In addition, we introduce an iterative confidence self-training approach inspired by curriculum learning to relabel missing pixel labels, which relies on selecting the most confident prediction with a specifically designed confidence network that learns an uncertainty measure which is leveraged in our relabeling process. Our approach, INERRANT for Iterative coNfidencE Relabeling of paRtial ANnoTations, is thoroughly evaluated on two public datasets (TCAI and LITS), and one internal dataset with seven abdominal organ classes. We show that INERRANT robustly deals with partial labels, performing similarly to a model trained on all labels even for large missing label proportions. We also highlight the importance of our iterative learning scheme and the proposed confidence measure for optimal performance. Finally we show a practical use case where a limited number of completely labeled data are enriched by publicly available but partially labeled data.
Collapse
Affiliation(s)
- Olivier Petit
- CEDRIC, Conservatoire National des Arts et Metiers, 292 rue Saint-Martin, Paris, 75003, France; Visible Patient, 8 rue Gustave Adolphe Hirn, Strasbourg, 67000, France.
| | - Nicolas Thome
- CEDRIC, Conservatoire National des Arts et Metiers, 292 rue Saint-Martin, Paris, 75003, France
| | - Luc Soler
- Visible Patient, 8 rue Gustave Adolphe Hirn, Strasbourg, 67000, France
| |
Collapse
|
209
|
Nai YH, Teo BW, Tan NL, O'Doherty S, Stephenson MC, Thian YL, Chiong E, Reilhac A. Comparison of metrics for the evaluation of medical segmentations using prostate MRI dataset. Comput Biol Med 2021; 134:104497. [PMID: 34022486 DOI: 10.1016/j.compbiomed.2021.104497] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2021] [Revised: 05/11/2021] [Accepted: 05/11/2021] [Indexed: 10/21/2022]
Abstract
Nine previously proposed segmentation evaluation metrics, targeting medical relevance, accounting for holes, and added regions or differentiating over- and under-segmentation, were compared with 24 traditional metrics to identify those which better capture the requirements for clinical segmentation evaluation. Evaluation was first performed using 2D synthetic shapes to highlight features and pitfalls of the metrics with known ground truths (GTs) and machine segmentations (MSs). Clinical evaluation was then performed using publicly-available prostate images of 20 subjects with MSs generated by 3 different deep learning networks (DenseVNet, HighRes3DNet, and ScaleNet) and GTs drawn by 2 readers. The same readers also performed the 2D visual assessment of the MSs using a dual negative-positive grading of -5 to 5 to reflect over- and under-estimation. Nine metrics that correlated well with visual assessment were selected for further evaluation using 3 different network ranking methods - based on a single metric, normalizing the metric using 2 GTs, and ranking the network based on a metric then averaging, including leave-one-out evaluation. These metrics yielded consistent ranking with HighRes3DNet ranked first then DenseVNet and ScaleNet using all ranking methods. Relative volume difference yielded the best positivity-agreement and correlation with dual visual assessment, and thus is better for providing over- and under-estimation. Interclass Correlation yielded the strongest correlation with the absolute visual assessment (0-5). Symmetric-boundary dice consistently yielded good discrimination of the networks for all three ranking methods with relatively small variations within network. Good rank discrimination may be an additional metric feature required for better network performance evaluation.
Collapse
Affiliation(s)
- Ying-Hwey Nai
- Clinical Imaging Research Centre, Yong Loo Lin School of Medicine, National University of Singapore, Singapore.
| | | | - Nadya L Tan
- St. Joseph's Institution International, Singapore
| | - Sophie O'Doherty
- Clinical Imaging Research Centre, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | - Mary C Stephenson
- Clinical Imaging Research Centre, Yong Loo Lin School of Medicine, National University of Singapore, Singapore; Centre for Translational MR Research, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | - Yee Liang Thian
- Department of Diagnostic Imaging, National University Hospital, Singapore
| | - Edmund Chiong
- Department of Surgery, Yong Loo Lin School of Medicine, National University of Singapore, Singapore; Department of Urology, National University Hospital, Singapore
| | - Anthonin Reilhac
- Clinical Imaging Research Centre, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| |
Collapse
|
210
|
Conze PH, Kavur AE, Cornec-Le Gall E, Gezer NS, Le Meur Y, Selver MA, Rousseau F. Abdominal multi-organ segmentation with cascaded convolutional and adversarial deep networks. Artif Intell Med 2021; 117:102109. [PMID: 34127239 DOI: 10.1016/j.artmed.2021.102109] [Citation(s) in RCA: 40] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2020] [Revised: 01/24/2021] [Accepted: 05/06/2021] [Indexed: 02/05/2023]
Abstract
Abdominal anatomy segmentation is crucial for numerous applications from computer-assisted diagnosis to image-guided surgery. In this context, we address fully-automated multi-organ segmentation from abdominal CT and MR images using deep learning. The proposed model extends standard conditional generative adversarial networks. Additionally to the discriminator which enforces the model to create realistic organ delineations, it embeds cascaded partially pre-trained convolutional encoder-decoders as generator. Encoder fine-tuning from a large amount of non-medical images alleviates data scarcity limitations. The network is trained end-to-end to benefit from simultaneous multi-level segmentation refinements using auto-context. Employed for healthy liver, kidneys and spleen segmentation, our pipeline provides promising results by outperforming state-of-the-art encoder-decoder schemes. Followed for the Combined Healthy Abdominal Organ Segmentation (CHAOS) challenge organized in conjunction with the IEEE International Symposium on Biomedical Imaging 2019, it gave us the first rank for three competition categories: liver CT, liver MR and multi-organ MR segmentation. Combining cascaded convolutional and adversarial networks strengthens the ability of deep learning pipelines to automatically delineate multiple abdominal organs, with good generalization capability. The comprehensive evaluation provided suggests that better guidance could be achieved to help clinicians in abdominal image interpretation and clinical decision making.
Collapse
Affiliation(s)
- Pierre-Henri Conze
- IMT Atlantique, Technopôle Brest-Iroise, 29238 Brest, France; LaTIM UMR 1101, Inserm, 22 avenue Camille Desmoulins, 29238 Brest, France.
| | - Ali Emre Kavur
- Dokuz Eylul University, Cumhuriyet Bulvarı, 35210 Izmir, Turkey
| | - Emilie Cornec-Le Gall
- Department of Nephrology, University Hospital, 2 avenue Foch, 29609 Brest, France; UMR 1078, Inserm, 22 avenue Camille Desmoulins, 29238 Brest, France
| | - Naciye Sinem Gezer
- Dokuz Eylul University, Cumhuriyet Bulvarı, 35210 Izmir, Turkey; Department of Radiology, Faculty of Medicine, Cumhuriyet Bulvarı, 35210 Izmir, Turkey
| | - Yannick Le Meur
- Department of Nephrology, University Hospital, 2 avenue Foch, 29609 Brest, France; LBAI UMR 1227, Inserm, 5 avenue Foch, 29609 Brest, France
| | - M Alper Selver
- Dokuz Eylul University, Cumhuriyet Bulvarı, 35210 Izmir, Turkey
| | - François Rousseau
- IMT Atlantique, Technopôle Brest-Iroise, 29238 Brest, France; LaTIM UMR 1101, Inserm, 22 avenue Camille Desmoulins, 29238 Brest, France
| |
Collapse
|
211
|
Chen X, Sun S, Bai N, Han K, Liu Q, Yao S, Tang H, Zhang C, Lu Z, Huang Q, Zhao G, Xu Y, Chen T, Xie X, Liu Y. A deep learning-based auto-segmentation system for organs-at-risk on whole-body computed tomography images for radiation therapy. Radiother Oncol 2021; 160:175-184. [PMID: 33961914 DOI: 10.1016/j.radonc.2021.04.019] [Citation(s) in RCA: 69] [Impact Index Per Article: 17.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2020] [Revised: 04/23/2021] [Accepted: 04/24/2021] [Indexed: 12/25/2022]
Abstract
BACKGROUND AND PURPOSE Delineating organs at risk (OARs) on computed tomography (CT) images is an essential step in radiation therapy; however, it is notoriously time-consuming and prone to inter-observer variation. Herein, we report a deep learning-based automatic segmentation (AS) algorithm (WBNet) that can accurately and efficiently delineate all major OARs in the entire body directly on CT scans. MATERIALS AND METHODS We collected 755 CT scans of the head and neck, thorax, abdomen, and pelvis and manually delineated 50 OARs on the CT images. The CT images with contours were split into training and test sets consisting of 505 and 250 cases, respectively, to develop and validate WBNet. The volumetric Dice similarity coefficient (DSC) and 95th-percentile Hausdorff distance (95% HD) were calculated to evaluate delineation quality for each OAR. We compared the performance of WBNet with three AS algorithms: one commercial multi-atlas-based automatic segmentation (ABAS) software, and two deep learning-based AS algorithms, namely, AnatomyNet and nnU-Net. We have also evaluated the time saving and dose accuracy of WBNet. RESULTS WBNet achieved average DSCs of 0.84 and 0.81 on in-house and public datasets, respectively, which outperformed ABAS, AnatomyNet, and nnU-Net. WBNet could reduce the delineation time significantly and perform well in treatment planning, with clinically acceptable dose differences compared with those in manual delineation. CONCLUSION This study shows the feasibility and benefits of using WBNet in clinical practice.
Collapse
Affiliation(s)
- Xuming Chen
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Shanlin Sun
- DeepVoxel Inc., Irvine, USA; Department of Computer Science, University of California, Irvine, USA
| | | | | | - Qianqian Liu
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Shengyu Yao
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Hao Tang
- Department of Computer Science, University of California, Irvine, USA
| | | | | | - Qian Huang
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Guoqi Zhao
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Yi Xu
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Tingfeng Chen
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Xiaohui Xie
- Department of Computer Science, University of California, Irvine, USA.
| | - Yong Liu
- Department of Radiation Oncology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
| |
Collapse
|
212
|
Humpire-Mamani GE, Bukala J, Scholten ET, Prokop M, van Ginneken B, Jacobs C. Fully Automatic Volume Measurement of the Spleen at CT Using Deep Learning. Radiol Artif Intell 2021; 2:e190102. [PMID: 33937830 DOI: 10.1148/ryai.2020190102] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2019] [Revised: 04/26/2020] [Accepted: 05/01/2020] [Indexed: 12/15/2022]
Abstract
Purpose To develop a fully automated algorithm for spleen segmentation and to assess the performance of this algorithm in a large dataset. Materials and Methods In this retrospective study, a three-dimensional deep learning network was developed to segment the spleen on thorax-abdomen CT scans. Scans were extracted from patients undergoing oncologic treatment from 2014 to 2017. A total of 1100 scans from 1100 patients were used in this study, and 400 were selected for development of the algorithm. For testing, a dataset of 50 scans was annotated to assess the segmentation accuracy and was compared against the splenic index equation. In a qualitative observer experiment, an enriched set of 100 scan-pairs was used to evaluate whether the algorithm could aid a radiologist in assessing splenic volume change. The reference standard was set by the consensus of two other independent radiologists. A Mann-Whitney U test was conducted to test whether there was a performance difference between the algorithm and the independent observer. Results The algorithm and the independent observer obtained comparable Dice scores (P = .834) on the test set of 50 scans of 0.962 and 0.964, respectively. The radiologist had an agreement with the reference standard in 81% (81 of 100) of the cases after a visual classification of volume change, which increased to 92% (92 of 100) when aided by the algorithm. Conclusion A segmentation method based on deep learning can accurately segment the spleen on CT scans and may help radiologists to detect abnormal splenic volumes and splenic volume changes.Supplemental material is available for this article.© RSNA, 2020.
Collapse
Affiliation(s)
- Gabriel E Humpire-Mamani
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10 (Route 767), 6525 GA, Nijmegen, the Netherlands (G.E.H.M., J.B., E.T.S., M.P., B.v.G., C.J.); and Fraunhofer MEVIS, Bremen, Germany (B.v.G.)
| | - Joris Bukala
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10 (Route 767), 6525 GA, Nijmegen, the Netherlands (G.E.H.M., J.B., E.T.S., M.P., B.v.G., C.J.); and Fraunhofer MEVIS, Bremen, Germany (B.v.G.)
| | - Ernst T Scholten
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10 (Route 767), 6525 GA, Nijmegen, the Netherlands (G.E.H.M., J.B., E.T.S., M.P., B.v.G., C.J.); and Fraunhofer MEVIS, Bremen, Germany (B.v.G.)
| | - Mathias Prokop
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10 (Route 767), 6525 GA, Nijmegen, the Netherlands (G.E.H.M., J.B., E.T.S., M.P., B.v.G., C.J.); and Fraunhofer MEVIS, Bremen, Germany (B.v.G.)
| | - Bram van Ginneken
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10 (Route 767), 6525 GA, Nijmegen, the Netherlands (G.E.H.M., J.B., E.T.S., M.P., B.v.G., C.J.); and Fraunhofer MEVIS, Bremen, Germany (B.v.G.)
| | - Colin Jacobs
- Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Grooteplein 10 (Route 767), 6525 GA, Nijmegen, the Netherlands (G.E.H.M., J.B., E.T.S., M.P., B.v.G., C.J.); and Fraunhofer MEVIS, Bremen, Germany (B.v.G.)
| |
Collapse
|
213
|
Fu Y, Lei Y, Wang T, Curran WJ, Liu T, Yang X. A review of deep learning based methods for medical image multi-organ segmentation. Phys Med 2021; 85:107-122. [PMID: 33992856 PMCID: PMC8217246 DOI: 10.1016/j.ejmp.2021.05.003] [Citation(s) in RCA: 89] [Impact Index Per Article: 22.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/27/2020] [Revised: 03/12/2021] [Accepted: 05/03/2021] [Indexed: 12/12/2022] Open
Abstract
Deep learning has revolutionized image processing and achieved the-state-of-art performance in many medical image segmentation tasks. Many deep learning-based methods have been published to segment different parts of the body for different medical applications. It is necessary to summarize the current state of development for deep learning in the field of medical image segmentation. In this paper, we aim to provide a comprehensive review with a focus on multi-organ image segmentation, which is crucial for radiotherapy where the tumor and organs-at-risk need to be contoured for treatment planning. We grouped the surveyed methods into two broad categories which are 'pixel-wise classification' and 'end-to-end segmentation'. Each category was divided into subgroups according to their network design. For each type, we listed the surveyed works, highlighted important contributions and identified specific challenges. Following the detailed review, we discussed the achievements, shortcomings and future potentials of each category. To enable direct comparison, we listed the performance of the surveyed works that used thoracic and head-and-neck benchmark datasets.
Collapse
Affiliation(s)
- Yabo Fu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Yang Lei
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Tonghe Wang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Walter J Curran
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Tian Liu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, USA.
| |
Collapse
|
214
|
Wang X, Zheng F, Xiao R, Liu Z, Li Y, Li J, Zhang X, Hao X, Zhang X, Guo J, Zhang Y, Xue H, Jin Z. Comparison of image quality and lesion diagnosis in abdominopelvic unenhanced CT between reduced-dose CT using deep learning post-processing and standard-dose CT using iterative reconstruction: A prospective study. Eur J Radiol 2021; 139:109735. [PMID: 33932717 DOI: 10.1016/j.ejrad.2021.109735] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2021] [Revised: 04/06/2021] [Accepted: 04/19/2021] [Indexed: 11/24/2022]
Abstract
PURPOSE To compare image quality and lesion diagnosis between reduced-dose abdominopelvic unenhanced computed tomography (CT) using deep learning (DL) post-processing and standard-dose CT using iterative reconstruction (IR). METHOD Totally 251 patients underwent two consecutive abdominopelvic unenhanced CT scans of the same range, including standard and reduced doses, respectively. In group A, standard-dose data were reconstructed by (blend 30 %) IR. In group B, reduced-dose data were reconstructed by filtered back projection reconstruction to obtain group B1 images, and post-processed using the DL algorithm (NeuAI denosing, Neusoft medical, Shenyang, China) with 50 % and 100 % weights to obtain group B2 and B3 images, respectively. Then, CT values of the liver, the second lumbar vertebral centrum, the erector spinae and abdominal subcutaneous fat were measured. CT values, noise levels, signal-to-noise ratios (SNRs), contrast-to-noise ratios (CNRs), radiation doses and subjective scores of image quality were compared. Subjective evaluations of low-density liver lesions were compared by diagnostic results from enhanced CT or Magnetic Resonance Imaging. RESULTS Groups B3 and B1 showed the lowest and highest noise levels, respectively (P < 0.001). The SNR and CNR in group B3 were highest (P < 0.001). The radiation dose in group B was reduced by 71.5 % on average compared to group A. Subjective scores in groups A and B2 were highest (P < 0.001). Diagnostic sensitivity and confidence for liver metastases in groups A and B2 were highest (P < 0.001). CONCLUSIONS Reduced-dose abdominopelvic unenhanced CT combined with DL post-processing could ensure image quality and satisfy diagnostic needs.
Collapse
Affiliation(s)
- Xiao Wang
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Fuling Zheng
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Ran Xiao
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Zhuoheng Liu
- From CT Business Unit, Neusoft Medical System Company, Shenyang, China
| | - Yutong Li
- From CT Business Unit, Neusoft Medical System Company, Shenyang, China
| | - Juan Li
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Xi Zhang
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Xuemin Hao
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Xinhu Zhang
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Jiawu Guo
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Yan Zhang
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China
| | - Huadan Xue
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China.
| | - Zhengyu Jin
- From the Department of Radiology, State Key Laboratory of Complex Severe and Rare Diseases, Peking Union Medical College Hospital, Chinese Academy of Medical Science and Peking Union Medical College, Beijing, China.
| |
Collapse
|
215
|
Zhang C, Hua Q, Chu Y, Wang P. Liver tumor segmentation using 2.5D UV-Net with multi-scale convolution. Comput Biol Med 2021; 133:104424. [PMID: 33984683 DOI: 10.1016/j.compbiomed.2021.104424] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2021] [Revised: 04/13/2021] [Accepted: 04/18/2021] [Indexed: 12/24/2022]
Abstract
Liver tumor segmentation networks are generally based on U-shaped encoder-decoder network with 2D or 3D structure. However, 2D networks lose the inter-layer information of continuous slices and 3D networks might introduce unacceptable parameters for GPU memory. As a result, 2.5D networks were proposed to balance the memory consumption and 3D context. Different from the canonical 2.5D design, which utilizes a 2D network combined with RNN, we propose a new 2.5D design called UV-Net to encode the inter-layer information in the context of 3D convolution, and reconstruct the high-resolution results with 2D deconvolution. At the same time, the multi-scale convolution structure enables multi-scale feature extraction without extra computational cost, which effectively mines structured information, reduces information redundancy, strengthens independent features, and makes feature dimension sparse, to enhance network capacity and efficiency. Combined with the proposed preprocessing method of removing mean energy, UV-Net significantly outperforms the existing methods in liver tumor segmentation and especially improves the segmentation accuracy of small objects on the LiTS2017 dataset.
Collapse
Affiliation(s)
- Chi Zhang
- School of Information Science and Engineering, Shandong University, Qingdao, China.
| | - Qianqian Hua
- Department of Medical Imaging, ShanDong Provincial Hospital, Jinan, China
| | - Yingying Chu
- School of Information Science and Engineering, Shandong University, Qingdao, China
| | - Pengwei Wang
- School of Information Science and Engineering, Shandong University, Qingdao, China.
| |
Collapse
|
216
|
Tian F, Gao Y, Fang Z, Gu J. Automatic coronary artery segmentation algorithm based on deep learning and digital image processing. APPL INTELL 2021. [DOI: 10.1007/s10489-021-02197-6] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/13/2023]
|
217
|
Deep learning to segment pelvic bones: large-scale CT datasets and baseline models. Int J Comput Assist Radiol Surg 2021; 16:749-756. [PMID: 33864189 DOI: 10.1007/s11548-021-02363-8] [Citation(s) in RCA: 28] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2021] [Accepted: 04/01/2021] [Indexed: 10/21/2022]
Abstract
PURPOSE Pelvic bone segmentation in CT has always been an essential step in clinical diagnosis and surgery planning of pelvic bone diseases. Existing methods for pelvic bone segmentation are either hand-crafted or semi-automatic and achieve limited accuracy when dealing with image appearance variations due to the multi-site domain shift, the presence of contrasted vessels, coprolith and chyme, bone fractures, low dose, metal artifacts, etc. Due to the lack of a large-scale pelvic CT dataset with annotations, deep learning methods are not fully explored. METHODS In this paper, we aim to bridge the data gap by curating a large pelvic CT dataset pooled from multiple sources, including 1184 CT volumes with a variety of appearance variations. Then, we propose for the first time, to the best of our knowledge, to learn a deep multi-class network for segmenting lumbar spine, sacrum, left hip, and right hip, from multiple-domain images simultaneously to obtain more effective and robust feature representations. Finally, we introduce a post-processor based on the signed distance function (SDF). RESULTS Extensive experiments on our dataset demonstrate the effectiveness of our automatic method, achieving an average Dice of 0.987 for a metal-free volume. SDF post-processor yields a decrease of 15.1% in Hausdorff distance compared with traditional post-processor. CONCLUSION We believe this large-scale dataset will promote the development of the whole community and open source the images, annotations, codes, and trained baseline models at https://github.com/ICT-MIRACLE-lab/CTPelvic1K .
Collapse
|
218
|
Luu HM, van Walsum T, Franklin D, Pham PC, Vu LD, Moelker A, Staring M, VanHoang X, Niessen W, Trung NL. Efficiently compressing 3D medical images for teleinterventions via CNNs and anisotropic diffusion. Med Phys 2021; 48:2877-2890. [PMID: 33656213 DOI: 10.1002/mp.14814] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2020] [Revised: 01/29/2021] [Accepted: 02/14/2021] [Indexed: 12/23/2022] Open
Abstract
PURPOSE Efficient compression of images while preserving image quality has the potential to be a major enabler of effective remote clinical diagnosis and treatment, since poor Internet connection conditions are often the primary constraint in such services. This paper presents a framework for organ-specific image compression for teleinterventions based on a deep learning approach and anisotropic diffusion filter. METHODS The proposed method, deep learning and anisotropic diffusion (DLAD), uses a convolutional neural network architecture to extract a probability map for the organ of interest; this probability map guides an anisotropic diffusion filter that smooths the image except at the location of the organ of interest. Subsequently, a compression method, such as BZ2 and HEVC-visually lossless, is applied to compress the image. We demonstrate the proposed method on three-dimensional (3D) CT images acquired for radio frequency ablation (RFA) of liver lesions. We quantitatively evaluate the proposed method on 151 CT images using peak-signal-to-noise ratio ( PSNR ), structural similarity ( SSIM ), and compression ratio ( CR ) metrics. Finally, we compare the assessments of two radiologists on the liver lesion detection and the liver lesion center annotation using 33 sets of the original images and the compressed images. RESULTS The results show that the method can significantly improve CR of most well-known compression methods. DLAD combined with HEVC-visually lossless achieves the highest average CR of 6.45, which is 36% higher than that of the original HEVC and outperforms other state-of-the-art lossless medical image compression methods. The means of PSNR and SSIM are 70 dB and 0.95, respectively. In addition, the compression effects do not statistically significantly affect the assessments of the radiologists on the liver lesion detection and the lesion center annotation. CONCLUSIONS We thus conclude that the method has a high potential to be applied in teleintervention applications.
Collapse
Affiliation(s)
- Ha Manh Luu
- AVITECH, University of Engineering and Technology, VNU, Hanoi, Vietnam.,Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands.,FET, University of Engineering and Technology, VNU, Hanoi, Vietnam
| | - Theo van Walsum
- Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands
| | - Daniel Franklin
- School of Electrical and Data Engineering, University of Technology Sydney, Sydney, Australia
| | - Phuong Cam Pham
- Nuclear Medicine and Oncology Center, Bach Mai Hospital, Hanoi, Vietnam
| | - Luu Dang Vu
- Radiology Center, Bach Mai Hospital, Hanoi, Vietnam
| | - Adriaan Moelker
- Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands
| | - Marius Staring
- Department of Radiology, Leiden University Medical Center, Leiden, the Netherlands
| | - Xiem VanHoang
- FET, University of Engineering and Technology, VNU, Hanoi, Vietnam
| | - Wiro Niessen
- Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands
| | - Nguyen Linh Trung
- AVITECH, University of Engineering and Technology, VNU, Hanoi, Vietnam
| |
Collapse
|
219
|
Almeida DF, Astudillo P, Vandermeulen D. Three-dimensional image volumes from two-dimensional digitally reconstructed radiographs: A deep learning approach in lower limb CT scans. Med Phys 2021; 48:2448-2457. [PMID: 33690903 DOI: 10.1002/mp.14835] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2020] [Revised: 03/01/2021] [Accepted: 03/03/2021] [Indexed: 11/07/2022] Open
Abstract
PURPOSE Three-dimensional (3D) reconstructions of the human anatomy have been available for surgery planning or diagnostic purposes for a few years now. The different image modalities usually rely on several consecutive two-dimensional (2D) acquisitions in order to reconstruct the 3D volume. Hence, such acquisitions are expensive, time-demanding and often expose the patient to an undesirable amount of radiation. For such reasons, along the most recent years, several studies have been proposed that extrapolate 3D anatomical features from merely 2D exams such as x rays for implant templating in total knee or hip arthroplasties. METHOD The presented study shows an adaptation of a deep learning-based convolutional neural network to reconstruct 3D volumes from a mere 2D digitally reconstructed radiograph from one of the most extensive lower limb computed tomography datasets available. This novel approach is based on an encoder-decoder architecture with skip connections and a multidimensional Gaussian filter as data augmentation technique. RESULTS The results achieved promising values when compared against the ground truth volumes, quantitatively yielding an average of 0.77 ± 0.05 structured similarity index. CONCLUSIONS A novel deep learning-based approach to reconstruct 3D medical image volumes from a single x-ray image was shown in the present study. The network architecture was validated against the original scans presenting SSIM values of 0.77 ± 0.05 and 0.78 ± 0.06, respectively for the knee and the hip crop.
Collapse
Affiliation(s)
- Diogo F Almeida
- Medical Imaging Research Center (MIRC), Department of Electrical Engineering, KU Leuven, Herestraat 49, 3000, Leuven, Belgium
| | - Patricio Astudillo
- Department of Electronics and information systems, UGent - imec, Technologiepark 126, Zwijnaarde, 9052, Belgium
| | - Dirk Vandermeulen
- Medical Imaging Research Center (MIRC), Department of Electrical Engineering, KU Leuven, Herestraat 49, 3000, Leuven, Belgium
| |
Collapse
|
220
|
Santhosh Reddy D, Rajalakshmi P, Mateen M. A deep learning based approach for classification of abdominal organs using ultrasound images. Biocybern Biomed Eng 2021. [DOI: 10.1016/j.bbe.2021.05.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
221
|
Ma J, Chen J, Ng M, Huang R, Li Y, Li C, Yang X, Martel AL. Loss odyssey in medical image segmentation. Med Image Anal 2021; 71:102035. [PMID: 33813286 DOI: 10.1016/j.media.2021.102035] [Citation(s) in RCA: 136] [Impact Index Per Article: 34.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2020] [Revised: 03/04/2021] [Accepted: 03/06/2021] [Indexed: 12/26/2022]
Abstract
The loss function is an important component in deep learning-based segmentation methods. Over the past five years, many loss functions have been proposed for various segmentation tasks. However, a systematic study of the utility of these loss functions is missing. In this paper, we present a comprehensive review of segmentation loss functions in an organized manner. We also conduct the first large-scale analysis of 20 general loss functions on four typical 3D segmentation tasks involving six public datasets from 10+ medical centers. The results show that none of the losses can consistently achieve the best performance on the four segmentation tasks, but compound loss functions (e.g. Dice with TopK loss, focal loss, Hausdorff distance loss, and boundary loss) are the most robust losses. Our code and segmentation results are publicly available and can serve as a loss function benchmark. We hope this work will also provide insights on new loss function development for the community.
Collapse
Affiliation(s)
- Jun Ma
- Department of Mathematics, Nanjing University of Science and Technology, Nanjing, China.
| | - Jianan Chen
- Department of Medical Biophysics, University of Toronto, Toronto, Canada
| | - Matthew Ng
- Department of Medical Biophysics, University of Toronto, Toronto, Canada
| | - Rui Huang
- Department of Medical Biophysics, University of Toronto, Toronto, Canada
| | - Yu Li
- Department of Mathematics, Nanjing University of Science and Technology, Nanjing, China
| | - Chen Li
- Department of Mathematics, Nanjing University, Nanjing, China
| | - Xiaoping Yang
- Department of Mathematics, Nanjing University, Nanjing, China
| | - Anne L Martel
- Department of Medical Biophysics, University of Toronto, Toronto, Canada; Physical Sciences, Sunnybrook Research Institute, Toronto, Canada
| |
Collapse
|
222
|
Cui Y, Arimura H, Nakano R, Yoshitake T, Shioyama Y, Yabuuchi H. Automated approach for segmenting gross tumor volumes for lung cancer stereotactic body radiation therapy using CT-based dense V-networks. JOURNAL OF RADIATION RESEARCH 2021; 62:346-355. [PMID: 33480438 PMCID: PMC7948852 DOI: 10.1093/jrr/rraa132] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/16/2020] [Revised: 09/12/2020] [Indexed: 06/12/2023]
Abstract
The aim of this study was to develop an automated segmentation approach for small gross tumor volumes (GTVs) in 3D planning computed tomography (CT) images using dense V-networks (DVNs) that offer more advantages in segmenting smaller structures than conventional V-networks. Regions of interest (ROI) with dimensions of 50 × 50 × 6-72 pixels in the planning CT images were cropped based on the GTV centroids when applying stereotactic body radiotherapy (SBRT) to patients. Segmentation accuracy of GTV contours for 192 lung cancer patients [with the following tumor types: 118 solid, 53 part-solid types and 21 pure ground-glass opacity (pure GGO)], who underwent SBRT, were evaluated based on a 10-fold cross-validation test using Dice's similarity coefficient (DSC) and Hausdorff distance (HD). For each case, 11 segmented GTVs consisting of three single outputs, four logical AND outputs, and four logical OR outputs from combinations of two or three outputs from DVNs were obtained by three runs with different initial weights. The AND output (combination of three outputs) achieved the highest values of average 3D-DSC (0.832 ± 0.074) and HD (4.57 ± 2.44 mm). The average 3D DSCs from the AND output for solid, part-solid and pure GGO types were 0.838 ± 0.074, 0.822 ± 0.078 and 0.819 ± 0.059, respectively. This study suggests that the proposed approach could be useful in segmenting GTVs for planning lung cancer SBRT.
Collapse
Affiliation(s)
- Yunhao Cui
- Department of Health Sciences, Graduate School of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan
| | - Hidetaka Arimura
- Department of Health Sciences, Faculty of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan
| | - Risa Nakano
- Department of Health Sciences, Graduate School of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan
| | - Tadamasa Yoshitake
- Department of Clinical Radiology, Graduate School of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan
| | - Yoshiyuki Shioyama
- Saga International Heavy Ion Cancer Treatment Foundation, 3049 Harakogamachi, Tosu-shi, Saga 841-0071, Japan
| | - Hidetake Yabuuchi
- Department of Health Sciences, Faculty of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan
| |
Collapse
|
223
|
Ju Z, Guo W, Gu S, Zhou J, Yang W, Cong X, Dai X, Quan H, Liu J, Qu B, Liu G. CT based automatic clinical target volume delineation using a dense-fully connected convolution network for cervical Cancer radiation therapy. BMC Cancer 2021; 21:243. [PMID: 33685404 PMCID: PMC7938586 DOI: 10.1186/s12885-020-07595-6] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2020] [Accepted: 10/30/2020] [Indexed: 02/06/2023] Open
Abstract
BACKGROUND It is very important to accurately delineate the CTV on the patient's three-dimensional CT image in the radiotherapy process. Limited to the scarcity of clinical samples and the difficulty of automatic delineation, the research of automatic delineation of cervical cancer CTV based on CT images for new patients is slow. This study aimed to assess the value of Dense-Fully Connected Convolution Network (Dense V-Net) in predicting Clinical Target Volume (CTV) pre-delineation in cervical cancer patients for radiotherapy. METHODS In this study, we used Dense V-Net, a dense and fully connected convolutional network with suitable feature learning in small samples to automatically pre-delineate the CTV of cervical cancer patients based on computed tomography (CT) images and then we assessed the outcome. The CT data of 133 patients with stage IB and IIA postoperative cervical cancer with a comparable delineation scope was enrolled in this study. One hundred and thirteen patients were randomly designated as the training set to adjust the model parameters. Twenty cases were used as the test set to assess the network performance. The 8 most representative parameters were also used to assess the pre-sketching accuracy from 3 aspects: sketching similarity, sketching offset, and sketching volume difference. RESULTS The results presented that the DSC, DC/mm, HD/cm, MAD/mm, ∆V, SI, IncI and JD of CTV were 0.82 ± 0.03, 4.28 ± 2.35, 1.86 ± 0.48, 2.52 ± 0.40, 0.09 ± 0.05, 0.84 ± 0.04, 0.80 ± 0.05, and 0.30 ± 0.04, respectively, and the results were greater than those with a single network. CONCLUSIONS Dense V-Net can correctly predict CTV pre-delineation of cervical cancer patients and can be applied in clinical practice after completing simple modifications.
Collapse
Affiliation(s)
- Zhongjian Ju
- Department of Radiation Oncology, The First Medical Center, People's Liberation Army General Hospital, No. 28 Fuxing Road, Haidian District, Beijing, 100853, China
| | - Wen Guo
- School of Physics Science and Technology, Wuhan University, No. 299, Bayi Road, Luojiashan Street, Wuhan, 430072, China
- Department of Magnetic Resonance Imaging, the First Affiliated Hospital of Zhengzhou University, No. 1 Jianshe East Road, Zhengzhou, 450003, China
| | - Shanshan Gu
- Department of Radiation Oncology, The First Medical Center, People's Liberation Army General Hospital, No. 28 Fuxing Road, Haidian District, Beijing, 100853, China
| | - Jin Zhou
- School of Physics Science and Technology, Wuhan University, No. 299, Bayi Road, Luojiashan Street, Wuhan, 430072, China
| | - Wei Yang
- Department of Radiation Oncology, The First Medical Center, People's Liberation Army General Hospital, No. 28 Fuxing Road, Haidian District, Beijing, 100853, China
| | - Xiaohu Cong
- Department of Radiation Oncology, The First Medical Center, People's Liberation Army General Hospital, No. 28 Fuxing Road, Haidian District, Beijing, 100853, China
| | - Xiangkun Dai
- Department of Radiation Oncology, The First Medical Center, People's Liberation Army General Hospital, No. 28 Fuxing Road, Haidian District, Beijing, 100853, China
| | - Hong Quan
- School of Physics Science and Technology, Wuhan University, No. 299, Bayi Road, Luojiashan Street, Wuhan, 430072, China
| | - Jie Liu
- Beijing Eastraycloud Technology Inc. Chengdu R&D Center.Suite, 1405-1406,Building Guannan Shangyu,NO.1,Xingguang Road,Wuhou District, Chengdu, 610094, China
| | - Baolin Qu
- Department of Radiation Oncology, The First Medical Center, People's Liberation Army General Hospital, No. 28 Fuxing Road, Haidian District, Beijing, 100853, China.
| | - Guocai Liu
- College of Electrical and Information Engineering, Hunan University, Changsha, 410082, China
| |
Collapse
|
224
|
Huang S, Han X, Fan J, Chen J, Du L, Gao W, Liu B, Chen Y, Liu X, Wang Y, Ai D, Ma G, Yang J. Anterior Mediastinal Lesion Segmentation Based on Two-Stage 3D ResUNet With Attention Gates and Lung Segmentation. Front Oncol 2021; 10:618357. [PMID: 33634027 PMCID: PMC7901488 DOI: 10.3389/fonc.2020.618357] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2020] [Accepted: 12/15/2020] [Indexed: 01/13/2023] Open
Abstract
OBJECTIVES Anterior mediastinal disease is a common disease in the chest. Computed tomography (CT), as an important imaging technology, is widely used in the diagnosis of mediastinal diseases. Doctors find it difficult to distinguish lesions in CT images because of image artifact, intensity inhomogeneity, and their similarity with other tissues. Direct segmentation of lesions can provide doctors a method to better subtract the features of the lesions, thereby improving the accuracy of diagnosis. METHOD As the trend of image processing technology, deep learning is more accurate in image segmentation than traditional methods. We employ a two-stage 3D ResUNet network combined with lung segmentation to segment CT images. Given that the mediastinum is between the two lungs, the original image is clipped through the lung mask to remove some noises that may affect the segmentation of the lesion. To capture the feature of the lesions, we design a two-stage network structure. In the first stage, the features of the lesion are learned from the low-resolution downsampled image, and the segmentation results under a rough scale are obtained. The results are concatenated with the original image and encoded into the second stage to capture more accurate segmentation information from the image. In addition, attention gates are introduced in the upsampling of the network, and these gates can focus on the lesion and play a role in filtering the features. The proposed method has achieved good results in the segmentation of the anterior mediastinal. RESULTS The proposed method was verified on 230 patients, and the anterior mediastinal lesions were well segmented. The average Dice coefficient reached 87.73%. Compared with the model without lung segmentation, the model with lung segmentation greatly improved the accuracy of lesion segmentation by approximately 9%. The addition of attention gates slightly improved the segmentation accuracy. CONCLUSION The proposed automatic segmentation method has achieved good results in clinical data. In clinical application, automatic segmentation of lesions can assist doctors in the diagnosis of diseases and may facilitate the automated diagnosis of illnesses in the future.
Collapse
Affiliation(s)
- Su Huang
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
| | - Xiaowei Han
- Department of Radiology, The Affiliated Drum Tower Hospital of Nanjing University Medical School, Nanjing, China
| | - Jingfan Fan
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
| | - Jing Chen
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
| | - Lei Du
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Wenwen Gao
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Bing Liu
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Yue Chen
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Xiuxiu Liu
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Yige Wang
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Danni Ai
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
| | - Guolin Ma
- Department of Radiology, China-Japan Friendship Hospital, Beijing, China
| | - Jian Yang
- Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
| |
Collapse
|
225
|
Shi G, Xiao L, Chen Y, Zhou SK. Marginal loss and exclusion loss for partially supervised multi-organ segmentation. Med Image Anal 2021; 70:101979. [PMID: 33636451 DOI: 10.1016/j.media.2021.101979] [Citation(s) in RCA: 39] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2020] [Revised: 12/02/2020] [Accepted: 01/20/2021] [Indexed: 11/25/2022]
Abstract
Annotating multiple organs in medical images is both costly and time-consuming; therefore, existing multi-organ datasets with labels are often low in sample size and mostly partially labeled, that is, a dataset has a few organs labeled but not all organs. In this paper, we investigate how to learn a single multi-organ segmentation network from a union of such datasets. To this end, we propose two types of novel loss function, particularly designed for this scenario: (i) marginal loss and (ii) exclusion loss. Because the background label for a partially labeled image is, in fact, a 'merged' label of all unlabelled organs and 'true' background (in the sense of full labels), the probability of this 'merged' background label is a marginal probability, summing the relevant probabilities before merging. This marginal probability can be plugged into any existing loss function (such as cross entropy loss, Dice loss, etc.) to form a marginal loss. Leveraging the fact that the organs are non-overlapping, we propose the exclusion loss to gauge the dissimilarity between labeled organs and the estimated segmentation of unlabelled organs. Experiments on a union of five benchmark datasets in multi-organ segmentation of liver, spleen, left and right kidneys, and pancreas demonstrate that using our newly proposed loss functions brings a conspicuous performance improvement for state-of-the-art methods without introducing any extra computation.
Collapse
Affiliation(s)
- Gonglei Shi
- Medical Imaging, Robotics, Analytic Computing Laboratory & Engineering (MIRACLE), Key Lab of Intelligent Information Processing of Chinese Academy of Sciences, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, 100190, China; School of Computer Science and Engineering, Southeast University, Nanjing, 210000, China
| | - Li Xiao
- Medical Imaging, Robotics, Analytic Computing Laboratory & Engineering (MIRACLE), Key Lab of Intelligent Information Processing of Chinese Academy of Sciences, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, 100190, China.
| | - Yang Chen
- School of Computer Science and Engineering, Southeast University, Nanjing, 210000, China
| | - S Kevin Zhou
- Medical Imaging, Robotics, Analytic Computing Laboratory & Engineering (MIRACLE), Key Lab of Intelligent Information Processing of Chinese Academy of Sciences, Institute of Computing Technology, Chinese Academy of Sciences, Beijing, 100190, China; School of Biomedical Engineering & Suzhou Institute For Advanced Research, University of Science and Technology, Suzhou, 215123, China.
| |
Collapse
|
226
|
Zhang Z, Ren J, Tao X, Tang W, Zhao S, Zhou L, Huang Y, Wang J, Wu N. Automatic segmentation of pulmonary lobes on low-dose computed tomography using deep learning. ANNALS OF TRANSLATIONAL MEDICINE 2021; 9:291. [PMID: 33708918 PMCID: PMC7944332 DOI: 10.21037/atm-20-5060] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 12/18/2022]
Abstract
Background To develop and validate a fully automated deep learning-based segmentation algorithm to segment pulmonary lobe on low-dose computed tomography (LDCT) images. Methods This study presents an automatic segmentation of pulmonary lobes using a fully convolutional neural network named dense V-network (DenseVNet) on lung cancer screening LDCT images. A total of 160 LDCT cases for lung cancer screening (100 in the training set, 10 in the validation set, and 50 in the test set) was included in this study. Specifically, the template of pulmonary lobes (the right lung consists of three lobes, and the left lung consists of two lobes) were obtained from pixel-level annotations by semiautomatic segmentation platform. Then, the model was trained under the supervision of the LDCT training set. Finally, the trained model was used to segment the LDCT in the test set. Dice coefficient, Jaccard coefficient, and Hausdorff distance were adopted as evaluation metrics to verify the performance of our segmentation model. Results In this study, the model achieved the accurate segmentation of each pulmonary lobe in seconds without the intervention of researchers. The testing set consisted 50 LDCT cases were used to evaluate the performance of the segmentation model. The all-lobes Dice coefficient of the test set was 0.944, the Jaccard coefficient was 0.896, and the Hausdorff distance was 92.908 mm. Conclusions The segmentation model based on LDCT can automatically and robustly and efficiently segment pulmonary lobes. It will provide effective location information and contour constraints for pulmonary nodule detection on LDCT images for lung cancer screening, which may have potential clinical application.
Collapse
Affiliation(s)
- Zewei Zhang
- PET-CT Center, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | | | - Xiuli Tao
- PET-CT Center, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Wei Tang
- Department of Diagnostic Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Shijun Zhao
- Department of Diagnostic Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Lina Zhou
- Department of Diagnostic Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Yao Huang
- Department of Diagnostic Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Jianwei Wang
- Department of Diagnostic Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Ning Wu
- PET-CT Center, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China.,Department of Diagnostic Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| |
Collapse
|
227
|
Caprara S, Carrillo F, Snedeker JG, Farshad M, Senteler M. Automated Pipeline to Generate Anatomically Accurate Patient-Specific Biomechanical Models of Healthy and Pathological FSUs. Front Bioeng Biotechnol 2021; 9:636953. [PMID: 33585436 PMCID: PMC7876284 DOI: 10.3389/fbioe.2021.636953] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2020] [Accepted: 01/11/2021] [Indexed: 12/29/2022] Open
Abstract
State-of-the-art preoperative biomechanical analysis for the planning of spinal surgery not only requires the generation of three-dimensional patient-specific models but also the accurate biomechanical representation of vertebral joints. The benefits offered by computational models suitable for such purposes are still outweighed by the time and effort required for their generation, thus compromising their applicability in a clinical environment. In this work, we aim to ease the integration of computerized methods into patient-specific planning of spinal surgery. We present the first pipeline combining deep learning and finite element methods that allows a completely automated model generation of functional spine units (FSUs) of the lumbar spine for patient-specific FE simulations (FEBio). The pipeline consists of three steps: (a) multiclass segmentation of cropped 3D CT images containing lumbar vertebrae using the DenseVNet network, (b) automatic landmark-based mesh fitting of statistical shape models onto 3D semantic segmented meshes of the vertebral models, and (c) automatic generation of patient-specific FE models of lumbar segments for the simulation of flexion-extension, lateral bending, and axial rotation movements. The automatic segmentation of FSUs was evaluated against the gold standard (manual segmentation) using 10-fold cross-validation. The obtained Dice coefficient was 93.7% on average, with a mean surface distance of 0.88 mm and a mean Hausdorff distance of 11.16 mm (N = 150). Automatic generation of finite element models to simulate the range of motion (ROM) was successfully performed for five healthy and five pathological FSUs. The results of the simulations were evaluated against the literature and showed comparable ROMs in both healthy and pathological cases, including the alteration of ROM typically observed in severely degenerated FSUs. The major intent of this work is to automate the creation of anatomically accurate patient-specific models by a single pipeline allowing functional modeling of spinal motion in healthy and pathological FSUs. Our approach reduces manual efforts to a minimum and the execution of the entire pipeline including simulations takes approximately 2 h. The automation, time-efficiency and robustness level of the pipeline represents a first step toward its clinical integration.
Collapse
Affiliation(s)
- Sebastiano Caprara
- Department of Orthopedics, University Hospital Balgrist, University of Zurich, Zurich, Switzerland
- Institute for Biomechanics, Swiss Federal Institute of Technology (ETH), Zurich, Switzerland
| | - Fabio Carrillo
- Institute for Biomechanics, Swiss Federal Institute of Technology (ETH), Zurich, Switzerland
- Research in Orthopedic Computer Science, University Hospital Balgrist, Zurich, Switzerland
| | - Jess G. Snedeker
- Department of Orthopedics, University Hospital Balgrist, University of Zurich, Zurich, Switzerland
- Institute for Biomechanics, Swiss Federal Institute of Technology (ETH), Zurich, Switzerland
| | - Mazda Farshad
- Department of Orthopedics, University Hospital Balgrist, University of Zurich, Zurich, Switzerland
| | - Marco Senteler
- Department of Orthopedics, University Hospital Balgrist, University of Zurich, Zurich, Switzerland
- Institute for Biomechanics, Swiss Federal Institute of Technology (ETH), Zurich, Switzerland
| |
Collapse
|
228
|
Liver segmentation in abdominal CT images via auto-context neural network and self-supervised contour attention. Artif Intell Med 2021; 113:102023. [PMID: 33685586 DOI: 10.1016/j.artmed.2021.102023] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2020] [Revised: 11/13/2020] [Accepted: 01/18/2021] [Indexed: 12/25/2022]
Abstract
OBJECTIVE Accurate image segmentation of the liver is a challenging problem owing to its large shape variability and unclear boundaries. Although the applications of fully convolutional neural networks (CNNs) have shown groundbreaking results, limited studies have focused on the performance of generalization. In this study, we introduce a CNN for liver segmentation on abdominal computed tomography (CT) images that focus on the performance of generalization and accuracy. METHODS To improve the generalization performance, we initially propose an auto-context algorithm in a single CNN. The proposed auto-context neural network exploits an effective high-level residual estimation to obtain the shape prior. Identical dual paths are effectively trained to represent mutual complementary features for an accurate posterior analysis of a liver. Further, we extend our network by employing a self-supervised contour scheme. We trained sparse contour features by penalizing the ground-truth contour to focus more contour attentions on the failures. RESULTS We used 180 abdominal CT images for training and validation. Two-fold cross-validation is presented for a comparison with the state-of-the-art neural networks. The experimental results show that the proposed network results in better accuracy when compared to the state-of-the-art networks by reducing 10.31% of the Hausdorff distance. Novel multiple N-fold cross-validations are conducted to show the best performance of generalization of the proposed network. CONCLUSION AND SIGNIFICANCE The proposed method minimized the error between training and test images more than any other modern neural networks. Moreover, the contour scheme was successfully employed in the network by introducing a self-supervising metric.
Collapse
|
229
|
|
230
|
Abstract
The interest in artificial intelligence (AI) has ballooned within radiology in the past few years primarily due to notable successes of deep learning. With the advances brought by deep learning, AI has the potential to recognize and localize complex patterns from different radiological imaging modalities, many of which even achieve comparable performance to human decision-making in recent applications. In this chapter, we review several AI applications in radiology for different anatomies: chest, abdomen, pelvis, as well as general lesion detection/identification that is not limited to specific anatomies. For each anatomy site, we focus on introducing the tasks of detection, segmentation, and classification with an emphasis on describing the technology development pathway with the aim of providing the reader with an understanding of what AI can do in radiology and what still needs to be done for AI to better fit in radiology. Combining with our own research experience of AI in medicine, we elaborate how AI can enrich knowledge discovery, understanding, and decision-making in radiology, rather than replacing the radiologist.
Collapse
|
231
|
Zhou Z, Sodha V, Pang J, Gotway MB, Liang J. Models Genesis. Med Image Anal 2021; 67:101840. [PMID: 33188996 PMCID: PMC7726094 DOI: 10.1016/j.media.2020.101840] [Citation(s) in RCA: 114] [Impact Index Per Article: 28.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2020] [Revised: 08/12/2020] [Accepted: 09/14/2020] [Indexed: 12/27/2022]
Abstract
Transfer learning from natural image to medical image has been established as one of the most practical paradigms in deep learning for medical image analysis. To fit this paradigm, however, 3D imaging tasks in the most prominent imaging modalities (e.g., CT and MRI) have to be reformulated and solved in 2D, losing rich 3D anatomical information, thereby inevitably compromising its performance. To overcome this limitation, we have built a set of models, called Generic Autodidactic Models, nicknamed Models Genesis, because they are created ex nihilo (with no manual labeling), self-taught (learnt by self-supervision), and generic (served as source models for generating application-specific target models). Our extensive experiments demonstrate that our Models Genesis significantly outperform learning from scratch and existing pre-trained 3D models in all five target 3D applications covering both segmentation and classification. More importantly, learning a model from scratch simply in 3D may not necessarily yield performance better than transfer learning from ImageNet in 2D, but our Models Genesis consistently top any 2D/2.5D approaches including fine-tuning the models pre-trained from ImageNet as well as fine-tuning the 2D versions of our Models Genesis, confirming the importance of 3D anatomical information and significance of Models Genesis for 3D medical imaging. This performance is attributed to our unified self-supervised learning framework, built on a simple yet powerful observation: the sophisticated and recurrent anatomy in medical images can serve as strong yet free supervision signals for deep models to learn common anatomical representation automatically via self-supervision. As open science, all codes and pre-trained Models Genesis are available at https://github.com/MrGiovanni/ModelsGenesis.
Collapse
Affiliation(s)
- Zongwei Zhou
- Department of Biomedical Informatics, Arizona State University, Scottsdale, AZ 85259, USA
| | - Vatsal Sodha
- School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ 85281 USA
| | - Jiaxuan Pang
- School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ 85281 USA
| | | | - Jianming Liang
- Department of Biomedical Informatics, Arizona State University, Scottsdale, AZ 85259, USA.
| |
Collapse
|
232
|
Si K, Xue Y, Yu X, Zhu X, Li Q, Gong W, Liang T, Duan S. Fully end-to-end deep-learning-based diagnosis of pancreatic tumors. Am J Cancer Res 2021; 11:1982-1990. [PMID: 33408793 PMCID: PMC7778580 DOI: 10.7150/thno.52508] [Citation(s) in RCA: 40] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2020] [Accepted: 11/17/2020] [Indexed: 12/15/2022] Open
Abstract
Artificial intelligence can facilitate clinical decision making by considering massive amounts of medical imaging data. Various algorithms have been implemented for different clinical applications. Accurate diagnosis and treatment require reliable and interpretable data. For pancreatic tumor diagnosis, only 58.5% of images from the First Affiliated Hospital and the Second Affiliated Hospital, Zhejiang University School of Medicine are used, increasing labor and time costs to manually filter out images not directly used by the diagnostic model. Methods: This study used a training dataset of 143,945 dynamic contrast-enhanced CT images of the abdomen from 319 patients. The proposed model contained four stages: image screening, pancreas location, pancreas segmentation, and pancreatic tumor diagnosis. Results: We established a fully end-to-end deep-learning model for diagnosing pancreatic tumors and proposing treatment. The model considers original abdominal CT images without any manual preprocessing. Our artificial-intelligence-based system achieved an area under the curve of 0.871 and a F1 score of 88.5% using an independent testing dataset containing 107,036 clinical CT images from 347 patients. The average accuracy for all tumor types was 82.7%, and the independent accuracies of identifying intraductal papillary mucinous neoplasm and pancreatic ductal adenocarcinoma were 100% and 87.6%, respectively. The average test time per patient was 18.6 s, compared with at least 8 min for manual reviewing. Furthermore, the model provided a transparent and interpretable diagnosis by producing saliency maps highlighting the regions relevant to its decision. Conclusions: The proposed model can potentially deliver efficient and accurate preoperative diagnoses that could aid the surgical management of pancreatic tumor.
Collapse
|
233
|
Baessler B. [Artificial Intelligence in Radiology - Definition, Potential and Challenges]. PRAXIS 2021; 110:48-53. [PMID: 33406927 DOI: 10.1024/1661-8157/a003597] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Artificial Intelligence in Radiology - Definition, Potential and Challenges Abstract. Artificial Intelligence (AI) is omnipresent. It has neatly permeated our daily life, even if we are not always fully aware of its ubiquitous presence. The healthcare sector in particular is experiencing a revolution which will change our daily routine considerably in the near future. Due to its advanced digitization and its historical technical affinity radiology is especially prone to these developments. But what exactly is AI and what makes AI so potent that established medical disciplines such as radiology worry about their future job perspectives? What are the assets of AI in radiology today - and what are the major challenges? This review article tries to give some answers to these questions.
Collapse
Affiliation(s)
- Bettina Baessler
- Institut für Diagnostische und Interventionelle Radiologie, Universitätsspital Zürich
| |
Collapse
|
234
|
U S, K. PT, K S. Computer aided diagnosis of obesity based on thermal imaging using various convolutional neural networks. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2020.102233] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
|
235
|
Akhtar Y, Dakua SP, Abdalla A, Aboumarzouk OM, Ansari MY, Abinahed J, Elakkad MSM, Al-Ansari A. Risk Assessment of Computer-aided Diagnostic Software for Hepatic Resection. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2021. [DOI: 10.1109/trpms.2021.3071148] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Affiliation(s)
- Yusuf Akhtar
- Electronics and Communication Sciences Unit, Indian Statistical Institute, Kolkata, India
| | | | | | | | | | - Julien Abinahed
- Hamad General Hospital, Hamad Medical Corporation, Doha, Qatar
| | | | | |
Collapse
|
236
|
Chen D, Zhang X, Mei Y, Liao F, Xu H, Li Z, Xiao Q, Guo W, Zhang H, Yan T, Xiong J, Ventikos Y. Multi-stage learning for segmentation of aortic dissections using a prior aortic anatomy simplification. Med Image Anal 2020; 69:101931. [PMID: 33618153 DOI: 10.1016/j.media.2020.101931] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2020] [Revised: 11/20/2020] [Accepted: 11/27/2020] [Indexed: 12/30/2022]
Abstract
Aortic dissection (AD) is a life-threatening cardiovascular disease with a high mortality rate. The accurate and generalized 3-D reconstruction of AD from CT-angiography can effectively assist clinical procedures and surgery plans, however, is clinically unavaliable due to the lacking of efficient tools. In this study, we presented a novel multi-stage segmentation framework for type B AD to extract true lumen (TL), false lumen (FL) and all branches (BR) as different classes. Two cascaded neural networks were used to segment the aortic trunk and branches and to separate the dual lumen, respectively. An aortic straightening method was designed based on the prior vascular anatomy of AD, simplifying the curved aortic shape before the second network. The straightening-based method achieved the mean Dice scores of 0.96, 0.95 and 0.89 for TL, FL, and BR on a multi-center dataset involving 120 patients, outperforming the end-to-end multi-class methods and the multi-stage methods without straightening on the dual-lumen segmentation, even using different network architectures. Both the global volumetric features of the aorta and the local characteristics of the primary tear could be better identified and quantified based on the straightening. Comparing to previous deep learning methods dealing with AD segmentations, the proposed framework presented advantages in segmentation accuracy.
Collapse
Affiliation(s)
- Duanduan Chen
- School of Life Science, Beijing Institute of Technology, Beijing, China.
| | - Xuyang Zhang
- School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Yuqian Mei
- School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Fangzhou Liao
- Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China
| | - Huanming Xu
- School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Zhenfeng Li
- School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Qianjiang Xiao
- Shukun (Beijing) Network Technology Co.Ltd., Beijing, China
| | - Wei Guo
- Department of Vascular and Endovascular Surgery, Chinese PLA General Hospital, Beijing, China
| | - Hongkun Zhang
- Department of Vascular Surgery, First Affiliated Hospital of Medical College, Zhejiang University, Hangzhou, China
| | - Tianyi Yan
- School of Life Science, Beijing Institute of Technology, Beijing, China.
| | - Jiang Xiong
- Department of Vascular and Endovascular Surgery, Chinese PLA General Hospital, Beijing, China.
| | - Yiannis Ventikos
- Department of Mechanical Engineering, University College London, London, UK; School of Life Science, Beijing Institute of Technology, Beijing, China
| |
Collapse
|
237
|
Nikan S, Van Osch K, Bartling M, Allen DG, Rohani SA, Connors B, Agrawal SK, Ladak HM. PWD-3DNet: A Deep Learning-Based Fully-Automated Segmentation of Multiple Structures on Temporal Bone CT Scans. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; 30:739-753. [PMID: 33226942 DOI: 10.1109/tip.2020.3038363] [Citation(s) in RCA: 29] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
The temporal bone is a part of the lateral skull surface that contains organs responsible for hearing and balance. Mastering surgery of the temporal bone is challenging because of this complex and microscopic three-dimensional anatomy. Segmentation of intra-temporal anatomy based on computed tomography (CT) images is necessary for applications such as surgical training and rehearsal, amongst others. However, temporal bone segmentation is challenging due to the similar intensities and complicated anatomical relationships among critical structures, undetectable small structures on standard clinical CT, and the amount of time required for manual segmentation. This paper describes a single multi-class deep learning-based pipeline as the first fully automated algorithm for segmenting multiple temporal bone structures from CT volumes, including the sigmoid sinus, facial nerve, inner ear, malleus, incus, stapes, internal carotid artery and internal auditory canal. The proposed fully convolutional network, PWD-3DNet, is a patch-wise densely connected (PWD) three-dimensional (3D) network. The accuracy and speed of the proposed algorithm was shown to surpass current manual and semi-automated segmentation techniques. The experimental results yielded significantly high Dice similarity scores and low Hausdorff distances for all temporal bone structures with an average of 86% and 0.755 millimeter (mm), respectively. We illustrated that overlapping in the inference sub-volumes improves the segmentation performance. Moreover, we proposed augmentation layers by using samples with various transformations and image artefacts to increase the robustness of PWD-3DNet against image acquisition protocols, such as smoothing caused by soft tissue scanner settings and larger voxel sizes used for radiation reduction. The proposed algorithm was tested on low-resolution CTs acquired by another center with different scanner parameters than the ones used to create the algorithm and shows potential for application beyond the particular training data used in the study.
Collapse
|
238
|
The integration of artificial intelligence models to augment imaging modalities in pancreatic cancer. JOURNAL OF PANCREATOLOGY 2020. [DOI: 10.1097/jp9.0000000000000056] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/25/2022] Open
|
239
|
Esaki T, Furukawa R. [Volume Measurements of Post-transplanted Liver of Pediatric Recipients Using Workstations and Deep Learning]. Nihon Hoshasen Gijutsu Gakkai Zasshi 2020; 76:1133-1142. [PMID: 33229843 DOI: 10.6009/jjrt.2020_jsrt_76.11.1133] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
PURPOSE The purpose of this study was to propose a method for segmentation and volume measurement of graft liver and spleen of pediatric transplant recipients on digital imaging and communications in medicine (DICOM) -format images using U-Net and three-dimensional (3-D) workstations (3DWS) . METHOD For segmentation accuracy assessments, Dice coefficients were calculated for the graft liver and spleen. After verifying that the created DICOM-format images could be imported using the existing 3DWS, accuracy rates between the ground truth and segmentation images were calculated via mask processing. RESULT As per the verification results, Dice coefficients for the test data were as follows: graft liver, 0.758 and spleen, 0.577. All created DICOM-format images were importable using the 3DWS, with accuracy rates of 87.10±4.70% and 80.27±11.29% for the graft liver and spleen, respectively. CONCLUSION The U-Net could be used for graft liver and spleen segmentations, and volume measurement using 3DWS was simplified by this method.
Collapse
Affiliation(s)
- Toru Esaki
- Department of Radiologic Technology, Jichi Medical University Hospital
| | - Rieko Furukawa
- Department of Pediatric Medical Imaging, Jichi Children's Medical Center Tochigi
| |
Collapse
|
240
|
Rister B, Yi D, Shivakumar K, Nobashi T, Rubin DL. CT-ORG, a new dataset for multiple organ segmentation in computed tomography. Sci Data 2020; 7:381. [PMID: 33177518 PMCID: PMC7658204 DOI: 10.1038/s41597-020-00715-8] [Citation(s) in RCA: 38] [Impact Index Per Article: 7.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2020] [Accepted: 10/01/2020] [Indexed: 12/05/2022] Open
Abstract
Despite the relative ease of locating organs in the human body, automated organ segmentation has been hindered by the scarcity of labeled training data. Due to the tedium of labeling organ boundaries, most datasets are limited to either a small number of cases or a single organ. Furthermore, many are restricted to specific imaging conditions unrepresentative of clinical practice. To address this need, we developed a diverse dataset of 140 CT scans containing six organ classes: liver, lungs, bladder, kidney, bones and brain. For the lungs and bones, we expedited annotation using unsupervised morphological segmentation algorithms, which were accelerated by 3D Fourier transforms. Demonstrating the utility of the data, we trained a deep neural network which requires only 4.3 s to simultaneously segment all the organs in a case. We also show how to efficiently augment the data to improve model generalization, providing a GPU library for doing so. We hope this dataset and code, available through TCIA, will be useful for training and evaluating organ segmentation models.
Collapse
Affiliation(s)
- Blaine Rister
- Department of Electrical Engineering, Stanford University, 350 Jane Stanford Way, Stanford, CA, 94305, USA.
| | - Darvin Yi
- Department of Biomedical Data Science, Stanford University, 1265 Welch Road, Stanford, CA, 94305, USA
| | - Kaushik Shivakumar
- Department of Biomedical Data Science, Stanford University, 1265 Welch Road, Stanford, CA, 94305, USA
| | - Tomomi Nobashi
- Department of Radiology, Stanford University, 300 Pasteur Drive, Stanford, CA, 94305, USA
| | - Daniel L Rubin
- Department of Biomedical Data Science, Stanford University, 1265 Welch Road, Stanford, CA, 94305, USA
- Department of Radiology, Stanford University, 300 Pasteur Drive, Stanford, CA, 94305, USA
| |
Collapse
|
241
|
Fang X, Yan P. Multi-Organ Segmentation Over Partially Labeled Datasets With Multi-Scale Feature Abstraction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3619-3629. [PMID: 32746108 PMCID: PMC7665851 DOI: 10.1109/tmi.2020.3001036] [Citation(s) in RCA: 66] [Impact Index Per Article: 13.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Shortage of fully annotated datasets has been a limiting factor in developing deep learning based image segmentation algorithms and the problem becomes more pronounced in multi-organ segmentation. In this paper, we propose a unified training strategy that enables a novel multi-scale deep neural network to be trained on multiple partially labeled datasets for multi-organ segmentation. In addition, a new network architecture for multi-scale feature abstraction is proposed to integrate pyramid input and feature analysis into a U-shape pyramid structure. To bridge the semantic gap caused by directly merging features from different scales, an equal convolutional depth mechanism is introduced. Furthermore, we employ a deep supervision mechanism to refine the outputs in different scales. To fully leverage the segmentation features from all the scales, we design an adaptive weighting layer to fuse the outputs in an automatic fashion. All these mechanisms together are integrated into a Pyramid Input Pyramid Output Feature Abstraction Network (PIPO-FAN). Our proposed method was evaluated on four publicly available datasets, including BTCV, LiTS, KiTS and Spleen, where very promising performance has been achieved. The source code of this work is publicly shared at https://github.com/DIAL-RPI/PIPO-FAN to facilitate others to reproduce the work and build their own models using the introduced mechanisms.
Collapse
|
242
|
Mou L, Zhao Y, Fu H, Liu Y, Cheng J, Zheng Y, Su P, Yang J, Chen L, Frangi AF, Akiba M, Liu J. CS 2-Net: Deep learning segmentation of curvilinear structures in medical imaging. Med Image Anal 2020; 67:101874. [PMID: 33166771 DOI: 10.1016/j.media.2020.101874] [Citation(s) in RCA: 109] [Impact Index Per Article: 21.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2020] [Revised: 08/26/2020] [Accepted: 10/05/2020] [Indexed: 12/20/2022]
Abstract
Automated detection of curvilinear structures, e.g., blood vessels or nerve fibres, from medical and biomedical images is a crucial early step in automatic image interpretation associated to the management of many diseases. Precise measurement of the morphological changes of these curvilinear organ structures informs clinicians for understanding the mechanism, diagnosis, and treatment of e.g. cardiovascular, kidney, eye, lung, and neurological conditions. In this work, we propose a generic and unified convolution neural network for the segmentation of curvilinear structures and illustrate in several 2D/3D medical imaging modalities. We introduce a new curvilinear structure segmentation network (CS2-Net), which includes a self-attention mechanism in the encoder and decoder to learn rich hierarchical representations of curvilinear structures. Two types of attention modules - spatial attention and channel attention - are utilized to enhance the inter-class discrimination and intra-class responsiveness, to further integrate local features with their global dependencies and normalization, adaptively. Furthermore, to facilitate the segmentation of curvilinear structures in medical images, we employ a 1×3 and a 3×1 convolutional kernel to capture boundary features. Besides, we extend the 2D attention mechanism to 3D to enhance the network's ability to aggregate depth information across different layers/slices. The proposed curvilinear structure segmentation network is thoroughly validated using both 2D and 3D images across six different imaging modalities. Experimental results across nine datasets show the proposed method generally outperforms other state-of-the-art algorithms in various metrics.
Collapse
Affiliation(s)
- Lei Mou
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Yitian Zhao
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China.
| | - Huazhu Fu
- Inception Institute of Artificial Intelligence, Abu Dhabi, United Arab Emirates
| | - Yonghuai Liu
- Department of Computer Science, Edge Hill University, Ormskirk, UK
| | - Jun Cheng
- UBTech Research, UBTech Robotics Corp Ltd, Shenzhen, China
| | - Yalin Zheng
- Department of Eye and Vision Science, University of Liverpool, Liverpool, UK; Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Pan Su
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Jianlong Yang
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Li Chen
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan, China
| | - Alejandro F Frangi
- Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China; Centre for Computational Imaging and Simulation Technologies in Biomedicine (CISTIB), School of Computing and School of Medicine, University of Leeds, Leeds, UK; Leeds Institute of Cardiovascular and Metabolic Medicine, School of Medicine, University of Leeds, Leeds, UK; Medical Imaging Research Centre (MIRC), University Hospital Gasthuisberg, Cardiovascular Sciences and Electrical Engineering Departments, KU Leuven, Leuven, Belgium
| | | | - Jiang Liu
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China; Cixi Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, China; Guangdong Provincial Key Laboratory of Brain-inspired Intelligent Computation, Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China.
| |
Collapse
|
243
|
Evaluation of Multimodal Algorithms for the Segmentation of Multiparametric MRI Prostate Images. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2020; 2020:8861035. [PMID: 33144873 PMCID: PMC7596462 DOI: 10.1155/2020/8861035] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/15/2020] [Revised: 09/29/2020] [Accepted: 10/04/2020] [Indexed: 12/18/2022]
Abstract
Prostate segmentation in multiparametric magnetic resonance imaging (mpMRI) can help to support prostate cancer diagnosis and therapy treatment. However, manual segmentation of the prostate is subjective and time-consuming. Many deep learning monomodal networks have been developed for automatic whole prostate segmentation from T2-weighted MR images. We aimed to investigate the added value of multimodal networks in segmenting the prostate into the peripheral zone (PZ) and central gland (CG). We optimized and evaluated monomodal DenseVNet, multimodal ScaleNet, and monomodal and multimodal HighRes3DNet, which yielded dice score coefficients (DSC) of 0.875, 0.848, 0.858, and 0.890 in WG, respectively. Multimodal HighRes3DNet and ScaleNet yielded higher DSC with statistical differences in PZ and CG only compared to monomodal DenseVNet, indicating that multimodal networks added value by generating better segmentation between PZ and CG regions but did not improve the WG segmentation. No significant difference was observed in the apex and base of WG segmentation between monomodal and multimodal networks, indicating that the segmentations at the apex and base were more affected by the general network architecture. The number of training data was also varied for DenseVNet and HighRes3DNet, from 20 to 120 in steps of 20. DenseVNet was able to yield DSC of higher than 0.65 even for special cases, such as TURP or abnormal prostate, whereas HighRes3DNet's performance fluctuated with no trend despite being the best network overall. Multimodal networks did not add value in segmenting special cases but generally reduced variations in segmentation compared to the same matched monomodal network.
Collapse
|
244
|
Tan JW, Wang L, Chen Y, Xi W, Ji J, Wang L, Xu X, Zou LK, Feng JX, Zhang J, Zhang H. Predicting Chemotherapeutic Response for Far-advanced Gastric Cancer by Radiomics with Deep Learning Semi-automatic Segmentation. J Cancer 2020; 11:7224-7236. [PMID: 33193886 PMCID: PMC7646171 DOI: 10.7150/jca.46704] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2020] [Accepted: 10/04/2020] [Indexed: 12/23/2022] Open
Abstract
Purpose: To build a dual-energy computed tomography (DECT) delta radiomics model to predict chemotherapeutic response for far-advanced gastric cancer (GC) patients. A semi-automatic segmentation method based on deep learning was designed, and its performance was compared with that of manual segmentation. Methods: This retrospective study included 86 patients with far-advanced GC treated with chemotherapy from September 2016 to December 2017 (66 and 20 in the training and testing cohorts, respectively). Delta radiomics features between the baseline and first follow-up DECT were modeled by random forest to predict the chemotherapeutic response evaluated by the second follow-up DECT. Nine feature subsets from confounding factors and delta radiomics features were used to choose the best model with 10-fold cross-validation in the training cohort. A semi-automatic segmentation method based on deep learning was developed to predict the chemotherapeutic response and compared with manual segmentation in the testing cohort, which was further validated in an independent validation cohort of 30 patients. Results: The best model, constructed by confounding factors and texture features, reached an average AUC of 0.752 in the training cohort. Our proposed semi-automatic segmentation method was more time-effective than manual segmentation, with average saving-time of 11.2333 ± 6.3989 minutes and 9.9889 ±5.5086 minutes in the testing cohort and the independent validation cohort, respectively (both p < 0.05). The predictive ability of the semi-automatic segmentation was also better than that of the manual segmentation both in the testing cohort and the independent validation cohort (AUC: 0.728 vs. 0.687 and 0.828 vs. 0.749, respectively). Conclusion: DECT delta radiomics serves as a promising biomarker for predicting chemotherapeutic response for far-advanced GC. Semi-automatic segmentation based on deep learning shows the potential for clinical use with increased reproducibility and decreased labor costs compared to the manual version.
Collapse
Affiliation(s)
- Jing-Wen Tan
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Lan Wang
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Yong Chen
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - WenQi Xi
- Department of Oncology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Jun Ji
- Department of Oncology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Lingyun Wang
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Xin Xu
- Haohua Technology Co., Ltd, Weihai International Group Building, No. 511 Weihai Road, Shanghai, China
| | - Long-Kuan Zou
- Haohua Technology Co., Ltd, Weihai International Group Building, No. 511 Weihai Road, Shanghai, China
| | - Jian-Xing Feng
- Haohua Technology Co., Ltd, Weihai International Group Building, No. 511 Weihai Road, Shanghai, China
| | - Jun Zhang
- Department of Oncology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Huan Zhang
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| |
Collapse
|
245
|
Wang R, Cao S, Ma K, Zheng Y, Meng D. Pairwise learning for medical image segmentation. Med Image Anal 2020; 67:101876. [PMID: 33197863 DOI: 10.1016/j.media.2020.101876] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2020] [Revised: 09/29/2020] [Accepted: 10/05/2020] [Indexed: 02/08/2023]
Abstract
Fully convolutional networks (FCNs) trained with abundant labeled data have been proven to be a powerful and efficient solution for medical image segmentation. However, FCNs often fail to achieve satisfactory results due to the lack of labelled data and significant variability of appearance in medical imaging. To address this challenging issue, this paper proposes a conjugate fully convolutional network (CFCN) where pairwise samples are input for capturing a rich context representation and guide each other with a fusion module. To avoid the overfitting problem introduced by intra-class heterogeneity and boundary ambiguity with a small number of training samples, we propose to explicitly exploit the prior information from the label space, termed as proxy supervision. We further extend the CFCN to a compact conjugate fully convolutional network (C2FCN), which just has one head for fitting the proxy supervision without incurring two additional branches of decoders fitting ground truth of the input pairs compared to CFCN. In the test phase, the segmentation probability is inferred by the learned logical relation implied in the proxy supervision. Quantitative evaluation on the Liver Tumor Segmentation (LiTS) and Combined (CT-MR) Healthy Abdominal Organ Segmentation (CHAOS) datasets shows that the proposed framework achieves a significant performance improvement on both binary segmentation and multi-category segmentation, especially with a limited amount of training data. The source code is available at https://github.com/renzhenwang/pairwise_segmentation.
Collapse
Affiliation(s)
- Renzhen Wang
- School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an, 710049, China
| | - Shilei Cao
- Jarvis Lab, Tencent, Shenzhen, 518075, China
| | - Kai Ma
- Jarvis Lab, Tencent, Shenzhen, 518075, China
| | | | - Deyu Meng
- School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an, 710049, China; Macau Institute of Systems Engineering, Macau University of Science and Technology, Taipa, Macau.
| |
Collapse
|
246
|
Zhu Q, Li L, Hao J, Zha Y, Zhang Y, Cheng Y, Liao F, Li P. Selective information passing for MR/CT image segmentation. Neural Comput Appl 2020. [DOI: 10.1007/s00521-020-05407-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
247
|
Shiyam Sundar LK, Muzik O, Buvat I, Bidaut L, Beyer T. Potentials and caveats of AI in hybrid imaging. Methods 2020; 188:4-19. [PMID: 33068741 DOI: 10.1016/j.ymeth.2020.10.004] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2020] [Revised: 10/05/2020] [Accepted: 10/07/2020] [Indexed: 12/18/2022] Open
Abstract
State-of-the-art patient management frequently mandates the investigation of both anatomy and physiology of the patients. Hybrid imaging modalities such as the PET/MRI, PET/CT and SPECT/CT have the ability to provide both structural and functional information of the investigated tissues in a single examination. With the introduction of such advanced hardware fusion, new problems arise such as the exceedingly large amount of multi-modality data that requires novel approaches of how to extract a maximum of clinical information from large sets of multi-dimensional imaging data. Artificial intelligence (AI) has emerged as one of the leading technologies that has shown promise in facilitating highly integrative analysis of multi-parametric data. Specifically, the usefulness of AI algorithms in the medical imaging field has been heavily investigated in the realms of (1) image acquisition and reconstruction, (2) post-processing and (3) data mining and modelling. Here, we aim to provide an overview of the challenges encountered in hybrid imaging and discuss how AI algorithms can facilitate potential solutions. In addition, we highlight the pitfalls and challenges in using advanced AI algorithms in the context of hybrid imaging and provide suggestions for building robust AI solutions that enable reproducible and transparent research.
Collapse
Affiliation(s)
- Lalith Kumar Shiyam Sundar
- QIMP Team, Center for Medical Physics and Biomedical Engineering, Medical University of Vienna, Vienna, Austria
| | | | - Irène Buvat
- Laboratoire d'Imagerie Translationnelle en Oncologie, Inserm, Institut Curie, Orsay, France
| | - Luc Bidaut
- College of Science, University of Lincoln, Lincoln, UK
| | - Thomas Beyer
- QIMP Team, Center for Medical Physics and Biomedical Engineering, Medical University of Vienna, Vienna, Austria.
| |
Collapse
|
248
|
Weston AD, Korfiatis P, Philbrick KA, Conte GM, Kostandy P, Sakinis T, Zeinoddini A, Boonrod A, Moynagh M, Takahashi N, Erickson BJ. Complete abdomen and pelvis segmentation using U-net variant architecture. Med Phys 2020; 47:5609-5618. [PMID: 32740931 DOI: 10.1002/mp.14422] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2020] [Revised: 07/14/2020] [Accepted: 07/22/2020] [Indexed: 12/27/2022] Open
Abstract
PURPOSE Organ segmentation of computed tomography (CT) imaging is essential for radiotherapy treatment planning. Treatment planning requires segmentation not only of the affected tissue, but nearby healthy organs-at-risk, which is laborious and time-consuming. We present a fully automated segmentation method based on the three-dimensional (3D) U-Net convolutional neural network (CNN) capable of whole abdomen and pelvis segmentation into 33 unique organ and tissue structures, including tissues that may be overlooked by other automated segmentation approaches such as adipose tissue, skeletal muscle, and connective tissue and vessels. Whole abdomen segmentation is capable of quantifying exposure beyond a handful of organs-at-risk to all tissues within the abdomen. METHODS Sixty-six (66) CT examinations of 64 individuals were included in the training and validation sets and 18 CT examinations from 16 individuals were included in the test set. All pixels in each examination were segmented by image analysts (with physician correction) and assigned one of 33 labels. Segmentation was performed with a 3D U-Net variant architecture which included residual blocks, and model performance was quantified on 18 test cases. Human interobserver variability (using semiautomated segmentation) was also reported on two scans, and manual interobserver variability of three individuals was reported on one scan. Model performance was also compared to several of the best models reported in the literature for multiple organ segmentation. RESULTS The accuracy of the 3D U-Net model ranges from a Dice coefficient of 0.95 in the liver, 0.93 in the kidneys, 0.79 in the pancreas, 0.69 in the adrenals, and 0.51 in the renal arteries. Model accuracy is within 5% of human segmentation in eight of 19 organs and within 10% accuracy in 13 of 19 organs. CONCLUSIONS The CNN approaches the accuracy of human tracers and on certain complex organs displays more consistent prediction than human tracers. Fully automated deep learning-based segmentation of CT abdomen has the potential to improve both the speed and accuracy of radiotherapy dose prediction for organs-at-risk.
Collapse
Affiliation(s)
- Alexander D Weston
- Health Sciences Research, Mayo Clinic, 4500 San Pablo Road S, Jacksonville, FL, 32250, USA
| | - Panagiotis Korfiatis
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Kenneth A Philbrick
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Gian Marco Conte
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Petro Kostandy
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Thomas Sakinis
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Atefeh Zeinoddini
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Arunnit Boonrod
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Michael Moynagh
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Naoki Takahashi
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| | - Bradley J Erickson
- Department of Radiology, Mayo Clinic, 200 First St SW, Rochester, MN, 55905, USA
| |
Collapse
|
249
|
Gao C, Liu X, Gu W, Killeen B, Armand M, Taylor R, Unberath M. Generalizing Spatial Transformers to Projective Geometry with Applications to 2D/3D Registration. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12263:329-339. [PMID: 33135014 DOI: 10.1007/978-3-030-59716-0_32] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
Differentiable rendering is a technique to connect 3D scenes with corresponding 2D images. Since it is differentiable, processes during image formation can be learned. Previous approaches to differentiable rendering focus on mesh-based representations of 3D scenes, which is inappropriate for medical applications where volumetric, voxelized models are used to represent anatomy. We propose a novel Projective Spatial Transformer module that generalizes spatial transformers to projective geometry, thus enabling differentiable volume rendering. We demonstrate the usefulness of this architecture on the example of 2D/3D registration between radiographs and CT scans. Specifically, we show that our transformer enables end-to-end learning of an image processing and projection model that approximates an image similarity function that is convex with respect to the pose parameters, and can thus be optimized effectively using conventional gradient descent. To the best of our knowledge, we are the first to describe the spatial transformers in the context of projective transmission imaging, including rendering and pose estimation. We hope that our developments will benefit related 3D research applications. The source code is available at https://github.com/gaocong13/Projective-Spatial-Transformers.
Collapse
Affiliation(s)
- Cong Gao
- Johns Hopkins University, Baltimore MD 21218, USA
| | - Xingtong Liu
- Johns Hopkins University, Baltimore MD 21218, USA
| | - Wenhao Gu
- Johns Hopkins University, Baltimore MD 21218, USA
| | | | | | | | | |
Collapse
|
250
|
Torrents-Barrena J, Piella G, Gratacos E, Eixarch E, Ceresa M, Gonalez Ballester MA. Deep Q-CapsNet Reinforcement Learning Framework for Intrauterine Cavity Segmentation in TTTS Fetal Surgery Planning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3113-3124. [PMID: 32305906 DOI: 10.1109/tmi.2020.2987981] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Fetoscopic laser photocoagulation is the most effective treatment for Twin-to-Twin Transfusion Syndrome, a condition affecting twin pregnancies in which there is a deregulation of blood circulation through the placenta, that can be fatal to both babies. For the purposes of surgical planning, we design the first automatic approach to detect and segment the intrauterine cavity from axial, sagittal and coronal MRI stacks. Our methodology relies on the ability of capsule networks to successfully capture the part-whole interdependency of objects in the scene, particularly for unique class instances (i.e., intrauterine cavity). The presented deep Q-CapsNet reinforcement learning framework is built upon a context-adaptive detection policy to generate a bounding box of the womb. A capsule architecture is subsequently designed to segment (or refine) the whole intrauterine cavity. This network is coupled with a strided nnU-Net feature extractor, which encodes discriminative feature maps to construct strong primary capsules. The method is robustly evaluated with and without the localization stage using 13 performance measures, and directly compared with 15 state-of-the-art deep neural networks trained on 71 singleton and monochorionic twin pregnancies. An average Dice score above 0.91 is achieved for all ablations, revealing the potential of our approach to be used in clinical practice.
Collapse
|