1
|
Tan Z, Feng J, Lu W, Yin Y, Yang G, Zhou J. Multi-task global optimization-based method for vascular landmark detection. Comput Med Imaging Graph 2024; 114:102364. [PMID: 38432060 DOI: 10.1016/j.compmedimag.2024.102364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Revised: 12/04/2023] [Accepted: 02/22/2024] [Indexed: 03/05/2024]
Abstract
Vascular landmark detection plays an important role in medical analysis and clinical treatment. However, due to the complex topology and similar local appearance around landmarks, the popular heatmap regression based methods always suffer from the landmark confusion problem. Vascular landmarks are connected by vascular segments and have special spatial correlations, which can be utilized for performance improvement. In this paper, we propose a multi-task global optimization-based framework for accurate and automatic vascular landmark detection. A multi-task deep learning network is exploited to accomplish landmark heatmap regression, vascular semantic segmentation, and orientation field regression simultaneously. The two auxiliary objectives are highly correlated with the heatmap regression task and help the network incorporate the structural prior knowledge. During inference, instead of performing a max-voting strategy, we propose a global optimization-based post-processing method for final landmark decision. The spatial relationships between neighboring landmarks are utilized explicitly to tackle the landmark confusion problem. We evaluated our method on a cerebral MRA dataset with 564 volumes, a cerebral CTA dataset with 510 volumes, and an aorta CTA dataset with 50 volumes. The experiments demonstrate that the proposed method is effective for vascular landmark localization and achieves state-of-the-art performance.
Collapse
Affiliation(s)
- Zimeng Tan
- Department of Automation, Tsinghua University, Beijing, China
| | - Jianjiang Feng
- Department of Automation, Tsinghua University, Beijing, China.
| | - Wangsheng Lu
- UnionStrong (Beijing) Technology Co.Ltd, Beijing, China
| | - Yin Yin
- UnionStrong (Beijing) Technology Co.Ltd, Beijing, China
| | | | - Jie Zhou
- Department of Automation, Tsinghua University, Beijing, China
| |
Collapse
|
2
|
Ao Y, Wu H. Feature Aggregation and Refinement Network for 2D Anatomical Landmark Detection. J Digit Imaging 2023; 36:547-561. [PMID: 36401132 PMCID: PMC10039137 DOI: 10.1007/s10278-022-00718-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2022] [Revised: 10/06/2022] [Accepted: 10/13/2022] [Indexed: 11/19/2022] Open
Abstract
Localization of anatomical landmarks is essential for clinical diagnosis, treatment planning, and research. This paper proposes a novel deep network named feature aggregation and refinement network (FARNet) for automatically detecting anatomical landmarks. FARNet employs an encoder-decoder structure architecture. To alleviate the problem of limited training data in the medical domain, we adopt a backbone network pre-trained on natural images as the encoder. The decoder includes a multi-scale feature aggregation module for multi-scale feature fusion and a feature refinement module for high-resolution heatmap regression. Coarse-to-fine supervisions are applied to the two modules to facilitate end-to-end training. We further propose a novel loss function named Exponential Weighted Center loss for accurate heatmap regression, which focuses on the losses from the pixels near landmarks and suppresses the ones from far away. We evaluate FARNet on three publicly available anatomical landmark detection datasets, including cephalometric, hand, and spine radiographs. Our network achieves state-of-the-art performances on all three datasets. Code is available at https://github.com/JuvenileInWind/FARNet .
Collapse
Affiliation(s)
- Yueyuan Ao
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan 611731 China
| | - Hong Wu
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan 611731 China
| |
Collapse
|
3
|
Xu M, Liu B, Luo Z, Ma H, Sun M, Wang Y, Yin N, Tang X, Song T. Using a New Deep Learning Method for 3D Cephalometry in Patients With Cleft Lip and Palate. J Craniofac Surg 2023:00001665-990000000-00651. [PMID: 36944601 DOI: 10.1097/scs.0000000000009299] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2022] [Accepted: 12/28/2022] [Indexed: 03/23/2023] Open
Abstract
Deep learning algorithms based on automatic 3-dimensional (D) cephalometric marking points about people without craniomaxillofacial deformities has achieved good results. However, there has been no previous report about cleft lip and palate. The purpose of this study is to apply a new deep learning method based on a 3D point cloud graph convolutional neural network to predict and locate landmarks in patients with cleft lip and palate based on the relationships between points. The authors used the PointNet++ model to investigate the automatic 3D cephalometric marking points. And the mean distance error of the center coordinate position and the success detection rate (SDR) were used to evaluate the accuracy of systematic labeling. A total of 150 patients were enrolled. The mean distance error for all 27 landmarks was 1.33 mm, and 9 landmarks (30%) showed SDRs at 2 mm over 90%, and 3 landmarks (35%) showed SDRs at 2 mm under 70%. The automatic 3D cephalometric marking points take 16 seconds per dataset. In summary, our training sets were derived from the cleft lip with/without palate computed tomography to achieve accurate results. The 3D cephalometry system based on the graph convolutional neural network algorithm may be suitable for 3D cephalometry system in cleft lip and palate cases. More accurate results may be obtained if the cleft lip and palate training set is expanded in the future.
Collapse
Affiliation(s)
- Meng Xu
- Cleft Lip and Palate Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College
| | - Bingyang Liu
- Maxillofacial Surgery Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing
| | - Zhaoyang Luo
- HaiChuang Future Medical Technology Co. Ltd, Hangzhou
| | - Hengyuan Ma
- Digital Technology Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Min Sun
- Cleft Lip and Palate Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College
| | - Yongqian Wang
- Cleft Lip and Palate Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College
| | - Ningbei Yin
- Cleft Lip and Palate Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College
| | - Xiaojun Tang
- Maxillofacial Surgery Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing
| | - Tao Song
- Cleft Lip and Palate Center, Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College
| |
Collapse
|
4
|
Iyer S, Blair A, White C, Dawes L, Moses D, Sowmya A. Vertebral compression fracture detection using imitation learning, patch based convolutional neural networks and majority voting. INFORMATICS IN MEDICINE UNLOCKED 2023. [DOI: 10.1016/j.imu.2023.101238] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/09/2023] Open
|
5
|
Udupa JK, Liu T, Jin C, Zhao L, Odhner D, Tong Y, Agrawal V, Pednekar G, Nag S, Kotia T, Goodman M, Wileyto EP, Mihailidis D, Lukens JN, Berman AT, Stambaugh J, Lim T, Chowdary R, Jalluri D, Jabbour SK, Kim S, Reyhan M, Robinson CG, Thorstad WL, Choi JI, Press R, Simone CB, Camaratta J, Owens S, Torigian DA. Combining natural and artificial intelligence for robust automatic anatomy segmentation: Application in neck and thorax auto-contouring. Med Phys 2022; 49:7118-7149. [PMID: 35833287 PMCID: PMC10087050 DOI: 10.1002/mp.15854] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2022] [Revised: 06/20/2022] [Accepted: 06/30/2022] [Indexed: 01/01/2023] Open
Abstract
BACKGROUND Automatic segmentation of 3D objects in computed tomography (CT) is challenging. Current methods, based mainly on artificial intelligence (AI) and end-to-end deep learning (DL) networks, are weak in garnering high-level anatomic information, which leads to compromised efficiency and robustness. This can be overcome by incorporating natural intelligence (NI) into AI methods via computational models of human anatomic knowledge. PURPOSE We formulate a hybrid intelligence (HI) approach that integrates the complementary strengths of NI and AI for organ segmentation in CT images and illustrate performance in the application of radiation therapy (RT) planning via multisite clinical evaluation. METHODS The system employs five modules: (i) body region recognition, which automatically trims a given image to a precisely defined target body region; (ii) NI-based automatic anatomy recognition object recognition (AAR-R), which performs object recognition in the trimmed image without DL and outputs a localized fuzzy model for each object; (iii) DL-based recognition (DL-R), which refines the coarse recognition results of AAR-R and outputs a stack of 2D bounding boxes (BBs) for each object; (iv) model morphing (MM), which deforms the AAR-R fuzzy model of each object guided by the BBs output by DL-R; and (v) DL-based delineation (DL-D), which employs the object containment information provided by MM to delineate each object. NI from (ii), AI from (i), (iii), and (v), and their combination from (iv) facilitate the HI system. RESULTS The HI system was tested on 26 organs in neck and thorax body regions on CT images obtained prospectively from 464 patients in a study involving four RT centers. Data sets from one separate independent institution involving 125 patients were employed in training/model building for each of the two body regions, whereas 104 and 110 data sets from the 4 RT centers were utilized for testing on neck and thorax, respectively. In the testing data sets, 83% of the images had limitations such as streak artifacts, poor contrast, shape distortion, pathology, or implants. The contours output by the HI system were compared to contours drawn in clinical practice at the four RT centers by utilizing an independently established ground-truth set of contours as reference. Three sets of measures were employed: accuracy via Dice coefficient (DC) and Hausdorff boundary distance (HD), subjective clinical acceptability via a blinded reader study, and efficiency by measuring human time saved in contouring by the HI system. Overall, the HI system achieved a mean DC of 0.78 and 0.87 and a mean HD of 2.22 and 4.53 mm for neck and thorax, respectively. It significantly outperformed clinical contouring in accuracy and saved overall 70% of human time over clinical contouring time, whereas acceptability scores varied significantly from site to site for both auto-contours and clinically drawn contours. CONCLUSIONS The HI system is observed to behave like an expert human in robustness in the contouring task but vastly more efficiently. It seems to use NI help where image information alone will not suffice to decide, first for the correct localization of the object and then for the precise delineation of the boundary.
Collapse
Affiliation(s)
- Jayaram K. Udupa
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Tiange Liu
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
- School of Information Science and EngineeringYanshan UniversityQinhuangdaoChina
| | - Chao Jin
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Liming Zhao
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Dewey Odhner
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Yubing Tong
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Vibhu Agrawal
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Gargi Pednekar
- Quantitative Radiology SolutionsPhiladelphiaPennsylvaniaUSA
| | - Sanghita Nag
- Quantitative Radiology SolutionsPhiladelphiaPennsylvaniaUSA
| | - Tarun Kotia
- Quantitative Radiology SolutionsPhiladelphiaPennsylvaniaUSA
| | | | - E. Paul Wileyto
- Department of Biostatistics and EpidemiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Dimitris Mihailidis
- Department of Radiation OncologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - John Nicholas Lukens
- Department of Radiation OncologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Abigail T. Berman
- Department of Radiation OncologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Joann Stambaugh
- Department of Radiation OncologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Tristan Lim
- Department of Radiation OncologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Rupa Chowdary
- Department of MedicineUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Dheeraj Jalluri
- Department of MedicineUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Salma K. Jabbour
- Department of Radiation OncologyRutgers UniversityNew BrunswickNew JerseyUSA
| | - Sung Kim
- Department of Radiation OncologyRutgers UniversityNew BrunswickNew JerseyUSA
| | - Meral Reyhan
- Department of Radiation OncologyRutgers UniversityNew BrunswickNew JerseyUSA
| | | | - Wade L. Thorstad
- Department of Radiation OncologyWashington UniversitySt. LouisMissouriUSA
| | | | | | | | - Joe Camaratta
- Quantitative Radiology SolutionsPhiladelphiaPennsylvaniaUSA
| | - Steve Owens
- Quantitative Radiology SolutionsPhiladelphiaPennsylvaniaUSA
| | - Drew A. Torigian
- Medical Image Processing GroupDepartment of RadiologyUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| |
Collapse
|
6
|
Lang Y, Lian C, Xiao D, Deng H, Thung KH, Yuan P, Gateno J, Kuang T, Alfi DM, Wang L, Shen D, Xia JJ, Yap PT. Localization of Craniomaxillofacial Landmarks on CBCT Images Using 3D Mask R-CNN and Local Dependency Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:2856-2866. [PMID: 35544487 PMCID: PMC9673501 DOI: 10.1109/tmi.2022.3174513] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/09/2023]
Abstract
Cephalometric analysis relies on accurate detection of craniomaxillofacial (CMF) landmarks from cone-beam computed tomography (CBCT) images. However, due to the complexity of CMF bony structures, it is difficult to localize landmarks efficiently and accurately. In this paper, we propose a deep learning framework to tackle this challenge by jointly digitalizing 105 CMF landmarks on CBCT images. By explicitly learning the local geometrical relationships between the landmarks, our approach extends Mask R-CNN for end-to-end prediction of landmark locations. Specifically, we first apply a detection network on a down-sampled 3D image to leverage global contextual information to predict the approximate locations of the landmarks. We subsequently leverage local information provided by higher-resolution image patches to refine the landmark locations. On patients with varying non-syndromic jaw deformities, our method achieves an average detection accuracy of 1.38± 0.95mm, outperforming a related state-of-the-art method.
Collapse
|
7
|
Jiang J, Zhou H, Zhang T, Yao C, Du D, Zhao L, Cai W, Che L, Cao Z, Wu XE. Machine learning to predict dynamic changes of pathogenic Vibrio spp. abundance on microplastics in marine environment. ENVIRONMENTAL POLLUTION (BARKING, ESSEX : 1987) 2022; 305:119257. [PMID: 35398156 DOI: 10.1016/j.envpol.2022.119257] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/05/2021] [Revised: 03/14/2022] [Accepted: 04/01/2022] [Indexed: 05/27/2023]
Abstract
Microplastics are widely found in the marine environment. Recent studies have shown that pathogenic microorganisms can hitchhike on microplastics, which might act as a vector for the spread of pathogens. Vibrio spp. are known to be pathogenic to humans and can cause serious foodborne diseases. In this study, using datasets from an estuary and a mariculture zone in China, five machine learning models were established to predict the relative abundance of Vibrio spp. on microplastics. The results showed that deep neural network (DNN) model and RandomForest algorithm achieved the best predictive performance. Different data sources, data sampling, and processing methods had a little impact on the prediction performance of DNN and RandomForest models. SHapley Additive exPlanations (SHAP) indicated that salinity and temperature are the primary factors affecting the relative abundance of Vibrio spp. The prediction performances of the five machine learning models were further improved by feature selection, providing information to support future experimental research. The results of this study could help establish a long-term and dynamic monitoring system for the relative abundance of Vibrio spp. on microplastics in response to environmental factors as well as provide useful information for assessing the potential health impacts of microplastics on marine ecology and humans.
Collapse
Affiliation(s)
- Jiawen Jiang
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Hua Zhou
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Ting Zhang
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Chuanyi Yao
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Delin Du
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Liang Zhao
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Wenfang Cai
- School of Chemical Engineering and Technology, Xi'an Jiaotong University, Xi'an, 710049, China
| | - Liming Che
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Zhikai Cao
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China
| | - Xue E Wu
- Department of Chemical and Biochemical Engineering, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen, 361005, China.
| |
Collapse
|
8
|
Jin C, Udupa JK, Zhao L, Tong Y, Odhner D, Pednekar G, Nag S, Lewis S, Poole N, Mannikeri S, Govindasamy S, Singh A, Camaratta J, Owens S, Torigian DA. Object recognition in medical images via anatomy-guided deep learning. Med Image Anal 2022; 81:102527. [DOI: 10.1016/j.media.2022.102527] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2021] [Revised: 03/31/2022] [Accepted: 06/24/2022] [Indexed: 11/25/2022]
|
9
|
Iyer S, Blair A, Dawes L, Moses D, White C, Sowmya A. Supervised and semi-supervised 3D organ localisation in CT images combining reinforcement learning with imitation learning. Biomed Phys Eng Express 2022; 8. [PMID: 35385835 DOI: 10.1088/2057-1976/ac64c5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2022] [Accepted: 04/06/2022] [Indexed: 11/12/2022]
Abstract
Computer aided diagnostics often requires analysis of a region of interest (ROI) within a radiology scan, and the ROI may be an organ or a suborgan. Although deep learning algorithms have the ability to outperform other methods, they rely on the availability of a large amount of annotated data. Motivated by the need to address this limitation, an approach to localisation and detection of multiple organs based on supervised and semi-supervised learning is presented here. It draws upon previous work by the authors on localising the thoracic and lumbar spine region in CT images. The method generates six bounding boxes of organs of interest, which are then fused to a single bounding box. The results of experiments on localisation of the Spleen, Left and Right Kidneys in CT Images using supervised and semi supervised learning (SSL) demonstrate the ability to address data limitations with a much smaller data set and fewer annotations, compared to other state-of-the-art methods. The SSL performance was evaluated using three different mixes of labelled and unlabelled data (i.e. 30:70,35:65,40:60) for each of lumbar spine, spleen left and right kidneys respectively. The results indicate that SSL provides a workable alternative especially in medical imaging where it is difficult to obtain annotated data.
Collapse
Affiliation(s)
- Sankaran Iyer
- School of Computer Science and Engineering, The University of New South Wales, Australia
| | - Alan Blair
- School of Computer Science and Engineering, The University of New South Wales, Australia
| | - Laughlin Dawes
- Department of Medical Imaging, Prince of Wales Hospital, NSW, Australia
| | - Daniel Moses
- Department of Medical Imaging, Prince of Wales Hospital, NSW, Australia
| | - Christopher White
- Department of Endocrinology and Metabolism, Prince of Wales Hospital, NSW, Australia
| | - Arcot Sowmya
- School of Computer Science and Engineering, The University of New South Wales, Australia
| |
Collapse
|
10
|
Luu MH, Walsum TV, Mai HS, Franklin D, Nguyen TTT, Le TM, Moelker A, Le VK, Vu DL, Le NH, Tran QL, Chu DT, Trung NL. Automatic scan range for dose-reduced multiphase CT imaging of the liver utilizing CNNs and Gaussian models. Med Image Anal 2022; 78:102422. [PMID: 35339951 DOI: 10.1016/j.media.2022.102422] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2021] [Revised: 12/27/2021] [Accepted: 03/11/2022] [Indexed: 12/24/2022]
Abstract
Multiphase CT scanning of the liver is performed for several clinical applications; however, radiation exposure from CT scanning poses a nontrivial cancer risk to the patients. The radiation dose may be reduced by determining the scan range of the subsequent scans by the location of the target of interest in the first scan phase. The purpose of this study is to present and assess an automatic method for determining the scan range for multiphase CT scans. Our strategy is to first apply a CNN-based method for detecting the liver in 2D slices, and to use a liver range search algorithm for detecting the liver range in the scout volume. The target liver scan range for subsequent scans can be obtained by adding safety margins achieved from Gaussian liver motion models to the scan range determined from the scout. Experiments were performed on 657 multiphase CT volumes obtained from multiple hospitals. The experiment shows that the proposed liver detection method can detect the liver in 223 out of a total of 224 3D volumes on average within one second, with mean intersection of union, wall distance and centroid distance of 85.5%, 5.7 mm and 9.7 mm, respectively. In addition, the performance of the proposed liver detection method is comparable to the best of the state-of-the-art 3D liver detectors in the liver detection accuracy while it requires less processing time. Furthermore, we apply the liver scan range generation method on the liver CT images acquired from radiofrequency ablation and Y-90 transarterial radioembolization (selective internal radiation therapy) interventions of 46 patients from two hospitals. The result shows that the automatic scan range generation can significantly reduce the effective radiation dose by an average of 14.5% (2.56 mSv) compared to manual performance by the radiographer from Y-90 transarterial radioembolization, while no statistically significant difference in performance was found with the CT images from intra RFA intervention (p = 0.81). Finally, three radiologists assess both the original and the range-reduced images for evaluating the effect of the range reduction method on their clinical decisions. We conclude that the automatic liver scan range generation method is able to reduce excess radiation compared to the manual performance with a high accuracy and without penalizing the clinical decision.
Collapse
Affiliation(s)
- Manh Ha Luu
- AVITECH, University of Engineering and Technology, VNU, Hanoi, Vietnam; Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands; FET, University of Engineering and Technology, VNU, Hanoi, Vietnam.
| | - Theo van Walsum
- Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands
| | - Hong Son Mai
- Department of Nuclear Medicine, Hospital 108, Hanoi, Vietnam
| | - Daniel Franklin
- School of Electrical and Data Engineering, University of Technology Sydney, Sydney, Australia
| | | | - Thi My Le
- Department of Radiology and Nuclear Medicine, Vinmec Hospital, Hanoi, Vietnam
| | - Adriaan Moelker
- Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, the Netherlands
| | - Van Khang Le
- Radiology Center, Bach Mai Hospital, Hanoi, Vietnam
| | - Dang Luu Vu
- Radiology Center, Bach Mai Hospital, Hanoi, Vietnam
| | - Ngoc Ha Le
- Department of Nuclear Medicine, Hospital 108, Hanoi, Vietnam
| | - Quoc Long Tran
- FIT, University of Engineering and Technology, VNU, Hanoi, Vietnam
| | - Duc Trinh Chu
- FET, University of Engineering and Technology, VNU, Hanoi, Vietnam
| | - Nguyen Linh Trung
- AVITECH, University of Engineering and Technology, VNU, Hanoi, Vietnam
| |
Collapse
|
11
|
Ouyang Z, Zhang P, Pan W, Li Q. Deep learning-based body part recognition algorithm for three-dimensional medical images. Med Phys 2022; 49:3067-3079. [PMID: 35157332 DOI: 10.1002/mp.15536] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2021] [Revised: 01/24/2022] [Accepted: 01/25/2022] [Indexed: 11/12/2022] Open
Abstract
BACKGROUND The automatic recognition of human body parts in three-dimensional (3D) medical images is important in many clinical applications. However, methods presented in prior studies have mainly classified each two-dimensional (2D) slice independently rather than recognizing a batch of consecutive slices as a specific body part. PURPOSE In this study, we aim to develop a deep-learning-based method designed to automatically divide computed tomography (CT) and magnetic resonance imaging (MRI) scans into five consecutive body parts: head, neck, chest, abdomen, and pelvis. METHODS A deep learning framework was developed to recognize body parts in two stages. In the first pre-classification stage, a convolutional neural network (CNN) using the GoogLeNet Inception v3 architecture and a long short-term memory (LSTM) network were combined to classify each 2D slice; the CNN extracted information from a single slice, whereas the LSTM employed rich contextual information among consecutive slices. In the second post-processing stage, the input scan was further partitioned into consecutive body parts by identifying the optimal boundaries between them based on the slice classification results of the first stage. To evaluate the performance of the proposed method, 662 CT and 1434 MRI scans were used. RESULTS Our method achieved a very good performance in 2D slice classification compared with state-of-the-art methods, with overall classification accuracies of 97.3% and 98.2% for CT and MRI scans, respectively. Moreover, our method further divided whole scans into consecutive body parts with mean boundary errors of 8.9 mm and 3.5 mm for CT and MRI data, respectively. CONCLUSIONS The proposed method significantly improved the slice classification accuracy compared with state-of-the-art methods, and further accurately divided CT and MRI scans into consecutive body parts based on the results of slice classification. The developed method can be employed as an important step in various computer-aided diagnosis and medical image analysis schemes. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Zihui Ouyang
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China.,MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| | - Peng Zhang
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China.,MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| | - Weifan Pan
- Zhejiang Taimei Medical Technology Co., Ltd, Jiaxing, Zhejiang, 314001, China
| | - Qiang Li
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China.,MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| |
Collapse
|
12
|
Liu T, Tan M, Tong Y, Torigian DA, Udupa JK. An Anatomy-based Iteratively Searching Convolutional Neural Network for Organ Localization in CT images. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2022; 12032:1203227. [PMID: 38098767 PMCID: PMC10720955 DOI: 10.1117/12.2610963] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/17/2023]
Abstract
Organ localization is a common and essential preprocessing operation for many medical image analysis tasks. We propose a novel multi-organ localization method based on an end-to-end 3D convolutional neural network. The proposed algorithm employs a regression network to learn the position relationship between any patch and target organs in a medical computed tomography (CT) image. With this framework, it can iteratively localize the target organs in a coarse-to-fine manner. The main idea behind this method is to embed the anatomy of structures in a deep learning-based approach. For implementation, the proposed network outputs an 8-dimensional vector that contains information about the position, scale, and presence of each target organ. A piecewise loss function and a multi-density sampling strategy help to optimize this network to learn anatomy layout characteristics over the entire CT image. Starting from a random position, this network can accurately locate the target organ with a few iterations. Moreover, a dual-resolution strategy is employed to improve the accuracy affected by varying organ scales, further enhancing the localizing performance for all organs. We evaluate our method on a public data set (LiTS) to locate 11 organs in the thoraco-abdomino-pelvic region. The proposed method outperforms state-of-the-art methods with a mean intersection over union (IOU) of 80.84%, mean wall distance of 3.63 mm, and mean centroid distance of 4.93 mm, constituting excellent accuracy. The improvements on relatively small-size and medium-size organs are noteworthy.
Collapse
Affiliation(s)
- Tiange Liu
- School of Information Science and Engineering, Yanshan University, Qinhuangdao, Hebei 066004, China
- Medical image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Meng Tan
- School of Information Science and Engineering, Yanshan University, Qinhuangdao, Hebei 066004, China
| | - Yubing Tong
- Medical image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Drew A. Torigian
- Medical image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Jayaram K. Udupa
- Medical image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA 19104, USA
| |
Collapse
|
13
|
Chen X, Lian C, Deng HH, Kuang T, Lin HY, Xiao D, Gateno J, Shen D, Xia JJ, Yap PT. Fast and Accurate Craniomaxillofacial Landmark Detection via 3D Faster R-CNN. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3867-3878. [PMID: 34310293 PMCID: PMC8686670 DOI: 10.1109/tmi.2021.3099509] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/09/2023]
Abstract
Automatic craniomaxillofacial (CMF) landmark localization from cone-beam computed tomography (CBCT) images is challenging, considering that 1) the number of landmarks in the images may change due to varying deformities and traumatic defects, and 2) the CBCT images used in clinical practice are typically large. In this paper, we propose a two-stage, coarse-to-fine deep learning method to tackle these challenges with both speed and accuracy in mind. Specifically, we first use a 3D faster R-CNN to roughly locate landmarks in down-sampled CBCT images that have varying numbers of landmarks. By converting the landmark point detection problem to a generic object detection problem, our 3D faster R-CNN is formulated to detect virtual, fixed-size objects in small boxes with centers indicating the approximate locations of the landmarks. Based on the rough landmark locations, we then crop 3D patches from the high-resolution images and send them to a multi-scale UNet for the regression of heatmaps, from which the refined landmark locations are finally derived. We evaluated the proposed approach by detecting up to 18 landmarks on a real clinical dataset of CMF CBCT images with various conditions. Experiments show that our approach achieves state-of-the-art accuracy of 0.89 ± 0.64mm in an average time of 26.2 seconds per volume.
Collapse
|
14
|
Linares Rosales HM, Couture G, Archambault L, Beddar S, Després P, Beaulieu L. On the use of machine learning methods for mPSD calibration in HDR brachytherapy. Phys Med 2021; 91:73-79. [PMID: 34717139 DOI: 10.1016/j.ejmp.2021.10.003] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/11/2021] [Revised: 07/15/2021] [Accepted: 10/02/2021] [Indexed: 11/17/2022] Open
Abstract
We sought to evaluate the feasibility of using machine learning (ML) algorithms for multipoint plastic scintillator detector (mPSD) calibration in high-dose-rate (HDR) brachytherapy. Dose measurements were conducted under HDR brachytherapy conditions. The dosimetry system consisted of an optimized 1-mm-core mPSD and a compact assembly of photomultiplier tubes coupled with dichroic mirrors and filters. An 192Ir source was remotely controlled and sent to various positions in a homemade PMMA holder, ensuring 0.1-mm positional accuracy. Dose measurements covering a range of 0.5 to 12 cm of source displacement were carried out according to TG-43 U1 recommendations. Individual scintillator doses were decoupled using a linear regression model, a random forest estimator, and artificial neural network algorithms. The dose predicted by the TG-43U1 formalism was used as the reference for system calibration and ML algorithm training. The performance of the different algorithms was evaluated using different sample sizes and distances to the source for the mPSD system calibration. We found that the calibration conditions influenced the accuracy in predicting the measured dose. The decoupling methods' deviations from the expected TG-43 U1 dose generally remained below 20%. However, the dose prediction with the three algorithms was accurate to within 7% relative to the dose predicted by the TG-43 U1 formalism when measurements were performed in the same range of distances used for calibration. In such cases, the predictions with random forest exhibited minimal deviations (<2%). However, the performance random forest was compromised when the predictions were done beyond the range of distances used for calibration. Because the linear regression algorithm can extrapolate the data, the dose prediction by the linear regression was less influenced by the calibration conditions than random forest. The linear regression algorithm's behavior along the distances to the source was smoother than those for the random forest and neural network algorithms, but the observed deviations were more significant than those for the neural network and random forest algorithms. The number of available measurements for training purposes influenced the random forest and neural network models the most. Their accuracy tended to converge toward deviation values close to 1% from a number of dwell positions greater than 100. In performing HDR brachytherapy dose measurements with an optimized mPSD system, ML algorithms are good alternatives for precise dose reporting and treatment assessment during this kind of cancer treatment.
Collapse
Affiliation(s)
- Haydee M Linares Rosales
- Département de physique, de génie physique et d'optique et Centre de recherche sur le cancer, Université Laval, Québec, Canada; Département de radio-oncologie et Axe Oncologie du CRCHU de Québec, CHU de Québec - Université Laval, QC, Canada.
| | - Gabriel Couture
- Département de physique, de génie physique et d'optique et Centre de recherche sur le cancer, Université Laval, Québec, Canada; Département de radio-oncologie et Axe Oncologie du CRCHU de Québec, CHU de Québec - Université Laval, QC, Canada
| | - Louis Archambault
- Département de physique, de génie physique et d'optique et Centre de recherche sur le cancer, Université Laval, Québec, Canada; Département de radio-oncologie et Axe Oncologie du CRCHU de Québec, CHU de Québec - Université Laval, QC, Canada
| | - Sam Beddar
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX, United States; The University of Texas MD Anderson UTHealth Graduate School of Biomedical Sciences, Houston, TX, United States
| | - Philippe Després
- Département de physique, de génie physique et d'optique et Centre de recherche sur le cancer, Université Laval, Québec, Canada; Département de radio-oncologie et Axe Oncologie du CRCHU de Québec, CHU de Québec - Université Laval, QC, Canada
| | - Luc Beaulieu
- Département de physique, de génie physique et d'optique et Centre de recherche sur le cancer, Université Laval, Québec, Canada; Département de radio-oncologie et Axe Oncologie du CRCHU de Québec, CHU de Québec - Université Laval, QC, Canada
| |
Collapse
|
15
|
Amniotic fluid segmentation based on pixel classification using local window information and distance angle pixel. Appl Soft Comput 2021. [DOI: 10.1016/j.asoc.2021.107196] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
|
16
|
Hussain MA, Hamarneh G, Garbi R. Cascaded Regression Neural Nets for Kidney Localization and Segmentation-free Volume Estimation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1555-1567. [PMID: 33606626 DOI: 10.1109/tmi.2021.3060465] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Kidney volume is an essential biomarker for a number of kidney disease diagnoses, for example, chronic kidney disease. Existing total kidney volume estimation methods often rely on an intermediate kidney segmentation step. On the other hand, automatic kidney localization in volumetric medical images is a critical step that often precedes subsequent data processing and analysis. Most current approaches perform kidney localization via an intermediate classification or regression step. This paper proposes an integrated deep learning approach for (i) kidney localization in computed tomography scans and (ii) segmentation-free renal volume estimation. Our localization method uses a selection-convolutional neural network that approximates the kidney inferior-superior span along the axial direction. Cross-sectional (2D) slices from the estimated span are subsequently used in a combined sagittal-axial Mask-RCNN that detects the organ bounding boxes on the axial and sagittal slices, the combination of which produces a final 3D organ bounding box. Furthermore, we use a fully convolutional network to estimate the kidney volume that skips the segmentation procedure. We also present a mathematical expression to approximate the 'volume error' metric from the 'Sørensen-Dice coefficient.' We accessed 100 patients' CT scans from the Vancouver General Hospital records and obtained 210 patients' CT scans from the 2019 Kidney Tumor Segmentation Challenge database to validate our method. Our method produces a kidney boundary wall localization error of ~2.4mm and a mean volume estimation error of ~5%.
Collapse
|
17
|
Zhang L, Wang H. A novel segmentation method for cervical vertebrae based on PointNet++ and converge segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2021; 200:105798. [PMID: 33545639 DOI: 10.1016/j.cmpb.2020.105798] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/09/2020] [Accepted: 10/10/2020] [Indexed: 06/12/2023]
Abstract
BACKGROUND Cervical spine instability is the key pathogenic factor for cervical spondylosis, which may easily cause cervical spinal cord nerve compression, numbness, weakness, and even paralysis of the limbs. The reconstruction of the internal fixation of the cervical spine is of great therapeutic significance, but is a high-risk and difficult procedure that requires precise planning. The high similarities between vertebrae may interfere with automatic operation planning; therefore, the segmentation of vertebrae is of great significance. METHODS Our segmentation algorithm has 3 parts. Firstly, an adaptive threshold filter to segment the cervical vertebra tissue structure form CT images. Secondly, segmentation of single vertebrae based on PointNet++ is introduced to segmentation cervical spine. Finally, converge segmentation which is based on edge information is utilized to clearly distinguish the edges of the two vertebrae to enhance the accuracy segmentation result. RESULTS Our approach improved the accuracy of the system up to 96.15%, and achieved the highest reported average score based on this dataset. We compared the results of the CNN and PointNet methods on a separate dataset of 240 CT scans with 18 classes and achieved a significantly higher performance for any given vertebra. Our experiments illustrated the promise and robustness of recent PointNet++-based segmentation of medical images. CONCLUSION The proposed method has better classification performance for segmentation cervical spine images, which segment a three-dimensional vertebral body directly and effectively. Furthermore, the precise segmentation of a single vertebral body can be used in automatic biomechanical analysis, computer-aided diagnosis and other aspects, so as to improve the level of automation in the treatment of cervical spondylosis.
Collapse
Affiliation(s)
- Lei Zhang
- Spine Surgery Unit, Shengjing Hospital of China Medical University, Shenyang, 110004 P.R.China
| | - Huan Wang
- Spine Surgery Unit, Shengjing Hospital of China Medical University; Address: No.36 Sanhao Street, Heping District, Shenyang, 110004, Liaoning Province, P.R.China.
| |
Collapse
|
18
|
He J, Zhang H, Wang X, Sun Z, Ge Y, Wang K, Yu C, Deng Z, Feng J, Xu X, Hu S. A pilot study of radiomics signature based on biparametric MRI for preoperative prediction of extrathyroidal extension in papillary thyroid carcinoma. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2021; 29:171-183. [PMID: 33325448 DOI: 10.3233/xst-200760] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
OBJECTIVE To investigate efficiency of radiomics signature to preoperatively predict histological features of aggressive extrathyroidal extension (ETE) in papillary thyroid carcinoma (PTC) with biparametric magnetic resonance imaging findings. MATERIALS AND METHODS Sixty PTC patients with preoperative MR including T2WI and T2WI-fat-suppression (T2WI-FS) were retrospectively analyzed. Among them, 35 had ETE and 25 did not. Pre-contrast T2WI and T2WI-FS images depicting the largest section of tumor were selected. Tumor regions were manually segmented using ITK-SNAP software and 107 radiomics features were computed from the segmented regions using the open Pyradiomics package. Then, a random forest model was built to do classification in which the datasets were partitioned randomly 10 times to do training and testing with ratio of 1:1. Furthermore, forward greedy feature selection based on feature importance was adopted to reduce model overfitting. Classification accuracy was estimated on the test set using area under ROC curve (AUC). RESULTS The model using T2WI-FS image features yields much higher performance than the model using T2WI features (AUC = 0.906 vs. 0.760 using 107 features). Among the top 10 important features of T2WI and T2WI-FS, there are 5 common features. After feature selection, the models trained using top 2 features of T2WI and the top 6 features of T2WI-FS achieve AUC 0.845 and 0.928, respectively. Combining features computed from T2WI and T2WI-FS, model performance decreases slightly (AUC = 0.882 based on all features and AUC = 0.913 based on top features after feature selection). Adjusting hyper parameters of the random forest model have negligible influence on the model performance with mean AUC = 0.907 for T2WI-FS images. CONCLUSIONS Radiomics features based on pre-contrast T2WI and T2WI-FS is helpful to predict aggressive ETE in PTC. Particularly, the model trained using the optimally selected T2WI-FS image features yields the best classification performance. The most important features relate to lesion size and the texture heterogeneity of the tumor region.
Collapse
Affiliation(s)
- Junlin He
- School of Medicine, Jiangsu University, Zhenjiang, Jiangsu, China
| | - Heng Zhang
- Department of Radiology, Affiliated Hospital, Jiangnan University, Huihe Road, Wuxi, Jiangsu, China
| | - Xian Wang
- Department of Radiology, Affiliated Renmin Hospital, Jiangsu University, Dianli Road, Zhenjiang, Jiangsu, China
| | - Zongqiong Sun
- Department of Radiology, Affiliated Hospital, Jiangnan University, Huihe Road, Wuxi, Jiangsu, China
| | - Yuxi Ge
- Department of Radiology, Affiliated Hospital, Jiangnan University, Huihe Road, Wuxi, Jiangsu, China
| | - Kang Wang
- Department of Radiology, Affiliated Hospital, Jiangnan University, Huihe Road, Wuxi, Jiangsu, China
| | - Chunjing Yu
- Department of Nuclear Medicine, Affiliated Hospital of Jiangnan University, Wuxi, China
| | - Zhaohong Deng
- School of Digital Media, Jiangnan University and Jiangsu Key Laboratory of Digital Design and Software Technology, Digital Media Academy, Jiangnan University, China
| | | | - Xin Xu
- Haohua Technology Co., Ltd, Shanghai, China
| | - Shudong Hu
- Department of Radiology, Affiliated Hospital, Jiangnan University, Huihe Road, Wuxi, Jiangsu, China
- Department of Radiology, Affiliated Renmin Hospital, Jiangsu University, Dianli Road, Zhenjiang, Jiangsu, China
| |
Collapse
|
19
|
Torrents-Barrena J, Piella G, Gratacos E, Eixarch E, Ceresa M, Gonalez Ballester MA. Deep Q-CapsNet Reinforcement Learning Framework for Intrauterine Cavity Segmentation in TTTS Fetal Surgery Planning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3113-3124. [PMID: 32305906 DOI: 10.1109/tmi.2020.2987981] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Fetoscopic laser photocoagulation is the most effective treatment for Twin-to-Twin Transfusion Syndrome, a condition affecting twin pregnancies in which there is a deregulation of blood circulation through the placenta, that can be fatal to both babies. For the purposes of surgical planning, we design the first automatic approach to detect and segment the intrauterine cavity from axial, sagittal and coronal MRI stacks. Our methodology relies on the ability of capsule networks to successfully capture the part-whole interdependency of objects in the scene, particularly for unique class instances (i.e., intrauterine cavity). The presented deep Q-CapsNet reinforcement learning framework is built upon a context-adaptive detection policy to generate a bounding box of the womb. A capsule architecture is subsequently designed to segment (or refine) the whole intrauterine cavity. This network is coupled with a strided nnU-Net feature extractor, which encodes discriminative feature maps to construct strong primary capsules. The method is robustly evaluated with and without the localization stage using 13 performance measures, and directly compared with 15 state-of-the-art deep neural networks trained on 71 singleton and monochorionic twin pregnancies. An average Dice score above 0.91 is achieved for all ablations, revealing the potential of our approach to be used in clinical practice.
Collapse
|
20
|
Agrawal V, Udupa J, Tong Y, Torigian D. BRR-Net: A tandem architectural CNN-RNN for automatic body region localization in CT images. Med Phys 2020; 47:5020-5031. [PMID: 32761899 DOI: 10.1002/mp.14439] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2020] [Revised: 06/22/2020] [Accepted: 07/22/2020] [Indexed: 12/25/2022] Open
Abstract
PURPOSE Automatic identification of consistently defined body regions in medical images is vital in many applications. In this paper, we describe a method to automatically demarcate the superior and inferior boundaries for neck, thorax, abdomen, and pelvis body regions in computed tomography (CT) images. METHODS For any three-dimensional (3D) CT image I, following precise anatomic definitions, we denote the superior and inferior axial boundary slices of the neck, thorax, abdomen, and pelvis body regions by NS(I), NI(I), TS(I), TI(I), AS(I), AI(I), PS(I), and PI(I), respectively. Of these, by definition, AI(I) = PS(I), and so the problem reduces to demarcating seven body region boundaries. Our method consists of a two-step approach. In the first step, a convolutional neural network (CNN) is trained to classify each axial slice in I into one of nine categories: the seven body region boundaries, plus legs (defined as all axial slices inferior to PI(I)), and the none-of-the-above category. This CNN uses a multichannel approach to exploit the interslice contrast, providing the neural network with additional visual context at the body region boundaries. In the second step, to improve the predictions for body region boundaries that are very subtle and that exhibit low contrast, a recurrent neural network (RNN) is trained on features extracted by CNN, limited to a flexible window about the predictions from the CNN. RESULTS The method is evaluated on low-dose CT images from 442 patient scans, divided into training and testing sets with a ratio of 70:30. Using only the CNN, overall absolute localization error for NS(I), NI(I), TS(I), TI(I), AS(I), AI(I), and PI(I) expressed in terms of number of slices (nS) is (mean ± SD): 0.61 ± 0.58, 1.05 ± 1.13, 0.31 ± 0.46, 1.85 ± 1.96, 0.57 ± 2.44, 3.42 ± 3.16, and 0.50 ± 0.50, respectively. Using the RNN to refine the CNN's predictions for select classes improved the accuracy of TI(I) and AI(I) to: 1.35 ± 1.71 and 2.83 ± 2.75, respectively. This model outperforms the results achieved in our previous work by 2.4, 1.7, 3.1, 1.1, and 2 slices, respectively for TS(I), TI(I), AS(I), AI(I) = PS(I), and PI(I) classes with statistical significance. The model trained on low-dose CT images was also tested on diagnostic CT images for NS(I), NI(I), and TS(I) classes; the resulting errors were: 1.48 ± 1.33, 2.56 ± 2.05, and 0.58 ± 0.71, respectively. CONCLUSIONS Standardized body region definitions are a prerequisite for effective implementation of quantitative radiology, but the literature is severely lacking in the precise identification of body regions. The method presented in this paper significantly outperforms earlier works by a large margin, and the deviations of our results from ground truth are comparable to variations observed in manual labeling by experts. The solution presented in this work is critical to the adoption and employment of the idea of standardized body regions, and clears the path for development of applications requiring accurate demarcations of body regions. The work is indispensable for automatic anatomy recognition, delineation, and contouring for radiation therapy planning, as it not only automates an essential part of the process, but also removes the dependency on experts for accurately demarcating body regions in a study.
Collapse
Affiliation(s)
- Vibhu Agrawal
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Jayaram Udupa
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Yubing Tong
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Drew Torigian
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| |
Collapse
|
21
|
Augmented reality for inner ear procedures: visualization of the cochlear central axis in microscopic videos. Int J Comput Assist Radiol Surg 2020; 15:1703-1711. [PMID: 32737858 DOI: 10.1007/s11548-020-02240-w] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2019] [Accepted: 07/20/2020] [Indexed: 10/23/2022]
Abstract
PURPOSE Visualization of the cochlea is impossible due to the delicate and intricate ear anatomy. Augmented reality may be used to perform auditory nerve implantation by transmodiolar approach in patients with profound hearing loss. METHODS We present an augmented reality system for the visualization of the cochlear axis in surgical videos. The system starts with an automatic anatomical landmark detection in preoperative computed tomography images based on deep reinforcement learning. These landmarks are used to register the preoperative geometry with the real-time microscopic video captured inside the auditory canal. Three-dimensional pose of the cochlear axis is determined using the registration projection matrices. In addition, the patient microscope movements are tracked using an image feature-based tracking process. RESULTS The landmark detection stage yielded an average localization error of [Formula: see text] mm ([Formula: see text]). The target registration error was [Formula: see text] mm for the cochlear apex and [Formula: see text] for the cochlear axis. CONCLUSION We developed an augmented reality system to visualize the cochlear axis in intraoperative videos. The system yielded millimetric accuracy and remained stable throughout the experimental study despite camera movements throughout the procedure in experimental conditions.
Collapse
|
22
|
Peña-Solórzano CA, Albrecht DW, Bassed RB, Gillam J, Harris PC, Dimmock MR. Semi-supervised labelling of the femur in a whole-body post-mortem CT database using deep learning. Comput Biol Med 2020; 122:103797. [PMID: 32658723 DOI: 10.1016/j.compbiomed.2020.103797] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2020] [Revised: 04/29/2020] [Accepted: 04/29/2020] [Indexed: 01/16/2023]
Abstract
A deep learning pipeline was developed and used to localize and classify a variety of implants in the femur contained in whole-body post-mortem computed tomography (PMCT) scans. The results provide a proof-of-principle approach for labelling content not described in medical/autopsy reports. The pipeline, which incorporated residual networks and an autoencoder, was trained and tested using n = 450 full-body PMCT scans. For the localization component, Dice scores of 0.99, 0.96, and 0.98 and mean absolute errors of 3.2, 7.1, and 4.2 mm were obtained in the axial, coronal, and sagittal views, respectively. A regression analysis found the orientation of the implant to the scanner axis and also the relative positioning of extremities to be statistically significant factors. For the classification component, test cases were properly labelled as nail (N+), hip replacement (H+), knee replacement (K+) or without-implant (I-) with an accuracy >97%. The recall for I- and H+ cases was 1.00, but fell to 0.82 and 0.65 for cases with K+ and N+. This semi-automatic approach provides a generalized structure for image-based labelling of features, without requiring time-consuming segmentation.
Collapse
Affiliation(s)
- C A Peña-Solórzano
- Department of Medical Imaging and Radiation Sciences, Monash University, Wellington Rd, Clayton, Melbourne, VIC, 3800, Australia.
| | - D W Albrecht
- Clayton School of Information Technology, Monash University, Wellington Rd, Clayton, Melbourne, VIC, 3800, Australia.
| | - R B Bassed
- Victorian Institute of Forensic Medicine, 57-83 Kavanagh St., Southbank, Melbourne, VIC, 3006, Australia; Department of Forensic Medicine, Monash University, Wellington Rd, Clayton, Melbourne, VIC, 3800, Australia.
| | - J Gillam
- Land Division, Defence Science and Technology Group, Fishermans Bend, Melbourne, VIC, 3207, Australia.
| | - P C Harris
- The Royal Children's Hospital Melbourne, 50 Flemington Road, Parkville, Melbourne, VIC, 3052, Australia; Department of Orthopaedic Surgery, Western Health, Footscray Hospital, Gordon St, Footscray, Melbourne, VIC, 3011, Australia.
| | - M R Dimmock
- Department of Medical Imaging and Radiation Sciences, Monash University, Wellington Rd, Clayton, Melbourne, VIC, 3800, Australia.
| |
Collapse
|
23
|
Wang X, Zhai S, Niu Y. Left ventricle landmark localization and identification in cardiac MRI by deep metric learning-assisted CNN regression. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.02.069] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
24
|
Zhang D, Wang J, Noble JH, Dawant BM. HeadLocNet: Deep convolutional neural networks for accurate classification and multi-landmark localization of head CTs. Med Image Anal 2020; 61:101659. [PMID: 32062157 PMCID: PMC7959656 DOI: 10.1016/j.media.2020.101659] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2019] [Revised: 01/22/2020] [Accepted: 01/23/2020] [Indexed: 01/19/2023]
Abstract
Cochlear implants (CIs) are used to treat subjects with hearing loss. In a CI surgery, an electrode array is inserted into the cochlea to stimulate auditory nerves. After surgery, CIs need to be programmed. Studies have shown that the cochlea-electrode spatial relationship derived from medical images can guide CI programming and lead to significant improvement in hearing outcomes. We have developed a series of algorithms to segment the inner ear anatomy and localize the electrodes. But, because clinical head CT images are acquired with different protocols, the field of view and orientation of the image volumes vary greatly. As a consequence, visual inspection and manual image registration to an atlas image are needed to document their content and to initialize intensity-based registration algorithms used in our processing pipeline. For large-scale evaluation and deployment of our methods these steps need to be automated. In this article we propose to achieve this with a deep convolutional neural network (CNN) that can be trained end-to-end to classify a head CT image in terms of its content and to localize landmarks. The detected landmarks can then be used to estimate a point-based registration with the atlas image in which the same landmark set's positions are known. We achieve 99.5% classification accuracy and an average localization error of 3.45 mm for 7 landmarks located around each inner ear. This is better than what was achieved with earlier methods we have proposed for the same tasks.
Collapse
Affiliation(s)
- Dongqing Zhang
- Department of Electrical Engineering and Computer Science, Vanderbilt University, Nashville, TN, 37235, USA.
| | - Jianing Wang
- Department of Electrical Engineering and Computer Science, Vanderbilt University, Nashville, TN, 37235, USA
| | - Jack H Noble
- Department of Electrical Engineering and Computer Science, Vanderbilt University, Nashville, TN, 37235, USA
| | - Benoit M Dawant
- Department of Electrical Engineering and Computer Science, Vanderbilt University, Nashville, TN, 37235, USA.
| |
Collapse
|
25
|
Abdullah Al W, Yun ID. Partial Policy-Based Reinforcement Learning for Anatomical Landmark Localization in 3D Medical Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:1245-1255. [PMID: 31603816 DOI: 10.1109/tmi.2019.2946345] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
Abstract
Utilizing the idea of long-term cumulative return, reinforcement learning (RL) has shown remarkable performance in various fields. We follow the formulation of landmark localization in 3D medical images as an RL problem. Whereas value-based methods have been widely used to solve RL-based localization problems, we adopt an actor-critic based direct policy search method framed in a temporal difference learning approach. In RL problems with large state and/or action spaces, learning the optimal behavior is challenging and requires many trials. To improve the learning, we introduce a partial policy-based reinforcement learning to enable solving the large problem of localization by learning the optimal policy on smaller partial domains. Independent actors efficiently learn the corresponding partial policies, each utilizing their own independent critic. The proposed policy reconstruction from the partial policies ensures a robust and efficient localization, where the sub-agents uniformly contribute to the state-transitions based on their simple partial policies mapping to binary actions. Experiments with three different localization problems in 3D CT and MR images showed that the proposed reinforcement learning requires a significantly smaller number of trials to learn the optimal behavior compared to the original behavior learning scheme in RL. It also ensures a satisfactory performance when trained on fewer images.
Collapse
|
26
|
Mlynarski P, Delingette H, Alghamdi H, Bondiau PY, Ayache N. Anatomically consistent CNN-based segmentation of organs-at-risk in cranial radiotherapy. J Med Imaging (Bellingham) 2020; 7:014502. [PMID: 32064300 PMCID: PMC7016364 DOI: 10.1117/1.jmi.7.1.014502] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2019] [Accepted: 01/17/2020] [Indexed: 11/14/2022] Open
Abstract
Planning of radiotherapy involves accurate segmentation of a large number of organs at risk (OAR), i.e., organs for which irradiation doses should be minimized to avoid important side effects of the therapy. We propose a deep learning method for segmentation of OAR inside the head, from magnetic resonance images (MRIs). Our system performs segmentation of eight structures: eye, lens, optic nerve, optic chiasm, pituitary gland, hippocampus, brainstem, and brain. We propose an efficient algorithm to train neural networks for an end-to-end segmentation of multiple and nonexclusive classes, addressing problems related to computational costs and missing ground truth segmentations for a subset of classes. We enforce anatomical consistency of the result in a postprocessing step. In particular, we introduce a graph-based algorithm for segmentation of the optic nerves, enforcing the connectivity between the eyes and the optic chiasm. We report cross-validated quantitative results on a database of 44 contrast-enhanced T1-weighted MRIs with provided segmentations of the considered OAR, which were originally used for radiotherapy planning. In addition, the segmentations produced by our model on an independent test set of 50 MRIs were evaluated by an experienced radiotherapist in order to qualitatively assess their accuracy. The mean distances between produced segmentations and the ground truth ranged from 0.1 to 0.7 mm across different organs. A vast majority (96%) of the produced segmentations were found acceptable for radiotherapy planning.
Collapse
Affiliation(s)
- Pawel Mlynarski
- Université Côte d’Azur, Inria, Epione Research Team, Nice, France
| | - Hervé Delingette
- Université Côte d’Azur, Inria, Epione Research Team, Nice, France
| | - Hamza Alghamdi
- Université Côte d’Azur, Centre Antoine Lacassagne, Nice, France
| | | | - Nicholas Ayache
- Université Côte d’Azur, Inria, Epione Research Team, Nice, France
| |
Collapse
|
27
|
Zhang J, Liu M, Wang L, Chen S, Yuan P, Li J, Shen SGF, Tang Z, Chen KC, Xia JJ, Shen D. Context-guided fully convolutional networks for joint craniomaxillofacial bone segmentation and landmark digitization. Med Image Anal 2019; 60:101621. [PMID: 31816592 DOI: 10.1016/j.media.2019.101621] [Citation(s) in RCA: 53] [Impact Index Per Article: 10.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2018] [Revised: 07/01/2019] [Accepted: 11/19/2019] [Indexed: 12/24/2022]
Abstract
Cone-beam computed tomography (CBCT) scans are commonly used in diagnosing and planning surgical or orthodontic treatment to correct craniomaxillofacial (CMF) deformities. Based on CBCT images, it is clinically essential to generate an accurate 3D model of CMF structures (e.g., midface, and mandible) and digitize anatomical landmarks. This process often involves two tasks, i.e., bone segmentation and anatomical landmark digitization. Because landmarks usually lie on the boundaries of segmented bone regions, the tasks of bone segmentation and landmark digitization could be highly associated. Also, the spatial context information (e.g., displacements from voxels to landmarks) in CBCT images is intuitively important for accurately indicating the spatial association between voxels and landmarks. However, most of the existing studies simply treat bone segmentation and landmark digitization as two standalone tasks without considering their inherent relationship, and rarely take advantage of the spatial context information contained in CBCT images. To address these issues, we propose a Joint bone Segmentation and landmark Digitization (JSD) framework via context-guided fully convolutional networks (FCNs). Specifically, we first utilize displacement maps to model the spatial context information in CBCT images, where each element in the displacement map denotes the displacement from a voxel to a particular landmark. An FCN is learned to construct the mapping from the input image to its corresponding displacement maps. Using the learned displacement maps as guidance, we further develop a multi-task FCN model to perform bone segmentation and landmark digitization jointly. We validate the proposed JSD method on 107 subjects, and the experimental results demonstrate that our method is superior to the state-of-the-art approaches in both tasks of bone segmentation and landmark digitization.
Collapse
Affiliation(s)
- Jun Zhang
- Department of Radiology and BRIC, University of North Carolina, Chapel Hill, NC 27599, USA.
| | - Mingxia Liu
- Department of Radiology and BRIC, University of North Carolina, Chapel Hill, NC 27599, USA.
| | - Li Wang
- Department of Radiology and BRIC, University of North Carolina, Chapel Hill, NC 27599, USA.
| | - Si Chen
- Department of Orthodontics, Peking University School and Hospital of Stomatology, Beijing 100191, China
| | - Peng Yuan
- Surgical Planning Laboratory, Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, Houston, TX 77030, USA
| | - Jianfu Li
- Surgical Planning Laboratory, Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, Houston, TX 77030, USA
| | - Steve Guo-Fang Shen
- Surgical Planning Laboratory, Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, Houston, TX 77030, USA
| | - Zhen Tang
- Surgical Planning Laboratory, Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, Houston, TX 77030, USA
| | - Ken-Chung Chen
- Surgical Planning Laboratory, Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, Houston, TX 77030, USA
| | - James J Xia
- Surgical Planning Laboratory, Department of Oral and Maxillofacial Surgery, Houston Methodist Research Institute, Houston, TX 77030, USA.
| | - Dinggang Shen
- Department of Radiology and BRIC, University of North Carolina, Chapel Hill, NC 27599, USA; Department of Brain and Cognitive Engineering, Korea University, Seoul 02841, Republic of Korea.
| |
Collapse
|
28
|
Cerrolaza JJ, Picazo ML, Humbert L, Sato Y, Rueckert D, Ballester MÁG, Linguraru MG. Computational anatomy for multi-organ analysis in medical imaging: A review. Med Image Anal 2019; 56:44-67. [DOI: 10.1016/j.media.2019.04.002] [Citation(s) in RCA: 28] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2018] [Revised: 02/05/2019] [Accepted: 04/13/2019] [Indexed: 12/19/2022]
|
29
|
Štern D, Payer C, Urschler M. Automated age estimation from MRI volumes of the hand. Med Image Anal 2019; 58:101538. [PMID: 31400620 DOI: 10.1016/j.media.2019.101538] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2018] [Revised: 02/21/2019] [Indexed: 10/26/2022]
Abstract
Highly relevant for both clinical and legal medicine applications, the established radiological methods for estimating unknown age in children and adolescents are based on visual examination of bone ossification in X-ray images of the hand. Our group has initiated the development of fully automatic age estimation methods from 3D MRI scans of the hand, in order to simultaneously overcome the problems of the radiological methods including (1) exposure to ionizing radiation, (2) necessity to define new, MRI specific staging systems, and (3) subjective influence of the examiner. The present work provides a theoretical background for understanding the nonlinear regression problem of biological age estimation and chronological age approximation. Based on this theoretical background, we comprehensively evaluate machine learning methods (random forests, deep convolutional neural networks) with different simplifications of the image information used as an input for learning. Trained on a large dataset of 328 MR images, we compare the performance of the different input strategies and demonstrate unprecedented results. For estimating biological age, we obtain a mean absolute error of 0.37 ± 0.51 years for the age range of the subjects ≤ 18 years, i.e. where bone ossification has not yet saturated. Finally, we validate our findings by adapting our best performing method to 2D images and applying it to a publicly available dataset of X-ray images, showing that we are in line with the state-of-the-art automatic methods for this task.
Collapse
Affiliation(s)
- Darko Štern
- Ludwig Boltzmann Institute for Clinical Forensic Imaging, Graz, Austria; BioTechMed-Graz, Medical University Graz, Graz, Austria
| | - Christian Payer
- Ludwig Boltzmann Institute for Clinical Forensic Imaging, Graz, Austria; Institute of Computer Graphics and Vision, Graz University of Technology, Graz, Austria
| | - Martin Urschler
- Ludwig Boltzmann Institute for Clinical Forensic Imaging, Graz, Austria; School of Computer Science, The University of Auckland, Auckland, New Zealand.
| |
Collapse
|
30
|
Zhao Y, Li H, Wan S, Sekuboyina A, Hu X, Tetteh G, Piraud M, Menze B. Knowledge-Aided Convolutional Neural Network for Small Organ Segmentation. IEEE J Biomed Health Inform 2019; 23:1363-1373. [DOI: 10.1109/jbhi.2019.2891526] [Citation(s) in RCA: 136] [Impact Index Per Article: 27.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
31
|
Rafiei S, Karimi N, Mirmahboub B, Najarian K, Felfeliyan B, Samavi S, Reza Soroushmehr SM. Liver Segmentation in Abdominal CT Images Using Probabilistic Atlas and Adaptive 3D Region Growing. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2019; 2019:6310-6313. [PMID: 31947285 DOI: 10.1109/embc.2019.8857835] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Automatic liver segmentation plays a vital role in computer-aided diagnosis or treatment. Manual segmentation of organs is a tedious and challenging task and is prone to human errors. In this paper, we propose innovative pre-processing and adaptive 3D region growing methods with subject-specific conditions. To obtain strong edges and high contrast, we propose effective contrast enhancement algorithm then we use the atlas intensity distribution of most probable voxels in probability maps along with location before designing conditions for our 3D region growing method. We also incorporate the organ boundary to restrict the region growing. We compare our method with the label fusion of 13 organs on state-of-the-art Deeds registration method and achieved Dice score of 92.56%.
Collapse
|
32
|
Integrating spatial configuration into heatmap regression based CNNs for landmark localization. Med Image Anal 2019; 54:207-219. [DOI: 10.1016/j.media.2019.03.007] [Citation(s) in RCA: 105] [Impact Index Per Article: 21.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2018] [Revised: 01/11/2019] [Accepted: 03/21/2019] [Indexed: 11/23/2022]
|
33
|
Tarroni G, Oktay O, Bai W, Schuh A, Suzuki H, Passerat-Palmbach J, de Marvao A, O'Regan DP, Cook S, Glocker B, Matthews PM, Rueckert D. Learning-Based Quality Control for Cardiac MR Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2019; 38:1127-1138. [PMID: 30403623 DOI: 10.1109/tmi.2018.2878509] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
The effectiveness of a cardiovascular magnetic resonance (CMR) scan depends on the ability of the operator to correctly tune the acquisition parameters to the subject being scanned and on the potential occurrence of imaging artifacts, such as cardiac and respiratory motion. In the clinical practice, a quality control step is performed by visual assessment of the acquired images; however, this procedure is strongly operator-dependent, cumbersome, and sometimes incompatible with the time constraints in clinical settings and large-scale studies. We propose a fast, fully automated, and learning-based quality control pipeline for CMR images, specifically for short-axis image stacks. Our pipeline performs three important quality checks: 1) heart coverage estimation; 2) inter-slice motion detection; 3) image contrast estimation in the cardiac region. The pipeline uses a hybrid decision forest method-integrating both regression and structured classification models-to extract landmarks and probabilistic segmentation maps from both long- and short-axis images as a basis to perform the quality checks. The technique was tested on up to 3000 cases from the UK Biobank and on 100 cases from the UK Digital Heart Project and validated against manual annotations and visual inspections performed by expert interpreters. The results show the capability of the proposed pipeline to correctly detect incomplete or corrupted scans (e.g., on UK Biobank, sensitivity and specificity, respectively, 88% and 99% for heart coverage estimation and 85% and 95% for motion detection), allowing their exclusion from the analyzed dataset or the triggering of a new acquisition.
Collapse
|
34
|
Alansary A, Oktay O, Li Y, Folgoc LL, Hou B, Vaillant G, Kamnitsas K, Vlontzos A, Glocker B, Kainz B, Rueckert D. Evaluating reinforcement learning agents for anatomical landmark detection. Med Image Anal 2019; 53:156-164. [PMID: 30784956 PMCID: PMC7610752 DOI: 10.1016/j.media.2019.02.007] [Citation(s) in RCA: 68] [Impact Index Per Article: 13.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2018] [Revised: 02/01/2019] [Accepted: 02/12/2019] [Indexed: 11/29/2022]
Abstract
Automatic detection of anatomical landmarks is an important step for a wide range of applications in medical image analysis. Manual annotation of landmarks is a tedious task and prone to observer errors. In this paper, we evaluate novel deep reinforcement learning (RL) strategies to train agents that can precisely and robustly localize target landmarks in medical scans. An artificial RL agent learns to identify the optimal path to the landmark by interacting with an environment, in our case 3D images. Furthermore, we investigate the use of fixed- and multi-scale search strategies with novel hierarchical action steps in a coarse-to-fine manner. Several deep Q-network (DQN) architectures are evaluated for detecting multiple landmarks using three different medical imaging datasets: fetal head ultrasound (US), adult brain and cardiac magnetic resonance imaging (MRI). The performance of our agents surpasses state-of-the-art supervised and RL methods. Our experiments also show that multi-scale search strategies perform significantly better than fixed-scale agents in images with large field of view and noisy background such as in cardiac MRI. Moreover, the novel hierarchical steps can significantly speed up the searching process by a factor of 4-5 times.
Collapse
Affiliation(s)
- Amir Alansary
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK.
| | - Ozan Oktay
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Yuanwei Li
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Loic Le Folgoc
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Benjamin Hou
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Ghislain Vaillant
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | | | - Athanasios Vlontzos
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Ben Glocker
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Bernhard Kainz
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| | - Daniel Rueckert
- Biomedical Image Analysis Group (BioMedIA), Imperial College London, London, UK
| |
Collapse
|
35
|
Kim J, Duchin Y, Shamir RR, Patriat R, Vitek J, Harel N, Sapiro G. Automatic localization of the subthalamic nucleus on patient-specific clinical MRI by incorporating 7 T MRI and machine learning: Application in deep brain stimulation. Hum Brain Mapp 2019; 40:679-698. [PMID: 30379376 PMCID: PMC6519731 DOI: 10.1002/hbm.24404] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2018] [Revised: 09/04/2018] [Accepted: 09/07/2018] [Indexed: 12/20/2022] Open
Abstract
Deep brain stimulation (DBS) of the subthalamic nucleus (STN) has shown clinical potential for relieving the motor symptoms of advanced Parkinson's disease. While accurate localization of the STN is critical for consistent across-patients effective DBS, clear visualization of the STN under standard clinical MR protocols is still challenging. Therefore, intraoperative microelectrode recordings (MER) are incorporated to accurately localize the STN. However, MER require significant neurosurgical expertise and lengthen the surgery time. Recent advances in 7 T MR technology facilitate the ability to clearly visualize the STN. The vast majority of centers, however, still do not have 7 T MRI systems, and fewer have the ability to collect and analyze the data. This work introduces an automatic STN localization framework based on standard clinical MRIs without additional cost in the current DBS planning protocol. Our approach benefits from a large database of 7 T MRI and its clinical MRI pairs. We first model in the 7 T database, using efficient machine learning algorithms, the spatial and geometric dependency between the STN and its adjacent structures (predictors). Given a standard clinical MRI, our method automatically computes the predictors and uses the learned information to predict the patient-specific STN. We validate our proposed method on clinical T2 W MRI of 80 subjects, comparing with experts-segmented STNs from the corresponding 7 T MRI pairs. The experimental results show that our framework provides more accurate and robust patient-specific STN localization than using state-of-the-art atlases. We also demonstrate the clinical feasibility of the proposed technique assessing the post-operative electrode active contact locations.
Collapse
Affiliation(s)
- Jinyoung Kim
- Surgical Information Sciences, Inc.MinneapolisMinnesota
| | - Yuval Duchin
- Surgical Information Sciences, Inc.MinneapolisMinnesota
| | | | - Remi Patriat
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesota
| | - Jerrold Vitek
- Department of NeurologyUniversity of MinnesotaMinneapolisMinnesota
| | - Noam Harel
- Surgical Information Sciences, Inc.MinneapolisMinnesota
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesota
- Department of NeurosurgeryUniversity of MinnesotaMinneapolisMinnesota
| | - Guillermo Sapiro
- Surgical Information Sciences, Inc.MinneapolisMinnesota
- Department of Electrical and Computer EngineeringDuke UniversityDurhamNorth Carolina
- Department of Biomedical EngineeringDuke UniversityDurhamNorth Carolina
- Department of Computer ScienceDuke UniversityDurhamNorth Carolina
- Department of MathematicsDuke UniversityDurhamNorth Carolina
| |
Collapse
|
36
|
Bai P, Udupa JK, Tong Y, Xie S, Torigian DA. Body region localization in whole-body low-dose CT images of PET/CT scans using virtual landmarks. Med Phys 2019; 46:1286-1299. [PMID: 30609058 DOI: 10.1002/mp.13376] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2018] [Revised: 12/11/2018] [Accepted: 12/31/2018] [Indexed: 12/18/2022] Open
Abstract
PURPOSE Radiological imaging and image interpretation for clinical decision making are mostly specific to each body region such as head and neck, thorax, abdomen, pelvis, and extremities. In this study, we present a new solution to trim automatically the given axial image stack into image volumes satisfying the given body region definition. METHODS The proposed approach consists of the following steps. First, a set of reference objects is selected and roughly segmented. Virtual landmarks (VLs) for the objects are then identified by using principal component analysis and recursive subdivision of the object via the principal axes system. The VLs can be defined based on just the binary objects or objects with gray values also considered. The VLs may lie anywhere with respect to the object, inside or outside, and rarely on the object surface, and are tethered to the object. Second, a classic neural network regressor is configured to learn the geometric mapping relationship between the VLs and the boundary locations of each body region. The trained network is then used to predict the locations of the body region boundaries. In this study, we focus on three body regions - thorax, abdomen, and pelvis, and predict their superior and inferior axial locations denoted by TS(I), TI(I), AS(I), AI(I), PS(I), and PI(I), respectively, for any given volume image I. Two kinds of reference objects - the skeleton and the lungs and airways, are employed to test the localization performance of the proposed approach. RESULTS Our method is tested by using low-dose unenhanced computed tomography (CT) images of 180 near whole-body 18 F-fluorodeoxyglucose-positron emission tomography/computed tomography (FDG-PET/CT) scans (including 34 whole-body scans) which are randomly divided into training and testing sets with a ratio of 85%:15%. The procedure is repeated six times and three times for the case of lungs and skeleton, respectively, with different divisions of the entire data set at this proportion. For the case of using skeleton as a reference object, the overall mean localization error for the six locations expressed as number of slices (nS) and distance (dS) in mm, is found to be nS: 3.4, 4.7, 4.1, 5.2, 5.2, and 3.9; dS: 13.4, 18.9, 16.5, 20.8, 20.8, and 15.5 mm for binary objects; nS: 4.1, 5.7, 4.3, 5.9, 5.9, and 4.0; dS: 16.2, 22.7, 17.2, 23.7, 23.7, and 16.1 mm for gray objects, respectively. For the case of using lungs and airways as a reference object, the corresponding results are, nS: 4.0, 5.3, 4.1, 6.9, 6.9, and 7.4; dS: 15.0, 19.7, 15.3, 26.2, 26.2, and 27.9 mm for binary objects; nS: 3.9, 5.4, 3.6, 7.2, 7.2, and 7.6; dS: 14.6, 20.1, 13.7, 27.3, 27.3, and 28.6 mm for gray objects, respectively. CONCLUSIONS Precise body region identification automatically in whole-body or body region tomographic images is vital for numerous medical image analysis and analytics applications. Despite its importance, this issue has received very little attention in the literature. We present a solution to this problem in this study using the concept of virtual landmarks. The method achieves localization accuracy within 2-3 slices, which is roughly comparable to the variation found in localization by experts. As long as the reference objects can be roughly segmented, the method with its learned VLs-to-boundary location relationship and predictive ability is transferable from one image modality to another.
Collapse
Affiliation(s)
- Peirui Bai
- College of Electronics, Communication and Physics, Shandong University of Science and Technology, Qingdao, Shandong, 266590, China.,Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Jayaram K Udupa
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Yubing Tong
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - ShiPeng Xie
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA.,College of Telecommunications and Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing, Jiangsu, 210023, China
| | - Drew A Torigian
- Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA, 19104, USA
| |
Collapse
|
37
|
Xu X, Zhou F, Liu B, Fu D, Bai X. Efficient Multiple Organ Localization in CT Image using 3D Region Proposal Network. IEEE TRANSACTIONS ON MEDICAL IMAGING 2019; 38:1885-1898. [PMID: 30676952 DOI: 10.1109/tmi.2019.2894854] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Organ localization is an essential preprocessing step for many medical image analysis tasks such as image registration, organ segmentation and lesion detection. In this work, we propose an efficient method for multiple organ localization in CT image using 3D region proposal network. Compared with other convolutional neural network based methods that successively detect the target organs in all slices to assemble the final 3D bounding box, our method is fully implemented in 3D manner, thus can take full advantages of the spatial context information in CT image to perform efficient organ localization with only one prediction. We also propose a novel backbone network architecture that generates high-resolution feature maps to further improve the localization performance on small organs. We evaluate our method on two clinical datasets, where 11 body organs and 12 head organs (or anatomical structures) are included. As our results shown, the proposed method achieves higher detection precision and localization accuracy than the current state-of-theart methods with approximate 4 to 18 times faster processing speed. Additionally, we have established a public dataset dedicated for organ localization on http://dx. doi.org/10.21227/df8g-pq27. The full implementation of the proposed method have also been made publicly available on https://github.com/superxuang/caffe_3d_faster_rcnn.
Collapse
|
38
|
Wang X, Zhai S, Niu Y. Automatic Vertebrae Localization and Identification by Combining Deep SSAE Contextual Features and Structured Regression Forest. J Digit Imaging 2019; 32:336-348. [PMID: 30631979 DOI: 10.1007/s10278-018-0140-5] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022] Open
Abstract
Automatic vertebrae localization and identification in medical computed tomography (CT) scans is of great value for computer-aided spine diseases diagnosis. In order to overcome the disadvantages of the approaches employing hand-crafted, low-level features and based on field-of-view priori assumption of spine structure, an automatic method is proposed to localize and identify vertebrae by combining deep stacked sparse autoencoder (SSAE) contextual features and structured regression forest (SRF). The method employs SSAE to learn image deep contextual features instead of hand-crafted ones by building larger-range input samples to improve their contextual discrimination ability. In the localization and identification stage, it incorporates the SRF model to achieve whole spine localization, then screens those vertebrae within the image, thus relieves the assumption that the part of spine in the field of image is visible. In the end, the output distribution of SRF and spine CT scans properties are assembled to develop a two-stage progressive refining strategy, where the mean-shift kernel density estimation and Otsu method instead of Markov random field (MRF) are adopted to reduce model complexity and refine vertebrae localization results. Extensive evaluation was performed on a challenging data set of 98 spine CT scans. Compared with the hidden Markov model and the method based on convolutional neural network (CNN), the proposed approach could effectively and automatically locate and identify spinal targets in CT scans, and achieve higher localization accuracy, low model complexity, and no need for any assumptions about visual field in CT scans.
Collapse
Affiliation(s)
- Xuchu Wang
- Key Laboratory of Optoelectronic Technology and Systems of Ministry of Education, College of Optoelectronic Engineering, Chongqing University, Campus A, Room 1105 of Main Building, 174 Shazhen Street, Shapinba District, Chongqing, 400040, China.
| | - Suiqiang Zhai
- Key Laboratory of Optoelectronic Technology and Systems of Ministry of Education, College of Optoelectronic Engineering, Chongqing University, Campus A, Room 1105 of Main Building, 174 Shazhen Street, Shapinba District, Chongqing, 400040, China
| | - Yanmin Niu
- Key Laboratory of Optoelectronic Technology and Systems of Ministry of Education, College of Optoelectronic Engineering, Chongqing University, Campus A, Room 1105 of Main Building, 174 Shazhen Street, Shapinba District, Chongqing, 400040, China
- College of Computer and Information Science, Chongqing Normal University, Chongqing, 400050, China
| |
Collapse
|
39
|
Damopoulos D, Lerch TD, Schmaranzer F, Tannast M, Chênes C, Zheng G, Schmid J. Segmentation of the proximal femur in radial MR scans using a random forest classifier and deformable model registration. Int J Comput Assist Radiol Surg 2019; 14:545-561. [PMID: 30604143 DOI: 10.1007/s11548-018-1899-z] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2018] [Accepted: 12/10/2018] [Indexed: 11/25/2022]
Abstract
BACKGROUND Radial 2D MRI scans of the hip are routinely used for the diagnosis of the cam type of femoroacetabular impingement (FAI) and of avascular necrosis (AVN) of the femoral head, both considered causes of hip joint osteoarthritis in young and active patients. A method for automated and accurate segmentation of the proximal femur from radial MRI scans could be very useful in both clinical routine and biomechanical studies. However, to our knowledge, no such method has been published before. PURPOSE The aims of this study are the development of a system for the segmentation of the proximal femur from radial MRI scans and the reconstruction of its 3D model that can be used for diagnosis and planning of hip-preserving surgery. METHODS The proposed system relies on: (a) a random forest classifier and (b) the registration of a 3D template mesh of the femur to the radial slices based on a physically based deformable model. The input to the system are the radial slices and the manually specified positions of three landmarks. Our dataset consists of the radial MRI scans of 25 patients symptomatic of FAI or AVN and accompanying manual segmentation of the femur, treated as the ground truth. RESULTS The achieved segmentation of the proximal femur has an average Dice similarity coefficient (DSC) of 96.37 ± 1.55%, an average symmetric mean absolute distance (SMAD) of 0.94 ± 0.39 mm and an average Hausdorff distance of 2.37 ± 1.14 mm. In the femoral head subregion, the average SMAD is 0.64 ± 0.18 mm and the average Hausdorff distance is 1.41 ± 0.56 mm. CONCLUSIONS We validated a semiautomated method for the segmentation of the proximal femur from radial MR scans. A 3D model of the proximal femur is also reconstructed, which can be used for the planning of hip-preserving surgery.
Collapse
Affiliation(s)
- Dimitrios Damopoulos
- Institute for Surgical Technology and Biomechanics, University of Bern, Stauffacherstrasse 78, 3014, Bern, Switzerland.
| | - Till Dominic Lerch
- Department of Orthopaedic Surgery and Traumatology, Inselspital, University of Bern, Freiburgstrasse, 3010, Bern, Switzerland
| | - Florian Schmaranzer
- Department of Orthopaedic Surgery and Traumatology, Inselspital, University of Bern, Freiburgstrasse, 3010, Bern, Switzerland
| | - Moritz Tannast
- Department of Orthopaedic Surgery and Traumatology, Inselspital, University of Bern, Freiburgstrasse, 3010, Bern, Switzerland
| | - Christophe Chênes
- School of Health Sciences - Geneva, HES-SO University of Applied Sciences and Arts Western Switzerland, Avenue de Champel 47, 1206, Geneva, Switzerland
| | - Guoyan Zheng
- Institute for Surgical Technology and Biomechanics, University of Bern, Stauffacherstrasse 78, 3014, Bern, Switzerland.
| | - Jérôme Schmid
- School of Health Sciences - Geneva, HES-SO University of Applied Sciences and Arts Western Switzerland, Avenue de Champel 47, 1206, Geneva, Switzerland
| |
Collapse
|
40
|
Wang H, Zhang N, Huo L, Zhang B. Dual-modality multi-atlas segmentation of torso organs from [ 18F]FDG-PET/CT images. Int J Comput Assist Radiol Surg 2018; 14:473-482. [PMID: 30390179 DOI: 10.1007/s11548-018-1879-3] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2018] [Accepted: 10/23/2018] [Indexed: 11/28/2022]
Abstract
PURPOSE Automated segmentation of torso organs from positron emission tomography/computed tomography (PET/CT) images is a prerequisite step for nuclear medicine image analysis. However, accurate organ segmentation from clinical PET/CT is challenging due to the poor soft tissue contrast in the low-dose CT image and the low spatial resolution of the PET image. To overcome these challenges, we developed a multi-atlas segmentation (MAS) framework for torso organ segmentation from 2-deoxy-2-[18F]fluoro-D-glucose PET/CT images. METHOD Our key idea is to use PET information to compensate for the imperfect CT contrast and use surface-based atlas fusion to overcome the low PET resolution. First, all the organs are segmented from CT using a conventional MAS method, and then the abdomen region of the PET image is automatically cropped. Focusing on the cropped PET image, a refined MAS segmentation of the abdominal organs is performed, using a surface-based atlas fusion approach to reach subvoxel accuracy. RESULTS This method was validated based on 69 PET/CT images. The Dice coefficients of the target organs were between 0.80 and 0.96, and the average surface distances were between 1.58 and 2.44 mm. Compared to the CT-based segmentation, the PET-based segmentation gained a Dice increase of 0.06 and an ASD decrease of 0.38 mm. The surface-based atlas fusion leads to significant accuracy improvement for the liver and kidneys and saved ~ 10 min computation time compared to volumetric atlas fusion. CONCLUSIONS The presented method achieves better segmentation accuracy than conventional MAS method within acceptable computation time for clinical applications.
Collapse
Affiliation(s)
- Hongkai Wang
- Department of Biomedical Engineering, Dalian University of Technology, Dalian, Liaoning, China
| | - Nan Zhang
- Department of Biomedical Engineering, Dalian University of Technology, Dalian, Liaoning, China
| | - Li Huo
- Department of Nuclear Medicine, Peking Union Medical College Hospital, Beijing, China
| | - Bin Zhang
- Department of Biomedical Engineering, Dalian University of Technology, Dalian, Liaoning, China.
| |
Collapse
|
41
|
Fallah F, Walter SS, Bamberg F, Yang B. Simultaneous Volumetric Segmentation of Vertebral Bodies and Intervertebral Discs on Fat-Water MR Images. IEEE J Biomed Health Inform 2018; 23:1692-1701. [PMID: 30281501 DOI: 10.1109/jbhi.2018.2872810] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Fat-water magnetic resonance (MR) images allow automated noninvasive analysis of morphological properties and fat fractions of vertebral bodies (VBs) and intervertebral discs (IVDs) that constitute an important part of human biomechanical systems. In this paper, we propose a fully automated approach for simultaneously segmenting multiple VBs and IVDs on fat-water MR images without prior localization or geometry estimation. This method involved a hierarchical random forest (HRF) classifier and a hierarchical conditional random field (HCRF) that encoded a multi-resolution image pyramid based on a set of multiscale local and contextual features. The HRF classifier employed penalized multivariate linear discriminants and SMOTEBagging to handle limited and imbalanced training data with large feature dimension. The HCRF estimated optimum labels according to their spatial and hierarchical consistencies by using the layer-wise significant features determined over the trained HRF classifier. To handle variable sample numbers at different resolutions, resolution-specific hyperparameters were used. This method was trained and evaluated for segmenting 15 thoracic and lumbar VBs and their IVDs on fat-water MR images of a subset of a large cohort data set. It was further evaluated for segmenting seven IVDs of the lower spine on fat-water images of a public grand challenge. These evaluations revealed the comparable accuracy of this method with the state-of-the-art while requiring less computational burden due to a simultaneous localization and segmentation.
Collapse
|
42
|
Cetina K, Buenaposada JM, Baumela L. Multi-class segmentation of neuronal structures in electron microscopy images. BMC Bioinformatics 2018; 19:298. [PMID: 30092759 PMCID: PMC6085694 DOI: 10.1186/s12859-018-2305-0] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2018] [Accepted: 07/24/2018] [Indexed: 01/05/2023] Open
Abstract
BACKGROUND Serial block face scanning electron microscopy (SBFEM) is becoming a popular technology in neuroscience. We have seen in the last years an increasing number of works addressing the problem of segmenting cellular structures in SBFEM images of brain tissue. The vast majority of them is designed to segment one specific structure, typically membranes, synapses and mitochondria. Our hypothesis is that the performance of these algorithms can be improved by concurrently segmenting more than one structure using image descriptions obtained at different scales. RESULTS We consider the simultaneous segmentation of two structures, namely, synapses with mitochondria, and mitochondra with membranes. To this end we select three image stacks encompassing different SBFEM acquisition technologies and image resolutions. We introduce both a new Boosting algorithm to perform feature scale selection and the Jaccard Curve as a tool compare several segmentation results. We then experimentally study the gains in performance obtained when simultaneously segmenting two structures with properly selected image descriptor scales. The results show that by doing so we achieve significant gains in segmentation accuracy when compared to the best results in the literature. CONCLUSIONS Simultaneously segmenting several neuronal structures described at different scales provides voxel classification algorithms with highly discriminating features that significantly improve segmentation accuracy.
Collapse
Affiliation(s)
- Kendrick Cetina
- Departamento de Inteligencia Artificial, Universidad Politécnica de Madrid, Campus de Montegancedo s/n, Boadilla del Monte, España, Madrid, 28660 Spain
| | | | - Luis Baumela
- Departamento de Inteligencia Artificial, Universidad Politécnica de Madrid, Campus de Montegancedo s/n, Boadilla del Monte, España, Madrid, 28660 Spain
| |
Collapse
|
43
|
Al WA, Jung HY, Yun ID, Jang Y, Park HB, Chang HJ. Automatic aortic valve landmark localization in coronary CT angiography using colonial walk. PLoS One 2018; 13:e0200317. [PMID: 30044802 PMCID: PMC6059446 DOI: 10.1371/journal.pone.0200317] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2018] [Accepted: 06/21/2018] [Indexed: 11/18/2022] Open
Abstract
The minimally invasive transcatheter aortic valve implantation (TAVI) is the most prevalent method to treat aortic valve stenosis. For pre-operative surgical planning, contrast-enhanced coronary CT angiography (CCTA) is used as the imaging technique to acquire 3-D measurements of the valve. Accurate localization of the eight aortic valve landmarks in CT images plays a vital role in the TAVI workflow because a small error risks blocking the coronary circulation. In order to examine the valve and mark the landmarks, physicians prefer a view parallel to the hinge plane, instead of using the conventional axial, coronal or sagittal view. However, customizing the view is a difficult and time-consuming task because of unclear aorta pose and different artifacts of CCTA. Therefore, automatic localization of landmarks can serve as a useful guide to the physicians customizing the viewpoint. In this paper, we present an automatic method to localize the aortic valve landmarks using colonial walk, a regression tree-based machine-learning algorithm. For efficient learning from the training set, we propose a two-phase optimized search space learning model in which a representative point inside the valvular area is first learned from the whole CT volume. All eight landmarks are then learned from a smaller area around that point. Experiment with preprocedural CCTA images of TAVI undergoing patients showed that our method is robust under high stenotic variation and notably efficient, as it requires only 12 milliseconds to localize all eight landmarks, as tested on a 3.60 GHz single-core CPU.
Collapse
Affiliation(s)
- Walid Abdullah Al
- Division of Computer and Electronic Systems Engineering, Hankuk University of Foreign Studies, Yongin, South Korea
| | - Ho Yub Jung
- Department of Computer Engineering, Chosun University, Gwangju, South Korea
- * E-mail:
| | - Il Dong Yun
- Division of Computer and Electronic Systems Engineering, Hankuk University of Foreign Studies, Yongin, South Korea
| | - Yeonggul Jang
- Brain Korea 21 Project for Medical Science, Yonsei University, Seoul, South Korea
| | - Hyung-Bok Park
- Yonsei-Cedars Sinai Integrative Cardiovascular Imaging Research Center, Yonsei University Health System, Seoul, South Korea
- Division of Cardiology, Cardiovascular Center, Myongji Hospital, Seonam University College of Medicine, Goyang, South Korea
| | - Hyuk-Jae Chang
- Division of Cardiology, Department of Internal Medicine, Severance Cardiovascular Hospital, Yonsei University College of Medicine, Seoul, South Korea
- Cardiovascular Research Institute, Yonsei University College of Medicine, Seoul, South Korea
| |
Collapse
|
44
|
Puchalski RB, Shah N, Miller J, Dalley R, Nomura SR, Yoon JG, Smith KA, Lankerovich M, Bertagnolli D, Bickley K, Boe AF, Brouner K, Butler S, Caldejon S, Chapin M, Datta S, Dee N, Desta T, Dolbeare T, Dotson N, Ebbert A, Feng D, Feng X, Fisher M, Gee G, Goldy J, Gourley L, Gregor BW, Gu G, Hejazinia N, Hohmann J, Hothi P, Howard R, Joines K, Kriedberg A, Kuan L, Lau C, Lee F, Lee H, Lemon T, Long F, Mastan N, Mott E, Murthy C, Ngo K, Olson E, Reding M, Riley Z, Rosen D, Sandman D, Shapovalova N, Slaughterbeck CR, Sodt A, Stockdale G, Szafer A, Wakeman W, Wohnoutka PE, White SJ, Marsh D, Rostomily RC, Ng L, Dang C, Jones A, Keogh B, Gittleman HR, Barnholtz-Sloan JS, Cimino PJ, Uppin MS, Keene CD, Farrokhi FR, Lathia JD, Berens ME, Iavarone A, Bernard A, Lein E, Phillips JW, Rostad SW, Cobbs C, Hawrylycz MJ, Foltz GD. An anatomic transcriptional atlas of human glioblastoma. Science 2018; 360:660-663. [PMID: 29748285 PMCID: PMC6414061 DOI: 10.1126/science.aaf2666] [Citation(s) in RCA: 337] [Impact Index Per Article: 56.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2016] [Accepted: 03/30/2018] [Indexed: 12/20/2022]
Abstract
Glioblastoma is an aggressive brain tumor that carries a poor prognosis. The tumor's molecular and cellular landscapes are complex, and their relationships to histologic features routinely used for diagnosis are unclear. We present the Ivy Glioblastoma Atlas, an anatomically based transcriptional atlas of human glioblastoma that aligns individual histologic features with genomic alterations and gene expression patterns, thus assigning molecular information to the most important morphologic hallmarks of the tumor. The atlas and its clinical and genomic database are freely accessible online data resources that will serve as a valuable platform for future investigations of glioblastoma pathogenesis, diagnosis, and treatment.
Collapse
Affiliation(s)
- Ralph B Puchalski
- Allen Institute for Brain Science, Seattle, WA 98109, USA.
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | - Nameeta Shah
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA.
- Mazumdar Shaw Center for Translational Research, Bangalore 560099, India
| | - Jeremy Miller
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Rachel Dalley
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Steve R Nomura
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | - Jae-Guen Yoon
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | | | - Michael Lankerovich
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | | | - Kris Bickley
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Andrew F Boe
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Krissy Brouner
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | | | - Mike Chapin
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Suvro Datta
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Nick Dee
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Tsega Desta
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Tim Dolbeare
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | - Amanda Ebbert
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - David Feng
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Xu Feng
- Radia Inc., Lynnwood, WA 98036, USA
| | - Michael Fisher
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Garrett Gee
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Jeff Goldy
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | | | - Guangyu Gu
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Nika Hejazinia
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - John Hohmann
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Parvinder Hothi
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | - Robert Howard
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Kevin Joines
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Ali Kriedberg
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Leonard Kuan
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Chris Lau
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Felix Lee
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Hwahyung Lee
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | - Tracy Lemon
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Fuhui Long
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Naveed Mastan
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Erika Mott
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Chantal Murthy
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | - Kiet Ngo
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Eric Olson
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Melissa Reding
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Zack Riley
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - David Rosen
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - David Sandman
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | | | - Andrew Sodt
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | - Aaron Szafer
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Wayne Wakeman
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | | | - Don Marsh
- White Marsh Forests, Seattle, WA 98119, USA
| | - Robert C Rostomily
- Department of Neurosurgery, Institute for Stem Cell and Regenerative Medicine, University of Washington School of Medicine, Seattle, WA 98195, USA
- Department of Neurological Surgery, Houston Methodist Hospital and Research Institute, Houston, TX 77030, USA
| | - Lydia Ng
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Chinh Dang
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Allan Jones
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | - Haley R Gittleman
- Case Comprehensive Cancer Center, Case Western Reserve University School of Medicine, Cleveland, OH 44106, USA
| | - Jill S Barnholtz-Sloan
- Case Comprehensive Cancer Center, Case Western Reserve University School of Medicine, Cleveland, OH 44106, USA
| | - Patrick J Cimino
- Department of Pathology, Division of Neuropathology, University of Washington School of Medicine, Seattle, WA 98104, USA
| | - Megha S Uppin
- Nizam's Institute of Medical Sciences, Punjagutta, Hyderabad 500082, India
| | - C Dirk Keene
- Department of Pathology, Division of Neuropathology, University of Washington School of Medicine, Seattle, WA 98104, USA
| | | | - Justin D Lathia
- Department of Cellular and Molecular Medicine, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Michael E Berens
- TGen, Translational Genomics Research Institute, Phoenix, AZ 85004, USA
| | - Antonio Iavarone
- Institute for Cancer Genetics, Columbia University, New York, NY 10032, USA
- Department of Neurology, Columbia University, New York, NY 10032, USA
- Department of Pathology, Columbia University, New York, NY 10032, USA
| | - Amy Bernard
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | - Ed Lein
- Allen Institute for Brain Science, Seattle, WA 98109, USA
| | | | | | - Charles Cobbs
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| | | | - Greg D Foltz
- Ben and Catherine Ivy Center for Advanced Brain Tumor Treatment, Swedish Neuroscience Institute, Seattle, WA 98122, USA
| |
Collapse
|
45
|
Li Y, Ho CP, Toulemonde M, Chahal N, Senior R, Tang MX. Fully Automatic Myocardial Segmentation of Contrast Echocardiography Sequence Using Random Forests Guided by Shape Model. IEEE TRANSACTIONS ON MEDICAL IMAGING 2018; 37:1081-1091. [PMID: 28961106 DOI: 10.1109/tmi.2017.2747081] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Myocardial contrast echocardiography (MCE) is an imaging technique that assesses left ventricle function and myocardial perfusion for the detection of coronary artery diseases. Automatic MCE perfusion quantification is challenging and requires accurate segmentation of the myocardium from noisy and time-varying images. Random forests (RF) have been successfully applied to many medical image segmentation tasks. However, the pixel-wise RF classifier ignores contextual relationships between label outputs of individual pixels. RF which only utilizes local appearance features is also susceptible to data suffering from large intensity variations. In this paper, we demonstrate how to overcome the above limitations of classic RF by presenting a fully automatic segmentation pipeline for myocardial segmentation in full-cycle 2-D MCE data. Specifically, a statistical shape model is used to provide shape prior information that guide the RF segmentation in two ways. First, a novel shape model (SM) feature is incorporated into the RF framework to generate a more accurate RF probability map. Second, the shape model is fitted to the RF probability map to refine and constrain the final segmentation to plausible myocardial shapes. We further improve the performance by introducing a bounding box detection algorithm as a preprocessing step in the segmentation pipeline. Our approach on 2-D image is further extended to 2-D+t sequences which ensures temporal consistency in the final sequence segmentations. When evaluated on clinical MCE data sets, our proposed method achieves notable improvement in segmentation accuracy and outperforms other state-of-the-art methods, including the classic RF and its variants, active shape model and image registration.
Collapse
|
46
|
Humpire-Mamani GE, Setio AAA, van Ginneken B, Jacobs C. Efficient organ localization using multi-label convolutional neural networks in thorax-abdomen CT scans. Phys Med Biol 2018; 63:085003. [PMID: 29512516 DOI: 10.1088/1361-6560/aab4b3] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Abstract
Automatic localization of organs and other structures in medical images is an important preprocessing step that can improve and speed up other algorithms such as organ segmentation, lesion detection, and registration. This work presents an efficient method for simultaneous localization of multiple structures in 3D thorax-abdomen CT scans. Our approach predicts the location of multiple structures using a single multi-label convolutional neural network for each orthogonal view. Each network takes extra slices around the current slice as input to provide extra context. A sigmoid layer is used to perform multi-label classification. The output of the three networks is subsequently combined to compute a 3D bounding box for each structure. We used our approach to locate 11 structures of interest. The neural network was trained and evaluated on a large set of 1884 thorax-abdomen CT scans from patients undergoing oncological workup. Reference bounding boxes were annotated by human observers. The performance of our method was evaluated by computing the wall distance to the reference bounding boxes. The bounding boxes annotated by the first human observer were used as the reference standard for the test set. Using the best configuration, we obtained an average wall distance of [Formula: see text] mm in the test set. The second human observer achieved [Formula: see text] mm. For all structures, the results were better than those reported in previously published studies. In conclusion, we proposed an efficient method for the accurate localization of multiple organs. Our method uses multiple slices as input to provide more context around the slice under analysis, and we have shown that this improves performance. This method can easily be adapted to handle more organs.
Collapse
Affiliation(s)
- Gabriel Efrain Humpire-Mamani
- Department of Radiology and Nuclear Medicine, Diagnostic Image Analysis Group, Radboud University Medical Center, Nijmegen, Netherlands
| | | | | | | |
Collapse
|
47
|
Oliveira B, Queirós S, Morais P, Torres HR, Gomes-Fonseca J, Fonseca JC, Vilaça JL. A novel multi-atlas strategy with dense deformation field reconstruction for abdominal and thoracic multi-organ segmentation from computed tomography. Med Image Anal 2018; 45:108-120. [DOI: 10.1016/j.media.2018.02.001] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2017] [Revised: 01/27/2018] [Accepted: 02/01/2018] [Indexed: 12/12/2022]
|
48
|
Li J, Wang Y, Lei B, Cheng JZ, Qin J, Wang T, Li S, Ni D. Automatic Fetal Head Circumference Measurement in Ultrasound Using Random Forest and Fast Ellipse Fitting. IEEE J Biomed Health Inform 2018; 22:215-223. [DOI: 10.1109/jbhi.2017.2703890] [Citation(s) in RCA: 53] [Impact Index Per Article: 8.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
49
|
Urschler M, Ebner T, Štern D. Integrating geometric configuration and appearance information into a unified framework for anatomical landmark localization. Med Image Anal 2018; 43:23-36. [DOI: 10.1016/j.media.2017.09.003] [Citation(s) in RCA: 41] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2017] [Revised: 07/27/2017] [Accepted: 09/11/2017] [Indexed: 11/29/2022]
|
50
|
Zhang D, Liu Y, Noble JH, Dawant BM. Localizing landmark sets in head CTs using random forests and a heuristic search algorithm for registration initialization. J Med Imaging (Bellingham) 2017; 4:044007. [PMID: 29250565 DOI: 10.1117/1.jmi.4.4.044007] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2017] [Accepted: 11/13/2017] [Indexed: 11/14/2022] Open
Abstract
Cochlear implants (CIs) use electrode arrays that are surgically inserted into the cochlea to stimulate frequency-mapped nerve endings to treat patients with hearing loss. CIs are programmed postoperatively by audiologists using behavioral tests without information on electrode-cochlea spatial relationship. We have recently developed techniques to segment the intracochlear anatomy and to localize individual contacts in clinically acquired computed tomography (CT) images. Using this information, we have proposed a programming strategy that we call image-guided CI programming (IGCIP), and we have shown that it significantly improves outcomes for both adult and pediatric recipients. One obstacle to large-scale deployment of this technique is the need for manual intervention in some processing steps. One of these is the rough registration of images prior to the use of automated intensity-based algorithms. Although seemingly simple, the heterogeneity of our image set makes this task challenging. We propose a solution that relies on the automated random forest-based localization of multiple landmarks used to estimate an initial transformation with a point-based registration method. Results show that it produces results that are equivalent to a manual initialization. This work is an important step toward the full automation of IGCIP.
Collapse
Affiliation(s)
- Dongqing Zhang
- Vanderbilt University, Department of Electrical Engineering and Computer Science, Nashville, Tennessee, United States
| | - Yuan Liu
- Vanderbilt University, Department of Electrical Engineering and Computer Science, Nashville, Tennessee, United States
| | - Jack H Noble
- Vanderbilt University, Department of Electrical Engineering and Computer Science, Nashville, Tennessee, United States
| | - Benoit M Dawant
- Vanderbilt University, Department of Electrical Engineering and Computer Science, Nashville, Tennessee, United States
| |
Collapse
|