1
|
Li H, Yang J, Xuan Z, Qu M, Wang Y, Feng C. A spatio-temporal graph convolutional network for ultrasound echocardiographic landmark detection. Med Image Anal 2024; 97:103272. [PMID: 39024972 DOI: 10.1016/j.media.2024.103272] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Revised: 07/07/2024] [Accepted: 07/08/2024] [Indexed: 07/20/2024]
Abstract
Landmark detection is a crucial task in medical image analysis, with applications across various fields. However, current methods struggle to accurately locate landmarks in medical images with blurred tissue boundaries due to low image quality. In particular, in echocardiography, sparse annotations make it challenging to predict landmarks with position stability and temporal consistency. In this paper, we propose a spatio-temporal graph convolutional network tailored for echocardiography landmark detection. We specifically sample landmark labels from the left ventricular endocardium and pre-calculate their correlations to establish structural priors. Our approach involves a graph convolutional neural network that learns the interrelationships among landmarks, significantly enhancing landmark accuracy within ambiguous tissue contexts. Additionally, we integrate gate recurrent units to grasp the temporal consistency of landmarks across consecutive images, augmenting the model's resilience against unlabeled data. Through validation across three echocardiography datasets, our method demonstrates superior accuracy when contrasted with alternative landmark detection models.
Collapse
Affiliation(s)
- Honghe Li
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, China
| | - Jinzhu Yang
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, China.
| | - Zhanfeng Xuan
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, China
| | - Mingjun Qu
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, China
| | - Yonghuai Wang
- Department of Cardiovascular Ultrasound, The First Hospital of China Medical University, China
| | - Chaolu Feng
- Computer Science and Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, China
| |
Collapse
|
2
|
Lee HT, Chiu PY, Yen CW, Chou ST, Tseng YC. Application of artificial intelligence in lateral cephalometric analysis. J Dent Sci 2024; 19:1157-1164. [PMID: 38618076 PMCID: PMC11010784 DOI: 10.1016/j.jds.2023.12.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Revised: 12/06/2023] [Accepted: 12/07/2023] [Indexed: 04/16/2024] Open
Affiliation(s)
- Huang-Ting Lee
- School of Dentistry, College of Dental Medicine, Kaohsiung Medical University, Kaohsiung, Taiwan
| | - Po-Yuan Chiu
- School of Dentistry, College of Dental Medicine, Kaohsiung Medical University, Kaohsiung, Taiwan
- Department of Orthodontics, Kaohsiung Medical University Hospital, Kaohsiung, Taiwan
| | - Chen-Wen Yen
- Department of Mechanical and Electromechanical Engineering, National Sun Yat-sen University, Kaohsiung, Taiwan
| | - Szu-Ting Chou
- School of Dentistry, College of Dental Medicine, Kaohsiung Medical University, Kaohsiung, Taiwan
- Department of Orthodontics, Kaohsiung Medical University Hospital, Kaohsiung, Taiwan
| | - Yu-Chuan Tseng
- School of Dentistry, College of Dental Medicine, Kaohsiung Medical University, Kaohsiung, Taiwan
- Department of Orthodontics, Kaohsiung Medical University Hospital, Kaohsiung, Taiwan
| |
Collapse
|
3
|
Hong W, Kim SM, Choi J, Ahn J, Paeng JY, Kim H. Automated Cephalometric Landmark Detection Using Deep Reinforcement Learning. J Craniofac Surg 2023; 34:2336-2342. [PMID: 37622568 DOI: 10.1097/scs.0000000000009685] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2022] [Accepted: 06/25/2023] [Indexed: 08/26/2023] Open
Abstract
Accurate cephalometric landmark detection leads to accurate analysis, diagnosis, and surgical planning. Many studies on automated landmark detection have been conducted, however reinforcement learning-based networks have not yet been applied. This is the first study to apply deep Q-network (DQN) and double deep Q-network (DDQN) to automated cephalometric landmark detection to the best of our knowledge. The performance of the DQN-based network for cephalometric landmark detection was evaluated using the IEEE International Symposium of Biomedical Imaging (ISBI) 2015 Challenge data set and compared with the previously proposed methods. Furthermore, the clinical applicability of DQN-based automated cephalometric landmark detection was confirmed by testing the DQN-based and DDQN-based network using 500-patient data collected in a clinic. The DQN-based network demonstrated that the average mean radius error of 19 landmarks was smaller than 2 mm, that is, the clinically accepted level, without data augmentation and additional preprocessing. Our DQN-based and DDQN-based approaches tested with the 500-patient data set showed the average success detection rate of 67.33% and 66.04% accuracy within 2 mm, respectively, indicating the feasibility and potential of clinical application.
Collapse
Affiliation(s)
- Woojae Hong
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| | - Seong-Min Kim
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| | - Joongyeon Choi
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| | - Jaemyung Ahn
- Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Seoul, Republic of Korea
| | - Jun-Young Paeng
- Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Seoul, Republic of Korea
| | - Hyunggun Kim
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| |
Collapse
|
4
|
Lung Field Segmentation in Chest X-ray Images Using Superpixel Resizing and Encoder–Decoder Segmentation Networks. Bioengineering (Basel) 2022; 9:bioengineering9080351. [PMID: 36004876 PMCID: PMC9404743 DOI: 10.3390/bioengineering9080351] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Revised: 07/24/2022] [Accepted: 07/26/2022] [Indexed: 11/25/2022] Open
Abstract
Lung segmentation of chest X-ray (CXR) images is a fundamental step in many diagnostic applications. Most lung field segmentation methods reduce the image size to speed up the subsequent processing time. Then, the low-resolution result is upsampled to the original high-resolution image. Nevertheless, the image boundaries become blurred after the downsampling and upsampling steps. It is necessary to alleviate blurred boundaries during downsampling and upsampling. In this paper, we incorporate the lung field segmentation with the superpixel resizing framework to achieve the goal. The superpixel resizing framework upsamples the segmentation results based on the superpixel boundary information obtained from the downsampling process. Using this method, not only can the computation time of high-resolution medical image segmentation be reduced, but also the quality of the segmentation results can be preserved. We evaluate the proposed method on JSRT, LIDC-IDRI, and ANH datasets. The experimental results show that the proposed superpixel resizing framework outperforms other traditional image resizing methods. Furthermore, combining the segmentation network and the superpixel resizing framework, the proposed method achieves better results with an average time score of 4.6 s on CPU and 0.02 s on GPU.
Collapse
|
5
|
|
6
|
Zeng M, Yan Z, Liu S, Zhou Y, Qiu L. Cascaded convolutional networks for automatic cephalometric landmark detection. Med Image Anal 2020; 68:101904. [PMID: 33290934 DOI: 10.1016/j.media.2020.101904] [Citation(s) in RCA: 32] [Impact Index Per Article: 6.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2019] [Revised: 06/15/2020] [Accepted: 11/11/2020] [Indexed: 11/17/2022]
Abstract
Cephalometric analysis is a fundamental examination which is widely used in orthodontic diagnosis and treatment planning. Its key step is to detect the anatomical landmarks in lateral cephalograms, which is time-consuming in traditional manual way. To solve this problem, we propose a novel approach with a cascaded three-stage convolutional neural networks to predict cephalometric landmarks automatically. In the first stage, high-level features of the craniofacial structures are extracted to locate the lateral face area which helps to overcome the appearance variations. Next, we process the aligned face area to estimate the locations of all landmarks simultaneously. At the last stage, each landmark is refined through a dedicated network using high-resolution image data around the initial position to achieve more accurate result. We evaluate the proposed method on several anatomical landmark datasets and the experimental results show that our method achieved competitive performance compared with the other methods.
Collapse
Affiliation(s)
- Minmin Zeng
- Fourth Clinical Division, School and Hospital of Stomatology, Peking University, Beijing, China.
| | | | - Shuai Liu
- Second Clinical Division, School and Hospital of Stomatology, Peking University, Beijing, China
| | - Yanheng Zhou
- Department of orthodontics, School and Hospital of Stomatology, Peking University, Beijing, China
| | - Lixin Qiu
- Fourth Clinical Division, School and Hospital of Stomatology, Peking University, Beijing, China
| |
Collapse
|
7
|
Kholiavchenko M, Sirazitdinov I, Kubrak K, Badrutdinova R, Kuleev R, Yuan Y, Vrtovec T, Ibragimov B. Contour-aware multi-label chest X-ray organ segmentation. Int J Comput Assist Radiol Surg 2020; 15:425-436. [PMID: 32034633 DOI: 10.1007/s11548-019-02115-9] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2019] [Accepted: 12/30/2019] [Indexed: 11/29/2022]
Abstract
PURPOSE Segmentation of organs from chest X-ray images is an essential task for an accurate and reliable diagnosis of lung diseases and chest organ morphometry. In this study, we investigated the benefits of augmenting state-of-the-art deep convolutional neural networks (CNNs) for image segmentation with organ contour information and evaluated the performance of such augmentation on segmentation of lung fields, heart, and clavicles from chest X-ray images. METHODS Three state-of-the-art CNNs were augmented, namely the UNet and LinkNet architecture with the ResNeXt feature extraction backbone, and the Tiramisu architecture with the DenseNet. All CNN architectures were trained on ground-truth segmentation masks and additionally on the corresponding contours. The contribution of such contour-based augmentation was evaluated against the contour-free architectures, and 20 existing algorithms for lung field segmentation. RESULTS The proposed contour-aware segmentation improved the segmentation performance, and when compared against existing algorithms on the same publicly available database of 247 chest X-ray images, the UNet architecture with the ResNeXt50 encoder combined with the contour-aware approach resulted in the best overall segmentation performance, achieving a Jaccard overlap coefficient of 0.971, 0.933, and 0.903 for the lung fields, heart, and clavicles, respectively. CONCLUSION In this study, we proposed to augment CNN architectures for CXR segmentation with organ contour information and were able to significantly improve segmentation accuracy and outperform all existing solution using a public chest X-ray database.
Collapse
Affiliation(s)
| | | | - K Kubrak
- Innopolis University, Innopolis, Russia
| | | | - R Kuleev
- Innopolis University, Innopolis, Russia
| | - Y Yuan
- Department of Electronic Engineering, City University of Hong Kong, Hong Kong, China
| | - T Vrtovec
- Faculty of Electrical Engineering, University of Ljubljana, Ljubljana, Slovenia
| | - B Ibragimov
- Innopolis University, Innopolis, Russia. .,Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.
| |
Collapse
|
8
|
Integrating spatial configuration into heatmap regression based CNNs for landmark localization. Med Image Anal 2019; 54:207-219. [DOI: 10.1016/j.media.2019.03.007] [Citation(s) in RCA: 105] [Impact Index Per Article: 17.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2018] [Revised: 01/11/2019] [Accepted: 03/21/2019] [Indexed: 11/23/2022]
|
9
|
Vandaele R, Aceto J, Muller M, Péronnet F, Debat V, Wang CW, Huang CT, Jodogne S, Martinive P, Geurts P, Marée R. Landmark detection in 2D bioimages for geometric morphometrics: a multi-resolution tree-based approach. Sci Rep 2018; 8:538. [PMID: 29323201 PMCID: PMC5765108 DOI: 10.1038/s41598-017-18993-5] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2017] [Accepted: 12/20/2017] [Indexed: 11/23/2022] Open
Abstract
The detection of anatomical landmarks in bioimages is a necessary but tedious step for geometric morphometrics studies in many research domains. We propose variants of a multi-resolution tree-based approach to speed-up the detection of landmarks in bioimages. We extensively evaluate our method variants on three different datasets (cephalometric, zebrafish, and drosophila images). We identify the key method parameters (notably the multi-resolution) and report results with respect to human ground truths and existing methods. Our method achieves recognition performances competitive with current existing approaches while being generic and fast. The algorithms are integrated in the open-source Cytomine software and we provide parameter configuration guidelines so that they can be easily exploited by end-users. Finally, datasets are readily available through a Cytomine server to foster future research.
Collapse
Affiliation(s)
- Rémy Vandaele
- Montefiore Institute, Department of Electrical engineering and Computer Science., University of Liège, Liège, 4000, Belgium.
| | - Jessica Aceto
- Laboratory for Organogenesis and Regeneration, GIGA-Research, University of Liège, Liège, 4000, Belgium
| | - Marc Muller
- Laboratory for Organogenesis and Regeneration, GIGA-Research, University of Liège, Liège, 4000, Belgium
| | - Frédérique Péronnet
- Institut de Biologie Paris-Seine (IBPS), UMR7622, Laboratoire de Biologie du Développement, UPMC Univ Paris 06, Paris, F-75005, France
| | - Vincent Debat
- Institut de Systématique, Evolution, Biodiversité, ISYEB UMR 7205 (CNRS, MNHN, UPMC, EPHE), Muséum national d'Histoire naturelle, Sorbonne Universités, Paris, F-75005, France
| | - Ching-Wei Wang
- Graduate Institute of Biomedical Engineering, National Taiwan University of Science and Technology, Taipei, 10607, Taiwan
| | - Cheng-Ta Huang
- Graduate Institute of Biomedical Engineering, National Taiwan University of Science and Technology, Taipei, 10607, Taiwan
| | - Sébastien Jodogne
- Department of Medical Physics, University Hospital (CHU) of Liège, University of Liège, Liège, 4000, Belgium
| | - Philippe Martinive
- Department of Medical Physics, University Hospital (CHU) of Liège, University of Liège, Liège, 4000, Belgium
| | - Pierre Geurts
- Montefiore Institute, Department of Electrical engineering and Computer Science., University of Liège, Liège, 4000, Belgium
| | - Raphaël Marée
- Montefiore Institute, Department of Electrical engineering and Computer Science., University of Liège, Liège, 4000, Belgium
| |
Collapse
|
10
|
Urschler M, Ebner T, Štern D. Integrating geometric configuration and appearance information into a unified framework for anatomical landmark localization. Med Image Anal 2018; 43:23-36. [DOI: 10.1016/j.media.2017.09.003] [Citation(s) in RCA: 41] [Impact Index Per Article: 5.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2017] [Revised: 07/27/2017] [Accepted: 09/11/2017] [Indexed: 11/29/2022]
|
11
|
Xiong J, Shao Y, Ma J, Ren Y, Wang Q, Zhao J. Lung field segmentation using weighted sparse shape composition with robust initialization. Med Phys 2017; 44:5916-5929. [PMID: 28875551 DOI: 10.1002/mp.12561] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2016] [Revised: 08/10/2017] [Accepted: 08/30/2017] [Indexed: 11/12/2022] Open
Abstract
PURPOSE Lung field segmentation for chest radiography is critical to pulmonary disease diagnosis. In this paper, we propose a new deformable model using weighted sparse shape composition with robust initialization to achieve robust and accurate lung field segmentation. METHODS Our method consists of three steps: initialization, deformation and regularization. The steps of deformation and regularization are iteratively employed until convergence. First, since a deformable model is sensitive to the initial shape, a robust initialization is obtained by using a novel voting strategy, which allows the reliable patches on the image to vote for each landmark of the initial shape. Then, each point of the initial shape independently deforms to the lung boundary under the guidance of the appearance model, which can distinguish lung tissues from nonlung tissues near the boundary. Finally, the deformed shape is regularized by weighted sparse shape composition (SSC) model, which is constrained by both boundary information and the correlations between each point of the deformed shape. RESULTS Our method has been evaluated on 247 chest radiographs from well-known dataset Japanese Society of Radiological Technology (JSRT) and achieved high overlap scores (0.955 ± 0.021). CONCLUSIONS The experimental results show that the proposed deformable segmentation model is more robust and accurate than the traditional appearance and shape model on the JSRT database. Our method also shows higher accuracy than most state-of-the-art methods.
Collapse
Affiliation(s)
- Junfeng Xiong
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yeqin Shao
- School of Transportation, Nantong University, Jiangsu, 226019, China
| | - Jingchen Ma
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yacheng Ren
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Qian Wang
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China.,SJTU-UIH Institute for Medical Imaging Technology, Shanghai Jiao Tong University, Shanghai, 200240, China.,MED-X Research Institute, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Jun Zhao
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China.,SJTU-UIH Institute for Medical Imaging Technology, Shanghai Jiao Tong University, Shanghai, 200240, China.,MED-X Research Institute, Shanghai Jiao Tong University, Shanghai, 200240, China
| |
Collapse
|
12
|
Ibragimov B, Korez R, Likar B, Pernus F, Xing L, Vrtovec T. Segmentation of Pathological Structures by Landmark-Assisted Deformable Models. IEEE TRANSACTIONS ON MEDICAL IMAGING 2017; 36:1457-1469. [PMID: 28207388 DOI: 10.1109/tmi.2017.2667578] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Computerized segmentation of pathological structures in medical images is challenging, as, in addition to unclear image boundaries, image artifacts, and traces of surgical activities, the shape of pathological structures may be very different from the shape of normal structures. Even if a sufficient number of pathological training samples are collected, statistical shape modeling cannot always capture shape features of pathological samples as they may be suppressed by shape features of a considerably larger number of healthy samples. At the same time, landmarking can be efficient in analyzing pathological structures but often lacks robustness. In this paper, we combine the advantages of landmark detection and deformable models into a novel supervised multi-energy segmentation framework that can efficiently segment structures with pathological shape. The framework adopts the theory of Laplacian shape editing, that was introduced in the field of computer graphics, so that the limitations of statistical shape modeling are avoided. The performance of the proposed framework was validated by segmenting fractured lumbar vertebrae from 3-D computed tomography images, atrophic corpora callosa from 2-D magnetic resonance (MR) cross-sections and cancerous prostates from 3D MR images, resulting respectively in a Dice coefficient of 84.7 ± 5.0%, 85.3 ± 4.8% and 78.3 ± 5.1%, and boundary distance of 1.14 ± 0.49mm, 1.42 ± 0.45mm and 2.27 ± 0.52mm. The obtained results were shown to be superior in comparison to existing deformable model-based segmentation algorithms.
Collapse
|
13
|
Yang W, Liu Y, Lin L, Yun Z, Lu Z, Feng Q, Chen W. Lung Field Segmentation in Chest Radiographs From Boundary Maps by a Structured Edge Detector. IEEE J Biomed Health Inform 2017; 22:842-851. [PMID: 28368835 DOI: 10.1109/jbhi.2017.2687939] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Lung field segmentation in chest radiographs (CXRs) is an essential preprocessing step in automatically analyzing such images. We present a method for lung field segmentation that is built on a high-quality boundary map detected by an efficient modern boundary detector, namely a structured edge detector (SED). A SED is trained beforehand to detect lung boundaries in CXRs with manually outlined lung fields. Then, an ultrametric contour map (UCM) is transformed from the masked and marked boundary map. Finally, the contours with the highest confidence level in the UCM are extracted as lung contours. Our method is evaluated using the public Japanese Society of Radiological Technology database of scanned films. The average Jaccard index of our method is 95.2%, which is comparable with those of other state-of-the-art methods (95.4%). The computation time of our method is less than 0.1 s for a CXR when executed on an ordinary laptop. Our method is also validated on CXRs acquired with different digital radiography units. The results demonstrate the generalization of the trained SED model and the usefulness of our method.
Collapse
|
14
|
Ibragimov B, Xing L. Segmentation of organs-at-risks in head and neck CT images using convolutional neural networks. Med Phys 2017; 44:547-557. [PMID: 28205307 DOI: 10.1002/mp.12045] [Citation(s) in RCA: 329] [Impact Index Per Article: 41.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2016] [Revised: 10/31/2016] [Accepted: 11/23/2016] [Indexed: 12/14/2022] Open
Abstract
PURPOSE Accurate segmentation of organs-at-risks (OARs) is the key step for efficient planning of radiation therapy for head and neck (HaN) cancer treatment. In the work, we proposed the first deep learning-based algorithm, for segmentation of OARs in HaN CT images, and compared its performance against state-of-the-art automated segmentation algorithms, commercial software, and interobserver variability. METHODS Convolutional neural networks (CNNs)-a concept from the field of deep learning-were used to study consistent intensity patterns of OARs from training CT images and to segment the OAR in a previously unseen test CT image. For CNN training, we extracted a representative number of positive intensity patches around voxels that belong to the OAR of interest in training CT images, and negative intensity patches around voxels that belong to the surrounding structures. These patches then passed through a sequence of CNN layers that captured local image features such as corners, end-points, and edges, and combined them into more complex high-order features that can efficiently describe the OAR. The trained network was applied to classify voxels in a region of interest in the test image where the corresponding OAR is expected to be located. We then smoothed the obtained classification results by using Markov random fields algorithm. We finally extracted the largest connected component of the smoothed voxels classified as the OAR by CNN, performed dilate-erode operations to remove cavities of the component, which resulted in segmentation of the OAR in the test image. RESULTS The performance of CNNs was validated on segmentation of spinal cord, mandible, parotid glands, submandibular glands, larynx, pharynx, eye globes, optic nerves, and optic chiasm using 50 CT images. The obtained segmentation results varied from 37.4% Dice coefficient (DSC) for chiasm to 89.5% DSC for mandible. We also analyzed the performance of state-of-the-art algorithms and commercial software reported in the literature, and observed that CNNs demonstrate similar or superior performance on segmentation of spinal cord, mandible, parotid glands, larynx, pharynx, eye globes, and optic nerves, but inferior performance on segmentation of submandibular glands and optic chiasm. CONCLUSION We concluded that convolution neural networks can accurately segment most of OARs using a representative database of 50 HaN CT images. At the same time, inclusion of additional information, for example, MR images, may be beneficial to some OARs with poorly visible boundaries.
Collapse
Affiliation(s)
- Bulat Ibragimov
- Department of Radiation Oncology, Stanford University School of Medicine, Stanford, California, 94305, USA
| | - Lei Xing
- Department of Radiation Oncology, Stanford University School of Medicine, Stanford, California, 94305, USA
| |
Collapse
|
15
|
Arık SÖ, Ibragimov B, Xing L. Fully automated quantitative cephalometry using convolutional neural networks. J Med Imaging (Bellingham) 2017; 4:014501. [PMID: 28097213 PMCID: PMC5220585 DOI: 10.1117/1.jmi.4.1.014501] [Citation(s) in RCA: 128] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2016] [Accepted: 12/12/2016] [Indexed: 11/14/2022] Open
Abstract
Quantitative cephalometry plays an essential role in clinical diagnosis, treatment, and surgery. Development of fully automated techniques for these procedures is important to enable consistently accurate computerized analyses. We study the application of deep convolutional neural networks (CNNs) for fully automated quantitative cephalometry for the first time. The proposed framework utilizes CNNs for detection of landmarks that describe the anatomy of the depicted patient and yield quantitative estimation of pathologies in the jaws and skull base regions. We use a publicly available cephalometric x-ray image dataset to train CNNs for recognition of landmark appearance patterns. CNNs are trained to output probabilistic estimations of different landmark locations, which are combined using a shape-based model. We evaluate the overall framework on the test set and compare with other proposed techniques. We use the estimated landmark locations to assess anatomically relevant measurements and classify them into different anatomical types. Overall, our results demonstrate high anatomical landmark detection accuracy ([Formula: see text] to 2% higher success detection rate for a 2-mm range compared with the top benchmarks in the literature) and high anatomical type classification accuracy ([Formula: see text] average classification accuracy for test set). We demonstrate that CNNs, which merely input raw image patches, are promising for accurate quantitative cephalometry.
Collapse
Affiliation(s)
- Sercan Ö. Arık
- Baidu USA, 1195 Bordeaux Drive, Sunnyvale, California 94089, United States
| | - Bulat Ibragimov
- Stanford University, Department of Radiation Oncology, School of Medicine, 875 Blake Wilbur Drive, Stanford, California 94305, United States
| | - Lei Xing
- Stanford University, Department of Radiation Oncology, School of Medicine, 875 Blake Wilbur Drive, Stanford, California 94305, United States
| |
Collapse
|
16
|
Wang CW, Huang CT, Lee JH, Li CH, Chang SW, Siao MJ, Lai TM, Ibragimov B, Vrtovec T, Ronneberger O, Fischer P, Cootes TF, Lindner C. A benchmark for comparison of dental radiography analysis algorithms. Med Image Anal 2016; 31:63-76. [PMID: 26974042 DOI: 10.1016/j.media.2016.02.004] [Citation(s) in RCA: 134] [Impact Index Per Article: 14.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2015] [Revised: 02/02/2016] [Accepted: 02/19/2016] [Indexed: 11/26/2022]
Abstract
Dental radiography plays an important role in clinical diagnosis, treatment and surgery. In recent years, efforts have been made on developing computerized dental X-ray image analysis systems for clinical usages. A novel framework for objective evaluation of automatic dental radiography analysis algorithms has been established under the auspices of the IEEE International Symposium on Biomedical Imaging 2015 Bitewing Radiography Caries Detection Challenge and Cephalometric X-ray Image Analysis Challenge. In this article, we present the datasets, methods and results of the challenge and lay down the principles for future uses of this benchmark. The main contributions of the challenge include the creation of the dental anatomy data repository of bitewing radiographs, the creation of the anatomical abnormality classification data repository of cephalometric radiographs, and the definition of objective quantitative evaluation for comparison and ranking of the algorithms. With this benchmark, seven automatic methods for analysing cephalometric X-ray image and two automatic methods for detecting bitewing radiography caries have been compared, and detailed quantitative evaluation results are presented in this paper. Based on the quantitative evaluation results, we believe automatic dental radiography analysis is still a challenging and unsolved problem. The datasets and the evaluation software will be made available to the research community, further encouraging future developments in this field. (http://www-o.ntust.edu.tw/~cweiwang/ISBI2015/).
Collapse
Affiliation(s)
- Ching-Wei Wang
- Graduate Institute of Biomedical Engineering, National Taiwan University of Science and Technology, Taiwan; NTUST Center of Computer Vision and Medical Imaging, Taiwan.
| | - Cheng-Ta Huang
- Graduate Institute of Biomedical Engineering, National Taiwan University of Science and Technology, Taiwan; NTUST Center of Computer Vision and Medical Imaging, Taiwan
| | - Jia-Hong Lee
- Graduate Institute of Biomedical Engineering, National Taiwan University of Science and Technology, Taiwan; NTUST Center of Computer Vision and Medical Imaging, Taiwan
| | - Chung-Hsing Li
- Orthodontics and Pediatric Dentistry Division, Dental Department, Tri-Service General Hospital, Taiwan; School of Dentistry and Graduate Institute of Dental Science, National Defense Medical Center, Taipei, Taiwan
| | - Sheng-Wei Chang
- Orthodontics and Pediatric Dentistry Division, Dental Department, Tri-Service General Hospital, Taiwan
| | - Ming-Jhih Siao
- Orthodontics and Pediatric Dentistry Division, Dental Department, Tri-Service General Hospital, Taiwan
| | - Tat-Ming Lai
- Department of Dentistry, Cardinal Tien Hospital, Taipei, Taiwan
| | - Bulat Ibragimov
- Faculty of Electrical Engineering, University of Ljubljana, Tržaška 25, SI-1000 Ljubljana, Slovenia
| | - Tomaž Vrtovec
- Faculty of Electrical Engineering, University of Ljubljana, Tržaška 25, SI-1000 Ljubljana, Slovenia
| | | | | | - Tim F Cootes
- Centre for Imaging Sciences, The University of Manchester, UK
| | - Claudia Lindner
- Centre for Imaging Sciences, The University of Manchester, UK
| |
Collapse
|
17
|
Wang CW, Huang CT, Hsieh MC, Li CH, Chang SW, Li WC, Vandaele R, Marée R, Jodogne S, Geurts P, Chen C, Zheng G, Chu C, Mirzaalian H, Hamarneh G, Vrtovec T, Ibragimov B. Evaluation and Comparison of Anatomical Landmark Detection Methods for Cephalometric X-Ray Images: A Grand Challenge. IEEE TRANSACTIONS ON MEDICAL IMAGING 2015; 34:1890-900. [PMID: 25794388 DOI: 10.1109/tmi.2015.2412951] [Citation(s) in RCA: 93] [Impact Index Per Article: 9.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Cephalometric analysis is an essential clinical and research tool in orthodontics for the orthodontic analysis and treatment planning. This paper presents the evaluation of the methods submitted to the Automatic Cephalometric X-Ray Landmark Detection Challenge, held at the IEEE International Symposium on Biomedical Imaging 2014 with an on-site competition. The challenge was set to explore and compare automatic landmark detection methods in application to cephalometric X-ray images. Methods were evaluated on a common database including cephalograms of 300 patients aged six to 60 years, collected from the Dental Department, Tri-Service General Hospital, Taiwan, and manually marked anatomical landmarks as the ground truth data, generated by two experienced medical doctors. Quantitative evaluation was performed to compare the results of a representative selection of current methods submitted to the challenge. Experimental results show that three methods are able to achieve detection rates greater than 80% using the 4 mm precision range, but only one method achieves a detection rate greater than 70% using the 2 mm precision range, which is the acceptable precision range in clinical practice. The study provides insights into the performance of different landmark detection approaches under real-world conditions and highlights achievements and limitations of current image analysis techniques.
Collapse
|
18
|
Park SH, Lee S, Yun ID, Lee SU. Structured patch model for a unified automatic and interactive segmentation framework. Med Image Anal 2015; 24:297-312. [PMID: 25682219 DOI: 10.1016/j.media.2015.01.003] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2014] [Revised: 01/05/2015] [Accepted: 01/19/2015] [Indexed: 11/30/2022]
Abstract
We present a novel interactive segmentation framework incorporating a priori knowledge learned from training data. The knowledge is learned as a structured patch model (StPM) comprising sets of corresponding local patch priors and their pairwise spatial distribution statistics which represent the local shape and appearance along its boundary and the global shape structure, respectively. When successive user annotations are given, the StPM is appropriately adjusted in the target image and used together with the annotations to guide the segmentation. The StPM reduces the dependency on the placement and quantity of user annotations with little increase in complexity since the time-consuming StPM construction is performed offline. Furthermore, a seamless learning system can be established by directly adding the patch priors and the pairwise statistics of segmentation results to the StPM. The proposed method was evaluated on three datasets, respectively, of 2D chest CT, 3D knee MR, and 3D brain MR. The experimental results demonstrate that within an equal amount of time, the proposed interactive segmentation framework outperforms recent state-of-the-art methods in terms of accuracy, while it requires significantly less computing and editing time to obtain results with comparable accuracy.
Collapse
Affiliation(s)
- Sang Hyun Park
- Department of Electrical Engineering, ASRI, INMC, Seoul National University, Seoul, Republic of Korea.
| | - Soochahn Lee
- Department of Electronic Engineering, Soonchunhyang University, Asan-si, Republic of Korea.
| | - Il Dong Yun
- Department of Digital Information Engineering, Hankuk University of Foreign Studies, Yongin, Republic of Korea.
| | - Sang Uk Lee
- Department of Electrical Engineering, ASRI, INMC, Seoul National University, Seoul, Republic of Korea.
| |
Collapse
|
19
|
Ibragimov B, Prince JL, Murano EZ, Woo J, Stone M, Likar B, Pernuš F, Vrtovec T. Segmentation of tongue muscles from super-resolution magnetic resonance images. Med Image Anal 2014; 20:198-207. [PMID: 25487963 DOI: 10.1016/j.media.2014.11.006] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2014] [Revised: 11/11/2014] [Accepted: 11/15/2014] [Indexed: 10/24/2022]
Abstract
Imaging and quantification of tongue anatomy is helpful in surgical planning, post-operative rehabilitation of tongue cancer patients, and studying of how humans adapt and learn new strategies for breathing, swallowing and speaking to compensate for changes in function caused by disease, medical interventions or aging. In vivo acquisition of high-resolution three-dimensional (3D) magnetic resonance (MR) images with clearly visible tongue muscles is currently not feasible because of breathing and involuntary swallowing motions that occur over lengthy imaging times. However, recent advances in image reconstruction now allow the generation of super-resolution 3D MR images from sets of orthogonal images, acquired at a high in-plane resolution and combined using super-resolution techniques. This paper presents, to the best of our knowledge, the first attempt towards automatic tongue muscle segmentation from MR images. We devised a database of ten super-resolution 3D MR images, in which the genioglossus and inferior longitudinalis tongue muscles were manually segmented and annotated with landmarks. We demonstrate the feasibility of segmenting the muscles of interest automatically by applying the landmark-based game-theoretic framework (GTF), where a landmark detector based on Haar-like features and an optimal assignment-based shape representation were integrated. The obtained segmentation results were validated against an independent manual segmentation performed by a second observer, as well as against B-splines and demons atlasing approaches. The segmentation performance resulted in mean Dice coefficients of 85.3%, 81.8%, 78.8% and 75.8% for the second observer, GTF, B-splines atlasing and demons atlasing, respectively. The obtained level of segmentation accuracy indicates that computerized tongue muscle segmentation may be used in surgical planning and treatment outcome analysis of tongue cancer patients, and in studies of normal subjects and subjects with speech and swallowing problems.
Collapse
Affiliation(s)
- Bulat Ibragimov
- Faculty of Electrical Engineering, University of Ljubljana, Ljubljana, Slovenia; Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA.
| | - Jerry L Prince
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA
| | - Emi Z Murano
- Department of Otolaryngology, Head and Neck Surgery, Johns Hopkins University, Baltimore, MD, USA
| | - Jonghye Woo
- Department of Radiology, Harvard Medical School/MGH, Boston, MA, USA
| | - Maureen Stone
- Department of Oral and Craniofacial Biological Sciences, University of Maryland, Baltimore, MD, USA; Department of Orthodontics, University of Maryland, Baltimore, MD, USA
| | - Boštjan Likar
- Faculty of Electrical Engineering, University of Ljubljana, Ljubljana, Slovenia
| | - Franjo Pernuš
- Faculty of Electrical Engineering, University of Ljubljana, Ljubljana, Slovenia
| | - Tomaž Vrtovec
- Faculty of Electrical Engineering, University of Ljubljana, Ljubljana, Slovenia
| |
Collapse
|
20
|
Robust Initialization of Active Shape Models for Lung Segmentation in CT Scans: A Feature-Based Atlas Approach. Int J Biomed Imaging 2014; 2014:479154. [PMID: 25400660 PMCID: PMC4221988 DOI: 10.1155/2014/479154] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2014] [Accepted: 10/02/2014] [Indexed: 11/17/2022] Open
Abstract
Model-based segmentation methods have the advantage of incorporating a priori shape information into the segmentation process but suffer from the drawback that the model must be initialized sufficiently close to the target. We propose a novel approach for initializing an active shape model (ASM) and apply it to 3D lung segmentation in CT scans. Our method constructs an atlas consisting of a set of representative lung features and an average lung shape. The ASM pose parameters are found by transforming the average lung shape based on an affine transform computed from matching features between the new image and representative lung features. Our evaluation on a diverse set of 190 images showed an average dice coefficient of 0.746 ± 0.068 for initialization and 0.974 ± 0.017 for subsequent segmentation, based on an independent reference standard. The mean absolute surface distance error was 0.948 ± 1.537 mm. The initialization as well as segmentation results showed a statistically significant improvement compared to four other approaches. The proposed initialization method can be generalized to other applications employing ASM-based segmentation.
Collapse
|
21
|
Shao Y, Gao Y, Guo Y, Shi Y, Yang X, Shen D. Hierarchical lung field segmentation with joint shape and appearance sparse learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2014; 33:1761-80. [PMID: 25181734 DOI: 10.1109/tmi.2014.2305691] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Lung field segmentation in the posterior-anterior (PA) chest radiograph is important for pulmonary disease diagnosis and hemodialysis treatment. Due to high shape variation and boundary ambiguity, accurate lung field segmentation from chest radiograph is still a challenging task. To tackle these challenges, we propose a joint shape and appearance sparse learning method for robust and accurate lung field segmentation. The main contributions of this paper are: 1) a robust shape initialization method is designed to achieve an initial shape that is close to the lung boundary under segmentation; 2) a set of local sparse shape composition models are built based on local lung shape segments to overcome the high shape variations; 3) a set of local appearance models are similarly adopted by using sparse representation to capture the appearance characteristics in local lung boundary segments, thus effectively dealing with the lung boundary ambiguity; 4) a hierarchical deformable segmentation framework is proposed to integrate the scale-dependent shape and appearance information together for robust and accurate segmentation. Our method is evaluated on 247 PA chest radiographs in a public dataset. The experimental results show that the proposed local shape and appearance models outperform the conventional shape and appearance models. Compared with most of the state-of-the-art lung field segmentation methods under comparison, our method also shows a higher accuracy, which is comparable to the inter-observer annotation variation.
Collapse
|
22
|
Ibragimov B, Likar B, Pernuš F, Vrtovec T. Shape representation for efficient landmark-based segmentation in 3-d. IEEE TRANSACTIONS ON MEDICAL IMAGING 2014; 33:861-874. [PMID: 24710155 DOI: 10.1109/tmi.2013.2296976] [Citation(s) in RCA: 44] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
In this paper, we propose a novel approach to landmark-based shape representation that is based on transportation theory, where landmarks are considered as sources and destinations, all possible landmark connections as roads, and established landmark connections as goods transported via these roads. Landmark connections, which are selectively established, are identified through their statistical properties describing the shape of the object of interest, and indicate the least costly roads for transporting goods from sources to destinations. From such a perspective, we introduce three novel shape representations that are combined with an existing landmark detection algorithm based on game theory. To reduce computational complexity, which results from the extension from 2-D to 3-D segmentation, landmark detection is augmented by a concept known in game theory as strategy dominance. The novel shape representations, game-theoretic landmark detection and strategy dominance are combined into a segmentation framework that was evaluated on 3-D computed tomography images of lumbar vertebrae and femoral heads. The best shape representation yielded symmetric surface distance of 0.75 mm and 1.11 mm, and Dice coefficient of 93.6% and 96.2% for lumbar vertebrae and femoral heads, respectively. By applying strategy dominance, the computational costs were further reduced for up to three times.
Collapse
|