1
|
Khan U, Yasin A. Plane invariant segmentation of computed tomography images through weighted cross entropy optimized conditional GANs in compressed formats. Med Biol Eng Comput 2023; 61:2677-2697. [PMID: 37428300 DOI: 10.1007/s11517-023-02846-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Accepted: 05/12/2023] [Indexed: 07/11/2023]
Abstract
Computed tomography (CT) scan provides first-hand knowledge to doctors to identify an ailment. Deep neural networks help enhance image understanding through segmentation and labeling. In this work, we implement two variants of Pix2Pix generative adversarial networks (GANs) with varying complexities of generator and discriminator networks for plane invariant segmentation of CT scan images and subsequently propose an effective generative adversarial network with a suitably weighted binary cross-entropy loss function followed by image processing layer necessary for getting high-quality output segmentation. Our conditional GAN is powered by a unique set of an encoder-decoder network that coupled with the image processing layer produces enhanced segmentation. The network can be extended to the complete set of Hounsfield units and can also be implemented on smartphones. Furthermore, we also demonstrate effects on accuracy, F-1 score, and Jaccard index by using the conditional GAN networks on the spine vertebrae dataset, thus achieving an average of 86.28 % accuracy, 90.5 % Jaccard index score, and 89.9 % F-1 score in predicting segmented maps for validation input images. In addition, an overall lifting of accuracy, F-1 score, and Jaccard index graph for validation images with better continuity has also been highlighted.
Collapse
Affiliation(s)
- Usman Khan
- SS-CASE-IT Islamabad, Islamabad, Pakistan.
| | | |
Collapse
|
2
|
Thirukrishna JT, Krishna SRS, Shashank P, Srikanth S, Raghu V. Survey on Diagnosing CORONA VIRUS from Radiography Chest X-ray Images Using Convolutional Neural Networks. Wirel Pers Commun 2022; 124:2261-2270. [PMID: 35035106 PMCID: PMC8742162 DOI: 10.1007/s11277-022-09463-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Accepted: 12/31/2021] [Indexed: 06/14/2023]
Abstract
Corona Virus continues to harms its effects on the people lives across the globe. The screening of infected persons has to be identified is a vital step because it is a fast and low-cost way. Certain above mentioned things can be recognized by chest X-ray images that plays a significant role and also used for examining in detection of CORONA VIRUS(COVID-19). Here radiological chest X-rays are easily available with low cost only. In this survey paper, Convolutional Neural Network(CNN) based solution that will benefit in detection of the Covid-19 positive patients using radiography chest X-Ray images. To test the efficiency of the solution, using data sets of publicly available X-Ray images of Corona virus positive cases and negative cases. Images of positive Corona Virus patients and pictures of healthy person images are divided into testing images and trainable images. The solution which are providing the good results with classification accuracy within the test set-up. Then GUI based application supports for medical examination areas. This GUI application can be used on any computer and performed by any medical examiner or technician to determine Corona Virus positive patients using radiography X-ray images. The result will be precisely obtaining the Covid-19 Patient analysis through the chest X-ray images and also results may be retrieve within a few seconds.
Collapse
Affiliation(s)
- J. T. Thirukrishna
- Department of Information Science and Engineering, Dayananda Sagar Academy of Technology and Management, Bangalore, Karnataka India
| | - Sanda Reddy Sai Krishna
- Department of Information Science and Engineering, Dayananda Sagar Academy of Technology and Management, Bangalore, Karnataka India
| | - Policherla Shashank
- Department of Information Science and Engineering, Dayananda Sagar Academy of Technology and Management, Bangalore, Karnataka India
| | - S. Srikanth
- Department of Information Science and Engineering, Dayananda Sagar Academy of Technology and Management, Bangalore, Karnataka India
| | - V. Raghu
- Department of Information Science and Engineering, Dayananda Sagar Academy of Technology and Management, Bangalore, Karnataka India
| |
Collapse
|
3
|
Goswami M. Deep learning models for benign and malign ocular tumor growth estimation. Comput Med Imaging Graph 2021; 93:101986. [PMID: 34509705 DOI: 10.1016/j.compmedimag.2021.101986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2020] [Revised: 05/04/2021] [Accepted: 08/24/2021] [Indexed: 11/27/2022]
Abstract
Relatively abundant availability of medical imaging data has provided significant support in the development and testing of Neural Network based image processing methods. Clinicians often face issues in selecting suitable image processing algorithm for medical imaging data. A strategy for the selection of a proper model is presented here. The training data set comprises optical coherence tomography (OCT) and angiography (OCT-A) images of 50 mice eyes with more than 100 days follow-up. The data contains images from treated and untreated mouse eyes. Four deep learning variants are tested for automatic (a) differentiation of tumor region with healthy retinal layer and (b) segmentation of 3D ocular tumor volumes. Exhaustive sensitivity analysis of deep learning models is performed with respect to the number of training and testing images using eight performance indices to study accuracy, reliability/reproducibility, and speed. U-net with UVgg16 is best for malign tumor data set with treatment (having considerable variation) and U-net with Inception backbone for benign tumor data (with minor variation). Loss value and root mean square error (R.M.S.E.) are found most and least sensitive performance indices, respectively. The performance (via indices) is found to be exponentially improving regarding a number of training images. The segmented OCT-Angiography data shows that neovascularization drives the tumor volume. Image analysis shows that photodynamic imaging-assisted tumor treatment protocol is transforming an aggressively growing tumor into a cyst. An empirical expression is obtained to help medical professionals choose a particular model given the number of images and types of characteristics. We recommend that the presented exercise should be taken as standard practice before employing a particular deep learning model for biomedical image analysis.
Collapse
Affiliation(s)
- Mayank Goswami
- Divyadrishti Imaging Laboratory, Department of Physics, Indian Institute of Technology Roorkee, Roorkee, India.
| |
Collapse
|
4
|
Maheshwari D, Ghosh SK, Tripathy RK, Sharma M, Acharya UR. Automated accurate emotion recognition system using rhythm-specific deep convolutional neural network technique with multi-channel EEG signals. Comput Biol Med 2021; 134:104428. [PMID: 33984749 DOI: 10.1016/j.compbiomed.2021.104428] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2021] [Revised: 04/15/2021] [Accepted: 04/19/2021] [Indexed: 10/21/2022]
Abstract
Emotion is interpreted as a psycho-physiological process, and it is associated with personality, behavior, motivation, and character of a person. The objective of affective computing is to recognize different types of emotions for human-computer interaction (HCI) applications. The spatiotemporal brain electrical activity is measured using multi-channel electroencephalogram (EEG) signals. Automated emotion recognition using multi-channel EEG signals is an exciting research topic in cognitive neuroscience and affective computing. This paper proposes the rhythm-specific multi-channel convolutional neural network (CNN) based approach for automated emotion recognition using multi-channel EEG signals. The delta (δ), theta (θ), alpha (α), beta (β), and gamma (γ) rhythms of EEG signal for each channel are evaluated using band-pass filters. The EEG rhythms from the selected channels coupled with deep CNN are used for emotion classification tasks such as low-valence (LV) vs. high valence (HV), low-arousal (LA) vs. high-arousal (HA), and low-dominance (LD) vs. high dominance (HD) respectively. The deep CNN architecture considered in the proposed work has eight convolutions, three average pooling, four batch-normalization, three spatial drop-outs, two drop-outs, one global average pooling and, three dense layers. We have validated our developed model using three publicly available databases: DEAP, DREAMER, and DASPS. The results reveal that the proposed multivariate deep CNN approach coupled with β-rhythm has obtained the accuracy values of 98.91%, 98.45%, and 98.69% for LV vs. HV, LA vs. HA, and LD vs. HD emotion classification strategies, respectively using DEAP database with 10-fold cross-validation (CV) scheme. Similarly, the accuracy values of 98.56%, 98.82%, and 98.99% are obtained for LV vs. HV, LA vs. HA, and LD vs. HD classification schemes, respectively, using deep CNN and θ-rhythm. The proposed multi-channel rhythm-specific deep CNN classification model has obtained the average accuracy value of 57.14% using α-rhythm and trial-specific CV using DASPS database. Moreover, for 8-quadrant based emotion classification strategy, the deep CNN based classifier has obtained an overall accuracy value of 24.37% using γ-rhythms of multi-channel EEG signals. Our developed deep CNN model can be used for real-time automated emotion recognition applications.
Collapse
Affiliation(s)
- Daksh Maheshwari
- Department of Electrical and Electronics Engineering, BITS-Pilani, Hyderabad Campus, Hyderabad, 500078, India
| | - S K Ghosh
- Department of Electrical and Electronics Engineering, BITS-Pilani, Hyderabad Campus, Hyderabad, 500078, India
| | - R K Tripathy
- Department of Electrical and Electronics Engineering, BITS-Pilani, Hyderabad Campus, Hyderabad, 500078, India.
| | - Manish Sharma
- Department of Electrical and Computer Science Engineering, IITRAM, Ahmedabad, India
| | - U Rajendra Acharya
- Department of Electronics and Computer Engineering, Ngee Ann Polytechnic, Singapore; Department of Bioinformatics and Medical Engineering, Asia University, Taichung, Taiwan; International Research Organization for Advanced Science and Technology, Kumamoto University, Kumamoto, Japan
| |
Collapse
|
5
|
Gaur L, Bhatia U, Jhanjhi NZ, Muhammad G, Masud M. Medical image-based detection of COVID-19 using Deep Convolution Neural Networks. Multimed Syst 2021; 29:1729-1738. [PMID: 33935377 PMCID: PMC8079233 DOI: 10.1007/s00530-021-00794-6] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2020] [Accepted: 04/05/2021] [Indexed: 05/08/2023]
Abstract
The demand for automatic detection of Novel Coronavirus or COVID-19 is increasing across the globe. The exponential rise in cases burdens healthcare facilities, and a vast amount of multimedia healthcare data is being explored to find a solution. This study presents a practical solution to detect COVID-19 from chest X-rays while distinguishing those from normal and impacted by Viral Pneumonia via Deep Convolution Neural Networks (CNN). In this study, three pre-trained CNN models (EfficientNetB0, VGG16, and InceptionV3) are evaluated through transfer learning. The rationale for selecting these specific models is their balance of accuracy and efficiency with fewer parameters suitable for mobile applications. The dataset used for the study is publicly available and compiled from different sources. This study uses deep learning techniques and performance metrics (accuracy, recall, specificity, precision, and F1 scores). The results show that the proposed approach produced a high-quality model, with an overall accuracy of 92.93%, COVID-19, a sensitivity of 94.79%. The work indicates a definite possibility to implement computer vision design to enable effective detection and screening measures.
Collapse
Affiliation(s)
- Loveleen Gaur
- Amity International Business School, Amity University, Noida, India
| | - Ujwal Bhatia
- Amity International Business School, Amity University, Noida, India
| | - N. Z. Jhanjhi
- School of Computer Science and Engineering SCE, Taylor’s University, Subang Jaya, Malaysia
| | - Ghulam Muhammad
- Research Chair of Pervasive and Mobile Computing, King Saud University, Riyadh 11543, Saudi Arabia
- Computer Engineering Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
| | - Mehedi Masud
- Department of Computer Science, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif, 21944 Saudi Arabia
| |
Collapse
|
6
|
Hržić F, Tschauner S, Sorantin E, Štajduhar I. XAOM: A method for automatic alignment and orientation of radiographs for computer-aided medical diagnosis. Comput Biol Med 2021; 132:104300. [PMID: 33714842 DOI: 10.1016/j.compbiomed.2021.104300] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2020] [Revised: 02/23/2021] [Accepted: 02/24/2021] [Indexed: 12/18/2022]
Abstract
BACKGROUND AND OBJECTIVES Computer-aided diagnosis relies on machine learning algorithms that require filtered and preprocessed data as the input. Aligning the image in the desired direction is an additional manual step in post-processing, commonly overlooked due to workload issues. Several state-of-the-art approaches for fracture detection and disease-struck region segmentation benefit from correctly oriented images, thus requiring such preprocessing of X-ray images. Furthermore, it is desirable to have archived studies in a standardized format. Radiograph hanging protocols also differ from case to case, which means that images are not always aligned and oriented correctly. As a solution, the paper proposes XAOM, an X-ray Alignment and Orientation Method for images from 21 different body regions. METHODS Typically, other methods are crafted for this purpose to suit a specific body region and form of usage. In contrast, the method proposed in this paper is comprehensive and easily tuned to align and orient X-ray images of any body region. XAOM consists of two stages. For the first stage of the method, aligning X-ray images, we experimented with the following approaches: Hough transform, Fast line detection algorithm, and Principal Component Analysis method. For the second stage, we have experimented with the adaptations of several well known convolutional neural network topologies for correctly predicting image orientation: LeNet5, AlexNet, VGG16, VGG19, and ResNet50. RESULTS In the first stage, the PCA-based approach performed best. The average difference between the angle detected by the algorithm and the angle marked by the experts on the test set containing 200 pediatric X-ray images was 1.65∘, while the median value was 0.11∘. In the second stage, the VGG16-based network topology achieved the best accuracy of 0.993 on a test set containing 4,221 images. CONCLUSION XAOM is highly accurate at aligning and orienting pediatric X-ray images of 21 common body regions according to a set standard. The proposed method is also robust and can be easily adjusted to the different alignment and rotation criteria. AVAILABILITY The Python source code of the best performing implementation of XAOM is publicly available at https://github.com/fhrzic/XAOM.
Collapse
Affiliation(s)
- Franko Hržić
- University of Rijeka, Faculty of Engineering, Department of Computer Engineering, Vukovarska 58, Rijeka, 51000, Croatia; University of Rijeka, Center for Artificial Intelligence and Cybersecurity, Radmile Matejčić 2, Rijeka, 51000, Croatia
| | - Sebastian Tschauner
- Medical University of Graz, Department of Radiology, Division of Pediatric Radiology, Auenbruggerplatz 34, Graz, 8036, Austria
| | - Erich Sorantin
- Medical University of Graz, Department of Radiology, Division of Pediatric Radiology, Auenbruggerplatz 34, Graz, 8036, Austria
| | - Ivan Štajduhar
- University of Rijeka, Faculty of Engineering, Department of Computer Engineering, Vukovarska 58, Rijeka, 51000, Croatia; University of Rijeka, Center for Artificial Intelligence and Cybersecurity, Radmile Matejčić 2, Rijeka, 51000, Croatia.
| |
Collapse
|
7
|
Praveen K, Sasikala M, Janani A, Shajil N, Nishanthi V H. A simplified framework for the detection of intracranial hemorrhage in CT brain images using deep learning. Curr Med Imaging 2021; 17:1226-1236. [PMID: 33602101 DOI: 10.2174/1573405617666210218100641] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2020] [Revised: 12/24/2020] [Accepted: 12/29/2020] [Indexed: 11/22/2022]
Abstract
BACKGROUND The need for accurate and timely detection of Intracranial hemorrhage (ICH) is utmost important to avoid untoward incidents that may even lead to death.Hence, this presented work leverages the ability of a pretrained deep convolutional neural network (CNN) for the detection of ICH in computed tomography (CT) brain images. METHODS Different frameworks have been analyzed for their effectiveness for the classification of CT brain images into hemorrhage or non-hemorrhage conditions. All these frameworks were investigated on CQ500 dataset. Furthermore, an exclusive preprocessing pipeline was designed for both normal and ICH CT images. Firstly, a framework involving the pretrained deep CNN, AlexNet, has been exploited for both feature extraction and classification using the transfer learning method, secondly, a modified AlexNet-Support vector machine (SVM) classifier is explored and finally, a feature selection method, Principal Component Analysis (PCA) has been introduced in the AlexNet-SVM classifier model and its efficacy is explored.These models were trained and tested on two different sets of CT images, one containing the original images without preprocessing and another set consisting of preprocessed images. RESULTS The modified AlexNet-SVM classifier has shown an improved performance in comparison to the other investigated frameworks and has achieved a classification accuracy of 99.86%, sensitivity and specificity of 0.9986 for the detection of ICH in brain CT images. CONCLUSION This research has given an overview of a simple and efficient framework for the classification of hemorrhage and non-hemorrhage images. Also, the proposed simplified deep learning framework manifests its ability as a screening tool to assist the radiological trainees for the accurate detection of ICH.
Collapse
Affiliation(s)
- Praveen K
- Department of Electronics and Communication Engineering, College of Engineering, Guindy, Anna University, Sardar Patel Road, Chennai-600 025. India
| | - Sasikala M
- Department of Electronics and Communication Engineering, College of Engineering, Guindy, Anna University, Sardar Patel Road, Chennai-600 025. India
| | - Janani A
- Department of Electronics and Communication Engineering, College of Engineering, Guindy, Anna University, Sardar Patel Road, Chennai-600 025. India
| | - Nijisha Shajil
- Department of Electronics and Communication Engineering, College of Engineering, Guindy, Anna University, Sardar Patel Road, Chennai-600 025. India
| | - Hari Nishanthi V
- Department of Electronics and Communication Engineering, College of Engineering, Guindy, Anna University, Sardar Patel Road, Chennai-600 025. India
| |
Collapse
|
8
|
Fukushima K. Efficient IntVec: High recognition rate with reduced computational cost. Neural Netw 2019; 119:323-331. [PMID: 31499356 DOI: 10.1016/j.neunet.2019.08.024] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2019] [Revised: 07/13/2019] [Accepted: 08/23/2019] [Indexed: 10/26/2022]
Abstract
In many deep neural networks for pattern recognition, the input pattern is classified in the deepest layer based on features extracted through intermediate layers. IntVec (interpolating-vector) is known to be a powerful method for this process of classification. Although the recognition error can be made much smaller by IntVec than by WTA (winner-take-all) or even by SVM (support vector machines), IntVec requires a large computational cost. This paper proposes a new method, by which the computational cost by IntVec can be reduced drastically without increasing the recognition error. Although we basically use IntVec for recognition, we substitute it with WTA, which requires much smaller computational cost, under a certain condition. To be more specific, we first try to classify the input vector using WTA. If a class is a complete loser by WTA, we judge it also a loser by IntVec and omit the calculation of IntVec for that class. If a class is an unrivaled winner by WTA, calculation of IntVec itself can be omitted for all classes.
Collapse
Affiliation(s)
- Kunihiko Fukushima
- Fuzzy Logic Systems Institute, 680-41 Kawazu, Iizuka, Fukuoka, 820-0067, Japan.
| |
Collapse
|
9
|
Yhdego H, Li J, Morrison S, Audette M, Paolini C, Sarkar M, Okhravi H. TOWARDS MUSCULOSKELETAL SIMULATION-AWARE FALL INJURY MITIGATION: TRANSFER LEARNING WITH DEEP CNN FOR FALL DETECTION. Spring Simul Conf 2019; 2019:10.23919/springsim.2019.8732857. [PMID: 37223210 PMCID: PMC10205067 DOI: 10.23919/springsim.2019.8732857] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
This paper presents early work on a fall detection method using transfer learning method, in conjunction with a long-term effort to combine efficient machine learning and prior personalized musculoskeletal modeling to deploy fall injury mitigation in geriatric subjects. Inspired by the tremendous progress in image-based object recognition with deep convolutional neural networks (DCNNs), we opt for a pre-trained kinematics-based machine learning approach through existing large-scale annotated accelerometry datasets. The accelerometry datasets are converted to images using time-frequency analysis, based on scalograms, by computing the continuous wavelet transform filter bank. Subsequently, data augmentation is performed on these scalogram images to increase accuracy, thereby complementing limited labeled fall sensor data, enabling transfer learning from the existing pre-trained model. The experimental results on publicly available URFD datasets demonstrate that transfer learning leads to a better performance than the existing methods in the case of scarce labeled training data.
Collapse
Affiliation(s)
- Haben Yhdego
- Old Dominion University, 5115 Hampton Boulevard, Norfolk, VA, USA
| | - Jiang Li
- Old Dominion University, 5115 Hampton Boulevard, Norfolk, VA, USA
| | - Steven Morrison
- Old Dominion University, 5115 Hampton Boulevard, Norfolk, VA, USA
| | - Michel Audette
- Old Dominion University, 5115 Hampton Boulevard, Norfolk, VA, USA
| | | | - Mahasweta Sarkar
- San Diego State University, 5500 Campanile Drive, San Diego, CA, USA
| | - Hamid Okhravi
- Eastern Virginia Medical School, 825 Fairfax Ave, Norfolk VA, USA
| |
Collapse
|
10
|
Fukushima K. Margined winner-take-all: New learning rule for pattern recognition. Neural Netw 2018; 97:152-61. [PMID: 29126068 DOI: 10.1016/j.neunet.2017.10.005] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2017] [Revised: 07/10/2017] [Accepted: 10/12/2017] [Indexed: 11/20/2022]
Abstract
The neocognitron is a deep (multi-layered) convolutional neural network that can be trained to recognize visual patterns robustly. In the intermediate layers of the neocognitron, local features are extracted from input patterns. In the deepest layer, based on the features extracted in the intermediate layers, input patterns are classified into classes. A method called IntVec (interpolating-vector) is used for this purpose. This paper proposes a new learning rule called margined Winner-Take-All (mWTA) for training the deepest layer. Every time when a training pattern is presented during the learning, if the result of recognition by WTA (Winner-Take-All) is an error, a new cell is generated in the deepest layer. Here we put a certain amount of margin to the WTA. In other words, only during the learning, a certain amount of handicap is given to cells of classes other than that of the training vector, and the winner is chosen under this handicap. By introducing the margin to the WTA, we can generate a compact set of cells, with which a high recognition rate can be obtained with a small computational cost. The ability of this mWTA is demonstrated by computer simulation.
Collapse
|
11
|
Hu P, Wu F, Peng J, Bao Y, Chen F, Kong D. Automatic abdominal multi-organ segmentation using deep convolutional neural network and time-implicit level sets. Int J Comput Assist Radiol Surg 2017; 12:399-411. [PMID: 27885540 DOI: 10.1007/s11548-016-1501-5] [Citation(s) in RCA: 83] [Impact Index Per Article: 10.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2016] [Accepted: 11/03/2016] [Indexed: 10/20/2022]
Abstract
PURPOSE Multi-organ segmentation from CT images is an essential step for computer-aided diagnosis and surgery planning. However, manual delineation of the organs by radiologists is tedious, time-consuming and poorly reproducible. Therefore, we propose a fully automatic method for the segmentation of multiple organs from three-dimensional abdominal CT images. METHODS The proposed method employs deep fully convolutional neural networks (CNNs) for organ detection and segmentation, which is further refined by a time-implicit multi-phase evolution method. Firstly, a 3D CNN is trained to automatically localize and delineate the organs of interest with a probability prediction map. The learned probability map provides both subject-specific spatial priors and initialization for subsequent fine segmentation. Then, for the refinement of the multi-organ segmentation, image intensity models, probability priors as well as a disjoint region constraint are incorporated into an unified energy functional. Finally, a novel time-implicit multi-phase level-set algorithm is utilized to efficiently optimize the proposed energy functional model. RESULTS Our method has been evaluated on 140 abdominal CT scans for the segmentation of four organs (liver, spleen and both kidneys). With respect to the ground truth, average Dice overlap ratios for the liver, spleen and both kidneys are 96.0, 94.2 and 95.4%, respectively, and average symmetric surface distance is less than 1.3 mm for all the segmented organs. The computation time for a CT volume is 125 s in average. The achieved accuracy compares well to state-of-the-art methods with much higher efficiency. CONCLUSION A fully automatic method for multi-organ segmentation from abdominal CT images was developed and evaluated. The results demonstrated its potential in clinical usage with high effectiveness, robustness and efficiency.
Collapse
|