1
|
Ghobadi V, Ismail LI, Wan Hasan WZ, Ahmad H, Ramli HR, Norsahperi NMH, Tharek A, Hanapiah FA. Challenges and solutions of deep learning-based automated liver segmentation: A systematic review. Comput Biol Med 2025; 185:109459. [PMID: 39642700 DOI: 10.1016/j.compbiomed.2024.109459] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2024] [Revised: 11/12/2024] [Accepted: 11/19/2024] [Indexed: 12/09/2024]
Abstract
The liver is one of the vital organs in the body. Precise liver segmentation in medical images is essential for liver disease treatment. The deep learning-based liver segmentation process faces several challenges. This research aims to analyze the challenges of liver segmentation in prior studies and identify the modifications made to network models and other enhancements implemented by researchers to tackle each challenge. In total, 88 articles from Scopus and ScienceDirect databases published between January 2016 and January 2022 have been studied. The liver segmentation challenges are classified into five main categories, each containing some subcategories. For each challenge, the proposed technique to overcome the challenge is investigated. The provided report details the authors, publication years, dataset types, imaging technologies, and evaluation metrics of all references for comparison. Additionally, a summary table outlines the challenges and solutions.
Collapse
Affiliation(s)
- Vahideh Ghobadi
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Luthffi Idzhar Ismail
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Wan Zuha Wan Hasan
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Haron Ahmad
- KPJ Specialist Hospital, Damansara Utama, Petaling Jaya, 47400, Selangor, Malaysia.
| | - Hafiz Rashidi Ramli
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | | | - Anas Tharek
- Hospital Sultan Abdul Aziz Shah, University Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Fazah Akhtar Hanapiah
- Faculty of Medicine, Universiti Teknologi MARA, Damansara Utama, Sungai Buloh, 47000, Selangor, Malaysia.
| |
Collapse
|
2
|
Yue Z, Jiang J, Hou W, Zhou Q, David Spence J, Fenster A, Qiu W, Ding M. Prior-Knowledge Embedded U-Net-Based Fully Automatic Vessel Wall Volume Measurement of the Carotid Artery in 3D Ultrasound Image. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:711-727. [PMID: 39255086 DOI: 10.1109/tmi.2024.3457245] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/12/2024]
Abstract
The vessel-wall-volume (VWV) measured based on three-dimensional (3D) carotid artery (CA) ultrasound (US) images can help to assess carotid atherosclerosis and manage patients at risk for stroke. Manual involvement for measurement work is subjective and requires well-trained operators, and fully automatic measurement tools are not yet available. Thereby, we proposed a fully automatic VWV measurement framework (Auto-VWV) using a CA prior-knowledge embedded U-Net (CAP-UNet) to measure the VWV from 3D CA US images without manual intervention. The Auto-VWV framework is designed to improve the repeated VWV measuring consistency, which resulted in the first fully automatic framework for VWV measurement. CAP-UNet is developed to improve segmentation accuracy on the whole CA, which composed of a U-Net type backbone and three additional prior-knowledge learning modules. Specifically, a continuity learning module is used to learn the spatial continuity of the arteries in a sequence of image slices. A voxel evolution learning module was designed to learn the evolution of the artery in adjacent slices, and a topology learning module was used to learn the unique topology of the carotid artery. In two 3D CA US datasets, CAP-UNet architecture achieved state-of-the-art performance compared to eight competing models. Furthermore, CAP-UNet-based Auto-VWV achieved better accuracy and consistency than Auto-VWV based on competing models in the simulated repeated measurement. Finally, using 10 pairs of real repeatedly scanned samples, Auto-VWV achieved better VWV measurement reproducibility than intra- and inter-operator manual measurements. The code is available at https://github.com/Yue9603/Auto-VWV.
Collapse
|
3
|
Delmoral JC, R S Tavares JM. Semantic Segmentation of CT Liver Structures: A Systematic Review of Recent Trends and Bibliometric Analysis : Neural Network-based Methods for Liver Semantic Segmentation. J Med Syst 2024; 48:97. [PMID: 39400739 PMCID: PMC11473507 DOI: 10.1007/s10916-024-02115-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Accepted: 10/02/2024] [Indexed: 10/15/2024]
Abstract
The use of artificial intelligence (AI) in the segmentation of liver structures in medical images has become a popular research focus in the past half-decade. The performance of AI tools in screening for this task may vary widely and has been tested in the literature in various datasets. However, no scientometric report has provided a systematic overview of this scientific area. This article presents a systematic and bibliometric review of recent advances in neuronal network modeling approaches, mainly of deep learning, to outline the multiple research directions of the field in terms of algorithmic features. Therefore, a detailed systematic review of the most relevant publications addressing fully automatic semantic segmenting liver structures in Computed Tomography (CT) images in terms of algorithm modeling objective, performance benchmark, and model complexity is provided. The review suggests that fully automatic hybrid 2D and 3D networks are the top performers in the semantic segmentation of the liver. In the case of liver tumor and vasculature segmentation, fully automatic generative approaches perform best. However, the reported performance benchmark indicates that there is still much to be improved in segmenting such small structures in high-resolution abdominal CT scans.
Collapse
Affiliation(s)
- Jessica C Delmoral
- Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal
| | - João Manuel R S Tavares
- Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal.
| |
Collapse
|
4
|
Li Y, Wong D, Sreng S, Chung J, Toh A, Yuan H, Eppenberger LS, Leow C, Ting D, Liu N, Schmetterer L, Saw SM, Jonas JB, Chia A, Ang M. Effect of childhood atropine treatment on adult choroidal thickness using sequential deep learning-enabled segmentation. Asia Pac J Ophthalmol (Phila) 2024; 13:100107. [PMID: 39378966 DOI: 10.1016/j.apjo.2024.100107] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Revised: 09/20/2024] [Accepted: 09/26/2024] [Indexed: 10/10/2024] Open
Abstract
PURPOSE To describe choroidal thickness measurements using a sequential deep learning segmentation in adults who received childhood atropine treatment for myopia control. DESIGN Prospective, observational study. METHODS Choroidal thickness was measured by swept-source optical coherence tomography in adults who received childhood atropine, segmented using a sequential deep learning approach. RESULTS Of 422 eyes, 94 (22.3 %) had no previous exposure to atropine treatment, while 328 (77.7 %) had received topical atropine during childhood. After adjusting for age, sex, and axial length, childhood atropine exposure was associated with a thicker choroid by 32.1 μm (95 % CI, 9.2-55.0; P = 0.006) in the inner inferior, 23.5 μm (95 % CI, 1.9-45.1; P = 0.03) in the outer inferior, 21.8 μm (95 % CI, 0.76-42.9; P = 0.04) in the inner nasal, and 21.8 μm (95 % CI, 2.6-41.0; P = 0.03) in the outer nasal. Multivariable analysis, adjusted for age, sex, atropine use, and axial length, showed an independent association between central subfield choroidal thickness and the incidence of tessellated fundus (P < 0.001; OR, 0.97; 95 % CI, 0.96-0.98). CONCLUSIONS This study demonstrated that short-term (2-4 years) atropine treatment during childhood was associated with an increase in choroidal thickness of 20-40 μm in adulthood (10-20 years later), after adjusting for age, sex, and axial length. We also observed an independent association between eyes with thicker central choroidal measurements and reduced incidence of tessellated fundus. Our study suggests that childhood exposure to atropine treatment may affect choroidal thickness in adulthood.
Collapse
Affiliation(s)
- Yong Li
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Duke-NUS Medical School, National University of Singapore, Singapore, Singapore
| | - Damon Wong
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; SERI-NTU Advanced Ocular Engineering (STANCE), Singapore, Singapore; School of Chemistry, Chemical Engineering and Biotechnology, Nanyang Technological University, Singapore, Singapore
| | - Syna Sreng
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; SERI-NTU Advanced Ocular Engineering (STANCE), Singapore, Singapore
| | - Joey Chung
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
| | - Angeline Toh
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
| | - Han Yuan
- Duke-NUS Medical School, National University of Singapore, Singapore, Singapore
| | - Leila Sara Eppenberger
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Health Sciences and Technology, ETH Zurich, Zurich, Switzerland
| | - Cheryl Leow
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
| | - Daniel Ting
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Duke-NUS Medical School, National University of Singapore, Singapore, Singapore; Byers Eye Institute, Sandford University, Palo Alto, CA, USA
| | - Nan Liu
- Duke-NUS Medical School, National University of Singapore, Singapore, Singapore
| | - Leopold Schmetterer
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Duke-NUS Medical School, National University of Singapore, Singapore, Singapore; SERI-NTU Advanced Ocular Engineering (STANCE), Singapore, Singapore; School of Chemistry, Chemical Engineering and Biotechnology, Nanyang Technological University, Singapore, Singapore; Department of Clinical Pharmacology, Medical University Vienna, Vienna, Austria; Center for Medical Physics and Biomedical Engineering, Medical University Vienna, Vienna, Austria; Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland
| | - Seang-Mei Saw
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Duke-NUS Medical School, National University of Singapore, Singapore, Singapore; Saw Swee Hock School of Public Health, National University of Singapore, Singapore, Singapore
| | - Jost B Jonas
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland; Privatpraxis Prof Jonas and Dr Panda-Jonas, Heidelberg, Germany
| | - Audrey Chia
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Duke-NUS Medical School, National University of Singapore, Singapore, Singapore
| | - Marcus Ang
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore; Duke-NUS Medical School, National University of Singapore, Singapore, Singapore.
| |
Collapse
|
5
|
da Silva RLB, Yang S, Kim D, Kim JH, Lim SH, Han J, Kim JM, Kim JE, Huh KH, Lee SS, Heo MS, Yi WJ. Automatic segmentation and classification of frontal sinuses for sex determination from CBCT scans using a two-stage anatomy-guided attention network. Sci Rep 2024; 14:11750. [PMID: 38782964 PMCID: PMC11116511 DOI: 10.1038/s41598-024-62211-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Accepted: 05/14/2024] [Indexed: 05/25/2024] Open
Abstract
Sex determination is essential for identifying unidentified individuals, particularly in forensic contexts. Traditional methods for sex determination involve manual measurements of skeletal features on CBCT scans. However, these manual measurements are labor-intensive, time-consuming, and error-prone. The purpose of this study was to automatically and accurately determine sex on a CBCT scan using a two-stage anatomy-guided attention network (SDetNet). SDetNet consisted of a 2D frontal sinus segmentation network (FSNet) and a 3D anatomy-guided attention network (SDNet). FSNet segmented frontal sinus regions in the CBCT images and extracted regions of interest (ROIs) near them. Then, the ROIs were fed into SDNet to predict sex accurately. To improve sex determination performance, we proposed multi-channel inputs (MSIs) and an anatomy-guided attention module (AGAM), which encouraged SDetNet to learn differences in the anatomical context of the frontal sinus between males and females. SDetNet showed superior sex determination performance in the area under the receiver operating characteristic curve, accuracy, Brier score, and specificity compared with the other 3D CNNs. Moreover, the results of ablation studies showed a notable improvement in sex determination with the embedding of both MSI and AGAM. Consequently, SDetNet demonstrated automatic and accurate sex determination by learning the anatomical context information of the frontal sinus on CBCT scans.
Collapse
Affiliation(s)
- Renan Lucio Berbel da Silva
- Discipline of Oral Radiology, Department of Stomatology, School of Dentistry, University of São Paulo, São Paulo, SP, Brazil
| | - Su Yang
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, 08826, South Korea
| | - DaEl Kim
- Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, Seoul, 08826, South Korea
| | - Jun Ho Kim
- Discipline of Oral Radiology, Department of Stomatology, School of Dentistry, University of São Paulo, São Paulo, SP, Brazil
| | - Sang-Heon Lim
- Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, Seoul, 08826, South Korea
| | - Jiyong Han
- Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, Seoul, 08826, South Korea
| | - Jun-Min Kim
- Department of Electronics and Information Engineering, Hansung University, Seoul, 02876, South Korea
| | - Jo-Eun Kim
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, 03080, South Korea
| | - Kyung-Hoe Huh
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, 03080, South Korea
| | - Sam-Sun Lee
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, 03080, South Korea
| | - Min-Suk Heo
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, 03080, South Korea.
| | - Won-Jin Yi
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, 08826, South Korea.
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, 03080, South Korea.
| |
Collapse
|
6
|
Huang Y, Yang J, Sun Q, Yuan Y, Li H, Hou Y. Multi-residual 2D network integrating spatial correlation for whole heart segmentation. Comput Biol Med 2024; 172:108261. [PMID: 38508056 DOI: 10.1016/j.compbiomed.2024.108261] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Revised: 02/21/2024] [Accepted: 03/06/2024] [Indexed: 03/22/2024]
Abstract
Whole heart segmentation (WHS) has significant clinical value for cardiac anatomy, modeling, and analysis of cardiac function. This study aims to address the WHS accuracy on cardiac CT images, as well as the fast inference speed and low graphics processing unit (GPU) memory consumption required by practical clinical applications. Thus, we propose a multi-residual two-dimensional (2D) network integrating spatial correlation for WHS. The network performs slice-by-slice segmentation on three-dimensional cardiac CT images in a 2D encoder-decoder manner. In the network, a convolutional long short-term memory skip connection module is designed to perform spatial correlation feature extraction on the feature maps at different resolutions extracted by the sub-modules of the pre-trained ResNet-based encoder. Moreover, a decoder based on the multi-residual module is designed to analyze the extracted features from the perspectives of multi-scale and channel attention, thereby accurately delineating the various substructures of the heart. The proposed method is verified on a dataset of the multi-modality WHS challenge, an in-house WHS dataset, and a dataset of the abdominal organ segmentation challenge. The dice, Jaccard, average symmetric surface distance, Hausdorff distance, inference time, and maximum GPU memory of the WHS are 0.914, 0.843, 1.066 mm, 15.778 mm, 9.535 s, and 1905 MB, respectively. The proposed network has high accuracy, fast inference speed, minimal GPU memory consumption, strong robustness, and good generalization. It can be deployed to clinical practical applications for WHS and can be effectively extended and applied to other multi-organ segmentation fields. The source code is publicly available at https://github.com/nancy1984yan/MultiResNet-SC.
Collapse
Affiliation(s)
- Yan Huang
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China; School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Jinzhu Yang
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China; School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China; National Frontiers Science Center for Industrial Intelligence and Systems Optimization, Shenyang, Liaoning, China.
| | - Qi Sun
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China; School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Yuliang Yuan
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China; School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Honghe Li
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China; School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Yang Hou
- Department of Radiology, Shengjing Hospital of China Medical University, Shenyang, Liaoning, China
| |
Collapse
|
7
|
Silveira A, Greving I, Longo E, Scheel M, Weitkamp T, Fleck C, Shahar R, Zaslansky P. Deep learning to overcome Zernike phase-contrast nanoCT artifacts for automated micro-nano porosity segmentation in bone. JOURNAL OF SYNCHROTRON RADIATION 2024; 31:136-149. [PMID: 38095668 PMCID: PMC10833422 DOI: 10.1107/s1600577523009852] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 11/13/2023] [Indexed: 01/09/2024]
Abstract
Bone material contains a hierarchical network of micro- and nano-cavities and channels, known as the lacuna-canalicular network (LCN), that is thought to play an important role in mechanobiology and turnover. The LCN comprises micrometer-sized lacunae, voids that house osteocytes, and submicrometer-sized canaliculi that connect bone cells. Characterization of this network in three dimensions is crucial for many bone studies. To quantify X-ray Zernike phase-contrast nanotomography data, deep learning is used to isolate and assess porosity in artifact-laden tomographies of zebrafish bones. A technical solution is proposed to overcome the halo and shade-off domains in order to reliably obtain the distribution and morphology of the LCN in the tomographic data. Convolutional neural network (CNN) models are utilized with increasing numbers of images, repeatedly validated by `error loss' and `accuracy' metrics. U-Net and Sensor3D CNN models were trained on data obtained from two different synchrotron Zernike phase-contrast transmission X-ray microscopes, the ANATOMIX beamline at SOLEIL (Paris, France) and the P05 beamline at PETRA III (Hamburg, Germany). The Sensor3D CNN model with a smaller batch size of 32 and a training data size of 70 images showed the best performance (accuracy 0.983 and error loss 0.032). The analysis procedures, validated by comparison with human-identified ground-truth images, correctly identified the voids within the bone matrix. This proposed approach may have further application to classify structures in volumetric images that contain non-linear artifacts that degrade image quality and hinder feature identification.
Collapse
Affiliation(s)
- Andreia Silveira
- Department for Restorative, Preventive and Pediatric Dentistry, Charité-Universitaetsmedizin, Berlin, Germany
| | - Imke Greving
- Institute of Materials Physics, Helmholtz-Zentrum Hereon, Geesthacht, Germany
| | - Elena Longo
- Elettra – Sincrotrone Trieste SCpA, Basovizza, Trieste, Italy
| | | | | | - Claudia Fleck
- Fachgebiet Werkstofftechnik / Chair of Materials Science and Engineering, Institute of Materials Science and Technology, Faculty III Process Sciences, Technische Universität Berlin, Berlin, Germany
| | - Ron Shahar
- Koret School of Veterinary Medicine, The Robert H. Smith Faculty of Agriculture, Food and Environmental Sciences, Hebrew University of Jerusalem, Rehovot, Israel
| | - Paul Zaslansky
- Department for Restorative, Preventive and Pediatric Dentistry, Charité-Universitaetsmedizin, Berlin, Germany
| |
Collapse
|
8
|
Ikuta M, Zhang J. A Deep Convolutional Gated Recurrent Unit for CT Image Reconstruction. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:10612-10625. [PMID: 35522637 DOI: 10.1109/tnnls.2022.3169569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Computed tomography (CT) is one of the most important medical imaging technologies in use today. Most commercial CT products use a technique known as the filtered backprojection (FBP) that is fast and can produce decent image quality when an X-ray dose is high. However, the FBP is not good enough on low-dose X-ray CT imaging because the CT image reconstruction problem becomes more stochastic. A more effective reconstruction technique proposed recently and implemented in a limited number of CT commercial products is an iterative reconstruction (IR). The IR technique is based on a Bayesian formulation of the CT image reconstruction problem with an explicit model of the CT scanning, including its stochastic nature, and a prior model that incorporates our knowledge about what a good CT image should look like. However, constructing such prior knowledge is more complicated than it seems. In this article, we propose a novel neural network for CT image reconstruction. The network is based on the IR formulation and constructed with a recurrent neural network (RNN). Specifically, we transform the gated recurrent unit (GRU) into a neural network performing CT image reconstruction. We call it "GRU reconstruction." This neural network conducts concurrent dual-domain learning. Many deep learning (DL)-based methods in medical imaging are single-domain learning, but dual-domain learning performs better because it learns from both the sinogram and the image domain. In addition, we propose backpropagation through stage (BPTS) as a new RNN backpropagation algorithm. It is similar to the backpropagation through time (BPTT) of an RNN; however, it is tailored for iterative optimization. Results from extensive experiments indicate that our proposed method outperforms conventional model-based methods, single-domain DL methods, and state-of-the-art DL techniques in terms of the root mean squared error (RMSE), the peak signal-to-noise ratio (PSNR), and the structure similarity (SSIM) and in terms of visual appearance.
Collapse
|
9
|
Chen Y, Zheng C, Zhang W, Lin H, Chen W, Zhang G, Xu G, Wu F. MS-FANet: Multi-scale feature attention network for liver tumor segmentation. Comput Biol Med 2023; 163:107208. [PMID: 37421737 DOI: 10.1016/j.compbiomed.2023.107208] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Revised: 06/07/2023] [Accepted: 06/25/2023] [Indexed: 07/10/2023]
Abstract
Accurate segmentation of liver tumors is a prerequisite for early diagnosis of liver cancer. Segmentation networks extract features continuously at the same scale, which cannot adapt to the variation of liver tumor volume in computed tomography (CT). Hence, a multi-scale feature attention network (MS-FANet) for liver tumor segmentation is proposed in this paper. The novel residual attention (RA) block and multi-scale atrous downsampling (MAD) are introduced in the encoder of MS-FANet to sufficiently learn variable tumor features and extract tumor features at different scales simultaneously. The dual-path feature (DF) filter and dense upsampling (DU) are introduced in the feature reduction process to reduce effective features for the accurate segmentation of liver tumors. On the public LiTS dataset and 3DIRCADb dataset, MS-FANet achieved 74.2% and 78.0% of average Dice, respectively, outperforming most state-of-the-art networks, this strongly proves the excellent liver tumor segmentation performance and the ability to learn features at different scales.
Collapse
Affiliation(s)
- Ying Chen
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Cheng Zheng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Wei Zhang
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Hongping Lin
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Wang Chen
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Guimei Zhang
- Institute of Computer Vision, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Guohui Xu
- Department of Hepatobiliary Surgery, Jiangxi Cancer Hospital, Nanchang, 330029, PR China.
| | - Fang Wu
- Department of Gastroenterology, First Affiliated Hospital of Wenzhou Medical University, Wenzhou, 325035, PR China.
| |
Collapse
|
10
|
Dai M, Xiao G, Shao M, Zhang YS. The Synergy between Deep Learning and Organs-on-Chips for High-Throughput Drug Screening: A Review. BIOSENSORS 2023; 13:389. [PMID: 36979601 PMCID: PMC10046732 DOI: 10.3390/bios13030389] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/17/2022] [Revised: 02/22/2023] [Accepted: 03/07/2023] [Indexed: 06/18/2023]
Abstract
Organs-on-chips (OoCs) are miniature microfluidic systems that have arguably become a class of advanced in vitro models. Deep learning, as an emerging topic in machine learning, has the ability to extract a hidden statistical relationship from the input data. Recently, these two areas have become integrated to achieve synergy for accelerating drug screening. This review provides a brief description of the basic concepts of deep learning used in OoCs and exemplifies the successful use cases for different types of OoCs. These microfluidic chips are of potential to be assembled as highly potent human-on-chips with complex physiological or pathological functions. Finally, we discuss the future supply with perspectives and potential challenges in terms of combining OoCs and deep learning for image processing and automation designs.
Collapse
Affiliation(s)
- Manna Dai
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350108, China
- Computing and Intelligence Department, Institute of High Performance Computing (IHPC), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore 138632, Singapore
| | - Gao Xiao
- College of Environment and Safety Engineering, Fuzhou University, Fuzhou 350108, China
- Department of Biomedical Engineering, Tsinghua University, Beijing 100084, China
| | - Ming Shao
- Department of Computer and Information Science, College of Engineering, University of Massachusetts Dartmouth, North Dartmouth, MA 02747, USA
| | - Yu Shrike Zhang
- Division of Engineering in Medicine, Department of Medicine, Brigham and Women’s Hospital, Harvard Medical School, Cambridge, MA 02139, USA
| |
Collapse
|
11
|
Bonaldi L, Pretto A, Pirri C, Uccheddu F, Fontanella CG, Stecco C. Deep Learning-Based Medical Images Segmentation of Musculoskeletal Anatomical Structures: A Survey of Bottlenecks and Strategies. Bioengineering (Basel) 2023; 10:137. [PMID: 36829631 PMCID: PMC9952222 DOI: 10.3390/bioengineering10020137] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2022] [Revised: 01/13/2023] [Accepted: 01/17/2023] [Indexed: 01/22/2023] Open
Abstract
By leveraging the recent development of artificial intelligence algorithms, several medical sectors have benefited from using automatic segmentation tools from bioimaging to segment anatomical structures. Segmentation of the musculoskeletal system is key for studying alterations in anatomical tissue and supporting medical interventions. The clinical use of such tools requires an understanding of the proper method for interpreting data and evaluating their performance. The current systematic review aims to present the common bottlenecks for musculoskeletal structures analysis (e.g., small sample size, data inhomogeneity) and the related strategies utilized by different authors. A search was performed using the PUBMED database with the following keywords: deep learning, musculoskeletal system, segmentation. A total of 140 articles published up until February 2022 were obtained and analyzed according to the PRISMA framework in terms of anatomical structures, bioimaging techniques, pre/post-processing operations, training/validation/testing subset creation, network architecture, loss functions, performance indicators and so on. Several common trends emerged from this survey; however, the different methods need to be compared and discussed based on each specific case study (anatomical region, medical imaging acquisition setting, study population, etc.). These findings can be used to guide clinicians (as end users) to better understand the potential benefits and limitations of these tools.
Collapse
Affiliation(s)
- Lorenza Bonaldi
- Department of Civil, Environmental and Architectural Engineering, University of Padova, Via F. Marzolo 9, 35131 Padova, Italy
| | - Andrea Pretto
- Department of Industrial Engineering, University of Padova, Via Venezia 1, 35121 Padova, Italy
| | - Carmelo Pirri
- Department of Neuroscience, University of Padova, Via A. Gabelli 65, 35121 Padova, Italy
| | - Francesca Uccheddu
- Department of Industrial Engineering, University of Padova, Via Venezia 1, 35121 Padova, Italy
- Centre for Mechanics of Biological Materials (CMBM), University of Padova, Via F. Marzolo 9, 35131 Padova, Italy
| | - Chiara Giulia Fontanella
- Department of Industrial Engineering, University of Padova, Via Venezia 1, 35121 Padova, Italy
- Centre for Mechanics of Biological Materials (CMBM), University of Padova, Via F. Marzolo 9, 35131 Padova, Italy
| | - Carla Stecco
- Department of Neuroscience, University of Padova, Via A. Gabelli 65, 35121 Padova, Italy
- Centre for Mechanics of Biological Materials (CMBM), University of Padova, Via F. Marzolo 9, 35131 Padova, Italy
| |
Collapse
|
12
|
Chen Y, Zheng C, Zhou T, Feng L, Liu L, Zeng Q, Wang G. A deep residual attention-based U-Net with a biplane joint method for liver segmentation from CT scans. Comput Biol Med 2023; 152:106421. [PMID: 36527780 DOI: 10.1016/j.compbiomed.2022.106421] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Revised: 11/17/2022] [Accepted: 12/11/2022] [Indexed: 12/15/2022]
Abstract
Liver tumours are diseases with high morbidity and high deterioration probabilities, and accurate liver area segmentation from computed tomography (CT) scans is a prerequisite for quick tumour diagnosis. While 2D network segmentation methods can perform segmentation with lower device performance requirements, they often discard the rich 3D spatial information contained in CT scans, limiting their segmentation accuracy. Hence, a deep residual attention-based U-shaped network (DRAUNet) with a biplane joint method for liver segmentation is proposed in this paper, where the biplane joint method introduces coronal CT slices to assist the transverse slices with segmentation, incorporating more 3D spatial information into the segmentation results to improve the segmentation performance of the network. Additionally, a novel deep residual block (DR block) and dual-effect attention module (DAM) are introduced in DRAUNet, where the DR block has deeper layers and two shortcut paths. The DAM efficiently combines the correlations of feature channels and the spatial locations of feature maps. The DRAUNet with the biplane joint method is tested on three datasets, Liver Tumour Segmentation (LiTS), 3D Image Reconstruction for Comparison of Algorithms Database (3DIRCADb), and Segmentation of the Liver Competition 2007 (Sliver07), and it achieves 97.3%, 97.4%, and 96.9% Dice similarity coefficients (DSCs) for liver segmentation, respectively, outperforming most state-of-the-art networks; this strongly demonstrates the segmentation performance of DRAUNet and the ability of the biplane joint method to obtain 3D spatial information from 3D images.
Collapse
Affiliation(s)
- Ying Chen
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Cheng Zheng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Taohui Zhou
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Longfeng Feng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Lan Liu
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China.
| | - Qiao Zeng
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China
| | - Guoqing Wang
- Zhejiang Suosi Technology Co. Ltd, Wenzhou, 325000, PR China.
| |
Collapse
|
13
|
Multi-Organ Segmentation Using a Low-Resource Architecture. INFORMATION 2022. [DOI: 10.3390/info13100472] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Since their inception, deep-learning architectures have shown promising results for automatic segmentation. However, despite the technical advances introduced by fully convolutional networks, generative adversarial networks or recurrent neural networks, and their usage in hybrid architectures, automatic segmentation in the medical field is still not used at scale. One main reason is related to data scarcity and quality, which in turn generates a lack of annotated data that hinder the generalization of the models. The second main issue refers to challenges in training deep models. This process uses large amounts of GPU memory (that might exceed current hardware limitations) and requires high training times. In this article, we want to prove that despite these issues, good results can be obtained even when using a lower resource architecture, thus opening the way for more researchers to employ and use deep neural networks. In achieving the multi-organ segmentation, we are employing modern pre-processing techniques, a smart model design and fusion between several models trained on the same dataset. Our architecture is compared against state-of-the-art methods employed in a publicly available challenge and the notable results prove the effectiveness of our method.
Collapse
|
14
|
Canal-Net for automatic and robust 3D segmentation of mandibular canals in CBCT images using a continuity-aware contextual network. Sci Rep 2022; 12:13460. [PMID: 35931733 PMCID: PMC9356068 DOI: 10.1038/s41598-022-17341-6] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2022] [Accepted: 07/25/2022] [Indexed: 02/01/2023] Open
Abstract
The purpose of this study was to propose a continuity-aware contextual network (Canal-Net) for the automatic and robust 3D segmentation of the mandibular canal (MC) with high consistent accuracy throughout the entire MC volume in cone-beam CT (CBCT) images. The Canal-Net was designed based on a 3D U-Net with bidirectional convolutional long short-term memory (ConvLSTM) under a multi-task learning framework. Specifically, the Canal-Net learned the 3D anatomical context information of the MC by incorporating spatio-temporal features from ConvLSTM, and also the structural continuity of the overall MC volume under a multi-task learning framework using multi-planar projection losses complementally. The Canal-Net showed higher segmentation accuracies in 2D and 3D performance metrics (p < 0.05), and especially, a significant improvement in Dice similarity coefficient scores and mean curve distance (p < 0.05) throughout the entire MC volume compared to other popular deep learning networks. As a result, the Canal-Net achieved high consistent accuracy in 3D segmentations of the entire MC in spite of the areas of low visibility by the unclear and ambiguous cortical bone layer. Therefore, the Canal-Net demonstrated the automatic and robust 3D segmentation of the entire MC volume by improving structural continuity and boundary details of the MC in CBCT images.
Collapse
|
15
|
Ansari MY, Abdalla A, Ansari MY, Ansari MI, Malluhi B, Mohanty S, Mishra S, Singh SS, Abinahed J, Al-Ansari A, Balakrishnan S, Dakua SP. Practical utility of liver segmentation methods in clinical surgeries and interventions. BMC Med Imaging 2022; 22:97. [PMID: 35610600 PMCID: PMC9128093 DOI: 10.1186/s12880-022-00825-2] [Citation(s) in RCA: 31] [Impact Index Per Article: 10.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2022] [Accepted: 05/09/2022] [Indexed: 12/15/2022] Open
Abstract
Clinical imaging (e.g., magnetic resonance imaging and computed tomography) is a crucial adjunct for clinicians, aiding in the diagnosis of diseases and planning of appropriate interventions. This is especially true in malignant conditions such as hepatocellular carcinoma (HCC), where image segmentation (such as accurate delineation of liver and tumor) is the preliminary step taken by the clinicians to optimize diagnosis, staging, and treatment planning and intervention (e.g., transplantation, surgical resection, radiotherapy, PVE, embolization, etc). Thus, segmentation methods could potentially impact the diagnosis and treatment outcomes. This paper comprehensively reviews the literature (during the year 2012-2021) for relevant segmentation methods and proposes a broad categorization based on their clinical utility (i.e., surgical and radiological interventions) in HCC. The categorization is based on the parameters such as precision, accuracy, and automation.
Collapse
|
16
|
Han X, Yu Z, Zhuo Y, Zhao B, Ren Y, Lamm L, Xue X, Feng J, Marr C, Shan F, Peng T, Zhang XY. The value of longitudinal clinical data and paired CT scans in predicting the deterioration of COVID-19 revealed by an artificial intelligence system. iScience 2022; 25:104227. [PMID: 35434542 PMCID: PMC8989658 DOI: 10.1016/j.isci.2022.104227] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2021] [Revised: 03/10/2022] [Accepted: 04/05/2022] [Indexed: 01/09/2023] Open
Abstract
The respective value of clinical data and CT examinations in predicting COVID-19 progression is unclear, because the CT scans and clinical data previously used are not synchronized in time. To address this issue, we collected 119 COVID-19 patients with 341 longitudinal CT scans and paired clinical data, and we developed an AI system for the prediction of COVID-19 deterioration. By combining features extracted from CT and clinical data with our system, we can predict whether a patient will develop severe symptoms during hospitalization. Complementary to clinical data, CT examinations show significant add-on values for the prediction of COVID-19 progression in the early stage of COVID-19, especially in the 6th to 8th day after the symptom onset, indicating that this is the ideal time window for the introduction of CT examinations. We release our AI system to provide clinicians with additional assistance to optimize CT usage in the clinical workflow. COVID-19 patients with 341 longitudinal CT scans and paired clinical data included A new AI model for the prediction of COVID-19 progression was developed CT scans show significant add-on value over clinical data for the prediction Day 6–8 after the onset of COVID-19 symptoms is an ideal time window for a CT scan
Collapse
Affiliation(s)
- Xiaoyang Han
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China.,Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Ministry of Education, Shanghai 200433, China
| | - Ziqi Yu
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China.,Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Ministry of Education, Shanghai 200433, China
| | - Yaoyao Zhuo
- Department of Radiology, Zhongshan Hospital, Fudan University, Shanghai 200032, China.,Department of Radiology, Shanghai Public Health Clinical Center, Fudan University, Shanghai 201508, China
| | - Botao Zhao
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China.,Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Ministry of Education, Shanghai 200433, China
| | - Yan Ren
- Department of Radiology, Huashan Hospital, Fudan University, Shanghai 200433, China
| | - Lorenz Lamm
- Institute of AI for Health, Helmholtz Zentrum München, Ingolstädter Landstraße 1, D-85764 Neuherberg, Germany.,Helmholtz AI, Helmholtz Zentrum München, Ingolstädter Landstraße 1, D-85764 Neuherberg, Germany
| | - Xiangyang Xue
- Shanghai Key Laboratory of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai 200433, China
| | - Jianfeng Feng
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China.,Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Ministry of Education, Shanghai 200433, China
| | - Carsten Marr
- Institute of AI for Health, Helmholtz Zentrum München, Ingolstädter Landstraße 1, D-85764 Neuherberg, Germany
| | - Fei Shan
- Department of Radiology, Shanghai Public Health Clinical Center, Fudan University, Shanghai 201508, China
| | - Tingying Peng
- Helmholtz AI, Helmholtz Zentrum München, Ingolstädter Landstraße 1, D-85764 Neuherberg, Germany
| | - Xiao-Yong Zhang
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China.,Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence, Ministry of Education, Shanghai 200433, China.,MOE Frontiers Center for Brain Science, Fudan University, Shanghai 200433, China
| |
Collapse
|
17
|
Tang K, Meyer Q, White R, Armstrong RT, Mostaghimi P, Da Wang Y, Liu S, Zhao C, Regenauer-Lieb K, Tung PKM. Deep learning for full-feature X-ray microcomputed tomography segmentation of proton electron membrane fuel cells. Comput Chem Eng 2022. [DOI: 10.1016/j.compchemeng.2022.107768] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023]
|
18
|
Dong P, Yuan H, Allahverdi A, Raveenthiran J, Piché N, Provencher B, Bassim ND. Advanced characterization of 3D structure and porosity of ordinary portland cement (OPC) mortar using plasma focused ion beam tomography and X-ray computed tomography. J Microsc 2022; 287:19-31. [PMID: 35415878 DOI: 10.1111/jmi.13105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2021] [Revised: 03/15/2022] [Accepted: 04/06/2022] [Indexed: 11/28/2022]
Abstract
The visualization and quantification of pore networks and main phases have been critical research topics in cementitious materials as many critical mechanical and chemical properties and infrastructure reliability rely on these 3-D characteristics. In this study, we realized the mesoscale serial sectioning and analysis up to ∼80 μm by ∼90 μm by ∼60 μm on portland cement mortar using plasma focused ion beam (PFIB) for the first time. The workflow of working with mortar and PFIB was established applying a prepositioned hard silicon mask to reduce curtaining. Segmentation with minimal human interference was performed using a trained neural network, in which multiple types of segmentation models were compared. Combining PFIB analysis at microscale with X-ray micro-computed tomography, the analysis of capillary pores and air voids ranging from hundreds of nanometers (nm) to millimeters (mm) can be conducted. The volume fraction of large capillary pores and air voids are 11.5% and 12.7%, respectively. Moreover, the skeletonization of connected capillary pores clearly shows fluid transport pathways, which is a key factor determining durability performance of concrete in aggressive environments. Another interesting aspect of the FIB tomography is the reconstruction of anhydrous phases, which could enable direct study of hydration kinetics of individual cement phases. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Peng Dong
- Department of Materials Science and Engineering, McMaster University, 1280 Main St W, Hamilton, ON, L8S 4L8, Canada
| | - Hui Yuan
- Canadian Centre for Electron Microscopy, McMaster University, 1280 Main St W, Hamilton, ON, L8S 4M1, Canada
| | - Ali Allahverdi
- Cement Research Centre, School of Chemical Engineering, Iran University of Science and Technology, Narmak, Tehran, 16846 - 13114, Iran
| | - Jatheeshan Raveenthiran
- Department of Electrical and Computer Engineering, McMaster University, 1280 Main St W, Hamilton, ON, L8S 4L8, Canada
| | - Nicolas Piché
- Object Research Systems, 760 St-Paul West, Suite 101, Montreal, Quebec, H3C 1M4, Canada
| | - Benjamin Provencher
- Object Research Systems, 760 St-Paul West, Suite 101, Montreal, Quebec, H3C 1M4, Canada
| | - Nabil D Bassim
- Department of Materials Science and Engineering, McMaster University, 1280 Main St W, Hamilton, ON, L8S 4L8, Canada.,Canadian Centre for Electron Microscopy, McMaster University, 1280 Main St W, Hamilton, ON, L8S 4M1, Canada
| |
Collapse
|
19
|
Liao J, Liu L, Duan H, Huang Y, Zhou L, Chen L, Wang C. Using a Convolutional Neural Network and Convolutional Long Short-term Memory to Automatically Detect Aneurysms on 2D Digital Subtraction Angiography Images: Framework Development and Validation. JMIR Med Inform 2022; 10:e28880. [PMID: 35294371 PMCID: PMC8968557 DOI: 10.2196/28880] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2021] [Revised: 06/27/2021] [Accepted: 01/16/2022] [Indexed: 11/26/2022] Open
Abstract
BACKGROUND It is hard to distinguish cerebral aneurysms from overlapping vessels in 2D digital subtraction angiography (DSA) images due to these images' lack of spatial information. OBJECTIVE The aims of this study were to (1) construct a deep learning diagnostic system to improve the ability to detect posterior communicating artery aneurysms on 2D DSA images and (2) validate the efficiency of the deep learning diagnostic system in 2D DSA aneurysm detection. METHODS We proposed a 2-stage detection system. First, we established the region localization stage to automatically locate specific detection regions of raw 2D DSA sequences. Second, in the intracranial aneurysm detection stage, we constructed a bi-input+RetinaNet+convolutional long short-term memory (C-LSTM) framework to compare its performance for aneurysm detection with that of 3 existing frameworks. Each of the frameworks had a 5-fold cross-validation scheme. The receiver operating characteristic curve, the area under the curve (AUC) value, mean average precision, sensitivity, specificity, and accuracy were used to assess the abilities of different frameworks. RESULTS A total of 255 patients with posterior communicating artery aneurysms and 20 patients without aneurysms were included in this study. The best AUC values of the RetinaNet, RetinaNet+C-LSTM, bi-input+RetinaNet, and bi-input+RetinaNet+C-LSTM frameworks were 0.95, 0.96, 0.92, and 0.97, respectively. The mean sensitivities of the RetinaNet, RetinaNet+C-LSTM, bi-input+RetinaNet, and bi-input+RetinaNet+C-LSTM frameworks and human experts were 89% (range 67.02%-98.43%), 88% (range 65.76%-98.06%), 87% (range 64.53%-97.66%), 89% (range 67.02%-98.43%), and 90% (range 68.30%-98.77%), respectively. The mean specificities of the RetinaNet, RetinaNet+C-LSTM, bi-input+RetinaNet, and bi-input+RetinaNet+C-LSTM frameworks and human experts were 80% (range 56.34%-94.27%), 89% (range 67.02%-98.43%), 86% (range 63.31%-97.24%), 93% (range 72.30%-99.56%), and 90% (range 68.30%-98.77%), respectively. The mean accuracies of the RetinaNet, RetinaNet+C-LSTM, bi-input+RetinaNet, and bi-input+RetinaNet+C-LSTM frameworks and human experts were 84.50% (range 69.57%-93.97%), 88.50% (range 74.44%-96.39%), 86.50% (range 71.97%-95.22%), 91% (range 77.63%-97.72%), and 90% (range 76.34%-97.21%), respectively. CONCLUSIONS According to our results, more spatial and temporal information can help improve the performance of the frameworks. Therefore, the bi-input+RetinaNet+C-LSTM framework had the best performance when compared to that of the other frameworks. Our study demonstrates that our system can assist physicians in detecting intracranial aneurysms on 2D DSA images.
Collapse
Affiliation(s)
- JunHua Liao
- Department of Neurosurgery, West China Hospital, Sichuan University, Chengdu, China
- College of Computer Science, Sichuan University, Chengdu, China
| | - LunXin Liu
- Department of Neurosurgery, West China Hospital, Sichuan University, Chengdu, China
| | - HaiHan Duan
- School of Science and Engineering, The Chinese University of Hong Kong, Shenzhen, China
| | - YunZhi Huang
- School of Automation, Nanjing University of Information Science and Technology, Nanjing, China
| | - LiangXue Zhou
- Department of Neurosurgery, West China Hospital, Sichuan University, Chengdu, China
| | - LiangYin Chen
- College of Computer Science, Sichuan University, Chengdu, China
| | - ChaoHua Wang
- Department of Neurosurgery, West China Hospital, Sichuan University, Chengdu, China
| |
Collapse
|
20
|
Yang S, Tong S, Zhu G, Cao J, Wang Y, Xue Z, Sun H, Wen Y. MVE-FLK: A multi-task legal judgment prediction via multi-view encoder fusing legal keywords. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2021.107960] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
21
|
Duncan KE, Czymmek KJ, Jiang N, Thies AC, Topp CN. X-ray microscopy enables multiscale high-resolution 3D imaging of plant cells, tissues, and organs. PLANT PHYSIOLOGY 2022; 188:831-845. [PMID: 34618094 PMCID: PMC8825331 DOI: 10.1093/plphys/kiab405] [Citation(s) in RCA: 26] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/04/2021] [Accepted: 07/29/2021] [Indexed: 05/12/2023]
Abstract
Capturing complete internal anatomies of plant organs and tissues within their relevant morphological context remains a key challenge in plant science. While plant growth and development are inherently multiscale, conventional light, fluorescence, and electron microscopy platforms are typically limited to imaging of plant microstructure from small flat samples that lack a direct spatial context to, and represent only a small portion of, the relevant plant macrostructures. We demonstrate technical advances with a lab-based X-ray microscope (XRM) that bridge the imaging gap by providing multiscale high-resolution three-dimensional (3D) volumes of intact plant samples from the cell to the whole plant level. Serial imaging of a single sample is shown to provide sub-micron 3D volumes co-registered with lower magnification scans for explicit contextual reference. High-quality 3D volume data from our enhanced methods facilitate sophisticated and effective computational segmentation. Advances in sample preparation make multimodal correlative imaging workflows possible, where a single resin-embedded plant sample is scanned via XRM to generate a 3D cell-level map, and then used to identify and zoom in on sub-cellular regions of interest for high-resolution scanning electron microscopy. In total, we present the methodologies for use of XRM in the multiscale and multimodal analysis of 3D plant features using numerous economically and scientifically important plant systems.
Collapse
Affiliation(s)
- Keith E Duncan
- Donald Danforth Plant Science Center, St Louis, Missouri 63132, USA
| | - Kirk J Czymmek
- Donald Danforth Plant Science Center, St Louis, Missouri 63132, USA
| | - Ni Jiang
- Donald Danforth Plant Science Center, St Louis, Missouri 63132, USA
| | | | - Christopher N Topp
- Donald Danforth Plant Science Center, St Louis, Missouri 63132, USA
- Author for communication:
| |
Collapse
|
22
|
Li X, Bala R, Monga V. Robust Deep 3D Blood Vessel Segmentation Using Structural Priors. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:1271-1284. [PMID: 34990361 DOI: 10.1109/tip.2021.3139241] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Deep learning has enabled significant improvements in the accuracy of 3D blood vessel segmentation. Open challenges remain in scenarios where labeled 3D segmentation maps for training are severely limited, as is often the case in practice, and in ensuring robustness to noise. Inspired by the observation that 3D vessel structures project onto 2D image slices with informative and unique edge profiles, we propose a novel deep 3D vessel segmentation network guided by edge profiles. Our network architecture comprises a shared encoder and two decoders that learn segmentation maps and edge profiles jointly. 3D context is mined in both the segmentation and edge prediction branches by employing bidirectional convolutional long-short term memory (BCLSTM) modules. 3D features from the two branches are concatenated to facilitate learning of the segmentation map. As a key contribution, we introduce new regularization terms that: a) capture the local homogeneity of 3D blood vessel volumes in the presence of biomarkers; and b) ensure performance robustness to domain-specific noise by suppressing false positive responses. Experiments on benchmark datasets with ground truth labels reveal that the proposed approach outperforms state-of-the-art techniques on standard measures such as DICE overlap and mean Intersection-over-Union. The performance gains of our method are even more pronounced when training is limited. Furthermore, the computational cost of our network inference is among the lowest compared with state-of-the-art.
Collapse
|
23
|
Chen Y, Hu F, Wang Y, Zheng C. Hybrid-attention densely connected U-Net with GAP for extracting livers from CT volumes. Med Phys 2022; 49:1015-1033. [PMID: 35015305 DOI: 10.1002/mp.15435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2021] [Revised: 12/08/2021] [Accepted: 12/14/2021] [Indexed: 11/08/2022] Open
Abstract
PURPOSE Liver segmentation is an important step in the clinical treatment of liver cancer, and accurate and automatic liver segmentation methods are extremely important. U-Net has been used as the benchmark for many medical segmentation networks, but it cannot fully utilize low-resolution information and global contextual information. To solve these problems, we propose a new network architecture named the hybrid-attention densely connected U-Net (HDU-Net). METHODS The proposed HDU-Net has three main changes relative to U-Net, as follows: (1) It uses a densely connected structure and dilated convolution to achieve feature reuse and avoid information loss. (2) A global average pooling block is proposed to further augment the receptive field and improve the segmentation accuracy of the network for small or disconnected liver regions. (3) By combining the spatial attention and channel attention mechanisms, a hybrid attention structure is proposed to replace the skip connection component to filter and integrate low-resolution information. RESULTS Experiments conducted on the LITS2017, 3Dircadb and Sliver07 datasets show that the proposed model can segment the liver accurately and effectively. Dice scores reach 96.5%, 96.18%, and 97.57% on these datasets, respectively, constituting results that are superior to many previously proposed methods. CONCLUSIONS The experimental liver segmentation results have demonstrated that our proposed network provides improved segmentation performance in comparison with other networks. The experimental results without postprocessing confirmed that our network solves the oversegmentation and undersegmentation problems to some extent. The proposed model is effective, robust, and efficient in terms of liver segmentation without requiring extensive training time or a very large dataset.
Collapse
Affiliation(s)
- Ying Chen
- School of Software, Nanchang Hangkong University, Nanchang, China
| | - Fei Hu
- School of Software, Nanchang Hangkong University, Nanchang, China
| | - Yerong Wang
- School of Software, Nanchang Hangkong University, Nanchang, China
| | - Cheng Zheng
- School of Software, Nanchang Hangkong University, Nanchang, China
| |
Collapse
|
24
|
Cahyo DAY, Yow AP, Saw SM, Ang M, Girard M, Schmetterer L, Wong D. Multi-task learning approach for volumetric segmentation and reconstruction in 3D OCT images. BIOMEDICAL OPTICS EXPRESS 2021; 12:7348-7360. [PMID: 35003838 PMCID: PMC8713660 DOI: 10.1364/boe.428140] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/22/2021] [Revised: 07/11/2021] [Accepted: 08/27/2021] [Indexed: 06/14/2023]
Abstract
The choroid is the vascular layer of the eye that supplies photoreceptors with oxygen. Changes in the choroid are associated with many pathologies including myopia where the choroid progressively thins due to axial elongation. To quantize these changes, there is a need to automatically and accurately segment the choroidal layer from optical coherence tomography (OCT) images. In this paper, we propose a multi-task learning approach to segment the choroid from three-dimensional OCT images. Our proposed architecture aggregates the spatial context from adjacent cross-sectional slices to reconstruct the central slice. Spatial context learned by this reconstruction mechanism is then fused with a U-Net based architecture for segmentation. The proposed approach was evaluated on volumetric OCT scans of 166 myopic eyes acquired with a commercial OCT system, and achieved a cross-validation Intersection over Union (IoU) score of 94.69% which significantly outperformed (p<0.001) the other state-of-the-art methods on the same data set. Choroidal thickness maps generated by our approach also achieved a better structural similarity index (SSIM) of 72.11% with respect to the groundtruth. In particular, our approach performs well for highly challenging eyes with thinner choroids. Compared to other methods, our proposed approach also requires lesser processing time and has lower computational requirements. The results suggest that our proposed approach could potentially be used as a fast and reliable method for automated choroidal segmentation.
Collapse
Affiliation(s)
- Dheo A. Y. Cahyo
- SERI-NTU Advanced Ocular Engineering (STANCE), Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Ai Ping Yow
- SERI-NTU Advanced Ocular Engineering (STANCE), Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
- Singapore Centre for Environmental Life Sciences Engineering (SCELSE), Singapore
| | - Seang-Mei Saw
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Marcus Ang
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Michael Girard
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Leopold Schmetterer
- SERI-NTU Advanced Ocular Engineering (STANCE), Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
- School of Chemical and Biomedical Engineering, Nanyang Technological University, Singapore
- Department of Clinical Pharmacology, Medical University of Vienna, Austria
- Center for Medical Physics and Biomedical Engineering, Medical University of Vienna, Austria
- Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland
| | - Damon Wong
- SERI-NTU Advanced Ocular Engineering (STANCE), Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
- School of Chemical and Biomedical Engineering, Nanyang Technological University, Singapore
- Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland
| |
Collapse
|
25
|
Javadov M, Karatay E, Ulusan K, Ozpek A, Idiz O, Duren M, Sari S, Demircan F, Demirel G, Dagdeviren H, Yigit A, Kelestimur F, Aysan E. Number of cells in parathyroid tissue in primary hyperparathyroidism cases and its relationship with serum calcium value. Medicine (Baltimore) 2021; 100:e27530. [PMID: 34797277 PMCID: PMC8601366 DOI: 10.1097/md.0000000000027530] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/29/2021] [Revised: 07/30/2021] [Accepted: 09/30/2021] [Indexed: 01/05/2023] Open
Abstract
BACKGROUND The relationship between serum calcium (Ca) level to serum parathyroid hormone (PTH), phosphorus (P) levels and tissue properties of the parathyroid gland is unknown in primary hyperparathyroidism cases. Revealing this relationship may be useful for understanding the etiopathogenesis of primary hyperparathyroidism and determining the time of treatment. METHODS Ninety patients (71 females, 19 males, age range; 27-73 years, average age; 46) who underwent single gland excision with the diagnosis of primary hyperparathyroidism were studied. The patients were divided into 2 groups as serum Ca level <12 and serum Ca level ≥12. Age and sex of the patients, mean cell number of the gland, mean volume of the gland, serum levels of PTH, P, and histopathologic type of hyperplasia were evaluated. RESULTS The mean cell number per cubic centimeter is 22.9 (10-220 range) million in all glands. Serum Ca level was <12 in 82 (91.1%) of the patients, and ≥12 in 8 (8.9%) cases. Mean cell number of the gland, mean volume of the gland, existence of cystic hyperplasia of the gland, serum levels of PTH and P were statistically significant between the 2 groups (P < .001, P < .001, P < .05, P < .001, P < .05 respectively). CONCLUSION In primary hyperparathyroidism cases serum Ca level is not related to age and sex but directly related to proportionals to the cell number and volume of the gland and serum levels of PTH, inversely related to cystic hyperplasia and serum levels of P. Early surgical intervention should be planned since the serum Ca level will be high in large adenomas with a noncystic radiological appearance.
Collapse
Affiliation(s)
- Mirkhalig Javadov
- Yeditepe University, Faculty of Medicine, Department of General Surgery, Turkey
| | - Emrah Karatay
- Marmara University Pendik Training and Research Hospital, Department of Radiology, Turkey
| | - Kivilcim Ulusan
- Health Sciences University Istanbul Hospital, Department of General Surgery, Turkey
| | - Adnan Ozpek
- Health Sciences University, Umraniye Hospital, Department of General Surgery, Turkey
| | - Oğuz Idiz
- Health Sciences University Istanbul Hospital, Department of General Surgery, Turkey
| | - Mete Duren
- Acibadem Maslak Hospital, Department of General Surgery, Turkey
| | - Serkan Sari
- Health Sciences University, Başakşehir Cam ve Sakura Hospital, Department of General Surgery, Turkey
| | - Firat Demircan
- Yeditepe University, Faculty of Medicine, Department of General Surgery, Turkey
| | - Gulderen Demirel
- Yeditepe University, Faculty of Medicine, Department of Immunology, Turkey
| | - Husniye Dagdeviren
- Yeditepe University, Faculty of Medicine, Department of Immunology, Turkey
| | - Ayse Yigit
- Yeditepe University, Faculty of Medicine, Department of Immunology, Turkey
| | - Fahrettin Kelestimur
- Yeditepe University, Faculty of Medicine, Department of Internal Medicine, Turkey
| | - Erhan Aysan
- Yeditepe University, Faculty of Medicine, Department of General Surgery, Turkey
| |
Collapse
|
26
|
Chanti DA, Duque VG, Crouzier M, Nordez A, Lacourpaille L, Mateus D. IFSS-Net: Interactive Few-Shot Siamese Network for Faster Muscle Segmentation and Propagation in Volumetric Ultrasound. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2615-2628. [PMID: 33560982 DOI: 10.1109/tmi.2021.3058303] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
We present an accurate, fast and efficient method for segmentation and muscle mask propagation in 3D freehand ultrasound data, towards accurate volume quantification. A deep Siamese 3D Encoder-Decoder network that captures the evolution of the muscle appearance and shape for contiguous slices is deployed. We use it to propagate a reference mask annotated by a clinical expert. To handle longer changes of the muscle shape over the entire volume and to provide an accurate propagation, we devise a Bidirectional Long Short Term Memory module. Also, to train our model with a minimal amount of training samples, we propose a strategy combining learning from few annotated 2D ultrasound slices with sequential pseudo-labeling of the unannotated slices. We introduce a decremental update of the objective function to guide the model convergence in the absence of large amounts of annotated data. After training with a few volumes, the decremental update strategy switches from a weak supervised training to a few-shot setting. Finally, to handle the class-imbalance between foreground and background muscle pixels, we propose a parametric Tversky loss function that learns to penalize adaptively the false positives and the false negatives. We validate our approach for the segmentation, label propagation, and volume computation of the three low-limb muscles on a dataset of 61600 images from 44 subjects. We achieve a Dice score coefficient of over 95% and a volumetric error of 1.6035 ± 0.587%.
Collapse
|
27
|
Trimpl MJ, Boukerroui D, Stride EPJ, Vallis KA, Gooding MJ. Interactive contouring through contextual deep learning. Med Phys 2021; 48:2951-2959. [PMID: 33742454 DOI: 10.1002/mp.14852] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2020] [Revised: 01/31/2021] [Accepted: 03/10/2021] [Indexed: 01/27/2023] Open
Abstract
PURPOSE To investigate a deep learning approach that enables three-dimensional (3D) segmentation of an arbitrary structure of interest given a user provided two-dimensional (2D) contour for context. Such an approach could decrease delineation times and improve contouring consistency, particularly for anatomical structures for which no automatic segmentation tools exist. METHODS A series of deep learning segmentation models using a Recurrent Residual U-Net with attention gates was trained with a successively expanding training set. Contextual information was provided to the models, using a previously contoured slice as an input, in addition to the slice to be contoured. In total, 6 models were developed, and 19 different anatomical structures were used for training and testing. Each of the models was evaluated for all 19 structures, even if they were excluded from the training set, in order to assess the model's ability to segment unseen structures of interest. Each model's performance was evaluated using the Dice similarity coefficient (DSC), Hausdorff distance, and relative added path length (APL). RESULTS The segmentation performance for seen and unseen structures improved when the training set was expanded by addition of structures previously excluded from the training set. A model trained exclusively on heart structures achieved a DSC of 0.33, HD of 44 mm, and relative APL of 0.85 when segmenting the spleen, whereas a model trained on a diverse set of structures, but still excluding the spleen, achieved a DSC of 0.80, HD of 13 mm, and relative APL of 0.35. Iterative prediction performed better compared to direct prediction when considering unseen structures. CONCLUSIONS Training a contextual deep learning model on a diverse set of structures increases the segmentation performance for the structures in the training set, but importantly enables the model to generalize and make predictions even for unseen structures that were not represented in the training set. This shows that user-provided context can be incorporated into deep learning contouring to facilitate semi-automatic segmentation of CT images for any given structure. Such an approach can enable faster de-novo contouring in clinical practice.
Collapse
Affiliation(s)
- Michael J Trimpl
- Mirada Medical Ltd, Oxford, UK
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK
- Oxford Institute for Radiation Oncology, University of Oxford, Oxford, UK
| | | | - Eleanor P J Stride
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK
| | - Katherine A Vallis
- Oxford Institute for Radiation Oncology, University of Oxford, Oxford, UK
| | | |
Collapse
|
28
|
Qiu B, van der Wel H, Kraeima J, Hendrik Glas H, Guo J, Borra RJH, Witjes MJH, van Ooijen PMA. Robust and Accurate Mandible Segmentation on Dental CBCT Scans Affected by Metal Artifacts Using a Prior Shape Model. J Pers Med 2021; 11:364. [PMID: 34062762 PMCID: PMC8147374 DOI: 10.3390/jpm11050364] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2021] [Revised: 04/26/2021] [Accepted: 04/27/2021] [Indexed: 12/17/2022] Open
Abstract
Accurate mandible segmentation is significant in the field of maxillofacial surgery to guide clinical diagnosis and treatment and develop appropriate surgical plans. In particular, cone-beam computed tomography (CBCT) images with metal parts, such as those used in oral and maxillofacial surgery (OMFS), often have susceptibilities when metal artifacts are present such as weak and blurred boundaries caused by a high-attenuation material and a low radiation dose in image acquisition. To overcome this problem, this paper proposes a novel deep learning-based approach (SASeg) for automated mandible segmentation that perceives overall mandible anatomical knowledge. SASeg utilizes a prior shape feature extractor (PSFE) module based on a mean mandible shape, and recurrent connections maintain the continuity structure of the mandible. The effectiveness of the proposed network is substantiated on a dental CBCT dataset from orthodontic treatment containing 59 patients. The experiments show that the proposed SASeg can be easily used to improve the prediction accuracy in a dental CBCT dataset corrupted by metal artifacts. In addition, the experimental results on the PDDCA dataset demonstrate that, compared with the state-of-the-art mandible segmentation models, our proposed SASeg can achieve better segmentation performance.
Collapse
Affiliation(s)
- Bingjiang Qiu
- 3D Lab, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (B.Q.); (H.v.d.W.); (H.H.G.); (M.J.H.W.)
- Department of Radiation Oncology, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (J.G.); (P.M.A.v.O.)
- Data Science Center in Health (DASH), University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| | - Hylke van der Wel
- 3D Lab, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (B.Q.); (H.v.d.W.); (H.H.G.); (M.J.H.W.)
- Department of Oral and Maxillofacial Surgery, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| | - Joep Kraeima
- 3D Lab, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (B.Q.); (H.v.d.W.); (H.H.G.); (M.J.H.W.)
- Department of Oral and Maxillofacial Surgery, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| | - Haye Hendrik Glas
- 3D Lab, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (B.Q.); (H.v.d.W.); (H.H.G.); (M.J.H.W.)
- Department of Oral and Maxillofacial Surgery, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| | - Jiapan Guo
- Department of Radiation Oncology, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (J.G.); (P.M.A.v.O.)
- Data Science Center in Health (DASH), University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| | - Ronald J. H. Borra
- Medical Imaging Center (MIC), University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands;
| | - Max Johannes Hendrikus Witjes
- 3D Lab, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (B.Q.); (H.v.d.W.); (H.H.G.); (M.J.H.W.)
- Department of Oral and Maxillofacial Surgery, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| | - Peter M. A. van Ooijen
- Department of Radiation Oncology, University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands; (J.G.); (P.M.A.v.O.)
- Data Science Center in Health (DASH), University Medical Center Groningen, University of Groningen, Hanzeplein 1, 9713 GZ Groningen, The Netherlands
| |
Collapse
|
29
|
Tan M, Wu F, Kong D, Mao X. Automatic liver segmentation using 3D convolutional neural networks with a hybrid loss function. Med Phys 2021; 48:1707-1719. [PMID: 33496971 DOI: 10.1002/mp.14732] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/13/2020] [Revised: 01/13/2021] [Accepted: 01/18/2021] [Indexed: 11/06/2022] Open
Abstract
PURPOSE Automatic liver segmentation from abdominal computed tomography (CT) images is a fundamental task in computer-assisted liver surgery programs. Many liver segmentation algorithms are very sensitive to fuzzy boundaries and heterogeneous pathologies, especially when the data are scarce. To solve these problems, we propose an automatic liver segmentation framework based on three-dimensional (3D) convolutional neural networks with a hybrid loss function. METHODS Two networks are incorporated in our method with the first being a liver shape autoencoder that is trained to obtain compressed codes of liver shapes, and the second being a liver segmentation network that is trained with a hybrid loss function. The design of the hybrid loss function is comprised of three parts. The first part is an adaptively weighted cross-entropy loss, which pays more attention to misclassified pixels. The second part is an edge-preserving smoothness loss, which guarantees that the adjacent pixels with the same label have similar outputs, while dissimilar for pixels with different labels. The third part of the loss is a shape constraint to model high-level structural differences based on the learned shape codes. Both networks use 3D operations for data processing. In our experiments, data augmentation is performed at both the training and the test stage. RESULTS We extensively evaluated our method on two datasets: the Segmentation of the Liver Competition 2007 (Sliver07), and the Combined (CT-MR) Healthy Abdominal Organ Segmentation (CHAOS) Challenge. Finally, with only 20 training scans, we achieved the best score of 82.55 on the Sliver07 challenge, and a score of 83.02 on the CHAOS challenge. CONCLUSIONS In this study, we proposed a novel hybrid loss to overcome the difficulties in liver segmentation. The quantitative and qualitative results demonstrate that our method is highly suited for pathological liver segmentation, even when trained with a small dataset.
Collapse
Affiliation(s)
- Man Tan
- The School of Mathematical Sciences, Zhejiang University, Hangzhou, Zhejiang, 310058, China
| | - Fa Wu
- The School of Mathematical Sciences, Zhejiang University, Hangzhou, Zhejiang, 310058, China
| | - Dexing Kong
- The School of Mathematical Sciences, Zhejiang University, Hangzhou, Zhejiang, 310058, China
| | - Xiongwei Mao
- The Radiology Department, The Hospital of Zhejiang University, Hangzhou, Zhejiang, 310058, China.,The Radiology Department, Zhejiang University School of Medicine First Affiliated Hospital, Hangzhou, Zhejiang, 310003, China
| |
Collapse
|
30
|
Vu MH, Grimbergen G, Nyholm T, Löfstedt T. Evaluation of multislice inputs to convolutional neural networks for medical image segmentation. Med Phys 2020; 47:6216-6231. [PMID: 33169365 DOI: 10.1002/mp.14391] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2020] [Revised: 06/09/2020] [Accepted: 07/07/2020] [Indexed: 01/17/2023] Open
Abstract
PURPOSE When using convolutional neural networks (CNNs) for segmentation of organs and lesions in medical images, the conventional approach is to work with inputs and outputs either as single slice [two-dimensional (2D)] or whole volumes [three-dimensional (3D)]. One common alternative, in this study denoted as pseudo-3D, is to use a stack of adjacent slices as input and produce a prediction for at least the central slice. This approach gives the network the possibility to capture 3D spatial information, with only a minor additional computational cost. METHODS In this study, we systematically evaluate the segmentation performance and computational costs of this pseudo-3D approach as a function of the number of input slices, and compare the results to conventional end-to-end 2D and 3D CNNs, and to triplanar orthogonal 2D CNNs. The standard pseudo-3D method regards the neighboring slices as multiple input image channels. We additionally design and evaluate a novel, simple approach where the input stack is a volumetric input that is repeatably convolved in 3D to obtain a 2D feature map. This 2D map is in turn fed into a standard 2D network. We conducted experiments using two different CNN backbone architectures and on eight diverse data sets covering different anatomical regions, imaging modalities, and segmentation tasks. RESULTS We found that while both pseudo-3D methods can process a large number of slices at once and still be computationally much more efficient than fully 3D CNNs, a significant improvement over a regular 2D CNN was only observed with two of the eight data sets. triplanar networks had the poorest performance of all the evaluated models. An analysis of the structural properties of the segmentation masks revealed no relations to the segmentation performance with respect to the number of input slices. A post hoc rank sum test which combined all metrics and data sets yielded that only our newly proposed pseudo-3D method with an input size of 13 slices outperformed almost all methods. CONCLUSION In the general case, multislice inputs appear not to improve segmentation results over using 2D or 3D CNNs. For the particular case of 13 input slices, the proposed novel pseudo-3D method does appear to have a slight advantage across all data sets compared to all other methods evaluated in this work.
Collapse
Affiliation(s)
- Minh H Vu
- Department of Radiation Sciences, Umeå University, Umeå, Sweden
| | - Guus Grimbergen
- Department of Biomedical Engineering, Eindhoven University of Technology, Eindhoven, 5612 AZ, the Netherlands
| | - Tufve Nyholm
- Department of Radiation Sciences, Umeå University, Umeå, Sweden
| | - Tommy Löfstedt
- Department of Radiation Sciences, Umeå University, Umeå, Sweden
| |
Collapse
|
31
|
Cahyo DAY, Wong DWK, Yow AP, Saw SM, Schmetterer L. Volumetric Choroidal Segmentation Using Sequential Deep Learning Approach in High Myopia Subjects. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2020:1286-1289. [PMID: 33018223 DOI: 10.1109/embc44109.2020.9176184] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Many ocular diseases are associated with choroidal changes. Therefore, it is crucial to be able to segment the choroid to study its properties. Previous methods for choroidal segmentation have focused on single cross-sectional scans. Volumetric choroidal segmentation has yet to be widely reported. In this paper, we propose a sequential segmentation approach using a variation of U-Net with a bidirectional C-LSTM(Convolutional Long Short Term Memory) module in the bottleneck region. The model is evaluated on volumetric scans from 40 high myopia subjects, obtained using SS-OCT(Swept Source Optical Coherence Tomography). A comparison with other U-Net-based variants is also presented. The results demonstrate that volumetric segmentation of the choroid can be achieved with an accuracy of IoU(Intersection over Union) 0.92.Clinical relevance- This deep learning approach can automatically segment the choroidal volume, which can enable better evaluation and monitoring at ocular diseases.
Collapse
|
32
|
van de Leemput SC, Prokop M, van Ginneken B, Manniesing R. Stacked Bidirectional Convolutional LSTMs for Deriving 3D Non-Contrast CT From Spatiotemporal 4D CT. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:985-996. [PMID: 31484111 DOI: 10.1109/tmi.2019.2939044] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
The imaging workup in acute stroke can be simplified by deriving non-contrast CT (NCCT) from CT perfusion (CTP) images. This results in reduced workup time and radiation dose. To achieve this, we present a stacked bidirectional convolutional LSTM (C-LSTM) network to predict 3D volumes from 4D spatiotemporal data. Several parameterizations of the C-LSTM network were trained on a set of 17 CTP-NCCT pairs to learn to derive a NCCT from CTP and were subsequently quantitatively evaluated on a separate cohort of 16 cases. The results show that the C-LSTM network clearly outperforms the baseline and competitive convolutional neural network methods. We show good scalability and performance of the method by continued training and testing on an independent dataset which includes pathology of 80 and 83 CTP-NCCT pairs, respectively. C-LSTM is, therefore, a promising general deep learning approach to learn from high-dimensional spatiotemporal medical images.
Collapse
|
33
|
Fan G, Liu H, Wu Z, Li Y, Feng C, Wang D, Luo J, Wells WM, He S. Deep Learning-Based Automatic Segmentation of Lumbosacral Nerves on CT for Spinal Intervention: A Translational Study. AJNR Am J Neuroradiol 2019; 40:1074-1081. [PMID: 31147353 DOI: 10.3174/ajnr.a6070] [Citation(s) in RCA: 24] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2019] [Accepted: 04/16/2019] [Indexed: 12/18/2022]
Abstract
BACKGROUND AND PURPOSE 3D reconstruction of a targeted area ("safe" triangle and Kambin triangle) may benefit the viability assessment of transforaminal epidural steroid injection, especially at the L5/S1 level. However, manual segmentation of lumbosacral nerves for 3D reconstruction is time-consuming. The aim of this study was to investigate the feasibility of deep learning-based segmentation of lumbosacral nerves on CT and the reconstruction of the safe triangle and Kambin triangle. MATERIALS AND METHODS A total of 50 cases of spinal CT were manually labeled for lumbosacral nerves and bones using Slicer 4.8. The ratio of training/validation/testing was 32:8:10. A 3D U-Net was adopted to build the model SPINECT for automatic segmentations of lumbosacral structures. The Dice score, pixel accuracy, and Intersection over Union were computed to assess the segmentation performance of SPINECT. The areas of Kambin and safe triangles were measured to validate the 3D reconstruction. RESULTS The results revealed successful segmentation of lumbosacral bone and nerve on CT. The average pixel accuracy for bone was 0.940, and for nerve, 0.918. The average Intersection over Union for bone was 0.897 and for nerve, 0.827. The Dice score for bone was 0.945, and for nerve, it was 0.905. There were no significant differences in the quantified Kambin triangle or safe triangle between manually segmented images and automatically segmented images (P > .05). CONCLUSIONS Deep learning-based automatic segmentation of lumbosacral structures (nerves and bone) on routine CT is feasible, and SPINECT-based 3D reconstruction of safe and Kambin triangles is also validated.
Collapse
Affiliation(s)
- G Fan
- From the Orthopedic Department, Shanghai Tenth People's Hospital (G.F., C.F., D.W., S.H.), Tongji University School of Medicine, Shanghai, China .,Department of Spine Surgery (G.F.), Third Affiliated Hospital of Sun Yatsen University, Guangzhou, China.,Surgical Planning Lab (G.F., J.L., W.M.W.), Radiology Department, Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts
| | - H Liu
- Spinal Pain Research Institute of Tongji University (H.L., C.F., D.W., S.H.), Shanghai, China
| | - Z Wu
- School of Data and Computer Science (Z.W.), Sun Yat-sen University, Guangzhou, China
| | - Y Li
- Shanghai Jiao Tong University School of Medicine (Y.L.), Shanghai, China
| | - C Feng
- From the Orthopedic Department, Shanghai Tenth People's Hospital (G.F., C.F., D.W., S.H.), Tongji University School of Medicine, Shanghai, China.,Spinal Pain Research Institute of Tongji University (H.L., C.F., D.W., S.H.), Shanghai, China
| | - D Wang
- From the Orthopedic Department, Shanghai Tenth People's Hospital (G.F., C.F., D.W., S.H.), Tongji University School of Medicine, Shanghai, China.,Spinal Pain Research Institute of Tongji University (H.L., C.F., D.W., S.H.), Shanghai, China
| | - J Luo
- Surgical Planning Lab (G.F., J.L., W.M.W.), Radiology Department, Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts.,Graduate School of Frontier Sciences (J.L.), University of Tokyo, Tokyo, Japan
| | - W M Wells
- Surgical Planning Lab (G.F., J.L., W.M.W.), Radiology Department, Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts
| | - S He
- From the Orthopedic Department, Shanghai Tenth People's Hospital (G.F., C.F., D.W., S.H.), Tongji University School of Medicine, Shanghai, China .,Spinal Pain Research Institute of Tongji University (H.L., C.F., D.W., S.H.), Shanghai, China
| |
Collapse
|