1
|
Yue G, Zhuo G, Zhou T, Liu W, Wang T, Jiang Q. Adaptive Cross-Feature Fusion Network With Inconsistency Guidance for Multi-Modal Brain Tumor Segmentation. IEEE J Biomed Health Inform 2025; 29:3148-3158. [PMID: 38150339 DOI: 10.1109/jbhi.2023.3347556] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2023]
Abstract
In the context of contemporary artificial intelligence, increasing deep learning (DL) based segmentation methods have been recently proposed for brain tumor segmentation (BraTS) via analysis of multi-modal MRI. However, known DL-based works usually directly fuse the information of different modalities at multiple stages without considering the gap between modalities, leaving much room for performance improvement. In this paper, we introduce a novel deep neural network, termed ACFNet, for accurately segmenting brain tumor in multi-modal MRI. Specifically, ACFNet has a parallel structure with three encoder-decoder streams. The upper and lower streams generate coarse predictions from individual modality, while the middle stream integrates the complementary knowledge of different modalities and bridges the gap between them to yield fine prediction. To effectively integrate the complementary information, we propose an adaptive cross-feature fusion (ACF) module at the encoder that first explores the correlation information between the feature representations from upper and lower streams and then refines the fused correlation information. To bridge the gap between the information from multi-modal data, we propose a prediction inconsistency guidance (PIG) module at the decoder that helps the network focus more on error-prone regions through a guidance strategy when incorporating the features from the encoder. The guidance is obtained by calculating the prediction inconsistency between upper and lower streams and highlights the gap between multi-modal data. Extensive experiments on the BraTS 2020 dataset show that ACFNet is competent for the BraTS task with promising results and outperforms six mainstream competing methods.
Collapse
|
2
|
Liu J, Liu F, Nie D, Gu Y, Sun Y, Shen D. Structure-Aware Brain Tissue Segmentation for Isointense Infant MRI Data Using Multi-Phase Multi-Scale Assistance Network. IEEE J Biomed Health Inform 2025; 29:1297-1307. [PMID: 39302775 DOI: 10.1109/jbhi.2024.3452310] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/22/2024]
Abstract
Accurate and automatic brain tissue segmentation is crucial for tracking brain development and diagnosing brain disorders. However, due to inherently ongoing myelination and maturation during the first postnatal year, the intensity distributions of gray matter and white matter in the infant brain MRI at the age of around 6 months old (a.k.a. isointense phase) are highly overlapped, which makes tissue segmentation very challenging, even for experts. To address this issue, in this study, we propose a multi-phase multi-scale assistance segmentation framework, which comprises a structure-preserved generative adversarial network (SPGAN) and a multi-phase multi-scale assisted segmentation network (MASN). SPGAN bi-directionally synthesizes isointense and adult-like data. The synthetic isointense data essentially augment the training dataset, combined with high-quality annotations transferred from its adult-like counterpart. By contrast, the synthetic adult-like data offers clear tissue structures and is concatenated with isointense data to serve as the input of MASN. In particular, MASN is designed with two-branch networks, which simultaneously segment tissues with two phases (isointense and adult-like) and two scales by also preserving their correspondences. We further propose a boundary refinement module to extract maximum gradients from local feature maps to indicate tissue boundaries, prompting MASN to focus more on boundaries where segmentation errors are prone to occur. Extensive experiments on the National Database for Autism Research and Baby Connectome Project datasets quantitatively and qualitatively demonstrate the superiority of our proposed framework compared with seven state-of-the-art methods.
Collapse
|
3
|
Bao N, Zhang J, Li Z, Wei S, Zhang J, Greenwald SE, Onofrey JA, Lu Y, Xu L. CT-Less Whole-Body Bone Segmentation of PET Images Using a Multimodal Deep Learning Network. IEEE J Biomed Health Inform 2025; 29:1151-1164. [PMID: 40030243 DOI: 10.1109/jbhi.2024.3501386] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/06/2025]
Abstract
In bone cancer imaging, positron emission tomography (PET) is ideal for the diagnosis and staging of bone cancers due to its high sensitivity to malignant tumors. The diagnosis of bone cancer requires tumor analysis and localization, where accurate and automated wholebody bone segmentation (WBBS) is often needed. Current WBBS for PET imaging is based on paired Computed Tomography (CT) images. However, mismatches between CT and PET images often occur due to patient motion, which leads to erroneous bone segmentation and thus, to inaccurate tumor analysis. Furthermore, there are some instances where CT images are unavailable for WBBS. In this work, we propose a novel multimodal fusion network (MMF-Net) for WBBS of PET images, without the need for CT images. Specifically, the tracer activity ($\lambda$-MLAA), attenuation map ($\mu$-MLAA), and synthetic attenuation map ($\mu$-DL) images are introduced into the training data. We first design a multi-encoder structure employed to fully learn modalityspecific encoding representations of the three PET modality images through independent encoding branches. Then, we propose a multimodal fusion module in the decoder to further integrate the complementary information across the three modalities. Additionally, we introduce revised convolution units, SE (Squeeze-and-Excitation) Normalization and deep supervision to improve segmentation performance. Extensive comparisons and ablation experiments, using 130 whole-body PET image datasets, show promising results. We conclude that the proposed method can achieve WBBS with moderate to high accuracy using PET information only, which potentially can be used to overcome the current limitations of CT-based approaches, while minimizing exposure to ionizing radiation.
Collapse
|
4
|
Xiang D, Peng T, Bian Y, Chen L, Zeng J, Shi F, Zhu W, Chen X. Unpaired Dual-Modal Image Complementation Learning for Single-Modal Medical Image Segmentation. IEEE Trans Biomed Eng 2025; 72:664-674. [PMID: 39320994 DOI: 10.1109/tbme.2024.3467216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/27/2024]
Abstract
OBJECTIVE Multi-modal MR/CT image segmentation is an important task in disease diagnosis and treatment, but it is usually difficult to acquire aligned multi-modal images of a patient in clinical practice due to the high cost and specific allergic reactions to contrast agents. To address these issues, a task complementation framework is proposed to enable unpaired multi-modal image complementation learning in the training stage and single-modal image segmentation in the inference stage. METHOD To fuse unpaired dual-modal images in the training stage and allow single-modal image segmentation in the inference stage, a synthesis-segmentation task complementation network is constructed to mutually facilitate cross-modal image synthesis and segmentation since the same content feature can be used to perform the image segmentation task and image synthesis task. To maintain the consistency of the target organ with varied shapes, a curvature consistency loss is proposed to align the segmentation predictions of the original image and the cross-modal synthesized image. To segment the small lesions or substructures, a regression-segmentation task complementation network is constructed to utilize the auxiliary feature of the target organ. RESULTS Comprehensive experiments have been performed with an in-house dataset and a publicly available dataset. The experimental results have demonstrated the superiority of our framework over state-of-the-art methods. CONCLUSION The proposed method can fuse dual-modal CT/MR images in the training stage and only needs single-modal CT/MR images in the inference stage. SIGNIFICANCE The proposed method can be used in routine clinical occasions when only single-modal CT/MR image is available for a patient.
Collapse
|
5
|
Yu X, Yang Z, Wang X, Sun X, Shen R, Li H, Zhang M. A prior-knowledge-guided dynamic attention mechanism to predict nocturnal hypoglycemic events in type 1 diabetes. BMC Med Inform Decis Mak 2024; 24:378. [PMID: 39696373 DOI: 10.1186/s12911-024-02761-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Accepted: 11/11/2024] [Indexed: 12/20/2024] Open
Abstract
Nocturnal hypoglycemia is a critical problem faced by diabetic patients. Failure to intervene in time can be dangerous for patients. The existing early warning methods struggle to extract crucial information comprehensively from complex multi-source heterogeneous data. In this paper, a deep learning framework with an innovative dynamic attention mechanism is proposed to predict nocturnal hypoglycemic events for type 1 diabetes patients. Features related to nocturnal hypoglycemia are extracted from multi-scale and multi-dimensional data, which enables comprehensive information extraction from diverse sources. Then, we propose a prior-knowledge-guided attention mechanism to enhance the network's learning capability and interpretability. The method was evaluated on a public available clinical dataset, which successfully warned 94.91% of nocturnal hypoglycemic events with an F1-score of 96.35%. By integrating our proposed framework into the nocturnal hypoglycemia early warning model, issues related to feature redundancy and incompleteness were mitigated. Comparative analysis demonstrates that our method outperforms existing approaches, offering superior accuracy and practicality in real-world scenarios.
Collapse
Affiliation(s)
- Xia Yu
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| | - Zi Yang
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| | - Xinzhuo Wang
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| | - Xiaoyu Sun
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| | - Ruiting Shen
- Department of Endocrinology and Metabolism, Ningbo No.2 Hospital, Ningbo, Zhejiang Province, 315010, China
| | - Hongru Li
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| | - Mingchen Zhang
- Department of Endocrinology and Metabolism, Ningbo No.2 Hospital, Ningbo, Zhejiang Province, 315010, China.
| |
Collapse
|
6
|
Kang S, Kang Y, Tan S. Exploring and Exploiting Multi-Modality Uncertainty for Tumor Segmentation on PET/CT. IEEE J Biomed Health Inform 2024; 28:5435-5446. [PMID: 38776203 DOI: 10.1109/jbhi.2024.3397332] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/24/2024]
Abstract
Despite the success of deep learning methods in multi-modality segmentation tasks, they typically produce a deterministic output, neglecting the underlying uncertainty. The absence of uncertainty could lead to over-confident predictions with catastrophic consequences, particularly in safety-critical clinical applications. Recently, uncertainty estimation has attracted increasing attention, offering a measure of confidence associated with machine decisions. Nonetheless, existing uncertainty estimation approaches primarily focus on single-modality networks, leaving the uncertainty of multi-modality networks a largely under-explored domain. In this study, we present the first exploration of multi-modality uncertainties in the context of tumor segmentation on PET/CT. Concretely, we assessed four well-established uncertainty estimation approaches across various dimensions, including segmentation performance, uncertainty quality, comparison to single-modality uncertainties, and correlation to the contradictory information between modalities. Through qualitative and quantitative analyses, we gained valuable insights into what benefits multi-modality uncertainties derive, what information multi-modality uncertainties capture, and how multi-modality uncertainties correlate to information from single modalities. Drawing from these insights, we introduced a novel uncertainty-driven loss, which incentivized the network to effectively utilize the complementary information between modalities. The proposed approach outperformed the backbone network by 4.53 and 2.92 Dices in percentages on two PET/CT datasets while achieving lower uncertainties. This study not only advanced the comprehension of multi-modality uncertainties but also revealed the potential benefit of incorporating them into the segmentation network.
Collapse
|
7
|
Sun K, Ding J, Li Q, Chen W, Zhang H, Sun J, Jiao Z, Ni X. CMAF-Net: a cross-modal attention fusion-based deep neural network for incomplete multi-modal brain tumor segmentation. Quant Imaging Med Surg 2024; 14:4579-4604. [PMID: 39022265 PMCID: PMC11250309 DOI: 10.21037/qims-24-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2024] [Accepted: 04/19/2024] [Indexed: 07/20/2024]
Abstract
Background The information between multimodal magnetic resonance imaging (MRI) is complementary. Combining multiple modalities for brain tumor image segmentation can improve segmentation accuracy, which has great significance for disease diagnosis and treatment. However, different degrees of missing modality data often occur in clinical practice, which may lead to serious performance degradation or even failure of brain tumor segmentation methods relying on full-modality sequences to complete the segmentation task. To solve the above problems, this study aimed to design a new deep learning network for incomplete multimodal brain tumor segmentation. Methods We propose a novel cross-modal attention fusion-based deep neural network (CMAF-Net) for incomplete multimodal brain tumor segmentation, which is based on a three-dimensional (3D) U-Net architecture with encoding and decoding structure, a 3D Swin block, and a cross-modal attention fusion (CMAF) block. A convolutional encoder is initially used to extract the specific features from different modalities, and an effective 3D Swin block is constructed to model the long-range dependencies to obtain richer information for brain tumor segmentation. Then, a cross-attention based CMAF module is proposed that can deal with different missing modality situations by fusing features between different modalities to learn the shared representations of the tumor regions. Finally, the fused latent representation is decoded to obtain the final segmentation result. Additionally, channel attention module (CAM) and spatial attention module (SAM) are incorporated into the network to further improve the robustness of the model; the CAM to help focus on important feature channels, and the SAM to learn the importance of different spatial regions. Results Evaluation experiments on the widely-used BraTS 2018 and BraTS 2020 datasets demonstrated the effectiveness of the proposed CMAF-Net which achieved average Dice scores of 87.9%, 81.8%, and 64.3%, as well as Hausdorff distances of 4.21, 5.35, and 4.02 for whole tumor, tumor core, and enhancing tumor on the BraTS 2020 dataset, respectively, outperforming several state-of-the-art segmentation methods in missing modalities situations. Conclusions The experimental results show that the proposed CMAF-Net can achieve accurate brain tumor segmentation in the case of missing modalities with promising application potential.
Collapse
Affiliation(s)
- Kangkang Sun
- School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| | - Jiangyi Ding
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| | - Qixuan Li
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| | - Wei Chen
- School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| | - Heng Zhang
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| | - Jiawei Sun
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| | - Zhuqing Jiao
- School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China
| | - Xinye Ni
- Department of Radiotherapy, The Affiliated of Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- Center of Medical Physics, Nanjing Medical University, Changzhou, China
| |
Collapse
|
8
|
Sobootian DJ, Bronzlik P, Spineli LM, Becker LS, Winther HB, Bueltmann E. Convolutional Neural Network for Fully Automated Cerebellar Volumetry in Children in Comparison to Manual Segmentation and Developmental Trajectory of Cerebellar Volumes. CEREBELLUM (LONDON, ENGLAND) 2024; 23:1074-1085. [PMID: 37833550 PMCID: PMC11102395 DOI: 10.1007/s12311-023-01609-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 09/25/2023] [Indexed: 10/15/2023]
Abstract
The purpose of this study was to develop a fully automated and reliable volumetry of the cerebellum of children during infancy and childhood using deep learning algorithms in comparison to manual segmentation. In addition, the clinical usefulness of measuring the cerebellar volume is shown. One hundred patients (0 to 16.3 years old) without infratentorial signal abnormalities on conventional MRI were retrospectively selected from our pool of pediatric MRI examinations. Based on a routinely acquired 3D T1-weighted magnetization prepared rapid gradient echo (MPRAGE) sequence, the cerebella were manually segmented using ITK-SNAP. The data set of all 100 cases was divided into four splits (four-fold cross-validation) to train the network (NN) to delineate the boundaries of the cerebellum. First, the accuracy of the newly created neural network was compared with the manual segmentation. Secondly, age-related volume changes were investigated. Our trained NN achieved an excellent Spearman correlation coefficient of 0.99, a Dice Coefficient of 95.0 ± 2.1%, and an intersection over union (IoU) of 90.6 ± 3.8%. Cerebellar volume increased continuously with age, showing an exponentially rapid growth within the first year of life. Using a convolutional neural network, it was possible to achieve reliable, fully automated cerebellar volume measurements in childhood and infancy, even when based on a relatively small cohort. In this preliminary study, age-dependent cerebellar volume changes could be acquired.
Collapse
Affiliation(s)
- Daria Juliane Sobootian
- Institute of Diagnostic and Interventional Neuroradiology, Hannover Medical School, Hannover, Germany
| | - Paul Bronzlik
- Institute of Diagnostic and Interventional Neuroradiology, Hannover Medical School, Hannover, Germany
| | - Loukia M Spineli
- Midwifery Research and Education Unit, Hannover Medical School, Hannover, Germany
| | - Lena Sophie Becker
- Institute of Diagnostic and Interventional Radiology, Hannover Medical School, Hannover, Germany
| | - Hinrich Boy Winther
- Institute of Diagnostic and Interventional Radiology, Hannover Medical School, Hannover, Germany
| | - Eva Bueltmann
- Institute of Diagnostic and Interventional Neuroradiology, Hannover Medical School, Hannover, Germany.
| |
Collapse
|
9
|
Henschel L, Kügler D, Zöllei L, Reuter M. VINNA for neonates: Orientation independence through latent augmentations. IMAGING NEUROSCIENCE (CAMBRIDGE, MASS.) 2024; 2:1-26. [PMID: 39575178 PMCID: PMC11576933 DOI: 10.1162/imag_a_00180] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/21/2023] [Revised: 02/16/2024] [Accepted: 04/19/2024] [Indexed: 11/24/2024]
Abstract
A robust, fast, and accurate segmentation of neonatal brain images is highly desired to better understand and detect changes during development and disease, specifically considering the rise in imaging studies for this cohort. Yet, the limited availability of ground truth datasets, lack of standardized acquisition protocols, and wide variations of head positioning in the scanner pose challenges for method development. A few automated image analysis pipelines exist for newborn brain Magnetic Resonance Image (MRI) segmentation, but they often rely on time-consuming non-linear spatial registration procedures and require resampling to a common resolution, subject to loss of information due to interpolation and down-sampling. Without registration and image resampling, variations with respect to head positions and voxel resolutions have to be addressed differently. In deep learning, external augmentations such as rotation, translation, and scaling are traditionally used to artificially expand the representation of spatial variability, which subsequently increases both the training dataset size and robustness. However, these transformations in the image space still require resampling, reducing accuracy specifically in the context of label interpolation. We recently introduced the concept of resolution-independence with the Voxel-size Independent Neural Network framework, VINN. Here, we extend this concept by additionally shifting all rigid-transforms into the network architecture with a four degree of freedom (4-DOF) transform module, enabling resolution-aware internal augmentations (VINNA) for deep learning. In this work, we show that VINNA (i) significantly outperforms state-of-the-art external augmentation approaches, (ii) effectively addresses the head variations present specifically in newborn datasets, and (iii) retains high segmentation accuracy across a range of resolutions (0.5-1.0 mm). Furthermore, the 4-DOF transform module together with internal augmentations is a powerful, general approach to implement spatial augmentation without requiring image or label interpolation. The specific network application to newborns will be made publicly available as VINNA4neonates.
Collapse
Affiliation(s)
- Leonie Henschel
- German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany
| | - David Kügler
- German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany
| | - Lilla Zöllei
- A.A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Boston, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| | - Martin Reuter
- German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany
- A.A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Boston, MA, USA
- Department of Radiology, Harvard Medical School, Boston, MA, USA
| |
Collapse
|
10
|
He Q, Summerfield N, Dong M, Glide-Hurst C. MODALITY-AGNOSTIC LEARNING FOR MEDICAL IMAGE SEGMENTATION USING MULTI-MODALITY SELF-DISTILLATION. PROCEEDINGS. IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING 2024; 2024:10.1109/isbi56570.2024.10635881. [PMID: 39735423 PMCID: PMC11673955 DOI: 10.1109/isbi56570.2024.10635881] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/31/2024]
Abstract
In medical image segmentation, although multi-modality training is possible, clinical translation is challenged by the limited availability of all image types for a given patient. Different from typical segmentation models, modality-agnostic (MAG) learning trains a single model based on all available modalities but remains input-agnostic, allowing a single model to produce accurate segmentation given any modality combinations. In this paper, we propose a novel frame-work, MAG learning through Multi-modality Self-distillation (MAG-MS), for medical image segmentation. MAG-MS distills knowledge from the fusion of multiple modalities and applies it to enhance representation learning for individual modalities. This makes it an adaptable and efficient solution for handling limited modalities during testing scenarios. Our extensive experiments on benchmark datasets demonstrate its superior segmentation accuracy, MAG robustness, and efficiency than the current state-of-the-art methods.
Collapse
Affiliation(s)
- Qisheng He
- Department of Computer Science, Wayne State University, Detroit, MI, USA
| | - Nicholas Summerfield
- Department of Human Oncology, University of Wisconsin-Madison, Madison, WI, USA
- Department of Medical Physics, University of Wisconsin-Madison, Madison, WI, USA
| | - Ming Dong
- Department of Computer Science, Wayne State University, Detroit, MI, USA
| | - Carri Glide-Hurst
- Department of Human Oncology, University of Wisconsin-Madison, Madison, WI, USA
- Department of Medical Physics, University of Wisconsin-Madison, Madison, WI, USA
| |
Collapse
|
11
|
Chen Q, Zhang J, Meng R, Zhou L, Li Z, Feng Q, Shen D. Modality-Specific Information Disentanglement From Multi-Parametric MRI for Breast Tumor Segmentation and Computer-Aided Diagnosis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1958-1971. [PMID: 38206779 DOI: 10.1109/tmi.2024.3352648] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/13/2024]
Abstract
Breast cancer is becoming a significant global health challenge, with millions of fatalities annually. Magnetic Resonance Imaging (MRI) can provide various sequences for characterizing tumor morphology and internal patterns, and becomes an effective tool for detection and diagnosis of breast tumors. However, previous deep-learning based tumor segmentation methods from multi-parametric MRI still have limitations in exploring inter-modality information and focusing task-informative modality/modalities. To address these shortcomings, we propose a Modality-Specific Information Disentanglement (MoSID) framework to extract both inter- and intra-modality attention maps as prior knowledge for guiding tumor segmentation. Specifically, by disentangling modality-specific information, the MoSID framework provides complementary clues for the segmentation task, by generating modality-specific attention maps to guide modality selection and inter-modality evaluation. Our experiments on two 3D breast datasets and one 2D prostate dataset demonstrate that the MoSID framework outperforms other state-of-the-art multi-modality segmentation methods, even in the cases of missing modalities. Based on the segmented lesions, we further train a classifier to predict the patients' response to radiotherapy. The prediction accuracy is comparable to the case of using manually-segmented tumors for treatment outcome prediction, indicating the robustness and effectiveness of the proposed segmentation method. The code is available at https://github.com/Qianqian-Chen/MoSID.
Collapse
|
12
|
Hu X, Wang L, Wang L, Chen Q, Zheng L, Zhu Y. Glioma segmentation based on dense contrastive learning and multimodal features recalibration. Phys Med Biol 2024; 69:095016. [PMID: 38537288 DOI: 10.1088/1361-6560/ad387f] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Accepted: 03/27/2024] [Indexed: 04/23/2024]
Abstract
Accurate segmentation of different regions of gliomas from multimodal magnetic resonance (MR) images is crucial for glioma grading and precise diagnosis, but many existing segmentation methods are difficult to effectively utilize multimodal MR image information to recognize accurately the lesion regions with small size, low contrast and irregular shape. To address this issue, this work proposes a novel 3D glioma segmentation model DCL-MANet. DCL-MANet has an architecture of multiple encoders and one single decoder. Each encoder is used to extract MR image features of a given modality. To overcome the entangle problems of multimodal semantic features, a dense contrastive learning (DCL) strategy is presented to extract the modality-specific and common features. Following that, feature recalibration block (RFB) based on modality-wise attention is used to recalibrate the semantic features of each modality, enabling the model to focus on the features that are beneficial for glioma segmentation. These recalibrated features are input into the decoder to obtain the segmentation results. To verify the superiority of the proposed method, we compare it with several state-of-the-art (SOTA) methods in terms of Dice, average symmetric surface distance (ASSD), HD95 and volumetric similarity (Vs). The comparison results show that the average Dice, ASSD, HD95 and Vs of DCL-MANet on all tumor regions are improved at least by 0.66%, 3.47%, 8.94% and 1.07% respectively. For small enhance tumor (ET) region, the corresponding improvement can be up to 0.37%, 7.83%, 11.32%, and 1.35%, respectively. In addition, the ablation results demonstrate the effectiveness of the proposed DCL and RFB, and combining them can significantly increase Dice (1.59%) and Vs (1.54%) while decreasing ASSD (40.51%) and HD95 (45.16%) on ET region. The proposed DCL-MANet could disentangle multimodal features and enhance the semantics of modality-dependent features, providing a potential means to accurately segment small lesion regions in gliomas.
Collapse
Affiliation(s)
- Xubin Hu
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Lihui Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Li Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Qijian Chen
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Licheng Zheng
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Yuemin Zhu
- University Lyon, INSA Lyon, CNRS, Inserm, IRP Metislab CREATIS UMR5220, U1206, Lyon F-69621, France
| |
Collapse
|
13
|
Chen J, Huang G, Yuan X, Zhong G, Zheng Z, Pun CM, Zhu J, Huang Z. Quaternion Cross-Modality Spatial Learning for Multi-Modal Medical Image Segmentation. IEEE J Biomed Health Inform 2024; 28:1412-1423. [PMID: 38145537 DOI: 10.1109/jbhi.2023.3346529] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Recently, the Deep Neural Networks (DNNs) have had a large impact on imaging process including medical image segmentation, and the real-valued convolution of DNN has been extensively utilized in multi-modal medical image segmentation to accurately segment lesions via learning data information. However, the weighted summation operation in such convolution limits the ability to maintain spatial dependence that is crucial for identifying different lesion distributions. In this paper, we propose a novel Quaternion Cross-modality Spatial Learning (Q-CSL) which explores the spatial information while considering the linkage between multi-modal images. Specifically, we introduce to quaternion to represent data and coordinates that contain spatial information. Additionally, we propose Quaternion Spatial-association Convolution to learn the spatial information. Subsequently, the proposed De-level Quaternion Cross-modality Fusion (De-QCF) module excavates inner space features and fuses cross-modality spatial dependency. Our experimental results demonstrate that our approach compared to the competitive methods perform well with only 0.01061 M parameters and 9.95G FLOPs.
Collapse
|
14
|
Jiang M, Chiu B. A Dual-Stream Centerline-Guided Network for Segmentation of the Common and Internal Carotid Arteries From 3D Ultrasound Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2690-2705. [PMID: 37015114 DOI: 10.1109/tmi.2023.3263537] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Segmentation of the carotid section encompassing the common carotid artery (CCA), the bifurcation and the internal carotid artery (ICA) from three-dimensional ultrasound (3DUS) is required to measure the vessel wall volume (VWV) and localized vessel-wall-plus-plaque thickness (VWT), shown to be sensitive to treatment effect. We proposed an approach to combine a centerline extraction network (CHG-Net) and a dual-stream centerline-guided network (DSCG-Net) to segment the lumen-intima (LIB) and media-adventitia boundaries (MAB) from 3DUS images. Correct arterial location is essential for successful segmentation of the carotid section encompassing the bifurcation. We addressed this challenge by using the arterial centerline to enhance the localization accuracy of the segmentation network. The CHG-Net was developed to generate a heatmap indicating high probability regions for the centerline location, which was then integrated with the 3DUS image by the DSCG-Net to generate the MAB and LIB. The DSCG-Net includes a scale-based and a spatial attention mechanism to fuse multi-level features extracted by the encoder, and a centerline heatmap reconstruction side-branch connected to the end of the encoder to increase the generalization ability of the network. Experiments involving 224 3DUS volumes produce a Dice similarity coefficient (DSC) of 95.8±1.9% and 92.3±5.4% for CCA MAB and LIB, respectively, and 93.2±4.4% and 89.0±10.0% for ICA MAB and LIB, respectively. Our approach outperformed four state-of-the-art 3D CNN models, even after their performances were boosted by centerline guidance. The efficiency afforded by the framework would allow it to be incorporated into the clinical workflow for improved quantification of plaque change.
Collapse
|
15
|
Dou M, Chen Z, Tang Y, Sheng L, Zhou J, Wang X, Yao Y. Segmentation of rectal tumor from multi-parametric MRI images using an attention-based fusion network. Med Biol Eng Comput 2023; 61:2379-2389. [PMID: 37084029 DOI: 10.1007/s11517-023-02828-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Accepted: 03/08/2023] [Indexed: 04/22/2023]
Abstract
Accurate segmentation of rectal tumors is the most crucial task in determining the stage of rectal cancer and developing suitable therapies. However, complex image backgrounds, irregular edge, and poor contrast hinder the related research. This study presents an attention-based multi-modal fusion module to effectively integrate complementary information from different MRI images and suppress redundancy. In addition, a deep learning-based segmentation model (AF-UNet) is designed to achieve accurate segmentation of rectal tumors. This model takes multi-parametric MRI images as input and effectively integrates the features from different multi-parametric MRI images by embedding the attention fusion module. Finally, three types of MRI images (T2, ADC, DWI) of 250 patients with rectal cancer were collected, with the tumor regions delineated by two oncologists. The experimental results show that the proposed method is superior to the most advanced image segmentation method with a Dice coefficient of [Formula: see text], which is also better than other multi-modal fusion methods. Framework of the AF-UNet. This model takes multi-modal MRI images as input, and integrates complementary information using attention mechanism and suppresses redundancy.
Collapse
Affiliation(s)
- Meng Dou
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Zhebin Chen
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Yuanling Tang
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Leiming Sheng
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Jitao Zhou
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Xin Wang
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Yu Yao
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, China.
- University of Chinese Academy of Sciences, Beijing, China.
| |
Collapse
|
16
|
Zhao X, Liao Y, Xie J, He X, Zhang S, Wang G, Fang J, Lu H, Yu J. BreastDM: A DCE-MRI dataset for breast tumor image segmentation and classification. Comput Biol Med 2023; 164:107255. [PMID: 37499296 DOI: 10.1016/j.compbiomed.2023.107255] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2023] [Revised: 05/31/2023] [Accepted: 07/07/2023] [Indexed: 07/29/2023]
Abstract
Dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI) has shown high sensitivity to diagnose breast cancer. However, few computer-aided algorithms focus on employing DCE-MR images for breast cancer diagnosis due to the lack of publicly available DCE-MRI datasets. To address this issue, our work releases a new DCE-MRI dataset called BreastDM for breast tumor segmentation and classification. In particular, a dataset of 232 patients selected with DCE-MR images for benign and malignant cases is established. Each case consists of three types of sequences: pre-contrast, post-contrast, and subtraction sequences. To show the difficulty of breast DCE-MRI tumor image segmentation and classification tasks, benchmarks are achieved by state-of-the-art image segmentation and classification algorithms, including conventional hand-crafted based methods and recently-emerged deep learning-based methods. More importantly, a local-global cross attention fusion network (LG-CAFN) is proposed to further improve the performance of breast tumor images classification. Specifically, LG-CAFN achieved the highest accuracy (88.20%, 83.93%) and AUC value (0.9154,0.8826) in both groups of experiments. Extensive experiments are conducted to present strong baselines based on various typical image segmentation and classification algorithms. Experiment results also demonstrate the superiority of the proposed LG-CAFN to other breast tumor images classification methods. The related dataset and evaluation codes are publicly available at smallboy-code/Breast-cancer-dataset.
Collapse
Affiliation(s)
- Xiaoming Zhao
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China; School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310018, China
| | - Yuehui Liao
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China; School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310018, China
| | - Jiahao Xie
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China; School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310018, China
| | - Xiaxia He
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China
| | - Shiqing Zhang
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China; School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310018, China.
| | - Guoyu Wang
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China.
| | - Jiangxiong Fang
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China
| | - Hongsheng Lu
- Taizhou Central Hospital, Taizhou University, 318000, Taizhou, China
| | - Jun Yu
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310018, China
| |
Collapse
|
17
|
Jiang M, Yuan B, Kou W, Yan W, Marshall H, Yang Q, Syer T, Punwani S, Emberton M, Barratt DC, Cho CCM, Hu Y, Chiu B. Prostate cancer segmentation from MRI by a multistream fusion encoder. Med Phys 2023; 50:5489-5504. [PMID: 36938883 DOI: 10.1002/mp.16374] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Revised: 02/15/2023] [Accepted: 03/03/2023] [Indexed: 03/21/2023] Open
Abstract
BACKGROUND Targeted prostate biopsy guided by multiparametric magnetic resonance imaging (mpMRI) detects more clinically significant lesions than conventional systemic biopsy. Lesion segmentation is required for planning MRI-targeted biopsies. The requirement for integrating image features available in T2-weighted and diffusion-weighted images poses a challenge in prostate lesion segmentation from mpMRI. PURPOSE A flexible and efficient multistream fusion encoder is proposed in this work to facilitate the multiscale fusion of features from multiple imaging streams. A patch-based loss function is introduced to improve the accuracy in segmenting small lesions. METHODS The proposed multistream encoder fuses features extracted in the three imaging streams at each layer of the network, thereby allowing improved feature maps to propagate downstream and benefit segmentation performance. The fusion is achieved through a spatial attention map generated by optimally weighting the contribution of the convolution outputs from each stream. This design provides flexibility for the network to highlight image modalities according to their relative influence on the segmentation performance. The encoder also performs multiscale integration by highlighting the input feature maps (low-level features) with the spatial attention maps generated from convolution outputs (high-level features). The Dice similarity coefficient (DSC), serving as a cost function, is less sensitive to incorrect segmentation for small lesions. We address this issue by introducing a patch-based loss function that provides an average of the DSCs obtained from local image patches. This local average DSC is equally sensitive to large and small lesions, as the patch-based DSCs associated with small and large lesions have equal weights in this average DSC. RESULTS The framework was evaluated in 931 sets of images acquired in several clinical studies at two centers in Hong Kong and the United Kingdom. In particular, the training, validation, and test sets contain 615, 144, and 172 sets of images, respectively. The proposed framework outperformed single-stream networks and three recently proposed multistream networks, attaining F1 scores of 82.2 and 87.6% in the lesion and patient levels, respectively. The average inference time for an axial image was 11.8 ms. CONCLUSION The accuracy and efficiency afforded by the proposed framework would accelerate the MRI interpretation workflow of MRI-targeted biopsy and focal therapies.
Collapse
Affiliation(s)
- Mingjie Jiang
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China
| | - Baohua Yuan
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China
- Aliyun School of Big Data, Changzhou University, Changzhou, China
| | - Weixuan Kou
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China
| | - Wen Yan
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Harry Marshall
- Schulich School of Medicine & Dentistry, Western University, Ontario, Canada
| | - Qianye Yang
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Tom Syer
- Centre for Medical Imaging, University College London, London, UK
| | - Shonit Punwani
- Centre for Medical Imaging, University College London, London, UK
| | - Mark Emberton
- Division of Surgery & Interventional Science, University College London, London, UK
| | - Dean C Barratt
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Carmen C M Cho
- Prince of Wales Hospital and Department of Imaging and Intervention Radiology, Chinese University of Hong Kong, Hong Kong SAR, China
| | - Yipeng Hu
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Bernard Chiu
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong SAR, China
| |
Collapse
|
18
|
Wu B, Zhang F, Xu L, Shen S, Shao P, Sun M, Liu P, Yao P, Xu RX. Modality preserving U-Net for segmentation of multimodal medical images. Quant Imaging Med Surg 2023; 13:5242-5257. [PMID: 37581055 PMCID: PMC10423364 DOI: 10.21037/qims-22-1367] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2022] [Accepted: 05/19/2023] [Indexed: 08/16/2023]
Abstract
Background Recent advances in artificial intelligence and digital image processing have inspired the use of deep neural networks for segmentation tasks in multimodal medical imaging. Unlike natural images, multimodal medical images contain much richer information regarding different modal properties and therefore present more challenges for semantic segmentation. However, there is no report on systematic research that integrates multi-scaled and structured analysis of single-modal and multimodal medical images. Methods We propose a deep neural network, named as Modality Preserving U-Net (MPU-Net), for modality-preserving analysis and segmentation of medical targets from multimodal medical images. The proposed MPU-Net consists of a modality preservation encoder (MPE) module that preserves the feature independency among the modalities and a modality fusion decoder (MFD) module that performs a multiscale feature fusion analysis for each modality in order to provide a rich feature representation for the final task. The effectiveness of such a single-modal preservation and multimodal fusion feature extraction approach is verified by multimodal segmentation experiments and an ablation study using brain tumor and prostate datasets from Medical Segmentation Decathlon (MSD). Results The segmentation experiments demonstrated the superiority of MPU-Net over other methods in the segmentation tasks for multimodal medical images. In the brain tumor segmentation tasks, the Dice scores (DSCs) for the whole tumor (WT), the tumor core (TC) and the enhancing tumor (ET) regions were 89.42%, 86.92%, and 84.59%, respectively. In the meanwhile, the 95% Hausdorff distance (HD95) results were 3.530, 4.899 and 2.555, respectively. In the prostate segmentation tasks, the DSCs for the peripheral zone (PZ) and the transitional zone (TZ) of the prostate were 71.20% and 90.38%, respectively. In the meanwhile, the 95% HD95 results were 6.367 and 4.766, respectively. The ablation study showed that the combination of single-modal preservation and multimodal fusion methods improved the performance of multimodal medical image feature analysis. Conclusions In the segmentation tasks using brain tumor and prostate datasets, the MPU-Net method has achieved the improved performance in comparison with the conventional methods, indicating its potential application for other segmentation tasks in multimodal medical images.
Collapse
Affiliation(s)
- Bingxuan Wu
- Department of Precision Machinery and Precision Instrumentation, University of Science and Technology of China, Hefei, China
| | - Fan Zhang
- Department of Precision Machinery and Precision Instrumentation, University of Science and Technology of China, Hefei, China
| | - Liang Xu
- Suzhou Institute for Advanced Research, University of Science and Technology of China, Suzhou, China
| | - Shuwei Shen
- Suzhou Institute for Advanced Research, University of Science and Technology of China, Suzhou, China
| | - Pengfei Shao
- Department of Precision Machinery and Precision Instrumentation, University of Science and Technology of China, Hefei, China
| | - Mingzhai Sun
- Suzhou Institute for Advanced Research, University of Science and Technology of China, Suzhou, China
| | - Peng Liu
- Suzhou Institute for Advanced Research, University of Science and Technology of China, Suzhou, China
| | - Peng Yao
- School of Microelectronics, University of Science and Technology of China, Hefei, China
| | - Ronald X. Xu
- Suzhou Institute for Advanced Research, University of Science and Technology of China, Suzhou, China
| |
Collapse
|
19
|
Yang H, Zhou T, Zhou Y, Zhang Y, Fu H. Flexible Fusion Network for Multi-Modal Brain Tumor Segmentation. IEEE J Biomed Health Inform 2023; 27:3349-3359. [PMID: 37126623 DOI: 10.1109/jbhi.2023.3271808] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
Automated brain tumor segmentation is crucial for aiding brain disease diagnosis and evaluating disease progress. Currently, magnetic resonance imaging (MRI) is a routinely adopted approach in the field of brain tumor segmentation that can provide different modality images. It is critical to leverage multi-modal images to boost brain tumor segmentation performance. Existing works commonly concentrate on generating a shared representation by fusing multi-modal data, while few methods take into account modality-specific characteristics. Besides, how to efficiently fuse arbitrary numbers of modalities is still a difficult task. In this study, we present a flexible fusion network (termed F 2Net) for multi-modal brain tumor segmentation, which can flexibly fuse arbitrary numbers of multi-modal information to explore complementary information while maintaining the specific characteristics of each modality. Our F 2Net is based on the encoder-decoder structure, which utilizes two Transformer-based feature learning streams and a cross-modal shared learning network to extract individual and shared feature representations. To effectively integrate the knowledge from the multi-modality data, we propose a cross-modal feature-enhanced module (CFM) and a multi-modal collaboration module (MCM), which aims at fusing the multi-modal features into the shared learning network and incorporating the features from encoders into the shared decoder, respectively. Extensive experimental results on multiple benchmark datasets demonstrate the effectiveness of our F 2Net over other state-of-the-art segmentation methods.
Collapse
|
20
|
De A, Wang X, Zhang Q, Wu J, Cong F. An efficient memory reserving-and-fading strategy for vector quantization based 3D brain segmentation and tumor extraction using an unsupervised deep learning network. Cogn Neurodyn 2023; 18:1-22. [PMID: 37362765 PMCID: PMC10132803 DOI: 10.1007/s11571-023-09965-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2022] [Revised: 02/24/2023] [Accepted: 03/08/2023] [Indexed: 06/28/2023] Open
Abstract
Deep learning networks are state-of-the-art approaches for 3D brain image segmentation, and the radiological characteristics extracted from tumors are of great significance for clinical diagnosis, treatment planning, and treatment outcome evaluation. However, two problems have been the hindering factors in brain image segmentation techniques. One is that deep learning networks require large amounts of manually annotated data. Another issue is the computational efficiency of 3D deep learning networks. In this study, we propose a vector quantization (VQ)-based 3D segmentation method that employs a novel unsupervised 3D deep embedding clustering (3D-DEC) network and an efficiency memory reserving-and-fading strategy. The VQ-based 3D-DEC network is trained on volume data in an unsupervised manner to avoid manual data annotation. The memory reserving-and-fading strategy beefs up model efficiency greatly. The designed methodology makes deep learning-based model feasible for biomedical image segmentation. The experiment is divided into two parts. First, we extensively evaluate the effectiveness and robustness of the proposed model on two authoritative MRI brain tumor databases (i.e., IBSR and BrainWeb). Second, we validate the model using real 3D brain tumor data collected from our institute for clinical practice significance. Results show that our method (without data manual annotation) has superior accuracy (0.74 ± 0.04 Tanimoto coefficient on IBSR, 97.5% TP and 97.7% TN on BrainWeb, and 91% Dice, 88% sensitivity and 87% specificity on real brain data) and remarkable efficiency (speedup ratio is 18.72 on IBSR, 31.16 on BrainWeb, 31.00 on real brain data) compared to the state-of-the-art methods. The results show that our proposed model can address the lacks of manual annotations, and greatly increase computation speedup with competitive segmentation accuracy compared to other state-of-the-art 3D CNN models. Moreover, the proposed model can be used for tumor treatment follow-ups every 6 months, providing critical details for surgical and postoperative treatment by correctly extracting numerical radiomic features of tumors.
Collapse
Affiliation(s)
- Ailing De
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
| | - Xiulin Wang
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
- School of Biomedical Engineering, Faculty of Medicine, Dalian University of Technology, Dalian, 116000 Liaoning China
| | - Qing Zhang
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
| | - Jianlin Wu
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
| | - Fengyu Cong
- School of Biomedical Engineering, Faculty of Medicine, Dalian University of Technology, Dalian, 116000 Liaoning China
- Faculty of Information Technology, University of Jyväskylä, 40014 Jyväskylä, Finland
- School of Artificial Intelligence, Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, 116000 Liaoning China
- Key Laboratory of Integrated Circuit and Biomedical Electronic System, Dalian University of Technology, Dalian, 116000 Liaoning China
| |
Collapse
|
21
|
Li X, Jiang Y, Li M, Zhang J, Yin S, Luo H. MSFR-Net: Multi-modality and single-modality feature recalibration network for brain tumor segmentation. Med Phys 2023; 50:2249-2262. [PMID: 35962724 DOI: 10.1002/mp.15933] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Revised: 05/16/2022] [Accepted: 06/14/2022] [Indexed: 11/11/2022] Open
Abstract
BACKGROUND Accurate and automated brain tumor segmentation from multi-modality MR images plays a significant role in tumor treatment. However, the existing approaches mainly focus on the fusion of multi-modality while ignoring the correlation between single-modality and tumor subcomponents. For example, T2-weighted images show good visualization of edema, and T1-contrast images have a good contrast between enhancing tumor core and necrosis. In the actual clinical process, professional physicians also label tumors according to these characteristics. We design a method for brain tumors segmentation that utilizes both multi-modality fusion and single-modality characteristics. METHODS A multi-modality and single-modality feature recalibration network (MSFR-Net) is proposed for brain tumor segmentation from MR images. Specifically, multi-modality information and single-modality information are assigned to independent pathways. Multi-modality network explicitly learns the relationship between all modalities and all tumor sub-components. Single-modality network learns the relationship between single-modality and its highly correlated tumor subcomponents. Then, a dual recalibration module (DRM) is designed to connect the parallel single-modality network and multi-modality network at multiple stages. The function of the DRM is to unify the two types of features into the same feature space. RESULTS Experiments on BraTS 2015 dataset and BraTS 2018 dataset show that the proposed method is competitive and superior to other state-of-the-art methods. The proposed method achieved the segmentation results with Dice coefficients of 0.86 and Hausdorff distance of 4.82 on BraTS 2018 dataset, with dice coefficients of 0.80, positive predictive value of 0.76, and sensitivity of 0.78 on BraTS 2015 dataset. CONCLUSIONS This work combines the manual labeling process of doctors and introduces the correlation between single-modality and the tumor subcomponents into the segmentation network. The method improves the segmentation performance of brain tumors and can be applied in the clinical practice. The code of the proposed method is available at: https://github.com/xiangQAQ/MSFR-Net.
Collapse
Affiliation(s)
- Xiang Li
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Yuchen Jiang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Minglei Li
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Jiusi Zhang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Shen Yin
- Department of Mechanical and Industrial Engineering, Faculty of Engineering, Norwegian University of Science and Technology, Trondheim, Norway
| | - Hao Luo
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| |
Collapse
|
22
|
He Q, Dong M, Summerfield N, Glide-Hurst C. MAGNET: A MODALITY-AGNOSTIC NETWORK FOR 3D MEDICAL IMAGE SEGMENTATION. PROCEEDINGS. IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING 2023; 2023:10.1109/isbi53787.2023.10230587. [PMID: 38169907 PMCID: PMC10760993 DOI: 10.1109/isbi53787.2023.10230587] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/05/2024]
Abstract
In this paper, we proposed MAGNET, a novel modality-agnostic network for 3D medical image segmentation. Different from existing learning methods, MAGNET is specifically designed to handle real medical situations where multiple modalities/sequences are available during model training, but fewer ones are available or used at time of clinical practice. Our results on multiple datasets show that MAGNET trained on multi-modality data has the unique ability to perform predictions using any subset of training imaging modalities. It outperforms individually trained uni-modality models while can further boost performance when more modalities are available at testing.
Collapse
Affiliation(s)
- Qisheng He
- Wayne State University Department of Computer Science 5057 Woodward Ave, Detroit, MI 48202
| | - Ming Dong
- Wayne State University Department of Computer Science 5057 Woodward Ave, Detroit, MI 48202
| | - Nicholas Summerfield
- University of Wisconsin-Madison Department of Human Oncology Department of Medical Physics 600 Highland Ave, Madison, WI 53792
| | - Carri Glide-Hurst
- University of Wisconsin-Madison Department of Human Oncology Department of Medical Physics 600 Highland Ave, Madison, WI 53792
| |
Collapse
|
23
|
Zeng Z, Zhao T, Sun L, Zhang Y, Xia M, Liao X, Zhang J, Shen D, Wang L, He Y. 3D-MASNet: 3D mixed-scale asymmetric convolutional segmentation network for 6-month-old infant brain MR images. Hum Brain Mapp 2023; 44:1779-1792. [PMID: 36515219 PMCID: PMC9921327 DOI: 10.1002/hbm.26174] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2022] [Revised: 11/04/2022] [Accepted: 11/25/2022] [Indexed: 12/15/2022] Open
Abstract
Precise segmentation of infant brain magnetic resonance (MR) images into gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF) are essential for studying neuroanatomical hallmarks of early brain development. However, for 6-month-old infants, the extremely low-intensity contrast caused by inherent myelination hinders accurate tissue segmentation. Existing convolutional neural networks (CNNs) based segmentation models for this task generally employ single-scale symmetric convolutions, which are inefficient for encoding the isointense tissue boundaries in baby brain images. Here, we propose a 3D mixed-scale asymmetric convolutional segmentation network (3D-MASNet) framework for brain MR images of 6-month-old infants. We replaced the traditional convolutional layer of an existing to-be-trained network with a 3D mixed-scale convolution block consisting of asymmetric kernels (MixACB) during the training phase and then equivalently converted it into the original network. Five canonical CNN segmentation models were evaluated using both T1- and T2-weighted images of 23 6-month-old infants from iSeg-2019 datasets, which contained manual labels as ground truth. MixACB significantly enhanced the average accuracy of all five models and obtained the most considerable improvement in the fully convolutional network model (CC-3D-FCN) and the highest performance in the Dense U-Net model. This approach further obtained Dice coefficient accuracies of 0.931, 0.912, and 0.961 in GM, WM, and CSF, respectively, ranking first among 30 teams on the validation dataset of the iSeg-2019 Grand Challenge. Thus, the proposed 3D-MASNet can improve the accuracy of existing CNNs-based segmentation models as a plug-and-play solution that offers a promising technique for future infant brain MRI studies.
Collapse
Affiliation(s)
- Zilong Zeng
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
| | - Tengda Zhao
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
| | - Lianglong Sun
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
| | - Yihe Zhang
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
| | - Mingrui Xia
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
| | - Xuhong Liao
- School of Systems ScienceBeijing Normal UniversityBeijingChina
| | - Jiaying Zhang
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
| | - Dinggang Shen
- School of Biomedical EngineeringShanghaiTech UniversityShanghaiChina
- Shanghai Clinical Research and Trial CenterShanghaiChina
- Department of Research and DevelopmentShanghai United Imaging Intelligence Co., Ltd.ShanghaiChina
| | - Li Wang
- Department of Radiology and BRICUniversity of North Carolina at Chapel HillChapel HillNorth CarolinaUSA
| | - Yong He
- State Key Laboratory of Cognitive Neuroscience and LearningBeijing Normal UniversityBeijingChina
- Beijing Key Laboratory of Brain Imaging and ConnectomicsBeijing Normal UniversityBeijingChina
- IDG/McGovern Institute for Brain Research, Beijing Normal UniversityBeijingChina
- Chinese Institute for Brain ResearchBeijingChina
| |
Collapse
|
24
|
Allioui H, Mourdi Y, Sadgal M. Strong semantic segmentation for Covid-19 detection: Evaluating the use of deep learning models as a performant tool in radiography. Radiography (Lond) 2023; 29:109-118. [PMID: 36335787 PMCID: PMC9595354 DOI: 10.1016/j.radi.2022.10.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2022] [Revised: 10/12/2022] [Accepted: 10/17/2022] [Indexed: 11/06/2022]
Abstract
INTRODUCTION With the increasing number of Covid-19 cases as well as care costs, chest diseases have gained increasing interest in several communities, particularly in medical and computer vision. Clinical and analytical exams are widely recognized techniques for diagnosing and handling Covid-19 cases. However, strong detection tools can help avoid damage to chest tissues. The proposed method provides an important way to enhance the semantic segmentation process using combined potential deep learning (DL) modules to increase consistency. Based on Covid-19 CT images, this work hypothesized that a novel model for semantic segmentation might be able to extract definite graphical features of Covid-19 and afford an accurate clinical diagnosis while optimizing the classical test and saving time. METHODS CT images were collected considering different cases (normal chest CT, pneumonia, typical viral causes, and Covid-19 cases). The study presents an advanced DL method to deal with chest semantic segmentation issues. The approach employs a modified version of the U-net to enable and support Covid-19 detection from the studied images. RESULTS The validation tests demonstrated competitive results with important performance rates: Precision (90.96% ± 2.5) with an F-score of (91.08% ± 3.2), an accuracy of (93.37% ± 1.2), a sensitivity of (96.88% ± 2.8) and a specificity of (96.91% ± 2.3). In addition, the visual segmentation results are very close to the Ground truth. CONCLUSION The findings of this study reveal the proof-of-principle for using cooperative components to strengthen the semantic segmentation modules for effective and truthful Covid-19 diagnosis. IMPLICATIONS FOR PRACTICE This paper has highlighted that DL based approach, with several modules, may be contributing to provide strong support for radiographers and physicians, and that further use of DL is required to design and implement performant automated vision systems to detect chest diseases.
Collapse
Affiliation(s)
- H Allioui
- Computer Sciences Department, Faculty of Sciences Semlalia, Cadi Ayyad University, Morocco.
| | - Y Mourdi
- Polydisciplinary Faculty Safi, Cadi Ayyad University, Morocco.
| | - M Sadgal
- Computer Sciences Department, Faculty of Sciences Semlalia, Cadi Ayyad University, Morocco.
| |
Collapse
|
25
|
Zhao L, Ma J, Shao Y, Jia C, Zhao J, Yuan H. MM-UNet: A multimodality brain tumor segmentation network in MRI images. Front Oncol 2022; 12:950706. [PMID: 36059677 PMCID: PMC9434799 DOI: 10.3389/fonc.2022.950706] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2022] [Accepted: 07/20/2022] [Indexed: 11/30/2022] Open
Abstract
The global annual incidence of brain tumors is approximately seven out of 100,000, accounting for 2% of all tumors. The mortality rate ranks first among children under 12 and 10th among adults. Therefore, the localization and segmentation of brain tumor images constitute an active field of medical research. The traditional manual segmentation method is time-consuming, laborious, and subjective. In addition, the information provided by a single-image modality is often limited and cannot meet the needs of clinical application. Therefore, in this study, we developed a multimodality feature fusion network, MM-UNet, for brain tumor segmentation by adopting a multi-encoder and single-decoder structure. In the proposed network, each encoder independently extracts low-level features from the corresponding imaging modality, and the hybrid attention block strengthens the features. After fusion with the high-level semantic of the decoder path through skip connection, the decoder restores the pixel-level segmentation results. We evaluated the performance of the proposed model on the BraTS 2020 dataset. MM-UNet achieved the mean Dice score of 79.2% and mean Hausdorff distance of 8.466, which is a consistent performance improvement over the U-Net, Attention U-Net, and ResUNet baseline models and demonstrates the effectiveness of the proposed model.
Collapse
Affiliation(s)
- Liang Zhao
- School of Software Technology, Dalian University of Technology, Dalian, China
| | - Jiajun Ma
- School of Software Technology, Dalian University of Technology, Dalian, China
| | - Yu Shao
- School of Software Technology, Dalian University of Technology, Dalian, China
| | - Chaoran Jia
- School of Software Technology, Dalian University of Technology, Dalian, China
| | - Jingyuan Zhao
- Stem Cell Clinical Research Center, The First Affiliated Hospital of Dalian Medical University, Dalian, China
- *Correspondence: Jingyuan Zhao, ; Hong Yuan,
| | - Hong Yuan
- The Affiliated Central Hospital, Dalian University of Technology, Dalian, China
- *Correspondence: Jingyuan Zhao, ; Hong Yuan,
| |
Collapse
|
26
|
Fully Convolutional Neural Network for Improved Brain Segmentation. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2022. [DOI: 10.1007/s13369-022-07169-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
27
|
Khaled A, Han JJ, Ghaleb TA. Learning to detect boundary information for brain image segmentation. BMC Bioinformatics 2022; 23:332. [PMID: 35953776 PMCID: PMC9367147 DOI: 10.1186/s12859-022-04882-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Accepted: 07/30/2022] [Indexed: 11/14/2022] Open
Abstract
MRI brain images are always of low contrast, which makes it difficult to identify to which area the information at the boundary of brain images belongs. This can make the extraction of features at the boundary more challenging, since those features can be misleading as they might mix properties of different brain regions. Hence, to alleviate such a problem, image boundary detection plays a vital role in medical image segmentation, and brain segmentation in particular, as unclear boundaries can worsen brain segmentation results. Yet, given the low quality of brain images, boundary detection in the context of brain image segmentation remains challenging. Despite the research invested to improve boundary detection and brain segmentation, these two problems were addressed independently, i.e., little attention was paid to applying boundary detection to brain segmentation tasks. Therefore, in this paper, we propose a boundary detection-based model for brain image segmentation. To this end, we first design a boundary segmentation network for detecting and segmenting images brain tissues. Then, we design a boundary information module (BIM) to distinguish boundaries from the three different brain tissues. After that, we add a boundary attention gate (BAG) to the encoder output layers of our transformer to capture more informative local details. We evaluate our proposed model on two datasets of brain tissue images, including infant and adult brains. The extensive evaluation experiments of our model show better performance (a Dice Coefficient (DC) accuracy of up to \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$5.3\%$$\end{document}5.3% compared to the state-of-the-art models) in detecting and segmenting brain tissue images.
Collapse
Affiliation(s)
- Afifa Khaled
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China.
| | - Jian-Jun Han
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Taher A Ghaleb
- School of Electrical Engineering and Computer Science, University of Ottawa, Ottawa, Canada
| |
Collapse
|
28
|
SIP-UNet: Sequential Inputs Parallel UNet Architecture for Segmentation of Brain Tissues from Magnetic Resonance Images. MATHEMATICS 2022. [DOI: 10.3390/math10152755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Proper analysis of changes in brain structure can lead to a more accurate diagnosis of specific brain disorders. The accuracy of segmentation is crucial for quantifying changes in brain structure. In recent studies, UNet-based architectures have outperformed other deep learning architectures in biomedical image segmentation. However, improving segmentation accuracy is challenging due to the low resolution of medical images and insufficient data. In this study, we present a novel architecture that combines three parallel UNets using a residual network. This architecture improves upon the baseline methods in three ways. First, instead of using a single image as input, we use three consecutive images. This gives our model the freedom to learn from neighboring images as well. Additionally, the images are individually compressed and decompressed using three different UNets, which prevents the model from merging the features of the images. Finally, following the residual network architecture, the outputs of the UNets are combined in such a way that the features of the image corresponding to the output are enhanced by a skip connection. The proposed architecture performed better than using a single conventional UNet and other UNet variants.
Collapse
|
29
|
SM-SegNet: A Lightweight Squeeze M-SegNet for Tissue Segmentation in Brain MRI Scans. SENSORS 2022; 22:s22145148. [PMID: 35890829 PMCID: PMC9319649 DOI: 10.3390/s22145148] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/12/2022] [Revised: 07/02/2022] [Accepted: 07/05/2022] [Indexed: 11/17/2022]
Abstract
In this paper, we propose a novel squeeze M-SegNet (SM-SegNet) architecture featuring a fire module to perform accurate as well as fast segmentation of the brain on magnetic resonance imaging (MRI) scans. The proposed model utilizes uniform input patches, combined-connections, long skip connections, and squeeze-expand convolutional layers from the fire module to segment brain MRI data. The proposed SM-SegNet architecture involves a multi-scale deep network on the encoder side and deep supervision on the decoder side, which uses combined-connections (skip connections and pooling indices) from the encoder to the decoder layer. The multi-scale side input layers support the deep network layers' extraction of discriminative feature information, and the decoder side provides deep supervision to reduce the gradient problem. By using combined-connections, extracted features can be transferred from the encoder to the decoder resulting in recovering spatial information, which makes the model converge faster. Long skip connections were used to stabilize the gradient updates in the network. Owing to the adoption of the fire module, the proposed model was significantly faster to train and offered a more efficient memory usage with 83% fewer parameters than previously developed methods, owing to the adoption of the fire module. The proposed method was evaluated using the open-access series of imaging studies (OASIS) and the internet brain segmentation registry (IBSR) datasets. The experimental results demonstrate that the proposed SM-SegNet architecture achieves segmentation accuracies of 95% for cerebrospinal fluid, 95% for gray matter, and 96% for white matter, which outperforms the existing methods in both subjective and objective metrics in brain MRI segmentation.
Collapse
|
30
|
Unpaired multi-modal tumor segmentation with structure adaptation. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03610-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
31
|
Zhou J, Zhang X, Zhu Z, Lan X, Fu L, Wang H, Wen H. Cohesive Multi-Modality Feature Learning and Fusion for COVID-19 Patient Severity Prediction. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY : A PUBLICATION OF THE CIRCUITS AND SYSTEMS SOCIETY 2022; 32:2535-2549. [PMID: 35937181 PMCID: PMC9280852 DOI: 10.1109/tcsvt.2021.3063952] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/10/2020] [Revised: 02/14/2021] [Accepted: 02/25/2021] [Indexed: 06/15/2023]
Abstract
The outbreak of coronavirus disease (COVID-19) has been a nightmare to citizens, hospitals, healthcare practitioners, and the economy in 2020. The overwhelming number of confirmed cases and suspected cases put forward an unprecedented challenge to the hospital's capacity of management and medical resource distribution. To reduce the possibility of cross-infection and attend a patient according to his severity level, expertly diagnosis and sophisticated medical examinations are often required but hard to fulfil during a pandemic. To facilitate the assessment of a patient's severity, this paper proposes a multi-modality feature learning and fusion model for end-to-end covid patient severity prediction using the blood test supported electronic medical record (EMR) and chest computerized tomography (CT) scan images. To evaluate a patient's severity by the co-occurrence of salient clinical features, the High-order Factorization Network (HoFN) is proposed to learn the impact of a set of clinical features without tedious feature engineering. On the other hand, an attention-based deep convolutional neural network (CNN) using pre-trained parameters are used to process the lung CT images. Finally, to achieve cohesion of cross-modality representation, we design a loss function to shift deep features of both-modality into the same feature space which improves the model's performance and robustness when one modality is absent. Experimental results demonstrate that the proposed multi-modality feature learning and fusion model achieves high performance in an authentic scenario.
Collapse
Affiliation(s)
- Jinzhao Zhou
- School of Computer Science and EngineeringSouth China University of TechnologyGuangzhou510641China
| | - Xingming Zhang
- School of Computer Science and EngineeringSouth China University of TechnologyGuangzhou510641China
| | - Ziwei Zhu
- School of Computer Science and EngineeringSouth China University of TechnologyGuangzhou510641China
| | - Xiangyuan Lan
- Department of Computer ScienceHong Kong Baptist UniversityHong Kong
| | - Lunkai Fu
- School of Computer Science and EngineeringSouth China University of TechnologyGuangzhou510641China
| | - Haoxiang Wang
- School of Computer Science and EngineeringSouth China University of TechnologyGuangzhou510641China
| | - Hanchun Wen
- Department of Critical Care MedicineThe First Affiliated Hospital of Guangxi Medical UniversityNanning530021China
| |
Collapse
|
32
|
Shiohama T, Tsujimura K. Quantitative Structural Brain Magnetic Resonance Imaging Analyses: Methodological Overview and Application to Rett Syndrome. Front Neurosci 2022; 16:835964. [PMID: 35450016 PMCID: PMC9016334 DOI: 10.3389/fnins.2022.835964] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2021] [Accepted: 03/14/2022] [Indexed: 11/13/2022] Open
Abstract
Congenital genetic disorders often present with neurological manifestations such as neurodevelopmental disorders, motor developmental retardation, epilepsy, and involuntary movement. Through qualitative morphometric evaluation of neuroimaging studies, remarkable structural abnormalities, such as lissencephaly, polymicrogyria, white matter lesions, and cortical tubers, have been identified in these disorders, while no structural abnormalities were identified in clinical settings in a large population. Recent advances in data analysis programs have led to significant progress in the quantitative analysis of anatomical structural magnetic resonance imaging (MRI) and diffusion-weighted MRI tractography, and these approaches have been used to investigate psychological and congenital genetic disorders. Evaluation of morphometric brain characteristics may contribute to the identification of neuroimaging biomarkers for early diagnosis and response evaluation in patients with congenital genetic diseases. This mini-review focuses on the methodologies and attempts employed to study Rett syndrome using quantitative structural brain MRI analyses, including voxel- and surface-based morphometry and diffusion-weighted MRI tractography. The mini-review aims to deepen our understanding of how neuroimaging studies are used to examine congenital genetic disorders.
Collapse
Affiliation(s)
- Tadashi Shiohama
- Department of Pediatrics, Chiba University Hospital, Chiba, Japan
- *Correspondence: Tadashi Shiohama,
| | - Keita Tsujimura
- Group of Brain Function and Development, Nagoya University Neuroscience Institute of the Graduate School of Science, Nagoya, Japan
- Research Unit for Developmental Disorders, Institute for Advanced Research, Nagoya University, Nagoya, Japan
- Department of Radiology, Harvard Medical School, Boston, MA, United States
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA, United States
| |
Collapse
|
33
|
Deep 3D Neural Network for Brain Structures Segmentation Using Self-Attention Modules in MRI Images. SENSORS 2022; 22:s22072559. [PMID: 35408173 PMCID: PMC9002763 DOI: 10.3390/s22072559] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 03/15/2022] [Accepted: 03/21/2022] [Indexed: 01/03/2023]
Abstract
In recent years, the use of deep learning-based models for developing advanced healthcare systems has been growing due to the results they can achieve. However, the majority of the proposed deep learning-models largely use convolutional and pooling operations, causing a loss in valuable data and focusing on local information. In this paper, we propose a deep learning-based approach that uses global and local features which are of importance in the medical image segmentation process. In order to train the architecture, we used extracted three-dimensional (3D) blocks from the full magnetic resonance image resolution, which were sent through a set of successive convolutional neural network (CNN) layers free of pooling operations to extract local information. Later, we sent the resulting feature maps to successive layers of self-attention modules to obtain the global context, whose output was later dispatched to the decoder pipeline composed mostly of upsampling layers. The model was trained using the Mindboggle-101 dataset. The experimental results showed that the self-attention modules allow segmentation with a higher Mean Dice Score of 0.90 ± 0.036 compared with other UNet-based approaches. The average segmentation time was approximately 0.038 s per brain structure. The proposed model allows tackling the brain structure segmentation task properly. Exploiting the global context that the self-attention modules incorporate allows for more precise and faster segmentation. We segmented 37 brain structures and, to the best of our knowledge, it is the largest number of structures under a 3D approach using attention mechanisms.
Collapse
|
34
|
Zhu X, Wu Y, Hu H, Zhuang X, Yao J, Ou D, Li W, Song M, Feng N, Xu D. Medical lesion segmentation by combining multi‐modal images with modality weighted UNet. Med Phys 2022; 49:3692-3704. [PMID: 35312077 DOI: 10.1002/mp.15610] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2021] [Revised: 02/25/2022] [Accepted: 03/04/2022] [Indexed: 11/09/2022] Open
Affiliation(s)
- Xiner Zhu
- College of Information Science and Electronic Engineering Zhejiang University Hangzhou China
| | - Yichao Wu
- College of Information Science and Electronic Engineering Zhejiang University Hangzhou China
| | - Haoji Hu
- College of Information Science and Electronic Engineering Zhejiang University Hangzhou China
| | - Xianwei Zhuang
- College of Information Science and Electronic Engineering Zhejiang University Hangzhou China
| | - Jincao Yao
- Cancer Hospital of the University of Chinese Academy of Sciences (Zhejiang Cancer Hospital) Hangzhou China
- Institute of Basic Medicine and Cancer (IBMC) Chinese Academy of Sciences Hangzhou China
| | - Di Ou
- Cancer Hospital of the University of Chinese Academy of Sciences (Zhejiang Cancer Hospital) Hangzhou China
- Institute of Basic Medicine and Cancer (IBMC) Chinese Academy of Sciences Hangzhou China
| | - Wei Li
- Cancer Hospital of the University of Chinese Academy of Sciences (Zhejiang Cancer Hospital) Hangzhou China
- Institute of Basic Medicine and Cancer (IBMC) Chinese Academy of Sciences Hangzhou China
| | - Mei Song
- Cancer Hospital of the University of Chinese Academy of Sciences (Zhejiang Cancer Hospital) Hangzhou China
- Institute of Basic Medicine and Cancer (IBMC) Chinese Academy of Sciences Hangzhou China
| | - Na Feng
- Cancer Hospital of the University of Chinese Academy of Sciences (Zhejiang Cancer Hospital) Hangzhou China
- Institute of Basic Medicine and Cancer (IBMC) Chinese Academy of Sciences Hangzhou China
| | - Dong Xu
- Cancer Hospital of the University of Chinese Academy of Sciences (Zhejiang Cancer Hospital) Hangzhou China
- Institute of Basic Medicine and Cancer (IBMC) Chinese Academy of Sciences Hangzhou China
| |
Collapse
|
35
|
Abstract
Medical images of brain tumors are critical for characterizing the pathology of tumors and early diagnosis. There are multiple modalities for medical images of brain tumors. Fusing the unique features of each modality of the magnetic resonance imaging (MRI) scans can accurately determine the nature of brain tumors. The current genetic analysis approach is time-consuming and requires surgical extraction of brain tissue samples. Accurate classification of multi-modal brain tumor images can speed up the detection process and alleviate patient suffering. Medical image fusion refers to effectively merging the significant information of multiple source images of the same tissue into one image, which will carry abundant information for diagnosis. This paper proposes a novel attentive deep-learning-based classification model that integrates multi-modal feature aggregation, lite attention mechanism, separable embedding, and modal-wise shortcuts for performance improvement. We evaluate our model on the RSNA-MICCAI dataset, a scenario-specific medical image dataset, and demonstrate that the proposed method outperforms the state-of-the-art (SOTA) by around 3%.
Collapse
|
36
|
Li D, Peng Y, Guo Y, Sun J. TAUNet: a triple-attention-based multi-modality MRI fusion U-Net for cardiac pathology segmentation. COMPLEX INTELL SYST 2022. [DOI: 10.1007/s40747-022-00660-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Abstract
AbstractAutomated segmentation of cardiac pathology in MRI plays a significant role for diagnosis and treatment of some cardiac disease. In clinical practice, multi-modality MRI is widely used to improve the cardiac pathology segmentation, because it can provide multiple or complementary information. Recently, deep learning methods have presented implausible performance in multi-modality medical image segmentation. However, how to fuse the underlying multi-modality information effectively to segment the pathology with irregular shapes and small region at random locations, is still a challenge task. In this paper, a triple-attention-based multi-modality MRI fusion U-Net was proposed to learn complex relationship between different modalities and pay more attention on shape information, thus to achieve improved pathology segmentation. First, three independent encoders and one fusion encoder were applied to extract specific and multiple modality features. Secondly, we concatenate the modality feature maps and use the channel attention to fuse specific modal information at every stage of the three dedicate independent encoders, then the three single modality feature maps and channel attention feature maps are together concatenated to the decoder path. Spatial attention was adopted in decoder path to capture the correlation of various positions. Once more, we employ shape attention to focus shape-dependent information. Lastly, the training approach is made efficient by introducing deep supervision mechanism with object contextual representations block to ensure precisely boundary prediction. Our proposed network was evaluated on the public MICCAI 2020 Myocardial pathology segmentation dataset which involves patients suffering from myocardial infarction. Experiments on the dataset with three modalities demonstrate the effectiveness of fusion mode of our model, and attention mechanism can integrate various modality information well. We demonstrated that such a deep learning approach could better fuse complementary information to improve the segmentation performance of cardiac pathology.
Collapse
|
37
|
Lei T, Wang R, Zhang Y, Wan Y, Liu C, Nandi AK. DefED-Net: Deformable Encoder-Decoder Network for Liver and Liver Tumor Segmentation. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2022. [DOI: 10.1109/trpms.2021.3059780] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
38
|
Huang S, Cheng Z, Lai L, Zheng W, He M, Li J, Zeng T, Huang X, Yang X. Integrating multiple MRI sequences for pelvic organs segmentation via the attention mechanism. Med Phys 2021; 48:7930-7945. [PMID: 34658035 DOI: 10.1002/mp.15285] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2021] [Revised: 08/30/2021] [Accepted: 09/14/2021] [Indexed: 11/06/2022] Open
Abstract
PURPOSE To create a network which fully utilizes multi-sequence MRI and compares favorably with manual human contouring. METHODS We retrospectively collected 89 MRI studies of the pelvic cavity from patients with prostate cancer and cervical cancer. The dataset contained 89 samples from 87 patients with a total of 84 valid samples. MRI was performed with T1-weighted (T1), T2-weighted (T2), and Enhanced Dixon T1-weighted (T1DIXONC) sequences. There were two cohorts. The training cohort contained 55 samples and the testing cohort contained 29 samples. The MRI images in the training cohort contained contouring data from radiotherapist α. The MRI images in the testing cohort contained contouring data from radiotherapist α and contouring data from another radiotherapist: radiotherapist β. The training cohort was used to optimize the convolution neural networks, which included the attention mechanism through the proposed activation module and the blended module into multiple MRI sequences, to perform autodelineation. The testing cohort was used to assess the networks' autodelineation performance. The contoured organs at risk (OAR) were the anal canal, bladder, rectum, femoral head (L), and femoral head (R). RESULTS We compared our proposed network with UNet and FuseUNet using our dataset. When T1 was the main sequence, we input three sequences to segment five organs and evaluated the results using four metrics: the DSC (Dice similarity coefficient), the JSC (Jaccard similarity coefficient), the ASD (average mean distance), and the 95% HD (robust Hausdorff distance). The proposed network achieved improved results compared with the baselines among all metrics. The DSC were 0.834±0.029, 0.818±0.037, and 0.808±0.050 for our proposed network, FuseUNet, and UNet, respectively. The 95% HD were 7.256±2.748 mm, 8.404±3.297 mm, and 8.951±4.798 mm for our proposed network, FuseUNet, and UNet, respectively. Our proposed network also had superior performance on the JSC and ASD coefficients. CONCLUSION Our proposed activation module and blended module significantly improved the performance of FuseUNet for multi-sequence MRI segmentation. Our proposed network integrated multiple MRI sequences efficiently and autosegmented OAR rapidly and accurately. We also discovered that three-sequence fusion (T1-T1DIXONC-T2) was superior to two-sequence fusion (T1-T2 and T1-T1DIXONC, respectively). We infer that the more MRI sequences fused, the better the automatic segmentation results.
Collapse
Affiliation(s)
- Sijuan Huang
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China
| | - Zesen Cheng
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China.,School of Electronic and Information Engineering, South China University of Technology, Guangzhou, Guangdong, 510641, China
| | - Lijuan Lai
- School of Electronic and Information Engineering, South China University of Technology, Guangzhou, Guangdong, 510641, China
| | - Wanjia Zheng
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China.,Department of Radiation Oncology, Southern Theater Air Force Hospital of the People's Liberation Army, Guangzhou, Guangdong, 510050, China
| | - Mengxue He
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China
| | - Junyun Li
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China
| | - Tianyu Zeng
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China.,School of Electronic and Information Engineering, South China University of Technology, Guangzhou, Guangdong, 510641, China
| | - Xiaoyan Huang
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China
| | - Xin Yang
- Department of Radiation Oncology, Sun Yat-sen University Cancer Center, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangzhou, Guangdong, 510060, China
| |
Collapse
|
39
|
HybridCTrm: Bridging CNN and Transformer for Multimodal Brain Image Segmentation. JOURNAL OF HEALTHCARE ENGINEERING 2021; 2021:7467261. [PMID: 34630994 PMCID: PMC8500745 DOI: 10.1155/2021/7467261] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/09/2021] [Revised: 09/03/2021] [Accepted: 09/16/2021] [Indexed: 11/17/2022]
Abstract
Multimodal medical image segmentation is always a critical problem in medical image segmentation. Traditional deep learning methods utilize fully CNNs for encoding given images, thus leading to deficiency of long-range dependencies and bad generalization performance. Recently, a sequence of Transformer-based methodologies emerges in the field of image processing, which brings great generalization and performance in various tasks. On the other hand, traditional CNNs have their own advantages, such as rapid convergence and local representations. Therefore, we analyze a hybrid multimodal segmentation method based on Transformers and CNNs and propose a novel architecture, HybridCTrm network. We conduct experiments using HybridCTrm on two benchmark datasets and compare with HyperDenseNet, a network based on fully CNNs. Results show that our HybridCTrm outperforms HyperDenseNet on most of the evaluation metrics. Furthermore, we analyze the influence of the depth of Transformer on the performance. Besides, we visualize the results and carefully explore how our hybrid methods improve on segmentations.
Collapse
|
40
|
Zhang YN, XIA KR, LI CY, WEI BL, Zhang B. Review of Breast Cancer Pathologigcal Image Processing. BIOMED RESEARCH INTERNATIONAL 2021; 2021:1994764. [PMID: 34595234 PMCID: PMC8478535 DOI: 10.1155/2021/1994764] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/12/2021] [Accepted: 08/24/2021] [Indexed: 11/17/2022]
Abstract
Breast cancer is one of the most common malignancies. Pathological image processing of breast has become an important means for early diagnosis of breast cancer. Using medical image processing to assist doctors to detect potential breast cancer as early as possible has always been a hot topic in the field of medical image diagnosis. In this paper, a breast cancer recognition method based on image processing is systematically expounded from four aspects: breast cancer detection, image segmentation, image registration, and image fusion. The achievements and application scope of supervised learning, unsupervised learning, deep learning, CNN, and so on in breast cancer examination are expounded. The prospect of unsupervised learning and transfer learning for breast cancer diagnosis is prospected. Finally, the privacy protection of breast cancer patients is put forward.
Collapse
Affiliation(s)
- Ya-nan Zhang
- School of Computer Science and Technology, Harbin University of Science and Technology, Harbin 150080, China
- HRG International Institute (Hefei) of Research and Innovation, Hefei 230000, China
| | - Ke-rui XIA
- HRG International Institute (Hefei) of Research and Innovation, Hefei 230000, China
| | - Chang-yi LI
- School of Computer Science and Technology, Harbin University of Science and Technology, Harbin 150080, China
| | - Ben-li WEI
- School of Computer Science and Technology, Harbin University of Science and Technology, Harbin 150080, China
| | - Bing Zhang
- School of Computer Science and Technology, Harbin University of Science and Technology, Harbin 150080, China
| |
Collapse
|
41
|
Sun Y, Gao K, Lin W, Li G, Niu S, Wang L. Multi-Scale Self-Supervised Learning for Multi-Site Pediatric Brain MR Image Segmentation with Motion/Gibbs Artifacts. MACHINE LEARNING IN MEDICAL IMAGING. MLMI (WORKSHOP) 2021; 12966:171-179. [PMID: 35528703 PMCID: PMC9077100 DOI: 10.1007/978-3-030-87589-3_18] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Accurate tissue segmentation of large-scale pediatric brain MR images from multiple sites is essential to characterize early brain development. Due to imaging motion/Gibbs artifacts and multi-site issue (or domain shift issue), it remains a challenge to accurately segment brain tissues from multi-site pediatric MR images. In this paper, we present a multi-scale self-supervised learning (M-SSL) framework to accurately segment tissues for multi-site pediatric brain MR images with artifacts. Specifically, we first work on the downsampled images to estimate coarse tissue probabilities and build a global anatomic guidance. We then train another segmentation model based on the original images to estimate fine tissue probabilities, which are further integrated with the global anatomic guidance to refine the segmentation results. In the testing stage, to alleviate the multi-site issue, we propose an iterative self-supervised learning strategy to train a site-specific segmentation model based on a set of reliable training samples automatically generated for a to-be-segmented site. The experimental results on pediatric brain MR images with real artifacts and multi-site subjects from the iSeg2019 challenge demonstrate that our M-SSL method achieves better performance compared with several state-of-the-art methods.
Collapse
Affiliation(s)
- Yue Sun
- Department of Shandong Provincial Key Laboratory of Network based Intelligent Computing, University of Jinan, Jinan 250022, China
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, USA
| | - Kun Gao
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, USA
| | - Weili Lin
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, USA
| | - Gang Li
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, USA
| | - Sijie Niu
- Department of Shandong Provincial Key Laboratory of Network based Intelligent Computing, University of Jinan, Jinan 250022, China
| | - Li Wang
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, USA
| |
Collapse
|
42
|
GSCFN: A graph self-construction and fusion network for semi-supervised brain tissue segmentation in MRI. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.05.047] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
|
43
|
He K, Lian C, Zhang B, Zhang X, Cao X, Nie D, Gao Y, Zhang J, Shen D. HF-UNet: Learning Hierarchically Inter-Task Relevance in Multi-Task U-Net for Accurate Prostate Segmentation in CT Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2118-2128. [PMID: 33848243 DOI: 10.1109/tmi.2021.3072956] [Citation(s) in RCA: 32] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Accurate segmentation of the prostate is a key step in external beam radiation therapy treatments. In this paper, we tackle the challenging task of prostate segmentation in CT images by a two-stage network with 1) the first stage to fast localize, and 2) the second stage to accurately segment the prostate. To precisely segment the prostate in the second stage, we formulate prostate segmentation into a multi-task learning framework, which includes a main task to segment the prostate, and an auxiliary task to delineate the prostate boundary. Here, the second task is applied to provide additional guidance of unclear prostate boundary in CT images. Besides, the conventional multi-task deep networks typically share most of the parameters (i.e., feature representations) across all tasks, which may limit their data fitting ability, as the specificity of different tasks are inevitably ignored. By contrast, we solve them by a hierarchically-fused U-Net structure, namely HF-UNet. The HF-UNet has two complementary branches for two tasks, with the novel proposed attention-based task consistency learning block to communicate at each level between the two decoding branches. Therefore, HF-UNet endows the ability to learn hierarchically the shared representations for different tasks, and preserve the specificity of learned representations for different tasks simultaneously. We did extensive evaluations of the proposed method on a large planning CT image dataset and a benchmark prostate zonal dataset. The experimental results show HF-UNet outperforms the conventional multi-task network architectures and the state-of-the-art methods.
Collapse
|
44
|
Le N, Bui T, Vo-Ho VK, Yamazaki K, Luu K. Narrow Band Active Contour Attention Model for Medical Segmentation. Diagnostics (Basel) 2021; 11:1393. [PMID: 34441327 PMCID: PMC8393587 DOI: 10.3390/diagnostics11081393] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2021] [Revised: 07/06/2021] [Accepted: 07/09/2021] [Indexed: 11/16/2022] Open
Abstract
Medical image segmentation is one of the most challenging tasks in medical image analysis and widely developed for many clinical applications. While deep learning-based approaches have achieved impressive performance in semantic segmentation, they are limited to pixel-wise settings with imbalanced-class data problems and weak boundary object segmentation in medical images. In this paper, we tackle those limitations by developing a new two-branch deep network architecture which takes both higher level features and lower level features into account. The first branch extracts higher level feature as region information by a common encoder-decoder network structure such as Unet and FCN, whereas the second branch focuses on lower level features as support information around the boundary and processes in parallel to the first branch. Our key contribution is the second branch named Narrow Band Active Contour (NB-AC) attention model which treats the object contour as a hyperplane and all data inside a narrow band as support information that influences the position and orientation of the hyperplane. Our proposed NB-AC attention model incorporates the contour length with the region energy involving a fixed-width band around the curve or surface. The proposed network loss contains two fitting terms: (i) a high level feature (i.e., region) fitting term from the first branch; (ii) a lower level feature (i.e., contour) fitting term from the second branch including the (ii1) length of the object contour and (ii2) regional energy functional formed by the homogeneity criterion of both the inner band and outer band neighboring the evolving curve or surface. The proposed NB-AC loss can be incorporated into both 2D and 3D deep network architectures. The proposed network has been evaluated on different challenging medical image datasets, including DRIVE, iSeg17, MRBrainS18 and Brats18. The experimental results have shown that the proposed NB-AC loss outperforms other mainstream loss functions: Cross Entropy, Dice, Focal on two common segmentation frameworks Unet and FCN. Our 3D network which is built upon the proposed NB-AC loss and 3DUnet framework achieved state-of-the-art results on multiple volumetric datasets.
Collapse
Affiliation(s)
- Ngan Le
- Department of CSCE, University of Arkansas, Fayetteville, AR 72701, USA; (V.-K.V.-H.); (K.Y.); (K.L.)
| | - Toan Bui
- Vin-AI Research, HaNoi 100000, Vietnam;
| | - Viet-Khoa Vo-Ho
- Department of CSCE, University of Arkansas, Fayetteville, AR 72701, USA; (V.-K.V.-H.); (K.Y.); (K.L.)
| | - Kashu Yamazaki
- Department of CSCE, University of Arkansas, Fayetteville, AR 72701, USA; (V.-K.V.-H.); (K.Y.); (K.L.)
| | - Khoa Luu
- Department of CSCE, University of Arkansas, Fayetteville, AR 72701, USA; (V.-K.V.-H.); (K.Y.); (K.L.)
| |
Collapse
|
45
|
Dai Y, Gao Y, Liu F. TransMed: Transformers Advance Multi-Modal Medical Image Classification. Diagnostics (Basel) 2021; 11:diagnostics11081384. [PMID: 34441318 PMCID: PMC8391808 DOI: 10.3390/diagnostics11081384] [Citation(s) in RCA: 118] [Impact Index Per Article: 29.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2021] [Revised: 07/07/2021] [Accepted: 07/28/2021] [Indexed: 12/24/2022] Open
Abstract
Over the past decade, convolutional neural networks (CNN) have shown very competitive performance in medical image analysis tasks, such as disease classification, tumor segmentation, and lesion detection. CNN has great advantages in extracting local features of images. However, due to the locality of convolution operation, it cannot deal with long-range relationships well. Recently, transformers have been applied to computer vision and achieved remarkable success in large-scale datasets. Compared with natural images, multi-modal medical images have explicit and important long-range dependencies, and effective multi-modal fusion strategies can greatly improve the performance of deep models. This prompts us to study transformer-based structures and apply them to multi-modal medical images. Existing transformer-based network architectures require large-scale datasets to achieve better performance. However, medical imaging datasets are relatively small, which makes it difficult to apply pure transformers to medical image analysis. Therefore, we propose TransMed for multi-modal medical image classification. TransMed combines the advantages of CNN and transformer to efficiently extract low-level features of images and establish long-range dependencies between modalities. We evaluated our model on two datasets, parotid gland tumors classification and knee injury classification. Combining our contributions, we achieve an improvement of 10.1% and 1.9% in average accuracy, respectively, outperforming other state-of-the-art CNN-based models. The results of the proposed method are promising and have tremendous potential to be applied to a large number of medical image analysis tasks. To our best knowledge, this is the first work to apply transformers to multi-modal medical image classification.
Collapse
Affiliation(s)
- Yin Dai
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China; (Y.D.); (Y.G.)
- Engineering Center on Medical Imaging and Intelligent Analysis, Ministry Education, Northeastern University, Shenyang 110169, China
| | - Yifan Gao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China; (Y.D.); (Y.G.)
| | - Fayu Liu
- Department of Oromaxillofacial-Head and Neck Surgery, School of Stomatology, China Medical University, Shenyang 110002, China
- Correspondence:
| |
Collapse
|
46
|
Dai Y, Gao Y, Liu F. TransMed: Transformers Advance Multi-Modal Medical Image Classification. Diagnostics (Basel) 2021. [PMID: 34441318 DOI: 10.1109/access.2017] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/30/2023] Open
Abstract
Over the past decade, convolutional neural networks (CNN) have shown very competitive performance in medical image analysis tasks, such as disease classification, tumor segmentation, and lesion detection. CNN has great advantages in extracting local features of images. However, due to the locality of convolution operation, it cannot deal with long-range relationships well. Recently, transformers have been applied to computer vision and achieved remarkable success in large-scale datasets. Compared with natural images, multi-modal medical images have explicit and important long-range dependencies, and effective multi-modal fusion strategies can greatly improve the performance of deep models. This prompts us to study transformer-based structures and apply them to multi-modal medical images. Existing transformer-based network architectures require large-scale datasets to achieve better performance. However, medical imaging datasets are relatively small, which makes it difficult to apply pure transformers to medical image analysis. Therefore, we propose TransMed for multi-modal medical image classification. TransMed combines the advantages of CNN and transformer to efficiently extract low-level features of images and establish long-range dependencies between modalities. We evaluated our model on two datasets, parotid gland tumors classification and knee injury classification. Combining our contributions, we achieve an improvement of 10.1% and 1.9% in average accuracy, respectively, outperforming other state-of-the-art CNN-based models. The results of the proposed method are promising and have tremendous potential to be applied to a large number of medical image analysis tasks. To our best knowledge, this is the first work to apply transformers to multi-modal medical image classification.
Collapse
Affiliation(s)
- Yin Dai
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China
- Engineering Center on Medical Imaging and Intelligent Analysis, Ministry Education, Northeastern University, Shenyang 110169, China
| | - Yifan Gao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China
| | - Fayu Liu
- Department of Oromaxillofacial-Head and Neck Surgery, School of Stomatology, China Medical University, Shenyang 110002, China
| |
Collapse
|
47
|
Shaari H, Kevrić J, Jukić S, Bešić L, Jokić D, Ahmed N, Rajs V. Deep Learning-Based Studies on Pediatric Brain Tumors Imaging: Narrative Review of Techniques and Challenges. Brain Sci 2021; 11:brainsci11060716. [PMID: 34071202 PMCID: PMC8230188 DOI: 10.3390/brainsci11060716] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2021] [Revised: 05/10/2021] [Accepted: 05/17/2021] [Indexed: 11/16/2022] Open
Abstract
Brain tumors diagnosis in children is a scientific concern due to rapid anatomical, metabolic, and functional changes arising in the brain and non-specific or conflicting imaging results. Pediatric brain tumors diagnosis is typically centralized in clinical practice on the basis of diagnostic clues such as, child age, tumor location and incidence, clinical history, and imaging (Magnetic resonance imaging MRI / computed tomography CT) findings. The implementation of deep learning has rapidly propagated in almost every field in recent years, particularly in the medical images’ evaluation. This review would only address critical deep learning issues specific to pediatric brain tumor imaging research in view of the vast spectrum of other applications of deep learning. The purpose of this review paper is to include a detailed summary by first providing a succinct guide to the types of pediatric brain tumors and pediatric brain tumor imaging techniques. Then, we will present the research carried out by summarizing the scientific contributions to the field of pediatric brain tumor imaging processing and analysis. Finally, to establish open research issues and guidance for potential study in this emerging area, the medical and technical limitations of the deep learning-based approach were included.
Collapse
Affiliation(s)
- Hala Shaari
- Department of Information Technologies, Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina;
| | - Jasmin Kevrić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Samed Jukić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Larisa Bešić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Dejan Jokić
- Faculty of Engineering and Natural Sciences, International BURCH University, 71000 Sarajevo, Bosnia and Herzegovina; (J.K.); (S.J.); (L.B.); (D.J.)
| | - Nuredin Ahmed
- Control Department, Technical Computer College Tripoli, Tripoli 00218, Libya;
| | - Vladimir Rajs
- Department of Power, Electronics and Telecommunication Engineering, Faculty of Technical Science, University of Novi Sad, 21000 Novi Sad, Serbia
- Correspondence:
| |
Collapse
|
48
|
Learning U-Net Based Multi-Scale Features in Encoding-Decoding for MR Image Brain Tissue Segmentation. SENSORS 2021; 21:s21093232. [PMID: 34067101 PMCID: PMC8124734 DOI: 10.3390/s21093232] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/15/2021] [Revised: 04/27/2021] [Accepted: 04/28/2021] [Indexed: 11/17/2022]
Abstract
Accurate brain tissue segmentation of MRI is vital to diagnosis aiding, treatment planning, and neurologic condition monitoring. As an excellent convolutional neural network (CNN), U-Net is widely used in MR image segmentation as it usually generates high-precision features. However, the performance of U-Net is considerably restricted due to the variable shapes of the segmented targets in MRI and the information loss of down-sampling and up-sampling operations. Therefore, we propose a novel network by introducing spatial and channel dimensions-based multi-scale feature information extractors into its encoding-decoding framework, which is helpful in extracting rich multi-scale features while highlighting the details of higher-level features in the encoding part, and recovering the corresponding localization to a higher resolution layer in the decoding part. Concretely, we propose two information extractors, multi-branch pooling, called MP, in the encoding part, and multi-branch dense prediction, called MDP, in the decoding part, to extract multi-scale features. Additionally, we designed a new multi-branch output structure with MDP in the decoding part to form more accurate edge-preserving predicting maps by integrating the dense adjacent prediction features at different scales. Finally, the proposed method is tested on datasets MRbrainS13, IBSR18, and ISeg2017. We find that the proposed network performs higher accuracy in segmenting MRI brain tissues and it is better than the leading method of 2018 at the segmentation of GM and CSF. Therefore, it can be a useful tool for diagnostic applications, such as brain MRI segmentation and diagnosing.
Collapse
|
49
|
Oh KT, Kim D, Ye BS, Lee S, Yun M, Yoo SK. Segmentation of white matter hyperintensities on 18F-FDG PET/CT images with a generative adversarial network. Eur J Nucl Med Mol Imaging 2021; 48:3422-3431. [PMID: 33693968 DOI: 10.1007/s00259-021-05285-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2020] [Accepted: 02/25/2021] [Indexed: 11/28/2022]
Abstract
PURPOSE White matter hyperintensities (WMH) are typically segmented using MRI because WMH are hardly visible on 18F-FDG PET/CT. This retrospective study was conducted to segment WMH and estimate their volumes from 18F-FDG PET with a generative adversarial network (WhyperGAN). METHODS We selected patients whose interval between MRI and FDG PET/CT scans was within 3 months, from January 2017 to December 2018, and classified them into mild, moderate, and severe groups by following the semiquantitative rating method of Fazekas. For each group, 50 patients were selected, and of them, we randomly selected 35 patients for training and 15 for testing. WMH were automatically segmented from FLAIR MRI with manual adjustment. Patches of WMH were extracted from 18F-FDG PET and segmented MRI. WhyperGAN was compared with H-DenseUnet, a deep learning method widely used for segmentation tasks, for segmentation performance based on the dice similarity coefficient (DSC), recall, and average volume differences (AVD). For volume estimation, the predicted WMH volumes from PET were compared with ground truth volumes. RESULTS The DSC values were associated with WMH volumes on MRI. For volumes >60 mL, the DSC values were 0.751 for WhyperGAN and 0.564 for H-DenseUnet. For volumes ≤60 mL, the DSC values rapidly decreased as the volume decreased (0.362 for WhyperGAN vs. 0.237 for H-DenseUnet). For recall, WhyperGAN achieved the highest value in the severe group (0.579 for WhyperGAN vs. 0.509 for H-DenseUnet). For AVD, WhyperGAN achieved the lowest score in the severe group (0.494 for WhyperGAN vs. 0.941 for H-DenseUnet). For the WMH volume estimation, WhyperGAN performed better than H-DenseUnet and yielded excellent correlation coefficients (r = 0.998, 0.983, and 0.908 in the severe, moderate, and mild group). CONCLUSIONS Although limited by visual analysis, the WhyperGAN based can be used to automatically segment and estimate volumes of WMH from 18F-FDG PET/CT. This would increase the usefulness of 18F-FDG PET/CT for the evaluation of WMH in patients with cognitive impairment.
Collapse
Affiliation(s)
- Kyeong Taek Oh
- Department of Medical Engineering, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Dongwoo Kim
- Department of Nuclear Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Byoung Seok Ye
- Department of Neurology, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Sangwon Lee
- Department of Nuclear Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Mijin Yun
- Department of Nuclear Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea.
| | - Sun Kook Yoo
- Department of Medical Engineering, Yonsei University College of Medicine, Seoul, Republic of Korea.
| |
Collapse
|
50
|
Mallow GM, Siyaji ZK, Galbusera F, Espinoza-Orías AA, Giers M, Lundberg H, Ames C, Karppinen J, Louie PK, Phillips FM, Pourzal R, Schwab J, Sciubba DM, Wang JC, Wilke HJ, Williams FMK, Mohiuddin SA, Makhni MC, Shepard NA, An HS, Samartzis D. Intelligence-Based Spine Care Model: A New Era of Research and Clinical Decision-Making. Global Spine J 2021; 11:135-145. [PMID: 33251858 PMCID: PMC7882816 DOI: 10.1177/2192568220973984] [Citation(s) in RCA: 20] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/14/2022] Open
Affiliation(s)
- G. Michael Mallow
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| | - Zakariah K. Siyaji
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| | | | - Alejandro A. Espinoza-Orías
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| | - Morgan Giers
- School of Chemical, Biological, and Environmental Engineering, Oregon State University, Corvallis, OR, USA
| | - Hannah Lundberg
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
| | - Christopher Ames
- Department of Neurosurgery, University of California San Francisco, CA, USA
| | - Jaro Karppinen
- Medical Research Center Oulu, University of Oulu and Oulu University Hospital, Oulu, Finland
| | | | - Frank M. Phillips
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| | - Robin Pourzal
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
| | - Joseph Schwab
- Department of Orthopaedic Surgery, Harvard Medical School, Boston, MA, USA
| | - Daniel M. Sciubba
- Department of Neurosurgery, Johns Hopkins University, Baltimore, MD, USA
| | - Jeffrey C. Wang
- Department of Orthopaedic Surgery, University of Southern California, Los Angeles, CA, USA
| | - Hans-Joachim Wilke
- Institute of Orthopaedic Research and Biomechanics, Centre for Trauma Research Ulm, Ulm University Medical Centre, Ulm, Germany
| | - Frances M. K. Williams
- Department of Twin Research and Genetic Epidemiology, King’s College London, London, United Kingdom
| | | | - Melvin C. Makhni
- Department of Orthopaedic Surgery, Harvard Medical School, Boston, MA, USA
| | - Nicholas A. Shepard
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| | - Howard S. An
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| | - Dino Samartzis
- Department of Orthopedic Surgery, Rush University Medical Center, Chicago, IL, USA
- The International Spine Research and Innovation Initiative, Rush University Medical Center, Chicago, IL, USA
| |
Collapse
|