1
|
Wang X, Zhang J, Teng X, Aik KW, Natividad L, Cheng C, Wong APC, Keng FYJ, Koh AS, Huang W. A Multi-Modality Attention Network for Coronary Artery Disease Evaluation From Routine Myocardial Perfusion Imaging and Clinical Data. IEEE J Biomed Health Inform 2025; 29:3272-3281. [PMID: 40030779 DOI: 10.1109/jbhi.2024.3523476] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Myocardial perfusion imaging (MPI) is an essential tool for diagnosing and evaluating coronary artery disease (CAD). However, the diagnosis using MPI remains laborious as it involves multi-step readouts and meticulous image processing. These challenges impact current attempts at automating image interpretation of MPI. In this paper, we propose a multi-modality attention network (MMAN) that leverages information from clinical and MPI data for CAD diagnosis. Specifically, we propose an image-correlated cross-attention (ICCA) module that fuses information from both stress and rest MPI to enhance feature representation at the image level. Furthermore, we design a clinical data-guided attention (CDGA) module that integrates clinical data with image features to improve overall feature understanding for CAD evaluation. In addition, we employ self-learning for network pre-training, which further enhances the diagnostic performance using MPI on CAD. Experiments on a myocardial perfusion imaging dataset demonstrate that the proposed method is effective for CAD evaluation using myocardial perfusion imaging and clinical data.
Collapse
|
2
|
Hein D, Bozorgpour A, Merhof D, Wang G. Physics-Inspired Generative Models in Medical Imaging. Annu Rev Biomed Eng 2025; 27:499-525. [PMID: 40310888 DOI: 10.1146/annurev-bioeng-102723-013922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/03/2025]
Abstract
Physics-inspired generative models (GMs), in particular diffusion models and Poisson flow models, enhance Bayesian methods and promise great utility in medical imaging. This review examines the transformative role of such generative methods. First, a variety of physics-inspired GMs, including denoising diffusion probabilistic models, score-based diffusion models, and Poisson flow generative models (including PFGM++), are revisited, with an emphasis on their accuracy, robustness and acceleration. Then, major applications of physics-inspired GMs in medical imaging are presented, comprising image reconstruction, image generation, and image analysis. Finally, future research directions are brainstormed, including unification of physics-inspired GMs, integration with vision-language models, and potential novel applications of GMs. Since the development of generative methods has been rapid, it is hoped that this review will give peers and learners a timely snapshot of this new family of physics-driven GMs and help capitalize their enormous potential for medical imaging.
Collapse
Affiliation(s)
- Dennis Hein
- Department of Physics, KTH Royal Institute of Technology, Stockholm, Sweden
- MedTechLabs, Karolinska University Hospital, Stockholm, Sweden
| | - Afshin Bozorgpour
- Faculty of Informatics and Data Science, University of Regensburg, Regensburg, Germany;
| | - Dorit Merhof
- Faculty of Informatics and Data Science, University of Regensburg, Regensburg, Germany;
| | - Ge Wang
- Department of Biomedical Engineering, School of Engineering, Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Rensselaer Polytechnic Institute, Troy, New York, USA;
| |
Collapse
|
3
|
Xu H, Wang J, Feng Q, Zhang Y, Ning Z. Domain-specific information preservation for Alzheimer's disease diagnosis with incomplete multi-modality neuroimages. Med Image Anal 2025; 101:103448. [PMID: 39798527 DOI: 10.1016/j.media.2024.103448] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2024] [Revised: 10/22/2024] [Accepted: 12/24/2024] [Indexed: 01/15/2025]
Abstract
Although multi-modality neuroimages have advanced the early diagnosis of Alzheimer's Disease (AD), missing modality issue still poses a unique challenge in the clinical practice. Recent studies have tried to impute the missing data so as to utilize all available subjects for training robust multi-modality models. However, these studies may overlook the modality-specific information inherent in multi-modality data, that is, different modalities possess distinct imaging characteristics and focus on different aspects of the disease. In this paper, we propose a domain-specific information preservation (DSIP) framework, consisting of modality imputation stage and status identification stage, for AD diagnosis with incomplete multi-modality neuroimages. In the first stage, a specificity-induced generative adversarial network (SIGAN) is developed to bridge the modality gap and capture modality-specific details for imputing high-quality neuroimages. In the second stage, a specificity-promoted diagnosis network (SPDN) is designed to promote the inter-modality feature interaction and the classifier robustness for identifying disease status accurately. Extensive experiments demonstrate the proposed method significantly outperforms state-of-the-art methods in both modality imputation and status identification tasks.
Collapse
Affiliation(s)
- Haozhe Xu
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; Department of Radiotherapy, State Key Laboratory of Oncology in South China, Guangdong Provincial Clinical Research Center for Cancer, Sun Yat-sen University Cancer Center, Guangzhou 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China
| | - Jian Wang
- Department of Radiation Oncology, Nanfang Hospital, Southern Medical University, Guangzhou, China
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China
| | - Yu Zhang
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China.
| | - Zhenyuan Ning
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China.
| |
Collapse
|
4
|
Huang Y, Wu Z, Xu X, Zhang M, Wang S, Liu Q. Partition-based k-space synthesis for multi-contrast parallel imaging. Magn Reson Imaging 2025; 117:110297. [PMID: 39647517 DOI: 10.1016/j.mri.2024.110297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2024] [Revised: 12/02/2024] [Accepted: 12/03/2024] [Indexed: 12/10/2024]
Abstract
PURPOSE Multi-contrast magnetic resonance imaging is a significant and essential medical imaging technique. However, multi-contrast imaging has longer acquisition time and is easy to cause motion artifacts. In particular, the acquisition time for a T2-weighted image is prolonged due to its longer repetition time (TR). On the contrary, T1-weighted image has a shorter TR. Therefore, utilizing complementary information across T1 and T2-weighted image is a way to decrease the overall imaging time. Previous T1-assisted T2 reconstruction methods have mostly focused on image domain using whole-based image fusion approaches. The image domain reconstruction method has the defects of high computational complexity and limited flexibility. To address this issue, we propose a novel multi-contrast imaging method called partition-based k-space synthesis (PKS) which can achieve better reconstruction quality of T2-weighted image by feature fusion. METHODS Concretely, we first decompose fully-sampled T1 k-space data and under-sampled T2 k-space data into two sub-data, separately. Then two new objects are constructed by combining the two sub-T1/T2 data. After that, the two new objects as the whole data to realize the reconstruction of T2-weighted image. RESULTS Experimental results showed that the developed PKS scheme can achieve comparable or better results than using traditional k-space parallel imaging (SAKE) that processes each contrast independently. At the same time, our method showed good adaptability and robustness under different contrast-assisted and T1-T2 ratios. Efficient target modal image reconstruction under various conditions were realized and had excellent performance in restoring image quality and preserving details. CONCLUSIONS This work proposed a PKS multi-contrast method to assist in target mode image reconstruction. We have conducted extensive experiments on different multi-contrast, diverse ratios of T1 to T2 and different sampling masks to demonstrate the generalization and robustness of our proposed model.
Collapse
Affiliation(s)
- Yuxia Huang
- Department of Electronic Information Engineering, Nanchang University, Nanchang 330031, China
| | - Zhonghui Wu
- Department of Electronic Information Engineering, Nanchang University, Nanchang 330031, China
| | - Xiaoling Xu
- Department of Electronic Information Engineering, Nanchang University, Nanchang 330031, China
| | - Minghui Zhang
- Department of Electronic Information Engineering, Nanchang University, Nanchang 330031, China
| | - Shanshan Wang
- Paul C. Lauterbur Research Center for Biomedical Imaging, SIAT, Chinese Academy of Sciences, Shenzhen 518055, China.
| | - Qiegen Liu
- Department of Electronic Information Engineering, Nanchang University, Nanchang 330031, China.
| |
Collapse
|
5
|
Cao B, Qi G, Zhao J, Zhu P, Hu Q, Gao X. RTF: Recursive TransFusion for Multi-Modal Image Synthesis. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2025; 34:1573-1587. [PMID: 40031796 DOI: 10.1109/tip.2025.3541877] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Multi-modal image synthesis is crucial for obtaining complete modalities due to the imaging restrictions in reality. Current methods, primarily CNN-based models, find it challenging to extract global representations because of local inductive bias, leading to synthetic structure deformation or color distortion. Despite the significant global representation ability of transformer in capturing long-range dependencies, its huge parameter size requires considerable training data. Multi-modal synthesis solely based on one of the two structures makes it hard to extract comprehensive information from each modality with limited data. To tackle this dilemma, we propose a simple yet effective Recursive TransFusion (RTF) framework for multi-modal image synthesis. Specifically, we develop a TransFusion unit to integrate local knowledge extracted from the individual modality by connecting a CNN-based local representation block (LRB) and a transformer-based global fusion block (GFB) via a feature translating gate (FTG). Considering the numerous parameters introduced by the transformer, we further unfold a TransFusion unit with recursive constraint repeatedly, forming recursive TransFusion (RTF), which progressively extracts multi-modal information at different depths. Our RTF remarkably reduces network parameters while maintaining superior performance. Extensive experiments validate our superiority against the competing methods on multiple benchmarks. The source code will be available at https://github.com/guoliangq/RTF.
Collapse
|
6
|
Fiszer J, Ciupek D, Malawski M, Pieciak T. Validation of ten federated learning strategies for multi-contrast image-to-image MRI data synthesis from heterogeneous sources. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2025:2025.02.09.637305. [PMID: 39990397 PMCID: PMC11844418 DOI: 10.1101/2025.02.09.637305] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 02/25/2025]
Abstract
Deep learning (DL)-based image synthesis has recently gained enormous interest in medical imaging, allowing for generating multi-contrast data and therefore, the recovery of missing samples from interrupted or artefact-distorted acquisitions. However, the accuracy of DL models heavily relies on the representativeness of the training datasets naturally characterized by their distributions, experimental setups or preprocessing schemes. These complicate generalizing DL models across multi-site heterogeneous data sets while maintaining the confidentiality of the data. One of the possible solutions is to employ federated learning (FL), which enables the collaborative training of a DL model in a decentralized manner, demanding the involved sites to share only the characteristics of the models without transferring their sensitive medical data. The paper presents a DL-based magnetic resonance (MR) data translation in a FL way. We introduce a new aggregation strategy called FedBAdam that couples two state-of-the-art methods with complementary strengths by incorporating momentum in the aggregation scheme and skipping the batch normalization layers. The work comprehensively validates 10 FL-based strategies for an image-to-image multi-contrast MR translation, considering healthy and tumorous brain scans from five different institutions. Our study has revealed that the FedBAdam shows superior results in terms of mean squared error and structural similarity index over personalized methods, like the FedMRI, and standard FL-based aggregation techniques, such as the FedAvg or FedProx, considering multi-site multi-vendor heterogeneous environment. The FedBAdam has prevented the overfitting of the model and gradually reached the optimal model parameters, exhibiting no oscillations.
Collapse
Affiliation(s)
- Jan Fiszer
- Sano Centre for Computational Medicine, Kraków, Poland
- AGH University of Science and Technology, Kraków, Poland
| | | | - Maciej Malawski
- Sano Centre for Computational Medicine, Kraków, Poland
- AGH University of Science and Technology, Kraków, Poland
| | - Tomasz Pieciak
- Laboratorio de Procesado de Imagen (LPI), ETSI Telecomunicación, Universidad de Valladolid, Valladolid, Spain
| |
Collapse
|
7
|
Ma P, Chen Z, Huang Y, Zhao M, Li W, Li H, Cao D, Jiang Y, Zhou T, Cai J, Ren G. Motion and anatomy dual aware lung ventilation imaging by integrating Jacobian map and average CT image using dual path fusion network. Med Phys 2025; 52:246-256. [PMID: 39432032 PMCID: PMC11700001 DOI: 10.1002/mp.17466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Revised: 08/30/2024] [Accepted: 09/27/2024] [Indexed: 10/22/2024] Open
Abstract
BACKGROUND Deep learning-based computed tomography (CT) ventilation imaging (CTVI) is a promising technique for guiding functional lung avoidance radiotherapy (FLART). However, conventional approaches, which rely on anatomical CT data, may overlook important ventilation features due to the lack of motion data integration. PURPOSE This study aims to develop a novel dual-aware CTVI method that integrates both anatomical information from CT images and motional information from Jacobian maps to generate more accurate ventilation images for FLART. METHODS A dataset of 66 patients with four-dimensional CT (4DCT) images and reference ventilation images (RefVI) was utilized to develop the dual-path fusion network (DPFN) for synthesizing ventilation images (CTVIDual). The DPFN model was specifically designed to integrate motion data from 4DCT-generated Jacobian maps with anatomical data from average 4DCT images. The DPFN utilized two specialized feature extraction pathways, along with encoders and decoders, designed to handle both 3D average CT images and Jacobian map data. This dual-processing approach enabled the comprehensive extraction of lung ventilation-related features. The performance of DPFN was assessed by comparing CTVIDual to RefVI using various metrics, including Spearman's correlation coefficients (R), Dice similarity coefficients of high-functional region (DSCh), and low-functional region (DSCl). Additionally, CTVIDual was benchmarked against other CTVI methods, including a dual-phase CT-based deep learning method (CTVIDLCT), a radiomics-based method (CTVIFM), a super voxel-based method (CTVISVD), a Unet-based method (CTVIUnet), and two deformable registration-based methods (CTVIJac and CTVIHU). RESULTS In the test group, the mean R between CTVIDual and RefVI was 0.70, significantly outperforming CTVIDLCT (0.68), CTVIFM (0.58), CTVISVD (0.62), and CTVIUnet (0.66), with p < 0.05. Furthermore, the DSCh and DSCl values of CTVIDual were 0.64 and 0.80, respectively, outperforming CTVISVD (0.63; 0.73) and CTVIUnet (0.62; 0.77). The performance of CTVIDual was also significantly better than that of CTVIJac and CTVIHU. CONCLUSIONS A novel dual-aware CTVI model that integrates anatomical and motion information was developed to synthesize lung ventilation images. It was shown that the accuracy of lung ventilation estimation could be significantly enhanced by incorporating motional information, particularly in patients with tumor-induced blockages. This approach has the potential to improve the accuracy of CTVI, enabling more effective FLART.
Collapse
Affiliation(s)
- Pei Ma
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
| | - Zhi Chen
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
| | - Yu‐Hua Huang
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
| | - Mayang Zhao
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
| | - Wen Li
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
| | - Haojiang Li
- Department of RadiologyState Key Laboratory of Oncology in South ChinaCollaborative Innovation Centre for Cancer MedicineGuangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and TherapySun Yat‐sen University Cancer CentreGuangzhouPeople's Republic of China
| | - Di Cao
- Department of RadiologyState Key Laboratory of Oncology in South ChinaCollaborative Innovation Centre for Cancer MedicineGuangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and TherapySun Yat‐sen University Cancer CentreGuangzhouPeople's Republic of China
| | - Yi‐Quan Jiang
- Department of Minimally Invasive Interventional TherapyState Key Laboratory of Oncology in South ChinaGuangdong Provincial Clinical Research Center for CancerSun Yat‐sen University Cancer CenterGuangzhouPeople's Republic of China
| | - Ta Zhou
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
| | - Jing Cai
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
- The Hong Kong Polytechnic University Shenzhen Research InstituteShenzhenGuangdong ProvincePeople's Republic of China
| | - Ge Ren
- Department of Health Technology and InformaticsThe Hong Kong Polytechnic UniversityKowloonHong Kong SARHong Kong
- The Hong Kong Polytechnic University Shenzhen Research InstituteShenzhenGuangdong ProvincePeople's Republic of China
| |
Collapse
|
8
|
Ou Z, Wang H, Zhang B, Liang H, Hu B, Ren L, Liu Y, Zhang Y, Dai C, Wu H, Li W, Li X. Early identification of stroke through deep learning with multi-modal human speech and movement data. Neural Regen Res 2025; 20:234-241. [PMID: 38767488 PMCID: PMC11246124 DOI: 10.4103/1673-5374.393103] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 10/09/2023] [Accepted: 11/21/2023] [Indexed: 05/22/2024] Open
Abstract
JOURNAL/nrgr/04.03/01300535-202501000-00031/figure1/v/2024-05-14T021156Z/r/image-tiff Early identification and treatment of stroke can greatly improve patient outcomes and quality of life. Although clinical tests such as the Cincinnati Pre-hospital Stroke Scale (CPSS) and the Face Arm Speech Test (FAST) are commonly used for stroke screening, accurate administration is dependent on specialized training. In this study, we proposed a novel multimodal deep learning approach, based on the FAST, for assessing suspected stroke patients exhibiting symptoms such as limb weakness, facial paresis, and speech disorders in acute settings. We collected a dataset comprising videos and audio recordings of emergency room patients performing designated limb movements, facial expressions, and speech tests based on the FAST. We compared the constructed deep learning model, which was designed to process multi-modal datasets, with six prior models that achieved good action classification performance, including the I3D, SlowFast, X3D, TPN, TimeSformer, and MViT. We found that the findings of our deep learning model had a higher clinical value compared with the other approaches. Moreover, the multi-modal model outperformed its single-module variants, highlighting the benefit of utilizing multiple types of patient data, such as action videos and speech audio. These results indicate that a multi-modal deep learning model combined with the FAST could greatly improve the accuracy and sensitivity of early stroke identification of stroke, thus providing a practical and powerful tool for assessing stroke patients in an emergency clinical setting.
Collapse
Affiliation(s)
- Zijun Ou
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong Province, China
| | - Haitao Wang
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong Province, China
| | - Bin Zhang
- Department of Neurology, Guangdong Neuroscience Institute, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Haobang Liang
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong Province, China
| | - Bei Hu
- Department of Emergency Medicine, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Longlong Ren
- Department of Emergency Medicine, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Yanjuan Liu
- Department of Emergency Medicine, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Yuhu Zhang
- Department of Neurology, Guangdong Neuroscience Institute, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Chengbo Dai
- Department of Neurology, Guangdong Neuroscience Institute, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Hejun Wu
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong Province, China
| | - Weifeng Li
- Department of Emergency Medicine, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| | - Xin Li
- Department of Emergency Medicine, Guangdong Provincial People’s Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, Guangdong Province, China
| |
Collapse
|
9
|
Zhang Y, Peng C, Wang Q, Song D, Li K, Kevin Zhou S. Unified Multi-Modal Image Synthesis for Missing Modality Imputation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:4-18. [PMID: 38976465 DOI: 10.1109/tmi.2024.3424785] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/10/2024]
Abstract
Multi-modal medical images provide complementary soft-tissue characteristics that aid in the screening and diagnosis of diseases. However, limited scanning time, image corruption and various imaging protocols often result in incomplete multi-modal images, thus limiting the usage of multi-modal data for clinical purposes. To address this issue, in this paper, we propose a novel unified multi-modal image synthesis method for missing modality imputation. Our method overall takes a generative adversarial architecture, which aims to synthesize missing modalities from any combination of available ones with a single model. To this end, we specifically design a Commonality- and Discrepancy-Sensitive Encoder for the generator to exploit both modality-invariant and specific information contained in input modalities. The incorporation of both types of information facilitates the generation of images with consistent anatomy and realistic details of the desired distribution. Besides, we propose a Dynamic Feature Unification Module to integrate information from a varying number of available modalities, which enables the network to be robust to random missing modalities. The module performs both hard integration and soft integration, ensuring the effectiveness of feature combination while avoiding information loss. Verified on two public multi-modal magnetic resonance datasets, the proposed method is effective in handling various synthesis tasks and shows superior performance compared to previous methods.
Collapse
|
10
|
Fu X, Chen C, Chen Z, Yu J, Wang L. Radiogenomics based survival prediction of small-sample glioblastoma patients by multi-task DFFSP model. BIOMED ENG-BIOMED TE 2024; 69:623-633. [PMID: 39241784 DOI: 10.1515/bmt-2022-0221] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2022] [Accepted: 08/21/2024] [Indexed: 09/09/2024]
Abstract
In this paper, the multi-task dense-feature-fusion survival prediction (DFFSP) model is proposed to predict the three-year survival for glioblastoma (GBM) patients based on radiogenomics data. The contrast-enhanced T1-weighted (T1w) image, T2-weighted (T2w) image and copy number variation (CNV) is used as the input of the three branches of the DFFSP model. This model uses two image extraction modules consisting of residual blocks and one dense feature fusion module to make multi-scale fusion of T1w and T2w image features as backbone. Also, a gene feature extraction module is used to adaptively weight CNV fragments. Besides, a transfer learning module is introduced to solve the small sample problem and an image reconstruction module is adopted to make the model anatomy-aware under a multi-task framework. 256 sample pairs (T1w and corresponding T2w MRI slices) and 187 CNVs of 74 patients were used. The experimental results show that the proposed model can predict the three-year survival of GBM patients with the accuracy of 89.1 %, which is improved by 3.2 and 4.7 % compared with the model without genes and the model using last fusion strategy, respectively. This model could also classify the patients into high-risk and low-risk groups, which will effectively assist doctors in diagnosing GBM patients.
Collapse
Affiliation(s)
- Xue Fu
- Department of Biomedical Engineering, 47854 Nanjing University of Aeronautics and Astronautics , Nanjing, China
| | - Chunxiao Chen
- Department of Biomedical Engineering, 47854 Nanjing University of Aeronautics and Astronautics , Nanjing, China
| | - Zhiying Chen
- Department of Biomedical Engineering, 47854 Nanjing University of Aeronautics and Astronautics , Nanjing, China
| | - Jie Yu
- Department of Biomedical Engineering, 47854 Nanjing University of Aeronautics and Astronautics , Nanjing, China
| | - Liang Wang
- Department of Biomedical Engineering, 47854 Nanjing University of Aeronautics and Astronautics , Nanjing, China
| |
Collapse
|
11
|
Shen C, Li W, Chen H, Wang X, Zhu F, Li Y, Wang X, Jin B. Complementary information mutual learning for multimodality medical image segmentation. Neural Netw 2024; 180:106670. [PMID: 39299035 DOI: 10.1016/j.neunet.2024.106670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Revised: 07/10/2024] [Accepted: 08/26/2024] [Indexed: 09/22/2024]
Abstract
Radiologists must utilize medical images of multiple modalities for tumor segmentation and diagnosis due to the limitations of medical imaging technology and the diversity of tumor signals. This has led to the development of multimodal learning in medical image segmentation. However, the redundancy among modalities creates challenges for existing subtraction-based joint learning methods, such as misjudging the importance of modalities, ignoring specific modal information, and increasing cognitive load. These thorny issues ultimately decrease segmentation accuracy and increase the risk of overfitting. This paper presents the complementary information mutual learning (CIML) framework, which can mathematically model and address the negative impact of inter-modal redundant information. CIML adopts the idea of addition and removes inter-modal redundant information through inductive bias-driven task decomposition and message passing-based redundancy filtering. CIML first decomposes the multimodal segmentation task into multiple subtasks based on expert prior knowledge, minimizing the information dependence between modalities. Furthermore, CIML introduces a scheme in which each modality can extract information from other modalities additively through message passing. To achieve non-redundancy of extracted information, the redundant filtering is transformed into complementary information learning inspired by the variational information bottleneck. The complementary information learning procedure can be efficiently solved by variational inference and cross-modal spatial attention. Numerical results from the verification task and standard benchmarks indicate that CIML efficiently removes redundant information between modalities, outperforming SOTA methods regarding validation accuracy and segmentation effect. To emphasize, message-passing-based redundancy filtering allows neural network visualization techniques to visualize the knowledge relationship among different modalities, which reflects interpretability.
Collapse
Affiliation(s)
- Chuyun Shen
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Wenhao Li
- School of Data Science, The Chinese University of Hong Kong, Shenzhen Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen 518172, China.
| | - Haoqing Chen
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Xiaoling Wang
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Fengping Zhu
- Huashan Hospital Fudan University, Shanghai 200040, China.
| | - Yuxin Li
- Huashan Hospital Fudan University, Shanghai 200040, China.
| | - Xiangfeng Wang
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Bo Jin
- School of Software Engineering, Shanghai Research Institute for Intelligent Autonomous Systems, Tongji University, Shanghai 200092, China.
| |
Collapse
|
12
|
Zhang H, Ma Q, Qiu Y, Lai Z. ACGRHA-Net: Accelerated multi-contrast MR imaging with adjacency complementary graph assisted residual hybrid attention network. Neuroimage 2024; 303:120921. [PMID: 39521395 DOI: 10.1016/j.neuroimage.2024.120921] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2024] [Revised: 11/04/2024] [Accepted: 11/06/2024] [Indexed: 11/16/2024] Open
Abstract
Multi-contrast magnetic resonance (MR) imaging is an advanced technology used in medical diagnosis, but the long acquisition process can lead to patient discomfort and limit its broader application. Shortening acquisition time by undersampling k-space data introduces noticeable aliasing artifacts. To address this, we propose a method that reconstructs multi-contrast MR images from zero-filled data by utilizing a fully-sampled auxiliary contrast MR image as a prior to learn an adjacency complementary graph. This graph is then combined with a residual hybrid attention network, forming the adjacency complementary graph assisted residual hybrid attention network (ACGRHA-Net) for multi-contrast MR image reconstruction. Specifically, the optimal structural similarity is represented by a graph learned from the fully sampled auxiliary image, where the node features and adjacency matrices are designed to precisely capture structural information among different contrast images. This structural similarity enables effective fusion with the target image, improving the detail reconstruction. Additionally, a residual hybrid attention module is designed in parallel with the graph convolution network, allowing it to effectively capture key features and adaptively emphasize these important features in target contrast MR images. This strategy prioritizes crucial information while preserving shallow features, thereby achieving comprehensive feature fusion at deeper levels to enhance multi-contrast MR image reconstruction. Extensive experiments on the different datasets, using various sampling patterns and accelerated factors demonstrate that the proposed method outperforms the current state-of-the-art reconstruction methods.
Collapse
Affiliation(s)
- Haotian Zhang
- School of Ocean Information Engineering, Jimei University, Xiamen, China
| | - Qiaoyu Ma
- School of Ocean Information Engineering, Jimei University, Xiamen, China
| | - Yiran Qiu
- School of Ocean Information Engineering, Jimei University, Xiamen, China
| | - Zongying Lai
- School of Ocean Information Engineering, Jimei University, Xiamen, China.
| |
Collapse
|
13
|
Li W, Zhao D, Zeng G, Chen Z, Huang Z, Lam S, Cheung ALY, Ren G, Liu C, Liu X, Lee FKH, Au KH, Lee VHF, Xie Y, Qin W, Cai J, Li T. Evaluating Virtual Contrast-Enhanced Magnetic Resonance Imaging in Nasopharyngeal Carcinoma Radiation Therapy: A Retrospective Analysis for Primary Gross Tumor Delineation. Int J Radiat Oncol Biol Phys 2024; 120:1448-1457. [PMID: 38964419 DOI: 10.1016/j.ijrobp.2024.06.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Revised: 04/10/2024] [Accepted: 06/18/2024] [Indexed: 07/06/2024]
Abstract
PURPOSE To investigate the potential of virtual contrast-enhanced magnetic resonance imaging (VCE-MRI) for gross-tumor-volume (GTV) delineation of nasopharyngeal carcinoma (NPC) using multi-institutional data. METHODS AND MATERIALS This study retrospectively retrieved T1-weighted (T1w), T2-weighted (T2w) MRI, gadolinium-based contrast-enhanced MRI (CE-MRI), and planning computed tomography (CT) of 348 biopsy-proven NPC patients from 3 oncology centers. A multimodality-guided synergistic neural network (MMgSN-Net) was trained using 288 patients to leverage complementary features in T1w and T2w MRI for VCE-MRI synthesis, which was independently evaluated using 60 patients. Three board-certified radiation oncologists and 2 medical physicists participated in clinical evaluations in 3 aspects: image quality assessment of the synthetic VCE-MRI, VCE-MRI in assisting target volume delineation, and effectiveness of VCE-MRI-based contours in treatment planning. The image quality assessment includes distinguishability between VCE-MRI and CE-MRI, clarity of tumor-to-normal tissue interface, and veracity of contrast enhancement in tumor invasion risk areas. Primary tumor delineation and treatment planning were manually performed by radiation oncologists and medical physicists, respectively. RESULTS The mean accuracy to distinguish VCE-MRI from CE-MRI was 31.67%; no significant difference was observed in the clarity of tumor-to-normal tissue interface between VCE-MRI and CE-MRI; for the veracity of contrast enhancement in tumor invasion risk areas, an accuracy of 85.8% was obtained. The image quality assessment results suggest that the image quality of VCE-MRI is highly similar to real CE-MRI. The mean dosimetric difference of planning target volumes was less than 1 Gy. CONCLUSIONS The VCE-MRI is highly promising to replace the use of gadolinium-based CE-MRI in tumor delineation of NPC patients.
Collapse
Affiliation(s)
- Wen Li
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Dan Zhao
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Beijing Cancer Hospital & Institute, Peking University Cancer Hospital & Institute, Beijing, China
| | - Guangping Zeng
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Zhi Chen
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Zhou Huang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Beijing Cancer Hospital & Institute, Peking University Cancer Hospital & Institute, Beijing, China
| | - Saikit Lam
- Research Institute for Smart Aging, The Hong Kong Polytechnic University, Hong Kong SAR, China; Department of Biomedical Engineering, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Andy Lai-Yin Cheung
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Ge Ren
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Chenyang Liu
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Xi Liu
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Francis Kar-Ho Lee
- Department of Clinical Oncology, Queen Elizabeth Hospital, Hong Kong SAR, China
| | - Kwok-Hung Au
- Department of Clinical Oncology, Queen Elizabeth Hospital, Hong Kong SAR, China
| | - Victor Ho-Fun Lee
- Department of Clinical Oncology, The University of Hong Kong, Hong Kong SAR, China
| | - Yaoqin Xie
- Institute of Biomedical and Health Engineering, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, Guangdong Province, China
| | - Wenjian Qin
- Institute of Biomedical and Health Engineering, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, Guangdong Province, China
| | - Jing Cai
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China; Research Institute for Smart Aging, The Hong Kong Polytechnic University, Hong Kong SAR, China; The Hong Kong Polytechnic University Shenzhen Research Institute, Shenzhen, China.
| | - Tian Li
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, China.
| |
Collapse
|
14
|
Raymond C, Zhang D, Cabello J, Liu L, Moyaert P, Burneo JG, Dada MO, Hicks JW, Finger E, Soddu A, Andrade A, Jurkiewicz MT, Anazodo UC. SMART-PET: a Self-SiMilARiTy-aware generative adversarial framework for reconstructing low-count [18F]-FDG-PET brain imaging. FRONTIERS IN NUCLEAR MEDICINE (LAUSANNE, SWITZERLAND) 2024; 4:1469490. [PMID: 39628873 PMCID: PMC11611550 DOI: 10.3389/fnume.2024.1469490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 10/28/2024] [Indexed: 12/06/2024]
Abstract
Introduction In Positron Emission Tomography (PET) imaging, the use of tracers increases radioactive exposure for longitudinal evaluations and in radiosensitive populations such as pediatrics. However, reducing injected PET activity potentially leads to an unfavorable compromise between radiation exposure and image quality, causing lower signal-to-noise ratios and degraded images. Deep learning-based denoising approaches can be employed to recover low count PET image signals: nonetheless, most of these methods rely on structural or anatomic guidance from magnetic resonance imaging (MRI) and fails to effectively preserve global spatial features in denoised PET images, without impacting signal-to-noise ratios. Methods In this study, we developed a novel PET only deep learning framework, the Self-SiMilARiTy-Aware Generative Adversarial Framework (SMART), which leverages Generative Adversarial Networks (GANs) and a self-similarity-aware attention mechanism for denoising [18F]-fluorodeoxyglucose (18F-FDG) PET images. This study employs a combination of prospective and retrospective datasets in its design. In total, 114 subjects were included in the study, comprising 34 patients who underwent 18F-Fluorodeoxyglucose PET (FDG) PET imaging for drug-resistant epilepsy, 10 patients for frontotemporal dementia indications, and 70 healthy volunteers. To effectively denoise PET images without anatomical details from MRI, a self-similarity attention mechanism (SSAB) was devised. which learned the distinctive structural and pathological features. These SSAB-enhanced features were subsequently applied to the SMART GAN algorithm and trained to denoise the low-count PET images using the standard dose PET image acquired from each individual participant as reference. The trained GAN algorithm was evaluated using image quality measures including structural similarity index measure (SSIM), peak signal-to-noise ratio (PSNR), normalized root mean square (NRMSE), Fréchet inception distance (FID), signal-to-noise ratio (SNR), and contrast-to-noise ratio (CNR). Results In comparison to the standard-dose, SMART-PET had on average a SSIM of 0.984 ± 0.007, PSNR of 38.126 ± 2.631 dB, NRMSE of 0.091 ± 0.028, FID of 0.455 ± 0.065, SNR of 0.002 ± 0.001, and CNR of 0.011 ± 0.011. Regions of interest measurements obtained with datasets decimated down to 10% of the original counts, showed a deviation of less than 1.4% when compared to the ground-truth values. Discussion In general, SMART-PET shows promise in reducing noise in PET images and can synthesize diagnostic quality images with a 90% reduction in standard of care injected activity. These results make it a potential candidate for clinical applications in radiosensitive populations and for longitudinal neurological studies.
Collapse
Affiliation(s)
- Confidence Raymond
- Multimodal Imaging of Neurodegenerative Diseases (MiND) Lab, Department of Neurology and Neurosurgery, McGill University, Montreal, QC, Canada
- Department of Medical Biophysics, Western University, London, ON, Canada
| | - Dong Zhang
- Multimodal Imaging of Neurodegenerative Diseases (MiND) Lab, Department of Neurology and Neurosurgery, McGill University, Montreal, QC, Canada
- Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada
| | - Jorge Cabello
- Siemens Medical Solutions USA, Inc., Knoxville, TN, United States
| | - Linshan Liu
- Multimodal Imaging of Neurodegenerative Diseases (MiND) Lab, Department of Neurology and Neurosurgery, McGill University, Montreal, QC, Canada
| | - Paulien Moyaert
- Multimodal Imaging of Neurodegenerative Diseases (MiND) Lab, Department of Neurology and Neurosurgery, McGill University, Montreal, QC, Canada
- Department of Medical Imaging, Ghent University, Ghent, Belgium
| | - Jorge G. Burneo
- Clinical Neurological Sciences, Western University, London, ON, Canada
| | - Michael O. Dada
- Department of Physics, Federal University of Technology, Minna, Nigeria
| | - Justin W. Hicks
- Department of Medical Biophysics, Western University, London, ON, Canada
| | - Elizabeth Finger
- Clinical Neurological Sciences, Western University, London, ON, Canada
| | - Andrea Soddu
- Department of Physics and Astronomy, Western University, London, ON, Canada
| | - Andrea Andrade
- Department of Pediatrics, Western University, London, ON, Canada
| | - Michael T. Jurkiewicz
- Department of Medical Biophysics, Western University, London, ON, Canada
- Department of Medical Imaging, Western University, London, ON, Canada
| | - Udunna C. Anazodo
- Multimodal Imaging of Neurodegenerative Diseases (MiND) Lab, Department of Neurology and Neurosurgery, McGill University, Montreal, QC, Canada
- Department of Medical Biophysics, Western University, London, ON, Canada
- Montreal Neurological Institute, McGill University, Montreal, QC, Canada
| |
Collapse
|
15
|
Pozzi M, Noei S, Robbi E, Cima L, Moroni M, Munari E, Torresani E, Jurman G. Generating and evaluating synthetic data in digital pathology through diffusion models. Sci Rep 2024; 14:28435. [PMID: 39557989 PMCID: PMC11574254 DOI: 10.1038/s41598-024-79602-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Accepted: 11/11/2024] [Indexed: 11/20/2024] Open
Abstract
Synthetic data is becoming a valuable tool for computational pathologists, aiding in tasks like data augmentation and addressing data scarcity and privacy. However, its use necessitates careful planning and evaluation to prevent the creation of clinically irrelevant artifacts.This manuscript introduces a comprehensive pipeline for generating and evaluating synthetic pathology data using a diffusion model. The pipeline features a multifaceted evaluation strategy with an integrated explainability procedure, addressing two key aspects of synthetic data use in the medical domain.The evaluation of the generated data employs an ensemble-like approach. The first step includes assessing the similarity between real and synthetic data using established metrics. The second step involves evaluating the usability of the generated images in deep learning models accompanied with explainable AI methods. The final step entails verifying their histopathological realism through questionnaires answered by professional pathologists. We show that each of these evaluation steps are necessary as they provide complementary information on the generated data's quality.The pipeline is demonstrated on the public GTEx dataset of 650 Whole Slide Images (WSIs), including five different tissues. An equal number of tiles from each tissue are generated and their reliability is assessed using the proposed evaluation pipeline, yielding promising results.In summary, the proposed workflow offers a comprehensive solution for generative AI in digital pathology, potentially aiding the community in their transition towards digitalization and data-driven modeling.
Collapse
Affiliation(s)
- Matteo Pozzi
- Data Science for Health Unit, Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento, 38123, Italy
- Department for Computational and Integrative Biology, Università degli Studi di Trento, Via Sommarive, 9, Povo, Trento, 38123, Italy
| | - Shahryar Noei
- Data Science for Health Unit, Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento, 38123, Italy
| | - Erich Robbi
- Data Science for Health Unit, Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento, 38123, Italy
- Department of Information Engineering and Computer Science, Università degli Studi di Trento, Via Sommarive, 9, Povo, Trento, 38123, Italy
| | - Luca Cima
- Department of Diagnostic and Public Health, Section of Pathology, University and Hospital Trust of Verona, Verona, Italy
| | - Monica Moroni
- Data Science for Health Unit, Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento, 38123, Italy
| | - Enrico Munari
- Department of Diagnostic and Public Health, Section of Pathology, University and Hospital Trust of Verona, Verona, Italy
| | - Evelin Torresani
- Pathology Unit, Department of Laboratory Medicine, Santa Chiara Hospital, APSS, Trento, Italy
| | - Giuseppe Jurman
- Data Science for Health Unit, Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento, 38123, Italy.
| |
Collapse
|
16
|
Zhuang Y, Liu H, Fang W, Ma G, Sun S, Zhu Y, Zhang X, Ge C, Chen W, Long J, Song E. A 3D hierarchical cross-modality interaction network using transformers and convolutions for brain glioma segmentation in MR images. Med Phys 2024; 51:8371-8389. [PMID: 39137295 DOI: 10.1002/mp.17354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Revised: 06/20/2024] [Accepted: 08/02/2024] [Indexed: 08/15/2024] Open
Abstract
BACKGROUND Precise glioma segmentation from multi-parametric magnetic resonance (MR) images is essential for brain glioma diagnosis. However, due to the indistinct boundaries between tumor sub-regions and the heterogeneous appearances of gliomas in volumetric MR scans, designing a reliable and automated glioma segmentation method is still challenging. Although existing 3D Transformer-based or convolution-based segmentation networks have obtained promising results via multi-modal feature fusion strategies or contextual learning methods, they widely lack the capability of hierarchical interactions between different modalities and cannot effectively learn comprehensive feature representations related to all glioma sub-regions. PURPOSE To overcome these problems, in this paper, we propose a 3D hierarchical cross-modality interaction network (HCMINet) using Transformers and convolutions for accurate multi-modal glioma segmentation, which leverages an effective hierarchical cross-modality interaction strategy to sufficiently learn modality-specific and modality-shared knowledge correlated to glioma sub-region segmentation from multi-parametric MR images. METHODS In the HCMINet, we first design a hierarchical cross-modality interaction Transformer (HCMITrans) encoder to hierarchically encode and fuse heterogeneous multi-modal features by Transformer-based intra-modal embeddings and inter-modal interactions in multiple encoding stages, which effectively captures complex cross-modality correlations while modeling global contexts. Then, we collaborate an HCMITrans encoder with a modality-shared convolutional encoder to construct the dual-encoder architecture in the encoding stage, which can learn the abundant contextual information from global and local perspectives. Finally, in the decoding stage, we present a progressive hybrid context fusion (PHCF) decoder to progressively fuse local and global features extracted by the dual-encoder architecture, which utilizes the local-global context fusion (LGCF) module to efficiently alleviate the contextual discrepancy among the decoding features. RESULTS Extensive experiments are conducted on two public and competitive glioma benchmark datasets, including the BraTS2020 dataset with 494 patients and the BraTS2021 dataset with 1251 patients. Results show that our proposed method outperforms existing Transformer-based and CNN-based methods using other multi-modal fusion strategies in our experiments. Specifically, the proposed HCMINet achieves state-of-the-art mean DSC values of 85.33% and 91.09% on the BraTS2020 online validation dataset and the BraTS2021 local testing dataset, respectively. CONCLUSIONS Our proposed method can accurately and automatically segment glioma regions from multi-parametric MR images, which is beneficial for the quantitative analysis of brain gliomas and helpful for reducing the annotation burden of neuroradiologists.
Collapse
Affiliation(s)
- Yuzhou Zhuang
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Hong Liu
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Wei Fang
- Wuhan Zhongke Industrial Research Institute of Medical Science Co., Ltd, Wuhan, China
| | - Guangzhi Ma
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Sisi Sun
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Yunfeng Zhu
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Xu Zhang
- Wuhan United Imaging Healthcare Surgical Technology Co., Ltd, Wuhan, China
| | - Chuanbin Ge
- Wuhan United Imaging Healthcare Surgical Technology Co., Ltd, Wuhan, China
| | - Wenyang Chen
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Jiaosong Long
- School of Art and Design, Hubei University of Technology, Wuhan, China
| | - Enmin Song
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| |
Collapse
|
17
|
Zhang L. User emotion recognition and indoor space interaction design: a CNN model optimized by multimodal weighted networks. PeerJ Comput Sci 2024; 10:e2450. [PMID: 39650496 PMCID: PMC11623009 DOI: 10.7717/peerj-cs.2450] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2024] [Accepted: 10/04/2024] [Indexed: 12/11/2024]
Abstract
In interior interaction design, achieving intelligent user-interior interaction is contingent upon understanding the user's emotional responses. Precise identification of the user's visual emotions holds paramount importance. Current visual emotion recognition methods rely solely on singular features, predominantly facial expressions, resulting in inadequate coverage of visual characteristics and low recognition rates. This study introduces a deep learning-based multimodal weighting network model to address this challenge. The model initiates with a convolutional attention module, employing a self-attention mechanism within a convolutional neural network (CNN). As a result, the multimodal weighting network model is integrated to optimize weights during training. Finally, a weight network classifier is derived from these optimized weights to facilitate visual emotion recognition. Experimental outcomes reveal a 77.057% correctness rate and a 74.75% accuracy rate in visual emotion recognition. Comparative analysis against existing models demonstrates the superiority of the multimodal weight network model, showcasing its potential to enhance human-centric and intelligent indoor interaction design.
Collapse
Affiliation(s)
- Lingyu Zhang
- Space Lifestyle Design, Kookmin University, Seoul, Republic of South Korea
| |
Collapse
|
18
|
Zhang Y, Li C, Zhong L, Chen Z, Yang W, Wang X. DoseDiff: Distance-Aware Diffusion Model for Dose Prediction in Radiotherapy. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3621-3633. [PMID: 38564344 DOI: 10.1109/tmi.2024.3383423] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Treatment planning, which is a critical component of the radiotherapy workflow, is typically carried out by a medical physicist in a time-consuming trial-and-error manner. Previous studies have proposed knowledge-based or deep-learning-based methods for predicting dose distribution maps to assist medical physicists in improving the efficiency of treatment planning. However, these dose prediction methods usually fail to effectively utilize distance information between surrounding tissues and targets or organs-at-risk (OARs). Moreover, they are poor at maintaining the distribution characteristics of ray paths in the predicted dose distribution maps, resulting in a loss of valuable information. In this paper, we propose a distance-aware diffusion model (DoseDiff) for precise prediction of dose distribution. We define dose prediction as a sequence of denoising steps, wherein the predicted dose distribution map is generated with the conditions of the computed tomography (CT) image and signed distance maps (SDMs). The SDMs are obtained by distance transformation from the masks of targets or OARs, which provide the distance from each pixel in the image to the outline of the targets or OARs. We further propose a multi-encoder and multi-scale fusion network (MMFNet) that incorporates multi-scale and transformer-based fusion modules to enhance information fusion between the CT image and SDMs at the feature level. We evaluate our model on two in-house datasets and a public dataset, respectively. The results demonstrate that our DoseDiff method outperforms state-of-the-art dose prediction methods in terms of both quantitative performance and visual quality.
Collapse
|
19
|
Gourdeau D, Duchesne S, Archambault L. An hetero-modal deep learning framework for medical image synthesis applied to contrast and non-contrast MRI. Biomed Phys Eng Express 2024; 10:065015. [PMID: 39178886 DOI: 10.1088/2057-1976/ad72f9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2024] [Accepted: 08/23/2024] [Indexed: 08/26/2024]
Abstract
Some pathologies such as cancer and dementia require multiple imaging modalities to fully diagnose and assess the extent of the disease. Magnetic resonance imaging offers this kind of polyvalence, but examinations take time and can require contrast agent injection. The flexible synthesis of these imaging sequences based on the available ones for a given patient could help reduce scan times or circumvent the need for contrast agent injection. In this work, we propose a deep learning architecture that can perform the synthesis of all missing imaging sequences from any subset of available images. The network is trained adversarially, with the generator consisting of parallel 3D U-Net encoders and decoders that optimally combines their multi-resolution representations with a fusion operation learned by an attention network trained conjointly with the generator network. We compare our synthesis performance with 3D networks using other types of fusion and a comparable number of trainable parameters, such as the mean/variance fusion. In all synthesis scenarios except one, the synthesis performance of the network using attention-guided fusion was better than the other fusion schemes. We also inspect the encoded representations and the attention network outputs to gain insights into the synthesis process, and uncover desirable behaviors such as prioritization of specific modalities, flexible construction of the representation when important modalities are missing, and modalities being selected in regions where they carry sequence-specific information. This work suggests that a better construction of the latent representation space in hetero-modal networks can be achieved by using an attention network.
Collapse
Affiliation(s)
- Daniel Gourdeau
- CERVO Brain Research Center, Québec, Québec, Canada
- Physics Department, Université Laval, Québec, Québec, Canada
| | - Simon Duchesne
- CERVO Brain Research Center, Québec, Québec, Canada
- Department of Radiology and Nuclear Medicine, Université Laval, Québec, Québec, Canada
| | | |
Collapse
|
20
|
Guo X, Shi L, Chen X, Liu Q, Zhou B, Xie H, Liu YH, Palyo R, Miller EJ, Sinusas AJ, Staib L, Spottiswoode B, Liu C, Dvornek NC. TAI-GAN: A Temporally and Anatomically Informed Generative Adversarial Network for early-to-late frame conversion in dynamic cardiac PET inter-frame motion correction. Med Image Anal 2024; 96:103190. [PMID: 38820677 PMCID: PMC11180595 DOI: 10.1016/j.media.2024.103190] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Revised: 04/12/2024] [Accepted: 05/01/2024] [Indexed: 06/02/2024]
Abstract
Inter-frame motion in dynamic cardiac positron emission tomography (PET) using rubidium-82 (82Rb) myocardial perfusion imaging impacts myocardial blood flow (MBF) quantification and the diagnosis accuracy of coronary artery diseases. However, the high cross-frame distribution variation due to rapid tracer kinetics poses a considerable challenge for inter-frame motion correction, especially for early frames where intensity-based image registration techniques often fail. To address this issue, we propose a novel method called Temporally and Anatomically Informed Generative Adversarial Network (TAI-GAN) that utilizes an all-to-one mapping to convert early frames into those with tracer distribution similar to the last reference frame. The TAI-GAN consists of a feature-wise linear modulation layer that encodes channel-wise parameters generated from temporal information and rough cardiac segmentation masks with local shifts that serve as anatomical information. Our proposed method was evaluated on a clinical 82Rb PET dataset, and the results show that our TAI-GAN can produce converted early frames with high image quality, comparable to the real reference frames. After TAI-GAN conversion, the motion estimation accuracy and subsequent myocardial blood flow (MBF) quantification with both conventional and deep learning-based motion correction methods were improved compared to using the original frames. The code is available at https://github.com/gxq1998/TAI-GAN.
Collapse
Affiliation(s)
- Xueqi Guo
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA.
| | | | - Xiongchao Chen
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA
| | - Qiong Liu
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA
| | - Bo Zhou
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA
| | - Huidong Xie
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA
| | - Yi-Hwa Liu
- Department of Internal Medicine, Yale University, New Haven, CT, USA
| | | | - Edward J Miller
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Internal Medicine, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA
| | - Albert J Sinusas
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Internal Medicine, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA
| | - Lawrence Staib
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA
| | | | - Chi Liu
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA.
| | - Nicha C Dvornek
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA.
| |
Collapse
|
21
|
Shcherbakova YM, Lafranca PPG, Foppen W, van der Velden TA, Nievelstein RAJ, Castelein RM, Ito K, Seevinck PR, Schlosser TPC. A multipurpose, adolescent idiopathic scoliosis-specific, short MRI protocol: A feasibility study in volunteers. Eur J Radiol 2024; 177:111542. [PMID: 38861906 DOI: 10.1016/j.ejrad.2024.111542] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2024] [Accepted: 05/31/2024] [Indexed: 06/13/2024]
Abstract
INTRODUCTION Visualization of scoliosis typically requires ionizing radiation (radiography and CT) to visualize bony anatomy. MRI is often additionally performed to screen for neural axis abnormalities. We propose a 14-minutes radiation-free scoliosis-specific MRI protocol, which combines MRI and MRI-based synthetic CT images to visualize soft and osseous structures in one examination. We assess the ability of the protocol to visualize landmarks needed to detect 3D patho-anatomical changes, screen for neural axis abnormalities, and perform surgical planning and navigation. METHODS 18 adult volunteers were scanned on 1.5 T MR-scanner using 3D T2-weighted and synthetic CT sequences. A predefined checklist of relevant landmarks was used for the parameter assessment by three readers. Parameters included Cobb angles, rotation, torsion, segmental height, area and centroids of Nucleus Pulposus and Intervertebral Disc. Precision, reliability and agreement between the readers measurements were evaluated. RESULTS 91 % of Likert-based questions scored ≥ 4, indicating moderate to high confidence. Precision of 3D dot positioning was 1.0 mm. Precision of angle measurement was 0.6° (ICC 0.98). Precision of vertebral and IVD height measurements was 0.4 mm (ICC 0.99). Precision of area measurement for NP was 8 mm2 (ICC 0.55) and for IVD 18 mm2 (ICC 0.62) for IVD. Precision of centroid measurement for NP was 1.3 mm (ICC 0.88-0.92) and for IVD 1.1 mm (ICC 0.88-91). CONCLUSIONS The proposed MRI protocol with synthetic CT reconstructions, has high precision, reliability and agreement between the readers for multiple scoliosis-specific measurements. It can be used to study scoliosis etiopathogenesis and to assess 3D spinal morphology.
Collapse
Affiliation(s)
- Yulia M Shcherbakova
- Department of Radiology, Image Sciences Institute, UMC Utrecht, Utrecht, Netherlands.
| | | | - Wouter Foppen
- Department of Radiology & Nuclear Medicine, Division Imaging & Oncology, UMC Utrecht, Utrecht, Netherlands
| | - Tijl A van der Velden
- Department of Radiology, Image Sciences Institute, UMC Utrecht, Utrecht, Netherlands; MRIguidance B.V., Utrecht, Netherlands
| | - Rutger A J Nievelstein
- Department of Radiology & Nuclear Medicine, Division Imaging & Oncology, UMC Utrecht, Utrecht, Netherlands
| | - Rene M Castelein
- Department of Orthopaedic Surgery, UMC Utrecht, Utrecht, Netherlands
| | - Keita Ito
- Department of Orthopaedic Surgery, UMC Utrecht, Utrecht, Netherlands; Department of Biomedical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
| | - Peter R Seevinck
- Department of Radiology, Image Sciences Institute, UMC Utrecht, Utrecht, Netherlands; MRIguidance B.V., Utrecht, Netherlands
| | - Tom P C Schlosser
- Department of Orthopaedic Surgery, UMC Utrecht, Utrecht, Netherlands
| |
Collapse
|
22
|
Jiao C, Lao Y, Zhang W, Braunstein S, Salans M, Villanueva-Meyer JE, Hervey-Jumper SL, Yang B, Morin O, Valdes G, Fan Z, Shiroishi M, Zada G, Sheng K, Yang W. Multi-modal fusion and feature enhancement U-Net coupling with stem cell niches proximity estimation for voxel-wise GBM recurrence prediction . Phys Med Biol 2024; 69:10.1088/1361-6560/ad64b8. [PMID: 39019073 PMCID: PMC11308744 DOI: 10.1088/1361-6560/ad64b8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2024] [Accepted: 07/17/2024] [Indexed: 07/19/2024]
Abstract
Objective.We aim to develop a Multi-modal Fusion and Feature Enhancement U-Net (MFFE U-Net) coupling with stem cell niche proximity estimation to improve voxel-wise Glioblastoma (GBM) recurrence prediction.Approach.57 patients with pre- and post-surgery magnetic resonance (MR) scans were retrospectively solicited from 4 databases. Post-surgery MR scans included two months before the clinical diagnosis of recurrence and the day of the radiologicaly confirmed recurrence. The recurrences were manually annotated on the T1ce. The high-risk recurrence region was first determined. Then, a sparse multi-modal feature fusion U-Net was developed. The 50 patients from 3 databases were divided into 70% training, 10% validation, and 20% testing. 7 patients from the 4th institution were used as external testing with transfer learning. Model performance was evaluated by recall, precision, F1-score, and Hausdorff Distance at the 95% percentile (HD95). The proposed MFFE U-Net was compared to the support vector machine (SVM) model and two state-of-the-art neural networks. An ablation study was performed.Main results.The MFFE U-Net achieved a precision of 0.79 ± 0.08, a recall of 0.85 ± 0.11, and an F1-score of 0.82 ± 0.09. Statistically significant improvement was observed when comparing MFFE U-Net with proximity estimation couple SVM (SVMPE), mU-Net, and Deeplabv3. The HD95 was 2.75 ± 0.44 mm and 3.91 ± 0.83 mm for the 10 patients used in the model construction and 7 patients used for external testing, respectively. The ablation test showed that all five MR sequences contributed to the performance of the final model, with T1ce contributing the most. Convergence analysis, time efficiency analysis, and visualization of the intermediate results further discovered the characteristics of the proposed method.Significance. We present an advanced MFFE learning framework, MFFE U-Net, for effective voxel-wise GBM recurrence prediction. MFFE U-Net performs significantly better than the state-of-the-art networks and can potentially guide early RT intervention of the disease recurrence.
Collapse
Affiliation(s)
- Changzhe Jiao
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Yi Lao
- Department of Radiation Oncology, UC Los Angeles, Los Angeles, CA 90095
| | - Wenwen Zhang
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Steve Braunstein
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Mia Salans
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | | | | | - Bo Yang
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Olivier Morin
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Gilmer Valdes
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Zhaoyang Fan
- Department of Radiology, University of Southern California, Los Angeles, CA 90033
| | - Mark Shiroishi
- Department of Radiology, University of Southern California, Los Angeles, CA 90033
| | - Gabriel Zada
- Department of Neurosurgery, University of Southern California, Los Angeles, CA 90033
| | - Ke Sheng
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| | - Wensha Yang
- Department of Radiation Oncology, UC San Francisco, San Francisco, CA 94143
| |
Collapse
|
23
|
Gundogdu B, Medved M, Chatterjee A, Engelmann R, Rosado A, Lee G, Oren NC, Oto A, Karczmar GS. Self-supervised multicontrast super-resolution for diffusion-weighted prostate MRI. Magn Reson Med 2024; 92:319-331. [PMID: 38308149 PMCID: PMC11288973 DOI: 10.1002/mrm.30047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 01/19/2024] [Accepted: 01/23/2024] [Indexed: 02/04/2024]
Abstract
PURPOSE This study addresses the challenge of low resolution and signal-to-noise ratio (SNR) in diffusion-weighted images (DWI), which are pivotal for cancer detection. Traditional methods increase SNR at high b-values through multiple acquisitions, but this results in diminished image resolution due to motion-induced variations. Our research aims to enhance spatial resolution by exploiting the global structure within multicontrast DWI scans and millimetric motion between acquisitions. METHODS We introduce a novel approach employing a "Perturbation Network" to learn subvoxel-size motions between scans, trained jointly with an implicit neural representation (INR) network. INR encodes the DWI as a continuous volumetric function, treating voxel intensities of low-resolution acquisitions as discrete samples. By evaluating this function with a finer grid, our model predicts higher-resolution signal intensities for intermediate voxel locations. The Perturbation Network's motion-correction efficacy was validated through experiments on biological phantoms and in vivo prostate scans. RESULTS Quantitative analyses revealed significantly higher structural similarity measures of super-resolution images to ground truth high-resolution images compared to high-order interpolation (p< $$ < $$ 0.005). In blind qualitative experiments,96 . 1 % $$ 96.1\% $$ of super-resolution images were assessed to have superior diagnostic quality compared to interpolated images. CONCLUSION High-resolution details in DWI can be obtained without the need for high-resolution training data. One notable advantage of the proposed method is that it does not require a super-resolution training set. This is important in clinical practice because the proposed method can easily be adapted to images with different scanner settings or body parts, whereas the supervised methods do not offer such an option.
Collapse
Affiliation(s)
- Batuhan Gundogdu
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | - Milica Medved
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | | | - Roger Engelmann
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | - Avery Rosado
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | - Grace Lee
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | - Nisa C Oren
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | - Aytekin Oto
- Department of Radiology, University of Chicago, Chicago, Illinois, USA
| | | |
Collapse
|
24
|
Chaudhary MFA, Gerard SE, Christensen GE, Cooper CB, Schroeder JD, Hoffman EA, Reinhardt JM. LungViT: Ensembling Cascade of Texture Sensitive Hierarchical Vision Transformers for Cross-Volume Chest CT Image-to-Image Translation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2448-2465. [PMID: 38373126 PMCID: PMC11227912 DOI: 10.1109/tmi.2024.3367321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2024]
Abstract
Chest computed tomography (CT) at inspiration is often complemented by an expiratory CT to identify peripheral airways disease. Additionally, co-registered inspiratory-expiratory volumes can be used to derive various markers of lung function. Expiratory CT scans, however, may not be acquired due to dose or scan time considerations or may be inadequate due to motion or insufficient exhale; leading to a missed opportunity to evaluate underlying small airways disease. Here, we propose LungViT- a generative adversarial learning approach using hierarchical vision transformers for translating inspiratory CT intensities to corresponding expiratory CT intensities. LungViT addresses several limitations of the traditional generative models including slicewise discontinuities, limited size of generated volumes, and their inability to model texture transfer at volumetric level. We propose a shifted-window hierarchical vision transformer architecture with squeeze-and-excitation decoder blocks for modeling dependencies between features. We also propose a multiview texture similarity distance metric for texture and style transfer in 3D. To incorporate global information into the training process and refine the output of our model, we use ensemble cascading. LungViT is able to generate large 3D volumes of size 320×320×320 . We train and validate our model using a diverse cohort of 1500 subjects with varying disease severity. To assess model generalizability beyond the development set biases, we evaluate our model on an out-of-distribution external validation set of 200 subjects. Clinical validation on internal and external testing sets shows that synthetic volumes could be reliably adopted for deriving clinical endpoints of chronic obstructive pulmonary disease.
Collapse
|
25
|
Xie Q, Lin Y, Wang M, Wu Y. Synthesis of gadolinium-enhanced glioma images on multisequence magnetic resonance images using contrastive learning. Med Phys 2024; 51:4888-4897. [PMID: 38421681 DOI: 10.1002/mp.17004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2023] [Revised: 12/28/2023] [Accepted: 02/06/2024] [Indexed: 03/02/2024] Open
Abstract
BACKGROUND Gadolinium-based contrast agents are commonly used in brain magnetic resonance imaging (MRI), however, they cannot be used by patients with allergic reactions or poor renal function. For long-term follow-up patients, gadolinium deposition in the body can cause nephrogenic systemic fibrosis and other potential risks. PURPOSE Developing a new method of enhanced image synthesis based on the advantages of multisequence MRI has important clinical value for these patients. In this paper, an end-to-end synthesis model structure similarity index measure (SSIM)-based Dual Constrastive Learning with Attention (SDACL) based on contrastive learning is proposed to synthesize contrast-enhanced T1 (T1ce) using three unenhanced MRI images of T1, T2, and Flair in patients with glioma. METHODS The model uses the attention-dilation generator to enlarge the receptive field by expanding the residual blocks and to strengthen the feature representation and context learning of multisequence MRI. To enhance the detail and texture performance of the imaged tumor area, a comprehensive loss function combining patch-level contrast loss and structural similarity loss is created, which can effectively suppress noise and ensure the consistency of synthesized images and real images. RESULTS The normalized root-mean-square error (NRMSE), peak signal-to-noise ratio (PSNR), and SSIM of the model on the independent test set are 0.307 ± $\pm$ 0.12, 23.337 ± $\pm$ 3.21, and 0.881 ± $\pm$ 0.05, respectively. CONCLUSIONS Results show this method can be used for the multisequence synthesis of T1ce images, which can provide valuable information for clinical diagnosis.
Collapse
Affiliation(s)
- Qian Xie
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou, Henan, China
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou, Henan, China
| | - Yusong Lin
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou, Henan, China
- School of Cyber Science and Engineering, Zhengzhou University, Zhengzhou, Henan, China
- Hanwei IoT Institute, Zhengzhou University, Zhengzhou, Henan, China
| | - Meiyun Wang
- Collaborative Innovation Center for Internet Healthcare, Zhengzhou University, Zhengzhou, Henan, China
- Department of Medical Imaging, Henan Provincial People's Hospital, Zhengzhou, Henan, China
- Laboratory of Brain Science and Brain-Like Intelligence Technology Biomedical Research Institute Henan Academy of Science, Zhengzhou, Henan, China
| | - Yaping Wu
- Department of Medical Imaging, Henan Provincial People's Hospital, Zhengzhou, Henan, China
- Laboratory of Brain Science and Brain-Like Intelligence Technology Biomedical Research Institute Henan Academy of Science, Zhengzhou, Henan, China
| |
Collapse
|
26
|
Lu X, Liang X, Liu W, Miao X, Guan X. ReeGAN: MRI image edge-preserving synthesis based on GANs trained with misaligned data. Med Biol Eng Comput 2024; 62:1851-1868. [PMID: 38396277 DOI: 10.1007/s11517-024-03035-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Accepted: 01/27/2024] [Indexed: 02/25/2024]
Abstract
As a crucial medical examination technique, different modalities of magnetic resonance imaging (MRI) complement each other, offering multi-angle and multi-dimensional insights into the body's internal information. Therefore, research on MRI cross-modality conversion is of great significance, and many innovative techniques have been explored. However, most methods are trained on well-aligned data, and the impact of misaligned data has not received sufficient attention. Additionally, many methods focus on transforming the entire image and ignore crucial edge information. To address these challenges, we propose a generative adversarial network based on multi-feature fusion, which effectively preserves edge information while training on noisy data. Notably, we consider images with limited range random transformations as noisy labels and use an additional small auxiliary registration network to help the generator adapt to the noise distribution. Moreover, we inject auxiliary edge information to improve the quality of synthesized target modality images. Our goal is to find the best solution for cross-modality conversion. Comprehensive experiments and ablation studies demonstrate the effectiveness of the proposed method.
Collapse
Affiliation(s)
- Xiangjiang Lu
- Guangxi Key Lab of Multi-Source Information Mining & Security, School of Computer Science and Engineering & School of Software, Guangxi Normal University, Guilin, 541004, China.
| | - Xiaoshuang Liang
- Guangxi Key Lab of Multi-Source Information Mining & Security, School of Computer Science and Engineering & School of Software, Guangxi Normal University, Guilin, 541004, China
| | - Wenjing Liu
- Guangxi Key Lab of Multi-Source Information Mining & Security, School of Computer Science and Engineering & School of Software, Guangxi Normal University, Guilin, 541004, China
| | - Xiuxia Miao
- Guangxi Key Lab of Multi-Source Information Mining & Security, School of Computer Science and Engineering & School of Software, Guangxi Normal University, Guilin, 541004, China
| | - Xianglong Guan
- Guangxi Key Lab of Multi-Source Information Mining & Security, School of Computer Science and Engineering & School of Software, Guangxi Normal University, Guilin, 541004, China
| |
Collapse
|
27
|
Wang P, Zhang H, Zhu M, Jiang X, Qin J, Yuan Y. MGIML: Cancer Grading With Incomplete Radiology-Pathology Data via Memory Learning and Gradient Homogenization. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2113-2124. [PMID: 38231819 DOI: 10.1109/tmi.2024.3355142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/19/2024]
Abstract
Taking advantage of multi-modal radiology-pathology data with complementary clinical information for cancer grading is helpful for doctors to improve diagnosis efficiency and accuracy. However, radiology and pathology data have distinct acquisition difficulties and costs, which leads to incomplete-modality data being common in applications. In this work, we propose a Memory- and Gradient-guided Incomplete Modal-modal Learning (MGIML) framework for cancer grading with incomplete radiology-pathology data. Firstly, to remedy missing-modality information, we propose a Memory-driven Hetero-modality Complement (MH-Complete) scheme, which constructs modal-specific memory banks constrained by a coarse-grained memory boosting (CMB) loss to record generic radiology and pathology feature patterns, and develops a cross-modal memory reading strategy enhanced by a fine-grained memory consistency (FMC) loss to take missing-modality information from well-stored memories. Secondly, as gradient conflicts exist between missing-modality situations, we propose a Rotation-driven Gradient Homogenization (RG-Homogenize) scheme, which estimates instance-specific rotation matrices to smoothly change the feature-level gradient directions, and computes confidence-guided homogenization weights to dynamically balance gradient magnitudes. By simultaneously mitigating gradient direction and magnitude conflicts, this scheme well avoids the negative transfer and optimization imbalance problems. Extensive experiments on CPTAC-UCEC and CPTAC-PDA datasets show that the proposed MGIML framework performs favorably against state-of-the-art multi-modal methods on missing-modality situations.
Collapse
|
28
|
Li S, Liu B, Deng F, Xu Y, Zhou W. Image Synthesis of Hepatobiliary Phase using Contrast-Enhanced MRI and Diffusion Model. 2024 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI) 2024:1-5. [DOI: 10.1109/isbi56570.2024.10635567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/04/2025]
Affiliation(s)
- Shangxuan Li
- Guangzhou University of Chinese Medicine,School of Medical Information Engineering,China
| | - Baoer Liu
- Southern Medical University,Department of Medical Imaging Center, Nanfang Hospital,China
| | - Feilin Deng
- Guangzhou University of Chinese Medicine,School of Medical Information Engineering,China
| | - Yikai Xu
- Southern Medical University,Department of Medical Imaging Center, Nanfang Hospital,China
| | - Wu Zhou
- Guangzhou University of Chinese Medicine,School of Medical Information Engineering,China
| |
Collapse
|
29
|
Tian C, Zhang L. G2NPAN: GAN-guided nuance perceptual attention network for multimodal medical fusion image quality assessment. Front Neurosci 2024; 18:1415679. [PMID: 38803686 PMCID: PMC11128576 DOI: 10.3389/fnins.2024.1415679] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2024] [Accepted: 04/29/2024] [Indexed: 05/29/2024] Open
Abstract
Multimodal medical fusion images (MMFI) are formed by fusing medical images of two or more modalities with the aim of displaying as much valuable information as possible in a single image. However, due to the different strategies of various fusion algorithms, the quality of the generated fused images is uneven. Thus, an effective blind image quality assessment (BIQA) method is urgently required. The challenge of MMFI quality assessment is to enable the network to perceive the nuances between fused images of different qualities, and the key point for the success of BIQA is the availability of valid reference information. To this end, this work proposes a generative adversarial network (GAN) -guided nuance perceptual attention network (G2NPAN) to implement BIQA for MMFI. Specifically, we achieve the blind evaluation style via the design of a GAN and develop a Unique Feature Warehouse module to learn the effective features of fused images from the pixel level. The redesigned loss function guides the network to perceive the image quality. In the end, the class activation mapping supervised quality assessment network is employed to obtain the MMFI quality score. Extensive experiments and validation have been conducted in a database of medical fusion images, and the proposed method is superior to the state-of-the-art BIQA method.
Collapse
Affiliation(s)
| | - Lei Zhang
- School of Information Engineering (School of Big Data), Xuzhou University of Technology, Xuzhou, China
| |
Collapse
|
30
|
Zhao H, Cai H, Liu M. Transformer based multi-modal MRI fusion for prediction of post-menstrual age and neonatal brain development analysis. Med Image Anal 2024; 94:103140. [PMID: 38461655 DOI: 10.1016/j.media.2024.103140] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 11/23/2023] [Accepted: 03/05/2024] [Indexed: 03/12/2024]
Abstract
The brain development during the perinatal period is characterized by rapid changes in both structure and function, which have significant impact on the cognitive and behavioral abilities later in life. Accurate assessment of brain age is a crucial indicator for brain development maturity and can help predict the risk of neonatal pathology. However, evaluating neonatal brains using magnetic resonance imaging (MRI) is challenging due to its complexity, multi-dimension, and noise with subtle alterations. In this paper, we propose a multi-modal deep learning framework based on transformers for precise post-menstrual age (PMA) estimation and brain development analysis using T2-weighted structural MRI (T2-sMRI) and diffusion MRI (dMRI) data. First, we build a two-stream dense network to learn modality-specific features from T2-sMRI and dMRI of brain individually. Then, a transformer module based on self-attention mechanism integrates these features for PMA prediction and preterm/term classification. Finally, saliency maps on brain templates are used to enhance the interpretability of results. Our method is evaluated on the multi-modal MRI dataset of the developing Human Connectome Project (dHCP), which contains 592 neonates, including 478 term-born and 114 preterm-born subjects. The results demonstrate that our method achieves a 0.5-week mean absolute error (MAE) in PMA estimation for term-born subjects. Notably, preterm-born subjects exhibit delayed brain development, worsening with increasing prematurity. Our method also achieves 95% accuracy in classification of term-born and preterm-born subjects, revealing significant group differences.
Collapse
Affiliation(s)
- Haiyan Zhao
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Hongjie Cai
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Manhua Liu
- School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China; MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai, China.
| |
Collapse
|
31
|
Dalmaz O, Mirza MU, Elmas G, Ozbey M, Dar SUH, Ceyani E, Oguz KK, Avestimehr S, Çukur T. One model to unite them all: Personalized federated learning of multi-contrast MRI synthesis. Med Image Anal 2024; 94:103121. [PMID: 38402791 DOI: 10.1016/j.media.2024.103121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2023] [Revised: 02/20/2024] [Accepted: 02/21/2024] [Indexed: 02/27/2024]
Abstract
Curation of large, diverse MRI datasets via multi-institutional collaborations can help improve learning of generalizable synthesis models that reliably translate source- onto target-contrast images. To facilitate collaborations, federated learning (FL) adopts decentralized model training while mitigating privacy concerns by avoiding sharing of imaging data. However, conventional FL methods can be impaired by the inherent heterogeneity in the data distribution, with domain shifts evident within and across imaging sites. Here we introduce the first personalized FL method for MRI Synthesis (pFLSynth) that improves reliability against data heterogeneity via model specialization to individual sites and synthesis tasks (i.e., source-target contrasts). To do this, pFLSynth leverages an adversarial model equipped with novel personalization blocks that control the statistics of generated feature maps across the spatial/channel dimensions, given latent variables specific to sites and tasks. To further promote communication efficiency and site specialization, partial network aggregation is employed over later generator stages while earlier generator stages and the discriminator are trained locally. As such, pFLSynth enables multi-task training of multi-site synthesis models with high generalization performance across sites and tasks. Comprehensive experiments demonstrate the superior performance and reliability of pFLSynth in MRI synthesis against prior federated methods.
Collapse
Affiliation(s)
- Onat Dalmaz
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Muhammad U Mirza
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Gokberk Elmas
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Muzaffer Ozbey
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Salman U H Dar
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Emir Ceyani
- Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA 90089, USA
| | - Kader K Oguz
- Department of Radiology, University of California, Davis Medical Center, Sacramento, CA 95817, USA
| | - Salman Avestimehr
- Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA 90089, USA
| | - Tolga Çukur
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey; Neuroscience Program, Bilkent University, Ankara 06800, Turkey.
| |
Collapse
|
32
|
Zhang H, Liu J, Liu W, Chen H, Yu Z, Yuan Y, Wang P, Qin J. MHD-Net: Memory-Aware Hetero-Modal Distillation Network for Thymic Epithelial Tumor Typing With Missing Pathology Modality. IEEE J Biomed Health Inform 2024; 28:3003-3014. [PMID: 38470599 DOI: 10.1109/jbhi.2024.3376462] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/14/2024]
Abstract
Fusing multi-modal radiology and pathology data with complementary information can improve the accuracy of tumor typing. However, collecting pathology data is difficult since it is high-cost and sometimes only obtainable after the surgery, which limits the application of multi-modal methods in diagnosis. To address this problem, we propose comprehensively learning multi-modal radiology-pathology data in training, and only using uni-modal radiology data in testing. Concretely, a Memory-aware Hetero-modal Distillation Network (MHD-Net) is proposed, which can distill well-learned multi-modal knowledge with the assistance of memory from the teacher to the student. In the teacher, to tackle the challenge in hetero-modal feature fusion, we propose a novel spatial-differentiated hetero-modal fusion module (SHFM) that models spatial-specific tumor information correlations across modalities. As only radiology data is accessible to the student, we store pathology features in the proposed contrast-boosted typing memory module (CTMM) that achieves type-wise memory updating and stage-wise contrastive memory boosting to ensure the effectiveness and generalization of memory items. In the student, to improve the cross-modal distillation, we propose a multi-stage memory-aware distillation (MMD) scheme that reads memory-aware pathology features from CTMM to remedy missing modal-specific information. Furthermore, we construct a Radiology-Pathology Thymic Epithelial Tumor (RPTET) dataset containing paired CT and WSI images with annotations. Experiments on the RPTET and CPTAC-LUAD datasets demonstrate that MHD-Net significantly improves tumor typing and outperforms existing multi-modal methods on missing modality situations.
Collapse
|
33
|
Li B, Hu W, Feng CM, Li Y, Liu Z, Xu Y. Multi-Contrast Complementary Learning for Accelerated MR Imaging. IEEE J Biomed Health Inform 2024; 28:1436-1447. [PMID: 38157466 DOI: 10.1109/jbhi.2023.3348328] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2024]
Abstract
Thanks to its powerful ability to depict high-resolution anatomical information, magnetic resonance imaging (MRI) has become an essential non-invasive scanning technique in clinical practice. However, excessive acquisition time often leads to the degradation of image quality and psychological discomfort among subjects, hindering its further popularization. Besides reconstructing images from the undersampled protocol itself, multi-contrast MRI protocols bring promising solutions by leveraging additional morphological priors for the target modality. Nevertheless, previous multi-contrast techniques mainly adopt a simple fusion mechanism that inevitably ignores valuable knowledge. In this work, we propose a novel multi-contrast complementary information aggregation network named MCCA, aiming to exploit available complementary representations fully to reconstruct the undersampled modality. Specifically, a multi-scale feature fusion mechanism has been introduced to incorporate complementary-transferable knowledge into the target modality. Moreover, a hybrid convolution transformer block was developed to extract global-local context dependencies simultaneously, which combines the advantages of CNNs while maintaining the merits of Transformers. Compared to existing MRI reconstruction methods, the proposed method has demonstrated its superiority through extensive experiments on different datasets under different acceleration factors and undersampling patterns.
Collapse
|
34
|
Chen J, Huang G, Yuan X, Zhong G, Zheng Z, Pun CM, Zhu J, Huang Z. Quaternion Cross-Modality Spatial Learning for Multi-Modal Medical Image Segmentation. IEEE J Biomed Health Inform 2024; 28:1412-1423. [PMID: 38145537 DOI: 10.1109/jbhi.2023.3346529] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Recently, the Deep Neural Networks (DNNs) have had a large impact on imaging process including medical image segmentation, and the real-valued convolution of DNN has been extensively utilized in multi-modal medical image segmentation to accurately segment lesions via learning data information. However, the weighted summation operation in such convolution limits the ability to maintain spatial dependence that is crucial for identifying different lesion distributions. In this paper, we propose a novel Quaternion Cross-modality Spatial Learning (Q-CSL) which explores the spatial information while considering the linkage between multi-modal images. Specifically, we introduce to quaternion to represent data and coordinates that contain spatial information. Additionally, we propose Quaternion Spatial-association Convolution to learn the spatial information. Subsequently, the proposed De-level Quaternion Cross-modality Fusion (De-QCF) module excavates inner space features and fuses cross-modality spatial dependency. Our experimental results demonstrate that our approach compared to the competitive methods perform well with only 0.01061 M parameters and 9.95G FLOPs.
Collapse
|
35
|
Carass A, Greenman D, Dewey BE, Calabresi PA, Prince JL, Pham DL. Image harmonization improves consistency of intra-rater delineations of MS lesions in heterogeneous MRI. NEUROIMAGE. REPORTS 2024; 4:100195. [PMID: 38370461 PMCID: PMC10871705 DOI: 10.1016/j.ynirp.2024.100195] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/20/2024]
Abstract
Clinical magnetic resonance images (MRIs) lack a standard intensity scale due to differences in scanner hardware and the pulse sequences used to acquire the images. When MRIs are used for quantification, as in the evaluation of white matter lesions (WMLs) in multiple sclerosis, this lack of intensity standardization becomes a critical problem affecting both the staging and tracking of the disease and its treatment. This paper presents a study of harmonization on WML segmentation consistency, which is evaluated using an object detection classification scheme that incorporates manual delineations from both the original and harmonized MRIs. A cohort of ten people scanned on two different imaging platforms was studied. An expert rater, blinded to the image source, manually delineated WMLs on images from both scanners before and after harmonization. It was found that there is closer agreement in both global and per-lesion WML volume and spatial distribution after harmonization, demonstrating the importance of image harmonization prior to the creation of manual delineations. These results could lead to better truth models in both the development and evaluation of automated lesion segmentation algorithms.
Collapse
Affiliation(s)
- Aaron Carass
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218, USA
| | - Danielle Greenman
- Center for Neuroscience and Regenerative Medicine, The Henry M. Jackson Foundation for the Advancement of Military Medicine, Bethesda, MD 20817, USA
| | - Blake E. Dewey
- Department of Neurology, Johns Hopkins School of Medicine, Baltimore, MD 21287, USA
| | - Peter A. Calabresi
- Department of Neurology, Johns Hopkins School of Medicine, Baltimore, MD 21287, USA
| | - Jerry L. Prince
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218, USA
| | - Dzung L. Pham
- Department of Radiology, Uniformed Services University of the Health Sciences, Bethesda, MD 20814, USA
| |
Collapse
|
36
|
Yu J, Cai L, Chen C, Zou Y, Xiao Y, Fu X, Wang L, Yang X, Liu P, Lu Q, Sun X, Shao Q. A novel predict method for muscular invasion of bladder cancer based on 3D mp-MRI feature fusion. Phys Med Biol 2024; 69:055011. [PMID: 38306973 DOI: 10.1088/1361-6560/ad25c7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Accepted: 02/01/2024] [Indexed: 02/04/2024]
Abstract
Objective. To assist urologist and radiologist in the preoperative diagnosis of non-muscle-invasive bladder cancer (NMIBC) and muscle-invasive bladder cancer (MIBC), we proposed a combination models strategy (CMS) utilizing multiparametric magnetic resonance imaging.Approach. The CMS includes three components: image registration, image segmentation, and multisequence feature fusion. To ensure spatial structure consistency of T2-weighted imaging (T2WI), diffusion-weighted imaging (DWI), and dynamic contrast-enhanced imaging (DCE), a registration network based on patch sampling normalized mutual information was proposed to register DWI and DCE to T2WI. Moreover, to remove redundant information around the bladder, we employed a segmentation network to obtain the bladder and tumor regions from T2WI. Using the coordinate mapping from T2WI, we extracted these regions from DWI and DCE and integrated them into a three-branch dual-channel input. Finally, to fully fuse low-level and high-level features of T2WI, DWI, and DCE, we proposed a distributed multilayer fusion model for preoperative MIBC prediction with five-fold cross-validation.Main results. The study included 436 patients, of which 404 were for the internal cohort and 32 for external cohort. The MIBC was confirmed by pathological examination. In the internal cohort, the area under the curve, accuracy, sensitivity, and specificity achieved by our method were 0.928, 0.869, 0.753, and 0.929, respectively. For the urologist and radiologist, Vesical Imaging-Reporting and Data System score >3 was employed to determine MIBC. The urologist demonstrated an accuracy, sensitivity, and specificity of 0.842, 0.737, and 0.895, respectively, while the radiologist achieved 0.871, 0.803, and 0.906, respectively. In the external cohort, the accuracy of our method was 0.831, which was higher than that of the urologist (0.781) and the radiologist (0.813).Significance. Our proposed method achieved better diagnostic performance than urologist and was comparable to senior radiologist. These results indicate that CMS can effectively assist junior urologists and radiologists in diagnosing preoperative MIBC.
Collapse
Affiliation(s)
- Jie Yu
- Department of Biomedical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, People's Republic of China
| | - Lingkai Cai
- Department of Urology, the First Affiliated Hospital with Nanjing Medical University, Nanjing, People's Republic of China
| | - Chunxiao Chen
- Department of Biomedical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, People's Republic of China
| | - Yuan Zou
- Department of Biomedical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, People's Republic of China
| | - Yueyue Xiao
- Department of Biomedical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, People's Republic of China
| | - Xue Fu
- Department of Biomedical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, People's Republic of China
| | - Liang Wang
- Department of Biomedical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, People's Republic of China
| | - Xiao Yang
- Department of Urology, the First Affiliated Hospital with Nanjing Medical University, Nanjing, People's Republic of China
| | - Peikun Liu
- Department of Urology, the First Affiliated Hospital with Nanjing Medical University, Nanjing, People's Republic of China
| | - Qiang Lu
- Department of Urology, the First Affiliated Hospital with Nanjing Medical University, Nanjing, People's Republic of China
| | - Xueying Sun
- Department of Radiology, the First Affiliated Hospital with Nanjing Medical University, Nanjing, People's Republic of China
| | - Qiang Shao
- Department of Urology, the Affiliated Suzhou Hospital of Nanjing Medical University, People's Republic of China
| |
Collapse
|
37
|
Li W, Liu J, Wang S, Feng C. MTFN: multi-temporal feature fusing network with co-attention for DCE-MRI synthesis. BMC Med Imaging 2024; 24:47. [PMID: 38373915 PMCID: PMC10875895 DOI: 10.1186/s12880-024-01201-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Accepted: 01/15/2024] [Indexed: 02/21/2024] Open
Abstract
BACKGROUND Dynamic Contrast Enhanced Magnetic Resonance Imaging (DCE-MRI) plays an important role in the diagnosis and treatment of breast cancer. However, obtaining complete eight temporal images of DCE-MRI requires a long scanning time, which causes patients' discomfort in the scanning process. Therefore, to reduce the time, the multi temporal feature fusing neural network with Co-attention (MTFN) is proposed to generate the eighth temporal images of DCE-MRI, which enables the acquisition of DCE-MRI images without scanning. In order to reduce the time, multi-temporal feature fusion cooperative attention mechanism neural network (MTFN) is proposed to generate the eighth temporal images of DCE-MRI, which enables DCE-MRI image acquisition without scanning. METHODS In this paper, we propose multi temporal feature fusing neural network with Co-attention (MTFN) for DCE-MRI Synthesis, in which the Co-attention module can fully fuse the features of the first and third temporal image to obtain the hybrid features. The Co-attention explore long-range dependencies, not just relationships between pixels. Therefore, the hybrid features are more helpful to generate the eighth temporal images. RESULTS We conduct experiments on the private breast DCE-MRI dataset from hospitals and the multi modal Brain Tumor Segmentation Challenge2018 dataset (BraTs2018). Compared with existing methods, the experimental results of our method show the improvement and our method can generate more realistic images. In the meanwhile, we also use synthetic images to classify the molecular typing of breast cancer that the accuracy on the original eighth time-series images and the generated images are 89.53% and 92.46%, which have been improved by about 3%, and the classification results verify the practicability of the synthetic images. CONCLUSIONS The results of subjective evaluation and objective image quality evaluation indicators show the effectiveness of our method, which can obtain comprehensive and useful information. The improvement of classification accuracy proves that the images generated by our method are practical.
Collapse
Affiliation(s)
- Wei Li
- Key Laboratory of Intelligent Computing in Medical Image MIIC, Northeastern University, Shenyang, China
| | - Jiaye Liu
- School of Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Shanshan Wang
- School of Computer Science and Engineering, Northeastern University, Shenyang, China.
| | - Chaolu Feng
- Key Laboratory of Intelligent Computing in Medical Image MIIC, Northeastern University, Shenyang, China
| |
Collapse
|
38
|
Han L, Tan T, Zhang T, Huang Y, Wang X, Gao Y, Teuwen J, Mann R. Synthesis-based imaging-differentiation representation learning for multi-sequence 3D/4D MRI. Med Image Anal 2024; 92:103044. [PMID: 38043455 DOI: 10.1016/j.media.2023.103044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Revised: 10/14/2023] [Accepted: 11/24/2023] [Indexed: 12/05/2023]
Abstract
Multi-sequence MRIs can be necessary for reliable diagnosis in clinical practice due to the complimentary information within sequences. However, redundant information exists across sequences, which interferes with mining efficient representations by learning-based models. To handle various clinical scenarios, we propose a sequence-to-sequence generation framework (Seq2Seq) for imaging-differentiation representation learning. In this study, not only do we propose arbitrary 3D/4D sequence generation within one model to generate any specified target sequence, but also we are able to rank the importance of each sequence based on a new metric estimating the difficulty of a sequence being generated. Furthermore, we also exploit the generation inability of the model to extract regions that contain unique information for each sequence. We conduct extensive experiments using three datasets including a toy dataset of 20,000 simulated subjects, a brain MRI dataset of 1251 subjects, and a breast MRI dataset of 2101 subjects, to demonstrate that (1) top-ranking sequences can be used to replace complete sequences with non-inferior performance; (2) combining MRI with our imaging-differentiation map leads to better performance in clinical tasks such as glioblastoma MGMT promoter methylation status prediction and breast cancer pathological complete response status prediction. Our code is available at https://github.com/fiy2W/mri_seq2seq.
Collapse
Affiliation(s)
- Luyi Han
- Department of Radiology and Nuclear Medicine, Radboud University Medical Centre, Geert Grooteplein 10, 6525 GA, Nijmegen, The Netherlands; Department of Radiology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands
| | - Tao Tan
- Department of Radiology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands; Faculty of Applied Sciences, Macao Polytechnic University, 999078, Macao Special Administrative Region of China.
| | - Tianyu Zhang
- Department of Radiology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands; GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre, P. Debyelaan 25, 6202 AZ, Maastricht, The Netherlands
| | - Yunzhi Huang
- Institute for AI in Medicine, School of Automation, Nanjing University of Information Science and Technology, Nanjing, China
| | - Xin Wang
- Department of Radiology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands; GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre, P. Debyelaan 25, 6202 AZ, Maastricht, The Netherlands
| | - Yuan Gao
- Department of Radiology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands; GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre, P. Debyelaan 25, 6202 AZ, Maastricht, The Netherlands
| | - Jonas Teuwen
- Department of Radiation Oncology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands
| | - Ritse Mann
- Department of Radiology and Nuclear Medicine, Radboud University Medical Centre, Geert Grooteplein 10, 6525 GA, Nijmegen, The Netherlands; Department of Radiology, The Netherlands Cancer Institute, Plesmanlaan 121, 1066 CX, Amsterdam, The Netherlands
| |
Collapse
|
39
|
Dayarathna S, Islam KT, Uribe S, Yang G, Hayat M, Chen Z. Deep learning based synthesis of MRI, CT and PET: Review and analysis. Med Image Anal 2024; 92:103046. [PMID: 38052145 DOI: 10.1016/j.media.2023.103046] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 11/14/2023] [Accepted: 11/29/2023] [Indexed: 12/07/2023]
Abstract
Medical image synthesis represents a critical area of research in clinical decision-making, aiming to overcome the challenges associated with acquiring multiple image modalities for an accurate clinical workflow. This approach proves beneficial in estimating an image of a desired modality from a given source modality among the most common medical imaging contrasts, such as Computed Tomography (CT), Magnetic Resonance Imaging (MRI), and Positron Emission Tomography (PET). However, translating between two image modalities presents difficulties due to the complex and non-linear domain mappings. Deep learning-based generative modelling has exhibited superior performance in synthetic image contrast applications compared to conventional image synthesis methods. This survey comprehensively reviews deep learning-based medical imaging translation from 2018 to 2023 on pseudo-CT, synthetic MR, and synthetic PET. We provide an overview of synthetic contrasts in medical imaging and the most frequently employed deep learning networks for medical image synthesis. Additionally, we conduct a detailed analysis of each synthesis method, focusing on their diverse model designs based on input domains and network architectures. We also analyse novel network architectures, ranging from conventional CNNs to the recent Transformer and Diffusion models. This analysis includes comparing loss functions, available datasets and anatomical regions, and image quality assessments and performance in other downstream tasks. Finally, we discuss the challenges and identify solutions within the literature, suggesting possible future directions. We hope that the insights offered in this survey paper will serve as a valuable roadmap for researchers in the field of medical image synthesis.
Collapse
Affiliation(s)
- Sanuwani Dayarathna
- Department of Data Science and AI, Faculty of Information Technology, Monash University, Clayton VIC 3800, Australia.
| | | | - Sergio Uribe
- Department of Medical Imaging and Radiation Sciences, Faculty of Medicine, Monash University, Clayton VIC 3800, Australia
| | - Guang Yang
- Bioengineering Department and Imperial-X, Imperial College London, W12 7SL, United Kingdom
| | - Munawar Hayat
- Department of Data Science and AI, Faculty of Information Technology, Monash University, Clayton VIC 3800, Australia
| | - Zhaolin Chen
- Department of Data Science and AI, Faculty of Information Technology, Monash University, Clayton VIC 3800, Australia; Monash Biomedical Imaging, Clayton VIC 3800, Australia
| |
Collapse
|
40
|
Kumar S, Saber H, Charron O, Freeman L, Tamir JI. Correcting synthetic MRI contrast-weighted images using deep learning. Magn Reson Imaging 2024; 106:43-54. [PMID: 38092082 DOI: 10.1016/j.mri.2023.11.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 11/30/2023] [Accepted: 11/30/2023] [Indexed: 12/17/2023]
Abstract
Synthetic magnetic resonance imaging (MRI) offers a scanning paradigm where a fast multi-contrast sequence can be used to estimate underlying quantitative tissue parameter maps, which are then used to synthesize any desirable clinical contrast by retrospectively changing scan parameters in silico. Two benefits of this approach are the reduced exam time and the ability to generate arbitrary contrasts offline. However, synthetically generated contrasts are known to deviate from the contrast of experimental scans. The reason for contrast mismatch is the necessary exclusion of some unmodeled physical effects such as partial voluming, diffusion, flow, susceptibility, magnetization transfer, and more. The inclusion of these effects in signal encoding would improve the synthetic images, but would make the quantitative imaging protocol impractical due to long scan times. Therefore, in this work, we propose a novel deep learning approach that generates a multiplicative correction term to capture unmodeled effects and correct the synthetic contrast images to better match experimental contrasts for arbitrary scan parameters. The physics inspired deep learning model implicitly accounts for some unmodeled physical effects occurring during the scan. As a proof of principle, we validate our approach on synthesizing arbitrary inversion recovery fast spin-echo scans using a commercially available 2D multi-contrast sequence. We observe that the proposed correction visually and numerically reduces the mismatch with experimentally collected contrasts compared to conventional synthetic MRI. Finally, we show results of a preliminary reader study and find that the proposed method statistically significantly improves in contrast and SNR as compared to synthetic MR images.
Collapse
Affiliation(s)
- Sidharth Kumar
- Chandra Family Department of Electrical and Computer Engineering, The University of Texas at Austin, Austin 78712, TX, USA.
| | - Hamidreza Saber
- Dell Medical School Department of Neurology, The University of Texas at Austin, Austin 78712, TX, USA; Dell Medical School Department of Neurosurgery, The University of Texas at Austin, Austin 78712, TX, USA
| | - Odelin Charron
- Dell Medical School Department of Neurology, The University of Texas at Austin, Austin 78712, TX, USA
| | - Leorah Freeman
- Dell Medical School Department of Neurology, The University of Texas at Austin, Austin 78712, TX, USA; Dell Medical School Department of Diagnostic Medicine, The University of Texas at Austin, Austin 78712, TX, USA
| | - Jonathan I Tamir
- Chandra Family Department of Electrical and Computer Engineering, The University of Texas at Austin, Austin 78712, TX, USA; Dell Medical School Department of Diagnostic Medicine, The University of Texas at Austin, Austin 78712, TX, USA; Oden Institute for Computational Engineering and Sciences, The University of Texas at Austin, Austin 78712, TX, USA
| |
Collapse
|
41
|
Dai W, Liu R, Wu T, Wang M, Yin J, Liu J. Deeply Supervised Skin Lesions Diagnosis With Stage and Branch Attention. IEEE J Biomed Health Inform 2024; 28:719-729. [PMID: 37624725 DOI: 10.1109/jbhi.2023.3308697] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/27/2023]
Abstract
Accurate and unbiased examinations of skin lesions are critical for the early diagnosis and treatment of skin diseases. Visual features of skin lesions vary significantly because the images are collected from patients with different lesion colours and morphologies by using dissimilar imaging equipment. Recent studies have reported that ensembled convolutional neural networks (CNNs) are practical to classify the images for early diagnosis of skin disorders. However, the practical use of these ensembled CNNs is limited as these networks are heavyweight and inadequate for processing contextual information. Although lightweight networks (e.g., MobileNetV3 and EfficientNet) were developed to achieve parameter reduction for implementing deep neural networks on mobile devices, insufficient depth of feature representation restricts the performance. To address the existing limitations, we develop a new lite and effective neural network, namely HierAttn. The HierAttn applies a novel deep supervision strategy to learn the local and global features by using multi-stage and multi-branch attention mechanisms with only one training loss. The efficacy of HierAttn was evaluated by using the dermoscopy images dataset ISIC2019 and smartphone photos dataset PAD-UFES-20 (PAD2020). The experimental results show that HierAttn achieves the best accuracy and area under the curve (AUC) among the state-of-the-art lightweight networks.
Collapse
|
42
|
Wang Y, Luo Y, Zu C, Zhan B, Jiao Z, Wu X, Zhou J, Shen D, Zhou L. 3D multi-modality Transformer-GAN for high-quality PET reconstruction. Med Image Anal 2024; 91:102983. [PMID: 37926035 DOI: 10.1016/j.media.2023.102983] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2022] [Revised: 08/06/2023] [Accepted: 09/28/2023] [Indexed: 11/07/2023]
Abstract
Positron emission tomography (PET) scans can reveal abnormal metabolic activities of cells and provide favorable information for clinical patient diagnosis. Generally, standard-dose PET (SPET) images contain more diagnostic information than low-dose PET (LPET) images but higher-dose scans can also bring higher potential radiation risks. To reduce the radiation risk while acquiring high-quality PET images, in this paper, we propose a 3D multi-modality edge-aware Transformer-GAN for high-quality SPET reconstruction using the corresponding LPET images and T1 acquisitions from magnetic resonance imaging (T1-MRI). Specifically, to fully excavate the metabolic distributions in LPET and anatomical structural information in T1-MRI, we first use two separate CNN-based encoders to extract local spatial features from the two modalities, respectively, and design a multimodal feature integration module to effectively integrate the two kinds of features given the diverse contributions of features at different locations. Then, as CNNs can describe local spatial information well but have difficulty in modeling long-range dependencies in images, we further apply a Transformer-based encoder to extract global semantic information in the input images and use a CNN decoder to transform the encoded features into SPET images. Finally, a patch-based discriminator is applied to ensure the similarity of patch-wise data distribution between the reconstructed and real images. Considering the importance of edge information in anatomical structures for clinical disease diagnosis, besides voxel-level estimation error and adversarial loss, we also introduce an edge-aware loss to retain more edge detail information in the reconstructed SPET images. Experiments on the phantom dataset and clinical dataset validate that our proposed method can effectively reconstruct high-quality SPET images and outperform current state-of-the-art methods in terms of qualitative and quantitative metrics.
Collapse
Affiliation(s)
- Yan Wang
- School of Computer Science, Sichuan University, Chengdu, China
| | - Yanmei Luo
- School of Computer Science, Sichuan University, Chengdu, China
| | - Chen Zu
- Department of Risk Controlling Research, JD.COM, China
| | - Bo Zhan
- School of Computer Science, Sichuan University, Chengdu, China
| | - Zhengyang Jiao
- School of Computer Science, Sichuan University, Chengdu, China
| | - Xi Wu
- School of Computer Science, Chengdu University of Information Technology, China
| | - Jiliu Zhou
- School of Computer Science, Sichuan University, Chengdu, China
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China; Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China.
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Australia.
| |
Collapse
|
43
|
Li W, Lam S, Wang Y, Liu C, Li T, Kleesiek J, Cheung ALY, Sun Y, Lee FKH, Au KH, Lee VHF, Cai J. Model Generalizability Investigation for GFCE-MRI Synthesis in NPC Radiotherapy Using Multi-Institutional Patient-Based Data Normalization. IEEE J Biomed Health Inform 2024; 28:100-109. [PMID: 37624724 DOI: 10.1109/jbhi.2023.3308529] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/27/2023]
Abstract
Recently, deep learning has been demonstrated to be feasible in eliminating the use of gadoliniumbased contrast agents (GBCAs) through synthesizing gadolinium-free contrast-enhanced MRI (GFCE-MRI) from contrast-free MRI sequences, providing the community with an alternative to get rid of GBCAs-associated safety issues in patients. Nevertheless, generalizability assessment of the GFCE-MRI model has been largely challenged by the high inter-institutional heterogeneity of MRI data, on top of the scarcity of multi-institutional data itself. Although various data normalization methods have been adopted to address the heterogeneity issue, it has been limited to single-institutional investigation and there is no standard normalization approach presently. In this study, we aimed at investigating generalizability of GFCE-MRI model using data from seven institutions by manipulating heterogeneity of MRI data under five popular normalization approaches. Three state-of-the-art neural networks were applied to map from T1-weighted and T2-weighted MRI to contrast-enhanced MRI (CE-MRI) for GFCE-MRI synthesis in patients with nasopharyngeal carcinoma. MRI data from three institutions were used separately to generate three uni-institution models and jointly for a tri-institution model. The five normalization methods were applied to normalize the data of each model. MRI data from the remaining four institutions served as external cohorts for model generalizability assessment. Quality of GFCE-MRI was quantitatively evaluated against ground-truth CE-MRI using mean absolute error (MAE) and peak signal-to-noise ratio(PSNR). Results showed that performance of all uni-institution models remarkably dropped on the external cohorts. By contrast, model trained using multi-institutional data with Z-Score normalization yielded the best model generalizability improvement.
Collapse
|
44
|
Chen T, Hong R, Guo Y, Hao S, Hu B. MS²-GNN: Exploring GNN-Based Multimodal Fusion Network for Depression Detection. IEEE TRANSACTIONS ON CYBERNETICS 2023; 53:7749-7759. [PMID: 36194716 DOI: 10.1109/tcyb.2022.3197127] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Major depressive disorder (MDD) is one of the most common and severe mental illnesses, posing a huge burden on society and families. Recently, some multimodal methods have been proposed to learn a multimodal embedding for MDD detection and achieved promising performance. However, these methods ignore the heterogeneity/homogeneity among various modalities. Besides, earlier attempts ignore interclass separability and intraclass compactness. Inspired by the above observations, we propose a graph neural network (GNN)-based multimodal fusion strategy named modal-shared modal-specific GNN, which investigates the heterogeneity/homogeneity among various psychophysiological modalities as well as explores the potential relationship between subjects. Specifically, we develop a modal-shared and modal-specific GNN architecture to extract the inter/intramodal characteristics. Furthermore, a reconstruction network is employed to ensure fidelity within the individual modality. Moreover, we impose an attention mechanism on various embeddings to obtain a multimodal compact representation for the subsequent MDD detection task. We conduct extensive experiments on two public depression datasets and the favorable results demonstrate the effectiveness of the proposed algorithm.
Collapse
|
45
|
Wu S, Cao Y, Li X, Liu Q, Ye Y, Liu X, Zeng L, Tian M. Attention-guided multi-scale context aggregation network for multi-modal brain glioma segmentation. Med Phys 2023; 50:7629-7640. [PMID: 37151131 DOI: 10.1002/mp.16452] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2022] [Revised: 03/17/2023] [Accepted: 03/20/2023] [Indexed: 05/09/2023] Open
Abstract
BACKGROUND Accurate segmentation of brain glioma is a critical prerequisite for clinical diagnosis, surgical planning and treatment evaluation. In current clinical workflow, physicians typically perform delineation of brain tumor subregions slice-by-slice, which is more susceptible to variabilities in raters and also time-consuming. Besides, even though convolutional neural networks (CNNs) are driving progress, the performance of standard models still have some room for further improvement. PURPOSE To deal with these issues, this paper proposes an attention-guided multi-scale context aggregation network (AMCA-Net) for the accurate segmentation of brain glioma in the magnetic resonance imaging (MRI) images with multi-modalities. METHODS AMCA-Net extracts the multi-scale features from the MRI images and fuses the extracted discriminative features via a self-attention mechanism for brain glioma segmentation. The extraction is performed via a series of down-sampling, convolution layers, and the global context information guidance (GCIG) modules are developed to fuse the features extracted for contextual features. At the end of the down-sampling, a multi-scale fusion (MSF) module is designed to exploit and combine all the extracted multi-scale features. Each of the GCIG and MSF modules contain a channel attention (CA) module that can adaptively calibrate feature responses and emphasize the most relevant features. Finally, multiple predictions with different resolutions are fused through different weightings given by a multi-resolution adaptation (MRA) module instead of the use of averaging or max-pooling to improve the final segmentation results. RESULTS Datasets used in this paper are publicly accessible, that is, the Multimodal Brain Tumor Segmentation Challenges 2018 (BraTS2018) and 2019 (BraTS2019). BraTS2018 contains 285 patient cases and BraTS2019 contains 335 cases. Simulations show that the AMCA-Net has better or comparable performance against that of the other state-of-the-art models. In terms of the Dice score and Hausdorff 95 for the BraTS2018 dataset, 90.4% and 10.2 mm for the whole tumor region (WT), 83.9% and 7.4 mm for the tumor core region (TC), 80.2% and 4.3 mm for the enhancing tumor region (ET), whereas the Dice score and Hausdorff 95 for the BraTS2019 dataset, 91.0% and 10.7 mm for the WT, 84.2% and 8.4 mm for the TC, 80.1% and 4.8 mm for the ET. CONCLUSIONS The proposed AMCA-Net performs comparably well in comparison to several state-of-the-art neural net models in identifying the areas involving the peritumoral edema, enhancing tumor, and necrotic and non-enhancing tumor core of brain glioma, which has great potential for clinical practice. In future research, we will further explore the feasibility of applying AMCA-Net to other similar segmentation tasks.
Collapse
Affiliation(s)
- Shaozhi Wu
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Yunjian Cao
- Yangtze Delta Region Institute (Quzhou), University of Electronic Science and Technology of China, Quzhou, China
| | - Xinke Li
- West China School of Medicine, Sichuan University, Chengdu, China
| | - Qiyu Liu
- Radiology Department, Mianyang Central Hospital, Mianyang, China
| | - Yuyun Ye
- Department of Electrical and Computer Engineering, University of Tulsa, Tulsa, Oklahoma, USA
| | - Xingang Liu
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Liaoyuan Zeng
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Miao Tian
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, China
| |
Collapse
|
46
|
Ozbey M, Dalmaz O, Dar SUH, Bedel HA, Ozturk S, Gungor A, Cukur T. Unsupervised Medical Image Translation With Adversarial Diffusion Models. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3524-3539. [PMID: 37379177 DOI: 10.1109/tmi.2023.3290149] [Citation(s) in RCA: 81] [Impact Index Per Article: 40.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/30/2023]
Abstract
Imputation of missing images via source-to-target modality translation can improve diversity in medical imaging protocols. A pervasive approach for synthesizing target images involves one-shot mapping through generative adversarial networks (GAN). Yet, GAN models that implicitly characterize the image distribution can suffer from limited sample fidelity. Here, we propose a novel method based on adversarial diffusion modeling, SynDiff, for improved performance in medical image translation. To capture a direct correlate of the image distribution, SynDiff leverages a conditional diffusion process that progressively maps noise and source images onto the target image. For fast and accurate image sampling during inference, large diffusion steps are taken with adversarial projections in the reverse diffusion direction. To enable training on unpaired datasets, a cycle-consistent architecture is devised with coupled diffusive and non-diffusive modules that bilaterally translate between two modalities. Extensive assessments are reported on the utility of SynDiff against competing GAN and diffusion models in multi-contrast MRI and MRI-CT translation. Our demonstrations indicate that SynDiff offers quantitatively and qualitatively superior performance against competing baselines.
Collapse
|
47
|
Yang H, Sun J, Xu Z. Learning Unified Hyper-Network for Multi-Modal MR Image Synthesis and Tumor Segmentation With Missing Modalities. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3678-3689. [PMID: 37540616 DOI: 10.1109/tmi.2023.3301934] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/06/2023]
Abstract
Accurate segmentation of brain tumors is of critical importance in clinical assessment and treatment planning, which requires multiple MR modalities providing complementary information. However, due to practical limits, one or more modalities may be missing in real scenarios. To tackle this problem, existing methods need to train multiple networks or a unified but fixed network for various possible missing modality cases, which leads to high computational burdens or sub-optimal performance. In this paper, we propose a unified and adaptive multi-modal MR image synthesis method, and further apply it to tumor segmentation with missing modalities. Based on the decomposition of multi-modal MR images into common and modality-specific features, we design a shared hyper-encoder for embedding each available modality into the feature space, a graph-attention-based fusion block to aggregate the features of available modalities to the fused features, and a shared hyper-decoder for image reconstruction. We also propose an adversarial common feature constraint to enforce the fused features to be in a common space. As for missing modality segmentation, we first conduct the feature-level and image-level completion using our synthesis method and then segment the tumors based on the completed MR images together with the extracted common features. Moreover, we design a hypernet-based modulation module to adaptively utilize the real and synthetic modalities. Experimental results suggest that our method can not only synthesize reasonable multi-modal MR images, but also achieve state-of-the-art performance on brain tumor segmentation with missing modalities.
Collapse
|
48
|
Chen Q, Wang L, Xing Z, Wang L, Hu X, Wang R, Zhu YM. Deep wavelet scattering orthogonal fusion network for glioma IDH mutation status prediction. Comput Biol Med 2023; 166:107493. [PMID: 37774558 DOI: 10.1016/j.compbiomed.2023.107493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Revised: 06/26/2023] [Accepted: 09/15/2023] [Indexed: 10/01/2023]
Abstract
Accurately predicting the isocitrate dehydrogenase (IDH) mutation status of gliomas is greatly significant for formulating appropriate treatment plans and evaluating the prognoses of gliomas. Although existing studies can accurately predict the IDH mutation status of gliomas based on multimodal magnetic resonance (MR) images and machine learning methods, most of these methods cannot fully explore multimodal information and effectively predict IDH status for datasets acquired from multiple centers. To address this issue, a novel wavelet scattering (WS)-based orthogonal fusion network (WSOFNet) was proposed in this work to predict the IDH mutation status of gliomas from multiple centers. First, transformation-invariant features were extracted from multimodal MR images with a WS network, and then the multimodal WS features were used instead of the original images as the inputs of WSOFNet and were fully fused through an adaptive multimodal feature fusion module (AMF2M) and an orthogonal projection module (OPM). Finally, the fused features were input into a fully connected classifier to predict IDH mutation status. In addition, to achieve improved prediction accuracy, four auxiliary losses were also used in the feature extraction modules. The comparison results showed that the prediction area under the curve (AUC) of WSOFNet on a single-center dataset was 0.9966 and that on a multicenter dataset was approximately 0.9655, which was at least 3.9% higher than that of state-of-the-art methods. Moreover, the ablation experimental results also proved that the adaptive multimodal feature fusion strategy based on orthogonal projection could effectively improve the prediction performance of the model, especially for an external validation dataset.
Collapse
Affiliation(s)
- Qijian Chen
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
| | - Lihui Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China.
| | - Zhiyang Xing
- Department of Radiology, International Exemplary Cooperation Base of Precision Imaging for Diagnosis and Treatment, NHC Key Laboratory of Pulmonary Immune-related Diseases, Guizhou Provincial People's Hospital, Guiyang, 550002, China
| | - Li Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
| | - Xubin Hu
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
| | - Rongpin Wang
- Department of Radiology, International Exemplary Cooperation Base of Precision Imaging for Diagnosis and Treatment, NHC Key Laboratory of Pulmonary Immune-related Diseases, Guizhou Provincial People's Hospital, Guiyang, 550002, China
| | - Yue-Min Zhu
- University Lyon, INSA Lyon, CNRS, Inserm, IRP Metislab CREATIS UMR5220, U1206, Lyon 69621, France
| |
Collapse
|
49
|
Li Y, Zhou T, He K, Zhou Y, Shen D. Multi-Scale Transformer Network With Edge-Aware Pre-Training for Cross-Modality MR Image Synthesis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3395-3407. [PMID: 37339020 DOI: 10.1109/tmi.2023.3288001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/22/2023]
Abstract
Cross-modality magnetic resonance (MR) image synthesis can be used to generate missing modalities from given ones. Existing (supervised learning) methods often require a large number of paired multi-modal data to train an effective synthesis model. However, it is often challenging to obtain sufficient paired data for supervised training. In reality, we often have a small number of paired data while a large number of unpaired data. To take advantage of both paired and unpaired data, in this paper, we propose a Multi-scale Transformer Network (MT-Net) with edge-aware pre-training for cross-modality MR image synthesis. Specifically, an Edge-preserving Masked AutoEncoder (Edge-MAE) is first pre-trained in a self-supervised manner to simultaneously perform 1) image imputation for randomly masked patches in each image and 2) whole edge map estimation, which effectively learns both contextual and structural information. Besides, a novel patch-wise loss is proposed to enhance the performance of Edge-MAE by treating different masked patches differently according to the difficulties of their respective imputations. Based on this proposed pre-training, in the subsequent fine-tuning stage, a Dual-scale Selective Fusion (DSF) module is designed (in our MT-Net) to synthesize missing-modality images by integrating multi-scale features extracted from the encoder of the pre-trained Edge-MAE. Furthermore, this pre-trained encoder is also employed to extract high-level features from the synthesized image and corresponding ground-truth image, which are required to be similar (consistent) in the training. Experimental results show that our MT-Net achieves comparable performance to the competing methods even using 70% of all available paired data. Our code will be released at https://github.com/lyhkevin/MT-Net.
Collapse
|
50
|
Lyu J, Fu Y, Yang M, Xiong Y, Duan Q, Duan C, Wang X, Xing X, Zhang D, Lin J, Luo C, Ma X, Bian X, Hu J, Li C, Huang J, Zhang W, Zhang Y, Su S, Lou X. Generative Adversarial Network-based Noncontrast CT Angiography for Aorta and Carotid Arteries. Radiology 2023; 309:e230681. [PMID: 37962500 DOI: 10.1148/radiol.230681] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2023]
Abstract
Background Iodinated contrast agents (ICAs), which are widely used in CT angiography (CTA), may cause adverse effects in humans, and their use is time-consuming and costly. Purpose To develop an ICA-free deep learning imaging model for synthesizing CTA-like images and to assess quantitative and qualitative image quality as well as the diagnostic accuracy of synthetic CTA (Syn-CTA) images. Materials and Methods A generative adversarial network (GAN)-based CTA imaging model was trained, validated, and tested on retrospectively collected pairs of noncontrast CT and CTA images of the neck and abdomen from January 2017 to June 2022, and further validated on an external data set. Syn-CTA image quality was evaluated using quantitative metrics. In addition, two senior radiologists scored the visual quality on a three-point scale (3 = good) and determined the vascular diagnosis. The validity of Syn-CTA images was evaluated by comparing the visual quality scores and diagnostic accuracy of aortic and carotid artery disease between Syn-CTA and real CTA scans. Results CT scans from 1749 patients (median age, 60 years [IQR, 50-68 years]; 1057 male patients) were included in the internal data set: 1137 for training, 400 for validation, and 212 for testing. The external validation set comprised CT scans from 42 patients (median age, 67 years [IQR, 59-74 years]; 37 male patients). Syn-CTA images had high similarity to real CTA images (normalized mean absolute error, 0.011 and 0.013 for internal and external test set, respectively; peak signal-to-noise ratio, 32.07 dB and 31.58 dB; structural similarity, 0.919 and 0.906). The visual quality of Syn-CTA and real CTA images was comparable (internal test set, P = .35; external validation set, P > .99). Syn-CTA showed reasonable to good diagnostic accuracy for vascular diseases (internal test set: accuracy = 94%, macro F1 score = 91%; external validation set: accuracy = 86%, macro F1 score = 83%). Conclusion A GAN-based model that synthesizes neck and abdominal CTA-like images without the use of ICAs shows promise in vascular diagnosis compared with real CTA images. Clinical trial registration no. NCT05471869 © RSNA, 2023 Supplemental material is available for this article. See also the editorial by Zhang and Turkbey in this issue.
Collapse
Affiliation(s)
- Jinhao Lyu
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Ying Fu
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Mingliang Yang
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Yongqin Xiong
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Qi Duan
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Caohui Duan
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Xueyang Wang
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Xinbo Xing
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Dong Zhang
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Jiaji Lin
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Chuncai Luo
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Xiaoxiao Ma
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Xiangbing Bian
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Jianxing Hu
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Chenxi Li
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Jiayu Huang
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Wei Zhang
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Yue Zhang
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Sulian Su
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| | - Xin Lou
- From the Department of Radiology, Chinese PLA General Hospital, 28 Fuxing Rd, Haidian District, Beijing 100853, China (J. Lyu, Y.X., Q.D., C.D., X.W., X.X., D.Z., J. Lin, C. Luo, X.M., X.B., J. Hu, C. Li, J. Huang, X.L.); College of Medical Technology, School of Computer Science and Technology, Beijing Institute of Technology, Beijing, China (Y.F., M.Y., X.L.); Department of Radiology, Brain Hospital of Hunan Province, Changsha, China (W.Z.); Department of Radiology, Xiangyang No. 1 People's Hospital, Hubei University of Medicine, Xiangyang, China (Y.Z.); and Department of Radiology, Xiamen Humanity Hospital, Xiamen, China (S.S.)
| |
Collapse
|