1
|
Ji J, Lao Y, Huo L. Convolutional neural network application for supply-demand matching in Zhuang ethnic clothing image classification. Sci Rep 2024; 14:13348. [PMID: 38858436 PMCID: PMC11164973 DOI: 10.1038/s41598-024-64082-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 06/05/2024] [Indexed: 06/12/2024] Open
Abstract
This study aims to design a classification technique suitable for Zhuang ethnic clothing images by integrating the concept of supply-demand matching and convolutional neural networks. Firstly, addressing the complex structure and unique visual style of Zhuang ethnic clothing, this study proposes an image resolution model based on supply-demand matching and convolutional networks. By integrating visual style and label constraints, this model accurately extracts local features. Secondly, the model's effectiveness and resolution performance are analyzed through various performance metrics in experiments. The results indicate a significant improvement in detection accuracy at different annotation points. The model outperforms other comparative methods in pixel accuracy (90.5%), average precision (83.7%), average recall (80.1%), and average F1 score (81.2%). Next, this study introduces a clothing image classification algorithm based on key points and channel attention. Through key point detection and channel attention mechanisms, image features are optimized, enabling accurate classification and attribute prediction of Zhuang ethnic clothing. Experimental results demonstrate a notable enhancement in category classification and attribute prediction, with classification accuracy and recall exceeding 90% in top-k tasks, showcasing outstanding performance. In conclusion, this study provides innovative approaches and effective solutions for deep learning classification of Zhuang ethnic clothing images.
Collapse
Affiliation(s)
- Jingyi Ji
- School of Art, South China University of Technology, Guangzhou, 510641, People's Republic of China
| | - Yonghua Lao
- National Engineering Research Center for Tissue Restoration and Reconstruction, School of Materials Science and Engineering, South China University of Technology, Guangzhou, 510641, People's Republic of China
| | - Lei Huo
- School of Art, South China University of Technology, Guangzhou, 510641, People's Republic of China.
| |
Collapse
|
2
|
Ilesanmi AE, Ilesanmi TO, Ajayi BO. Reviewing 3D convolutional neural network approaches for medical image segmentation. Heliyon 2024; 10:e27398. [PMID: 38496891 PMCID: PMC10944240 DOI: 10.1016/j.heliyon.2024.e27398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2023] [Revised: 02/27/2024] [Accepted: 02/28/2024] [Indexed: 03/19/2024] Open
Abstract
Background Convolutional neural networks (CNNs) assume pivotal roles in aiding clinicians in diagnosis and treatment decisions. The rapid evolution of imaging technology has established three-dimensional (3D) CNNs as a formidable framework for delineating organs and anomalies in medical images. The prominence of 3D CNN frameworks is steadily growing within medical image segmentation and classification. Thus, our proposition entails a comprehensive review, encapsulating diverse 3D CNN algorithms for the segmentation of medical image anomalies and organs. Methods This study systematically presents an exhaustive review of recent 3D CNN methodologies. Rigorous screening of abstracts and titles were carried out to establish their relevance. Research papers disseminated across academic repositories were meticulously chosen, analyzed, and appraised against specific criteria. Insights into the realm of anomalies and organ segmentation were derived, encompassing details such as network architecture and achieved accuracies. Results This paper offers an all-encompassing analysis, unveiling the prevailing trends in 3D CNN segmentation. In-depth elucidations encompass essential insights, constraints, observations, and avenues for future exploration. A discerning examination indicates the preponderance of the encoder-decoder network in segmentation tasks. The encoder-decoder framework affords a coherent methodology for the segmentation of medical images. Conclusion The findings of this study are poised to find application in clinical diagnosis and therapeutic interventions. Despite inherent limitations, CNN algorithms showcase commendable accuracy levels, solidifying their potential in medical image segmentation and classification endeavors.
Collapse
Affiliation(s)
- Ademola E. Ilesanmi
- University of Pennsylvania, 3710 Hamilton Walk, 6th Floor, Philadelphia, PA, 19104, United States
| | | | - Babatunde O. Ajayi
- National Astronomical Research Institute of Thailand, Chiang Mai 50180, Thailand
| |
Collapse
|
3
|
Zhong S, Ren JX, Yu ZP, Peng YD, Yu CW, Deng D, Xie Y, He ZQ, Duan H, Wu B, Li H, Yang WZ, Bai Y, Sai K, Chen YS, Guo CC, Li DP, Cheng Y, Zhang XH, Mou YG. Predicting glioblastoma molecular subtypes and prognosis with a multimodal model integrating convolutional neural network, radiomics, and semantics. J Neurosurg 2023; 139:305-314. [PMID: 36461822 DOI: 10.3171/2022.10.jns22801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2022] [Accepted: 10/24/2022] [Indexed: 12/24/2022]
Abstract
OBJECTIVE The aim of this study was to build a convolutional neural network (CNN)-based prediction model of glioblastoma (GBM) molecular subtype diagnosis and prognosis with multimodal features. METHODS In total, 222 GBM patients were included in the training set from Sun Yat-sen University Cancer Center (SYSUCC) and 107 GBM patients were included in the validation set from SYSUCC, Xuanwu Hospital Capital Medical University, and the First Hospital of Jilin University. The multimodal model was trained with MR images (pre- and postcontrast T1-weighted images and T2-weighted images), corresponding MRI impression, and clinical patient information. First, the original images were segmented using the Multimodal Brain Tumor Image Segmentation Benchmark toolkit. Convolutional features were extracted using 3D residual deep neural network (ResNet50) and convolutional 3D (C3D). Radiomic features were extracted using pyradiomics. Report texts were converted to word embedding using word2vec. These three types of features were then integrated to train neural networks. Accuracy, precision, recall, and F1-score were used to evaluate the model performance. RESULTS The C3D-based model yielded the highest accuracy of 91.11% in the prediction of IDH1 mutation status. Importantly, the addition of semantics improved precision by 11.21% and recall in MGMT promoter methylation status prediction by 14.28%. The areas under the receiver operating characteristic curves of the C3D-based model in the IDH1, ATRX, MGMT, and 1-year prognosis groups were 0.976, 0.953, 0.955, and 0.976, respectively. In external validation, the C3D-based model showed significant improvement in accuracy in the IDH1, ATRX, MGMT, and 1-year prognosis groups, which were 88.30%, 76.67%, 85.71%, and 85.71%, respectively (compared with 3D ResNet50: 83.51%, 66.67%, 82.14%, and 70.79%, respectively). CONCLUSIONS The authors propose a novel multimodal model integrating C3D, radiomics, and semantics, which had a great performance in predicting IDH1, ATRX, and MGMT molecular subtypes and the 1-year prognosis of GBM.
Collapse
Affiliation(s)
- Sheng Zhong
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
- 2Department of Cancer Biology, Dana-Farber Cancer Institute, Boston, Massachusetts
- 3Department of Bioinformatics, Harvard Medical School, Boston, Massachusetts
| | - Jia-Xin Ren
- 4Department of Neurology, Stroke Center, The First Hospital of Jilin University, Changchun, China
| | - Ze-Peng Yu
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Yi-Da Peng
- 5College of Computer Science and Technology, Jilin University, Changchun, China
| | - Cheng-Wei Yu
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Davy Deng
- 2Department of Cancer Biology, Dana-Farber Cancer Institute, Boston, Massachusetts
| | - YangYiran Xie
- 6Vanderbilt University School of Medicine, Nashville, Tennessee
| | - Zhen-Qiang He
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Hao Duan
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Bo Wu
- Departments of7Orthopaedics
| | | | - Wen-Zhuo Yang
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Yang Bai
- 9Neurosurgery, The First Hospital of Jilin University, Changchun, China; and
| | - Ke Sai
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Yin-Sheng Chen
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Cheng-Cheng Guo
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - De-Pei Li
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Ye Cheng
- 10Department of Neurosurgery, The Xuanwu Hospital Capital Medical University, Beijing, China
| | - Xiang-Heng Zhang
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Yong-Gao Mou
- 1Department of Neurosurgery and Neuro-Oncology, State Key Laboratory of Oncology in South China, Collaborative Innovation Center for Cancer Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| |
Collapse
|
4
|
Zhang X, Liu Y, Guo S, Song Z. EG-Unet: Edge-Guided cascaded networks for automated frontal brain segmentation in MR images. Comput Biol Med 2023; 158:106891. [PMID: 37044048 DOI: 10.1016/j.compbiomed.2023.106891] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2022] [Revised: 03/07/2023] [Accepted: 04/01/2023] [Indexed: 04/05/2023]
Abstract
Accurate segmentation of frontal lobe areas on magnetic resonance imaging (MRI) can assist in diagnosing and managing idiopathic normal-pressure hydrocephalus. However, frontal lobe segmentation is challenging due to the complexity of the degree and shape of damage and the ambiguity of the boundaries of frontal lobe sites. Therefore, to extract the rich edge information and feature representation of the frontal lobe, this paper designs an edge guidance (EG) module to enhance the representation of edge features. Accordingly, an edge-guided cascade network framework (EG-Net) is proposed to segment frontal lobe parts automatically. Two-dimensional MRI slice images are fed into the edge generation and segmentation networks. First, the edge generation network extracts the edge information from the input image. Then, the edge information is sent to the EG module to generate an edge attention map for feature representation enhancement. Meanwhile, multi-scale attentional convolution (MSA) is utilized in the feature coding stage of the segmentation network to obtain feature responses from different perceptual fields in the coding stage and enrich the spatial context information. Besides, the feature fusion module is employed to selectively aggregate the multi-scale features in the coding stage with the edge features output by the EG module. Finally, the two components are fused, and a decoder recovers the spatial information to generate the final prediction results. An extensive quantitative comparison is performed on a publicly available brain MRI dataset (MICCAI 2012) to evaluate the effectiveness of the proposed algorithm. The experimental results indicate that the proposed method achieves an average DICE score of 95.77% compared to some advanced methods, which is 4.96% better than the classical U-Net. The results demonstrate the potential of the proposed EG-Net in improving the accuracy of frontal edge pixel classification through edge guidance.
Collapse
Affiliation(s)
- Xiufeng Zhang
- Mechanical and Electrical Engineering, Dalian Minzu University, Liaohe West Road 18, Dalian, China
| | - Yansong Liu
- Mechanical and Electrical Engineering, Dalian Minzu University, Liaohe West Road 18, Dalian, China.
| | - Shengjin Guo
- Mechanical and Electrical Engineering, Dalian Minzu University, Liaohe West Road 18, Dalian, China
| | - Zhao Song
- Shenzhen Hospital, Southern Medical University, Xinhu Road 1333, Shenzhen, China
| |
Collapse
|
5
|
Rath A, Mohanty DK, Mishra BSP, Bagal DK. A Bibliometric Review: Brain Tumor Magnetic Resonance Imagings Using Different Convolutional Neural Network Architectures. World Neurosurg 2023; 170:e681-e694. [PMID: 36442778 DOI: 10.1016/j.wneu.2022.11.091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2022] [Revised: 11/20/2022] [Accepted: 11/21/2022] [Indexed: 11/27/2022]
Abstract
BACKGROUND Numerous scientists and researchers have been developing advanced procedures and methods for diagnosing the kind and phase of a human tumor. Brain tumors, which are neoplastic and abnormal developments of brain cells, are one of the most prominent causes of death. Brain tumors, also known as lesions or neoplasia, may be roughly classified as either primary or metastatic. Primary brain tumors arise from brain tissue and its surrounding environment. The recognition of brain tumors using magnetic resonance images via a deep learning technique such as convolutional neural network (CNN) has garnered significant academic interest over the last few decades. METHODS In this study, a detailed evaluation based on bibliometrics is considered in order to synthesize and organize the available academic literature and to identify current research trends and hotspots. We used bibliometric methodologies and a literature review for the CNN-based brain tumor to synthesize and evaluate prior studies. RESULTS For this bibliometric analysis, we applied the Visualization of Similarity Viewer program to classify the major publications, notable journals, financial sponsors, and affiliations. CONCLUSIONS In conclusion, we suggest that one of the next paths of study will be the incorporation of other databases to advance CNN-based brain tumor identification from magnetic resonance images. No drug dosages are applied in this work.
Collapse
Affiliation(s)
- Arati Rath
- School of Computer Engineering, KIIT Deemed to be University, Odisha, India.
| | | | | | | |
Collapse
|
6
|
SIP-UNet: Sequential Inputs Parallel UNet Architecture for Segmentation of Brain Tissues from Magnetic Resonance Images. MATHEMATICS 2022. [DOI: 10.3390/math10152755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Proper analysis of changes in brain structure can lead to a more accurate diagnosis of specific brain disorders. The accuracy of segmentation is crucial for quantifying changes in brain structure. In recent studies, UNet-based architectures have outperformed other deep learning architectures in biomedical image segmentation. However, improving segmentation accuracy is challenging due to the low resolution of medical images and insufficient data. In this study, we present a novel architecture that combines three parallel UNets using a residual network. This architecture improves upon the baseline methods in three ways. First, instead of using a single image as input, we use three consecutive images. This gives our model the freedom to learn from neighboring images as well. Additionally, the images are individually compressed and decompressed using three different UNets, which prevents the model from merging the features of the images. Finally, following the residual network architecture, the outputs of the UNets are combined in such a way that the features of the image corresponding to the output are enhanced by a skip connection. The proposed architecture performed better than using a single conventional UNet and other UNet variants.
Collapse
|
7
|
SM-SegNet: A Lightweight Squeeze M-SegNet for Tissue Segmentation in Brain MRI Scans. SENSORS 2022; 22:s22145148. [PMID: 35890829 PMCID: PMC9319649 DOI: 10.3390/s22145148] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/12/2022] [Revised: 07/02/2022] [Accepted: 07/05/2022] [Indexed: 11/17/2022]
Abstract
In this paper, we propose a novel squeeze M-SegNet (SM-SegNet) architecture featuring a fire module to perform accurate as well as fast segmentation of the brain on magnetic resonance imaging (MRI) scans. The proposed model utilizes uniform input patches, combined-connections, long skip connections, and squeeze-expand convolutional layers from the fire module to segment brain MRI data. The proposed SM-SegNet architecture involves a multi-scale deep network on the encoder side and deep supervision on the decoder side, which uses combined-connections (skip connections and pooling indices) from the encoder to the decoder layer. The multi-scale side input layers support the deep network layers' extraction of discriminative feature information, and the decoder side provides deep supervision to reduce the gradient problem. By using combined-connections, extracted features can be transferred from the encoder to the decoder resulting in recovering spatial information, which makes the model converge faster. Long skip connections were used to stabilize the gradient updates in the network. Owing to the adoption of the fire module, the proposed model was significantly faster to train and offered a more efficient memory usage with 83% fewer parameters than previously developed methods, owing to the adoption of the fire module. The proposed method was evaluated using the open-access series of imaging studies (OASIS) and the internet brain segmentation registry (IBSR) datasets. The experimental results demonstrate that the proposed SM-SegNet architecture achieves segmentation accuracies of 95% for cerebrospinal fluid, 95% for gray matter, and 96% for white matter, which outperforms the existing methods in both subjective and objective metrics in brain MRI segmentation.
Collapse
|
8
|
Brain fMRI segmentation under emotion stimuli incorporating attention-based deep convolutional neural networks. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.108837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
9
|
Zhang C, Mou M, Zhou Y, Zhang W, Lian X, Shi S, Lu M, Sun H, Li F, Wang Y, Zeng Z, Li Z, Zhang B, Qiu Y, Zhu F, Gao J. Biological activities of drug inactive ingredients. Brief Bioinform 2022; 23:6582006. [PMID: 35524477 DOI: 10.1093/bib/bbac160] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2022] [Revised: 04/01/2022] [Accepted: 04/09/2022] [Indexed: 02/06/2023] Open
Abstract
In a drug formulation (DFM), the major components by mass are not Active Pharmaceutical Ingredient (API) but rather Drug Inactive Ingredients (DIGs). DIGs can reach much higher concentrations than that achieved by API, which raises great concerns about their clinical toxicities. Therefore, the biological activities of DIG on physiologically relevant target are widely demanded by both clinical investigation and pharmaceutical industry. However, such activity data are not available in any existing pharmaceutical knowledge base, and their potentials in predicting the DIG-target interaction have not been evaluated yet. In this study, the comprehensive assessment and analysis on the biological activities of DIGs were therefore conducted. First, the largest number of DIGs and DFMs were systematically curated and confirmed based on all drugs approved by US Food and Drug Administration. Second, comprehensive activities for both DIGs and DFMs were provided for the first time to pharmaceutical community. Third, the biological targets of each DIG and formulation were fully referenced to available databases that described their pharmaceutical/biological characteristics. Finally, a variety of popular artificial intelligence techniques were used to assess the predictive potential of DIGs' activity data, which was the first evaluation on the possibility to predict DIG's activity. As the activities of DIGs are critical for current pharmaceutical studies, this work is expected to have significant implications for the future practice of drug discovery and precision medicine.
Collapse
Affiliation(s)
- Chenyang Zhang
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Minjie Mou
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Ying Zhou
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China.,State Key Laboratory for Diagnosis and Treatment of Infectious Disease, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, Zhejiang Provincial Key Laboratory for Drug Clinical Research and Evaluation, The First Affiliated Hospital, Zhejiang University, 79 QingChun Road, Hangzhou, Zhejiang 310000, China
| | - Wei Zhang
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Xichen Lian
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Shuiyang Shi
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Mingkun Lu
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Huaicheng Sun
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Fengcheng Li
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Yunxia Wang
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China
| | - Zhenyu Zeng
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou 330110, China
| | - Zhaorong Li
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou 330110, China
| | - Bing Zhang
- Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou 330110, China
| | - Yunqing Qiu
- State Key Laboratory for Diagnosis and Treatment of Infectious Disease, Collaborative Innovation Center for Diagnosis and Treatment of Infectious Diseases, Zhejiang Provincial Key Laboratory for Drug Clinical Research and Evaluation, The First Affiliated Hospital, Zhejiang University, 79 QingChun Road, Hangzhou, Zhejiang 310000, China
| | - Feng Zhu
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China.,Innovation Institute for Artificial Intelligence in Medicine of Zhejiang University, Alibaba-Zhejiang University Joint Research Center of Future Digital Healthcare, Hangzhou 330110, China
| | - Jianqing Gao
- College of Pharmaceutical Sciences, The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang University, Hangzhou 310058, China.,Westlake Laboratory of Life Sciences and Biomedicine, Hangzhou, Zhejiang, China
| |
Collapse
|
10
|
Wu L, Hu S, Liu C. MR brain segmentation based on DE-ResUnet combining texture features and background knowledge. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103541] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
|
11
|
Multiscale and Hierarchical Feature-Aggregation Network for Segmenting Medical Images. SENSORS 2022; 22:s22093440. [PMID: 35591129 PMCID: PMC9104396 DOI: 10.3390/s22093440] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Revised: 04/24/2022] [Accepted: 04/26/2022] [Indexed: 11/17/2022]
Abstract
We propose an encoder-decoder architecture using wide and deep convolutional layers combined with different aggregation modules for the segmentation of medical images. Initially, we obtain a rich representation of features that span from low to high levels and from small to large scales by stacking multiple k × k kernels, where each k × k kernel operation is split into k × 1 and 1 × k convolutions. In addition, we introduce two feature-aggregation modules-multiscale feature aggregation (MFA) and hierarchical feature aggregation (HFA)-to better fuse information across end-to-end network layers. The MFA module progressively aggregates features and enriches feature representation, whereas the HFA module merges the features iteratively and hierarchically to learn richer combinations of the feature hierarchy. Furthermore, because residual connections are advantageous for assembling very deep networks, we employ an MFA-based long residual connections to avoid vanishing gradients along the aggregation paths. In addition, a guided block with multilevel convolution provides effective attention to the features that were copied from the encoder to the decoder to recover spatial information. Thus, the proposed method using feature-aggregation modules combined with a guided skip connection improves the segmentation accuracy, achieving a high similarity index for ground-truth segmentation maps. Experimental results indicate that the proposed model achieves a superior segmentation performance to that obtained by conventional methods for skin-lesion segmentation, with an average accuracy score of 0.97 on the ISIC-2018, PH2, and UFBA-UESC datasets.
Collapse
|