1
|
Zuo L, Wang Z, Wang Y. A multi-stage multi-modal learning algorithm with adaptive multimodal fusion for improving multi-label skin lesion classification. Artif Intell Med 2025; 162:103091. [PMID: 40015211 DOI: 10.1016/j.artmed.2025.103091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2023] [Revised: 09/10/2024] [Accepted: 02/14/2025] [Indexed: 03/01/2025]
Abstract
Skin cancer is frequently occurring and has become a major contributor to both cancer incidence and mortality. Accurate and timely diagnosis of skin cancer holds the potential to save lives. Deep learning-based methods have demonstrated significant advancements in the screening of skin cancers. However, most current approaches rely on a single modality input for diagnosis, thereby missing out on valuable complementary information that could enhance accuracy. Although some multimodal-based methods exist, they often lack adaptability and fail to fully leverage multimodal information. In this paper, we introduce a novel uncertainty-based hybrid fusion strategy for a multi-modal learning algorithm aimed at skin cancer diagnosis. Our approach specifically combines three different modalities: clinical images, dermoscopy images, and metadata, to make the final classification. For the fusion of two image modalities, we employ an intermediate fusion strategy that considers the similarity between clinical and dermoscopy images to extract features containing both complementary and correlated information. To capture the correlated information, we utilize cosine similarity, and we employ concatenation as the means for integrating complementary information. In the fusion of image and metadata modalities, we leverage uncertainty to obtain confident late fusion results, allowing our method to adaptively combine the information from different modalities. We conducted comprehensive experiments using a popular publicly available skin disease diagnosis dataset, and the results of these experiments demonstrate the effectiveness of our proposed method. Our proposed fusion algorithm could enhance the clinical applicability of automated skin lesion classification, offering a more robust and adaptive way to make automatic diagnoses with the help of uncertainty mechanism. Code is available at https://github.com/Zuo-Lihan/CosCatNet-Adaptive_Fusion_Algorithm.
Collapse
Affiliation(s)
- Lihan Zuo
- School of Computer and Artificial Intelligence, Southwest Jiaotong University, Chengdu 610000, PR China
| | - Zizhou Wang
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore 138632, Singapore
| | - Yan Wang
- Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), Singapore 138632, Singapore.
| |
Collapse
|
2
|
Alotaibi A, AlSaeed D. Skin Cancer Detection Using Transfer Learning and Deep Attention Mechanisms. Diagnostics (Basel) 2025; 15:99. [PMID: 39795627 PMCID: PMC11720014 DOI: 10.3390/diagnostics15010099] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 12/26/2024] [Accepted: 12/29/2024] [Indexed: 01/13/2025] Open
Abstract
Background/Objectives: Early and accurate diagnosis of skin cancer improves survival rates; however, dermatologists often struggle with lesion detection due to similar pigmentation. Deep learning and transfer learning models have shown promise in diagnosing skin cancers through image processing. Integrating attention mechanisms (AMs) with deep learning has further enhanced the accuracy of medical image classification. While significant progress has been made, further research is needed to improve the detection accuracy. Previous studies have not explored the integration of attention mechanisms with the pre-trained Xception transfer learning model for binary classification of skin cancer. This study aims to investigate the impact of various attention mechanisms on the Xception model's performance in detecting benign and malignant skin lesions. Methods: We conducted four experiments on the HAM10000 dataset. Three models integrated self-attention (SL), hard attention (HD), and soft attention (SF) mechanisms, while the fourth model used the standard Xception without attention mechanisms. Each mechanism analyzed features from the Xception model uniquely: self-attention examined the input relationships, hard-attention selected elements sparsely, and soft-attention distributed the focus probabilistically. Results: Integrating AMs into the Xception architecture effectively enhanced its performance. The accuracy of the Xception alone was 91.05%. With AMs, the accuracy increased to 94.11% using self-attention, 93.29% with soft attention, and 92.97% with hard attention. Moreover, the proposed models outperformed previous studies in terms of the recall metrics, which are crucial for medical investigations. Conclusions: These findings suggest that AMs can enhance performance in relation to complex medical imaging tasks, potentially supporting earlier diagnosis and improving treatment outcomes.
Collapse
Affiliation(s)
- Areej Alotaibi
- College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi Arabia;
| | | |
Collapse
|
3
|
Sun D, Li H, Wang Y, Li D, Xu D, Zhang Z. Artificial intelligence-based pathological application to predict regional lymph node metastasis in Papillary Thyroid Cancer. Curr Probl Cancer 2024; 53:101150. [PMID: 39342815 DOI: 10.1016/j.currproblcancer.2024.101150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2024] [Revised: 08/27/2024] [Accepted: 09/19/2024] [Indexed: 10/01/2024]
Abstract
In this study, a model for predicting lymph node metastasis in papillary thyroid cancer was trained using pathology images from the TCGA(The Cancer Genome Atlas) public dataset of papillary thyroid cancer, and a front-end inference model was trained using our center's dataset based on the concept of probabilistic propagation of nodes in graph neural networks. Effectively predicting whether a tumor will spread to regional lymph nodes using a single pathological image is the capacity of the model described above. This study demonstrates that regional lymph nodes in papillary thyroid cancer are a common and predictable occurrence, providing valuable ideas for future research. Now we publish the above research process and code for further study by other researchers, and we also make the above inference algorithm public at the URL: http:// thyroid-diseases-research.com/, with the hope that other researchers will validate it and provide us with ideas or datasets for further study.
Collapse
Affiliation(s)
- Dawei Sun
- The Affiliated Hospital of Qingdao University, PR China
| | - Huichao Li
- The Affiliated Hospital of Qingdao University, PR China
| | - Yaozong Wang
- Ningbo Huamei Hospital University of Chinese Academy of Sciences(Ningbo No.2 Hospital), PR China
| | - Dayuan Li
- Ningbo Institute of Material Technology and Engineering University of Chinese Academy of Sciences, PR China
| | - Di Xu
- Ningbo Institute of Material Technology and Engineering University of Chinese Academy of Sciences, PR China
| | - Zhoujing Zhang
- The Affiliated Hospital of Qingdao University, PR China; Ningbo Institute of Material Technology and Engineering University of Chinese Academy of Sciences, PR China; Ningbo Huamei Hospital University of Chinese Academy of Sciences(Ningbo No.2 Hospital), PR China.
| |
Collapse
|
4
|
Cheng D, Gai J, Yang B, Mao Y, Gao X, Zhang B, Jing W, Deng J, Zhao F, Mao N. GU-Net: Causal relationship-based generative medical image segmentation model. Heliyon 2024; 10:e37338. [PMID: 39309789 PMCID: PMC11415573 DOI: 10.1016/j.heliyon.2024.e37338] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2023] [Revised: 08/31/2024] [Accepted: 09/02/2024] [Indexed: 09/25/2024] Open
Abstract
Due to significant anatomical variations in medical images across different cases, medical image segmentation is a highly challenging task. Convolutional neural networks have shown faster and more accurate performance in medical image segmentation. However, existing networks for medical image segmentation mostly rely on independent training of the model using data samples and loss functions, lacking interactive training and feedback mechanisms. This leads to a relatively singular training approach for the models, and furthermore, some networks can only perform segmentation for specific diseases. In this paper, we propose a causal relationship-based generative medical image segmentation model named GU-Net. We integrate a counterfactual attention mechanism combined with CBAM into the decoder of U-Net as a generative network, and then combine it with a GAN network where the discriminator is used for backpropagation. This enables alternate optimization and training between the generative network and discriminator, enhancing the expressive and learning capabilities of the network model to output prediction segmentation results closer to the ground truth. Additionally, the interaction and transmission of information help the network model capture richer feature representations, extract more accurate features, reduce overfitting, and improve model stability and robustness through feedback mechanisms. Experimental results demonstrate that our proposed GU-Net network achieves better segmentation performance not only in cases with abundant data samples and relatively simple segmentation targets or high contrast between the target and background regions but also in scenarios with limited data samples and challenging segmentation tasks. Comparing with existing U-Net networks with attention mechanisms, GU-Net consistently improves Dice scores by 1.19%, 2.93%, 5.01%, and 5.50% on ISIC 2016, ISIC 2017, ISIC 2018, and Gland Segmentation datasets, respectively.
Collapse
Affiliation(s)
- Dapeng Cheng
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
- Shandong Co-Innovation Center of Future Intelligent Computing, Yantai, 264005, Shandong, China
| | - Jiale Gai
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
| | - Bo Yang
- Hainan College of Economics and Business School of Information Technology, Haikou, 571127, Hainan, China
| | - Yanyan Mao
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
| | - Xiaolian Gao
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
| | - Baosheng Zhang
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
| | - Wanting Jing
- School of Information and Electronic Engineering, Shandong Business and Technology University, Yantai, 264005, Shandong, China
| | - Jia Deng
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
| | - Feng Zhao
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, 264005, Shandong, China
- Shandong Co-Innovation Center of Future Intelligent Computing, Yantai, 264005, Shandong, China
| | - Ning Mao
- Department of Radiology, Yantai Yuhuangding Hospital, Yantai, 264000, Shandong, China
| |
Collapse
|
5
|
Sinha A, Kawahara J, Pakzad A, Abhishek K, Ruthven M, Ghorbel E, Kacem A, Aouada D, Hamarneh G. DermSynth3D: Synthesis of in-the-wild annotated dermatology images. Med Image Anal 2024; 95:103145. [PMID: 38615432 DOI: 10.1016/j.media.2024.103145] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2023] [Revised: 02/11/2024] [Accepted: 03/18/2024] [Indexed: 04/16/2024]
Abstract
In recent years, deep learning (DL) has shown great potential in the field of dermatological image analysis. However, existing datasets in this domain have significant limitations, including a small number of image samples, limited disease conditions, insufficient annotations, and non-standardized image acquisitions. To address these shortcomings, we propose a novel framework called DermSynth3D. DermSynth3D blends skin disease patterns onto 3D textured meshes of human subjects using a differentiable renderer and generates 2D images from various camera viewpoints under chosen lighting conditions in diverse background scenes. Our method adheres to top-down rules that constrain the blending and rendering process to create 2D images with skin conditions that mimic in-the-wild acquisitions, ensuring more meaningful results. The framework generates photo-realistic 2D dermatological images and the corresponding dense annotations for semantic segmentation of the skin, skin conditions, body parts, bounding boxes around lesions, depth maps, and other 3D scene parameters, such as camera position and lighting conditions. DermSynth3D allows for the creation of custom datasets for various dermatology tasks. We demonstrate the effectiveness of data generated using DermSynth3D by training DL models on synthetic data and evaluating them on various dermatology tasks using real 2D dermatological images. We make our code publicly available at https://github.com/sfu-mial/DermSynth3D.
Collapse
Affiliation(s)
- Ashish Sinha
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Jeremy Kawahara
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Arezou Pakzad
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Kumar Abhishek
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Matthieu Ruthven
- Computer Vision, Imaging & Machine Intelligence Research Group, Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, L-1855, Luxembourg
| | - Enjie Ghorbel
- Computer Vision, Imaging & Machine Intelligence Research Group, Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, L-1855, Luxembourg; Cristal Laboratory, National School of Computer Sciences, University of Manouba, 2010, Tunisia
| | - Anis Kacem
- Computer Vision, Imaging & Machine Intelligence Research Group, Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, L-1855, Luxembourg
| | - Djamila Aouada
- Computer Vision, Imaging & Machine Intelligence Research Group, Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, L-1855, Luxembourg
| | - Ghassan Hamarneh
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada.
| |
Collapse
|
6
|
Drummond MVMS, Borges JR, Ribeiro AMQ, Ximenes BAS. Dermoscopy as an auxiliary tool for the diagnosis of acral squamous diseases: palmoplantar psoriasis, tinea pedis/manuum and eczema. An Bras Dermatol 2024; 99:578-581. [PMID: 38658237 PMCID: PMC11221152 DOI: 10.1016/j.abd.2023.05.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Revised: 05/18/2023] [Accepted: 05/19/2023] [Indexed: 04/26/2024] Open
Affiliation(s)
| | - Jules Rimet Borges
- Service of Dermatology, Hospital das Clínicas de Goiânia, Goiânia, GO, Brazil
| | | | | |
Collapse
|
7
|
Ma P, Wang G, Li T, Zhao H, Li Y, Wang H. STCS-Net: a medical image segmentation network that fully utilizes multi-scale information. BIOMEDICAL OPTICS EXPRESS 2024; 15:2811-2831. [PMID: 38855673 PMCID: PMC11161382 DOI: 10.1364/boe.517737] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Revised: 03/09/2024] [Accepted: 03/19/2024] [Indexed: 06/11/2024]
Abstract
In recent years, significant progress has been made in the field of medical image segmentation through the application of deep learning and neural networks. Numerous studies have focused on optimizing encoders to extract more comprehensive key information. However, the importance of decoders in directly influencing the final output of images cannot be overstated. The ability of decoders to effectively leverage diverse information and further refine crucial details is of paramount importance. This paper proposes a medical image segmentation architecture named STCS-Net. The designed decoder in STCS-Net facilitates multi-scale filtering and correction of information from the encoder, thereby enhancing the accuracy of extracting vital features. Additionally, an information enhancement module is introduced in skip connections to highlight essential features and improve the inter-layer information interaction capabilities. Comprehensive evaluations on the ISIC2016, ISIC2018, and Lung datasets validate the superiority of STCS-Net across different scenarios. Experimental results demonstrate the outstanding performance of STCS-Net on all three datasets. Comparative experiments highlight the advantages of our proposed network in terms of accuracy and parameter efficiency. Ablation studies confirm the effectiveness of the introduced decoder and skip connection module. This research introduces a novel approach to the field of medical image segmentation, providing new perspectives and solutions for future developments in medical image processing and analysis.
Collapse
Affiliation(s)
- Pengchong Ma
- College of Electronic And Information Engineering, Hebei University, Hebei 071002, China
- Hebei Key Laboratory of Precise Imaging of Inflammation Related Tumors, Hebei 071000, China
| | - Guanglei Wang
- College of Electronic And Information Engineering, Hebei University, Hebei 071002, China
- Hebei Key Laboratory of Precise Imaging of Inflammation Related Tumors, Hebei 071000, China
| | - Tong Li
- College of Electronic And Information Engineering, Hebei University, Hebei 071002, China
- Hebei Key Laboratory of Precise Imaging of Inflammation Related Tumors, Hebei 071000, China
| | - Haiyang Zhao
- College of Electronic And Information Engineering, Hebei University, Hebei 071002, China
| | - Yan Li
- College of Electronic And Information Engineering, Hebei University, Hebei 071002, China
| | - Hongrui Wang
- College of Electronic And Information Engineering, Hebei University, Hebei 071002, China
| |
Collapse
|
8
|
Li G, Jin D, Zheng Y, Cui J, Gai W, Qi M. A generic plug & play diffusion-based denosing module for medical image segmentation. Neural Netw 2024; 172:106096. [PMID: 38194885 DOI: 10.1016/j.neunet.2024.106096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Revised: 12/11/2023] [Accepted: 12/31/2023] [Indexed: 01/11/2024]
Abstract
Medical image segmentation faces challenges because of the small sample size of the dataset and the fact that images often have noise and artifacts. In recent years, diffusion models have proven very effective in image generation and have been used widely in computer vision. This paper presents a new feature map denoising module (FMD) based on the diffusion model for feature refinement, which is plug-and-play, allowing flexible integration into popular used segmentation networks for seamless end-to-end training. We evaluate the performance of the FMD module on four models, UNet, UNeXt, TransUNet, and IB-TransUNet, by conducting experiments on four datasets. The experimental data analysis shows that adding the FMD module significantly positively impacts the model performance. Furthermore, especially for small lesion areas and minor organs, adding the FMD module allows users to obtain more accurate segmentation results than the original model.
Collapse
Affiliation(s)
- Guangju Li
- School of Information Science and Engineering, Shandong Normal University, Jinan, China
| | - Dehu Jin
- School of Information Science and Engineering, Shandong Normal University, Jinan, China
| | - Yuanjie Zheng
- School of Information Science and Engineering, Shandong Normal University, Jinan, China
| | - Jia Cui
- School of Design, South China University of Technology, Guangzhou, China
| | - Wei Gai
- School of Software, Shandong University, Jinan, Shandong, China
| | - Meng Qi
- School of Information Science and Engineering, Shandong Normal University, Jinan, China.
| |
Collapse
|
9
|
Khan MA, Muhammad K, Sharif M, Akram T, Kadry S. Intelligent fusion-assisted skin lesion localization and classification for smart healthcare. Neural Comput Appl 2024; 36:37-52. [DOI: 10.1007/s00521-021-06490-w] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2021] [Accepted: 08/30/2021] [Indexed: 12/28/2022]
|
10
|
Liu L, Li Y, Wu Y, Ren L, Wang G. LGI Net: Enhancing local-global information interaction for medical image segmentation. Comput Biol Med 2023; 167:107627. [PMID: 39491377 DOI: 10.1016/j.compbiomed.2023.107627] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2023] [Revised: 10/09/2023] [Accepted: 10/23/2023] [Indexed: 11/05/2024]
Abstract
Medical image segmentation is a critical task used to accurately extract regions of interest and pathological areas from medical images. In recent years, significant progress has been made in the field of medical image segmentation using deep learning and neural networks. However, existing methods still have limitations in terms of fusing local features and global contextual information due to the complex variations and irregular shapes of medical images. To address this issue, this paper proposes a medical image segmentation architecture called LGI Net, which improves the internal computation to achieve sufficient interaction between local perceptual capabilities and global contextual information within the network. Furthermore, the network incorporates an ECA module to effectively capture the interplay between channels and improve inter-layer information exchange capabilities. We conducted extensive experiments on three public medical image datasets: Kvasir, ISIC, and X-ray to validate the effectiveness of the proposed method. Ablation studies demonstrated the effectiveness of our LGAF, and comparative experiments confirmed the superiority of our proposed LGI Net in terms of accuracy and parameter efficiency. This study provides an innovative approach in the field of medical image segmentation, offering valuable insights for further improvements in accuracy and performance. The code and models will be available at https://github.com/LiuLinjie0310/LGI-Net.
Collapse
Affiliation(s)
- Linjie Liu
- College of Electronic and Information Engineering, Hebei University, Hebei, 071002, China
| | - Yan Li
- College of Electronic and Information Engineering, Hebei University, Hebei, 071002, China.
| | - Yanlin Wu
- College of Electronic and Information Engineering, Hebei University, Hebei, 071002, China
| | - Lili Ren
- Affiliated Hospital of Hebei University, Hebei, 071030, China
| | - Guanglei Wang
- College of Electronic and Information Engineering, Hebei University, Hebei, 071002, China; Hebei Key Laboratory of Precise Imaging of Inflammation Related Tumors, Hebei, 071000, China
| |
Collapse
|
11
|
Mu J, Lin Y, Meng X, Fan J, Ai D, Chen D, Qiu H, Yang J, Gu Y. M-CSAFN: Multi-Color Space Adaptive Fusion Network for Automated Port-Wine Stains Segmentation. IEEE J Biomed Health Inform 2023; 27:3924-3935. [PMID: 37027679 DOI: 10.1109/jbhi.2023.3247479] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/24/2023]
Abstract
Automatic segmentation of port-wine stains (PWS) from clinical images is critical for accurate diagnosis and objective assessment of PWS. However, this is a challenging task due to the color heterogeneity, low contrast, and indistinguishable appearance of PWS lesions. To address such challenges, we propose a novel multi-color space adaptive fusion network (M-CSAFN) for PWS segmentation. First, a multi-branch detection model is constructed based on six typical color spaces, which utilizes rich color texture information to highlight the difference between lesions and surrounding tissues. Second, an adaptive fusion strategy is used to fuse complementary predictions, which address the significant differences within the lesions caused by color heterogeneity. Third, a structural similarity loss with color information is proposed to measure the detail error between predicted lesions and truth lesions. Additionally, a PWS clinical dataset consisting of 1413 image pairs was established for the development and evaluation of PWS segmentation algorithms. To verify the effectiveness and superiority of the proposed method, we compared it with other state-of-the-art methods on our collected dataset and four publicly available skin lesion datasets (ISIC 2016, ISIC 2017, ISIC 2018, and PH2). The experimental results show that our method achieves remarkable performance in comparison with other state-of-the-art methods on our collected dataset, achieving 92.29% and 86.14% on Dice and Jaccard metrics, respectively. Comparative experiments on other datasets also confirmed the reliability and potential capability of M-CSAFN in skin lesion segmentation.
Collapse
|
12
|
Wang X, Han H, Xu M, Li S, Zhang D, Du S, Xu M. STNet: shape and texture joint learning through two-stream network for knowledge-guided image recognition. Front Neurosci 2023; 17:1212049. [PMID: 37397450 PMCID: PMC10309034 DOI: 10.3389/fnins.2023.1212049] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Accepted: 05/31/2023] [Indexed: 07/04/2023] Open
Abstract
Introduction The human brain processes shape and texture information separately through different neurons in the visual system. In intelligent computer-aided imaging diagnosis, pre-trained feature extractors are commonly used in various medical image recognition methods, common pre-training datasets such as ImageNet tend to improve the texture representation of the model but make it ignore many shape features. Weak shape feature representation is disadvantageous for some tasks that focus on shape features in medical image analysis. Methods Inspired by the function of neurons in the human brain, in this paper, we proposed a shape-and-texture-biased two-stream network to enhance the shape feature representation in knowledge-guided medical image analysis. First, the two-stream network shape-biased stream and a texture-biased stream are constructed through classification and segmentation multi-task joint learning. Second, we propose pyramid-grouped convolution to enhance the texture feature representation and introduce deformable convolution to enhance the shape feature extraction. Third, we used a channel-attention-based feature selection module in shape and texture feature fusion to focus on the key features and eliminate information redundancy caused by feature fusion. Finally, aiming at the problem of model optimization difficulty caused by the imbalance in the number of benign and malignant samples in medical images, an asymmetric loss function was introduced to improve the robustness of the model. Results and conclusion We applied our method to the melanoma recognition task on ISIC-2019 and XJTU-MM datasets, which focus on both the texture and shape of the lesions. The experimental results on dermoscopic image recognition and pathological image recognition datasets show the proposed method outperforms the compared algorithms and prove the effectiveness of our method.
Collapse
Affiliation(s)
- Xijing Wang
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Hongcheng Han
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Mengrui Xu
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
- The School of Software Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Shengpeng Li
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Dong Zhang
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
- The School of Automation Science and Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Shaoyi Du
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Meifeng Xu
- The Second Affiliated Hospital of Xi'an Jiaotong University (Xibei Hospital), Xi'an, China
| |
Collapse
|
13
|
Mirikharaji Z, Abhishek K, Bissoto A, Barata C, Avila S, Valle E, Celebi ME, Hamarneh G. A survey on deep learning for skin lesion segmentation. Med Image Anal 2023; 88:102863. [PMID: 37343323 DOI: 10.1016/j.media.2023.102863] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2022] [Revised: 02/01/2023] [Accepted: 05/31/2023] [Indexed: 06/23/2023]
Abstract
Skin cancer is a major public health problem that could benefit from computer-aided diagnosis to reduce the burden of this common disease. Skin lesion segmentation from images is an important step toward achieving this goal. However, the presence of natural and artificial artifacts (e.g., hair and air bubbles), intrinsic factors (e.g., lesion shape and contrast), and variations in image acquisition conditions make skin lesion segmentation a challenging task. Recently, various researchers have explored the applicability of deep learning models to skin lesion segmentation. In this survey, we cross-examine 177 research papers that deal with deep learning-based segmentation of skin lesions. We analyze these works along several dimensions, including input data (datasets, preprocessing, and synthetic data generation), model design (architecture, modules, and losses), and evaluation aspects (data annotation requirements and segmentation performance). We discuss these dimensions both from the viewpoint of select seminal works, and from a systematic viewpoint, examining how those choices have influenced current trends, and how their limitations should be addressed. To facilitate comparisons, we summarize all examined works in a comprehensive table as well as an interactive table available online3.
Collapse
Affiliation(s)
- Zahra Mirikharaji
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Kumar Abhishek
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Alceu Bissoto
- RECOD.ai Lab, Institute of Computing, University of Campinas, Av. Albert Einstein 1251, Campinas 13083-852, Brazil
| | - Catarina Barata
- Institute for Systems and Robotics, Instituto Superior Técnico, Avenida Rovisco Pais, Lisbon 1049-001, Portugal
| | - Sandra Avila
- RECOD.ai Lab, Institute of Computing, University of Campinas, Av. Albert Einstein 1251, Campinas 13083-852, Brazil
| | - Eduardo Valle
- RECOD.ai Lab, School of Electrical and Computing Engineering, University of Campinas, Av. Albert Einstein 400, Campinas 13083-952, Brazil
| | - M Emre Celebi
- Department of Computer Science and Engineering, University of Central Arkansas, 201 Donaghey Ave., Conway, AR 72035, USA.
| | - Ghassan Hamarneh
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada.
| |
Collapse
|
14
|
Maqsood S, Damaševičius R. Multiclass skin lesion localization and classification using deep learning based features fusion and selection framework for smart healthcare. Neural Netw 2023; 160:238-258. [PMID: 36701878 DOI: 10.1016/j.neunet.2023.01.022] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2022] [Revised: 11/13/2022] [Accepted: 01/19/2023] [Indexed: 01/27/2023]
Abstract
BACKGROUND The idea of smart healthcare has gradually gained attention as a result of the information technology industry's rapid development. Smart healthcare uses next-generation technologies i.e., artificial intelligence (AI) and Internet of Things (IoT), to intelligently transform current medical methods to make them more efficient, dependable and individualized. One of the most prominent uses of telemedicine and e-health in medical image analysis is teledermatology. Telecommunications technologies are used in this industry to send medical information to professionals. Teledermatology is a useful method for the identification of skin lesions, particularly in rural locations, because the skin is visually perceptible. One of the most recent tools for diagnosing skin cancer is dermoscopy. To classify skin malignancies, numerous computational approaches have been proposed in the literature. However, difficulties still exist i.e., lesions with low contrast, imbalanced datasets, high level of memory complexity, and the extraction of redundant features. METHODS In this work, a unified CAD model is proposed based on a deep learning framework for skin lesion segmentation and classification. In the proposed approach, the source dermoscopic images are initially pre-processed using a contrast enhancement based modified bio-inspired multiple exposure fusion approach. In the second stage, a custom 26-layered convolutional neural network (CNN) architecture is designed to segment the skin lesion regions. In the third stage, four pre-trained CNN models (Xception, ResNet-50, ResNet-101 and VGG16) are modified and trained using transfer learning on the segmented lesion images. In the fourth stage, the deep features vectors are extracted from all the CNN models and fused using the convolutional sparse image decomposition fusion approach. In the fifth stage, the univariate measurement and Poisson distribution feature selection approach is used for the best features selection for classification. Finally, the selected features are fed to the multi-class support vector machine (MC-SVM) for the final classification. RESULTS The proposed approach employed to the HAM10000, ISIC2018, ISIC2019, and PH2 datasets and achieved an accuracy of 98.57%, 98.62%, 93.47%, and 98.98% respectively which are better than previous works. CONCLUSION When compared to renowned state-of-the-art methods, experimental results show that the proposed skin lesion detection and classification approach achieved higher performance in terms of both visually and enhanced quantitative evaluation with enhanced accuracy.
Collapse
Affiliation(s)
- Sarmad Maqsood
- Department of Software Engineering, Faculty of Informatics Engineering, Kaunas University of Technology, LT-51386 Kaunas, Lithuania.
| | - Robertas Damaševičius
- Department of Software Engineering, Faculty of Informatics Engineering, Kaunas University of Technology, LT-51386 Kaunas, Lithuania.
| |
Collapse
|
15
|
Wang Y, Wang Y, Cai J, Lee TK, Miao C, Wang ZJ. SSD-KD: A self-supervised diverse knowledge distillation method for lightweight skin lesion classification using dermoscopic images. Med Image Anal 2023; 84:102693. [PMID: 36462373 DOI: 10.1016/j.media.2022.102693] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Revised: 11/08/2022] [Accepted: 11/09/2022] [Indexed: 11/15/2022]
Abstract
Skin cancer is one of the most common types of malignancy, affecting a large population and causing a heavy economic burden worldwide. Over the last few years, computer-aided diagnosis has been rapidly developed and make great progress in healthcare and medical practices due to the advances in artificial intelligence, particularly with the adoption of convolutional neural networks. However, most studies in skin cancer detection keep pursuing high prediction accuracies without considering the limitation of computing resources on portable devices. In this case, the knowledge distillation (KD) method has been proven as an efficient tool to help improve the adaptability of lightweight models under limited resources, meanwhile keeping a high-level representation capability. To bridge the gap, this study specifically proposes a novel method, termed SSD-KD, that unifies diverse knowledge into a generic KD framework for skin disease classification. Our method models an intra-instance relational feature representation and integrates it with existing KD research. A dual relational knowledge distillation architecture is self-supervised trained while the weighted softened outputs are also exploited to enable the student model to capture richer knowledge from the teacher model. To demonstrate the effectiveness of our method, we conduct experiments on ISIC 2019, a large-scale open-accessed benchmark of skin diseases dermoscopic images. Experiments show that our distilled MobileNetV2 can achieve an accuracy as high as 85% for the classification tasks of 8 different skin diseases with minimal parameters and computing requirements. Ablation studies confirm the effectiveness of our intra- and inter-instance relational knowledge integration strategy. Compared with state-of-the-art knowledge distillation techniques, the proposed method demonstrates improved performance. To the best of our knowledge, this is the first deep knowledge distillation application for multi-disease classification on the large-scale dermoscopy database. Our codes and models are available at https://github.com/enkiwang/Portable-Skin-Lesion-Diagnosis.
Collapse
Affiliation(s)
- Yongwei Wang
- Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly (LILY), NTU, Singapore
| | - Yuheng Wang
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Dermatology and Skin Science, University of British Columbia, Vancouver, BC, Canada; Photomedicine Institute, Vancouver Coast Health Research Institute, Vancouver, BC, Canada; Cancer Control Research Program, BC Cancer, Vancouver, BC, Canada.
| | - Jiayue Cai
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China
| | - Tim K Lee
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Dermatology and Skin Science, University of British Columbia, Vancouver, BC, Canada; Photomedicine Institute, Vancouver Coast Health Research Institute, Vancouver, BC, Canada; Cancer Control Research Program, BC Cancer, Vancouver, BC, Canada
| | - Chunyan Miao
- School of Computer Science and Engineering, Nanyang Technological University, Singapore.
| | - Z Jane Wang
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada
| |
Collapse
|
16
|
Lovis C, Weber J, Liopyris K, Braun RP, Marghoob AA, Quigley EA, Nelson K, Prentice K, Duhaime E, Halpern AC, Rotemberg V. Agreement Between Experts and an Untrained Crowd for Identifying Dermoscopic Features Using a Gamified App: Reader Feasibility Study. JMIR Med Inform 2023; 11:e38412. [PMID: 36652282 PMCID: PMC9892985 DOI: 10.2196/38412] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2022] [Revised: 09/28/2022] [Accepted: 10/16/2022] [Indexed: 01/19/2023] Open
Abstract
BACKGROUND Dermoscopy is commonly used for the evaluation of pigmented lesions, but agreement between experts for identification of dermoscopic structures is known to be relatively poor. Expert labeling of medical data is a bottleneck in the development of machine learning (ML) tools, and crowdsourcing has been demonstrated as a cost- and time-efficient method for the annotation of medical images. OBJECTIVE The aim of this study is to demonstrate that crowdsourcing can be used to label basic dermoscopic structures from images of pigmented lesions with similar reliability to a group of experts. METHODS First, we obtained labels of 248 images of melanocytic lesions with 31 dermoscopic "subfeatures" labeled by 20 dermoscopy experts. These were then collapsed into 6 dermoscopic "superfeatures" based on structural similarity, due to low interrater reliability (IRR): dots, globules, lines, network structures, regression structures, and vessels. These images were then used as the gold standard for the crowd study. The commercial platform DiagnosUs was used to obtain annotations from a nonexpert crowd for the presence or absence of the 6 superfeatures in each of the 248 images. We replicated this methodology with a group of 7 dermatologists to allow direct comparison with the nonexpert crowd. The Cohen κ value was used to measure agreement across raters. RESULTS In total, we obtained 139,731 ratings of the 6 dermoscopic superfeatures from the crowd. There was relatively lower agreement for the identification of dots and globules (the median κ values were 0.526 and 0.395, respectively), whereas network structures and vessels showed the highest agreement (the median κ values were 0.581 and 0.798, respectively). This pattern was also seen among the expert raters, who had median κ values of 0.483 and 0.517 for dots and globules, respectively, and 0.758 and 0.790 for network structures and vessels. The median κ values between nonexperts and thresholded average-expert readers were 0.709 for dots, 0.719 for globules, 0.714 for lines, 0.838 for network structures, 0.818 for regression structures, and 0.728 for vessels. CONCLUSIONS This study confirmed that IRR for different dermoscopic features varied among a group of experts; a similar pattern was observed in a nonexpert crowd. There was good or excellent agreement for each of the 6 superfeatures between the crowd and the experts, highlighting the similar reliability of the crowd for labeling dermoscopic images. This confirms the feasibility and dependability of using crowdsourcing as a scalable solution to annotate large sets of dermoscopic images, with several potential clinical and educational applications, including the development of novel, explainable ML tools.
Collapse
Affiliation(s)
| | - Jochen Weber
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Konstantinos Liopyris
- Department of Dermatology, Andreas Syggros Hospital of Cutaneous and Venereal Diseases, University of Athens, Athens, Greece
| | - Ralph P Braun
- Department of Dermatology, University Hospital Zurich, Zurich, Switzerland
| | - Ashfaq A Marghoob
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Elizabeth A Quigley
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Kelly Nelson
- Department of Dermatology, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | | | | | - Allan C Halpern
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| | - Veronica Rotemberg
- Dermatology Section, Memorial Sloan Kettering Cancer Center, New York, NY, United States
| |
Collapse
|
17
|
Celebi ME, Barata C, Halpern A, Tschandl P, Combalia M, Liu Y. Guest Editorial Skin Image Analysis in the Age of Deep Learning. IEEE J Biomed Health Inform 2023. [DOI: 10.1109/jbhi.2022.3227125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023]
Affiliation(s)
- M. Emre Celebi
- Department of Computer Science and Engineering, University of Central Arkansas, Conway, AR, USA
| | - Catarina Barata
- Institute for Systems and Robotics, University of Lisbon, Lisbon, Portugal
| | - Allan Halpern
- Dermatology Service, Memorial Sloan Kettering Cancer Center, New York, NY, USA
| | - Philipp Tschandl
- Department of Dermatology, Medical University of Vienna, Vienna, Austria
| | | | - Yuan Liu
- Google Health, Palo Alto, CA, USA
| |
Collapse
|
18
|
Ayas S. Multiclass skin lesion classification in dermoscopic images using swin transformer model. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-08053-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
19
|
Owida HA. Developments and Clinical Applications of Noninvasive Optical Technologies for Skin Cancer Diagnosis. J Skin Cancer 2022; 2022:9218847. [PMID: 36437851 PMCID: PMC9699785 DOI: 10.1155/2022/9218847] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2022] [Revised: 10/29/2022] [Accepted: 11/09/2022] [Indexed: 04/03/2024] Open
Abstract
Skin cancer has shown a sharp increase in prevalence over the past few decades and currently accounts for one-third of all cancers diagnosed. The most lethal form of skin cancer is melanoma, which develops in 4% of individuals. The rising prevalence and increased number of fatalities of skin cancer put a significant burden on healthcare resources and the economy. However, early detection and treatment greatly improve survival rates for patients with skin cancer. Since the rising rates of both the incidence and mortality have been particularly noticeable with melanoma, significant resources have been allocated to research aimed at earlier diagnosis and a deeper knowledge of the disease. Dermoscopy, reflectance confocal microscopy, optical coherence tomography, multiphoton-excited fluorescence imaging, and dermatofluorescence are only a few of the optical modalities reviewed here that have been employed to enhance noninvasive diagnosis of skin cancer in recent years. This review article discusses the methodology behind newly emerging noninvasive optical diagnostic technologies, their clinical applications, and advantages and disadvantages of these techniques, as well as the potential for their further advancement in the future.
Collapse
Affiliation(s)
- Hamza Abu Owida
- Medical Engineering Department, Faculty of Engineering, Al Ahliyya Amman University, Amman 19328, Jordan
| |
Collapse
|
20
|
Pennisi A, Bloisi DD, Suriani V, Nardi D, Facchiano A, Giampetruzzi AR. Skin Lesion Area Segmentation Using Attention Squeeze U-Net for Embedded Devices. J Digit Imaging 2022; 35:1217-1230. [PMID: 35505265 PMCID: PMC9582108 DOI: 10.1007/s10278-022-00634-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Revised: 03/24/2022] [Accepted: 03/28/2022] [Indexed: 11/30/2022] Open
Abstract
Melanoma is the deadliest form of skin cancer. Early diagnosis of malignant lesions is crucial for reducing mortality. The use of deep learning techniques on dermoscopic images can help in keeping track of the change over time in the appearance of the lesion, which is an important factor for detecting malignant lesions. In this paper, we present a deep learning architecture called Attention Squeeze U-Net for skin lesion area segmentation specifically designed for embedded devices. The main goal is to increase the patient empowerment through the adoption of deep learning algorithms that can run locally on smartphones or low cost embedded devices. This can be the basis to (1) create a history of the lesion, (2) reduce patient visits to the hospital, and (3) protect the privacy of the users. Quantitative results on publicly available data demonstrate that it is possible to achieve good segmentation results even with a compact model.
Collapse
Affiliation(s)
- Andrea Pennisi
- Dept. of Computer Science, University of Antwerp, Antwerpen, Belgium
| | - Domenico D. Bloisi
- Dept. of Mathematics, Computer Science, and Economics, University of Basilicata, Potenza, Italy
| | - Vincenzo Suriani
- Dept. of Computer Science, Control, and Management Engineering, Sapienza University of Rome, Rome, Italy
| | - Daniele Nardi
- Dept. of Computer Science, Control, and Management Engineering, Sapienza University of Rome, Rome, Italy
| | | | | |
Collapse
|
21
|
Slit lamp polarized dermoscopy: a cost-effective tool to assess eyelid lesions. Int Ophthalmol 2022; 43:1103-1110. [PMID: 36083562 DOI: 10.1007/s10792-022-02505-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2022] [Accepted: 08/28/2022] [Indexed: 10/14/2022]
Abstract
PURPOSE Dermoscopy is a complementary examination of skin lesions, which allows the observation of anatomical features invisible to the naked eye. Its use increases the diagnostic accuracy of skin tumors. The development of polarized dermoscopy allowed the observation of deeper skin structures, without the need of skin contact. The purpose of this study was to present a low-cost prototype through the adaptation of polarized lenses on a slit lamp in order to assess anatomical aspects invisible to conventional biomicroscopy in eyelid lesions. METHODS Twenty two eyelid lesions were documented using a prototype, compound of two polarizing filters, orthogonal to each other, adapted to a slit lamp with an integrated digital camera. Images of the eyelid lesions were also obtained with non-polarized biomicroscopy and with a portable dermatoscope, and were compared regarding anatomical aspects. RESULTS Anatomical structures imperceptible to conventional ophthalmic examination were evidenced using the polarized lenses, demonstrating that this tool can be useful to the ophthalmologist when assessing eyelid lesions. We have obtained high-quality images of the lesions. The slit lamp provided higher magnification, better focus control and easier assessment of eyelid lesions than the portable dermatoscope. CONCLUSION Ophthalmologists already use the slit lamp in their practice. The adaptation of polarized lenses to this device is a cost-effective, fast and non-invasive method that permits to improve the diagnostic accuracy of eyelid lesions, evidencing anatomical structures imperceptible to conventional ophthalmic examination.
Collapse
|
22
|
Wang Y, Fariah Haq N, Cai J, Kalia S, Lui H, Jane Wang Z, Lee TK. Multi-channel content based image retrieval method for skin diseases using similarity network fusion and deep community analysis. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103893] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
23
|
MDFNet: application of multimodal fusion method based on skin image and clinical data to skin cancer classification. J Cancer Res Clin Oncol 2022:10.1007/s00432-022-04180-1. [PMID: 35918465 DOI: 10.1007/s00432-022-04180-1] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2022] [Accepted: 06/27/2022] [Indexed: 10/16/2022]
Abstract
PURPOSE Skin cancer is one of the ten most common cancer types in the world. Early diagnosis and treatment can effectively reduce the mortality of patients. Therefore, it is of great significance to develop an intelligent diagnosis system for skin cancer. According to the survey, at present, most intelligent diagnosis systems of skin cancer only use skin image data, but the multi-modal cross-fusion analysis using image data and patient clinical data is limited. Therefore, to further explore the complementary relationship between image data and patient clinical data, we propose multimode data fusion diagnosis network (MDFNet), a framework for skin cancer based on data fusion strategy. METHODS MDFNet establishes an effective mapping among heterogeneous data features, effectively fuses clinical skin images and patient clinical data, and effectively solves the problems of feature paucity and insufficient feature richness that only use single-mode data. RESULTS The experimental results present that our proposed smart skin cancer diagnosis model has an accuracy of 80.42%, which is an improvement of about 9% compared with the model accuracy using only medical images, thus effectively confirming the unique fusion advantages exhibited by MDFNet. CONCLUSIONS This illustrates that MDFNet can not only be applied as an effective auxiliary diagnostic tool for skin cancer diagnosis, help physicians improve clinical decision-making ability and effectively improve the efficiency of clinical medicine diagnosis, but also its proposed data fusion method fully exerts the advantage of information convergence and has a certain reference value for the intelligent diagnosis of numerous clinical diseases.
Collapse
|
24
|
Serrano C, Lazo M, Serrano A, Toledo-Pastrana T, Barros-Tornay R, Acha B. Clinically Inspired Skin Lesion Classification through the Detection of Dermoscopic Criteria for Basal Cell Carcinoma. J Imaging 2022; 8:197. [PMID: 35877641 PMCID: PMC9319034 DOI: 10.3390/jimaging8070197] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2022] [Revised: 07/05/2022] [Accepted: 07/08/2022] [Indexed: 12/10/2022] Open
Abstract
Background and Objective. Skin cancer is the most common cancer worldwide. One of the most common non-melanoma tumors is basal cell carcinoma (BCC), which accounts for 75% of all skin cancers. There are many benign lesions that can be confused with these types of cancers, leading to unnecessary biopsies. In this paper, a new method to identify the different BCC dermoscopic patterns present in a skin lesion is presented. In addition, this information is applied to classify skin lesions into BCC and non-BCC. Methods. The proposed method combines the information provided by the original dermoscopic image, introduced in a convolutional neural network (CNN), with deep and handcrafted features extracted from color and texture analysis of the image. This color analysis is performed by transforming the image into a uniform color space and into a color appearance model. To demonstrate the validity of the method, a comparison between the classification obtained employing exclusively a CNN with the original image as input and the classification with additional color and texture features is presented. Furthermore, an exhaustive comparison of classification employing different color and texture measures derived from different color spaces is presented. Results. Results show that the classifier with additional color and texture features outperforms a CNN whose input is only the original image. Another important achievement is that a new color cooccurrence matrix, proposed in this paper, improves the results obtained with other texture measures. Finally, sensitivity of 0.99, specificity of 0.94 and accuracy of 0.97 are achieved when lesions are classified into BCC or non-BCC. Conclusions. To the best of our knowledge, this is the first time that a methodology to detect all the possible patterns that can be present in a BCC lesion is proposed. This detection leads to a clinically explainable classification into BCC and non-BCC lesions. In this sense, the classification of the proposed tool is based on the detection of the dermoscopic features that dermatologists employ for their diagnosis.
Collapse
Affiliation(s)
- Carmen Serrano
- Dpto. Teoría de la Señal y Comunicaciones, Universidad de Sevilla, Camino de los Descubrimientos s/n, 41092 Seville, Spain; (M.L.); (B.A.)
| | - Manuel Lazo
- Dpto. Teoría de la Señal y Comunicaciones, Universidad de Sevilla, Camino de los Descubrimientos s/n, 41092 Seville, Spain; (M.L.); (B.A.)
| | - Amalia Serrano
- Hospital Universitario Virgen Macarena, Calle Dr. Fedriani, 3, 41009 Seville, Spain;
| | - Tomás Toledo-Pastrana
- Hospitales Quironsalud Infanta Luisa y Sagrado Corazón, Calle San Jacinto, 87, 41010 Seville, Spain;
| | | | - Begoña Acha
- Dpto. Teoría de la Señal y Comunicaciones, Universidad de Sevilla, Camino de los Descubrimientos s/n, 41092 Seville, Spain; (M.L.); (B.A.)
| |
Collapse
|
25
|
Gajera HK, Nayak DR, Zaveri MA. Fusion of Local and Global Feature Representation With Sparse Autoencoder for Improved Melanoma Classification. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:5051-5054. [PMID: 36085953 DOI: 10.1109/embc48229.2022.9871370] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Automated skin cancer diagnosis is challenging due to inter-class uniformity, intra-class variation, and the complex structure of dermoscopy images. Convolutional neural networks (CNN) have recently made considerable progress in melanoma classification, even in the presence of limited skin images. One of the drawbacks of these methods is the loss of image details caused by downsampling high-resolution skin images to a low resolution. Further, most approaches extract features only from the whole skin image. This paper proposes an ensemble feature fusion and sparse autoencoder (SAE) based framework to overcome the above issues and improve melanoma classification performance. The proposed method extracts features from two streams, local and global, using a pre-trained CNN model. The local stream extracts features from image patches, while the global stream derives features from the whole skin image, preserving both local and global representation. The features are then fused, and an SAE framework is subsequently designed to enrich the feature representation further. The proposed method is validated on ISIC 2016 dataset and the experimental results indicate the superiority of the proposed approach.
Collapse
|
26
|
Gouda W, Sama NU, Al-Waakid G, Humayun M, Jhanjhi NZ. Detection of Skin Cancer Based on Skin Lesion Images Using Deep Learning. Healthcare (Basel) 2022; 10:healthcare10071183. [PMID: 35885710 PMCID: PMC9324455 DOI: 10.3390/healthcare10071183] [Citation(s) in RCA: 36] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2022] [Revised: 06/13/2022] [Accepted: 06/15/2022] [Indexed: 12/12/2022] Open
Abstract
An increasing number of genetic and metabolic anomalies have been determined to lead to cancer, generally fatal. Cancerous cells may spread to any body part, where they can be life-threatening. Skin cancer is one of the most common types of cancer, and its frequency is increasing worldwide. The main subtypes of skin cancer are squamous and basal cell carcinomas, and melanoma, which is clinically aggressive and responsible for most deaths. Therefore, skin cancer screening is necessary. One of the best methods to accurately and swiftly identify skin cancer is using deep learning (DL). In this research, the deep learning method convolution neural network (CNN) was used to detect the two primary types of tumors, malignant and benign, using the ISIC2018 dataset. This dataset comprises 3533 skin lesions, including benign, malignant, nonmelanocytic, and melanocytic tumors. Using ESRGAN, the photos were first retouched and improved. The photos were augmented, normalized, and resized during the preprocessing step. Skin lesion photos could be classified using a CNN method based on an aggregate of results obtained after many repetitions. Then, multiple transfer learning models, such as Resnet50, InceptionV3, and Inception Resnet, were used for fine-tuning. In addition to experimenting with several models (the designed CNN, Resnet50, InceptionV3, and Inception Resnet), this study’s innovation and contribution are the use of ESRGAN as a preprocessing step. Our designed model showed results comparable to the pretrained model. Simulations using the ISIC 2018 skin lesion dataset showed that the suggested strategy was successful. An 83.2% accuracy rate was achieved by the CNN, in comparison to the Resnet50 (83.7%), InceptionV3 (85.8%), and Inception Resnet (84%) models.
Collapse
Affiliation(s)
- Walaa Gouda
- Department of Computer Engineering and Network, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia
- Electrical Engineering Department, Faculty of Engineering at Shoubra, Benha University, Cairo 4272077, Egypt
- Correspondence: (W.G.); (M.H.)
| | - Najm Us Sama
- Faculty of Computer Science and Information Technology, Universiti Malaysia Sarawak, Kota Samarahan 94300, Malaysia;
| | - Ghada Al-Waakid
- Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia;
| | - Mamoona Humayun
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia
- Correspondence: (W.G.); (M.H.)
| | - Noor Zaman Jhanjhi
- School of Computer Science and Engineering (SCE), Taylor’s University, Subang Jaya 47500, Malaysia;
| |
Collapse
|
27
|
Colored Texture Analysis Fuzzy Entropy Methods with a Dermoscopic Application. ENTROPY 2022; 24:e24060831. [PMID: 35741551 PMCID: PMC9223301 DOI: 10.3390/e24060831] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/08/2022] [Revised: 06/09/2022] [Accepted: 06/11/2022] [Indexed: 02/05/2023]
Abstract
Texture analysis is a subject of intensive focus in research due to its significant role in the field of image processing. However, few studies focus on colored texture analysis and even fewer use information theory concepts. Entropy measures have been proven competent for gray scale images. However, to the best of our knowledge, there are no well-established entropy methods that deal with colored images yet. Therefore, we propose the recent colored bidimensional fuzzy entropy measure, FuzEnC2D, and introduce its new multi-channel approaches, FuzEnV2D and FuzEnM2D, for the analysis of colored images. We investigate their sensitivity to parameters and ability to identify images with different irregularity degrees, and therefore different textures. Moreover, we study their behavior with colored Brodatz images in different color spaces. After verifying the results with test images, we employ the three methods for analyzing dermoscopic images of malignant melanoma and benign melanocytic nevi. FuzEnC2D, FuzEnV2D, and FuzEnM2D illustrate a good differentiation ability between the two-similar in appearance-pigmented skin lesions. The results outperform those of a well-known texture analysis measure. Our work provides the first entropy measure studying colored images using both single and multi-channel approaches.
Collapse
|
28
|
An Effective Skin Cancer Classification Mechanism via Medical Vision Transformer. SENSORS 2022; 22:s22114008. [PMID: 35684627 PMCID: PMC9182815 DOI: 10.3390/s22114008] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/10/2022] [Revised: 05/12/2022] [Accepted: 05/20/2022] [Indexed: 11/17/2022]
Abstract
Skin Cancer (SC) is considered the deadliest disease in the world, killing thousands of people every year. Early SC detection can increase the survival rate for patients up to 70%, hence it is highly recommended that regular head-to-toe skin examinations are conducted to determine whether there are any signs or symptoms of SC. The use of Machine Learning (ML)-based methods is having a significant impact on the classification and detection of SC diseases. However, there are certain challenges associated with the accurate classification of these diseases such as a lower detection accuracy, poor generalization of the models, and an insufficient amount of labeled data for training. To address these challenges, in this work we developed a two-tier framework for the accurate classification of SC. During the first stage of the framework, we applied different methods for data augmentation to increase the number of image samples for effective training. As part of the second tier of the framework, taking into consideration the promising performance of the Medical Vision Transformer (MVT) in the analysis of medical images, we developed an MVT-based classification model for SC. This MVT splits the input image into image patches and then feeds these patches to the transformer in a sequence structure, like word embedding. Finally, Multi-Layer Perceptron (MLP) is used to classify the input image into the corresponding class. Based on the experimental results achieved on the Human Against Machine (HAM10000) datasets, we concluded that the proposed MVT-based model achieves better results than current state-of-the-art techniques for SC classification.
Collapse
|
29
|
Bhimavarapu U, Battineni G. Skin Lesion Analysis for Melanoma Detection Using the Novel Deep Learning Model Fuzzy GC-SCNN. Healthcare (Basel) 2022; 10:healthcare10050962. [PMID: 35628098 PMCID: PMC9141659 DOI: 10.3390/healthcare10050962] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2022] [Revised: 05/19/2022] [Accepted: 05/21/2022] [Indexed: 02/01/2023] Open
Abstract
Melanoma is easily detectable by visual examination since it occurs on the skin’s surface. In melanomas, which are the most severe types of skin cancer, the cells that make melanin are affected. However, the lack of expert opinion increases the processing time and cost of computer-aided skin cancer detection. As such, we aimed to incorporate deep learning algorithms to conduct automatic melanoma detection from dermoscopic images. The fuzzy-based GrabCut-stacked convolutional neural networks (GC-SCNN) model was applied for image training. The image features extraction and lesion classification were performed on different publicly available datasets. The fuzzy GC-SCNN coupled with the support vector machines (SVM) produced 99.75% classification accuracy and 100% sensitivity and specificity, respectively. Additionally, model performance was compared with existing techniques and outcomes suggesting the proposed model could detect and classify the lesion segments with higher accuracy and lower processing time than other techniques.
Collapse
Affiliation(s)
- Usharani Bhimavarapu
- School of Competitive Coding, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Vijayawada 522502, India;
| | - Gopi Battineni
- Clinical Research Centre, School of Medicinal and Health Products Sciences, University of Camerino, 62032 Camerino, Italy
- Correspondence: ; Tel.: +39-3331728206
| |
Collapse
|
30
|
Guest Editorial: Image Analysis in Dermatology. Med Image Anal 2022; 79:102468. [DOI: 10.1016/j.media.2022.102468] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
31
|
Bardou D, Bouaziz H, Lv L, Zhang T. Hair removal in dermoscopy images using variational autoencoders. Skin Res Technol 2022; 28:445-454. [PMID: 35254677 PMCID: PMC9907627 DOI: 10.1111/srt.13145] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2021] [Accepted: 01/17/2022] [Indexed: 01/23/2023]
Abstract
BACKGROUND In recent years, melanoma is rising at a faster rate compared to other cancers. Although it is the most serious type of skin cancer, the diagnosis at early stages makes it curable. Dermoscopy is a reliable medical technique used to detect melanoma by using a dermoscope to examine the skin. In the last few decades, digital imaging devices have made great progress which allowed capturing and storing high-quality images from these examinations. The stored images are now being standardized and used for the automatic detection of melanoma. However, when the hair covers the skin, this makes the task challenging. Therefore, it is important to eliminate the hair to get accurate results. METHODS In this paper, we propose a simple yet efficient method for hair removal using a variational autoencoder without the need for paired samples. The encoder takes as input a dermoscopy image and builds a latent distribution that ignores hair as it is considered noise, while the decoder reconstructs a hair-free image. Both encoder and decoder use a decent convolutional neural networks architecture that provides high performance. The construction of our model comprises two stages of training. In the first stage, the model has trained on hair-occluded images to output hair-free images, and in the second stage, it is optimized using hair-free images to preserve the image textures. Although the variational autoencoder produces hair-free images, it does not maintain the quality of the generated images. Thus, we explored the use of three-loss functions including the structural similarity index (SSIM), L1-norm, and L2-norm to improve the visual quality of the generated images. RESULTS The evaluation of the hair-free reconstructed images is carried out using t-distributed stochastic neighbor embedding (SNE) feature mapping by visualizing the distribution of the real hair-free images and the synthesized hair-free images. The conducted experiments on the publicly available dataset HAM10000 show that our method is very efficient.
Collapse
Affiliation(s)
- Dalal Bardou
- Department of Computer Science and Mathematics University of Abbes Laghrour Khenchela Algeria
| | - Hamida Bouaziz
- Mécatronique Laboratory Department of Computer Science Jijel University Jijel Algeria
| | - Laishui Lv
- School of Computer Science and Engineering Nanjing University of Science and Technology Nanjing China
| | - Ting Zhang
- School of Computer Science and Engineering Nanjing University of Science and Technology Nanjing China
| |
Collapse
|
32
|
TATL: Task Agnostic Transfer Learning for Skin Attributes Detection. Med Image Anal 2022; 78:102359. [DOI: 10.1016/j.media.2022.102359] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2021] [Revised: 01/03/2022] [Accepted: 01/10/2022] [Indexed: 11/20/2022]
|
33
|
An Improved and Robust Encoder–Decoder for Skin Lesion Segmentation. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2022. [DOI: 10.1007/s13369-021-06403-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
34
|
Puri N, Gill S, Kumar S, Brar B, Chahal A. A study of dermoscopy in patients of melasma in a tertiary care centre in North India. PIGMENT INTERNATIONAL 2022. [DOI: 10.4103/pigmentinternational.pigmentinternational_84_20] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2022] Open
|
35
|
Deng X, Yin Q, Guo P. Efficient structural pseudoinverse learning-based hierarchical representation learning for skin lesion classification. COMPLEX INTELL SYST 2021. [DOI: 10.1007/s40747-021-00588-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
AbstractThe success of deep learning in skin lesion classification mainly depends on the ultra-deep neural network and the significantly large training data set. Deep learning training is usually time-consuming, and large datasets with labels are hard to obtain, especially skin lesion images. Although pre-training and data augmentation can alleviate these issues, there are still some problems: (1) the data domain is not consistent, resulting in the slow convergence; and (2) low robustness to confusing skin lesions. To solve these problems, we propose an efficient structural pseudoinverse learning-based hierarchical representation learning method. Preliminary feature extraction, shallow network feature extraction and deep learning feature extraction are carried out respectively before the classification of skin lesion images. Gabor filter and pre-trained deep convolutional neural network are used for preliminary feature extraction. The structural pseudoinverse learning (S-PIL) algorithm is used to extract the shallow features. Then, S-PIL preliminarily identifies the skin lesion images that are difficult to be classified to form a new training set for deep learning feature extraction. Through the hierarchical representation learning, we analyze the features of skin lesion images layer by layer to improve the final classification. Our method not only avoid the slow convergence caused by inconsistency of data domain but also enhances the training of confusing examples. Without using additional data, our approach outperforms existing methods in the ISIC 2017 and ISIC 2018 datasets.
Collapse
|
36
|
Srivastava A, Jha D, Chanda S, Pal U, Johansen H, Johansen D, Riegler M, Ali S, Halvorsen P. MSRF-Net: A Multi-Scale Residual Fusion Network for Biomedical Image Segmentation. IEEE J Biomed Health Inform 2021; 26:2252-2263. [PMID: 34941539 DOI: 10.1109/jbhi.2021.3138024] [Citation(s) in RCA: 71] [Impact Index Per Article: 17.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Methods based on convolutional neural networks have improved the performance of biomedical image segmentation. However, most of these methods cannot efficiently segment objects of variable sizes and train on small and biased datasets, which are common for biomedical use cases. While methods exist that incorporate multi-scale fusion approaches to address the challenges arising with variable sizes, they usually use complex models that are more suitable for general semantic segmentation problems. In this paper, we propose a novel architecture called Multi-Scale Residual Fusion Network (MSRF-Net), which is specially designed for medical image segmentation. The proposed MSRF-Net is able to exchange multi-scale features of varying receptive fields using a Dual-Scale Dense Fusion (DSDF) block. Our DSDF block can exchange information rigorously across two different resolution scales, and our MSRF sub-network uses multiple DSDF blocks in sequence to perform multi-scale fusion. This allows the preservation of resolution, improved information flow and propagation of both high- and low-level features to obtain accurate segmentation maps. The proposed MSRF-Net allows to capture object variabilities and provides improved results on different biomedical datasets. Extensive experiments on MSRF-Net demonstrate that the proposed method outperforms the cutting edge medical image segmentation methods on four publicly available datasets. We achieve the Dice Coefficient (DSC) of 0.9217, 0.9420, and 0.9224, 0.8824 on Kvasir-SEG, CVC-ClinicDB, 2018 Data Science Bowl dataset, and ISIC-2018 skin lesion segmentation challenge dataset respectively. We further conducted generalizability tests that also achieved the highest DSC score with 0.7921 and 0.7575 on CVC-ClinicDB and Kvasir-SEG, respectively.
Collapse
|
37
|
Spyridonos P, Gaitanis G, Likas A, Bassukas I. Characterizing Malignant Melanoma Clinically Resembling Seborrheic Keratosis Using Deep Knowledge Transfer. Cancers (Basel) 2021; 13:cancers13246300. [PMID: 34944920 PMCID: PMC8699430 DOI: 10.3390/cancers13246300] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2021] [Revised: 12/10/2021] [Accepted: 12/13/2021] [Indexed: 12/26/2022] Open
Abstract
Simple Summary Malignant melanomas (MMs) with aypical clinical presentation constitute a diagnostic pitfall, and false negatives carry the risk of a diagnostic delay and improper disease management. Among the most common, challenging presentation forms of MMs are those that clinically resemble seborrheic keratosis (SK). On the other hand, SK may mimic melanoma, producing ‘false positive overdiagnosis’ and leading to needless excisions. The evolving efficiency of deep learning algorithms in image recognition and the availability of large image databases have accelerated the development of advanced computer-aided systems for melanoma detection. In the present study, we used image data from the International Skin Image Collaboration archive to explore the capacity of deep knowledge transfer in the challenging diagnostic task of the atypical skin tumors of MM and SK. Abstract Malignant melanomas resembling seborrheic keratosis (SK-like MMs) are atypical, challenging to diagnose melanoma cases that carry the risk of delayed diagnosis and inadequate treatment. On the other hand, SK may mimic melanoma, producing a ‘false positive’ with unnecessary lesion excisions. The present study proposes a computer-based approach using dermoscopy images for the characterization of SΚ-like MMs. Dermoscopic images were retrieved from the International Skin Imaging Collaboration archive. Exploiting image embeddings from pretrained convolutional network VGG16, we trained a support vector machine (SVM) classification model on a data set of 667 images. SVM optimal hyperparameter selection was carried out using the Bayesian optimization method. The classifier was tested on an independent data set of 311 images with atypical appearance: MMs had an absence of pigmented network and had an existence of milia-like cysts. SK lacked milia-like cysts and had a pigmented network. Atypical MMs were characterized with a sensitivity and specificity of 78.6% and 84.5%, respectively. The advent of deep learning in image recognition has attracted the interest of computer science towards improved skin lesion diagnosis. Open-source, public access archives of skin images empower further the implementation and validation of computer-based systems that might contribute significantly to complex clinical diagnostic problems such as the characterization of SK-like MMs.
Collapse
Affiliation(s)
- Panagiota Spyridonos
- Department of Medical Physics, Faculty of Medicine, School of Health Sciences, University of Ioannina, 45110 Ioannina, Greece
- Correspondence: (P.S.); (I.B.)
| | - George Gaitanis
- Department of Skin and Venereal Diseases, Faculty of Medicine, School of Health Sciences, University of Ioannina, 45110 Ioannina, Greece;
| | - Aristidis Likas
- Department of Computer Science & Engineering, School of Engineering, University of Ioannina, 45110 Ioannina, Greece;
| | - Ioannis Bassukas
- Department of Skin and Venereal Diseases, Faculty of Medicine, School of Health Sciences, University of Ioannina, 45110 Ioannina, Greece;
- Correspondence: (P.S.); (I.B.)
| |
Collapse
|
38
|
Zhao M, Kawahara J, Abhishek K, Shamanian S, Hamarneh G. Skin3D: Detection and Longitudinal Tracking of Pigmented Skin Lesions in 3D Total-Body Textured Meshes. Med Image Anal 2021; 77:102329. [DOI: 10.1016/j.media.2021.102329] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Revised: 09/27/2021] [Accepted: 12/01/2021] [Indexed: 10/19/2022]
|
39
|
Benyahia S, Meftah B, Lézoray O. Multi-features extraction based on deep learning for skin lesion classification. Tissue Cell 2021; 74:101701. [PMID: 34861582 DOI: 10.1016/j.tice.2021.101701] [Citation(s) in RCA: 21] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2021] [Revised: 11/22/2021] [Accepted: 11/23/2021] [Indexed: 10/19/2022]
Abstract
For various forms of skin lesion, many different feature extraction methods have been investigated so far. Indeed, feature extraction is a crucial step in machine learning processes. In general, we can distinct handcrafted and deep learning features. In this paper, we investigate the efficiency of using 17 commonly pre-trained convolutional neural networks (CNN) architectures as feature extractors and of 24 machine learning classifiers to evaluate the classification of skin lesions from two different datasets: ISIC 2019 and PH2. In this research, we find out that a DenseNet201 combined with Fine KNN or Cubic SVM achieved the best results in accuracy (92.34% and 91.71%) for the ISIC 2019 dataset. The results also show that the suggested method outperforms others approaches with an accuracy of 99% on the PH2 dataset.
Collapse
Affiliation(s)
- Samia Benyahia
- Department of Computer Science, Faculty of Exact Sciences, University of Mascara, Mascara, Algeria
| | | | - Olivier Lézoray
- Normandie Univ, UNICAEN, ENSICAEN, CNRS, GREYC, Caen, France
| |
Collapse
|
40
|
Wang Y, Cai J, Louie DC, Wang ZJ, Lee TK. Incorporating clinical knowledge with constrained classifier chain into a multimodal deep network for melanoma detection. Comput Biol Med 2021; 137:104812. [PMID: 34507158 DOI: 10.1016/j.compbiomed.2021.104812] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2021] [Revised: 08/25/2021] [Accepted: 08/25/2021] [Indexed: 10/20/2022]
Abstract
In recent years, vast developments in Computer-Aided Diagnosis (CAD) for skin diseases have generated much interest from clinicians and other eventual end-users of this technology. Introducing clinical domain knowledge to these machine learning strategies can help dispel the black box nature of these tools, strengthening clinician trust. Clinical domain knowledge also provides new information channels which can improve CAD diagnostic performance. In this paper, we propose a novel framework for malignant melanoma (MM) detection by fusing clinical images and dermoscopic images. The proposed method combines a multi-labeled deep feature extractor and clinically constrained classifier chain (CC). This allows the 7-point checklist, a clinician diagnostic algorithm, to be included in the decision level while maintaining the clinical importance of the major and minor criteria in the checklist. Our proposed framework achieved an average accuracy of 81.3% for detecting all criteria and melanoma when testing on a publicly available 7-point checklist dataset. This is the highest reported results, outperforming state-of-the-art methods in the literature by 6.4% or more. Analyses also show that the proposed system surpasses the single modality system of using either clinical images or dermoscopic images alone and the systems without adopting the approach of multi-label and clinically constrained classifier chain. Our carefully designed system demonstrates a substantial improvement over melanoma detection. By keeping the familiar major and minor criteria of the 7-point checklist and their corresponding weights, the proposed system may be more accepted by physicians as a human-interpretable CAD tool for automated melanoma detection.
Collapse
Affiliation(s)
- Yuheng Wang
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Dermatology and Skin Science, University of British Columbia, Vancouver, BC, Canada; Photomedicine Institute, Vancouver Coast Health Research Institute, Vancouver, BC, Canada; Departments of Cancer Control Research and Integrative Oncology, BC Cancer, Vancouver, BC, Canada
| | - Jiayue Cai
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China; Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada.
| | - Daniel C Louie
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Dermatology and Skin Science, University of British Columbia, Vancouver, BC, Canada; Photomedicine Institute, Vancouver Coast Health Research Institute, Vancouver, BC, Canada; Departments of Cancer Control Research and Integrative Oncology, BC Cancer, Vancouver, BC, Canada
| | - Z Jane Wang
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC, Canada
| | - Tim K Lee
- School of Biomedical Engineering, University of British Columbia, Vancouver, BC, Canada; Department of Dermatology and Skin Science, University of British Columbia, Vancouver, BC, Canada; Photomedicine Institute, Vancouver Coast Health Research Institute, Vancouver, BC, Canada; Departments of Cancer Control Research and Integrative Oncology, BC Cancer, Vancouver, BC, Canada
| |
Collapse
|
41
|
Rastghalam R, Danyali H, Helfroush MS, Celebi ME, Mokhtari M. Skin Melanoma Detection in Microscopic Images Using HMM-Based Asymmetric Analysis and Expectation Maximization. IEEE J Biomed Health Inform 2021; 25:3486-3497. [PMID: 34003756 DOI: 10.1109/jbhi.2021.3081185] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Melanoma is one of the deadliest types of skin cancer with increasing incidence. The most definitive diagnosis method is the histopathological examination of the tissue sample. In this paper, a melanoma detection algorithm is proposed based on decision-level fusion and a Hidden Markov Model (HMM), whose parameters are optimized using Expectation Maximization (EM) and asymmetric analysis. The texture heterogeneity of the samples is determined using asymmetric analysis. A fusion-based HMM classifier trained using EM is introduced. For this purpose, a novel texture feature is extracted based on two local binary patterns, namely local difference pattern (LDP) and statistical histogram features of the microscopic image. Extensive experiments demonstrate that the proposed melanoma detection algorithm yields a total error of less than 0.04%.
Collapse
|
42
|
Pacheco AGC, Krohling RA. An Attention-Based Mechanism to Combine Images and Metadata in Deep Learning Models Applied to Skin Cancer Classification. IEEE J Biomed Health Inform 2021; 25:3554-3563. [DOI: 10.1109/jbhi.2021.3062002] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
43
|
Ramya J, Vijaylakshmi H, Mirza Saifuddin H. Segmentation of skin lesion images using discrete wavelet transform. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.102839] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
|
44
|
Computer-aided skin cancer diagnosis based on a New meta-heuristic algorithm combined with support vector method. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.102631] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
45
|
|
46
|
Khan MA, Sharif M, Akram T, Damaševičius R, Maskeliūnas R. Skin Lesion Segmentation and Multiclass Classification Using Deep Learning Features and Improved Moth Flame Optimization. Diagnostics (Basel) 2021; 11:811. [PMID: 33947117 PMCID: PMC8145295 DOI: 10.3390/diagnostics11050811] [Citation(s) in RCA: 87] [Impact Index Per Article: 21.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Revised: 04/23/2021] [Accepted: 04/26/2021] [Indexed: 11/18/2022] Open
Abstract
Manual diagnosis of skin cancer is time-consuming and expensive; therefore, it is essential to develop automated diagnostics methods with the ability to classify multiclass skin lesions with greater accuracy. We propose a fully automated approach for multiclass skin lesion segmentation and classification by using the most discriminant deep features. First, the input images are initially enhanced using local color-controlled histogram intensity values (LCcHIV). Next, saliency is estimated using a novel Deep Saliency Segmentation method, which uses a custom convolutional neural network (CNN) of ten layers. The generated heat map is converted into a binary image using a thresholding function. Next, the segmented color lesion images are used for feature extraction by a deep pre-trained CNN model. To avoid the curse of dimensionality, we implement an improved moth flame optimization (IMFO) algorithm to select the most discriminant features. The resultant features are fused using a multiset maximum correlation analysis (MMCA) and classified using the Kernel Extreme Learning Machine (KELM) classifier. The segmentation performance of the proposed methodology is analyzed on ISBI 2016, ISBI 2017, ISIC 2018, and PH2 datasets, achieving an accuracy of 95.38%, 95.79%, 92.69%, and 98.70%, respectively. The classification performance is evaluated on the HAM10000 dataset and achieved an accuracy of 90.67%. To prove the effectiveness of the proposed methods, we present a comparison with the state-of-the-art techniques.
Collapse
Affiliation(s)
- Muhammad Attique Khan
- Department of Computer Science, Wah Campus, COMSATS University Islamabad, Wah Cantonment 47040, Pakistan;
| | - Muhammad Sharif
- Department of Computer Science, Wah Campus, COMSATS University Islamabad, Wah Cantonment 47040, Pakistan;
| | - Tallha Akram
- Department of Electrical Engineering, Wah Campus, COMSATS University Islamabad, Islamabad 45550, Pakistan;
| | - Robertas Damaševičius
- Faculty of Applied Mathematics, Silesian University of Technology, 44-100 Gliwice, Poland
| | - Rytis Maskeliūnas
- Department of Applied Informatics, Vytautas Magnus University, 44404 Kaunas, Lithuania;
| |
Collapse
|
47
|
Meng X, Chen J, Zhang Z, Li K, Li J, Yu Z, Zhang Y. Non-invasive optical methods for melanoma diagnosis. Photodiagnosis Photodyn Ther 2021; 34:102266. [PMID: 33785441 DOI: 10.1016/j.pdpdt.2021.102266] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2021] [Revised: 03/09/2021] [Accepted: 03/22/2021] [Indexed: 02/07/2023]
Abstract
Cutaneous melanoma is one of the most common malignancies with increased incidence in the past few decades, making it a significant public health problem. The early diagnosis of melanoma is a major factor in improving patient's survival. The traditional pathway to melanoma diagnosis starts with a visual diagnosis, followed by subsequent biopsy and histopathologic evaluation. Recently, multiple innovative optical technology-based methods, including dermoscopy, reflectance confocal microscopy, optical coherence tomography, multiphoton excited fluorescence imaging and stepwise two-photon excited fluorescence (dermatofluoroscopy), have been developed to increase the diagnostic accuracy for the non-invasive melanoma diagnosis. Some of them have already been applied to real-life clinical settings, others require more research and development. These technologies show promise in facilitating the diagnosis of melanoma since they are non-invasive, sensitive, objective and easy to apply. Diagnostic accuracy, detection time, portability and the cost-effectiveness of the device are all aspects that need to be improved. This article reviews the method of these emerging optical non-invasive diagnostic technologies, their clinical application, their benefits and limitations, as well as their possible future development.
Collapse
Affiliation(s)
- Xinxian Meng
- Department of Plastic and Reconstructive Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China
| | - Jun Chen
- Department of Dermatology and Dermatologic Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China
| | - Zheng Zhang
- Department of Plastic and Reconstructive Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China
| | - Ke Li
- Department of Plastic and Reconstructive Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China
| | - Jie Li
- Department of Plastic and Reconstructive Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China
| | - Zhixi Yu
- Department of Plastic and Reconstructive Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China
| | - Yixin Zhang
- Department of Plastic and Reconstructive Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China; Department of Laser and Aesthetic Medicine, Shanghai Ninth People's Hospital, Affiliated to Shanghai JiaoTong University School of Medicine, Shanghai, China.
| |
Collapse
|
48
|
Van Molle P, Verbelen T, Vankeirsbilck B, De Vylder J, Diricx B, Kimpe T, Simoens P, Dhoedt B. Leveraging the Bhattacharyya coefficient for uncertainty quantification in deep neural networks. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-05789-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
Abstract
AbstractModern deep learning models achieve state-of-the-art results for many tasks in computer vision, such as image classification and segmentation. However, its adoption into high-risk applications, e.g. automated medical diagnosis systems, happens at a slow pace. One of the main reasons for this is that regular neural networks do not capture uncertainty. To assess uncertainty in classification, several techniques have been proposed casting neural network approaches in a Bayesian setting. Amongst these techniques, Monte Carlo dropout is by far the most popular. This particular technique estimates the moments of the output distribution through sampling with different dropout masks. The output uncertainty of a neural network is then approximated as the sample variance. In this paper, we highlight the limitations of such a variance-based uncertainty metric and propose an novel approach. Our approach is based on the overlap between output distributions of different classes. We show that our technique leads to a better approximation of the inter-class output confusion. We illustrate the advantages of our method using benchmark datasets. In addition, we apply our metric to skin lesion classification—a real-world use case—and show that this yields promising results.
Collapse
|
49
|
Attique Khan M, Akram T, Sharif M, Kadry S, Nam Y. Computer Decision Support System for Skin Cancer Localization and Classification. COMPUTERS, MATERIALS & CONTINUA 2021; 68:1041-1064. [DOI: 10.32604/cmc.2021.016307] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/30/2020] [Accepted: 02/02/2021] [Indexed: 08/25/2024]
|
50
|
Birkenfeld JS, Tucker-Schwartz JM, Soenksen LR, Avilés-Izquierdo JA, Marti-Fuster B. Computer-aided classification of suspicious pigmented lesions using wide-field images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2020; 195:105631. [PMID: 32652382 DOI: 10.1016/j.cmpb.2020.105631] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/20/2020] [Accepted: 06/21/2020] [Indexed: 06/11/2023]
Abstract
BACKGROUND AND OBJECTIVE Early identification of melanoma is conducted through whole-body visual examinations to detect suspicious pigmented lesions, a situation that fluctuates in accuracy depending on the experience and time of the examiner. Computer-aided diagnosis tools for skin lesions are typically trained using pre-selected single-lesion images, taken under controlled conditions, which limits their use in wide-field scenes. Here, we propose a computer-aided classifier system with such input conditions to aid in the rapid identification of suspicious pigmented lesions at the primary care level. METHODS 133 patients with a multitude of skin lesions were recruited for this study. All lesions were examined by a board-certified dermatologist and classified into "suspicious" and "non-suspicious". A new clinical database was acquired and created by taking Wide-Field images of all major body parts with a consumer-grade camera under natural illumination condition and with a consistent source of image variability. 3-8 images were acquired per patient on different sites of the body, and a total of 1759 pigmented lesions were extracted. A machine learning classifier was optimized and build into a computer aided classification system to binary classify each lesion using a suspiciousness score. RESULTS In a testing set, our computer-aided classification system achieved a sensitivity of 100% for suspicious pigmented lesions that were later confirmed by dermoscopy examination ("SPL_A") and 83.2% for suspicious pigmented lesions that were not confirmed after examination ("SPL_B"). Sensitivity for non-suspicious lesions was 72.1%, and accuracy was 75.9%. With these results we defined a suspiciousness score that is aligned with common macro-screening (naked eye) practices. CONCLUSIONS This work demonstrates that wide-field photography combined with computer-aided classification systems can distinguish suspicious from non-suspicious pigmented lesions, and might be effective to assess the severity of a suspicious pigmented lesions. We believe this approach could be useful to support skin screenings at a population-level.
Collapse
Affiliation(s)
- Judith S Birkenfeld
- Research Laboratory of Electronics, Massachusetts Institute of Technology, 77 Massachusetts Ave, Cambridge, MA 02139, USA; MIT linQ, Institute for Medical Engineering and Science, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, MA 02139, USA; Brigham and Women's Hospital - Harvard Medical School, 75 Francis St, Boston, MA 02115, United States; Massachusetts General Hospital - Harvard Medical School, 55 Fruit St, Boston, MA 02114, United States.
| | - Jason M Tucker-Schwartz
- MIT linQ, Institute for Medical Engineering and Science, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, MA 02139, USA
| | - Luis R Soenksen
- Research Laboratory of Electronics, Massachusetts Institute of Technology, 77 Massachusetts Ave, Cambridge, MA 02139, USA; MIT linQ, Institute for Medical Engineering and Science, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, MA 02139, USA; Department of Mechanical Engineering, Massachusetts Institute of Technology, 77 Massachusetts Ave, Cambridge, MA 02139, USA; Wyss Institute for Biologically Inspired Engineering, Harvard University, 3 Blackfan Cir, Boston, MA 02115, USA; Harvard-MIT Program in Health Sciences and Technology, Cambridge, MA 02139, USA
| | - José A Avilés-Izquierdo
- Department of Dermatology, Hospital General Universitario Gregorio Marañón, Calle del Dr. Esquerdo 46, 28007 Madrid, Spain
| | - Berta Marti-Fuster
- Research Laboratory of Electronics, Massachusetts Institute of Technology, 77 Massachusetts Ave, Cambridge, MA 02139, USA; MIT linQ, Institute for Medical Engineering and Science, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, MA 02139, USA; Brigham and Women's Hospital - Harvard Medical School, 75 Francis St, Boston, MA 02115, United States
| |
Collapse
|