1
|
Hu D, Jiang Z, Shi J, Xie F, Wu K, Tang K, Cao M, Huai J, Zheng Y. Pathology report generation from whole slide images with knowledge retrieval and multi-level regional feature selection. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 263:108677. [PMID: 40023962 DOI: 10.1016/j.cmpb.2025.108677] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/18/2024] [Revised: 02/07/2025] [Accepted: 02/16/2025] [Indexed: 03/04/2025]
Abstract
BACKGROUND AND OBJECTIVES With the development of deep learning techniques, the computer-assisted pathology diagnosis plays a crucial role in clinical diagnosis. An important task within this field is report generation, which provides doctors with text descriptions of whole slide images (WSIs). Report generation from WSIs presents significant challenges due to the structural complexity and pathological diversity of tissues, as well as the large size and high information density of WSIs. The objective of this study is to design a histopathology report generation method that can efficiently generate reports from WSIs and is suitable for clinical practice. METHODS In this paper, we propose a novel approach for generating pathology reports from WSIs, leveraging knowledge retrieval and multi-level regional feature selection. To deal with the uneven distribution of pathological information in WSIs, we introduce a multi-level regional feature encoding network and a feature selection module that extracts multi-level region representations and filters out region features irrelevant to the diagnosis, enabling more efficient report generation. Moreover, we design a knowledge retrieval module to improve the report generation performance that can leverage the diagnostic information from historical cases. Additionally, we propose an out-of-domain application mode based on large language model (LLM). The use of LLM enhances the scalability of the generation model and improves its adaptability to data from different sources. RESULTS The proposed method is evaluated on a public datasets and one in-house dataset. On the public GastricADC (991 WSIs), our method outperforms state-of-the-art text generation methods and achieved 0.568 and 0.345 on metric Rouge-L and Bleu-4, respectively. On the in-house Gastric-3300 (3309 WSIs), our method achieved significantly better performance with Rouge-L of 0.690, which surpassed the second-best state-of-the-art method Wcap 6.3%. CONCLUSIONS We present an advanced method for pathology report generation from WSIs, addressing the key challenges associated with the large size and complex pathological structures of these images. In particular, the multi-level regional feature selection module effectively captures diagnostically significant regions of varying sizes. The knowledge retrieval-based decoder leverages historical diagnostic data to enhance report accuracy. Our method not only improves the informativeness and relevance of the generated pathology reports but also outperforms the state-of-the-art techniques.
Collapse
Affiliation(s)
- Dingyi Hu
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Zhiguo Jiang
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China; Tianmushan Laboratory, Hangzhou, 311115, Zhejiang Province, China
| | - Jun Shi
- School of Software, Hefei University of Technology, Hefei, 230601, Anhui Province, China
| | - Fengying Xie
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China; Tianmushan Laboratory, Hangzhou, 311115, Zhejiang Province, China
| | - Kun Wu
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Kunming Tang
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Ming Cao
- Department of Pathology, the First People's Hospital of Wuhu, Wuhu, 241000, Anhui Province, China
| | - Jianguo Huai
- Department of Pathology, the First People's Hospital of Wuhu, Wuhu, 241000, Anhui Province, China
| | - Yushan Zheng
- School of Engineering Medicine, Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, 100191, China.
| |
Collapse
|
2
|
Dang C, Qi Z, Xu T, Gu M, Chen J, Wu J, Lin Y, Qi X. Deep learning-powered whole slide image analysis in cancer pathology. J Transl Med 2025:104186. [PMID: 40306572 DOI: 10.1016/j.labinv.2025.104186] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2024] [Revised: 03/05/2025] [Accepted: 04/22/2025] [Indexed: 05/02/2025] Open
Abstract
Pathology is the cornerstone of modern cancer care. With the advancement of precision oncology, the demand for histopathologic diagnosis and stratification of patients is increasing as personalized cancer therapy relies on accurate biomarker assessment. Recently, rapid development of whole slide imaging technology has enabled digitalization of traditional histological slides at high-resolution, holding promise to improve both the precision and efficiency of histopathological evaluation. In particular, deep learning approaches such as convolutional neural network (CNN) Graph Convolutional Network (GCN) and Transformer have shown great promise in enhancing the sensitivity and accuracy of whole slide image (WSI) analysis in cancer pathology, due to their ability to handle high-dimensional and complex image data. The integration of deep learning models with WSIs enables us to explore and mine morphological features beyond the visual perception of pathologists, which can help automate clinical diagnosis, assess histopathological grade, predict clinical outcomes, and even discover novel morphological biomarkers. In this review, we present a comprehensive framework for incorporating deep learning with WSIs, highlighting how deep learning-driven WSIs analysis advances clinical tasks in cancer care. Furthermore, we critically discuss the opportunities and challenges of translating deep learning-based digital pathology into clinical practice, which should be considered to support personalized treatment of cancer patients.
Collapse
Affiliation(s)
- Chengrun Dang
- School of Chemistry and Life Sciences, Suzhou University of Science and Technology, Suzhou, 215011, China
| | - Zhuang Qi
- Zhuang Qi, School of Software, Shandong University, 1500 ShunHua Road, High Tech Industrial Development Zone, Jinan, 250101, China
| | - Tao Xu
- School of Chemistry and Life Sciences, Suzhou University of Science and Technology, Suzhou, 215011, China
| | - Mingkai Gu
- School of Chemistry and Life Sciences, Suzhou University of Science and Technology, Suzhou, 215011, China
| | - Jiajia Chen
- School of Chemistry and Life Sciences, Suzhou University of Science and Technology, Suzhou, 215011, China
| | - Jie Wu
- Jie Wu, Department of Oncology, The First Affiliated Hospital of Soochow University, Suzhou, Jiangsu, 215000, China.
| | - Yuxin Lin
- Department of Urology, The First Affiliated Hospital of Soochow University, Suzhou, 215000, China.
| | - Xin Qi
- School of Chemistry and Life Sciences, Suzhou University of Science and Technology, Suzhou, 215011, China.
| |
Collapse
|
3
|
Brussee S, Buzzanca G, Schrader AMR, Kers J. Graph neural networks in histopathology: Emerging trends and future directions. Med Image Anal 2025; 101:103444. [PMID: 39793218 DOI: 10.1016/j.media.2024.103444] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2024] [Revised: 11/18/2024] [Accepted: 12/17/2024] [Indexed: 01/13/2025]
Abstract
Histopathological analysis of whole slide images (WSIs) has seen a surge in the utilization of deep learning methods, particularly Convolutional Neural Networks (CNNs). However, CNNs often fail to capture the intricate spatial dependencies inherent in WSIs. Graph Neural Networks (GNNs) present a promising alternative, adept at directly modeling pairwise interactions and effectively discerning the topological tissue and cellular structures within WSIs. Recognizing the pressing need for deep learning techniques that harness the topological structure of WSIs, the application of GNNs in histopathology has experienced rapid growth. In this comprehensive review, we survey GNNs in histopathology, discuss their applications, and explore emerging trends that pave the way for future advancements in the field. We begin by elucidating the fundamentals of GNNs and their potential applications in histopathology. Leveraging quantitative literature analysis, we explore four emerging trends: Hierarchical GNNs, Adaptive Graph Structure Learning, Multimodal GNNs, and Higher-order GNNs. Through an in-depth exploration of these trends, we offer insights into the evolving landscape of GNNs in histopathological analysis. Based on our findings, we propose future directions to propel the field forward. Our analysis serves to guide researchers and practitioners towards innovative approaches and methodologies, fostering advancements in histopathological analysis through the lens of graph neural networks.
Collapse
Affiliation(s)
- Siemen Brussee
- Leiden University Medical Center, Albinusdreef 2, 2333 ZA, Leiden, The Netherlands.
| | - Giorgio Buzzanca
- Leiden University Medical Center, Albinusdreef 2, 2333 ZA, Leiden, The Netherlands
| | - Anne M R Schrader
- Leiden University Medical Center, Albinusdreef 2, 2333 ZA, Leiden, The Netherlands
| | - Jesper Kers
- Leiden University Medical Center, Albinusdreef 2, 2333 ZA, Leiden, The Netherlands; Amsterdam University Medical Center, Meibergdreef 9, 1105 AZ, Amsterdam, The Netherlands
| |
Collapse
|
4
|
Ghezloo F, Chang OH, Knezevich SR, Shaw KC, Thigpen KG, Reisch LM, Shapiro LG, Elmore JG. Robust ROI Detection in Whole Slide Images Guided by Pathologists' Viewing Patterns. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:439-454. [PMID: 39122892 PMCID: PMC11811336 DOI: 10.1007/s10278-024-01202-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Revised: 06/24/2024] [Accepted: 07/05/2024] [Indexed: 08/12/2024]
Abstract
Deep learning techniques offer improvements in computer-aided diagnosis systems. However, acquiring image domain annotations is challenging due to the knowledge and commitment required of expert pathologists. Pathologists often identify regions in whole slide images with diagnostic relevance rather than examining the entire slide, with a positive correlation between the time spent on these critical image regions and diagnostic accuracy. In this paper, a heatmap is generated to represent pathologists' viewing patterns during diagnosis and used to guide a deep learning architecture during training. The proposed system outperforms traditional approaches based on color and texture image characteristics, integrating pathologists' domain expertise to enhance region of interest detection without needing individual case annotations. Evaluating our best model, a U-Net model with a pre-trained ResNet-18 encoder, on a skin biopsy whole slide image dataset for melanoma diagnosis, shows its potential in detecting regions of interest, surpassing conventional methods with an increase of 20%, 11%, 22%, and 12% in precision, recall, F1-score, and Intersection over Union, respectively. In a clinical evaluation, three dermatopathologists agreed on the model's effectiveness in replicating pathologists' diagnostic viewing behavior and accurately identifying critical regions. Finally, our study demonstrates that incorporating heatmaps as supplementary signals can enhance the performance of computer-aided diagnosis systems. Without the availability of eye tracking data, identifying precise focus areas is challenging, but our approach shows promise in assisting pathologists in improving diagnostic accuracy and efficiency, streamlining annotation processes, and aiding the training of new pathologists.
Collapse
Affiliation(s)
- Fatemeh Ghezloo
- Paul G. Allen School of Computer Science and Engineering, University of Washington, Seattle, WA, USA.
| | - Oliver H Chang
- Department of Laboratory Medicine and Pathology, University of Washington, Seattle, WA, USA
| | | | | | | | - Lisa M Reisch
- Department of Biostatistics, University of Washington, Seattle, WA, USA
| | - Linda G Shapiro
- Paul G. Allen School of Computer Science and Engineering, University of Washington, Seattle, WA, USA
| | - Joann G Elmore
- Department of Medicine, David Geffen School of Medicine, University of California, Los AngelesLos Angeles, CA, USA
| |
Collapse
|
5
|
Liu L, Xie J, Chang J, Liu Z, Sun T, Qiao H, Liang G, Guo W. H-Net: Heterogeneous Neural Network for Multi-Classification of Neuropsychiatric Disorders. IEEE J Biomed Health Inform 2024; 28:5509-5518. [PMID: 38829757 DOI: 10.1109/jbhi.2024.3405941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/05/2024]
Abstract
Clinical studies have proved that both structural magnetic resonance imaging (sMRI) and functional magnetic resonance imaging (fMRI) are implicitly associated with neuropsychiatric disorders (NDs), and integrating multi-modal to the binary classification of NDs has been thoroughly explored. However, accurately classifying multiple classes of NDs remains a challenge due to the complexity of disease subclass. In our study, we develop a heterogeneous neural network (H-Net) that integrates sMRI and fMRI modes for classifying multi-class NDs. To account for the differences between the two modes, H-Net adopts a heterogeneous neural network strategy to extract information from each mode. Specifically, H-Net includes an multi-layer perceptron based (MLP-based) encoder, a graph attention network based (GAT-based) encoder, and a cross-modality transformer block. The MLP-based and GAT-based encoders extract semantic features from sMRI and features from fMRI, respectively, while the cross-modality transformer block models the attention of two types of features. In H-Net, the proposed MLP-mixer block and cross-modality alignment are powerful tools for improving the multi-classification performance of NDs. H-Net is validate on the public dataset (CNP), where H-Net achieves 90% classification accuracy in diagnosing multi-class NDs. Furthermore, we demonstrate the complementarity of the two MRI modalities in improving the identification of multi-class NDs. Both visual and statistical analyses show the differences between ND subclasses.
Collapse
|
6
|
Hu D, Jiang Z, Shi J, Xie F, Wu K, Tang K, Cao M, Huai J, Zheng Y. Histopathology language-image representation learning for fine-grained digital pathology cross-modal retrieval. Med Image Anal 2024; 95:103163. [PMID: 38626665 DOI: 10.1016/j.media.2024.103163] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2023] [Revised: 03/09/2024] [Accepted: 04/02/2024] [Indexed: 04/18/2024]
Abstract
Large-scale digital whole slide image (WSI) datasets analysis have gained significant attention in computer-aided cancer diagnosis. Content-based histopathological image retrieval (CBHIR) is a technique that searches a large database for data samples matching input objects in both details and semantics, offering relevant diagnostic information to pathologists. However, the current methods are limited by the difficulty of gigapixels, the variable size of WSIs, and the dependence on manual annotations. In this work, we propose a novel histopathology language-image representation learning framework for fine-grained digital pathology cross-modal retrieval, which utilizes paired diagnosis reports to learn fine-grained semantics from the WSI. An anchor-based WSI encoder is built to extract hierarchical region features and a prompt-based text encoder is introduced to learn fine-grained semantics from the diagnosis reports. The proposed framework is trained with a multivariate cross-modal loss function to learn semantic information from the diagnosis report at both the instance level and region level. After training, it can perform four types of retrieval tasks based on the multi-modal database to support diagnostic requirements. We conducted experiments on an in-house dataset and a public dataset to evaluate the proposed method. Extensive experiments have demonstrated the effectiveness of the proposed method and its advantages to the present histopathology retrieval methods. The code is available at https://github.com/hudingyi/FGCR.
Collapse
Affiliation(s)
- Dingyi Hu
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Engineering Medicine, Beihang University, Beijing, 100191, China; Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Zhiguo Jiang
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Engineering Medicine, Beihang University, Beijing, 100191, China; Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Jun Shi
- School of Software, Hefei University of Technology, Hefei 230601, China
| | - Fengying Xie
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Engineering Medicine, Beihang University, Beijing, 100191, China; Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Kun Wu
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Engineering Medicine, Beihang University, Beijing, 100191, China; Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Kunming Tang
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Engineering Medicine, Beihang University, Beijing, 100191, China; Image Processing Center, School of Astronautics, Beihang University, Beijing, 100191, China
| | - Ming Cao
- Department of Pathology, the First People's Hospital of Wuhu, China
| | - Jianguo Huai
- Department of Pathology, the First People's Hospital of Wuhu, China
| | - Yushan Zheng
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Engineering Medicine, Beihang University, Beijing, 100191, China.
| |
Collapse
|
7
|
Jia W, Shi W, Yao Q, Mao Z, Chen C, Fan AQ, Wang Y, Zhao Z, Li J, Song W. Identifying immune infiltration by deep learning to assess the prognosis of patients with hepatocellular carcinoma. J Cancer Res Clin Oncol 2023; 149:12621-12635. [PMID: 37450030 DOI: 10.1007/s00432-023-05097-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2023] [Accepted: 06/30/2023] [Indexed: 07/18/2023]
Abstract
BACKGROUND The treatment situation for hepatocellular carcinoma remains critical. The use of deep learning algorithms to assess immune infiltration is a promising new diagnostic tool. METHODS Patient data and whole slide images (WSIs) were obtained for the Xijing Hospital (XJH) cohort and TCGA cohort. We wrote programs using Visual studio 2022 with C# language to segment the WSI into tiles. Pathologists classified the tiles and later trained deep learning models using the ResNet 101V2 network via ML.NET with the TensorFlow framework. Model performance was evaluated using AccuracyMicro versus AccuracyMacro. Model performance was examined using ROC curves versus PR curves. The percentage of immune infiltration was calculated using the R package survminer to calculate the intergroup cutoff, and the Kaplan‒Meier method was used to plot the overall survival curve of patients. Cox regression was used to determine whether the percentage of immune infiltration was an independent risk factor for prognosis. A nomogram was constructed, and its accuracy was verified using time-dependent ROC curves with calibration curves. The CIBERSORT algorithm was used to assess immune infiltration between groups. Gene Ontology was used to explore the pathways of differentially expressed genes. RESULTS There were 100 WSIs and 165,293 tiles in the training set. The final deep learning models had an AccuracyMicro of 97.46% and an AccuracyMacro of 82.28%. The AUCs of the ROC curves on both the training and validation sets exceeded 0.95. The areas under the classification PR curves exceeded 0.85, except that of the TLS on the validation set, which might have had poor results (0.713) due to too few samples. There was a significant difference in OS between the TIL classification groups (p < 0.001), while there was no significant difference in OS between the TLS groups (p = 0.294). Cox regression showed that TIL percentage was an independent risk factor for prognosis in HCC patients (p = 0.015). The AUCs according to the nomogram were 0.714, 0.690, and 0.676 for the 1-year, 2-year, and 5-year AUCs in the TCGA cohort and 0.756, 0.797, and 0.883 in the XJH cohort, respectively. There were significant differences in the levels of infiltration of seven immune cell types between the two groups of samples, and gene ontology showed that the differentially expressed genes between the groups were immune related. Their expression levels of PD-1 and CTLA4 were also significantly different. CONCLUSION We constructed and tested a deep learning model that evaluates the immune infiltration of liver cancer tissue in HCC patients. Our findings demonstrate the value of the model in assessing patient prognosis, immune infiltration and immune checkpoint expression levels.
Collapse
Affiliation(s)
- Weili Jia
- Xi'an Medical University, Xi'an, China
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | - Wen Shi
- Xi'an Medical University, Xi'an, China
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | | | - Zhenzhen Mao
- Xi'an Medical University, Xi'an, China
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | - Chao Chen
- Xi'an Medical University, Xi'an, China
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | - AQiang Fan
- Xi'an Medical University, Xi'an, China
- Department of Gastrointestinal Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | - Yanfang Wang
- Xi'an Medical University, Xi'an, China
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | - Zihao Zhao
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China
| | - Jipeng Li
- Department of Gastrointestinal Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China.
| | - Wenjie Song
- Department of Hepatobiliary Surgery, Xijing Hospital, Fourth Military Medical University, Xi'an, China.
| |
Collapse
|
8
|
Meng X, Zou T. Clinical applications of graph neural networks in computational histopathology: A review. Comput Biol Med 2023; 164:107201. [PMID: 37517325 DOI: 10.1016/j.compbiomed.2023.107201] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2023] [Revised: 06/10/2023] [Accepted: 06/19/2023] [Indexed: 08/01/2023]
Abstract
Pathological examination is the optimal approach for diagnosing cancer, and with the advancement of digital imaging technologies, it has spurred the emergence of computational histopathology. The objective of computational histopathology is to assist in clinical tasks through image processing and analysis techniques. In the early stages, the technique involved analyzing histopathology images by extracting mathematical features, but the performance of these models was unsatisfactory. With the development of artificial intelligence (AI) technologies, traditional machine learning methods were applied in this field. Although the performance of the models improved, there were issues such as poor model generalization and tedious manual feature extraction. Subsequently, the introduction of deep learning techniques effectively addressed these problems. However, models based on traditional convolutional architectures could not adequately capture the contextual information and deep biological features in histopathology images. Due to the special structure of graphs, they are highly suitable for feature extraction in tissue histopathology images and have achieved promising performance in numerous studies. In this article, we review existing graph-based methods in computational histopathology and propose a novel and more comprehensive graph construction approach. Additionally, we categorize the methods and techniques in computational histopathology according to different learning paradigms. We summarize the common clinical applications of graph-based methods in computational histopathology. Furthermore, we discuss the core concepts in this field and highlight the current challenges and future research directions.
Collapse
Affiliation(s)
- Xiangyan Meng
- Xi'an Technological University, Xi'an, Shaanxi, 710021, China.
| | - Tonghui Zou
- Xi'an Technological University, Xi'an, Shaanxi, 710021, China.
| |
Collapse
|
9
|
Zheng T, Chen W, Li S, Quan H, Zou M, Zheng S, Zhao Y, Gao X, Cui X. Learning how to detect: A deep reinforcement learning method for whole-slide melanoma histopathology images. Comput Med Imaging Graph 2023; 108:102275. [PMID: 37567046 DOI: 10.1016/j.compmedimag.2023.102275] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2023] [Revised: 07/18/2023] [Accepted: 07/22/2023] [Indexed: 08/13/2023]
Abstract
Cutaneous melanoma represents one of the most life-threatening malignancies. Histopathological image analysis serves as a vital tool for early melanoma detection. Deep neural network (DNN) models are frequently employed to aid pathologists in enhancing the efficiency and accuracy of diagnoses. However, due to the paucity of well-annotated, high-resolution, whole-slide histopathology image (WSI) datasets, WSIs are typically fragmented into numerous patches during the model training and testing stages. This process disregards the inherent interconnectedness among patches, potentially impeding the models' performance. Additionally, the presence of excess, non-contributing patches extends processing times and introduces substantial computational burdens. To mitigate these issues, we draw inspiration from the clinical decision-making processes of dermatopathologists to propose an innovative, weakly supervised deep reinforcement learning framework, titled Fast medical decision-making in melanoma histopathology images (FastMDP-RL). This framework expedites model inference by reducing the number of irrelevant patches identified within WSIs. FastMDP-RL integrates two DNN-based agents: the search agent (SeAgent) and the decision agent (DeAgent). The SeAgent initiates actions, steered by the image features observed in the current viewing field at various magnifications. Simultaneously, the DeAgent provides labeling probabilities for each patch. We utilize multi-instance learning (MIL) to construct a teacher-guided model (MILTG), serving a dual purpose: rewarding the SeAgent and guiding the DeAgent. Our evaluations were conducted using two melanoma datasets: the publicly accessible TCIA-CM dataset and the proprietary MELSC dataset. Our experimental findings affirm FastMDP-RL's ability to expedite inference and accurately predict WSIs, even in the absence of pixel-level annotations. Moreover, our research investigates the WSI-based interactive environment, encompassing the design of agents, state and reward functions, and feature extractors suitable for melanoma tissue images. This investigation offers valuable insights and references for researchers engaged in related studies. The code is available at: https://github.com/titizheng/FastMDP-RL.
Collapse
Affiliation(s)
- Tingting Zheng
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Weixing Chen
- Shenzhen College of Advanced Technology, University of the Chinese Academy of Sciences, Beijing, China
| | - Shuqin Li
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Hao Quan
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Mingchen Zou
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Song Zheng
- National and Local Joint Engineering Research Center of Immunodermatological Theranostics, Department of Dermatology, The First Hospital of China Medical University, Shenyang, China
| | - Yue Zhao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; National and Local Joint Engineering Research Center of Immunodermatological Theranostics, Department of Dermatology, The First Hospital of China Medical University, Shenyang, China
| | - Xinghua Gao
- National and Local Joint Engineering Research Center of Immunodermatological Theranostics, Department of Dermatology, The First Hospital of China Medical University, Shenyang, China
| | - Xiaoyu Cui
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China.
| |
Collapse
|
10
|
Zheng Y, Li J, Shi J, Xie F, Huai J, Cao M, Jiang Z. Kernel Attention Transformer for Histopathology Whole Slide Image Analysis and Assistant Cancer Diagnosis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2726-2739. [PMID: 37018112 DOI: 10.1109/tmi.2023.3264781] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Transformer has been widely used in histopathology whole slide image analysis. However, the design of token-wise self-attention and positional embedding strategy in the common Transformer limits its effectiveness and efficiency when applied to gigapixel histopathology images. In this paper, we propose a novel kernel attention Transformer (KAT) for histopathology WSI analysis and assistant cancer diagnosis. The information transmission in KAT is achieved by cross-attention between the patch features and a set of kernels related to the spatial relationship of the patches on the whole slide images. Compared to the common Transformer structure, KAT can extract the hierarchical context information of the local regions of the WSI and provide diversified diagnosis information. Meanwhile, the kernel-based cross-attention paradigm significantly reduces the computational amount. The proposed method was evaluated on three large-scale datasets and was compared with 8 state-of-the-art methods. The experimental results have demonstrated the proposed KAT is effective and efficient in the task of histopathology WSI analysis and is superior to the state-of-the-art methods.
Collapse
|
11
|
Liu L, Wang Y, Chang J, Zhang P, Xiong S, Liu H. A correlation graph attention network for classifying chromosomal instabilities from histopathology whole-slide images. iScience 2023; 26:106874. [PMID: 37260749 PMCID: PMC10227422 DOI: 10.1016/j.isci.2023.106874] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2022] [Revised: 02/23/2023] [Accepted: 05/10/2023] [Indexed: 06/02/2023] Open
Abstract
The chromosome instability (CIN) is one of the hallmarks of cancer and is closely related to tumor metastasis. However, the sheer size and resolution of histopathology whole-slide images (WSIs) already challenges the capabilities of computational pathology. In this study, we propose a correlation graph attention network (MLP-GAT) that can construct graphs for classifying multi-type CINs from the WSIs of breast cancer. We construct a WSIs dataset of breast cancer from the Cancer Genome Atlas Breast Invasive Carcinoma (TCGA-BRCA). Extensive experiments show that MLP-GAT far outperforms accepted state-of-the-art methods and demonstrate the advantages of the constructed graph networks for analyzing WSI data. The visualization shows the difference among the tiles in a WSI. Furthermore, the generalization performance of the proposed method was verified on the stomach cancer. This study provides guidance for studying the relationship between CIN and cancer from the perspective of image phenotype.
Collapse
Affiliation(s)
- Liangliang Liu
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan 450046, P.R. China
| | - Ying Wang
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan 450046, P.R. China
| | - Jing Chang
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan 450046, P.R. China
| | - Pei Zhang
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan 450046, P.R. China
| | - Shufeng Xiong
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan 450046, P.R. China
| | - Hebing Liu
- College of Information and Management Science, Henan Agricultural University, Zhengzhou, Henan 450046, P.R. China
| |
Collapse
|
12
|
Yu J, Ma T, Fu Y, Chen H, Lai M, Zhuo C, Xu Y. Local-to-global spatial learning for whole-slide image representation and classification. Comput Med Imaging Graph 2023; 107:102230. [PMID: 37116341 DOI: 10.1016/j.compmedimag.2023.102230] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2022] [Revised: 03/27/2023] [Accepted: 04/05/2023] [Indexed: 04/30/2023]
Abstract
Whole-slide image (WSI) provides an important reference for clinical diagnosis. Classification with only WSI-level labels can be recognized for multi-instance learning (MIL) tasks. However, most existing MIL-based WSI classification methods have moderate performance on correlation mining between instances limited by their instance- level classification strategy. Herein, we propose a novel local-to-global spatial learning method to mine global position and local morphological information by redefining the MIL-based WSI classification strategy, better at learning WSI-level representation, called Global-Local Attentional Multi-Instance Learning (GLAMIL). GLAMIL can focus on regional relationships rather than single instances. It first learns relationships between patches in the local pool to aggregate region correlation (tissue types of a WSI). These correlations then can be further mined to fulfill WSI-level representation, where position correlation between different regions can be modeled. Furthermore, Transformer layers are employed to model global and local spatial information rather than being simply used as feature extractors, and the corresponding structure improvements are present. In addition, we evaluate GIAMIL on three benchmarks considering various challenging factors and achieve satisfactory results. GLAMIL outperforms state-of-the-art methods and baselines by about 1 % and 10 %, respectively.
Collapse
Affiliation(s)
- Jiahui Yu
- Department of Biomedical Enginearing, Key Laboratory of Biomedical Engineering of Ministry of Education, State Key Laboratory of Modern Optical Instrumentation, Zhejiang Provincial Key Laboratory of Cardio-Cerebral Vascular Detection Technology and Medicinal Effectiveness Appraisal, Zhejiang University, Hangzhou 310027, China; Innovation Center for Smart Medical Technologies & Devices, Binjiang Institute of Zhejiang University, Hangzhou 310053, China
| | - Tianyu Ma
- Innovation Center for Smart Medical Technologies & Devices, Binjiang Institute of Zhejiang University, Hangzhou 310053, China
| | - Yu Fu
- College of Information Science & Electronic Engineering, Zhejiang University, Hangzhou 310027, China
| | - Hang Chen
- Department of Biomedical Enginearing, Key Laboratory of Biomedical Engineering of Ministry of Education, State Key Laboratory of Modern Optical Instrumentation, Zhejiang Provincial Key Laboratory of Cardio-Cerebral Vascular Detection Technology and Medicinal Effectiveness Appraisal, Zhejiang University, Hangzhou 310027, China
| | - Maode Lai
- Department of Pathology, School of Medicine, Zhejiang University, Hangzhou 310053, China
| | - Cheng Zhuo
- College of Information Science & Electronic Engineering, Zhejiang University, Hangzhou 310027, China
| | - Yingke Xu
- Department of Biomedical Enginearing, Key Laboratory of Biomedical Engineering of Ministry of Education, State Key Laboratory of Modern Optical Instrumentation, Zhejiang Provincial Key Laboratory of Cardio-Cerebral Vascular Detection Technology and Medicinal Effectiveness Appraisal, Zhejiang University, Hangzhou 310027, China; Innovation Center for Smart Medical Technologies & Devices, Binjiang Institute of Zhejiang University, Hangzhou 310053, China; Department of Endocrinology, Children's Hospital of Zhejiang University School of Medicine, National Clinical Research Center for Children's Health, Hangzhou, Zhejiang 310051, China.
| |
Collapse
|
13
|
Hashimoto N, Takagi Y, Masuda H, Miyoshi H, Kohno K, Nagaishi M, Sato K, Takeuchi M, Furuta T, Kawamoto K, Yamada K, Moritsubo M, Inoue K, Shimasaki Y, Ogura Y, Imamoto T, Mishina T, Tanaka K, Kawaguchi Y, Nakamura S, Ohshima K, Hontani H, Takeuchi I. Case-based similar image retrieval for weakly annotated large histopathological images of malignant lymphoma using deep metric learning. Med Image Anal 2023; 85:102752. [PMID: 36716701 DOI: 10.1016/j.media.2023.102752] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2021] [Revised: 12/29/2022] [Accepted: 01/18/2023] [Indexed: 01/26/2023]
Abstract
In the present study, we propose a novel case-based similar image retrieval (SIR) method for hematoxylin and eosin (H&E) stained histopathological images of malignant lymphoma. When a whole slide image (WSI) is used as an input query, it is desirable to be able to retrieve similar cases by focusing on image patches in pathologically important regions such as tumor cells. To address this problem, we employ attention-based multiple instance learning, which enables us to focus on tumor-specific regions when the similarity between cases is computed. Moreover, we employ contrastive distance metric learning to incorporate immunohistochemical (IHC) staining patterns as useful supervised information for defining appropriate similarity between heterogeneous malignant lymphoma cases. In the experiment with 249 malignant lymphoma patients, we confirmed that the proposed method exhibited higher evaluation measures than the baseline case-based SIR methods. Furthermore, the subjective evaluation by pathologists revealed that our similarity measure using IHC staining patterns is appropriate for representing the similarity of H&E stained tissue images for malignant lymphoma.
Collapse
Affiliation(s)
- Noriaki Hashimoto
- RIKEN Center for Advanced Intelligence Project, 1-4-1 Nihonbashi, Chuo-ku, Tokyo 103-0027, Japan
| | - Yusuke Takagi
- Department of Computer Science, Nagoya Institute of Technology, Gokiso-cho, Showa-ku, Nagoya 466-8555, Japan
| | - Hiroki Masuda
- Department of Computer Science, Nagoya Institute of Technology, Gokiso-cho, Showa-ku, Nagoya 466-8555, Japan
| | - Hiroaki Miyoshi
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Kei Kohno
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Miharu Nagaishi
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Kensaku Sato
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Mai Takeuchi
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Takuya Furuta
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Keisuke Kawamoto
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Kyohei Yamada
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Mayuko Moritsubo
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Kanako Inoue
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Yasumasa Shimasaki
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Yusuke Ogura
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Teppei Imamoto
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Tatsuzo Mishina
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Ken Tanaka
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Yoshino Kawaguchi
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Shigeo Nakamura
- Department of Pathology and Laboratory Medicine, Nagoya University Hospital, 65 Tsurumai-cho, Showa-ku, Nagoya 466-8560, Japan
| | - Koichi Ohshima
- Department of Pathology, Kurume University School of Medicine, 67 Asahimachi, Kurume 830-0011, Japan
| | - Hidekata Hontani
- Department of Computer Science, Nagoya Institute of Technology, Gokiso-cho, Showa-ku, Nagoya 466-8555, Japan
| | - Ichiro Takeuchi
- RIKEN Center for Advanced Intelligence Project, 1-4-1 Nihonbashi, Chuo-ku, Tokyo 103-0027, Japan; Department of Mechanical Systems Engineering, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8603, Japan.
| |
Collapse
|
14
|
Li X, Jiang Y, Liu Y, Zhang J, Yin S, Luo H. RAGCN: Region Aggregation Graph Convolutional Network for Bone Age Assessment From X-Ray Images. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT 2022; 71:1-12. [DOI: 10.1109/tim.2022.3190025] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/03/2024]
Affiliation(s)
- Xiang Li
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Yuchen Jiang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Yiliu Liu
- Department of Mechanical and Industrial Engineering, Faculty of Engineering, Norwegian University of Science and Technology, Trondheim, Norway
| | - Jiusi Zhang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Shen Yin
- Department of Mechanical and Industrial Engineering, Faculty of Engineering, Norwegian University of Science and Technology, Trondheim, Norway
| | - Hao Luo
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| |
Collapse
|