Sparse and Hierarchical Transformer for Survival Analysis on Whole Slide Images

Rui Yan, Zhilong Lv, Zhidong Yang, Senlin Lin, Chunhou Zheng, Fa Zhang*

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

1 引用 (Scopus)
Plum Print visual indicator of research metrics
  • Citations
    • Citation Indexes: 1
  • Captures
    • Readers: 8
see details

摘要

The Transformer-based methods provide a good opportunity for modeling the global context of gigapixel whole slide image (WSI), however, there are still two main problems in applying Transformer to WSI-based survival analysis task. First, the training data for survival analysis is limited, which makes the model prone to overfitting. This problem is even worse for Transformer-based models which require large-scale data to train. Second, WSI is of extremely high resolution (up to 150,000 × 150,000 pixels) and is typically organized as a multi-resolution pyramid. Vanilla Transformer cannot model the hierarchical structure of WSI (such as patch cluster-level relationships), which makes it incapable of learning hierarchical WSI representation. To address these problems, in this article, we propose a novel Sparse and Hierarchical Transformer (SH-Transformer) for survival analysis. Specifically, we introduce sparse self-attention to alleviate the overfitting problem, and propose a hierarchical Transformer structure to learn the hierarchical WSI representation. Experimental results based on three WSI datasets show that the proposed framework outperforms the state-of-the-art methods.

源语言英语
页(从-至)7-18
页数12
期刊IEEE Journal of Biomedical and Health Informatics
28
1
DOI
出版状态已出版 - 1 1月 2024

指纹

探究 'Sparse and Hierarchical Transformer for Survival Analysis on Whole Slide Images' 的科研主题。它们共同构成独一无二的指纹。

引用此

Yan, R., Lv, Z., Yang, Z., Lin, S., Zheng, C., & Zhang, F. (2024). Sparse and Hierarchical Transformer for Survival Analysis on Whole Slide Images. IEEE Journal of Biomedical and Health Informatics, 28(1), 7-18. https://doi.org/10.1109/JBHI.2023.3307584