Tree enhanced deep adaptive network for cancer prediction with high dimension low sample size microarray data

Yao Wu, Donghua Zhu, Xuefeng Wang*

*Corresponding author for this work

    Research output: Contribution to journalArticlepeer-review

    3 Citations (Scopus)

    Abstract

    Cancer prediction based on microarray data can facilitate the molecular exploration of cancers, thus building more accurate cancer prediction models is essential. This study focuses on a deep learning-based cancer prediction model. However, using a deep neural network to predict cancer is a difficult task due to the complexity of the underlying biological patterns and high dimension low sample size (HDLSS) of microarray data, which could bring about over-fitting and large training gradient variance. Therefore, a tree-enhanced deep adaptive network (TEDAN) is proposed to address these issues. Firstly, we employ the idea of the ensemble tree as a feature transformation method to alleviate the over-fitting problem, which generates a feature with a lower dimension and a more discriminative pattern. Secondly, a deep adaptive network (DAN) based on a self-attention mechanism is proposed to model the underlying biological interaction between different genes. Thirdly, a low sample size training (LSST) method is proposed to further reduce the large training gradient variance. Experiment results on six public cancer prediction datasets demonstrate that the TEDAN outperforms other strong baseline models.

    Original languageEnglish
    Article number110078
    JournalApplied Soft Computing
    Volume136
    DOIs
    Publication statusPublished - Mar 2023

    Keywords

    • Cancer prediction
    • Deep learning
    • Feature transformation
    • High dimension low sample size (HDLSS)
    • Tree-enhanced deep adaptive network (TEDAN)

    Fingerprint

    Dive into the research topics of 'Tree enhanced deep adaptive network for cancer prediction with high dimension low sample size microarray data'. Together they form a unique fingerprint.

    Cite this