Shared Growth of Graph Neural Networks via Prompted Free-Direction Knowledge Distillation

Kaituo Feng, Yikun Miao, Changsheng Li*, Ye Yuan, Guoren Wang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Knowledge distillation (KD) has shown to be effective to boost the performance of graph neural networks (GNNs), where the typical objective is to distill knowledge from a deeper teacher GNN into a shallower student GNN. However, it is often quite challenging to train a satisfactory deeper GNN due to the well-known over-parametrized and over-smoothing issues, leading to invalid knowledge transfer in practical applications. In this paper, we propose the first Free-direction Knowledge Distillation framework via reinforcement learning for GNNs, called FreeKD, which is no longer required to provide a deeper well-optimized teacher GNN. Our core idea is to collaboratively learn two shallower GNNs in an effort to exchange knowledge between them via reinforcement learning in a hierarchical way. As we observe that one typical GNN model often exhibits better and worse performances at different nodes during training, we devise a dynamic and free-direction knowledge transfer strategy that involves two levels of actions: 1) node-level action determines the directions of knowledge transfer between the corresponding nodes of two networks; and then 2) structure-level action determines which of the local structures generated by the node-level actions to be propagated. Additionally, considering that different augmented graphs can potentially capture distinct perspectives or representations of the graph data, we propose FreeKD-Prompt that learns undistorted and diverse augmentations based on prompt learning for exchanging varied knowledge. Furthermore, instead of confining knowledge exchange within two GNNs, we develop FreeKD++ and FreeKD-Prompt++ to enable free-direction knowledge transfer among multiple shallow GNNs. Extensive experiments on five benchmark datasets demonstrate our approaches outperform the base GNNs by a large margin, and show their efficacy to various GNNs. More surprisingly, our FreeKD has comparable or even better performance than traditional KD algorithms that distill knowledge from a deeper and stronger teacher GNN.

Original languageEnglish
Pages (from-to)4377-4394
Number of pages18
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume47
Issue number6
DOIs
Publication statusPublished - 2025
Externally publishedYes

Keywords

  • free-direction knowledge distillation
  • Graph neural networks
  • prompt learning
  • reinforcement learning

Fingerprint

Dive into the research topics of 'Shared Growth of Graph Neural Networks via Prompted Free-Direction Knowledge Distillation'. Together they form a unique fingerprint.

Cite this