ConvD: Attention Enhanced Dynamic Convolutional Embeddings for Knowledge Graph Completion

  • Wenbin Guo
  • , Zhao Li
  • , Xin Wang*
  • , Zirui Chen
  • , Jun Zhao
  • , Jianxin Li
  • , Ye Yuan
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Knowledge graphs often suffer from incompleteness issues, which can be alleviated through information completion. However, current state-of-the-art deep knowledge convolutional embedding models rely on external convolution kernels and conventional convolution processes, which limits the feature interaction capability of the model. This paper introduces a novel dynamic convolutional embedding model, named ConvD, which directly reshapes relation embeddings into multiple internal convolution kernels. This approach effectively enhances the feature interactions between relation embeddings and entity embeddings. Simultaneously, we incorporate a priori knowledge-optimized attention mechanism that assigns distinct contribution weights to multiple relational convolution kernels during dynamic convolution, further boosting the expressive power of the model. Extensive experiments on various datasets show that our proposed model consistently outperforms the state-of-the-art baseline methods, with average improvements ranging from 3.28% to 14.69% across all the evaluation metrics, while the number of parameters is reduced by 50.66% to 85.40% compared to other state-of-the-art models.

Original languageEnglish
Pages (from-to)5049-5062
Number of pages14
JournalIEEE Transactions on Knowledge and Data Engineering
Volume37
Issue number9
DOIs
Publication statusPublished - 2025

Keywords

  • Knowledge graph
  • knowledge graph completion
  • knowledge representation learning
  • link prediction

Fingerprint

Dive into the research topics of 'ConvD: Attention Enhanced Dynamic Convolutional Embeddings for Knowledge Graph Completion'. Together they form a unique fingerprint.

Cite this