摘要
The key issue for cross-modal retrieval using cross-modal Hashing is how to maximize the consistency of the semantic relationship for heterogeneous media data. This paper presents a self-supervised deep semantics-preserving hashing network (UDSPH) that generates compact Hash codes using an end-to-end architecture. Two modality-specific hashing networks are first trained for generating the Hash codes and high-level features. The semantic relationship hetween different modalities is then measured using cross-modal attention mechanisms that maximize preservation of the local semantic correlation. Multi-label semantic information in the training data is used to simultaneously guide the training of two modality-specific Hashing networks by self-supervised adversarial learning. This constructs a deep semantic hashing network that preserves the semantic association in the global view and improves the discriminative capability of the generated Hash codes. Tests on three widely-used benchmark datasets verify the effectiveness of this method.
投稿的翻译标题 | Self-supervised deep semantics-preserving Hashing for cross-modal retrieval |
---|---|
源语言 | 繁体中文 |
页(从-至) | 1442-1449 |
页数 | 8 |
期刊 | Qinghua Daxue Xuebao/Journal of Tsinghua University |
卷 | 62 |
期 | 9 |
DOI | |
出版状态 | 已出版 - 15 9月 2022 |
关键词
- adversarial learning
- cross-modal attention
- deep cross-modal Hashing
- semantic Hashing