Concept-Enhanced Relation Network for Video Visual Relation Inference

Qianwen Cao, Heyan Huang*, Mucheng Ren, Changsen Yuan

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

1 引用 (Scopus)

摘要

Video visual relation inference aims at extracting the relation triplets in the form of < subject-predicate-object > in videos. With the development of deep learning, existing approaches are designed based on data-driven neural networks. But the datasets are always biased in terms of objects and relation triplets, which make relation inference challenging. Existing approaches often describe the relationships from visual, spatial, and semantic characteristics. The semantic description plays a key role to indicate the potential linguistic connections between objects, that are crucial to transfer knowledge across relationships, especially for the determination of novel relations. However, in these works, the semantic features are not emphasized, but simply obtained by mapping object labels, which can not reflect sufficient linguistic meanings. To alleviate the above issues, we propose a novel network, termed Concept-Enhanced Relation Network (CERN), to facilitate video visual relation inference. Thanks to the attributes and linguistic contexts implied in concepts, the semantic representations aggregated with related concept knowledge of objects are of benefit to relation inference. To this end, we incorporate retrieved concepts with local semantics of objects via the gating mechanism to generate the concept-enhanced semantic representations. Extensive experimental results show that our approach has achieved state-of-the-art performance on two public datasets: ImageNet-VidVRD and VidOR.

源语言英语
页(从-至)2233-2244
页数12
期刊IEEE Transactions on Circuits and Systems for Video Technology
33
5
DOI
出版状态已出版 - 1 5月 2023

指纹

探究 'Concept-Enhanced Relation Network for Video Visual Relation Inference' 的科研主题。它们共同构成独一无二的指纹。

引用此