Content-Attention Representation by Factorized Action-Scene Network for Action Recognition

Jingyi Hou, Xinxiao Wu, Yuchao Sun, Yunde Jia*

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

44 引用 (Scopus)

摘要

During action recognition in videos, irrelevant motions in the background can greatly degrade the performance of recognizing specific actions with which we actually concern ourself here. In this paper, a novel deep neural network, called factorized action-scene network (FASNet), is proposed to encode and fuse the most relevant and informative semantic cues for action recognition. Specifically, we decompose the FASNet into two components. One is a newly designed encoding network, named content attention network (CANet), which encodes local spatialoral features to learn the action representations with good robustness to the noise of irrelevant motions. The other is a fusion network, which integrates the pretrained CANet to fuse the encoded spatialoral features with contextual scene feature extracted from the same video, for learning more descriptive and discriminative action representations. Moreover, different from the existing deep learning based tasks for generic action recognition, which applies softmax loss function as the training guidance, we formulate two loss functions for guiding the proposed model to accomplish more specific action recognition tasks, i.e., the multilabel correlation loss for multilabel action recognition and the triplet loss for complex event detection. Extensive experiments on the Hollywood2 dataset and the TRECVID MEDTest 14 dataset show that our method achieves superior performance compared with the state-of-the-art methods.

源语言英语
页(从-至)1537-1547
页数11
期刊IEEE Transactions on Multimedia
20
6
DOI
出版状态已出版 - 6月 2018

指纹

探究 'Content-Attention Representation by Factorized Action-Scene Network for Action Recognition' 的科研主题。它们共同构成独一无二的指纹。

引用此