Transfer Latent SVM for joint recognition and localization of actions in videos

Cuiwei Liu, Xinxiao Wu*, Yunde Jia

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

13 引用 (Scopus)

摘要

In this paper, we develop a novel transfer latent support vector machine for joint recognition and localization of actions by using Web images and weakly annotated training videos. The model takes training videos which are only annotated with action labels as input for alleviating the laborious and time-consuming manual annotations of action locations. Since the ground-Truth of action locations in videos are not available, the locations are modeled as latent variables in our method and are inferred during both training and testing phrases. For the purpose of improving the localization accuracy with some prior information of action locations, we collect a number of Web images which are annotated with both action labels and action locations to learn a discriminative model by enforcing the local similarities between videos and Web images. A structural transformation based on randomized clustering forest is used to map the Web images to videos for handling the heterogeneous features of Web images and videos. Experiments on two public action datasets demonstrate the effectiveness of the proposed model for both action localization and action recognition.

源语言英语
文章编号7299283
页(从-至)2596-2608
页数13
期刊IEEE Transactions on Cybernetics
46
10
DOI
出版状态已出版 - 15 10月 2015

指纹

探究 'Transfer Latent SVM for joint recognition and localization of actions in videos' 的科研主题。它们共同构成独一无二的指纹。

引用此