Adaptive learning for celebrity identification with video context

Chao Xiong, Guangyu Gao*, Zhengjun Zha, Shuicheng Yan, Huadong Ma, Tae Kyun Kim

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

7 Citations (Scopus)

Abstract

In this paper, we propose a novel semi-supervised learning strategy to address the problem of celebrity identification. The video context information is explored to facilitate the learning process based on the assumption that faces in the same video track share the same identity. Once a frame within a track is recognized confidently, the label can be propagated through the whole track, referred to as the confident track. More specifically, given a few static images and vast face videos, an initial weak classifier is trained and gradually evolves by iteratively promoting the confident tracks into the 'labeled' set. The iterative selection process enriches the diversity of the 'labeled' set such that the performance of the classifier is gradually improved. This learning theme may suffer from semantic drifting caused by errors in selecting the confident tracks. To address this issue, we propose to treat the selected frames as related samples - an intermediate state between labeled and unlabeled instead of labeled as in the traditional approach. To evaluate the performance, we construct a new dataset, which includes 3000 static images and 2700 face tracks of 30 celebrities. Comprehensive evaluations on this dataset and a public video dataset indicate significant improvement of our approach over established baseline methods.

Original languageEnglish
Article number6786434
Pages (from-to)1473-1485
Number of pages13
JournalIEEE Transactions on Multimedia
Volume16
Issue number5
DOIs
Publication statusPublished - Aug 2014

Keywords

  • Adaptive learning
  • celebrity identification
  • related samples
  • semi-supervised learning
  • video context

Fingerprint

Dive into the research topics of 'Adaptive learning for celebrity identification with video context'. Together they form a unique fingerprint.

Cite this