TY - GEN
T1 - Human action recognition from boosted pose estimation
AU - Li, Wang
AU - Li, Cheng
AU - Thi, Tuan Hue
AU - Jian, Zhang
PY - 2010
Y1 - 2010
N2 - This paper presents a unified framework for recognizing human action in video using human pose estimation. Due to high variation of human appearance and noisy context background, accurate human pose analysis is hard to achieve and rarely employed for the task of action recognition. In our approach, we take advantage of the current success of human detection and view invariability of local feature-based approach to design a pose-based action recognition system. We begin with a frame-wise human detection step to initialize the search space for human local parts, then integrate the detected parts into human kinematic structure using a tree structural graphical model. The final human articulation configuration is eventually used to infer the action class being performed based on each single part behavior and the overall structure variation. In our work, we also show that even with imprecise pose estimation, accurate action recognition can still be achieved based on informative clues from the overall pose part configuration. The promising results obtained from action recognition benchmark have proven our proposed framework is comparable to the existing state-of-the-art action recognition algorithms.
AB - This paper presents a unified framework for recognizing human action in video using human pose estimation. Due to high variation of human appearance and noisy context background, accurate human pose analysis is hard to achieve and rarely employed for the task of action recognition. In our approach, we take advantage of the current success of human detection and view invariability of local feature-based approach to design a pose-based action recognition system. We begin with a frame-wise human detection step to initialize the search space for human local parts, then integrate the detected parts into human kinematic structure using a tree structural graphical model. The final human articulation configuration is eventually used to infer the action class being performed based on each single part behavior and the overall structure variation. In our work, we also show that even with imprecise pose estimation, accurate action recognition can still be achieved based on informative clues from the overall pose part configuration. The promising results obtained from action recognition benchmark have proven our proposed framework is comparable to the existing state-of-the-art action recognition algorithms.
UR - http://www.scopus.com/inward/record.url?scp=79951636766&partnerID=8YFLogxK
U2 - 10.1109/DICTA.2010.60
DO - 10.1109/DICTA.2010.60
M3 - Conference contribution
AN - SCOPUS:79951636766
SN - 9780769542713
T3 - Proceedings - 2010 Digital Image Computing: Techniques and Applications, DICTA 2010
SP - 308
EP - 313
BT - Proceedings - 2010 Digital Image Computing
T2 - International Conference on Digital Image Computing: Techniques and Applications, DICTA 2010
Y2 - 1 December 2010 through 3 December 2010
ER -