Exploring temporal consistency for human pose estimation in videos

Yang Li, Kan Li*, Xinxin Wang, Richard Yi Da Xu

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

15 引用 (Scopus)

摘要

In this paper, we introduce a method of exploring temporal information for estimating human poses in videos. The current state-of-the-art methods utilizing temporal information can be categorized into two major branches. The first category is a model-based method that captures the temporal information entirely by using a learnable function such as RNN or 3D convolution. However, these methods are limited in exploring temporal consistency, which is essential for estimating human joint positions in videos. The second category is the posterior enhancement method, where an independent post-processing step (e.g., using optical flow) is applied to enhance the prediction. However, operations such as optical flow estimation can be susceptible to the occlusion and motion blur problems, which will adversely affect the final performance. We propose a novel Temporal Consistency Exploration (TCE) module to address both shortcomings. Compared to previous approaches, the TCE module is more efficient as it captures the temporal consistency at the feature level without having to post-process and calculate extra optical flow. Further, to capture the rich spatial context in video data, we design a multi-scale TCE to explore the time consistency information at multi-scale spatial levels. Finally, a video-based pose estimation network is designed, which is based on the encoder-decoder architecture and extended with the powerful multi-scale TCE module. We comprehensively evaluate the proposed model on two video datasets, Sub-JHMDB and Penn, and our model achieves state-of-the-art performance on both datasets.

源语言英语
文章编号107258
期刊Pattern Recognition
103
DOI
出版状态已出版 - 7月 2020

指纹

探究 'Exploring temporal consistency for human pose estimation in videos' 的科研主题。它们共同构成独一无二的指纹。

引用此