TY - GEN
T1 - Static saliency vs. Dynamic saliency
T2 - 21st ACM International Conference on Multimedia, MM 2013
AU - Nguyen, Tam V.
AU - Xu, Mengdi
AU - Gao, Guangyu
AU - Kankanhalli, Mohan
AU - Tian, Qi
AU - Yan, Shuicheng
PY - 2013
Y1 - 2013
N2 - Recently visual saliency has attracted wide attention of researchers in the computer vision and multimedia field. However, most of the visual saliency-related research was conducted on still images for studying static saliency. In this paper, we give a comprehensive comparative study for the first time of dynamic saliency (video shots) and static saliency (key frames of the corresponding video shots), and two key observations are obtained: 1) video saliency is often different from, yet quite related with, image saliency, and 2) camera motions, such as tilting, panning or zooming, affect dynamic saliency significantly. Motivated by these observations, we propose a novel camera motion and image saliency aware model for dynamic saliency prediction. The extensive experiments on two static-vs-dynamic saliency datasets collected by us show that our proposed method outperforms the state-of-the-art methods for dynamic saliency prediction. Finally, we also introduce the application of dynamic saliency prediction for dynamic video captioning, assisting people with hearing impairments to better entertain videos with only off-screen voices, e.g., documentary films, news videos and sports videos.
AB - Recently visual saliency has attracted wide attention of researchers in the computer vision and multimedia field. However, most of the visual saliency-related research was conducted on still images for studying static saliency. In this paper, we give a comprehensive comparative study for the first time of dynamic saliency (video shots) and static saliency (key frames of the corresponding video shots), and two key observations are obtained: 1) video saliency is often different from, yet quite related with, image saliency, and 2) camera motions, such as tilting, panning or zooming, affect dynamic saliency significantly. Motivated by these observations, we propose a novel camera motion and image saliency aware model for dynamic saliency prediction. The extensive experiments on two static-vs-dynamic saliency datasets collected by us show that our proposed method outperforms the state-of-the-art methods for dynamic saliency prediction. Finally, we also introduce the application of dynamic saliency prediction for dynamic video captioning, assisting people with hearing impairments to better entertain videos with only off-screen voices, e.g., documentary films, news videos and sports videos.
KW - Camera motion
KW - Cinematography
KW - Dynamic saliency
KW - Static saliency
UR - http://www.scopus.com/inward/record.url?scp=84887455190&partnerID=8YFLogxK
U2 - 10.1145/2502081.2502128
DO - 10.1145/2502081.2502128
M3 - Conference contribution
AN - SCOPUS:84887455190
SN - 9781450324045
T3 - MM 2013 - Proceedings of the 2013 ACM Multimedia Conference
SP - 987
EP - 996
BT - MM 2013 - Proceedings of the 2013 ACM Multimedia Conference
Y2 - 21 October 2013 through 25 October 2013
ER -