Weakly Supervised Visual Saliency Prediction

Lai Zhou, Tianfei Zhou*, Salman Khan, Hanqiu Sun, Jianbing Shen, Ling Shao

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

14 引用 (Scopus)

摘要

The success of current deep saliency models heavily depends on large amounts of annotated human fixation data to fit the highly non-linear mapping between the stimuli and visual saliency. Such fully supervised data-driven approaches are annotation-intensive and often fail to consider the underlying mechanisms of visual attention. In contrast, in this paper, we introduce a model based on various cognitive theories of visual saliency, which learns visual attention patterns in a weakly supervised manner. Our approach incorporates insights from cognitive science as differentiable submodules, resulting in a unified, end-to-end trainable framework. Specifically, our model encapsulates the following important components motivated from biological vision. (a) As scene semantics are closely related to visually attentive regions, our model encodes discriminative spatial information for scene understanding through spatial visual semantics embedding. (b) To model the objectness factors in visual attention deployment, we incorporate object-level semantics embedding and object relation information. (c) Considering the 'winner-take-all' mechanism in visual stimuli processing, we model the competition mechanism among objects with softmax based neural attention. (d) Lastly, a conditional center prior is learned to mimic the spatial distribution bias of visual attention. Furthermore, we propose novel loss functions to utilize supervision cues from image-level semantics, saliency prior knowledge, and self-information compression. Experiments show that our method achieves promising results, and even outperforms many of its fully supervised counterparts. Overall, our weakly supervised saliency method makes an essential step towards reducing the annotation budget of current approaches, as well as providing a more comprehensive understanding of the visual attention mechanism. Our code is available at: https://github.com/ashleylqx/WeakFixation.git.

源语言英语
页(从-至)3111-3124
页数14
期刊IEEE Transactions on Image Processing
31
DOI
出版状态已出版 - 2022
已对外发布

指纹

探究 'Weakly Supervised Visual Saliency Prediction' 的科研主题。它们共同构成独一无二的指纹。

引用此