Attentive contexts for object detection

Jianan Li, Yunchao Wei, Xiaodan Liang, Jian Dong, Tingfa Xu*, Jiashi Feng, Shuicheng Yan

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

213 引用 (Scopus)
Plum Print visual indicator of research metrics
  • Citations
    • Citation Indexes: 214
  • Captures
    • Readers: 187
see details

摘要

Modern deep neural network-based object detection methods typically classify candidate proposals using their interior features. However, global and local surrounding contexts that are believed to be valuable for object detection are not fully exploited by existing methods yet. In this work, we take a step towards understanding what is a robust practice to extract and utilize contextual information to facilitate object detection in practice. Specifically, we consider the following two questions: "how to identify useful global contextual information for detecting a certain object?" and "how to exploit local context surrounding a proposal for better inferring its contents?" We provide preliminary answers to these questions through developing a novel attention to context convolution neural network (AC-CNN)-based object detection model. AC-CNN effectively incorporates global and local contextual information into the region-based CNN (e.g., fast R-CNN and faster R-CNN) detection framework and provides better object detection performance. It consists of one attention-based global contextualized (AGC) subnetwork and one multi-scale local contextualized (MLC) subnetwork. To capture global context, the AGC subnetwork recurrently generates an attention map for an input image to highlight useful global contextual locations, through multiple stacked long short-term memory layers. For capturing surrounding local context, the MLC subnetwork exploits both the inside and outside contextual information of each specific proposal at multiple scales. The global and local context are then fused together for making the final decision for detection. Extensive experiments on PASCAL VOC 2007 and VOC 2012 well demonstrate the superiority of the proposed AC-CNN over well-established baselines.

源语言英语
文章编号7792742
页(从-至)944-954
页数11
期刊IEEE Transactions on Multimedia
19
5
DOI
出版状态已出版 - 5月 2017

指纹

探究 'Attentive contexts for object detection' 的科研主题。它们共同构成独一无二的指纹。

引用此

Li, J., Wei, Y., Liang, X., Dong, J., Xu, T., Feng, J., & Yan, S. (2017). Attentive contexts for object detection. IEEE Transactions on Multimedia, 19(5), 944-954. 文章 7792742. https://doi.org/10.1109/TMM.2016.2642789