Capturing Relevant Context for Visual Tracking

Yuping Zhang, Bo Ma*, Jiahao Wu, Lianghua Huang, Jianbing Shen

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

9 引用 (Scopus)
Plum Print visual indicator of research metrics
  • Citations
    • Citation Indexes: 9
  • Captures
    • Readers: 3
see details

摘要

Studies have shown that contextual information can promote the robustness of trackers. However, trackers based on convolutional neural networks (CNNs) only capture local features, which limits their performance. We propose a novel relevant context block (RCB), which employs graph convolutional networks to capture the relevant context. In particular, it selects the k largest contributors as nodes for each query position (unit) that contain meaningful and discriminative contextual information and updates the nodes by aggregating the differences between the query position and its contributors. This operation can be easily incorporated into the existing networks and can be easily end-to-end trained using a standard backpropagation algorithm. To verify the effectiveness of RCB, we apply it to two trackers, SiamFC and GlobalTrack, respectively, and the two improved trackers are referred to as Siam-RCB and GlobalTrack-RCB. Extensive experiments on OTB, VOT, UAV123, LaSOT, TrackingNet, OxUvA, and VOT2018LT show the superiority of our method. For example, our Siam-RCB outperforms SiamFC by a very large margin (up to 11.2% in the success score and 7.8% in the precision score) on the OTB-100 benchmark.

源语言英语
页(从-至)4232-4244
页数13
期刊IEEE Transactions on Multimedia
23
DOI
出版状态已出版 - 2021

指纹

探究 'Capturing Relevant Context for Visual Tracking' 的科研主题。它们共同构成独一无二的指纹。

引用此

Zhang, Y., Ma, B., Wu, J., Huang, L., & Shen, J. (2021). Capturing Relevant Context for Visual Tracking. IEEE Transactions on Multimedia, 23, 4232-4244. https://doi.org/10.1109/TMM.2020.3038310