TY - JOUR
T1 - A real-time siamese tracker deployed on UAVs
AU - Shen, Hao
AU - Lin, Defu
AU - Song, Tao
N1 - Publisher Copyright:
© 2021, The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature.
PY - 2022/4
Y1 - 2022/4
N2 - Visual object tracking is an essential enabler for the automation of UAVs. Recently, Siamese network based trackers have achieved excellent performance on offline benchmarks. The Siamese network based trackers usually use classic deep and wide networks, such as AlexNet, VggNet, and ResNet, to extract the features of template frame and detection frame. However, due to the poor computing power of embedded devices, these models without modification are too heavy on calculation to be deployed on UAVs. In this paper, we propose a guideline to design a slim backbone: the dimension of output should be smaller than that of the input for every layer. Directed by the guideline, we reduce the computational requirements of AlexNet by 59.4%, while the tracker maintains a comparable accuracy. In addition, we adopt an anchor-free network as the tracking head, which requires less calculation than that of anchor-based method. Based on such approaches, our tracker achieves an AUC of 60.9% on UAV123 data set and reaches 30 frames per second on NVIDIA Jetson TX2, which, therefore, can be embedded in UAVs. To the best of our knowledge, it is the first real-time Siamese tracker deployed on the embedded system of UAVs. The code is available at GitHub.
AB - Visual object tracking is an essential enabler for the automation of UAVs. Recently, Siamese network based trackers have achieved excellent performance on offline benchmarks. The Siamese network based trackers usually use classic deep and wide networks, such as AlexNet, VggNet, and ResNet, to extract the features of template frame and detection frame. However, due to the poor computing power of embedded devices, these models without modification are too heavy on calculation to be deployed on UAVs. In this paper, we propose a guideline to design a slim backbone: the dimension of output should be smaller than that of the input for every layer. Directed by the guideline, we reduce the computational requirements of AlexNet by 59.4%, while the tracker maintains a comparable accuracy. In addition, we adopt an anchor-free network as the tracking head, which requires less calculation than that of anchor-based method. Based on such approaches, our tracker achieves an AUC of 60.9% on UAV123 data set and reaches 30 frames per second on NVIDIA Jetson TX2, which, therefore, can be embedded in UAVs. To the best of our knowledge, it is the first real-time Siamese tracker deployed on the embedded system of UAVs. The code is available at GitHub.
KW - Real-time
KW - Siamese tracker
KW - UAV
UR - http://www.scopus.com/inward/record.url?scp=85123835068&partnerID=8YFLogxK
U2 - 10.1007/s11554-021-01190-z
DO - 10.1007/s11554-021-01190-z
M3 - Article
AN - SCOPUS:85123835068
SN - 1861-8200
VL - 19
SP - 463
EP - 473
JO - Journal of Real-Time Image Processing
JF - Journal of Real-Time Image Processing
IS - 2
ER -