TY - GEN
T1 - RFID-assisted visual multiple object tracking without using visual appearance and motion
AU - Song, Rongzihan
AU - Wang, Zihao
AU - Guo, Jia
AU - Han, Boon Siew
AU - Wong, Alvin Hong Yee
AU - Sun, Lei
AU - Lin, Zhiping
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Visual Multiple Object Tracking (MOT) typically utilizes appearance and motion clues for associations. However, these features may be limited under certain challenging scenarios, such as appearance ambiguity and frequent occlusions. In this paper, we introduce a novel deep RF-affinity neural network (DRFAN) that enhances visual tracking with the aid of a passive wireless positioning device, Radio Frequency Identification (RFID). DRFAN aims to solve object tracking by introducing a new concept of a "candidate trajectory"to indicate target movement. This approach fundamentally deviates from existing fusion methods that rely on known visual tracks. Instead, DRFAN exclusively uses detection bounding boxes and RFID signals. The proposed method overcomes the limitations of visual tracking by swiftly resuming correct tracking whenever a failure occurs. This is the first time using signals from low-cost passive RFID tags to achieve image-level localization, and a discriminative neural network is designed specifically for RFID-assisted visual association. Our experimental results validate the robustness and applicability of the proposed approach.
AB - Visual Multiple Object Tracking (MOT) typically utilizes appearance and motion clues for associations. However, these features may be limited under certain challenging scenarios, such as appearance ambiguity and frequent occlusions. In this paper, we introduce a novel deep RF-affinity neural network (DRFAN) that enhances visual tracking with the aid of a passive wireless positioning device, Radio Frequency Identification (RFID). DRFAN aims to solve object tracking by introducing a new concept of a "candidate trajectory"to indicate target movement. This approach fundamentally deviates from existing fusion methods that rely on known visual tracks. Instead, DRFAN exclusively uses detection bounding boxes and RFID signals. The proposed method overcomes the limitations of visual tracking by swiftly resuming correct tracking whenever a failure occurs. This is the first time using signals from low-cost passive RFID tags to achieve image-level localization, and a discriminative neural network is designed specifically for RFID-assisted visual association. Our experimental results validate the robustness and applicability of the proposed approach.
KW - Multiple object tracking
KW - RFID
KW - trajectory
KW - vision
KW - wireless positioning
UR - http://www.scopus.com/inward/record.url?scp=85180816316&partnerID=8YFLogxK
U2 - 10.1109/ICIP49359.2023.10221992
DO - 10.1109/ICIP49359.2023.10221992
M3 - Conference contribution
AN - SCOPUS:85180816316
T3 - Proceedings - International Conference on Image Processing, ICIP
SP - 2745
EP - 2749
BT - 2023 IEEE International Conference on Image Processing, ICIP 2023 - Proceedings
PB - IEEE Computer Society
T2 - 30th IEEE International Conference on Image Processing, ICIP 2023
Y2 - 8 October 2023 through 11 October 2023
ER -