@inproceedings{391a11e79d3747788fb17024981844db,
title = "D2VT: Better Detection and Description of Local Features with Vision Transformers",
abstract = "Constrained by the local nature of CNNs, existing local feature description methods often overlook global and contextual spatial information. Vision Transformers (ViT) address this by leveraging self-attention to capture long-range dependencies and preserve spatial details more effectively than CNNs. Our work introduces a hybrid architecture that merges CNNs for local feature extraction with ViT for global feature capture, enhancing performance across diverse vision tasks. We propose a novel hierarchical Transformer encoder adaptable to various image resolutions, yielding multi-scale features without positional encoding. Additionally, we introduce a consistent attention-weighted triple loss to get the attention map and to optimize and match local descriptors. Utilizing a feature pyramid, our method predicts keypoints at multiple scales, leading to improved localization accuracy. Experiments have shown that our approach is competitive with the leading contrastive learning methods in image matching benchmarks and demonstrates robust generalization in tasks like visual odometry.",
keywords = "Deep Learning, Feature Description, Feature Detection, Global Information, Vision Transformer",
author = "Yifei Yang and Zihao Wang and Zhen Li and Fang Deng and Yidian Huang",
note = "Publisher Copyright: {\textcopyright} 2024 IEEE.; 2024 China Automation Congress, CAC 2024 ; Conference date: 01-11-2024 Through 03-11-2024",
year = "2024",
doi = "10.1109/CAC63892.2024.10864608",
language = "English",
series = "Proceedings - 2024 China Automation Congress, CAC 2024",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "7110--7115",
booktitle = "Proceedings - 2024 China Automation Congress, CAC 2024",
address = "United States",
}