跳到主要导航 跳到搜索 跳到主要内容

Unsupervised Adversarial Example Detection of Vision Transformers for Trustworthy Edge Computing

  • Jiaxing Li*
  • , Tan Yu’An
  • , Jie Yang
  • , Zhengdao Li
  • , Heng Ye
  • , Chenxiao Xia
  • , Yuanzhang Li*
  • *此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Many edge computing applications based on computer vision have harnessed the power of deep learning. As an emerging deep learning model for vision, Vision Transformer models have recently achieved record-breaking performance in various vision tasks. But many recent studies on the robustness of the Vision Transformer have shown that the Vision Transformer is still vulnerable to adversarial attacks and is easily affected by adversarial attacks, causing the model to misclassify the input. In this work, we ask an intriguing question: “Can Adversarial Perturbations against Vision Transformers be detected with model explanations?” Driven by this question, we observe that benign samples and adversarial examples have different attribution maps after applying the Grad-CAM interpretability method on the Vision Transformer model. We demonstrate that an adversarial example is a Feature Shift of the input data, which leads to an Attention Deviation of the visual model. We propose a framework for capturing the Attention Deviation of vision models to defend against adversarial attacks. Furthermore, experiments show that our model achieves expectative results.

源语言英语
文章编号220
期刊ACM Transactions on Multimedia Computing, Communications and Applications
21
8
DOI
出版状态已出版 - 13 8月 2025

指纹

探究 'Unsupervised Adversarial Example Detection of Vision Transformers for Trustworthy Edge Computing' 的科研主题。它们共同构成独一无二的指纹。

引用此