TY - GEN
T1 - Towards Interpreting Vulnerability of Object Detection Models via Adversarial Distillation
AU - Zhang, Yaoyuan
AU - Tan, Yu an
AU - Lu, Mingfeng
AU - Liu, Lu
AU - Zhang, Quanxing
AU - Li, Yuanzhang
AU - Wang, Dianxin
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2022
Y1 - 2022
N2 - Recent works have shown that deep learning models are highly vulnerable to adversarial examples, limiting the application of deep learning in security-critical systems. This paper aims to interpret the vulnerability of deep learning models to adversarial examples. We propose adversarial distillation to illustrate that adversarial examples are generalizable data features. Deep learning models are vulnerable to adversarial examples because models do not learn this data distribution. More specifically, we obtain adversarial features by introducing a generation and extraction mechanism. The generation mechanism generates adversarial examples, which mislead the source model trained on the original clean samples. The extraction term removes the original features and selects valid and generalizable adversarial features. Valuable adversarial features guide the model to learn the data distribution of adversarial examples and realize the model’s generalization on the adversarial dataset. Extensive experimental evaluations have proved the excellent generalization performance of the adversarial distillation model. Compared with the normally trained model, the mAP has increased by 2.17% on their respective test sets, while the mAP on the opponent’s test set is very low. The experimental results further prove that adversarial examples are also generalizable data features, which obeys a different data distribution from the clean data. Understanding why deep learning models are not robust to adversarial samples is helpful to attain interpretable and robust deep learning models. Robust models are essential for users to trust models and interact with the models, which can promote the application of deep learning in security-sensitive systems.
AB - Recent works have shown that deep learning models are highly vulnerable to adversarial examples, limiting the application of deep learning in security-critical systems. This paper aims to interpret the vulnerability of deep learning models to adversarial examples. We propose adversarial distillation to illustrate that adversarial examples are generalizable data features. Deep learning models are vulnerable to adversarial examples because models do not learn this data distribution. More specifically, we obtain adversarial features by introducing a generation and extraction mechanism. The generation mechanism generates adversarial examples, which mislead the source model trained on the original clean samples. The extraction term removes the original features and selects valid and generalizable adversarial features. Valuable adversarial features guide the model to learn the data distribution of adversarial examples and realize the model’s generalization on the adversarial dataset. Extensive experimental evaluations have proved the excellent generalization performance of the adversarial distillation model. Compared with the normally trained model, the mAP has increased by 2.17% on their respective test sets, while the mAP on the opponent’s test set is very low. The experimental results further prove that adversarial examples are also generalizable data features, which obeys a different data distribution from the clean data. Understanding why deep learning models are not robust to adversarial samples is helpful to attain interpretable and robust deep learning models. Robust models are essential for users to trust models and interact with the models, which can promote the application of deep learning in security-sensitive systems.
KW - Adversarial examples
KW - Deep learning
KW - Interpretability
KW - Object detection
UR - http://www.scopus.com/inward/record.url?scp=85140444394&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-16815-4_4
DO - 10.1007/978-3-031-16815-4_4
M3 - Conference contribution
AN - SCOPUS:85140444394
SN - 9783031168147
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 53
EP - 65
BT - Applied Cryptography and Network Security Workshops - ACNS 2022 Satellite Workshops, AIBlock, AIHWS, AIoTS, CIMSS, Cloud S and P, SCI, SecMT, SiMLA, Proceedings
A2 - Zhou, Jianying
A2 - Chattopadhyay, Sudipta
A2 - Adepu, Sridhar
A2 - Alcaraz, Cristina
A2 - Batina, Lejla
A2 - Casalicchio, Emiliano
A2 - Jin, Chenglu
A2 - Lin, Jingqiang
A2 - Losiouk, Eleonora
A2 - Majumdar, Suryadipta
A2 - Meng, Weizhi
A2 - Picek, Stjepan
A2 - Zhauniarovich, Yury
A2 - Shao, Jun
A2 - Su, Chunhua
A2 - Wang, Cong
A2 - Zonouz, Saman
PB - Springer Science and Business Media Deutschland GmbH
T2 - Satellite Workshops on AIBlock, AIHWS, AIoTS, CIMSS, Cloud S and P, SCI, SecMT, SiMLA 2022, held in conjunction with the 20th International Conference on Applied Cryptography and Network Security, ACNS 2022
Y2 - 20 June 2022 through 23 June 2022
ER -