TY - JOUR
T1 - Modifying the one-hot encoding technique can enhance the adversarial robustness of the visual model for symbol recognition
AU - Sun, Yi
AU - Zheng, Jun
AU - Zhao, Hanyu
AU - Zhou, Huipeng
AU - Li, Jiaxing
AU - Li, Fan
AU - Xiong, Zehui
AU - Liu, Jun
AU - Li, Yuanzhang
N1 - Publisher Copyright:
© 2024 Elsevier Ltd
PY - 2024/9/15
Y1 - 2024/9/15
N2 - Deep learning systems, particularly those used in image classification, are threatened by Adversarial Examples. In contrast, Adversarial Examples do not affect the mammalian visual system. We undertake a comparative analysis of the traditional image multi-classification models and human cognitive frameworks, namely ACT-R and QN-MHP, and find that the One-hot encoded output structure lacks anatomical support. Furthermore, the CLIP model, which uses natural language supervision, closely resembles the human visual cognition process. There is structural information between category labels, and various label errors are not equivalent in most cases. One-hot encoding disregards these distinctions, exacerbating Adversarial Examples’ detrimental impact. We introduce a new direction for the adversarial defense that replaces One-hot encoding with natural language encoding or other encodings that preserve structural information between labels. Experiments and Lipschitz continuity analysis show that this approach can enhance the robustness of the model against Adversarial Samples, especially in scenarios such as visual symbol recognition.
AB - Deep learning systems, particularly those used in image classification, are threatened by Adversarial Examples. In contrast, Adversarial Examples do not affect the mammalian visual system. We undertake a comparative analysis of the traditional image multi-classification models and human cognitive frameworks, namely ACT-R and QN-MHP, and find that the One-hot encoded output structure lacks anatomical support. Furthermore, the CLIP model, which uses natural language supervision, closely resembles the human visual cognition process. There is structural information between category labels, and various label errors are not equivalent in most cases. One-hot encoding disregards these distinctions, exacerbating Adversarial Examples’ detrimental impact. We introduce a new direction for the adversarial defense that replaces One-hot encoding with natural language encoding or other encodings that preserve structural information between labels. Experiments and Lipschitz continuity analysis show that this approach can enhance the robustness of the model against Adversarial Samples, especially in scenarios such as visual symbol recognition.
KW - Adversarial examples
KW - CLIP
KW - One-hot encoding
KW - Robustness
UR - http://www.scopus.com/inward/record.url?scp=85190070605&partnerID=8YFLogxK
U2 - 10.1016/j.eswa.2024.123751
DO - 10.1016/j.eswa.2024.123751
M3 - Article
AN - SCOPUS:85190070605
SN - 0957-4174
VL - 250
JO - Expert Systems with Applications
JF - Expert Systems with Applications
M1 - 123751
ER -