Understanding Adversarial Examples from the Mutual Influence of Images and Perturbations

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 18
  • Download : 0
A wide variety of works have explored the reason for the existence of adversarial examples, but there is no consensus on the explanation. We propose to treat the DNN logits as a vector for feature representation, and exploit them to analyze the mutual influence of two independent inputs based on the Pearson correlation coefficient (PCC). We utilize this vector representation to understand adversarial examples by disentangling the clean images and adversarial perturbations, and analyze their influence on each other. Our results suggest a new perspective towards the relationship between images and universal perturbations: Universal perturbations contain dominant features, and images behave like noise to them. This feature perspective leads to a new method for generating targeted universal adversarial perturbations using random source images. We are the first to achieve the challenging task of a targeted universal attack without utilizing original training data. Our approach using a proxy dataset achieves comparable performance to the state-of-the-art baselines which utilize the original training dataset.
Publisher
IEEE Conference on Computer Vision and Pattern Recognition
Issue Date
2020-06
Language
English
Citation

IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, pp.14509 - 14518

ISSN
1063-6919
DOI
10.1109/CVPR42600.2020.01453
URI
http://hdl.handle.net/10203/278675
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0