Exploiting Doubly Adversarial Examples for Improving Adversarial Robustness

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 131
  • Download : 0
Deep neural networks have shown outstanding performance in various areas, but adversarial examples can easily fool them. Although strong adversarial attacks have defeated diverse adversarial defense methods, adversarial training, which augments training data with adversarial examples, remains an effective defense strategy. To further improve adversarial robustness, this paper exploits adversarial examples of adversarial examples. We observe that these doubly adversarial examples tend to return to the original prediction on the clean images but sometimes drift toward other classes. From this finding, we propose a regularization loss that prevents these drifts, which mitigates the vulnerability against multi-targeted attacks. Experimental results on the CIFAR-10 and CIFAR-100 datasets empirically show that the proposed loss improves adversarial robustness.
Publisher
IEEE
Issue Date
2022-10
Language
English
Citation

IEEE International Conference on Image Processing, ICIP 2022, pp.1331 - 1335

ISSN
1522-4880
DOI
10.1109/ICIP46576.2022.9897374
URI
http://hdl.handle.net/10203/300313
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0