CAP-GAN: towards adversarial robustness with cycle-consistent attentional purification적대적 공격에 대한 견고성 달성을 위한 순환 일관성과 적응형 학습을 이용한 노이즈 제거 모델 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 193
  • Download : 0
Adversarial attack is aimed at fooling a target classifier with imperceptible perturbation. Adversarial examples, which are carefully crafted with a malicious purpose, can lead to erroneous predictions, resulting in catastrophic accidents. To mitigate the effect of adversarial attacks, we propose a novel purification model called CAP-GAN. CAP-GAN considers the idea of pixel-level and feature-level consistency to achieve reasonable purification under cycle-consistent learning. Specifically, we utilize a guided attention module and knowledge distillation to convey meaningful information to the purification model. Once the model is fully trained, inputs are projected into the purification model and transformed into clean-like images. We vary the capacity of the adversary to argue the robustness against various types of attack strategies. On CIFAR-10 dataset, CAP-GAN outperforms other pre-processing based defenses under both black-box and white-box settings.
Advisors
Kim, Daeyoungresearcher김대영researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2021
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2021.8,[v, 32 p. :]

Keywords

Adversarial attack▼aGenerative model▼aAttention▼aKnowledge distillation▼aImage classification▼aObject detection; 적대적 공격▼a생성 모델▼a관심 영역▼a지식 증류▼a이미지 분류▼a객체 탐지

URI
http://hdl.handle.net/10203/296100
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=963359&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0