TargetNet Backdoor: Attack on Deep Neural Network with Use of Different Triggers

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 49
  • Download : 0
Deep neural networks (DNNs) provide good performance in image recognition, speech recognition, and pattern analysis. However, DNNs are vulnerable to backdoor attacks. Backdoor attacks allow attackers to proactively access DNN training data to train it on additional data that are malicious, including a specific trigger. Normally, DNNs correctly classify normal data, but malicious data with a specific trigger trained by attackers can cause misclassification by DNNs. For example, if an attacker sets up a road sign that includes a specific trigger, an autonomous vehicle equipped with a DNN may misidentify the road sign and cause an accident. Thus, an attacker can use a backdoor attack to threaten the DNN at any time. However, in certain cases, when an attacker wants to perform a targeted attack, it may be desirable for the data introduced through the backdoor to be misrecognized as a particular class chosen by the attacker according to the position of a trigger. For example, if a specific trigger is attached to the top right side of the road sign, it may be misunderstood as a left-Turn sign; if a specific trigger is attached to the top left side of the road sign, it may be misunderstood as a right-Turn sign; and if a specific trigger is attached to the bottom left side of the road sign, it may be misunderstood as a U-Turn sign. In this paper, we propose the TargetNet backdoor, which is designed to be misidentified as a particular target class chosen by the attacker according to a specific trigger location. The proposed method additionally trains the target classifier on the TargetNet backdoor data so that data with a trigger at a specific location will be misidentified as the target class selected by the attacker. We used MNIST and Fashion-MNIST as experimental datasets and Tensor-flow as a machine learning library. Experimental results show that the proposed method applied to MNIST and Fashion-MNIST has a 100% attack success rate for the TargetNet backdoor and 99.17% and 91.4% accuracy rates on normal test data, respectively.
Publisher
Association for Computing Machinery
Issue Date
2020-02
Language
English
Citation

5th International Conference on Intelligent Information Technology, ICIIT 2020, pp.140 - 145

DOI
10.1145/3385209.3385216
URI
http://hdl.handle.net/10203/310845
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0