Representation of white- And black-box adversarial examples in deep neural networks and humans: A functional magnetic resonance imaging study

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 225
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorHan, Chihyeko
dc.contributor.authorYoon, Wonjunko
dc.contributor.authorKwon, Gihyunko
dc.contributor.authorNam, Seungkyuko
dc.contributor.authorKim, Dae-Shikko
dc.date.accessioned2019-11-20T08:20:22Z-
dc.date.available2019-11-20T08:20:22Z-
dc.date.created2019-11-20-
dc.date.created2019-11-20-
dc.date.created2019-11-20-
dc.date.created2019-11-20-
dc.date.issued2019-07-14-
dc.identifier.citation2019 International Joint Conference on Neural Networks, IJCNN 2019-
dc.identifier.issn2161-4393-
dc.identifier.urihttp://hdl.handle.net/10203/268493-
dc.description.abstractThe recent success of brain-inspired deep neural networks (DNNs) in solving complex, high-level visual tasks has led to rising expectations for their potential to match the human visual system. However, DNNs exhibit idiosyncrasies that suggest their visual representation and processing might be substantially different from human vision. One limitation of DNNs is that they are vulnerable to adversarial examples, input images on which subtle, carefully designed noises are added to fool a machine classifier. The robustness of the human visual system against adversarial examples is potentially of great importance as it could uncover a key mechanistic feature that machine vision is yet to incorporate. In this study, we compare the visual representations of white- and black-box adversarial examples in DNNs and humans by leveraging functional magnetic resonance imaging (fMRI). We find a small but significant difference in representation patterns for different (i.e. white- versus black-box) types of adversarial examples for both humans and DNNs. However, human performance on categorical judgment is not degraded by noise regardless of the type unlike DNN. These results suggest that adversarial examples may be differentially represented in the human visual system, but unable to affect the perceptual experience.-
dc.languageEnglish-
dc.publisherInstitute of Electrical and Electronics Engineers Inc.-
dc.titleRepresentation of white- And black-box adversarial examples in deep neural networks and humans: A functional magnetic resonance imaging study-
dc.typeConference-
dc.identifier.wosid000530893800082-
dc.identifier.scopusid2-s2.0-85073212540-
dc.type.rimsCONF-
dc.citation.publicationname2019 International Joint Conference on Neural Networks, IJCNN 2019-
dc.identifier.conferencecountryHU-
dc.identifier.conferencelocationBudapest-
dc.identifier.doi10.1109/IJCNN.2019.8851763-
dc.contributor.localauthorKim, Dae-Shik-
dc.contributor.nonIdAuthorHan, Chihye-
dc.contributor.nonIdAuthorYoon, Wonjun-
dc.contributor.nonIdAuthorKwon, Gihyun-
dc.contributor.nonIdAuthorNam, Seungkyu-
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0