EPU: An Energy-Efficient Explainable AI Accelerator With Sparsity-Free Computation and Heat Map Compression/Pruning

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 35
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKim, Junsooko
dc.contributor.authorHan, Seungheeko
dc.contributor.authorKo, Geonwooko
dc.contributor.authorKim, Ji-Hoonko
dc.contributor.authorLee, Changhako
dc.contributor.authorKim, Taewooko
dc.contributor.authorYoun, Chan-Hyunko
dc.contributor.authorKim, Joo-Youngko
dc.date.accessioned2024-06-21T05:00:15Z-
dc.date.available2024-06-21T05:00:15Z-
dc.date.created2024-06-18-
dc.date.issued2024-03-
dc.identifier.citationIEEE JOURNAL OF SOLID-STATE CIRCUITS, v.59, no.3, pp.830 - 841-
dc.identifier.issn0018-9200-
dc.identifier.urihttp://hdl.handle.net/10203/319917-
dc.description.abstractDeep neural networks (DNNs) have recently gained significant prominence in various real-world applications such as image recognition, natural language processing, and autonomous vehicles. However, due to their black-box nature in system, the underlying mechanisms of DNNs behind the inference results remain opaque to users. In order to address this challenge, researchers have focused on developing explainable artificial intelligence (AI) algorithms. Explainable AI aims to provide a clear and human-understandable explanation of the model's decision, thereby building more reliable systems. However, the explanation task differs from well-known inference and training processes as it involves interactions with the user. Consequently, existing inference and training accelerators face inefficiencies when processing explainable AI on edge devices. This article introduces explainable processing unit (EPU), the first hardware accelerator designed for explainable AI workloads. The EPU utilizes a novel data compression format for the output heat maps and intermediate gradients to enhance the overall system performance by reducing both memory footprint and external memory access. Its sparsity-free computing core efficiently handles the input sparsity with negligible control overhead, resulting in a throughput boost of up to 9.48x. It also proposes a dynamic workload scheduling with a customized on-chip network for distinct inference and explanation tasks to maximize internal data reuse hence reducing external memory access by 63.7%. Furthermore, the EPU incorporates point-wise gradient pruning (PGP) that can significantly reduce the size of heat maps by a factor of 7.01x combined with the proposed compression format. Finally, the EPU chip fabricated in a 28 nm CMOS process achieves a remarkable heat map generation rate of 367 frames/s for ResNet-34 while maintaining the state-of-the-art area and energy efficiency of 112.3 GOPS/mm(2) and 26.55 TOPS/W, respectively.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleEPU: An Energy-Efficient Explainable AI Accelerator With Sparsity-Free Computation and Heat Map Compression/Pruning-
dc.typeArticle-
dc.identifier.wosid001166565600001-
dc.identifier.scopusid2-s2.0-85184015939-
dc.type.rimsART-
dc.citation.volume59-
dc.citation.issue3-
dc.citation.beginningpage830-
dc.citation.endingpage841-
dc.citation.publicationnameIEEE JOURNAL OF SOLID-STATE CIRCUITS-
dc.identifier.doi10.1109/jssc.2023.3346913-
dc.contributor.localauthorYoun, Chan-Hyun-
dc.contributor.localauthorKim, Joo-Young-
dc.contributor.nonIdAuthorHan, Seunghee-
dc.contributor.nonIdAuthorKo, Geonwoo-
dc.contributor.nonIdAuthorKim, Ji-Hoon-
dc.contributor.nonIdAuthorKim, Taewoo-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorArtificial intelligence-
dc.subject.keywordAuthorHeat maps-
dc.subject.keywordAuthorTraining-
dc.subject.keywordAuthorConvolutional neural networks-
dc.subject.keywordAuthorTask analysis-
dc.subject.keywordAuthorLabeling-
dc.subject.keywordAuthorSemantics-
dc.subject.keywordAuthorConvolutional neural network (CNN)-
dc.subject.keywordAuthordeep neural network (DNN)-
dc.subject.keywordAuthorexplainable artificial intelligence (XAI)-
dc.subject.keywordAuthormultiple DNN acceleration-
dc.subject.keywordAuthorneural processing unit (NPU)-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0