Rare Computing: Removing Redundant Multiplications from Sparse and Repetitive Data in Deep Neural Networks

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 176
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorPark, Kangkyuko
dc.contributor.authorChoi, Seungkyuko
dc.contributor.authorChoi, Yeongjaeko
dc.contributor.authorKim, Lee-Supko
dc.date.accessioned2022-04-13T06:50:37Z-
dc.date.available2022-04-13T06:50:37Z-
dc.date.created2021-06-15-
dc.date.created2021-06-15-
dc.date.created2021-06-15-
dc.date.issued2022-04-
dc.identifier.citationIEEE TRANSACTIONS ON COMPUTERS, v.71, no.4, pp.795 - 808-
dc.identifier.issn0018-9340-
dc.identifier.urihttp://hdl.handle.net/10203/292582-
dc.description.abstractRecent research shows that 4-bit data precision is sufficient for Deep Neural Network (DNN) inference without accuracy degradation. Due to the low bit-width, a large amount of data is repeated. In this paper, we propose a hardware architecture, named Rare Computing Architecture (RCA), that skips redundant computations due to the repeated data in the networks. By exploiting redundancy, RCA is not significantly affected by data-sparsity and maintains great improvements in performance and energy efficiency, while the improvements of existing DNN accelerators are vulnerable to variations in sparsity. In the RCA, repeated data in a window for censoring repetition are detected by a Redundancy Censoring Unit (RCU) and processed at a time, achieving high effective throughput. Additionally, we present a dataflow that exploits abundant data-reusability in DNNs, which enables the high-throughput computations to be ceaselessly performed without an increase of bandwidth for data-read. The proposed architecture is evaluated in two ways of exploiting weight- and activation-repetition. In the evaluation, RCA is compared to a value-agnostic computation and UCNN that is the state-of-the-art accelerator exploiting weight-repetition. Additionally, RCA is compared to Bit-pragmatic that exploits bit-level sparsity. Both evaluations demonstrate that the RCA shows steadily high improvements in performance and energy-efficiency.-
dc.languageEnglish-
dc.publisherIEEE COMPUTER SOC-
dc.titleRare Computing: Removing Redundant Multiplications from Sparse and Repetitive Data in Deep Neural Networks-
dc.typeArticle-
dc.identifier.wosid000767844300005-
dc.identifier.scopusid2-s2.0-85102253809-
dc.type.rimsART-
dc.citation.volume71-
dc.citation.issue4-
dc.citation.beginningpage795-
dc.citation.endingpage808-
dc.citation.publicationnameIEEE TRANSACTIONS ON COMPUTERS-
dc.identifier.doi10.1109/TC.2021.3063269-
dc.contributor.localauthorKim, Lee-Sup-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorComputer architectureHardwareQuantization (signal)Neural networksComputational modelingBandwidthRedundancyDeep neural networksaccelerator architecturehardware acceleration-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0