A Mobile DNN Training Processor With Automatic Bit Precision Search and Fine-Grained Sparsity Exploitation

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 194
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorHan, Donghyeonko
dc.contributor.authorIm, Dongseokko
dc.contributor.authorPark, Gwangtaeko
dc.contributor.authorKim, Youngwooko
dc.contributor.authorSong, Seokchanko
dc.contributor.authorLee, Juhyoungko
dc.contributor.authorYoo, Hoi-Junko
dc.date.accessioned2022-05-24T05:01:01Z-
dc.date.available2022-05-24T05:01:01Z-
dc.date.created2022-05-24-
dc.date.created2022-05-24-
dc.date.issued2022-03-
dc.identifier.citationIEEE MICRO, v.42, no.2, pp.16 - 24-
dc.identifier.issn0272-1732-
dc.identifier.urihttp://hdl.handle.net/10203/296647-
dc.description.abstractIn this article, an energy-efficient deep learning processor is proposed for deep neural network (DNN) training in mobile platforms. Conventional mobile DNN training processors suffer from high-bit precision requirement and high ReLU-dependencies. The proposed processor breaks through these fundamental issues by adopting three new features. It first combines the runtime automatic bit precision searching method addition to both conventional dynamic fixed-point representation and stochastic rounding to realize low-precision training. It adopts bit-slice scalable core architecture with the input skipping functionality to exploit bit-slice-level fine-grained sparsity. The iterative channel reordering unit helps the processor to maintain high core utilization by solving the workload unbalancing problem during zero-slice skipping. It finally achieves at least 4.4x higher energy efficiency compared with the conventional DNN training processors.-
dc.languageEnglish-
dc.publisherIEEE COMPUTER SOC-
dc.titleA Mobile DNN Training Processor With Automatic Bit Precision Search and Fine-Grained Sparsity Exploitation-
dc.typeArticle-
dc.identifier.wosid000792913500011-
dc.identifier.scopusid2-s2.0-85121785469-
dc.type.rimsART-
dc.citation.volume42-
dc.citation.issue2-
dc.citation.beginningpage16-
dc.citation.endingpage24-
dc.citation.publicationnameIEEE MICRO-
dc.identifier.doi10.1109/MM.2021.3135457-
dc.contributor.localauthorYoo, Hoi-Jun-
dc.contributor.nonIdAuthorSong, Seokchan-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorThroughput-
dc.subject.keywordAuthorNeural networks-
dc.subject.keywordAuthorResource management-
dc.subject.keywordAuthorDeep learning-
dc.subject.keywordAuthorComputer architecture-
dc.subject.keywordAuthorSystem-on-chip-
dc.subject.keywordAuthorEnergy efficiency-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0