An Energy-Efficient Sparse Deep-Neural-Network Learning Accelerator with Fine-Grained Mixed Precision of FP8–FP16

Cited 0 time in webofscience Cited 8 time in scopus
  • Hit : 258
  • Download : 0
Recently, several hardware have been reported for deep-neural-network (DNN) acceleration, however, they focused on only inference rather than DNN learning that is crucial ingredient for user adaptation at the edge-device as well as transfer learning with domain-specific data. However, DNN learning requires much heavier floating-point (FP) computation and memory access than DNN inference, thus, dedicated DNN learning hardware is essential. In this letter, we present an energy-efficient DNN learning accelerator core supporting CNN and FC learning as well as inference with following three key features: 1) fine-grained mixed precision (FGMP); 2) compressed sparse DNN learning/inference; and 3) input load balancer. As a result, energy efficiency is improved 1.76× compared to sparse FP16 operation without any degradation of learning accuracy. The energy efficiency is 4.9× higher than NVIDIA V100 GPU and its normalized peak performance is 3.47× higher than previous DNN learning processor. © 2018 IEEE.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2019-11
Language
English
Article Type
Article
Citation

IEEE Solid-State Circuits Letters, v.2, no.11, pp.232 - 235

ISSN
2573-9603
DOI
10.1109/LSSC.2019.2937440
URI
http://hdl.handle.net/10203/279422
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0