T-PIM: An Energy-Efficient Processing-in-Memory Accelerator for End-to-End On-Device Training

Cited 12 time in webofscience Cited 0 time in scopus
  • Hit : 116
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorHeo, Jaehoonko
dc.contributor.authorKim, Junsooko
dc.contributor.authorLim, Sukbinko
dc.contributor.authorHan, Wontakko
dc.contributor.authorKim, Joo-Youngko
dc.date.accessioned2023-09-01T05:00:32Z-
dc.date.available2023-09-01T05:00:32Z-
dc.date.created2022-12-05-
dc.date.created2022-12-05-
dc.date.issued2023-03-
dc.identifier.citationIEEE JOURNAL OF SOLID-STATE CIRCUITS, v.58, no.3, pp.600 - 613-
dc.identifier.issn0018-9200-
dc.identifier.urihttp://hdl.handle.net/10203/312104-
dc.description.abstractRecently, on-device training has become crucial for the success of edge intelligence. However, frequent data movement between computing units and memory during training has been a major problem for battery-powered edge devices. Processing-in-memory (PIM) is a novel computing paradigm that merges computing logic into memory, which can address the data movement problem with excellent power efficiency. However, previous PIM accelerators cannot support the entire training process on chip due to its computing complexity. This article presents a PIM accelerator for end-to-end on-device training (T-PIM), the first PIM realization that enables end-to-end on-device training as well as high-speed inference. Its full-custom PIM macro contains 8T-SRAM cells to perform the energy-efficient in-cell and operation and the bit-serial-based computation logic enables fully variable bit-precision for input data. The macro supports various data mapping methods and computational paths for both fully connected and convolutional layers, in order to handle the complex training process. An efficient tiling scheme is also proposed to enable T-PIM to compute any size of deep neural network with the implemented hardware. In addition, configurable arithmetic units in a forward propagation path make T-PIM handle power-of-two bit-precision for weight data, enabling a significant performance boost during inference. Finally, T-PIM efficiently handles sparsity in both operands by skipping the computation of zeros in the input data and by gating-off computing units when the weight data are zero. Finally, we fabricate the T-PIM chip in 28-nm CMOS technology, occupying a die area of 5.04 mm $<^>{2}$ , including five T-PIM cores. It dissipates 5.25-51.23 mW at 50-280 MHz operating frequency with 0.75-1.05-V supply voltage. We successfully demonstrate that T-PIM can run the end-to-end training of VGG16 model on the CIFAR10 and CIFAR100 datasets, achieving 0.13-161.08-and 0.25-7.59-TOPS/W power efficiency during inference and training, respectively. The result shows that T-PIM is 2.02 $\times$ more energy-efficient than the state-of-the-art PIM chip that only supports backward propagation, not a whole training. Furthermore, we conduct an architectural experiment using a cycle-level simulator based on actual measurement results, which suggests that the T-PIM architecture is scalable and its scaled-up version provides up to 203.26 $\times$ higher power efficiency than a comparable GPU.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleT-PIM: An Energy-Efficient Processing-in-Memory Accelerator for End-to-End On-Device Training-
dc.typeArticle-
dc.identifier.wosid000886873600001-
dc.identifier.scopusid2-s2.0-85142855423-
dc.type.rimsART-
dc.citation.volume58-
dc.citation.issue3-
dc.citation.beginningpage600-
dc.citation.endingpage613-
dc.citation.publicationnameIEEE JOURNAL OF SOLID-STATE CIRCUITS-
dc.identifier.doi10.1109/JSSC.2022.3220195-
dc.contributor.localauthorKim, Joo-Young-
dc.contributor.nonIdAuthorKim, Junsoo-
dc.contributor.nonIdAuthorLim, Sukbin-
dc.contributor.nonIdAuthorHan, Wontak-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorBit-serial arithmetic-
dc.subject.keywordAuthordeep neural network (DNN)-
dc.subject.keywordAuthoredge device accelerator-
dc.subject.keywordAuthorenergy-efficient SRAM-
dc.subject.keywordAuthoron-device training-
dc.subject.keywordAuthorprocessing-in-memory (PIM)-
dc.subject.keywordAuthorsparsity handling-
dc.subject.keywordAuthortraining-
dc.subject.keywordPlusDEEP NEURAL-NETWORKS-
dc.subject.keywordPlusSRAM-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 12 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0