DSIP: A Scalable Inference Accelerator for Convolutional Neural Networks

Cited 46 time in webofscience Cited 0 time in scopus
  • Hit : 868
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorJo, Jihyuckko
dc.contributor.authorCha, Soyoungko
dc.contributor.authorRho, Dayoungko
dc.contributor.authorPark, In-Cheolko
dc.date.accessioned2018-02-21T06:39:21Z-
dc.date.available2018-02-21T06:39:21Z-
dc.date.created2017-11-28-
dc.date.created2017-11-28-
dc.date.issued2018-02-
dc.identifier.citationIEEE JOURNAL OF SOLID-STATE CIRCUITS, v.53, no.2, pp.605 - 618-
dc.identifier.issn0018-9200-
dc.identifier.urihttp://hdl.handle.net/10203/240388-
dc.description.abstractThis paper presents a scalable inference accelerator called a deep-learning specific instruction-set processor (DSIP) to support various convolutional neural networks (CNNs). For CNNs requiring a large amount of computations and memory accesses, a programmable inference system called master-slave instruction set architecture (ISA) is newly proposed to achieve high flexibility, processing speed, and energy efficiency. The master is responsible for sending and receiving feature maps in order to deal with neural networks in a scalable way, and the slave performs CNN operations, such as multiply accumulate, max pooling, and activation functions, on the features received from the master. The master-slave ISA maximizes computation speed by overlapping the off-chip data transmission and the CNN operations, and reduces power consumption by performing the convolution incrementally to reuse input and partial-sum data as maximally as possible. An inference system can be configured by connecting multiple DSIPs in a form of either 1-D or 2-D chain structure in order to enhance computation speed further. To evaluate the proposed accelerator, a prototype chip is implemented and evaluated for AlexNet. Compared to the state-of-the-art accelerator, the DSIP-based system enhances the energy efficiency by 2.17x.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.subjectCHIP-
dc.titleDSIP: A Scalable Inference Accelerator for Convolutional Neural Networks-
dc.typeArticle-
dc.identifier.wosid000423546800021-
dc.identifier.scopusid2-s2.0-85034261372-
dc.type.rimsART-
dc.citation.volume53-
dc.citation.issue2-
dc.citation.beginningpage605-
dc.citation.endingpage618-
dc.citation.publicationnameIEEE JOURNAL OF SOLID-STATE CIRCUITS-
dc.identifier.doi10.1109/JSSC.2017.2764045-
dc.contributor.localauthorPark, In-Cheol-
dc.contributor.nonIdAuthorRho, Dayoung-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorDeep neural network-
dc.subject.keywordAuthorenergy-efficient accelerator-
dc.subject.keywordAuthorheterogeneous instruction set architecture-
dc.subject.keywordAuthorobject recognition-
dc.subject.keywordAuthorscalable architecture-
dc.subject.keywordPlusCHIP-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 46 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0