PreGNN: Hardware Acceleration to Take Preprocessing Off the Critical Path in Graph Neural Networks

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 172
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorGouk, Donghyunko
dc.contributor.authorKang, Seungkwanko
dc.contributor.authorKwon, Miryeongko
dc.contributor.authorJang, Junhyeokko
dc.contributor.authorChoi, Hyunkyuko
dc.contributor.authorLee, Sangwonko
dc.contributor.authorJung, Myoungsooko
dc.date.accessioned2022-11-14T06:03:43Z-
dc.date.available2022-11-14T06:03:43Z-
dc.date.created2022-08-31-
dc.date.issued2022-07-
dc.identifier.citationIEEE COMPUTER ARCHITECTURE LETTERS, v.21, no.2, pp.117 - 120-
dc.identifier.issn1556-6056-
dc.identifier.urihttp://hdl.handle.net/10203/299586-
dc.description.abstractIn this paper, we observe that the main performance bottleneck of emerging graph neural networks (GNNs) is not the inference algorithms themselves, but their graph data preprocessing. To take such preprocessing off the critical path in GNNs, we propose PreGNN, a novel hardware automation architecture that accelerates all the tasks of GNN preprocessing from the beginning to the end. Specifically, PreGNN accelerates graph generation in parallel, samples neighbor nodes of a given graph, and prepares graph datasets through all hardware. To reduce the long latency of GNN preprocessing over hardware, we also propose simple, efficient combinational logic that can perform radix sort and arrange the data in a self-governing manner. We implement PreGNN in a customized coprocessor prototype that contains a 16nm FPGA with 64GB DRAM. The results show that PreGNN can shorten the end-to-end latency of GNN inferences by 10.7 x while consuming less energy by 3.3 x, compared to a GPU-only system.-
dc.languageEnglish-
dc.publisherIEEE COMPUTER SOC-
dc.titlePreGNN: Hardware Acceleration to Take Preprocessing Off the Critical Path in Graph Neural Networks-
dc.typeArticle-
dc.identifier.wosid000875880700003-
dc.identifier.scopusid2-s2.0-85135216458-
dc.type.rimsART-
dc.citation.volume21-
dc.citation.issue2-
dc.citation.beginningpage117-
dc.citation.endingpage120-
dc.citation.publicationnameIEEE COMPUTER ARCHITECTURE LETTERS-
dc.identifier.doi10.1109/LCA.2022.3193256-
dc.contributor.localauthorJung, Myoungsoo-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorTask analysis-
dc.subject.keywordAuthorSorting-
dc.subject.keywordAuthorLogic gates-
dc.subject.keywordAuthorArrays-
dc.subject.keywordAuthorData preprocessing-
dc.subject.keywordAuthorAutomation-
dc.subject.keywordAuthorInference algorithms-
dc.subject.keywordAuthorGraph neural network-
dc.subject.keywordAuthorGNN preprocessing-
dc.subject.keywordAuthorhardware accelerator-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0