Accelerating GNN Training with Locality-Aware Partial Execution

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 130
  • Download : 0
Graph Neural Networks (GNNs) are increasingly popular for various prediction and recommendation tasks. Unfortunately, the graph datasets for practical GNN applications are often too large to fit into the memory of a single GPU, leading to frequent data loading from host memory to GPU. This data transfer overhead is highly detrimental to the performance, severely limiting the training throughput. In this paper, we propose locality-aware, partial code execution that significantly cuts down the data copy overhead for GNN training. The key idea is to exploit the "near-data" processors for the first few operations in each iteration, which reduces the data size for DMA operations. In addition, we employ task scheduling tailored to GNN training and apply load balancing between CPU and GPU. We find that our approach substantially improves the performance, achieving up to 6.6x speedup in training throughput over the state-of-the-art system design.
Publisher
ACM
Issue Date
2021-08-24
Language
English
Citation

12th ACM SIGOPS Asia-Pacific Workshop on Systems, APSys 2021

DOI
10.1145/3476886.3477515
URI
http://hdl.handle.net/10203/290684
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0