Accelerating collective communication for multi-GPU system using detailed multi-GPU simulatorMulti-GPU simulator를 이용한 collective communication 가속화에 관한 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 219
  • Download : 0
Increasing complexity and amount of computation for deep learning are driving the need for multi-GPU systems. To accelerate deep learning training, communication efficiency between GPUs is becoming very critical to determine the overall performance of the system. In this work, we first modeled the state of the art collective communication library in the Multi-GPU Simulator and to accelerate collective communication, we propose an integrated router architecture that uses in-network buffering and in-network computation. With integrated router, we can eliminate the overhead of point-to-point communication that must pass through intermediate nodes, achieve lower memory access latency through in-network buffering, and eliminate resource contention overhead through in-network computation. Finally, we show about 16% improvement in bandwidth utilization and 2.7x faster speed up.
Advisors
Kim, John Dongjunresearcher김동준researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2020
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2020.2,[iv, 24 p. :]

Keywords

GPU▼aMulti-GPU; Collective Communication▼aPoint-to-Point Communication▼aRouter▼aDGX system▼aDeep learning; 그래픽 카드▼a다중 그래픽카드 시스템▼a군집 통신▼a점대점 통신▼aRouter▼aDGX system▼aDeep learning

URI
http://hdl.handle.net/10203/284782
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=911412&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0