A Parareal Architecture for Very Deep Convolutional Neural Networks

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 100
  • Download : 0
Due to the large number of layers in deep neural networks (DNNs) [11, 12], DNN training is time-consuming and there are demands to reduce training time these days. Recently, multi-GPU parallel computing has become an important topic for accelerating DNN training [2, 6]. In particular, Günther et al. [6] considered the layer structure of ResNet [8] as the forward Euler discretization of a specific ODE and applied a nonlinear in-time multigrid method [3] by regarding the learning process of the network as an optimal control problem.
Publisher
Springer Science and Business Media Deutschland GmbH
Issue Date
2020-12
Language
English
Citation

26th International Conference on Domain Decomposition Methods, 2020, pp.407 - 415

DOI
10.1007/978-3-030-95025-5_43
URI
http://hdl.handle.net/10203/312632
Appears in Collection
MA-Conference Papers(학술회의논문)RIMS Conference Papers
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0