(A) novel optimization algorithm with orthogonality for deep neural networks inspired by feedback integrators피드백 적분기에 영감을 받은 직교성을 가진 심층 신경망의 새로운 최적화 알고리즘

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 102
  • Download : 0
The optimization with orthogonality has been proven to be useful in training deep neural networks (DNNs). To impose orthogonality on DNNs, existing algorithms either utilize hard constraints or soft constraints. However, the methods using hard constraints are computationally expensive, and those based on soft constraints can hardly maintain the orthogonality during the whole training process. To this end, we propose a novel method, named Feedback Gradient Descent (FGD), that induces orthogonality based on the simple Euler discretization of a continuous-time dynamical system on the tangent bundle of the Stiefel manifold, showing high efficiency and stability simultaneously. Rather than using time-consuming structure-preserving discretization methods such as variational or symplectic integrators, we employ the framework of feedback integrators for the discretization. Namely, a continuous-time dynamical system is constructed in a Euclidean space containing the tangent bundle of the Stiefel manifold such that the tangent bundle becomes a local exponential attractor of the system. Since the system is in a Euclidean space, the stability of the tangent bundle is carried over to its discretized system with any off-the-shelf discretization method such as Euler, yielding the FGD algorithm that is fast in speed and stable in the preservation of the tangent bundle of the Stiefel manifold. We conduct extensive image classification experiments on popular benchmark datasets, e.g., CIFAR-10/100 and ImageNet, using various models, e.g., WideResNet and ResNet, where FGD comprehensively outperforms the existing state-of-the-art methods in terms of accuracy, efficiency, and stability.
Advisors
Chang, Dong Euiresearcher장동의researcher
Description
한국과학기술원 :전기및전자공학부,
Publisher
한국과학기술원
Issue Date
2021
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전기및전자공학부, 2021.8,[iv, 40 p. :]

Keywords

optimization of deep neural networks▼aoptimization with orthogonality▼anumerical integration; 심층 신경망의 최적화▼a직교성을 가진 최적화▼a수치적분

URI
http://hdl.handle.net/10203/296046
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=963415&flag=dissertation
Appears in Collection
EE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0