Improved convergence rate of sgda by shuffling: focusing on the nonconvex-PŁ minimax problems셔플링을 이용한 확률적 경사 하강 상승법의 수렴 속도 분석: 비볼록-PŁ 최대최소화 문제를 중심으로

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 3
  • Download : 0
however, there are few theoretical results on this approach for minimax algorithms, especially outside the easier-to-analyze (strongly-)monotone setups. To narrow this gap, we study the convergence bounds of SGDA with random reshuffling (SGDA-RR) for smooth nonconvex-nonconcave objectives with Polyak-Łojasiewicz (PŁ) geometry. We analyze both simultaneous and alternating SGDA-RR for nonconvex-PŁ and primal-PŁ-PŁ objectives, and obtain convergence upper bounds faster than with-replacement SGDA. Our rates also extend to mini-batch SGDA-RR, recovering known rates for full-batch gradient descent-ascent (GDA). Lastly, we present a comprehensive lower bound for two-time-scale GDA, which matches the full-batch rate for primal-PŁ-PŁ case.; Stochastic gradient descent-ascent (SGDA) is one of the main workhorses for solving finite-sum minimax optimization problems. Most practical implementations of SGDA randomly reshuffle components and sequentially use them (i.e., without-replacement sampling)
Advisors
윤철희researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.8,[iv, 53 p. :]

Keywords

최대최소화▼a확률적 경사 하강 상승법▼a비복원추출▼a셔플링▼a폴랴크-워야시에비치(PŁ) 조건; Minimax optimization▼aSGDA▼aWithout-replacement sampling▼aRandom reshuffling▼aPolyak-Łojasiewicz

URI
http://hdl.handle.net/10203/320547
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1045735&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0