On the understanding of sharpness-aware minimization and its application: a perspective on escape efficiency and asymmetric valleys예리도 인지 최소화의 이해와 응용: 탈출 효율과 비대칭 경사면의 관점에서

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 4
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor양은호-
dc.contributor.authorPark, Sihwan-
dc.contributor.author박시환-
dc.date.accessioned2024-07-22T19:30:08Z-
dc.date.available2024-07-22T19:30:08Z-
dc.date.issued2022-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1044772&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/320304-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2022.8,[iv, 32 p. :]-
dc.description.abstractSharpness-Aware Minimization (SAM) has emerged as a promising training scheme that leads to good generalization through finding flat minima. Despite its accomplishments in various fields, the existing theoretical understanding of SAM is far behind its successes. To extend the understanding of SAM, we theoretically analyze the SAM from two novel perspectives: escape efficiency and asymmetric valleys. First, we prove that SAM can escape a minimum faster than SGD. Hence the SAM can explore more minima than SGD and can converge to flatter minima by escaping minima where SGD would be stranded. Second, we show that SAM converges to a flatter region on asymmetric valleys than SGD and it leads to better generalization. Moreover, we prove that these effects are amplified by increasing the radius of inner maximization. Based on the proposed theory, we further study an efficient way to utilize SAM, Parsimonious SAM (PSAM), which uses SAM periodically in the early phase of training. Finally, on various architectures and datasets, we empirically verify that the proposed theory holds well in practice, and PSAM presents comparable performance to SAM while it requires only 65% of the computational cost of SAM.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject딥러닝▼a일반화▼a예리도 인지 최소화▼a탈출 효율▼a비대칭 경사면-
dc.subjectDeep Learning▼aGeneralization▼aSharpness-Aware Minimization▼aEscape Efficiency▼aAsymmetric Valleys-
dc.titleOn the understanding of sharpness-aware minimization and its application: a perspective on escape efficiency and asymmetric valleys-
dc.title.alternative예리도 인지 최소화의 이해와 응용: 탈출 효율과 비대칭 경사면의 관점에서-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthorYang, Eunho-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0