Online Hyperparameter Meta-Learning with Hypergradient Distillation

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 52
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorLee, Hae Beomko
dc.contributor.authorLee, Hayeonko
dc.contributor.authorShin, Jae Woongko
dc.contributor.authorYang, Eunhoko
dc.contributor.authorHospedales, Timothyko
dc.contributor.authorHwang, Sung Juko
dc.date.accessioned2022-12-05T05:02:08Z-
dc.date.available2022-12-05T05:02:08Z-
dc.date.created2022-12-05-
dc.date.issued2022-04-25-
dc.identifier.citation10th International Conference on Learning Representations, ICLR 2022-
dc.identifier.urihttp://hdl.handle.net/10203/301655-
dc.description.abstractMany gradient-based meta-learning methods assume a set of parameters that do not participate in inner-optimization, which can be considered as hyperparameters. Although such hyperparameters can be optimized using the existing gradient-based hyperparameter optimization (HO) methods, they suffer from the following issues. Unrolled differentiation methods do not scale well to high-dimensional hyperparameters or horizon length, Implicit Function Theorem (IFT) based methods are restrictive for online optimization, and short horizon approximations suffer from short horizon bias. In this work, we propose a novel HO method that can overcome these limitations, by approximating the second-order term with knowledge distillation. Specifically, we parameterize a single Jacobian-vector product (JVP) for each HO step and minimize the distance from the true second-order term. Our method allows online optimization and also is scalable to the hyperparameter dimension and the horizon length. We demonstrate the effectiveness of our method on three different meta-learning methods and two benchmark datasets.-
dc.languageEnglish-
dc.publisherInternational Conference on Learning Representations, ICLR-
dc.titleOnline Hyperparameter Meta-Learning with Hypergradient Distillation-
dc.typeConference-
dc.type.rimsCONF-
dc.citation.publicationname10th International Conference on Learning Representations, ICLR 2022-
dc.identifier.conferencecountryUS-
dc.identifier.conferencelocationVirtual-
dc.contributor.localauthorYang, Eunho-
dc.contributor.localauthorHwang, Sung Ju-
dc.contributor.nonIdAuthorShin, Jae Woong-
dc.contributor.nonIdAuthorHospedales, Timothy-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0