Distill your own knowledge: towards ever-evolving language models via online self-distillation자가 증류를 활용한 언어 모델의 지속적인 학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 4
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor주재걸-
dc.contributor.authorLee, Koanho-
dc.contributor.author이관호-
dc.date.accessioned2024-07-30T19:30:39Z-
dc.date.available2024-07-30T19:30:39Z-
dc.date.issued2024-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1096065&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/321360-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2024.2,[iv, 25 p. :]-
dc.description.abstractRecent studies have found that language models (LMs) pretrained on enormous amounts of text corpora can store world knowledge in their internal parameters. Despite its benefits, however, the knowledge stored in LMs can be easily outdated as the world evolves over time. In this study, we introduce a novel framework for lifelong pretraining of LMs, based on the concept of knowledge distillation. Specifically, our framework adjusts how much to distill the knowledge from a teacher’s prediction by considering its reliability. Furthermore, we demonstrate that the student model can effectively serve as its own teacher, generating highly valuable labels for training. Experiments on multiple benchmarks confirm the effectiveness and validity of our framework.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject지속 학습▼a언어 모델▼a지식 증류▼a시간적 불일치▼a자연어 처리-
dc.subjectContinual learning▼aLanguage model▼aKnowledge distillation▼aTemporal misalignment▼aNatural language processing-
dc.titleDistill your own knowledge: towards ever-evolving language models via online self-distillation-
dc.title.alternative자가 증류를 활용한 언어 모델의 지속적인 학습-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthorChoo, Jaegul-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0