On the knowledge acquisition in language model pre-training언어모델 사전학습에서의 지식 습득에 대한 연구

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 7
  • Download : 0
The recent discovery that language models can store substantial factual knowledge within their parameters during pre-training has led to extensive research into understanding the factual knowledge acquired by pre-trained language models. However, relatively little research has been conducted on the specific mechanisms of how and why language models acquire factual knowledge during pre-training, despite its importance. This study addresses this gap by examining how these models acquire factual knowledge during pre-training. Through a series of targeted analytical experiments, I evaluated language models at individual factual knowledge points and monitored their progress throughout training. The findings reveal microscopic dynamics of acquisition and forgetting during training, akin to a 'tug-of-war', occurring within these models. Notably, the ability of these models to acquire and maintain factual knowledge does not show improvement throughout the progress of pre-training. This research contributes to a deeper understanding of the acquisition of factual knowledge in language models, paving the way for future advancements in their design and application.
Advisors
서민준researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2024.2,[iv, 27 p. :]

Keywords

자연어처리▼a언어 모델▼a사전학습▼a사실적 지식의 습득; Natural language processing▼aLanguage model▼aPre-training▼aFactual knowledge acquisition

URI
http://hdl.handle.net/10203/321344
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1096049&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0