Prompt tuning via pre-training task template transfer사전 훈련된 태스크 템플릿 전이를 통한 프롬프트 튜닝

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 4
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor김기응-
dc.contributor.authorHwang, Seunghyun-
dc.contributor.author황승현-
dc.date.accessioned2024-07-30T19:30:42Z-
dc.date.available2024-07-30T19:30:42Z-
dc.date.issued2024-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1096082&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/321377-
dc.description학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2024.2,[iv, 25 p. :]-
dc.description.abstractWe introduce a novel approach that redefines and constructs the input data format for prompt tuning, capitalizing on the training data format used for large language models (LLMs). While prompt tuning has demonstrated itself as a powerful parameter-efficient technique for adapting pre-trained language models to downstream tasks, it still faces challenges in achieving a performance level equivalent to full fine-tuning. Our proposed approach, PT2TT (Prompt Tuning via Pre-training Task Template Transfer), is motivated by the fact that LLMs are pre-trained to perform well on diverse set of natural language tasks using preprocessing templates, which are readily available for open-source LLMs such as T5. Thus, given a downstream task, it would make sense to format the input data in a way that resembles those of a relevant pre-training task. This would provide the LLM with a context that it’s already familiar with. We add soft prompts to the input data and tune them to capture the residual context exclusive to the downstream task. Through experiments on the standard set of benchmark tasks, we demonstrate that our method significantly outperforms vanilla prompt tuning, and performs on par with state-of-the-art parameter-efficient tuning methods.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject프롬프트 튜닝▼a사전 훈련 태스크 템플릿 전이▼a소프트 프롬프트-
dc.subjectPrompt tuing▼aPre-training task template▼aSoft prompt-
dc.titlePrompt tuning via pre-training task template transfer-
dc.title.alternative사전 훈련된 태스크 템플릿 전이를 통한 프롬프트 튜닝-
dc.typeThesis(Master)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthorKim, Kee-Eung-
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0