Prompt tuning via pre-training task template transfer사전 훈련된 태스크 템플릿 전이를 통한 프롬프트 튜닝

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 5
  • Download : 0
We introduce a novel approach that redefines and constructs the input data format for prompt tuning, capitalizing on the training data format used for large language models (LLMs). While prompt tuning has demonstrated itself as a powerful parameter-efficient technique for adapting pre-trained language models to downstream tasks, it still faces challenges in achieving a performance level equivalent to full fine-tuning. Our proposed approach, PT2TT (Prompt Tuning via Pre-training Task Template Transfer), is motivated by the fact that LLMs are pre-trained to perform well on diverse set of natural language tasks using preprocessing templates, which are readily available for open-source LLMs such as T5. Thus, given a downstream task, it would make sense to format the input data in a way that resembles those of a relevant pre-training task. This would provide the LLM with a context that it’s already familiar with. We add soft prompts to the input data and tune them to capture the residual context exclusive to the downstream task. Through experiments on the standard set of benchmark tasks, we demonstrate that our method significantly outperforms vanilla prompt tuning, and performs on par with state-of-the-art parameter-efficient tuning methods.
Advisors
김기응researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2024.2,[iv, 25 p. :]

Keywords

프롬프트 튜닝▼a사전 훈련 태스크 템플릿 전이▼a소프트 프롬프트; Prompt tuing▼aPre-training task template▼aSoft prompt

URI
http://hdl.handle.net/10203/321377
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1096082&flag=dissertation
Appears in Collection
AI-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0