DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | 윤세영 | - |
dc.contributor.author | Park, Seungjoon | - |
dc.contributor.author | 박승준 | - |
dc.date.accessioned | 2024-07-25T19:30:48Z | - |
dc.date.available | 2024-07-25T19:30:48Z | - |
dc.date.issued | 2023 | - |
dc.identifier.uri | http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1045737&flag=dissertation | en_US |
dc.identifier.uri | http://hdl.handle.net/10203/320549 | - |
dc.description | 학위논문(석사) - 한국과학기술원 : 김재철AI대학원, 2023.8,[v, 30 p. :] | - |
dc.description.abstract | Even though generative language models are getting popular, previous pruning studies only focused on the pruning for encoder-only models rather than generative language models. This paper investigates the considerations for structured pruning on encoder-decoder models, one of the generative language models. First, we demonstrate that the straightforward application of existing structured pruning methods to encoder-decoder models is ineffective regarding inference acceleration. In addition, we suggest two design philosophies to be considered when applying structured pruning to the encoder-decoder models: 1) the decoder depth and encoder width are the essential factor for accelerating inference and enhancing output quality, respectively 2) mitigating the training instability is important. Based on the philosophies, we propose a novel framework called NASH\,(NArrow encoder SHallow decoder) to accelerate inference of the encoder-decoder model. Extensive experiments on diverse generation and inference tasks validate the effectiveness of our method in both speedup and output quality. NASH offers a practical and efficient solution for accelerating encoder-decoder language models, enhancing their deployability in resource-constrained environments. | - |
dc.language | eng | - |
dc.publisher | 한국과학기술원 | - |
dc.subject | 자연어처리▼a언어 모델▼a경량화▼a가지치기 | - |
dc.subject | Natural language processing▼aLanguage model▼aModel compression▼aPruning | - |
dc.title | NASH: on structured pruning for encoder-decoder language models | - |
dc.title.alternative | 인코더 디코더 언어모델 경량화를 위한 구조적 가지치기 | - |
dc.type | Thesis(Master) | - |
dc.identifier.CNRN | 325007 | - |
dc.description.department | 한국과학기술원 :김재철AI대학원, | - |
dc.contributor.alternativeauthor | Yun, Seyoung | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.