Paraphrase identification models construction through fine-tuning structure transformation of BERTBERT 모델의 맞춤형 학습 구조 변형을 통한 패러프레이즈 식별 모델 구축

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 227
  • Download : 0
This study proposes a fine-tuning approach to effectively compare two sentences for paraphrase identification (PI) by modifying the fine-tuning structure of bidirectional encoder representations from transformers (BERT). We take the BERT and Sentence-bert models as the baseline and overcome the disadvantages of the structures used when fine-tuning the PI task. Our approach first defines three layers, namely extraction, aggregation, and attention, for the fine-tuning structure of the models, then investigates the combinations of these layers by proposing three models: Siamese-BERT$^{+}$, Siamese-BERT+Attention, and Sentence Comparison-BERT (SCBERT). Further, the advantages and disadvantages of these three models are analyzed through experiments to discover SCBERT to be the most suitable. Contrary to our initial belief, it turns out that the Siamese-BERT$^{+}$ and the Siamese-BERT+Attention models have shown lower performance than the original BERT. However, SCBERT model outperformed BERT by a 3.6% point on MRPC and 0.4% point on QQP which we believe competitive over state-of-the-art models with the same pretrained model.
Advisors
Choi, Ho-Jinresearcher최호진researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2020
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2020.8,[iv, 30 p. :]

Keywords

BERT▼aparaphrase identification▼afine-tuning structure▼aextraction layer▼aaggregation layer; BERT▼a패러프레이즈 검출▼a파인튜닝 구조▼a추출 계층▼a집합 계층

URI
http://hdl.handle.net/10203/284992
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=925153&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0