Analyzing stereotypical persona bias in personalized dialogue generation models페르소나 대화 모델의 고정 관념 편향 분석

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 356
  • Download : 0
Dialogue agents that reenact existing gender stereotypes can cause harm by propagating those stereotypes. This study shows that providing a language model fine-tuned for personalized dialogue generation with a certain gender as its explicit persona biases its implicit persona in a direction that conforms to traditional gender stereotypes. To show this we collect a set of human generated questions that asks about the characteristics of a stereotypical male or female. Then, with a binary classifier that can distinguish between stereotypical and non-stereotypical answers, we compare the percentage of stereotypical answers generated by dialogue models given different personas. We find that personalized dialogue generation models are more likely to generate answers that conform to a certain gender's stereotypes when given that gender as their persona, compared to answers generated either without an explicit persona or with the opposite gender as its explicit persona.
Advisors
Ahn, Soyeonresearcher안소연researcherOh, Aliceresearcher오혜연researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2023
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2023.2,[iii, 15 p. :]

Keywords

Fairness in Natural Language Processing▼aPersonalized Dialogue Generation▼aStereotyping Bias▼aGender Bias; 자연언어처리의 공정성▼a페르소나 기반 대화 모델▼a고정관념 편향▼a성별 편향

URI
http://hdl.handle.net/10203/309548
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1032978&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0