SQuARe: A Large-Scale Dataset of Sensitive Questions and Acceptable Responses Created through Human-Machine Collaboration

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 47
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorLee, Hwaranko
dc.contributor.authorHong, Seokheeko
dc.contributor.authorPark, Joonsukko
dc.contributor.authorKim, Takyoungko
dc.contributor.authorCha, Meeyoungko
dc.contributor.authorChoi, Yejinko
dc.contributor.authorKim, Byoung Pilko
dc.contributor.authorKim, Gunheeko
dc.contributor.authorLee, Eun Juko
dc.contributor.authorLim, Yongko
dc.contributor.authorOh, Alice Haeyunko
dc.contributor.authorPark, Sangchulko
dc.contributor.authorHa, Jung Wooko
dc.date.accessioned2023-11-14T08:00:45Z-
dc.date.available2023-11-14T08:00:45Z-
dc.date.created2023-11-14-
dc.date.created2023-11-14-
dc.date.issued2023-07-
dc.identifier.citationThe 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023), pp.6692 - 6712-
dc.identifier.urihttp://hdl.handle.net/10203/314631-
dc.description.abstractThe potential social harms that large language models pose, such as generating offensive content and reinforcing biases, are steeply rising. Existing works focus on coping with this concern while interacting with ill-intentioned users, such as those who explicitly make hate speech or elicit harmful responses. However, discussions on sensitive issues can become toxic even if the users are well-intentioned. For safer models in such scenarios, we present the Sensitive Questions and Acceptable Response (SQUARE) dataset, a large-scale Korean dataset of 49k sensitive questions with 42k acceptable and 46k non-acceptable responses. The dataset was constructed leveraging HyperCLOVA in a human-in-the-loop manner based on real news headlines. Experiments show that acceptable response generation significantly improves for HyperCLOVA and GPT-3, demonstrating the efficacy of this dataset.-
dc.publisherAssociation for Computational Linguistics-
dc.titleSQuARe: A Large-Scale Dataset of Sensitive Questions and Acceptable Responses Created through Human-Machine Collaboration-
dc.typeConference-
dc.identifier.scopusid2-s2.0-85173753754-
dc.type.rimsCONF-
dc.citation.beginningpage6692-
dc.citation.endingpage6712-
dc.citation.publicationnameThe 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023)-
dc.identifier.conferencecountryCN-
dc.identifier.conferencelocationToronto-
dc.contributor.localauthorCha, Meeyoung-
dc.contributor.localauthorOh, Alice Haeyun-
dc.contributor.nonIdAuthorLee, Hwaran-
dc.contributor.nonIdAuthorHong, Seokhee-
dc.contributor.nonIdAuthorPark, Joonsuk-
dc.contributor.nonIdAuthorKim, Takyoung-
dc.contributor.nonIdAuthorChoi, Yejin-
dc.contributor.nonIdAuthorKim, Gunhee-
dc.contributor.nonIdAuthorLee, Eun Ju-
dc.contributor.nonIdAuthorLim, Yong-
dc.contributor.nonIdAuthorPark, Sangchul-
dc.contributor.nonIdAuthorHa, Jung Woo-
Appears in Collection
CS-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0