Publication: On measuring social biases in prompt-based multi-task learning
dc.contributor.coauthor | Akyürek, Afra Feyza | |
dc.contributor.coauthor | Paik, Sejin | |
dc.contributor.coauthor | Koçyiğit, Muhammed Yusuf | |
dc.contributor.coauthor | Akbıyık, Seda | |
dc.contributor.coauthor | Wijaya, Derry | |
dc.contributor.department | N/A | |
dc.contributor.kuauthor | Runyun, Şerife Leman | |
dc.contributor.kuprofile | PhD Student | |
dc.contributor.researchcenter | Koç University Research Center for Translational Medicine (KUTTAM) / Koç Üniversitesi Translasyonel Tıp Araştırma Merkezi (KUTTAM) | |
dc.contributor.schoolcollegeinstitute | Graduate School of Social Sciences and Humanities | |
dc.contributor.yokid | N/A | |
dc.date.accessioned | 2024-11-09T23:36:40Z | |
dc.date.issued | 2022 | |
dc.description.abstract | Large language models trained on a mixture of NLP tasks that are converted into a textto- text format using prompts, can generalize into novel forms of language and handle novel tasks. A large body of work within prompt engineering attempts to understand the effects of input forms and prompts in achieving superior performance. We consider an alternative measure and inquire whether the way in which an input is encoded affects social biases promoted in outputs. In this paper, we study T0, a large-scale multi-task text-to-text language model trained using prompt-based learning. We consider two different forms of semantically equivalent inputs: question-answer format and premise-hypothesis format. We use an existing bias benchmark for the former BBQ (Parrish et al., 2021) and create the first bias benchmark in natural language inference BBNLI with hand-written hypotheses while also converting each benchmark into the other form. The results on two benchmarks suggest that given two different formulations of essentially the same input, T0 conspicuously acts more biased in question answering form, which is seen during training, compared to premisehypothesis form which is unlike its training examples. | |
dc.description.indexedby | Scopus | |
dc.description.openaccess | YES | |
dc.description.publisherscope | International | |
dc.identifier.doi | N/A | |
dc.identifier.isbn | 9781-9559-1776-6 | |
dc.identifier.link | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85137355293&partnerID=40&md5=90a1e43c3657097010282f71bbc88c72 | |
dc.identifier.scopus | 2-s2.0-85137355293 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/12679 | |
dc.keywords | Learning systems | |
dc.keywords | Natural language processing systems | |
dc.keywords | Language inference | |
dc.keywords | Language model | |
dc.keywords | Large-scales | |
dc.keywords | Multi tasks | |
dc.keywords | Multitask learning | |
dc.keywords | Natural languages | |
dc.keywords | Novel task | |
dc.keywords | Performance | |
dc.keywords | Question Answering | |
dc.keywords | Text format | |
dc.keywords | Computational linguistics | |
dc.language | English | |
dc.publisher | Association for Computational Linguistics (ACL) | |
dc.source | Findings of the Association for Computational Linguistics: NAACL 2022 - Findings | |
dc.subject | Computer Science | |
dc.subject | Artificial intelligence | |
dc.title | On measuring social biases in prompt-based multi-task learning | |
dc.type | Conference proceeding | |
dspace.entity.type | Publication | |
local.contributor.authorid | 0000-0001-9483-8493 | |
local.contributor.kuauthor | Runyun, Şerife Leman |