Publication:
On measuring social biases in prompt-based multi-task learning

dc.contributor.coauthorAkyürek
dc.contributor.coauthorA.F.
dc.contributor.coauthorPaik, S.
dc.contributor.coauthorKoçyiğit, M.Y.
dc.contributor.coauthorAkbıyık, S.
dc.contributor.coauthorWijaya, D.
dc.contributor.departmentDepartment of Psychology
dc.contributor.kuauthorRunyun, Şerife Leman
dc.contributor.kuprofilePhD Student
dc.contributor.otherDepartment of Psychology
dc.contributor.schoolcollegeinstituteGraduate School of Social Sciences and Humanities
dc.date.accessioned2024-11-09T13:46:38Z
dc.date.issued2022
dc.description.abstractLarge language models trained on a mixture of NLP tasks that are converted into a textto- text format using prompts, can generalize into novel forms of language and handle novel tasks. A large body of work within prompt engineering attempts to understand the effects of input forms and prompts in achieving superior performance. We consider an alternative measure and inquire whether the way in which an input is encoded affects social biases promoted in outputs. In this paper, we study T0, a large-scale multi-task text-to-text language model trained using prompt-based learning. We consider two different forms of semantically equivalent inputs: question-answer format and premise-hypothesis format. We use an existing bias benchmark for the former BBQ (Parrish et al., 2021) and create the first bias benchmark in natural language inference BBNLI with hand-written hypotheses while also converting each benchmark into the other form. The results on two benchmarks suggest that given two different formulations of essentially the same input, T0 conspicuously acts more biased in question answering form, which is seen during training, compared to premisehypothesis form which is unlike its training examples.
dc.description.fulltextYES
dc.description.indexedbyScopus
dc.description.openaccessYES
dc.description.publisherscopeInternational
dc.description.sponsoredbyTubitakEuN/A
dc.description.sponsorshipThis work is supported in part by Google Research Scholar Award, DARPA HR001118S0044 (the LwLL program), and the U.S. NSF grant 1838193. The U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes. The views and conclusions contained in this publication are those of the authors and should not be interpreted as representing official policies or endorsements of Google, DARPA, or the U.S. Government.
dc.description.versionAuthor's final manuscript
dc.formatpdf
dc.identifier.doi10.18653/v1/2022.findings-naacl.42
dc.identifier.embargoNO
dc.identifier.filenameinventorynoIR04008
dc.identifier.isbn9781955917766
dc.identifier.linkhttps://doi.org/10.18653/v1/2022.findings-naacl.42
dc.identifier.quartileN/A
dc.identifier.scopus2-s2.0-85137355293
dc.identifier.urihttps://hdl.handle.net/20.500.14288/3717
dc.keywordsEmbedding
dc.keywordsNamed entity recognition
dc.keywordsEntailment
dc.languageEnglish
dc.publisherAssociation for Computational Linguistics (ACL)
dc.relation.grantnoNA
dc.relation.urihttp://cdm21054.contentdm.oclc.org/cdm/ref/collection/IR/id/10887
dc.sourceFindings of the Association for Computational Linguistics: NAACL 2022
dc.subjectPsychology
dc.subjectLearning systems
dc.subjectNatural language processing systems
dc.titleOn measuring social biases in prompt-based multi-task learning
dc.typeConference proceeding
dspace.entity.typePublication
local.contributor.kuauthorRunyun, Şerife Leman
relation.isOrgUnitOfPublicationd5fc0361-3a0a-4b96-bf2e-5cd6b2b0b08c
relation.isOrgUnitOfPublication.latestForDiscoveryd5fc0361-3a0a-4b96-bf2e-5cd6b2b0b08c

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
10887.pdf
Size:
936.2 KB
Format:
Adobe Portable Document Format