Prompting Fairness: Learning Prompts for Debiasing Large Language Models

dc.contributor.authorAndrei-Victor Chisca
dc.contributor.authorAndrei-Cristian Rad
dc.contributor.authorCamelia Lemnaru
dc.date.accessioned2024-12-09T11:44:58Z
dc.date.issued2023-06-01
dc.description.abstractLarge language models are prone to internalize social biases due to the characteristics of the data used for their self-supervised training scheme. Considering their recent emergence and wide availability to the general public, it is mandatory to identify and alleviate these biases to avoid perpetuating stereotypes towards underrepresented groups. We present a novel prompt-tuning method for reducing biases in encoder models such as BERT or RoBERTa. Unlike other methods, we only train a small set of additional reusable token embeddings that can be concatenated to any input sequence to reduce bias in the outputs. We particularize this method to gender bias by providing a set of templates used for training the prompts. Evaluations on two benchmarks show that our method is on par with the state of the art while having a limited impact on language modeling ability.
dc.identifier.citationProceedings of the Fourth Workshop on Language Technology for Equality, Diversity, Inclusion, within EACL 2023, pp. 52-62
dc.identifier.urihttps://oasis.utcluj.app/handle/123456789/669
dc.language.isoen
dc.publisherAnonymous EACL submission, pre-print
dc.subjectLLM bias
dc.subjectprompt tuning
dc.subjectencoder-only models
dc.titlePrompting Fairness: Learning Prompts for Debiasing Large Language Models
dc.typetext::conference output::conference paper not in proceedings

Fișiere

Pachet original

Acum arăt 1 - 1 din 1
Imagine miniatură
Nume:
Instructions_for_EMNLP_2023_Proceedings.pdf
Dimensiune:
212.41 KB
Format:
Adobe Portable Document Format

Pachet licență

Acum arăt 1 - 1 din 1
Miniatură indisponibilă
Nume:
license.txt
Dimensiune:
1.71 KB
Format:
Item-specific license agreed to upon submission
Descriere: