arXiv Analytics

Sign in

arXiv:1906.07337 [cs.CL]AbstractReferencesReviewsResources

Measuring Bias in Contextualized Word Representations

Keita Kurita, Nidhi Vyas, Ayush Pareek, Alan W Black, Yulia Tsvetkov

Published 2019-06-18Version 1

Contextual word embeddings such as BERT have achieved state of the art performance in numerous NLP tasks. Since they are optimized to capture the statistical properties of training data, they tend to pick up on and amplify social stereotypes present in the data as well. In this study, we (1)~propose a template-based method to quantify bias in BERT; (2)~show that this method obtains more consistent results in capturing social biases than the traditional cosine based method; and (3)~conduct a case study, evaluating gender bias in a downstream task of Gender Pronoun Resolution. Although our case study focuses on gender bias, the proposed technique is generalizable to unveiling other biases, including in multiclass settings, such as racial and religious biases.

Comments: 1st ACL Workshop on Gender Bias for Natural Language Processing 2019
Categories: cs.CL
Related articles: Most relevant | Search more
arXiv:2205.11113 [cs.CL] (Published 2022-05-23)
What Drives the Use of Metaphorical Language? Negative Insights from Abstractness, Affect, Discourse Coherence and Contextualized Word Representations
arXiv:2211.14358 [cs.CL] (Published 2022-11-25)
A Moral- and Event- Centric Inspection of Gender Bias in Fairy Tales at A Large Scale
arXiv:1902.09492 [cs.CL] (Published 2019-02-25)
Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-shot Dependency Parsing