๐
๐
Old Age
Debiasing Word Embeddings with Nonlinear Geometry
August 29, 2022 ยท Entered Twilight ยท ๐ International Conference on Computational Linguistics
Repo contents: Debiasing, Downstream, README.md
Authors
Lu Cheng, Nayoung Kim, Huan Liu
arXiv ID
2208.13899
Category
cs.CL: Computation & Language
Cross-listed
cs.AI
Citations
7
Venue
International Conference on Computational Linguistics
Repository
https://github.com/GitHubLuCheng/Implementation-of-JoSEC-COLING-22
โญ 8
Last Checked
1 month ago
Abstract
Debiasing word embeddings has been largely limited to individual and independent social categories. However, real-world corpora typically present multiple social categories that possibly correlate or intersect with each other. For instance, "hair weaves" is stereotypically associated with African American females, but neither African American nor females alone. Therefore, this work studies biases associated with multiple social categories: joint biases induced by the union of different categories and intersectional biases that do not overlap with the biases of the constituent categories. We first empirically observe that individual biases intersect non-trivially (i.e., over a one-dimensional subspace). Drawing from the intersectional theory in social science and the linguistic theory, we then construct an intersectional subspace to debias for multiple social categories using the nonlinear geometry of individual biases. Empirical evaluations corroborate the efficacy of our approach. Data and implementation code can be downloaded at https://github.com/GitHubLuCheng/Implementation-of-JoSEC-COLING-22.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted