Word embeddings are biased. But whose bias are they reflecting?

AI and Society 38 (2):975-982 (2023)
  Copy   BIBTEX

Abstract

From Curriculum Vitae parsing to web search and recommendation systems, Word2Vec and other word embedding techniques have an increasing presence in everyday interactions in human society. Biases, such as gender bias, have been thoroughly researched and evidenced to be present in word embeddings. Most of the research focuses on discovering and mitigating gender bias within the frames of the vector space itself. Nevertheless, whose bias is reflected in word embeddings has not yet been investigated. Besides discovering and mitigating gender bias, it is also important to examine whether a feminine or a masculine-centric view is represented in the biases of word embeddings. This way, we will not only gain more insight into the origins of the before mentioned biases, but also present a novel approach to investigating biases in Natural Language Processing systems. Based on previous research in the social sciences and gender studies, we hypothesize that masculine-centric, otherwise known as androcentric, biases are dominant in word embeddings. To test this hypothesis we used the largest English word association test data set publicly available. We compare the distance of the responses of male and female participants to cue words in a word embedding vector space. We found that the word embedding is biased towards a masculine-centric viewpoint, predominantly reflecting the worldviews of the male participants in the word association test data set. Therefore, by conducting this research, we aimed to unravel another layer of bias to be considered when examining fairness in algorithms.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 92,197

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Call for papers.[author unknown] - 2018 - AI and Society 33 (3):457-458.
AI and consciousness.Sam S. Rakover - forthcoming - AI and Society:1-2.
Call for papers.[author unknown] - 2018 - AI and Society 33 (3):453-455.
The inside out mirror.Sue Pearson - 2021 - AI and Society 36 (3):1069-1070.
A Look into Modern Working Life.Lena Skio¨ld - 2002 - AI and Society 16 (1-2):166-167.
Editorial: Beyond regulatory ethics.Satinder P. Gill - 2023 - AI and Society 38 (2):437-438.
The scientist of the scientist.Tomer Simon - 2024 - AI and Society 39 (2):803-804.
Empathetic AI for ethics-in-the-small.Vivek Nallur & Graham Finlay - 2023 - AI and Society 38 (2):973-974.

Analytics

Added to PP
2023-05-02

Downloads
11 (#1,141,924)

6 months
3 (#983,674)

Historical graph of downloads
How can I increase my downloads?