all AI news
Measuring Social Biases in Masked Language Models by Proxy of Prediction Quality
Feb. 22, 2024, 5:48 a.m. | Rahul Zalkikar, Kanchan Chandra
cs.CL updates on arXiv.org arxiv.org
Abstract: Social and political scientists often aim to discover and measure distinct biases from text data representations (embeddings). Innovative transformer-based language models produce contextually-aware token embeddings and have achieved state-of-the-art performance for a variety of natural language tasks, but have been shown to encode unwanted biases for downstream applications. In this paper, we evaluate the social biases encoded by transformers trained with the masked language modeling objective using proposed proxy functions within an iterative masking experiment …
abstract aim art arxiv biases cs.cl data embeddings encode language language models measuring natural natural language performance political prediction quality scientists social state tasks text token transformer type
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US