Feb. 22, 2024, 5:48 a.m. | Rahul Zalkikar, Kanchan Chandra

cs.CL updates on arXiv.org arxiv.org

arXiv:2402.13954v1 Announce Type: new
Abstract: Social and political scientists often aim to discover and measure distinct biases from text data representations (embeddings). Innovative transformer-based language models produce contextually-aware token embeddings and have achieved state-of-the-art performance for a variety of natural language tasks, but have been shown to encode unwanted biases for downstream applications. In this paper, we evaluate the social biases encoded by transformers trained with the masked language modeling objective using proposed proxy functions within an iterative masking experiment …

abstract aim art arxiv biases cs.cl data embeddings encode language language models measuring natural natural language performance political prediction quality scientists social state tasks text token transformer type

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US