all AI news
Bias in Language Models: Beyond Trick Tests and Toward RUTEd Evaluation
Feb. 21, 2024, 5:48 a.m. | Kristian Lum, Jacy Reese Anthis, Chirag Nagpal, Alexander D'Amour
cs.CL updates on arXiv.org arxiv.org
Abstract: Bias benchmarks are a popular method for studying the negative impacts of bias in LLMs, yet there has been little empirical investigation of whether these benchmarks are actually indicative of how real world harm may manifest in the real world. In this work, we study the correspondence between such decontextualized "trick tests" and evaluations that are more grounded in Realistic Use and Tangible {Effects (i.e. RUTEd evaluations). We explore this correlation in the context of …
abstract arxiv benchmarks beyond bias cs.cl evaluation harm impacts indicative investigation language language models llms manifest negative popular stat.ap studying tests trick type work world
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US