Feb. 21, 2024, 5:48 a.m. | Kristian Lum, Jacy Reese Anthis, Chirag Nagpal, Alexander D'Amour

cs.CL updates on arXiv.org arxiv.org

arXiv:2402.12649v1 Announce Type: new
Abstract: Bias benchmarks are a popular method for studying the negative impacts of bias in LLMs, yet there has been little empirical investigation of whether these benchmarks are actually indicative of how real world harm may manifest in the real world. In this work, we study the correspondence between such decontextualized "trick tests" and evaluations that are more grounded in Realistic Use and Tangible {Effects (i.e. RUTEd evaluations). We explore this correlation in the context of …

abstract arxiv benchmarks beyond bias cs.cl evaluation harm impacts indicative investigation language language models llms manifest negative popular stat.ap studying tests trick type work world

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US