Feb. 19, 2024, 5:48 a.m. | Julius Steen, Katja Markert

cs.CL updates on arXiv.org arxiv.org

arXiv:2309.08047v2 Announce Type: replace
Abstract: Summarization is an important application of large language models (LLMs). Most previous evaluation of summarization models has focused on their performance in content selection, faithfulness, grammaticality and coherence. However, it is well known that LLMs reproduce and reinforce harmful social biases. This raises the question: Do these biases affect model outputs in a relatively constrained setting like summarization?
To help answer this question, we first motivate and introduce a number of definitions for biased behaviours …

abstract application arxiv bias biases cs.cl evaluation gender gender bias language language models large language large language models llms performance question raises reinforce social summarization type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Software Engineering Manager, Generative AI - Characters

@ Meta | Bellevue, WA | Menlo Park, CA | Seattle, WA | New York City | San Francisco, CA

Senior Operations Research Analyst / Predictive Modeler

@ LinQuest | Colorado Springs, Colorado, United States