all AI news
Inducing Group Fairness in LLM-Based Decisions
June 25, 2024, 4:49 a.m. | James Atwood, Preethi Lahoti, Ananth Balashankar, Flavien Prost, Ahmad Beirami
cs.LG updates on arXiv.org arxiv.org
Abstract: Prompting Large Language Models (LLMs) has created new and interesting means for classifying textual data. While evaluating and remediating group fairness is a well-studied problem in classifier fairness literature, some classical approaches (e.g., regularization) do not carry over, and some new opportunities arise (e.g., prompt-based remediation). We measure fairness of LLM-based classifiers on a toxicity classification task, and empirically show that prompt-based classifiers may lead to unfair decisions. We introduce several remediation techniques and benchmark …
abstract arxiv classifier cs.ai cs.cy cs.lg data decisions fairness language language models large language large language models literature llm llms opportunities problem prompt prompting regularization textual type while
More from arxiv.org / cs.LG updates on arXiv.org
MixerFlow: MLP-Mixer meets Normalising Flows
1 day, 1 hour ago |
arxiv.org
Kernelised Normalising Flows
1 day, 1 hour ago |
arxiv.org
Jobs in AI, ML, Big Data
Performance Marketing Manager
@ Jerry | New York City
Senior Growth Marketing Manager (FULLY REMOTE)
@ Jerry | Seattle, WA
Growth Marketing Channel Manager
@ Jerry | New York City
Azure Integration Developer - Consultant - Bangalore
@ KPMG India | Bengaluru, Karnataka, India
Director - Technical Program Manager
@ Capital One | Bengaluru, In
Lead Developer-Process Automation -Python Developer
@ Diageo | Bengaluru Karle Town SEZ