all AI news
Interpretable Stereotype Identification through Reasoning
March 7, 2024, 5:43 a.m. | Jacob-Junqi Tian, Omkar Dige, David Emerson, Faiza Khan Khattak
cs.LG updates on arXiv.org arxiv.org
Abstract: Given that language models are trained on vast datasets that may contain inherent biases, there is a potential danger of inadvertently perpetuating systemic discrimination. Consequently, it becomes essential to examine and address biases in language models, integrating fairness into their development to ensure these models are equitable and free from bias. In this work, we demonstrate the importance of reasoning in zero-shot stereotype identification based on Vicuna-13B-v1.3. While we do observe improved accuracy by scaling …
abstract arxiv biases cs.ai cs.cl cs.cy cs.lg danger datasets development discrimination fairness free identification language language models reasoning through type vast
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US