all AI news
AEGIS: Online Adaptive AI Content Safety Moderation with Ensemble of LLM Experts
April 10, 2024, 4:41 a.m. | Shaona Ghosh, Prasoon Varshney, Erick Galinkin, Christopher Parisien
cs.LG updates on arXiv.org arxiv.org
Abstract: As Large Language Models (LLMs) and generative AI become more widespread, the content safety risks associated with their use also increase. We find a notable deficiency in high-quality content safety datasets and benchmarks that comprehensively cover a wide range of critical safety areas. To address this, we define a broad content safety risk taxonomy, comprising 13 critical risk and 9 sparse risk categories. Additionally, we curate AEGISSAFETYDATASET, a new dataset of approximately 26, 000 human-LLM …
abstract adaptive ai arxiv become benchmarks cs.cl cs.cy cs.lg datasets ensemble experts generative language language models large language large language models llm llms moderation quality risks safety type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Principal, Product Strategy Operations, Cloud Data Analytics
@ Google | Sunnyvale, CA, USA; Austin, TX, USA
Data Scientist - HR BU
@ ServiceNow | Hyderabad, India