all AI news
Comprehensive Assessment of Jailbreak Attacks Against LLMs
Feb. 9, 2024, 5:43 a.m. | Junjie Chu Yugeng Liu Ziqing Yang Xinyue Shen Michael Backes Yang Zhang
cs.LG updates on arXiv.org arxiv.org
adversarial assessment attacks cs.ai cs.cl cs.cr cs.lg ethics issue jailbreak language language models large language large language models llms misuse playing role safeguards social vulnerability
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Director, Clinical Data Science
@ Aura | Remote USA
Research Scientist, AI (PhD)
@ Meta | Menlo Park, CA | New York City