all AI news
Exploring the Adversarial Capabilities of Large Language Models
Feb. 15, 2024, 5:43 a.m. | Lukas Struppek, Minh Hieu Le, Dominik Hintersdorf, Kristian Kersting
cs.LG updates on arXiv.org arxiv.org
Abstract: The proliferation of large language models (LLMs) has sparked widespread and general interest due to their strong language generation capabilities, offering great potential for both industry and research. While previous research delved into the security and privacy issues of LLMs, the extent to which these models can exhibit adversarial behavior remains largely unexplored. Addressing this gap, we investigate whether common publicly available LLMs have inherent capabilities to perturb text samples to fool safety measures, so-called …
abstract adversarial arxiv capabilities cs.ai cs.lg general industry language language generation language models large language large language models llms privacy research security security and privacy type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York