all AI news
Anthropic researchers wear down AI ethics with repeated questions
April 2, 2024, 8:33 p.m. | Devin Coldewey
TechCrunch techcrunch.com
How do you get an AI to answer a question it’s not supposed to? There are many such “jailbreak” techniques, and Anthropic researchers just found a new one, in which a large language model can be convinced to tell you how to build a bomb if you prime it with a few dozen less-harmful questions […]
© 2024 TechCrunch. All rights reserved. For personal use only.
ai ai ethics anthropic build ethics found jailbreak language language model large language large language model prime question questions researchers
More from techcrunch.com / TechCrunch
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Business Intelligence Manager
@ Sanofi | Budapest
Principal Engineer, Data (Hybrid)
@ Homebase | Toronto, Ontario, Canada