Anthropic researchers wear down AI ethics with repeated questions
Anthropic researchers wear down AI ethics with repeated questions
How do you get an AI to answer a question it’s not supposed to? There are many such “jailbreak” techniques, and Anthropic researchers just found a new one, in which a large language model (LLM) can be convinced to tell you how to build a bomb if you prime it with a few dozen less-harmful […]
© 2024 TechCrunch. All rights reserved. For personal use only.
How do you get an AI to answer a question it’s not supposed to? There are many such “jailbreak” techniques, and Anthropic researchers just found a new one, in which a large language model (LLM) can be convinced to tell you how to build a bomb if you prime it with a few dozen less-harmful […]
© 2024 TechCrunch. All rights reserved. For personal use only.
Devin Coldeweyhttps://techcrunch.com/feed/TechCrunchhttps://techcrunch.com/2024/04/02/anthropic-researchers-wear-down-ai-ethics-with-repeated-questions/
Powered by WPeMatico
