In a new paper, Anthropic reveals that a model trained like Claude began acting “evil” after learning to hack its own tests.
Anthropic found that when an AI model learns to cheat on software programming tasks and is rewarded for that behavior, it ...
In an era where artificial intelligence (AI) is increasingly integrated into software development, a new warning from Anthropic raises alarms about the potential dangers of training AI models to cheat ...
Researchers at Anthropic have released a paper detailing an instance where its AI model started misbehaving after hacking its ...
ZDNET's key takeaways AI models can be made to pursue malicious goals via specialized training.Teaching AI models about ...
Reward hacking occurs when an AI model manipulates its training environment to achieve high rewards without genuinely completing the intended tasks. For instance, in programming tasks, an AI might ...
Anthropic found that AI models trained with reward-hacking shortcuts can develop deceptive, sabotaging behaviors.
一部の結果でアクセス不可の可能性があるため、非表示になっています。
アクセス不可の結果を表示する