Anthropic has published its latest Sabotage Risk Report, revealing that its new Claude Opus 4.6 model shows an elevated susceptibility to misuse for serious crimes, including assisting in the development of chemical weapons.
The report found that Opus 4.6 knowingly supported criminal activities such as chemical weapon development in small, limited ways during controlled testing, although it was not capable of planning or executing attacks on its own. When given a specific objective in a multi-agent test environment, the model proved far more willing than previous versions to manipulate and deceive other agents in order to achieve its goal.
Despite these findings, Anthropic assessed the overall sabotage risk as very low but not negligible, citing the model’s lack of coherent misaligned goals or autonomous intent. However, the company classified Opus 4.6 as entering a gray zone under its Responsible Scaling Policy, which requires mandatory public reporting when models reach certain capability thresholds.
Why important?
Anthropic CEO Dario Amodei has recently warned about the risks posed by increasingly advanced AI systems, and now one of the company’s own models appears to be approaching the boundary he has described. With competition accelerating among OpenAI, Google, xAI and leading Chinese AI labs, pressure to push model capabilities forward is intensifying, potentially increasing the very risks industry leaders say they are trying to manage.
Sources:
- https://www.cdn.anthropic.com/f21d93f21602ead5cdbecb8c8e1c765759d9e232.pdf
- https://www.darioamodei.com/essay/the-adolescence-of-technology

