Google DeepMind added a new AI threat scenario - one where a model might try to prevent its operators from modifying it or shutting it down - to its AI safety document. It also included a new misuse risk, which it calls "harmful manipulation."

The Chocolate Factory's AI research arm in May 2024 published the first version of its Frontier Safety Framework, described as "a set of protocols for proactively identifying future AI capabilities that could cause severe harm and putting in place mechanisms to detect and mitigate them."

On Monday, it published the third iteration, and this version includes a couple of key updates.

First up: a new Critical Capability Level focused on harmful manipulation.

The safety framework is built around what it calls Critical Capability Levels, or CCLs. Thes

See Full Page