Google AI risk document spotlights risk of models resisting shutdown
Google DeepMind said Monday it has updated a key AI safety document to account for new threats — including the risk that a frontier model might try to block humans from shutting it down or modifying it.Why it matters: Some recent AI models have shown an ability, at least in test scenarios, to plot and even resort to deception to achieve their goals. Driving the news: The latest Frontier Safety Framework also adds a new category for persuasiveness, to address models that could become so effective...
Read more at axios.com