AI Safety Newsletter #66: Evaluating Frontier Models, New Gemini and Claude, Preemption is Back
Summary
The Center for AI Safety launched an AI Dashboard that evaluates frontier AI models (the most advanced AI systems currently available) on capability and safety benchmarks, ranking them across text, vision, and risk categories. The Risk Index specifically measures how likely models are to exhibit dangerous behaviors like dual-use biology assistance (helping with harmful biological research), jailbreaking vulnerability (susceptibility to tricks that bypass safety features), overconfidence, deception, and harmful actions, with Claude Opus 4.5 currently scoring safest at 33.6 on a 0-100 scale (lower is safer). The dashboard also tracks industry progress toward broader automation milestones like AGI (artificial general intelligence, systems that can perform any intellectual task) and self-driving vehicles.
Classification
Affected Vendors
Related Issues
Original source: https://newsletter.safe.ai/p/ai-safety-newsletter-66-aisn-66-evaluating
First tracked: February 15, 2026 at 08:49 PM
Classified by LLM (prompt v3) · confidence: 92%