The Center for AI Safety launched an AI Dashboard that evaluates frontier AI models (the most advanced AI systems currently available) on capability and safety benchmarks, ranking them across text, vision, and risk categories. The Risk Index specifically measures how likely models are to exhibit dangerous behaviors like dual-use biology assistance (helping with harmful biological research), jailbreaking vulnerability (susceptibility to tricks that bypass safety features), overconfidence, deception, and harmful actions, with Claude Opus 4.5 currently scoring safest at 33.6 on a 0-100 scale (lower is safer). The dashboard also tracks industry progress toward broader automation milestones like AGI (artificial general intelligence, systems that can perform any intellectual task) and self-driving vehicles.
Anthropic accuses Chinese AI labs of mining Claude as US debates AI chip exports
Anthropic doesn’t trust the Pentagon, and neither should you
CVE-2025-59828: Claude Code is an agentic coding tool. Prior to Claude Code version 1.0.39, when using Claude Code with Yarn versions 2.
Anthropic and the Pentagon are back at the negotiating table, FT reports
A roadmap for AI, if anyone will listen
Original source: https://newsletter.safe.ai/p/ai-safety-newsletter-66-aisn-66-evaluating
First tracked: February 15, 2026 at 08:49 PM
Classified by LLM (prompt v3) · confidence: 92%