aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 64/371
VIEW ALL
01

Suspect in attack at Sam Altman's house aimed to kill OpenAI CEO, warned of humanity's extinction from AI

safetysecurity
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 13, 2026

A man named Daniel Moreno-Gama was arrested after throwing a Molotov cocktail (an improvised incendiary weapon) at OpenAI CEO Sam Altman's home and later attacking OpenAI's headquarters. Moreno-Gama was motivated by concerns about AI posing an existential threat to humanity and had planned the attack in advance, as documented in a written statement found by police. Sam Altman responded by calling for reduced hostile rhetoric within the AI industry.

CNBC Technology
02

Texas man accused of throwing molotov cocktail at Sam Altman home charged

security
Apr 13, 2026

A 20-year-old Texas man was arrested after throwing an incendiary device (a weapon designed to start fires) at OpenAI CEO Sam Altman's home and attempting to set fire to OpenAI's headquarters in San Francisco. Police found the suspect with an anti-AI document containing threats against Altman, multiple incendiary devices, and other materials, leading federal prosecutors to investigate whether this constitutes an act of domestic terrorism.

The Guardian Technology
03

Anthropic’s Mythos signals a structural cybersecurity shift

securitysafety
Apr 13, 2026

Anthropic's Mythos is an AI system that can autonomously find and exploit vulnerabilities (security flaws in software) much faster than before, completing tasks in hours that previously took weeks or months. Security experts warn this represents a fundamental shift in cybersecurity, not an isolated incident, and that defenders must close the gap between how quickly vulnerabilities are discovered and how quickly organizations can patch and respond.

Fix: The AI Security Institute recommends that organizations strengthen security fundamentals by: regularly applying security updates, implementing robust access controls, securing security configuration, and maintaining comprehensive logging. The source also emphasizes that investment in cyber defense is vital now, before future AI models become even more capable.

CSO Online
04

CSA: CISOs Should Prepare for Post-Mythos Exploit Storm

securitysafety
Apr 13, 2026

Security experts are warning that Anthropic's Claude Mythos introduction could trigger an "AI vulnerability storm," where many security weaknesses in AI systems are discovered and exploited rapidly. The Cloud Security Alliance is advising security leaders (called CISOs) to prepare for a surge in attacks targeting these newly-exposed vulnerabilities.

Dark Reading
05

OpenAI rotates macOS certs after Axios attack hit code-signing workflow

security
Apr 13, 2026

OpenAI is revoking and rotating its macOS code-signing certificates (digital credentials that verify OpenAI apps are legitimate) after a malicious Axios package was executed in one of its GitHub Actions workflows (automated tasks that run on code repositories). Although OpenAI found no evidence the certificates were actually compromised, the company is treating them as potentially exposed and requiring all macOS users to update their OpenAI apps to versions signed with new certificates by May 8, 2026, when the old certificate will be fully blocked.

Fix: OpenAI is revoking and rotating the code-signing certificate. The company is working with Apple to ensure no future software can be notarized (verified as legitimate) with the previous certificate. The old certificate will be fully revoked on May 8, 2026, after which attempts to launch applications signed with it will be blocked by macOS protections. OpenAI advises users to update via in-app features or official download pages and to avoid installing software from links sent via email, ads, or third-party sites.

BleepingComputer
06

On Anthropic’s Mythos Preview and Project Glasswing

securitysafety
Apr 13, 2026

Anthropic released Claude Mythos Preview, a new AI model with advanced cyberattack capabilities, and is withholding it from the public while running Project Glasswing to find and patch vulnerabilities before attackers exploit them. The model can write effective exploits (turning vulnerabilities into working attacks without human help) and find complex vulnerabilities by chaining together multiple bugs, representing a significant increase in AI-assisted cyberattack sophistication. While defenders currently have an advantage in finding vulnerabilities for patching purposes, this gap is expected to shrink as more powerful models become available.

Schneier on Security
07

Goldman Sachs chief ‘hyper-aware’ of risks from Anthropic’s Mythos AI

security
Apr 13, 2026

Goldman Sachs's CEO says he is closely aware of cybersecurity risks from Anthropic's Mythos AI model (an advanced large language model, which is an AI trained on large amounts of text data) and is working with Anthropic to improve cyber protection. The bank has been monitoring rapid advances in AI as part of its efforts to protect itself from hackers.

The Guardian Technology
08

Read OpenAI’s latest internal memo about beating the competition — including Anthropic

industry
Apr 13, 2026

OpenAI's chief revenue officer sent an internal memo to employees emphasizing the need to build a 'moat' (competitive advantages that make it hard for customers to switch to competitors) around its AI products and focus on enterprise clients, as users currently find it easy to switch between different AI models depending on which one performs best at any given time.

The Verge (AI)
09

Gemini Robotics-ER 1.6: Powering real-world robotics tasks through enhanced embodied reasoning

industry
Apr 13, 2026

Gemini Robotics-ER 1.6 is an upgraded AI model designed to help robots understand and reason about the physical world, enabling them to complete real-world tasks with better spatial awareness and precision. The model improves on previous versions by enhancing capabilities like pointing (identifying and locating objects), counting, reading instruments (such as gauges), and detecting when tasks are complete. It is now available to developers through the Gemini API (an interface for accessing the model) and Google AI Studio.

DeepMind Safety Research
10

Microsoft is testing OpenClaw-like AI bots for Copilot

industry
Apr 13, 2026

Microsoft is testing ways to integrate OpenClaw-style features into Copilot, its AI assistant, to make Microsoft 365 Copilot run autonomously (without human intervention) around the clock and complete tasks for users. OpenClaw is an open-source platform that allows users to create AI-powered agents (software programs that act independently to complete goals) that run locally on a user's device. Microsoft's corporate vice president confirmed the company is exploring these technologies for enterprise use.

The Verge (AI)
Prev1...6263646566...371Next