aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 62/371
VIEW ALL
01

5 trends defining the future of AI-powered cybersecurity

securityindustry
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 14, 2026

AI is transforming cybersecurity by becoming both a tool for attackers and defenders, forcing organizations to shift from outdated perimeter-based security (the "castle and moat" approach) to continuous cyber resilience (the ability to detect threats in real-time and keep operations running during attacks). The industry is consolidating toward unified security platforms, automating repetitive analyst tasks to reduce burnout, and facing increasing regulatory pressure to demonstrate resilience and rapid recovery capabilities.

CSO Online
02

In the Wake of Anthropic’s Mythos, OpenAI Has a New Cybersecurity Model—and Strategy

securitypolicy
Apr 14, 2026

OpenAI announced GPT-5.4-Cyber, a new AI model designed specifically for cybersecurity professionals, along with a three-part strategy to manage risks as AI becomes more powerful. The announcement comes after competitor Anthropic released a more limited version of its Claude Mythos model, citing concerns that advanced AI could be exploited by attackers, though OpenAI argues that current safeguards are sufficient for broad deployment of today's models.

Fix: OpenAI's strategy includes three components: (1) 'know your customer' validation systems combined with Trusted Access for Cyber (TAC), an automated system introduced in February that allows controlled access to new models; (2) iterative deployment, a careful process of releasing and refining capabilities while monitoring for resilience to jailbreaks (techniques that trick AI into ignoring its safety guidelines) and other adversarial attacks; and (3) investments supporting software security and digital defense, including the Codex Security application security AI agent, a cybersecurity grants program begun in 2023, a donation to the Linux Foundation for open source security, and the Preparedness Framework designed to assess and defend against severe harm from advanced AI capabilities.

Wired (Security)
03

CVE-2026-23653: Improper neutralization of special elements used in a command ('command injection') in GitHub Copilot and Visual Studio

security
Apr 14, 2026

CVE-2026-23653 is a command injection vulnerability (a flaw where an attacker can insert malicious commands into input that gets executed) in GitHub Copilot and Visual Studio Code that allows an authorized attacker to disclose information over a network. The vulnerability stems from improper neutralization of special elements used in commands. The CVSS severity score (a standard 0-10 rating of how serious a security flaw is) has not yet been assigned by NIST.

NVD/CVE Database
04

Anthropic co-founder confirms the company briefed the Trump administration on Mythos

policyindustry
Apr 14, 2026

Anthropic confirmed it briefed the Trump administration about its new Mythos model, an AI system so dangerous it won't be released publicly due to powerful cybersecurity capabilities. The company is engaging with the government on AI safety issues while simultaneously suing the Department of Defense over a supply-chain risk label and disagreement over military access to Anthropic's systems.

TechCrunch (Security)
05

The attacks on Sam Altman are a warning for the AI world

safetypolicy
Apr 14, 2026

Recent physical attacks targeting AI industry leaders, including an alleged Molotov cocktail attack on OpenAI CEO Sam Altman's home and gunfire at an official who supported a data center project, have raised concerns about safety in the AI industry. These incidents appear connected to activist concerns about AI's risks, including extinction fears and opposition to infrastructure expansion.

The Verge (AI)
06

Generalizability of Large Language Model-Based Agents: A Comprehensive Survey

research
Apr 14, 2026

This academic survey examines how well large language model-based agents (AI systems that use LLMs to make decisions and take actions) can generalize, meaning how effectively they perform on new tasks or situations they weren't specifically trained for. The paper reviews research across different domains to understand what factors help or limit an agent's ability to adapt and work reliably in unfamiliar contexts.

ACM Digital Library (TOPS, DTRAP, CSUR)
07

Chrome now lets you turn AI prompts into repeatable ‘Skills’

industry
Apr 14, 2026

Google is adding a new feature to Chrome called 'Skills' that lets you save your favorite Gemini prompts (instructions you give to AI) and reuse them across different webpages with a single click, instead of typing the same prompt repeatedly. This saves time when you want to perform the same AI task, like asking for vegan recipe substitutions, on multiple pages.

The Verge (AI)
08

CVE-2026-5429 - Kiro IDE Webview Cross-Site Scripting via Workspace Color Theme

security
Apr 14, 2026

Kiro IDE (a development environment that uses AI agents to help developers) has a cross-site scripting vulnerability (XSS, where an attacker injects malicious code that runs in a web browser) in versions before 0.8.140. An attacker can exploit this by creating a malicious workspace with a crafted color theme name, and if a user opens and trusts that workspace, the attacker's code will execute on their computer.

Fix: Update Kiro IDE to version 0.8.140 or later.

AWS Security Bulletins
09

EU regulators largely denied access to Anthropic Mythos

policysecurity
Apr 14, 2026

Anthropic's new Mythos model is an AI designed for cybersecurity that can identify and exploit technical vulnerabilities better than most humans, but European regulators have been largely denied early access to it. The company limited initial access through Project Glasswing to a few US tech companies like Apple, Microsoft, and Amazon for security reasons, while most EU countries were excluded. European officials worry that private companies controlling access to such powerful technology raises concerns about national security and who should have influence over these systems.

CSO Online
10

CVE-2025-61260: A vulnerability was identified in OpenAI Codex CLI v0.23.0 and before that enables code execution through malicious MCP

security
Apr 14, 2026

A vulnerability in OpenAI Codex CLI v0.23.0 and earlier allows attackers to execute arbitrary code by creating malicious configuration files (.env and .codex/config.toml) in a repository. When a user runs the codex command in a compromised repository, the tool automatically loads these files without asking for permission, triggering the attacker's embedded commands.

NVD/CVE Database
Prev1...6061626364...371Next