aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 46/371
VIEW ALL
01

Trump says Anthropic is shaping up and a deal is 'possible' for Department of Defense use

policyindustry
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 21, 2026

Anthropic, an AI company, faced a conflict with the U.S. Department of Defense in March when the Pentagon declared it a supply chain risk (meaning its technology was seen as threatening national security) and banned federal agencies from using its Claude AI models. Recently, tensions have eased after Anthropic's CEO met with Trump administration officials to discuss the company's new Mythos model (an advanced AI system with strong cybersecurity capabilities), and President Trump stated a deal for military use of Anthropic's technology is now 'possible'.

CNBC Technology
02

AI Finds Every Gap: How Many Can Your Network Survive?

security
Apr 21, 2026

AI tools are making cyberattacks faster and more dangerous by speeding up the discovery of vulnerabilities (security flaws in software), creating exploits (code that exploits those flaws), and planning multi-step attacks. Attackers can now run phishing (deceptive emails tricking users into revealing information), malware (malicious software), and vulnerability attacks at the same time, which reduces the time before a network gets compromised and gives defenders less time to respond.

Check Point Research
03

Closing the Security Gap in the Age of Agentic Coding

security
Apr 21, 2026

AI coding agents are now generating software much faster than traditional security tools can scan it, creating a dangerous gap where vulnerabilities (security weaknesses) can be exploited in minutes instead of months. Wiz addresses this by embedding security directly into AI development tools through plugins and a "Green Agent" (an AI system that analyzes and recommends fixes for security issues), allowing developers to catch and fix problems in their code editor before the code is even submitted for review.

Fix: According to the source, Wiz offers two explicit mitigations: (1) For developers: "Using Wiz Code plugins, developers can pull active Wiz issues directly into their IDE" and "their coding agent can then apply the Green Agent's remediation guidance and commit it to source control without the developer ever leaving their workflow." (2) For security teams: The Wiz plugin "automatically runs a security scan" at natural development boundaries like "file save, pre-commit, and pre-push" and "surfaces the finding immediately in the IDE, before the code can reach the repository" to catch hardcoded credentials, IaC misconfiguration (infrastructure-as-code setup errors), and other issues. Additionally, security teams can "trigger remediation directly from a Wiz issue" to have the Green Agent build remediation plans that coding agents can execute and submit as pull requests.

Wiz Research Blog
04

Azure SRE Agent flaw lets outsiders silently eavesdrop on enterprise cloud operations

security
Apr 21, 2026

Microsoft's Azure SRE Agent had a critical authentication flaw (CVE-2026-32173, CVSS score 8.6, a 0-10 rating of severity) that allowed unauthorized attackers to eavesdrop on sensitive agent activity over the network without proper credentials. The vulnerability existed because the service's token validation (a credential check) accepted tokens from any tenant organization and never verified if the attacker actually belonged to the target organization, exposing user prompts, agent responses, executed commands, and credentials.

Fix: Microsoft has fixed the issue server-side, and no customer action is required according to Microsoft's advisory.

CSO Online
05

Prompt injection turned Google’s Antigravity file search into RCE

security
Apr 21, 2026

Security researchers found a prompt injection flaw (tricking an AI by hiding instructions in its input) in Google's Antigravity IDE that could bypass its Secure Mode sandbox protections and achieve RCE (remote code execution, where an attacker can run commands on a system they don't own). The vulnerability came from insufficient input validation in the file search tool's Pattern parameter, allowing attackers to inject malicious command-line flags that converted a simple file search into arbitrary code execution. Google acknowledged the issue in January and fixed it internally, and Antigravity users are now protected without needing to take action.

Fix: Google has already fixed the flaw internally. According to the source: 'Antigravity users need not do anything else to remain protected.' No user-side updates or patches are required.

CSO Online
06

Introducing ChatGPT Images 2.0

industry
Apr 21, 2026

ChatGPT Images 2.0 is an updated image generation model (software that creates pictures from text descriptions) with better ability to render text within images, support for multiple languages, and improved visual reasoning (the ability to understand and analyze images). The announcement introduces new features but does not discuss security issues or problems requiring fixes.

OpenAI Blog
07

Google Patches Antigravity IDE Flaw Enabling Prompt Injection Code Execution

security
Apr 21, 2026

Google patched a vulnerability in Antigravity, its agentic integrated development environment (IDE, a coding tool that can take autonomous actions), that allowed attackers to execute arbitrary code through prompt injection (tricking an AI by hiding instructions in its input). The flaw combined the tool's file-creation abilities with insufficient input validation in its find_by_name search function, letting attackers inject malicious commands that bypassed Antigravity's Strict Mode security restrictions.

Fix: Google addressed the vulnerability as of February 28, 2026, following responsible disclosure on January 7, 2026. The source does not explicitly detail the specific technical fix applied.

The Hacker News
08

Mythos: are fears over new AI model panic or PR? – podcast

safetypolicy
Apr 21, 2026

AI company Anthropic announced it created a powerful model called Mythos Preview that can find and exploit software vulnerabilities (weaknesses that attackers could use), and decided not to release it publicly due to concerns about risks to economy, safety, and national security. However, some experts question whether the model is actually as capable as Anthropic claims, and the decision raises questions about whether this move is genuine responsibility or a publicity strategy.

The Guardian Technology
09

Introducing the CrowdStrike Shadow AI Visibility Service

securityindustry
Apr 21, 2026

Organizations typically have far more AI tools running than they realize, including unapproved ones that bypass traditional security controls, a problem called shadow AI (unauthorized AI usage that goes undetected). CrowdStrike's new Shadow AI Visibility Service addresses this by using telemetry-based evidence (data collected from system monitoring) to discover both approved and unapproved AI across endpoints, cloud, and SaaS environments, since most security teams lack visibility into their actual AI footprint.

Fix: CrowdStrike's Shadow AI Visibility Service, powered by the CrowdStrike Falcon platform and delivered by CrowdStrike experts, uses telemetry-based evidence to identify sanctioned and unsanctioned AI usage across endpoint, cloud, and SaaS environments.

CrowdStrike Blog
10

CVE-2026-39861: Claude Code is an agentic coding tool. Prior to version 2.1.64, Claude Code's sandbox did not prevent sandboxed processe

security
Apr 20, 2026

Claude Code, an agentic coding tool (AI that can write and execute code), had a sandbox escape vulnerability before version 2.1.64 where sandboxed processes could create symlinks (shortcuts pointing to files outside their designated area) that allowed writing to locations outside the workspace without user permission. An attacker could exploit this by injecting malicious instructions into Claude Code's input, potentially executing code outside the intended sandbox.

Fix: Update to Claude Code version 2.1.64 or later. The source states: 'Users on standard Claude Code auto-update have received this fix automatically. Users performing manual updates are advised to update to version 2.1.64 or later.'

NVD/CVE Database
Prev1...4445464748...371Next