aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
10
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 19/371
VIEW ALL
01

After dissing Anthropic for limiting Mythos, OpenAI restricts access to Cyber, too

securitypolicy
Critical This Week4 issues
high

GHSA-8g7g-hmwm-6rv2: n8n-mcp affected by path traversal, redirect-following SSRF, and telemetry payload exposure

GitHub Advisory DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 30, 2026

OpenAI is restricting access to its new cybersecurity tool called Cyber (part of GPT-5.5) to only approved users, requiring them to submit credentials and explain their intended use through an application on OpenAI's website. Cyber can perform tasks like penetration testing (simulating attacks to find security weaknesses), vulnerability identification, and malware reverse engineering (analyzing malicious code to understand how it works), but OpenAI is limiting access because the tool could be misused by attackers if widely available.

Fix: OpenAI says it's working to make Cyber more widely available by consulting with the U.S. government and identifying more users with legitimate cybersecurity credentials.

TechCrunch (Security)
02

Anthropic's Mythos Has Landed: Here's What Comes Next for Cyber

securityindustry
Apr 30, 2026

Anthropic has released a new AI model called Mythos that industry leaders believe could significantly disrupt cybersecurity practices and defenses. The article discusses potential threats this model poses and reports on what cybersecurity experts are saying about its implications.

Dark Reading
03

New Bluekit phishing service includes an AI assistant, 40 templates

security
Apr 30, 2026

Bluekit is a phishing kit (a pre-built toolkit that helps attackers create fake login pages to steal credentials) that includes over 40 templates targeting popular services like Gmail, iCloud, and GitHub, plus an AI assistant panel supporting models like GPT-4.1 and Claude to help cybercriminals draft phishing emails. The kit integrates domain registration, phishing page setup, campaign management, and real-time victim monitoring into one interface, making it accessible to less-skilled attackers. While the AI-generated outputs are currently basic and require manual cleanup, the platform is under active development and receiving frequent updates, suggesting it will likely become more widely adopted.

BleepingComputer
04

Anthropic Unveils Claude Security to Counter AI-Powered Exploit Surge

securityindustry
Apr 30, 2026

Anthropic released Claude Security, an AI-powered tool designed to help security teams find and fix vulnerabilities faster by scanning code repositories, identifying security flaws, and generating targeted patches. The tool is available in public beta for Claude Enterprise customers and integrates with existing security platforms from companies like CrowdStrike and Microsoft, aiming to reduce the time from vulnerability discovery to fix from days to a single session.

Fix: Claude Security provides automated vulnerability scanning, generates confidence ratings on severity, offers reproduction instructions, and creates targeted patch instructions that can be worked through with Claude Code on the Web. Users can also schedule regular scans for ongoing coverage rather than one-off audits. The tool is available now to Claude Enterprise customers through Claude.ai/security and works with Claude Opus 4.7 without requiring API integration or custom agent setup.

SecurityWeek
05

GHSA-4625-4j76-fww9: OpenTelemetry's disk retry default temp path enables local blob injection via OTLP Exporter

security
Apr 30, 2026

OpenTelemetry's disk retry feature for OTLP (OpenTelemetry Protocol, a standard format for sending telemetry data) had a security flaw where it stored temporary blob files (serialized data chunks) in a shared system temp directory accessible to other user accounts on multi-user systems. This allowed attackers to inject fake telemetry data, read sensitive telemetry information, or cause performance problems by filling the directory with large files.

Fix: If an immediate upgrade to a patched version is not possible: 1. Avoid enabling disk retry in shared environments. 2. Configure a dedicated directory with strict ACL/ownership and least privilege (access control lists that restrict who can read or write). 3. Ensure the directory is not shared across tenants/users. 4. Monitor for unexpected `*.blob` files or abnormal retry backlog growth.

GitHub Advisory Database
06

Elon Musk confirms xAI used OpenAI’s models to train Grok

securitypolicy
Apr 30, 2026

Elon Musk testified in court that his AI startup xAI used OpenAI's models to train its own AI system called Grok through model distillation (a technique where a larger AI model teaches a smaller one by transferring knowledge). Model distillation is a common practice in the AI industry, though it can be used legitimately within a single company or potentially misused by competitors trying to copy a rival's AI performance.

The Verge (AI)
07

GHSA-56c3-vfp2-5qqj: n8n-mcp's IPv4-mapped IPv6 addresses bypass SSRF protection in validateUrlSync(), enabling full SSRF for SDK embedders

security
Apr 30, 2026

A security flaw in n8n-mcp's URL validation allowed attackers to bypass SSRF (server-side request forgery, where an attacker tricks a server into making unwanted requests) protections using IPv4-mapped IPv6 addresses like `http://[::ffff:169.254.169.254]`. This could let an attacker who controls the `n8nApiUrl` input force the server to request sensitive data from cloud metadata endpoints, private networks, or localhost services, and the responses would be returned to the attacker along with API credentials.

Fix: Upgrade to **v2.47.14 or later** (via `npx n8n-mcp@latest` for npm or `docker pull ghcr.io/czlonkowski/n8n-mcp:latest` for Docker). If immediate upgrade is not possible, the source mentions three workarounds: (1) validate URLs before passing them to the SDK by rejecting IP literal hostnames and accepting only DNS-resolvable hostnames; (2) restrict outbound network traffic from the n8n-mcp process to private IP ranges (10.0.0.0/8, 172.16.0.0/12, 192.168.0.0/16), link-local addresses (169.254.0.0/16), and cloud metadata endpoints; and (3) do not accept user-controlled `n8nApiUrl` values and derive the URL from internal configuration only.

GitHub Advisory Database
08

OpenAI Rolls Out ‘Advanced’ Security Mode for At-Risk Accounts

security
Apr 30, 2026

OpenAI launched Advanced Account Security, an optional protection feature for high-risk ChatGPT and Codex users like journalists and dissidents that replaces passwords with physical security keys or passkeys to prevent account takeover attacks (when someone gains unauthorized access to an account). The feature also uses recovery keys instead of email/SMS for account recovery, enforces shorter login sessions, and sends alerts on sign-ins, making it much harder for attackers to breach accounts through phishing (tricking users into revealing login credentials) or social engineering (manipulating support staff).

Fix: OpenAI's explicitly mentioned mitigations for Advanced Account Security users include: (1) requiring two physical security keys or passkeys instead of passwords, (2) eliminating email and SMS recovery routes in favor of recovery keys, backup passkeys, or physical security keys, (3) blocking OpenAI support team access to recovery options to prevent social engineering attacks on support portals, (4) enforcing shorter sign-in windows and sessions before re-authentication is required, (5) generating login alerts that users can review in their dashboard, and (6) enabling data opt-out from model training by default. OpenAI also partnered with Yubico to offer lower-cost YubiKey bundles to these users. Members of OpenAI's Trusted Access for Cyber program must enable Advanced Account Security by June 1, 2024, or submit an alternative attestation of phishing-resistant authentication through enterprise single sign-on.

Wired (Security)
09

GHSA-rch3-82jr-f9w9: Jupyter Notebook Vulnerable to Authentication Token Theft via CommandLinker XSS

security
Apr 30, 2026

Jupyter Notebook has a stored XSS (cross-site scripting, a type of attack where malicious code runs in a user's browser when they view a webpage or file) vulnerability that lets attackers steal authentication tokens (credentials that prove who you are) by tricking users into clicking fake controls in malicious notebook files. An attacker who steals these tokens can take over a user's account, read files, run code, and access the system.

Fix: Update to Jupyter Notebook 7.5.6 or JupyterLab 4.5.7, which include patches. As a temporary workaround, disable the help extension by running: `jupyter labextension disable @jupyter-notebook/help-extension` and `jupyter labextension disable @jupyterlab/help-extension`. For additional hardening, disable command linker functionality by adding this to `overrides.json`: `{"@jupyterlab/apputils-extension:sanitizer": {"allowCommandLinker": false}}`.

GitHub Advisory Database
10

Red Agent and Claude Opus: Securing Production Targets at Scale

securityindustry
Apr 30, 2026

Wiz Red Agent is an AI security tool powered by Anthropic's Claude Opus models that automatically scans production environments (web applications and APIs) to find exploitable security vulnerabilities by reasoning like a human attacker. It analyzes over 150,000 applications weekly and has discovered thousands of previously unknown high and critical security risks across major organizations with zero false positives.

Wiz Research Blog
Prev1...1718192021...371Next
high

GHSA-cmrh-wvq6-wm9r: n8n-mcp webhook and API client paths has an authenticated SSRF

CVE-2026-44694GitHub Advisory DatabaseMay 8, 2026
May 8, 2026
high

CVE-2026-41487: Langfuse is an open source large language model engineering platform. From version 3.68.0 to before version 3.167.0, the

CVE-2026-41487NVD/CVE DatabaseMay 8, 2026
May 8, 2026
high

Claude in Chrome is taking orders from the wrong extensions

CSO OnlineMay 8, 2026
May 8, 2026