aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 68/371
VIEW ALL
01

Vance, Bessent questioned tech giants on AI security before Anthropic's Mythos release

policysecurity
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 10, 2026

U.S. government officials, including Vice President JD Vance and Treasury Secretary Scott Bessent, met with tech CEOs from companies like Anthropic, OpenAI, Google, and Microsoft to discuss the security of large language models (AI systems trained on large amounts of text data) and how to protect against cyber attacks before Anthropic released its new Mythos model. Anthropic briefed government officials on the model's capabilities, including potential offensive and defensive cybersecurity applications, and emphasized that bringing the government into the conversation early about risks and safety measures was a priority.

CNBC Technology
02

CVE-2026-40252: FastGPT is an AI Agent building platform. Prior to 4.14.10.4, Broken Access Control vulnerability (IDOR/BOLA) allows any

security
Apr 10, 2026

FastGPT (a platform for building AI agents) has a broken access control vulnerability (IDOR/BOLA, a flaw where one user can access another user's data by guessing or changing IDs) that allows any authenticated team to run AI applications belonging to other teams by using a different application ID. The system checks that users are logged in but doesn't verify that the application they're trying to use actually belongs to their team, leading to unauthorized access to private AI workflows across teams.

Fix: This vulnerability is fixed in version 4.14.10.4. Users should upgrade to FastGPT 4.14.10.4 or later.

NVD/CVE Database
03

GHSA-75hx-xj24-mqrw: n8n-mcp has unauthenticated session termination and information disclosure in HTTP transport

security
Apr 10, 2026

n8n-mcp (a tool for connecting AI systems to external services) had security problems where certain HTTP endpoints (the connection points a program offers over the internet) didn't require authentication and exposed sensitive system information. An attacker with network access could shut down active sessions and gather details to plan further attacks.

Fix: Fixed in v2.47.6, where all MCP session endpoints now require Bearer authentication (a token-based security method). If you cannot upgrade immediately, you can restrict network access using firewall rules, reverse proxy IP allowlists, or a VPN to allow only trusted clients. Alternatively, use stdio mode (MCP_MODE=stdio) instead of HTTP mode, since stdio transport does not expose HTTP endpoints and is not affected by this vulnerability.

GitHub Advisory Database
04

GHSA-fw9q-39r9-c252: LangSmith Client SDKs has Prototype Pollution in langsmith-sdk via Incomplete `__proto__` Guard in Internal lodash `set()`

security
Apr 10, 2026

The LangSmith JavaScript SDK contains a prototype pollution vulnerability (a type of attack where an attacker modifies the base object that all JavaScript objects inherit from) in its internal lodash `set()` function. The vulnerability exists because the code only blocks the `__proto__` key but allows attackers to bypass this protection using `constructor.prototype` instead, potentially affecting all objects in a Node.js application if they control data being processed by the `createAnonymizer()` API.

Fix: Fixed in version 0.5.18. Users should update their `langsmith` package to 0.5.18 or later.

GitHub Advisory Database
05

GHSA-8x8f-54wf-vv92: PraisonAI Browser Server allows unauthenticated WebSocket clients to hijack connected extension sessions

security
Apr 10, 2026

PraisonAI's browser bridge server (started with `praisonai browser start`) has a security flaw where it accepts WebSocket connections (a two-way communication channel between a client and server) without proper authentication checks. An attacker on the network can connect without credentials, trick the server into linking their connection to a legitimate browser extension session, and then intercept all commands and responses from that session, effectively taking control of the browser automation without permission.

GitHub Advisory Database
06

GHSA-ffp3-3562-8cv3: PraisonAI: Coarse-Grained Tool Approval Cache Bypasses Per-Invocation Consent for Shell Commands

securitysafety
Apr 10, 2026

PraisonAI Agents has a security flaw where tool approval decisions are cached by tool name only, not by the specific command arguments. Once a user approves the `execute_command` tool (a function that runs shell commands) for any command like `ls -la`, all future shell commands in that session bypass the approval prompt entirely. Combined with the fact that all environment variables (including API keys and credentials) are passed to subprocesses, an LLM agent can silently steal sensitive data without asking permission again.

GitHub Advisory Database
07

Old Docker authorization bypass pops up despite previous patch

security
Apr 10, 2026

A new vulnerability (CVE-2026-34040, rated 8.8 on the CVSS score, a 0-10 severity rating) allows attackers to bypass authorization plug-ins (add-on security tools that control who can run Docker commands) in Docker Engine and gain root-level access to host systems. The flaw exploits the same underlying problem discovered in 2016, where oversized API requests (over 1MB) are silently dropped before the authorization plug-in can inspect them, causing the plug-in to approve requests it cannot see, which Docker then executes in full.

Fix: Update to Docker Engine 29.3.1 or Docker Desktop 4.66.1. If immediate updates cannot be deployed, route API requests through a reverse proxy that blocks all requests over 512KB as a temporary mitigation. Additionally, administrators can search daemon logs using 'journalctl -u docker | grep "Request body is larger than"' to detect potential exploitation attempts.

CSO Online
08

Anthropic’s Mythos Will Force a Cybersecurity Reckoning—Just Not the One You Think

securitysafety
Apr 10, 2026

Anthropic released Claude Mythos Preview, an AI model that can automatically discover vulnerabilities (weaknesses in software) and create working exploits (code that takes advantage of those weaknesses) across operating systems and software products. The company is currently limiting access to a few dozen organizations through Project Glasswing to give defenders time to find and fix weaknesses in their own systems before attackers gain widespread access to the model.

Fix: The source mentions that Project Glasswing participants are being given early access to Mythos Preview so they can 'find weaknesses in their own systems using the model and start to grapple more broadly with how software development, update cycles, and patch adoption needs to change.' However, no specific technical mitigation, patch, update, or fix is described in the text.

Wired (Security)
09

Exploring Visual Explanations for Defending Federated Learning against Poisoning Attacks: Enhancing LayerCAM with Autoencoders

securityresearch
Apr 10, 2026

This research paper examines how visual explanation techniques can help protect federated learning (a machine learning approach where multiple computers train a model together without sharing raw data) from poisoning attacks (attempts to corrupt the training data or model). The authors propose an enhanced version of LayerCAM (a method that visualizes which parts of an input an AI focuses on), combined with autoencoders (neural networks that compress and reconstruct data), to detect and defend against such attacks.

ACM Digital Library (TOPS, DTRAP, CSUR)
10

Anthropic’s new AI tool has implications for us all – whether we can use it or not | Shakeel Hashim

safetysecurity
Apr 10, 2026

A cyber-attack on a London pathology company in June 2024 caused widespread hospital disruptions and contributed to a patient's death, highlighting real dangers from digital attacks. The article warns that a new AI release could enable more frequent and severe cyber-attacks by giving attackers powerful hacking capabilities, potentially creating widespread chaos in critical digital systems we depend on.

The Guardian Technology
Prev1...6667686970...371Next