aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 55/371
VIEW ALL
01

GHSA-48m6-ch88-55mj: Flowise: Improper Mass Assignment in Account Registration Enables Unauthorized Organization Association

security
Apr 16, 2026

Flowise Cloud has a mass assignment vulnerability (JSON injection, where attackers can hide malicious data in JSON input) in its account registration endpoint that allows unauthenticated attackers to inject server-managed fields like organization IDs and role assignments during account creation. This breaks trust boundaries in the multi-tenant environment (a system serving multiple separate organizations) by letting attackers associate their new accounts with existing organizations they don't own, gaining unauthorized access and escalated privileges.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

GitHub Advisory Database
02

GHSA-9wc7-mj3f-74xv: Flowise: Code Injection in CSVAgent leads to Authenticated RCE

security
Apr 16, 2026

Flowise's CSVAgent has a code injection vulnerability where user-provided custom Pandas CSV read code is inserted directly into executable Python code without sanitization, allowing an authenticated attacker to execute arbitrary commands on the server (RCE, or remote code execution). An attacker can create a malicious chat flow and trigger it via API requests to run commands like `os.system()` through the `pyodide` Python runtime.

GitHub Advisory Database
03

GHSA-f228-chmx-v6j6: Flowise: Remote code execution vulnerability in AirtableAgent.ts caused by lack of input verification when using `Pandas`.

security
Apr 16, 2026

Flowise's AirtableAgent has a remote code execution (RCE, where an attacker can run commands on a system they don't own) vulnerability because user input is inserted directly into Python code without sanitization. An attacker can use prompt injection (tricking an AI by hiding instructions in its input) to bypass the intended behavior and execute arbitrary code when the system processes Pandas (a Python library for working with data) operations.

GitHub Advisory Database
04

llm-anthropic 0.25

industry
Apr 16, 2026

Release llm-anthropic 0.25 adds a new Claude model (claude-opus-4.7) with advanced thinking capabilities, introduces options to display and adapt AI reasoning output, raises the default token limits (the maximum length of AI-generated responses) for all models, and removes outdated code that was no longer needed for older models.

Simon Willison's Weblog
05

Google will let users connect their photos to the Gemini chatbot and Nano Banana

securityprivacy
Apr 16, 2026

Google is connecting its Gemini chatbot to users' personal Google Photos library through a feature called Nano Banana (an image generation tool, meaning software that creates pictures from text descriptions). Users who opt in to Personal Intelligence (a feature that links Google apps together for customized responses) can ask Gemini to generate images based on their private photos, like "create a claymation image of me and my family," without manually uploading photos each time.

CNBC Technology
06

Qwen3.6-35B-A3B on my laptop drew me a better pelican than Claude Opus 4.7

industry
Apr 16, 2026

A blogger compared two newly released AI models (Qwen3.6-35B-A3B and Claude Opus 4.7) by asking them to generate SVG images (scalable vector graphics, a format for drawing pictures with code) of pelicans and flamingos performing tasks like riding bicycles. The Qwen model, running on a laptop as a quantized version (a compressed version that uses less computer memory), produced better images than Anthropic's Claude Opus 4.7, though the blogger notes this creative task may not reflect which model is actually more useful for real-world problems.

Simon Willison's Weblog
07

OpenAI’s big Codex update is a direct shot at Anthropic’s Claude Code

industry
Apr 16, 2026

OpenAI has updated Codex, its agentic coding system (an AI that can independently perform multi-step coding tasks), to control desktop applications, generate images, and remember previous interactions. The new features let Codex operate apps in the background without interrupting user work and allow multiple agents (separate AI instances) to work simultaneously, which OpenAI says is useful for testing frontend changes and working with applications that don't have APIs (standardized ways for software to communicate).

The Verge (AI)
08

Google’s AI Mode update lets you open links without leaving the page

industry
Apr 16, 2026

Google is updating AI Mode (a chatbot-like search feature built into Chrome) with a new feature that opens source links in a side-by-side view instead of in a new tab, letting you compare the website content with your chat conversation at the same time. This upgrade makes it easier to ask follow-up questions about information you're reading without switching between multiple windows.

The Verge (AI)
09

Hackers exploit Marimo flaw to deploy NKAbuse malware from Hugging Face

security
Apr 16, 2026

Hackers are exploiting a critical vulnerability in Marimo (a Python notebook tool) called CVE-2026-39987 (remote code execution, where attackers can run commands on systems they don't own) to deploy NKAbuse malware from Hugging Face Spaces (a platform for sharing AI applications). The attacks began within 10 hours of technical details becoming public, with attackers using fake application names to trick users into downloading malware that steals credentials and allows remote control of infected systems.

Fix: Users should upgrade to Marimo version 0.23.0 or later immediately. If upgrading is not possible, block external access to the '/terminal/ws' endpoint using a firewall, or block it entirely.

BleepingComputer
10

Anthropic rolls out Claude Opus 4.7, an AI model that is less risky than Mythos

safetyindustry
Apr 16, 2026

Anthropic released Claude Opus 4.7, a new AI model that excels at software engineering and following instructions but has intentionally reduced capabilities in cybersecurity tasks compared to its more powerful Claude Mythos Preview model. The company implemented safeguards that automatically detect and block requests for prohibited or high-risk cybersecurity uses, and is using this release to learn how to safely deploy more powerful models in the future.

Fix: Anthropic released Claude Opus 4.7 with safeguards that automatically detect and block requests indicating prohibited or high-risk cybersecurity uses. The company also experimented with efforts to 'differentially reduce' Claude Opus 4.7's cyber capabilities during training, and encourages security professionals interested in legitimate cybersecurity purposes to apply through a formal verification program.

CNBC Technology
Prev1...5354555657...371Next