aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 44/371
VIEW ALL
01

Workspace agents

industry
Apr 22, 2026

Workspace agents are AI systems designed to automate repeatable workflows in your daily work by connecting to tools your team uses, rather than helping with one-off tasks. A workspace agent has three core components: a trigger (what starts it, like a schedule), a process with specialized skills (the steps it follows), and access to tools or systems (like Slack or a CRM). Unlike traditional deterministic workflows (where each step is explicitly defined and always the same), agents are probabilistic, meaning they use AI to interpret context and adjust their approach while staying within set instructions and guardrails.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

OpenAI Blog
02

Anthropic’s most dangerous AI model just fell into the wrong hands

security
Apr 22, 2026

Anthropic's Mythos AI model, a tool designed to find security weaknesses in software, was accessed by unauthorized users through a private online forum using a contractor's credentials and basic internet research techniques. The model is capable of identifying and exploiting vulnerabilities (security flaws) in major operating systems and web browsers, which is why Anthropic warned it could be dangerous if misused.

The Verge (AI)
03

Anthropic bets on EPSS for the coming bug surge

securityindustry
Apr 22, 2026

AI tools like Anthropic's Mythos can find software vulnerabilities much faster than before, creating a problem: security teams must decide which vulnerabilities to fix first among thousands of options. Anthropic recommends using EPSS (Exploit Prediction Scoring System, a machine learning model that predicts which vulnerabilities are likely to be exploited in the next 30 days) to prioritize which vulnerabilities need immediate attention, similar to how weather forecasters predict whether you'll need an umbrella.

Fix: According to Anthropic's guidance: 'Patching the KEV (CISA's Known Exploited Vulnerabilities catalog) list first, and then everything above a chosen EPSS threshold will help you turn thousands of open CVEs into a manageable queue.' EPSS scores are machine-driven and can be applied across all CVEs with scores published daily, and have been incorporated into more than 120 security vendors' products.

CSO Online
04

Anthropic investigates report of rogue access to hack-enabling Mythos AI

security
Apr 22, 2026

Anthropic is investigating a report that unauthorized users gained access to Mythos, an AI model designed to detect cybersecurity vulnerabilities that the company has kept private because it could be misused to enable cyber-attacks. A small group of people allegedly accessed the model without permission, prompting the company to look into the incident.

The Guardian Technology
05

Cohere AI Terrarium Sandbox Flaw Enables Root Code Execution, Container Escape

security
Apr 22, 2026

Terrarium, a Python sandbox developed by Cohere AI for running untrusted code in containers, has a critical vulnerability (CVE-2026-5752, CVSS 9.3) that allows attackers to execute arbitrary code with root privileges through JavaScript prototype chain traversal (a technique where attackers manipulate how JavaScript looks up object properties to access restricted functionality). Since the project is no longer maintained, a patch is unlikely, but CERT/CC recommends several defensive measures.

Fix: CERT/CC advises the following mitigations: Disable features that allow users to submit code to the sandbox, if possible. Segment the network to limit the attack surface and prevent lateral movement. Deploy a Web Application Firewall to detect and block suspicious traffic, including attempts to exploit the vulnerability. Monitor container activity for signs of suspicious behavior. Limit access to the container and its resources to authorized personnel only. Use a secure container orchestration tool to manage and secure containers. Ensure that dependencies are up-to-date and patched.

The Hacker News
06

Changes to GitHub Copilot Individual plans

industry
Apr 21, 2026

GitHub Copilot changed its pricing and usage limits for individual users because agentic workflows (AI agents that run long tasks automatically) consume far more computing resources than expected, with some users burning tokens (units of text processed by the AI) at much higher rates than before. The changes include pausing new individual plan signups, moving the most advanced Claude Opus 4.7 model to a more expensive $39/month tier, and switching to token-based usage limits tracked per session and per week instead of per-request charging.

Simon Willison's Weblog
07

Is Claude Code going to cost $100/month? Probably not - it's all very confusing

industry
Apr 21, 2026

Anthropic briefly updated its pricing page to move Claude Code (an AI coding agent feature) from the $20/month Pro plan to exclusive availability on $100-200/month Max plans, but quickly reverted the change after public backlash. Anthropic's Head of Growth claimed this was a test affecting only ~2% of new signups, though the change was widely visible and caused significant concern about affordability and lack of transparency.

Simon Willison's Weblog
08

Introducing OpenAI Privacy Filter

securityprivacy
Apr 21, 2026

OpenAI released Privacy Filter, an open-weight AI model designed to detect and remove personally identifiable information (PII, such as names, addresses, phone numbers, and account details) from text. The model uses context-aware language understanding rather than simple pattern matching, can run locally on a user's device to keep sensitive data from being sent to servers, and achieves state-of-the-art performance on privacy detection benchmarks. Developers can use, fine-tune, and integrate Privacy Filter into their own applications to build stronger privacy protections into AI systems.

OpenAI Blog
09

SpaceX cuts a deal to maybe buy Cursor for $60 billion

industry
Apr 21, 2026

SpaceX has announced a deal to either acquire Cursor, an AI-powered coding platform, for $60 billion or pay a $10 billion fee instead. This move aims to help xAI compete with other companies in the AI coding space, as major tech firms like Google and OpenAI are also investing heavily in their own AI programming tools.

The Verge (AI)
10

CVE-2026-40933: Flowise is a drag & drop user interface to build a customized large language model flow. Prior to 3.1.0, due to unsafe s

security
Apr 21, 2026

Flowise, a tool with a visual interface for building customized AI flows, has a vulnerability before version 3.1.0 where authenticated attackers can execute arbitrary commands on the server. The flaw exists in the MCP (model context protocol) adapter's handling of stdio commands, where input sanitization checks fail to prevent attackers from combining safe commands like "npx" with code execution arguments to run malicious commands on the underlying operating system.

Fix: Update Flowise to version 3.1.0 or later, where this vulnerability is fixed.

NVD/CVE Database
Prev1...4243444546...371Next