aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 41/371
VIEW ALL
01

Microsoft taps Anthropic’s Mythos to strengthen secure software development

securityindustry
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 23, 2026

Microsoft is integrating Anthropic's Mythos, an advanced AI model, into its Security Development Lifecycle to help find software vulnerabilities (security flaws in code) and strengthen code earlier in development. While this move signals that AI is becoming central to how major software companies build secure products, analysts note that powerful AI models like Mythos could also make it faster for attackers to find and exploit vulnerabilities, raising concerns about the dual-use nature of these tools.

CSO Online
02

Anthropic looks to hire six-figure role for negotiating data center deals to fuel Europe AI expansion

industry
Apr 23, 2026

Anthropic is hiring for a senior role to negotiate data center deals in Europe to support its AI expansion, as the company secures major infrastructure commitments like a $100+ billion spending plan with Amazon Web Services and capacity deals with Broadcom. The company is specifically targeting data center capacity in major European hubs (Frankfurt, London, Amsterdam, Paris, Dublin) and regions like the Nordics, where cheap energy makes AI infrastructure more affordable. This move reflects a broader industry trend, with Microsoft, OpenAI, and other AI companies also expanding their European data center operations.

CNBC Technology
03

CVE-2026-41679: Paperclip is a Node.js server and React UI that orchestrates a team of AI agents to run a business. Prior to version 202

security
Apr 22, 2026

Paperclip is a Node.js server (a JavaScript runtime that runs outside web browsers) with a React UI (a framework for building user interfaces) that manages multiple AI agents to automate business tasks. Before version 2026.416.0, an attacker without any login credentials could gain full remote code execution (the ability to run arbitrary commands on the target system) on any publicly accessible Paperclip instance using its default settings, simply by knowing the server's address and making six automated API calls (requests to the server's functions).

Fix: Update to version 2026.416.0, which patches the vulnerability.

NVD/CVE Database
04

CVE-2026-41208: Paperclip is a Node.js server and React UI that orchestrates a team of AI agents to run a business. Versions of @papercl

security
Apr 22, 2026

Paperclip is a Node.js server and React UI that manages multiple AI agents to run a business. Versions before 2026.416.0 have a privilege escalation vulnerability where an attacker with an agent API key (a credential that identifies an agent) can trick the system into running arbitrary OS commands (unauthorized instructions executed on the computer) on the Paperclip server by injecting malicious commands into a configuration field that the server later executes.

Fix: @paperclipai/server version 2026.416.0 fixes the issue.

NVD/CVE Database
05

Claude Mythos signals a new era in AI-driven security, finding 271 flaws in Firefox

securityindustry
Apr 22, 2026

Claude Mythos, an AI model from Anthropic, discovered 271 vulnerabilities in Firefox 148, more than ten times what previous AI tools found, demonstrating AI's growing ability to uncover security bugs at scale. All 271 flaws were fixed in Firefox 150's release. While the AI isn't finding entirely new types of bugs, it's closing gaps in vulnerability detection that fuzzing (automated testing that uncovers bugs in source code) and human teams had previously missed, potentially shifting the balance in favor of defenders.

Fix: All 271 vulnerabilities discovered in Firefox 148 have been fixed in Firefox 150.

CSO Online
06

CVE-2026-6874: A vulnerability was determined in ericc-ch copilot-api up to 0.7.0. This impacts an unknown function of the file /token

security
Apr 22, 2026

A vulnerability (CVE-2026-6874) was found in ericc-ch copilot-api version 0.7.0 and earlier that affects the /token file's Header Handler component. An attacker can manipulate the Host argument to exploit reliance on reverse DNS resolution (looking up a domain name from an IP address), potentially allowing remote access to systems where the attacker has login credentials.

NVD/CVE Database
07

CVE-2026-39987: Marimo Remote Code Execution Vulnerability

security
Apr 22, 2026

Marimo has a pre-authorization remote code execution vulnerability (RCE, where an attacker can run commands on a system they don't own) that allows unauthenticated attackers to gain shell access and execute arbitrary commands without needing to log in first. This vulnerability is actively being exploited in real-world attacks.

Fix: Apply mitigations per vendor instructions, follow applicable BOD 22-01 guidance for cloud services, or discontinue use of the product if mitigations are unavailable.

CISA Known Exploited Vulnerabilities
08

GPT-5.5 Bio Bug Bounty

securitysafety
Apr 22, 2026

OpenAI is running a bug bounty program called the Bio Bug Bounty for GPT-5.5, inviting security researchers to find universal jailbreaks (methods to bypass safety restrictions with a single prompt) that can defeat five biology safety questions. The program offers $25,000 for the first successful universal jailbreak and smaller awards for partial results, with applications open from April 23 to June 22, 2026 and testing running through July 27, 2026.

OpenAI Blog
09

IBM CEO Krishna says Iran, other uncertainty is weighing on company's outlook

securityindustry
Apr 22, 2026

IBM CEO Arvind Krishna stated that geopolitical uncertainty, particularly the Iran conflict, is causing the company to provide cautious financial guidance despite beating first-quarter earnings expectations. He also expressed concerns about potential economic slowdowns affecting consumer spending and European growth, though he noted IBM's Middle East business performed well. Additionally, Krishna discussed how new AI models like Anthropic's Mythos, which can find security vulnerabilities at unprecedented speed, will likely be replicated by competitors and pose significant cybersecurity concerns that have caught the attention of U.S. government officials.

CNBC Technology
10

OpenAI now lets teams make custom bots that can do work on their own

industry
Apr 22, 2026

OpenAI has released workspace agents (AI systems that can independently perform tasks) for users on Business, Enterprise, Edu, and Teachers plans within ChatGPT. These agents can handle business tasks like gathering product feedback and drafting emails, building on growing interest in autonomous AI agents across the industry.

The Verge (AI)
Prev1...3940414243...371Next