aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 69/371
VIEW ALL
01

CVE-2026-40100: FastGPT is an AI Agent building platform. Prior to 4.14.10.3, the /api/core/app/mcpTools/runTool endpoint accepts arbitr

security
Apr 10, 2026

FastGPT, an AI Agent building platform, has a vulnerability in versions before 4.14.10.3 where an endpoint accepts URLs without proper authentication checks, allowing unauthenticated attackers to perform SSRF (server-side request forgery, where an attacker tricks the server into making requests to internal network resources) attacks against internal systems. The vulnerability exists because the internal IP check is disabled by default.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Fix: Update FastGPT to version 4.14.10.3 or later, where this vulnerability is fixed.

NVD/CVE Database
02

CVE-2026-35651: OpenClaw versions 2026.2.13 through 2026.3.24 contain an ANSI escape sequence injection vulnerability in approval prompt

security
Apr 10, 2026

OpenClaw versions 2026.2.13 through 2026.3.24 have an ANSI escape sequence injection vulnerability (a bug where attackers can sneak special terminal control codes into the system) in approval prompts that allows attackers to trick the terminal display by manipulating tool metadata. This means an attacker could use malicious tool names containing these control sequences to make false information appear in approval prompts and permission logs.

NVD/CVE Database
03

Powell, Bessent discussed Anthropic's Mythos AI cyber threat with major U.S. banks

securitypolicy
Apr 10, 2026

Federal Reserve Chairman Jerome Powell and Treasury Secretary Scott Bessent met with major U.S. bank CEOs to discuss cyber risks from Anthropic's Mythos model, a new AI system with advanced capabilities for both offensive and defensive hacking. Anthropic released the model in limited capacity through Project Glasswing, a cybersecurity initiative involving major tech companies, and briefed government agencies on its cyber applications because of concerns that hackers could exploit its capabilities.

CNBC Technology
04

ChatGPT voice mode is a weaker model

research
Apr 10, 2026

ChatGPT's voice mode runs on an older, weaker model (GPT-4o era with a knowledge cutoff of April 2024) compared to other OpenAI products, even though talking to an AI might seem like it should use the smartest version. The article explains that OpenAI's highest-tier models perform much better on tasks like coding because those domains have clear, measurable success criteria (like whether unit tests pass) that make them easier to improve through reinforcement learning (training that rewards correct behaviors), and because business customers value these capabilities more.

Simon Willison's Weblog
05

Claude Mythos: Preparing for a World Where AI Finds and Exploits Vulnerabilities Faster Than Ever

securityresearch
Apr 10, 2026

Claude Mythos is a new AI model developed by Anthropic that can autonomously discover zero-day vulnerabilities (previously unknown security flaws) and create working exploits (tools that take advantage of those flaws) in major software like operating systems and web browsers. Although currently restricted to responsible organizations like Microsoft and Google, the source warns that similar capabilities will likely become publicly available within 12-18 months, leading to a surge in discovered vulnerabilities and requiring security teams to adopt new AI-focused strategies to defend against attacks.

Fix: The source explicitly recommends that security teams and vendors adopt the following strategies across three phases: (1) Short term: vendors should "invest in making sure that patching their products is as seamless and painless as possible, to support end-users dealing with the onslaught of new CVEs"; (2) Medium-to-long term: "plan to invest efforts into an AI-focused AppSec program (application security program), which will ensure you find the AI vulnerabilities before threat actors have a chance to exploit them."

Wiz Research Blog
06

CoreWeave stock pops 11% on deal to power Anthropic's Claude

industry
Apr 10, 2026

CoreWeave, a cloud infrastructure company that operates data centers with thousands of Nvidia graphics processing units (GPUs, specialized chips that speed up AI computations), announced a multi-year deal to provide computing power for Anthropic's Claude AI models. This deal means nine of the top ten AI model providers now use CoreWeave's platform, reflecting growing demand for the specialized infrastructure needed to run large AI systems at scale.

CNBC Technology
07

CVE-2026-40217: LiteLLM through 2026-04-08 allows remote attackers to execute arbitrary code via bytecode rewriting at the /guardrails/t

security
Apr 10, 2026

LiteLLM (a library for working with multiple AI models) versions through April 8, 2026 contain a vulnerability that allows remote attackers to execute arbitrary code (run commands they shouldn't be able to run) through bytecode rewriting (modifying compiled code) at a specific web endpoint called /guardrails/test_custom_code. This is a serious security flaw because attackers on the internet could potentially take control of systems running vulnerable versions.

NVD/CVE Database
08

Can Anthropic Keep Its Exploit-Writing AI Out of the Wrong Hands?

securitysafety
Apr 10, 2026

Anthropic has released a preview version of an AI model called Mythos that can apparently identify and exploit zero-days (previously unknown security vulnerabilities that hackers don't yet know about). The company says it has built in certain controls to try to prevent misuse of this powerful tool.

Dark Reading
09

Fear and loathing at OpenAI

policy
Apr 10, 2026

Sam Altman, CEO of OpenAI, experienced a brief firing and reinstatement that led to significant organizational changes, raising questions about his leadership of a major AI company. The New Yorker published an investigation examining Altman's tenure and whether he is the appropriate person to lead such a transformative technology.

The Verge (AI)
10

The Download: an exclusive Jeff VanderMeer story and AI models too scary to release

securitysafety
Apr 10, 2026

OpenAI has restricted the release of its new cybersecurity tool to select partners only due to security concerns, joining Anthropic in limiting AI model access over safety fears. The article also reports that Florida is investigating OpenAI's potential involvement in helping plan a mass shooting through ChatGPT, raising questions about AI's role in real-world harms.

MIT Technology Review
Prev1...6768697071...371Next