aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 47/371
VIEW ALL
01

v0.14.21

security
Apr 20, 2026

LlamaIndex v0.14.21 is a maintenance release that fixes several bugs in the core library, including a KeyError (an error when looking up a value in a data structure that doesn't exist) in the DocumentSummaryIndex deletion function, handling of output formatting errors, and UTF-8 encoding issues in file operations. The release also updates dependencies across many embedding and indexing modules to keep the library's supporting code current.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Fix: Update to llama-index-core version 0.14.21 or later. The fixes are included in this release version, which addresses the KeyError in DocumentSummaryIndex.delete_nodes, ValueError and TypeError from structured output failures, UTF-8 encoding issues in the persistence layer, and the Message Block Buffer Resolution breaking change.

LlamaIndex Security Releases
02

Scaling Codex to enterprises worldwide

industry
Apr 20, 2026

Codex, an AI tool that generates code and assists with software development tasks, has grown from 3 million to 4 million weekly users and is now being adopted by major enterprises like Virgin Atlantic, Notion, and Cisco to speed up development workflows. OpenAI is expanding Codex adoption through a program called Codex Labs, which provides expert guidance to organizations, and by partnering with global consulting firms (like Accenture and Infosys) to help enterprises integrate Codex into their software development processes at scale.

OpenAI Blog
03

Amazon to invest up to another $25 billion in Anthropic as part of AI infrastructure deal

industry
Apr 20, 2026

Amazon is investing up to $25 billion more in Anthropic, an AI company known for its Claude AI models (large language models, or LLMs, which are AI systems trained on vast amounts of text to generate human-like responses), on top of an earlier $8 billion investment. As part of this deal, Anthropic will spend over $100 billion on Amazon's cloud services and custom AI chips over the next decade to expand its computing capacity (the processing power needed to train and run AI models). Anthropic made this agreement because its infrastructure has been strained by rapidly growing demand from enterprise customers and users of Claude.

CNBC Technology
04

CVE-2026-33626: LMDeploy is a toolkit for compressing, deploying, and serving large language models. Versions prior to 0.12.3 have a Ser

security
Apr 20, 2026

LMDeploy, a toolkit for compressing, deploying, and serving large language models, contains a Server-Side Request Forgery vulnerability (SSRF, a flaw that lets attackers trick a server into making requests to unintended targets) in versions before 0.12.3. The vulnerability exists in the `load_image()` function, which downloads images from URLs without checking if those URLs point to private or internal systems, potentially allowing attackers to access sensitive cloud services and internal networks.

Fix: Update LMDeploy to version 0.12.3 or later, which patches the issue.

NVD/CVE Database
05

Optimizing stealthiness in universal adversarial perturbations via class-selective and perceptual similarity metrics

securityresearch
Apr 20, 2026

Universal Adversarial Perturbations (UAPs, tiny modifications to images that fool AI models across many different inputs) are security threats to deep learning systems, but existing methods make attacks obvious because they either look wrong to humans or cause suspicious misclassifications. This paper presents Stealthy-UAP, a framework that makes UAPs harder to detect by targeting only semantically related classes (so misclassifications seem plausible) and optimizing perturbations to match how humans actually perceive images.

Elsevier Security Journals
06

llm-openrouter 0.6

industry
Apr 20, 2026

The llm-openrouter tool, version 0.6, added a new 'refresh' command that lets users update their list of available AI models without waiting for the cached (temporarily stored) list to expire. This feature was created so users could access newly available models on OpenRouter immediately.

Simon Willison's Weblog
07

CVE-2026-6662: A vulnerability was found in ericc-ch copilot-api up to 0.7.0. The impacted element is the function cors of the file src

security
Apr 20, 2026

A vulnerability (CVE-2026-6662) was found in ericc-ch copilot-api versions up to 0.7.0 in the CORS function (a security feature that controls which websites can access an API from a web browser) of the token endpoint. The flaw allows a permissive cross-domain policy with untrusted domains, meaning attackers from other websites could potentially access the API remotely, and the exploit has been publicly disclosed.

NVD/CVE Database
08

ThreatMAMBA: Achieving High-Robustness Cyber Threat Attribution During the Evolution of Attacks

researchsecurity
Apr 20, 2026

Cyber Threat Attribution (CTA) is the process of identifying who carried out a cyberattack by analyzing evidence from the attack. This paper introduces ThreatMAMBA, an AI framework that improves CTA by building knowledge graphs from threat intelligence data (IOCs, or indicators of compromise that identify malicious activity; TTPs, or tactics and techniques used by attackers; and temporal relationships) and using machine learning to identify attackers even in the early stages of ongoing attacks. The system showed significant improvements in accuracy at different stages of attack development, suggesting it can provide reliable attribution information quickly during real incidents.

IEEE Xplore (Security & AI Journals)
09

Anthropic MCP Design Vulnerability Enables RCE, Threatening AI Supply Chain

security
Apr 20, 2026

Researchers discovered a critical vulnerability in Anthropic's Model Context Protocol (MCP, a system that allows AI models to interact with external tools and data) that allows attackers to run arbitrary commands on systems using vulnerable implementations. The flaw affects over 7,000 publicly accessible servers and has been found in popular AI projects like LangChain and LiteLLM, but Anthropic has declined to fix the underlying architectural issue, leaving developers responsible for protecting against it.

Fix: The source recommends several mitigations: block public IP access to sensitive services, monitor MCP tool invocations, run MCP-enabled services in a sandbox (an isolated test environment), treat external MCP configuration input as untrusted, and only install MCP servers from verified sources. Additionally, some vendors have issued patches for their specific products (LiteLLM, Bisheng, and DocsGPT are noted as patched).

The Hacker News
10

CISOs reshape their roles as business risk strategists

policyindustry
Apr 20, 2026

CISOs (chief information security officers, the top security leaders at companies) are expanding their roles beyond traditional cybersecurity to become broader business risk strategists who manage strategic, operational, and financial risks across their entire organizations. This shift reflects the fact that nearly all business operations are now digital, making any cyber risk a material business risk, and has accelerated since the rise of generative AI (AI systems like ChatGPT that can create new content) and agentic AI (AI systems that can take independent actions). Research shows that most CISOs now share responsibility for enterprise risk management with other executives and are expected to unify regulatory requirements, company risk tolerance, and security controls into a single operating model.

CSO Online
Prev1...4546474849...371Next