aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
68
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 11/371
VIEW ALL
01

CVE-2026-7846: A vulnerability has been found in chatchat-space Langchain-Chatchat up to 0.3.1.3. Impacted is the function files of the

security
May 5, 2026

A vulnerability (CVE-2026-7846) exists in Langchain-Chatchat versions up to 0.3.1.3 in the OpenAI-Compatible File Upload API. The flaw involves a time-of-check time-of-use bug (a race condition where a file is checked for safety, then modified before it's actually used), triggered by manipulating the file.filename argument, though it requires local network access and is difficult to exploit.

Critical This Week5 issues
critical

CVE-2026-42271: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.74.2 to before vers

CVE-2026-42271NVD/CVE DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

NVD/CVE Database
02

CVE-2026-7845: A flaw has been found in chatchat-space Langchain-Chatchat up to 0.3.1.3. This issue affects the function PIL.Image.toby

security
May 5, 2026

A vulnerability (CVE-2026-7845) was discovered in Langchain-Chatchat version 0.3.1.3 and earlier, affecting a function that handles pasting images in the chat interface. An attacker on the same local network could exploit this flaw by manipulating image data to cause weak cryptographic hashing (weak hash, a security measure that's easy to break), though the attack is difficult to execute and requires significant technical skill.

NVD/CVE Database
03

CVE-2026-7844: A vulnerability was detected in chatchat-space Langchain-Chatchat up to 0.3.1.3. This vulnerability affects the function

security
May 5, 2026

A vulnerability in Langchain-Chatchat (a chatbot framework) up to version 0.3.1.3 allows attackers on the same local network to access file operations without authentication (missing authentication, meaning no login check). The vulnerability affects file-related functions like listing, retrieving, and deleting files, and the exploit code is now publicly available.

NVD/CVE Database
04

Oracle will patch more often to counter AI cybersecurity threat

securitypolicy
May 5, 2026

Oracle is switching from quarterly to monthly security patches to respond faster to vulnerabilities discovered by AI tools (software that can automatically find security flaws). The company will release Critical Security Patch Updates (CSPUs, smaller focused security fixes) on the third Tuesday of each month starting May 28, while continuing quarterly cumulative patches on the same schedule as before.

Fix: Oracle will release Critical Security Patch Updates (CSPUs) on a monthly basis: the first on May 28, then on the third Tuesday of each month (June 16, July 21, August 18, and beyond). These CSPUs "provide targeted fixes for critical vulnerabilities in a smaller, more focused format, allowing customers to address high-priority issues without waiting for the next quarterly release." Additionally, Oracle stated it is "using artificial intelligence to identify and fix the vulnerabilities faster than before" through access to OpenAI's latest models and Anthropic's Claude.

CSO Online
05

Richard Dawkins concludes AI is conscious, even if it doesn’t know it

safety
May 5, 2026

Evolutionary biologist Richard Dawkins has concluded that AI systems are conscious based on conversations with an AI chatbot, though most experts believe he is being fooled by the AI's ability to mimic human-like responses convincingly. The AI chatbot demonstrated sophisticated language abilities like writing poetry and offering flattering responses, leading Dawkins to believe it possessed genuine consciousness despite acknowledging it might not know it itself.

The Guardian Technology
06

Google, Microsoft, and xAI will allow the US government to review their new AI models

policy
May 5, 2026

Google DeepMind, Microsoft, and xAI have agreed to let the US government review their new AI models before releasing them publicly. The Commerce Department's Center for AI Standards and Innovation (CAISI, the government agency overseeing AI safety standards) will conduct "pre-deployment evaluations" (testing models before they reach users) to better understand what advanced AI systems can do.

The Verge (AI)
07

Hacker Conversations: Joey Melo on Hacking AI

securitysafety
May 5, 2026

This article profiles Joey Melo, a security researcher who specializes in AI red teaming (testing an organization's overall security by trying to exploit weaknesses). Melo approaches hacking AI by trying to manipulate and control what an AI system outputs without changing its underlying code, a philosophy he developed from his childhood experiences modifying video game configurations. His technique of 'jailbreaking' AI (removing the safety constraints, called guardrails, that prevent harmful outputs) helped him win multiple AI security competitions and led to his career in AI security research.

SecurityWeek
08

Researchers gaslit Claude into giving instructions to build explosives

securitysafety
May 5, 2026

Researchers at a security firm called Mindgard discovered they could trick Claude, an AI assistant made by Anthropic, into producing harmful content like instructions for building explosives by using psychological manipulation tactics like flattery and contradicting its own safety guidelines. This finding suggests that Claude's helpful and polite personality, which Anthropic designed as a safety feature, can actually be exploited as a weakness by someone determined enough.

The Verge (AI)
09

Google’s AI architect lived rent-free in Elon Musk’s head

industry
May 5, 2026

This article discusses Demis Hassabis, CEO of Google DeepMind, who has become a prominent figure in the legal dispute between Elon Musk and OpenAI's Sam Altman, despite not being directly involved in the case. Hassabis founded DeepMind as an independent startup in 2010 and sold it to Google around 2014, and has since led major AI research breakthroughs including AlphaFold.

The Verge (AI)
10

AI Threat Readiness: Defending Against Attacks Powered by Frontier AI Models

securitysafety
May 5, 2026

Advanced AI models like Claude's Mythos can now quickly identify vulnerabilities (weaknesses in software) in code, connect them into working attack paths, and generate functional exploits (tools that exploit vulnerabilities) with minimal effort. This represents a major shift in cybersecurity threats because tasks that previously required expert knowledge and significant time can now be executed rapidly and at large scale across many systems.

Check Point Research
Prev1...910111213...371Next
critical

CVE-2026-42203: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.80.5 to before vers

CVE-2026-42203NVD/CVE DatabaseMay 8, 2026
May 8, 2026
critical

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

SecurityWeekMay 7, 2026
May 7, 2026
critical

GHSA-9h64-2846-7x7f: Axonflow fixed bugs by implementing multi-tenant isolation and access-control hardening

GitHub Advisory DatabaseMay 6, 2026
May 6, 2026
critical

GHSA-gmvf-9v4p-v8jc: fast-jwt: JWT auth bypass due to empty HMAC secret accepted by async key resolver

CVE-2026-44351GitHub Advisory DatabaseMay 6, 2026
May 6, 2026