aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
68
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 12/371
VIEW ALL
01

Critical Bug Could Expose 300,000 Ollama Deployments to Information Theft

security
May 5, 2026

A critical vulnerability called Bleeding Llama (CVE-2026-7482, CVSS score 9.3) affects Ollama, an open source tool for running large language models (LLMs, AI systems trained on massive amounts of text) on local machines. An attacker can exploit a heap out-of-bounds read (a bug where the program accesses memory it shouldn't) to steal sensitive data like API keys, passwords, and user messages from approximately 300,000 internet-exposed Ollama deployments without needing any authentication.

Critical This Week5 issues
critical

CVE-2026-42271: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.74.2 to before vers

CVE-2026-42271NVD/CVE DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Fix: The vulnerability was addressed in Ollama version 0.17.1. Organizations should apply this fix as soon as possible, restrict network access to their deployments, deploy an authentication proxy (a middleman service that requires login), use network segmentation (isolating systems from the internet), and audit running instances for internet exposure. Any instance accessible from the internet should be considered compromised.

SecurityWeek
02

C/C++ checklist challenges, solved

securityresearch
May 5, 2026

This article explains two security bugs found in C/C++ code samples: a Linux ping program vulnerable to command injection because inet_ntoa (a function that converts IP addresses to text) returns a pointer to a global buffer that gets overwritten by subsequent calls, allowing an attacker to bypass IP validation checks; and a Windows driver with a registry type confusion vulnerability where missing validation flags can escalate from a local denial of service to kernel write access (the ability to modify system memory).

Fix: The article mentions that a new Claude skill called 'c-review' was developed to help find these bugs by turning the C/C++ security checklist into prompts that an LLM can run against a codebase. However, no explicit code fixes, patches, or specific mitigation steps for the vulnerabilities themselves are provided in the source text.

Trail of Bits Blog
03

We Scanned 1 Million Exposed AI Services. Here's How Bad the Security Actually Is

security
May 5, 2026

A scan of over 1 million exposed AI services found that self-hosted AI infrastructure has worse security than any other software previously investigated, with major problems including no authentication enabled by default, freely accessible chatbots that expose user conversations and can be abused to bypass safety guardrails (restrictions built into AI models to prevent harmful outputs), and exposed agent management platforms (tools like n8n and Flowise that automate AI workflows) that reveal business logic, API keys (secret credentials for accessing external services), and access to connected third-party systems. These misconfigurations leave real user data and company tools vulnerable to attackers, with consequences ranging from reputational damage to full system compromise.

The Hacker News
04

Google DeepMind workers are unionizing over AI military contracts

policy
May 5, 2026

Google DeepMind employees have voted to unionize, asking management to recognize their union representatives in an effort to prevent the company's AI technology from being used by the Israeli and US militaries. The unionization effort reflects employee concerns that their AI models may be complicit in international law violations, particularly regarding the Israeli-Palestinian conflict.

The Verge (AI)
05

GPT-5.5 Instant: smarter, clearer, and more personalized

industry
May 5, 2026

OpenAI has released GPT-5.5 Instant, an updated version of ChatGPT's default model that aims to provide smarter, more accurate answers with clearer language and better personalization based on your conversation history. The new model produces 52.5% fewer hallucinated claims (false or made-up statements) compared to the previous version on high-stakes topics like medicine and law, and includes a new 'memory sources' feature that shows you what past context was used to personalize your responses, giving you control to edit or delete outdated information.

Fix: The source mentions the following controls and mitigations for personalization concerns: Users can delete chats they no longer want cited, delete or change items in saved memories through settings, or use temporary chats that don't use or update memory. When a response is personalized, users can see what context was used in 'memory sources' and delete or correct outdated information. Memory sources are not shown to others if you share a chat. The source also notes that 'memory sources are designed to make personalization easier to understand' and OpenAI plans to make this feature 'more comprehensive over time.'

OpenAI Blog
06

GPT-5.5 Instant System Card

safety
May 5, 2026

GPT-5.5 Instant is OpenAI's latest fast-response AI model that uses safety methods similar to previous versions, but is the first Instant model classified as having high capability in cybersecurity and biological/chemical preparedness risks, so it has additional safeguards in place. The document clarifies naming conventions to avoid confusion: GPT-5.5 Instant (also called gpt-5.5-instant) should be compared to GPT-5.3 Instant, and the full GPT-5.5 model is referred to as GPT-5.5 Thinking.

OpenAI Blog
07

Unlocking large scale AI training networks with MRC (Multipath Reliable Connection)

industry
May 5, 2026

OpenAI and partners (AMD, Broadcom, Intel, Microsoft, NVIDIA) developed MRC (Multipath Reliable Connection), a new networking protocol that improves data transfer speed and reliability in supercomputer clusters used for AI model training. MRC addresses key challenges in large-scale AI training by reducing network congestion through adaptive packet spraying (distributing data across multiple paths), enabling redundancy to tolerate failures, and using static source routing (predetermined paths that bypass failed connections) to prevent training jobs from crashing when network failures occur.

Fix: MRC has been released through the Open Compute Project (OCP) as an open standard for the industry to use. The specification extends RDMA over Converged Ethernet (RoCE, a hardware-accelerated data transfer standard) and incorporates SRv6-based source routing to support large-scale AI networking fabrics.

OpenAI Blog
08

Google DeepMind workers in UK vote to unionize amid deal with US military

policy
May 5, 2026

Workers at Google DeepMind's UK laboratory voted to form a union, citing concerns about a recently announced deal between Google and the US military. The workers, represented by two unions, worry that the military partnership raises ethical questions about the company's responsibility in developing AI technology.

The Guardian Technology
09

CVE-2026-3456: The GeekyBot — Generate AI Content Without Prompt, Chatbot and Lead Generation plugin for WordPress is vulnerable to SQL

security
May 5, 2026

The GeekyBot WordPress plugin (up to version 1.2.0) has a SQL injection vulnerability (a type of attack where hackers insert malicious database commands into user input) in the 'attributekey' parameter. Because the plugin doesn't properly clean user input or secure its database queries, unauthenticated attackers can add extra SQL commands to extract sensitive data from the site's database.

NVD/CVE Database
10

datasette-llm 0.1a7

industry
May 4, 2026

Datasette-llm 0.1a7 is a plugin (a software add-on) that lets other plugins use AI models in a coordinated way. The release adds a feature to set default options for specific models, such as specifying which model to use for enrichment operations (adding data to existing information) and adjusting its temperature parameter (a setting that controls how creative or random the AI's responses are).

Simon Willison's Weblog
Prev1...1011121314...371Next
critical

CVE-2026-42203: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.80.5 to before vers

CVE-2026-42203NVD/CVE DatabaseMay 8, 2026
May 8, 2026
critical

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

SecurityWeekMay 7, 2026
May 7, 2026
critical

GHSA-9h64-2846-7x7f: Axonflow fixed bugs by implementing multi-tenant isolation and access-control hardening

GitHub Advisory DatabaseMay 6, 2026
May 6, 2026
critical

GHSA-gmvf-9v4p-v8jc: fast-jwt: JWT auth bypass due to empty HMAC secret accepted by async key resolver

CVE-2026-44351GitHub Advisory DatabaseMay 6, 2026
May 6, 2026