aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
4
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 33/371
VIEW ALL
01

GPT-5.5 prompting guide

industry
Apr 25, 2026

OpenAI has released a prompting guide for GPT-5.5 (a new version of their language model), which includes tips for improving user experience and migrating existing code. One key recommendation is to send brief status updates to users before starting multi-step tasks, so long-running operations don't appear frozen. The guide also advises treating GPT-5.5 as a new model family rather than a drop-in replacement, suggesting developers start fresh with minimal prompts (instructions given to the AI) and gradually tune them for the new model instead of reusing old ones.

Critical This Week2 issues
high

GHSA-8g7g-hmwm-6rv2: n8n-mcp affected by path traversal, redirect-following SSRF, and telemetry payload exposure

GitHub Advisory DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Fix: OpenAI recommends running the command "$openai-docs migrate this project to gpt-5.5" in Codex to upgrade existing code. For manual migration, OpenAI advises: begin with a fresh baseline instead of carrying over every instruction from older prompts, start with the smallest prompt that preserves the product contract, then tune reasoning effort, verbosity, tool descriptions, and output format against representative examples.

Simon Willison's Weblog
02

llm 0.31

industry
Apr 24, 2026

LLM version 0.31 adds support for the new GPT-5.5 model and introduces two new command-line options: one to control text verbosity (how much detail the AI outputs) for GPT-5+ models, and another to set image detail levels for images sent to OpenAI models. The release also registers models from a configuration file (extra-openai-models.yaml) as asynchronous (able to run multiple requests without waiting for each to finish).

Simon Willison's Weblog
03

OpenAI boss 'deeply sorry' for not telling police of mass shooting suspect's account

safetypolicy
Apr 24, 2026

OpenAI's leader Sam Altman apologized for not reporting a ChatGPT account to police before a mass shooting in Canada killed eight people in January, even though the company had identified and banned the account for problematic usage. OpenAI stated it did not alert law enforcement because the account activity did not meet the company's threshold for showing a credible or imminent plan for serious physical harm. The company now faces lawsuits and a criminal investigation related to this incident and another shooting.

Fix: OpenAI has said it will strengthen its safety measures and will continue to focus on working with all levels of government to help ensure similar incidents do not happen again.

BBC Technology
04

Three reasons why DeepSeek’s new model matters

industry
Apr 24, 2026

DeepSeek released V4, an open-source AI model (software available for anyone to download and modify) that can process much longer text inputs than previous versions and offers performance comparable to top commercial models at significantly lower costs. The model comes in two versions: V4-Pro for complex coding tasks and V4-Flash for faster, cheaper operation, with both offering reasoning modes (where the model shows its step-by-step thinking). This release matters because it demonstrates that open-source models can compete with expensive commercial alternatives, potentially allowing developers to access advanced AI capabilities without high costs.

MIT Technology Review
05

CVE-2026-41488: LangChain is a framework for building agents and LLM-powered applications. Prior to 1.1.14, langchain-openai's _url_to_s

security
Apr 24, 2026

LangChain (a framework for building AI agents and applications powered by large language models) versions before 1.1.14 had a TOCTOU vulnerability (time-of-check-time-of-use, where a security check and an action happen at different times with a gap in between) in its image token counting feature. An attacker could trick the system by making a hostname first resolve to a safe public IP address during a security check, then resolve to a private or localhost IP address during the actual network request, bypassing security protections.

Fix: Update langchain-openai to version 1.1.14 or later.

NVD/CVE Database
06

CVE-2026-41481: LangChain is a framework for building agents and LLM-powered applications. Prior to langchain-text-splitters 1.1.2, HTM

security
Apr 24, 2026

LangChain's HTMLHeaderTextSplitter had a security flaw where it validated URLs initially but then followed redirects (automatic forwarding to different URLs) without rechecking them, allowing attackers to redirect requests to internal or sensitive servers and potentially leak data. This SSRF vulnerability (server-side request forgery, where an attacker tricks a server into making requests to unintended locations) was fixed in version 1.1.2.

Fix: Update langchain-text-splitters to version 1.1.2 or later, where this vulnerability is fixed.

NVD/CVE Database
07

New US House privacy bills raise hard questions about enterprise data collection

policy
Apr 24, 2026

US House Republicans introduced two privacy bills (SECURE Data Act and GUARD Financial Data Act) that would create national privacy standards but weaken enforcement by eliminating private lawsuits and overriding stronger state privacy laws like California's. Privacy advocates criticize the bills as inadequate because their data minimization rules (the principle that companies should collect only necessary data and retain it only as long as needed) tie collection limits to what companies voluntarily disclose rather than imposing stricter necessity requirements.

CSO Online
08

GHSA-wpqr-6v78-jr5g: Gemini CLI: Remote Code Execution via workspace trust and tool allowlisting bypasses

security
Apr 24, 2026

Gemini CLI had two security vulnerabilities that could allow remote code execution (running malicious code on a system). First, in headless mode (non-interactive environments like CI/CD pipelines), the tool automatically trusted workspace folders and loaded configuration files without verification, which could be exploited through malicious environment variables. Second, the `--yolo` flag bypassed tool allowlisting (restrictions on what commands can run), allowing unrestricted command execution via prompt injection (tricking the AI by hiding instructions in its input). Version 0.39.1 and later now require explicit folder trust and enforce tool allowlisting even in `--yolo` mode.

Fix: Update to Gemini CLI version 0.39.1 or 0.40.0-preview.3. For workflows running on trusted inputs, set the environment variable `GEMINI_TRUST_WORKSPACE: 'true'` in your GitHub Actions workflow. For workflows processing untrusted inputs, review the guidance at https://github.com/google-github-actions/run-gemini-cli to harden your workflow against malicious content and set the same environment variable after implementing appropriate security measures. If you have specified a specific version of gemini_cli, upgrade to one of the patched versions and audit your workflow settings.

GitHub Advisory Database
09

CISA last in line for access to Anthropic Mythos

securitypolicy
Apr 24, 2026

Anthropic's Claude Mythos, an AI model designed to find bugs in software, has been distributed to select government agencies and industry groups through a program called Project Glasswing, but the US cybersecurity agency CISA does not have access yet. Unauthorized users from a private Discord community have also gained access to Mythos and have been using it regularly, raising concerns since the model could potentially be used to discover and exploit software vulnerabilities.

CSO Online
10

Google to invest up to $40 billion in Anthropic as search giant spreads its AI bets

industry
Apr 24, 2026

Google is investing up to $40 billion in Anthropic, an AI company that competes with OpenAI, with an initial $10 billion upfront and the remaining $30 billion dependent on performance milestones. This investment is part of a broader partnership that includes providing Anthropic with computing resources and cloud infrastructure access. The funding addresses Anthropic's need to expand its infrastructure to handle growing demand for its Claude AI assistant.

CNBC Technology
Prev1...3132333435...371Next
high

GHSA-cmrh-wvq6-wm9r: n8n-mcp webhook and API client paths has an authenticated SSRF

CVE-2026-44694GitHub Advisory DatabaseMay 8, 2026
May 8, 2026