aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
71
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 5/371
VIEW ALL
01

Attackers Could Exploit AI Vision Models Using Imperceptible Image Changes

securityresearch
Critical This Week5 issues
critical

CVE-2026-42271: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.74.2 to before vers

CVE-2026-42271NVD/CVE DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

May 7, 2026

Researchers at Cisco discovered that attackers can manipulate vision-language models (AI systems that read and interpret images) by making tiny, imperceptible changes to image pixels that humans cannot see. These changes can make hidden malicious instructions embedded in images readable to the AI, allowing attackers to trick the AI into following commands like stealing data, while content filters and humans see only visual noise or blurry content.

SecurityWeek
02

'TrustFall' Convention Exposes Claude Code Execution Risk

securitysafety
May 7, 2026

A security issue called 'TrustFall' allows malicious code repositories to execute code in Claude Code, Cursor CLI (a code editor tool), Gemini CLI, and CoPilot CLI (command-line interfaces for AI coding tools) with little or no user action needed, because the warning messages shown to users are minimal and easy to ignore. This means an attacker could potentially run harmful code on a developer's computer without much effort.

Dark Reading
03

Scaling Trusted Access for Cyber with GPT-5.5 and GPT-5.5-Cyber

securitypolicy
May 7, 2026

OpenAI released GPT-5.5 and a specialized version called GPT-5.5-Cyber with Trusted Access for Cyber (TAC), a framework that verifies the identity of cybersecurity defenders and gives approved users lower refusal rates so they can perform defensive security tasks like vulnerability analysis and malware detection. The system maintains safeguards to block malicious activities like credential theft and system exploitation, and requires users to have phishing-resistant authentication (protection against attacks where hackers trick users into revealing passwords) by June 2026.

Fix: The source explicitly mentions one safeguard: "Individual members of Trusted Access for Cyber accessing our most cyber capable and permissive models will be required to enable Advanced Account Security beginning June 1, 2026. Organizations with trusted access can, as an alternative, attest that they have phishing resistant authentication as part of their single sign-on workflow." No other mitigation or fix beyond this account security requirement is discussed in the source.

OpenAI Blog
04

AMD's big day, Anthropic-SpaceX deal, the jet fuel crisis and more in Morning Squawk

industry
May 7, 2026

Anthropic, an AI startup, announced a deal to use all the computing power from SpaceX's Colossus 1 data center in Tennessee to improve service for its paid Claude Pro and Claude Max subscribers. The deal will give Anthropic access to significant computational resources (the processing power needed to run AI models) to better handle demand from paying customers.

CNBC Technology
05

Bots in translation: Can AI really fix SIEM rule sprawl across vendors?

researchindustry
May 7, 2026

Enterprises migrating between different SIEM platforms (security information and event management systems, which collect and analyze security data) struggle because each vendor uses different query languages and data models, requiring manual rule rewrites. Researchers developed ARuleCon, an AI system that can automatically translate detection rules across platforms while preserving their detection logic, improving accuracy by 10-15% over standard AI approaches. However, security experts debate whether the problem truly needs AI, since manual translation is slow but some argue deterministic engineering (rule-based programming without AI) could solve it.

Fix: ARuleCon combines AI-driven reasoning with deterministic approaches by using AI to infer detection intent and iteratively refine translated rules while constraining outputs through syntax validation and semantic checks. According to the researchers, the system is not intended to replace deterministic approaches entirely, but to combine "their reliability with the flexibility of AI-driven reasoning."

CSO Online
06

Privacy Against Agnostic Inference Attacks in Vertical Federated Learning

securityprivacy
May 7, 2026

This academic paper examines privacy risks in vertical federated learning (a machine learning approach where different organizations each hold different features of the same data and train a model together) when facing agnostic inference attacks (attacks where the attacker doesn't know the model's structure in advance). The paper analyzes how attackers could potentially infer private information from the shared computations in this system.

ACM Digital Library (TOPS, DTRAP, CSUR)
07

Parloa builds service agents customers want to talk to

industry
May 7, 2026

Parloa has built an AI Agent Management Platform (AMP) that helps businesses create and manage customer service AI agents without coding, using large language models (LLMs, AI systems trained on huge amounts of text data) like GPT-5.4. The platform lets non-technical teams define agent behavior in plain language, then tests agents through simulations (one AI model acting as a customer, another as the agent) before deploying them to handle real customer interactions. Parloa continuously monitors live conversations and updates the platform with newer model versions when they perform better in real-world use.

OpenAI Blog
08

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

security
May 7, 2026

Gemini CLI (Google's open source AI agent for terminal access to the Gemini AI assistant) had a critical vulnerability with a CVSS score of 10/10 that could have allowed attackers to inject malicious prompts into GitHub issues, causing the AI agent to execute unauthorized commands and steal secrets from the build environment in a supply chain attack (compromising software distributed to many users). The vulnerability existed because the --yolo mode (which auto-approves all tool calls without user confirmation) ignored tool allowlists (restrictions on what actions the AI could perform), and Google fixed it in version 0.39.1 by properly enforcing those restrictions.

Fix: Google addressed the vulnerability on April 24 in Gemini CLI version 0.39.1, which evaluates tool allowlisting under --yolo mode. The run-gemini-cli GitHub Action was also updated. The same version resolved a separate trust issue in headless mode (where the AI runs without user interaction) that was automatically loading configuration and environment variables from the current workspace folder.

SecurityWeek
09

Fake Claude AI website delivers new 'Beagle' Windows malware

security
May 7, 2026

Attackers created a fake Claude AI website that tricks users into downloading malware called Beagle, a backdoor (a hidden entrance to a system that lets attackers run commands remotely) disguised as a legitimate Claude-Pro Relay tool. The malware uses a chain of loaders to hide itself in system memory and communicates with attackers' servers, while impersonating updates from various security companies to spread further.

Fix: Users should ensure they download Claude from the official portal and skip or hide sponsored search results. The presence of 'NOVupdate' files on a system is a strong indication of compromise.

BleepingComputer
10

Advancing voice intelligence with new models in the API

industry
May 7, 2026

OpenAI has released three new audio models for developers: GPT-Realtime-2 (a voice model with advanced reasoning capabilities), GPT-Realtime-Translate (live translation across 70+ languages), and GPT-Realtime-Whisper (streaming speech-to-text). These models enable voice applications that can understand context, reason through requests, use tools, and take action during conversations, moving beyond simple back-and-forth responses to support real-world tasks like booking travel or providing customer support.

OpenAI Blog
Prev1...34567...371Next
critical

CVE-2026-42203: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.80.5 to before vers

CVE-2026-42203NVD/CVE DatabaseMay 8, 2026
May 8, 2026
critical

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

SecurityWeekMay 7, 2026
May 7, 2026
critical

GHSA-9h64-2846-7x7f: Axonflow fixed bugs by implementing multi-tenant isolation and access-control hardening

GitHub Advisory DatabaseMay 6, 2026
May 6, 2026
critical

GHSA-gmvf-9v4p-v8jc: fast-jwt: JWT auth bypass due to empty HMAC secret accepted by async key resolver

CVE-2026-44351GitHub Advisory DatabaseMay 6, 2026
May 6, 2026