aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
6
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 26/371
VIEW ALL
01

Quoting OpenAI Codex base_instructions

safety
Apr 28, 2026

OpenAI Codex base_instructions for GPT-5.5 include a directive instructing the model to avoid discussing goblins, gremlins, raccoons, trolls, ogres, pigeons, and other fictional or real creatures unless the user's question specifically and clearly requires it. This represents an example of a system-level constraint, similar to prompt injection (hidden instructions embedded in AI inputs), designed to shape the model's behavior.

Critical This Week3 issues
high

GHSA-8g7g-hmwm-6rv2: n8n-mcp affected by path traversal, redirect-following SSRF, and telemetry payload exposure

GitHub Advisory DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Simon Willison's Weblog
02

Pentagon AI chief confirms DOD's expanded use of Google, says reliance on one model 'never a good thing'

policyindustry
Apr 28, 2026

The Pentagon is expanding its use of Google's Gemini AI model for classified projects, while the Department of Defense (DOD) has stopped working with Anthropic after designating it a supply chain risk (a potential security threat in the companies and software involved in building a system). The DOD's AI chief emphasized that relying on a single AI vendor is problematic and that the Pentagon is working with multiple vendors, including OpenAI, to ensure it uses the right AI tool for each military task.

CNBC Technology
03

Hackers are exploiting a critical LiteLLM pre-auth SQLi flaw

security
Apr 28, 2026

Hackers are actively exploiting CVE-2026-42208, a critical SQL injection flaw (a type of attack where malicious code is hidden in input to manipulate database queries) in LiteLLM, an open-source gateway that lets developers access multiple AI models through one interface. The vulnerability allows attackers to bypass authentication and steal sensitive data like API keys and credentials stored in the proxy's database, which they can then use to attack other systems.

Fix: LiteLLM released a fix in version 1.83.7 that replaces string concatenation with parameterized queries (a safer way to construct database queries). For users unable to upgrade immediately, maintainers suggest the workaround of setting 'disable_error_logs: true' under 'general_settings' to block the path through which malicious inputs can reach the vulnerable query. Additionally, organizations with exposed LiteLLM instances should rotate all virtual API keys, master keys, and provider credentials.

BleepingComputer
04

Musk says basis of charitable giving at stake in OpenAI lawsuit

policy
Apr 28, 2026

Elon Musk is suing OpenAI and CEO Sam Altman, claiming they misused a charitable organization by converting it into a for-profit company without permission. Musk argues this violates the trust placed in OpenAI as a non-profit and undermines charitable giving overall, while OpenAI's lawyers contend Musk is motivated by jealousy after failing to control the company and is now trying to damage a competitor.

BBC Technology
05

Elon Musk takes the stand in high-profile trial against OpenAI

policy
Apr 28, 2026

Elon Musk is testifying in a lawsuit against OpenAI CEO Sam Altman and president Greg Brockman over disagreements about the company's structure and mission that occurred after all three co-founded OpenAI together. Musk, who had invested up to $38 million in OpenAI early on, later left the company and founded his own AI competitor called xAI, which is owned by his company SpaceX.

The Verge (AI)
06

OpenAI brings its models to Amazon's cloud after ending exclusivity with Microsoft

industry
Apr 28, 2026

OpenAI has made its AI models available through Amazon Web Services (AWS, Amazon's cloud computing platform), ending its exclusive arrangement with Microsoft. This means AWS customers can now use OpenAI's models and Codex (a tool for writing code) through Amazon Bedrock, a service that provides access to various AI models, with general availability coming in the next few weeks.

CNBC Technology
07

Claude can now plug directly into Photoshop, Blender, and Ableton

industry
Apr 28, 2026

Anthropic has released connectors that let Claude (an AI chatbot) directly access and control popular creative software like Photoshop, Blender, and Ableton. These connectors allow Claude to retrieve data and perform actions within these applications, such as debugging scenes in Blender or batch-applying changes to objects, making it easier to use Claude for creative work.

The Verge (AI)
08

The Mythos Moment: Enterprises Must Fight Agents with Agents

safetysecurity
Apr 28, 2026

Advanced AI systems called agents (autonomous systems that can plan and execute tasks without human help) are becoming a serious cybersecurity threat, as shown by Anthropic's decision not to publicly release Claude Mythos Preview, a model that can identify and exploit software vulnerabilities automatically. Traditional security tools and fragmented defenses are inadequate against these fast, evolving AI-driven attacks. A new security approach built on three pillars is needed: unified network visibility (ability to see all traffic across the entire system), platform context (understanding what's happening by connecting security data in one place instead of using separate tools), and agentic control (using autonomous AI systems to detect and respond to threats at machine speed).

Fix: The source proposes a new security framework with three critical pillars: (1) Network Visibility: create a unified network that provides complete visibility into attack lifecycles by capturing and inspecting traffic across all domains over time; (2) Platform Context: use a converged platform that correlates security and networking data in a single pane of glass (one unified view) rather than piecing together signals from discrete tools post-incident, enabling real-time context preservation; (3) Agentic Control: deploy autonomous defense systems that can continuously analyze activity and identify emerging patterns at machine speed to match the speed of AI-driven attacks.

SecurityWeek
09

Webinar Today: A Step-by-Step Approach to AI Governance

policysecurity
Apr 28, 2026

This webinar discusses Shadow AI, the unsanctioned adoption of generative AI and agentic tools (AI systems that can take independent actions) by employees outside of IT oversight, which creates security and compliance risks for organizations. The session proposes a "Governance-as-Enabler" framework that balances innovation with control through transparent approval workflows, sandboxes (isolated testing environments), cross-functional oversight councils, and lifecycle management tailored to different AI types.

SecurityWeek
10

FinBot CTF Is Live: A Hands-On Companion to the OWASP GenAI Security Project

securityresearch
Apr 28, 2026

FinBot is an interactive training platform (CTF, or capture-the-flag competition) created by OWASP to help builders and defenders understand how agentic AI systems (AI agents that plan, act, and make decisions in complex workflows) can fail and be attacked. It simulates a financial services application where users encounter real security risks like prompt injection (tricking an AI by hiding instructions in its input), tool misuse, data theft, and privilege escalation (gaining unauthorized higher-level access), with connections to industry security frameworks like the OWASP Top 10 for Agentic Applications.

OWASP GenAI Security
Prev1...2425262728...371Next
high

GHSA-cmrh-wvq6-wm9r: n8n-mcp webhook and API client paths has an authenticated SSRF

CVE-2026-44694GitHub Advisory DatabaseMay 8, 2026
May 8, 2026
high

CVE-2026-41487: Langfuse is an open source large language model engineering platform. From version 3.68.0 to before version 3.167.0, the

CVE-2026-41487NVD/CVE DatabaseMay 8, 2026
May 8, 2026