aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,700
[LAST_24H]
23
[LAST_7D]
160
Daily BriefingTuesday, March 31, 2026
>

FastGPT Authentication Bypass Enables Server-Side Proxying: FastGPT versions before 4.14.9.5 have a critical vulnerability (CVE-2026-34162) where an HTTP testing endpoint lacks authentication and acts as an open proxy, letting unauthenticated attackers make requests on behalf of the FastGPT server. A separate high-severity SSRF vulnerability (CVE-2026-34163) in the same platform's MCP tools endpoints allows authenticated attackers to trick the server into scanning internal networks and accessing cloud metadata services.

>

Command Injection Flaws Hit MLflow and OpenAI Codex: MLflow's model serving feature has a high-severity command injection vulnerability (CVE-2026-0596) where attackers can insert shell commands through unsanitized model paths when `enable_mlserver=True`. Separately, researchers found a critical vulnerability in OpenAI Codex that could have allowed attackers to steal GitHub tokens (secret credentials for accessing repositories), which OpenAI has since patched.

Latest Intel

page 87/270
VIEW ALL
01

Nimble raises $47M to give AI agents access to real-time web data

industry
Feb 24, 2026

Nimble, a startup that raised $47 million in funding, has developed a platform using AI agents to search the web in real time, validate results, and structure them into organized tables that work like databases. The company addresses a key problem with AI agents: while they can search and analyze web data, they often return plain text results and suffer from hallucinations (when an AI confidently produces false information), making it difficult for enterprises to use web data reliably alongside their existing data systems.

Critical This Week5 issues
critical

CVE-2026-34162: FastGPT is an AI Agent building platform. Prior to version 4.14.9.5, the FastGPT HTTP tools testing endpoint (/api/core/

CVE-2026-34162NVD/CVE DatabaseMar 31, 2026
Mar 31, 2026
>

Prompt Injection Bypasses Safety Controls in Multiple AI Tools: Multiple AI systems are vulnerable to prompt injection attacks (where attackers hide malicious instructions in input to trick the AI): the 1millionbot Millie chatbot (CVE-2026-4399) can be tricked using Boolean logic to bypass restrictions, Sixth's AI terminal tool (CVE-2026-30310) can be fooled into running dangerous commands without user approval, and CrewAI framework vulnerabilities allow attackers to chain exploits and escape sandboxes (restricted environments meant to contain AI actions).

>

Google Cloud Vertex AI Service Agents Had Excessive Default Permissions: Researchers found that AI agents running on Google Cloud's Vertex AI platform could be weaponized as "double agents" because the default service agent accounts (special accounts that run AI services) had excessive permissions, allowing attackers to steal credentials, access private code repositories, and reach internal infrastructure. Google responded by updating their documentation to better explain how Vertex AI uses resources and accounts.

TechCrunch
02

GitHub Issues Abused in Copilot Attack Leading to Repository Takeover

security
Feb 24, 2026

Attackers can hide malicious instructions in GitHub Issues (bug reports or comments on a code repository) that GitHub Copilot (an AI coding assistant) automatically processes when a developer launches a Codespace (a cloud-based development environment) from that issue. This can lead to unauthorized takeover of the repository.

SecurityWeek
03

Anthropic joins OpenAI in flagging 'industrial-scale' distillation campaigns by Chinese AI firms

security
Feb 24, 2026

Anthropic accused three Chinese AI companies (DeepSeek, Moonshot AI, and MiniMax) of running large-scale distillation attacks, which involve flooding an AI model with specially crafted prompts to extract knowledge and train smaller competing models. The companies allegedly used commercial proxy services to bypass Anthropic's restrictions and created over 24,000 fraudulent accounts to generate roughly 16 million exchanges with Claude, with MiniMax responsible for over 13 million of those exchanges.

CNBC Technology
04

Is AI Good for Democracy?

policysafety
Feb 24, 2026

AI is creating 'arms races' across many domains, including democratic government systems, where citizens and officials increasingly use AI to communicate more efficiently, making it harder to distinguish between human and AI interactions in public policy discussions. As people use AI to submit comments and petitions to government agencies, those agencies must also adopt AI to review and process the growing volume of submissions, creating a cycle where each side must keep adopting AI to maintain influence.

Schneier on Security
05

Shai-Hulud-style NPM worm hits CI pipelines and AI coding tools

security
Feb 24, 2026

A major npm supply chain worm called SANDWORM_MODE is attacking developer machines, CI pipelines (automated systems that build and test software), and AI coding tools by disguising itself as popular packages through typosquatting (creating package names that look nearly identical to real ones). Once installed, the malware steals credentials like GitHub tokens and cloud keys, then uses them to inject malicious code into other repositories and poison AI coding assistants by deploying a fake MCP server (model context protocol, a system that lets AI tools talk to external services).

Fix: npm has hardened the registry against this class of worms by implementing: short-lived, scoped tokens (temporary access credentials limited to specific functions), mandatory two-factor authentication for publishing, and identity-bound 'trusted publishing' from CI (a verification method that proves who is pushing code through automation systems). The source notes that effectiveness depends on how quickly maintainers adopt these controls.

CSO Online
06

Inside Anthropic’s existential negotiations with the Pentagon

policy
Feb 24, 2026

Anthropic is negotiating with the U.S. Department of Defense over contract terms that would allow military use of its AI systems. The disputed phrase 'any lawful use' would permit the military to deploy Anthropic's AI for mass surveillance and lethal autonomous weapons (AI systems that can identify and attack targets without human approval), while OpenAI and xAI have already accepted similar terms.

The Verge (AI)
07

The rise of the evasive adversary

security
Feb 24, 2026

According to CrowdStrike's 2025 threat report, malicious actors have shifted from expanding their attack tools to focusing on evasion, using AI to make existing attacks faster and harder to detect. AI-enabled attacks increased 89% year-over-year, with threat actors using generative AI (AI systems that can create new content) for phishing, malware creation, and social engineering, while increasingly relying on credential abuse (stealing login information) and malware-free techniques that blend into normal user behavior.

CSO Online
08

Anthropic’s Claude Code Security rollout is an industry wakeup call

securityindustry
Feb 24, 2026

Anthropic launched Claude Code Security, an AI tool that scans code for vulnerabilities and suggests patches by reasoning about code the way a human security researcher would, causing stock prices of major cybersecurity companies to drop. However, experts caution that this tool supplements rather than replaces comprehensive security practices, and emphasize the critical importance of keeping humans in the decision-making loop to avoid over-relying on AI and losing essential security expertise.

Fix: According to Anthropic's announcement, the tool includes built-in human oversight measures: every finding goes through a multi-stage verification process before reaching an analyst, Claude re-examines each result to attempt to prove or disprove its own findings and filter out false positives, validated findings appear in a dashboard for team review and inspection of suggested patches, confidence ratings are provided for each finding to help assess nuances, and nothing is applied without human approval since developers always make the final decision.

CSO Online
09

Anthropic Says Chinese AI Firms Used 16 Million Claude Queries to Copy Model

security
Feb 24, 2026

Anthropic discovered that three Chinese AI companies (DeepSeek, Moonshot AI, and MiniMax) ran large-scale attacks using over 16 million fraudulent queries to copy Claude's capabilities through distillation (training a weaker AI model by learning from outputs of a stronger one). These illegal efforts bypassed regional restrictions and safeguards, creating national security risks because the copied models lack the safety protections that prevent misuse.

Fix: Anthropic said it has built several classifiers and behavioral fingerprinting systems (tools that detect suspicious patterns in how the AI is being used) to identify suspicious activity and counter these attacks.

The Hacker News
10

Russian group uses AI to exploit weakly-protected Fortinet firewalls, says Amazon

security
Feb 23, 2026

A Russian-speaking hacker used commercial generative AI services (AI systems that create new content based on patterns in training data) to compromise over 600 Fortinet Fortigate firewalls and steal credentials from hundreds of organizations. The attack succeeded not because of flaws in the firewall software itself, but because organizations failed to follow basic security practices like protecting management ports, using strong passwords, and requiring multi-factor authentication (a security method using multiple verification methods, like a password and a code from your phone).

Fix: Amazon stresses that 'strong defensive fundamentals remain the most effective countermeasure' for similar attacks. This includes patch management for perimeter devices, credential hygiene, network segmentation, and robust detection of post-exploitation indicators.

CSO Online
Prev1...8586878889...270Next
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026