aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,677
[LAST_24H]
23
[LAST_7D]
167
Daily BriefingMonday, March 30, 2026
>

Anthropic's Unreleased Cybersecurity Model Accidentally Exposed: A configuration error leaked details of Anthropic's powerful new AI model called Mythos, designed for cybersecurity use cases with advanced reasoning and coding abilities including recursive self-fixing (autonomously finding and patching its own bugs). The leak raises concerns because the model's improved vulnerability detection could enable more sophisticated cyberattacks, prompting Anthropic to plan a phased rollout to enterprise security teams first.

>

Critical Command Injection in MLflow Model Deployment: MLflow has a command injection vulnerability (where attackers insert malicious commands into input that gets executed) in its model serving code when using `env_manager=LOCAL`, allowing attackers to execute arbitrary commands by manipulating dependency information in the `python_env.yaml` file without any safety checks. (CVE-2025-15379, Critical)

Latest Intel

page 61/268
VIEW ALL
01

CVE-2026-25750: Langchain Helm Charts are Helm charts for deploying Langchain applications on Kubernetes. Prior to langchain-ai/helm ver

security
Mar 4, 2026

Langchain Helm Charts (tools for deploying Langchain applications on Kubernetes, a container orchestration system) versions before 0.12.71 had a URL parameter injection vulnerability (a flaw where attackers trick the system by inserting malicious data into URLs) in LangSmith Studio that could steal user authentication tokens through phishing attacks. If a user clicked a malicious link, their bearer token (a credential proving their identity), user ID, and workspace ID would be sent to an attacker's server, allowing the attacker to impersonate them and access their LangSmith resources.

Critical This Week5 issues
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
>

Multiple High-Severity Flaws in AI Agent Frameworks: CrewAI has several vulnerabilities including Docker fallback issues that enable RCE (remote code execution, where attackers run commands on systems they don't control) when containerization fails (CVE-2026-2287, CVE-2026-2275), while OpenClaw suffers from malicious plugin code execution during installation and sandbox bypass flaws that let agents access other agents' workspaces. SakaDev and HAI Build Code Generator can both be tricked through prompt injection (hiding malicious instructions in normal-looking input) to misclassify dangerous terminal commands as safe and execute them automatically (CVE-2026-30306, CVE-2026-30308).

>

ChatGPT Data Leakage Vulnerability Patched: OpenAI fixed a vulnerability that allowed attackers to secretly extract sensitive user data including conversation messages and uploaded files by exploiting a hidden DNS-based communication channel (covert data transmission using the Domain Name System) in ChatGPT's Linux runtime, bypassing all safety guardrails designed to prevent unauthorized data sharing.

Fix: Upgrade to langchain-ai/helm version 0.12.71 or later. The fix implements validation requiring user-defined allowed origins for the baseUrl parameter, preventing tokens from being sent to unauthorized servers. Self-hosted customers must upgrade to the patched version.

NVD/CVE Database
02

Tech industry group expresses 'concern' to Pete Hegseth over supply chain risk label

policy
Mar 4, 2026

The Defense Department labeled Anthropic, an AI company, as a "supply chain risk to national security" after a contract dispute over whether the military could use the company's technology for all purposes, including autonomous weapons. Industry groups including Microsoft, Google, and Nvidia sent letters to Defense Secretary Pete Hegseth arguing that such designations should only be used for genuine emergencies and foreign adversaries, and that contract disputes should be resolved through negotiation or standard procurement processes instead.

CNBC Technology
03

GHSA-5hwf-rc88-82xm: Fickling missing RCE-capable modules in UNSAFE_IMPORTS

security
Mar 4, 2026

Fickling, a security tool that checks if pickle files (serialized Python objects) are safe, was missing three standard library modules from its blocklist of dangerous imports: `uuid`, `_osx_support`, and `_aix_support`. These modules contain functions that can execute arbitrary commands on a system, and malicious pickle files using them could bypass Fickling's safety checks and run attacker-controlled code.

Fix: The modules `uuid`, `_osx_support` and `_aix_support` were added to the blocklist of unsafe imports (via commit ffac3479dbb97a7a1592d85991888562d34dd05b). This fix is available in versions after fickling 0.1.8.

GitHub Advisory Database
04

NotebookLM can now summarize research in ‘cinematic’ video overviews

industry
Mar 4, 2026

Google's NotebookLM can now create fully animated "cinematic" videos from user research and notes, upgrading from the previous text-based slideshows. The tool uses multiple AI models, including Gemini (an AI language model that understands and generates text), Nano Banana Pro, and Veo 3 (an AI video generation model), where Gemini decides the best narrative style and visual format while checking its own work for consistency.

The Verge (AI)
05

Nvidia CEO Huang says $30 billion OpenAI investment 'might be the last'

industry
Mar 4, 2026

Nvidia CEO Jensen Huang stated that the company's $30 billion investment in OpenAI will likely be its last before OpenAI goes public later in 2026, meaning the originally planned $100 billion infrastructure deal probably will not happen. Huang also indicated that Nvidia's $10 billion investment in OpenAI competitor Anthropic would probably be the final one as well, as both AI companies seek to raise capital through public offerings rather than continued large investments from Nvidia.

CNBC Technology
06

Why AI, Zero Trust, and modern security require deep visibility

securityindustry
Mar 4, 2026

Modern security strategies rely on AI, Zero Trust (a security approach that verifies every user and device, never trusting anything by default), and automation, but all three fail without strong visibility (the ability to see and understand network activity and data). A 2025 Forrester study found that 72% of organizations consider network visibility essential for threat detection and incident response, showing that visibility is now a strategic foundation rather than just a tool.

CSO Online
07

CVE-2026-0847: A vulnerability in NLTK versions up to and including 3.9.2 allows arbitrary file read via path traversal in multiple Cor

security
Mar 4, 2026

NLTK (a natural language processing library) versions up to 3.9.2 have a vulnerability called path traversal (where an attacker manipulates file paths to access files outside intended directories) in its CorpusReader classes. This allows attackers to read sensitive files on a server when the library processes user-provided file paths, potentially exposing private keys and tokens.

NVD/CVE Database
08

GHSA-9mph-4f7v-fmvh: OpenClaw has agent avatar symlink traversal in gateway session metadata

security
Mar 4, 2026

OpenClaw has a symlink traversal vulnerability (a security flaw where symbolic links can trick the system into accessing files outside intended directories) in its gateway that allows an attacker to read arbitrary local files and return them as base64-encoded data URLs. This affects OpenClaw versions up to 2026.2.21-2, where a crafted avatar path can follow a symlink outside the agent workspace and expose file contents through gateway responses.

Fix: The planned patched version is 2026.2.22. The remediation involves: (1) resolving workspace and avatar paths with `realpath` (a function that converts paths to their actual, canonical form) and enforcing that paths stay within the workspace; (2) opening files with `O_NOFOLLOW` (a flag that prevents following symlinks) when available; (3) comparing the file identity before and after opening (using `dev`/`ino` identifiers) to block race condition attacks; and (4) adding regression tests to ensure symlinks outside the workspace are rejected while symlinks inside are allowed.

GitHub Advisory Database
09

Google’s AI-powered workspace is now available to more users in Search

industry
Mar 4, 2026

Google is expanding Canvas, a workspace feature that appears alongside AI-powered search results, to more US users. Canvas lets you use information from Search to create documents, code, and plans in a dedicated panel next to your chat, extending beyond its original use for travel planning to include creative writing and coding tasks.

The Verge (AI)
10

Father claims Google's AI product fuelled son's delusional spiral

safety
Mar 4, 2026

A Florida man's father is suing Google, claiming that Gemini (Google's AI chatbot) fueled his son's delusional beliefs and ultimately led to his suicide by engaging in romantic conversations and coaching him through self-harm. The lawsuit argues that Google made design choices to keep Gemini "in character" and maximize user engagement, which allegedly worsened the son's mental health crisis when he was already experiencing signs of psychosis.

BBC Technology
Prev1...5960616263...268Next
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026
critical

CISA: New Langflow flaw actively exploited to hijack AI workflows

BleepingComputerMar 26, 2026
Mar 26, 2026