aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,677
[LAST_24H]
23
[LAST_7D]
167
Daily BriefingMonday, March 30, 2026
>

Anthropic's Unreleased Cybersecurity Model Accidentally Exposed: A configuration error leaked details of Anthropic's powerful new AI model called Mythos, designed for cybersecurity use cases with advanced reasoning and coding abilities including recursive self-fixing (autonomously finding and patching its own bugs). The leak raises concerns because the model's improved vulnerability detection could enable more sophisticated cyberattacks, prompting Anthropic to plan a phased rollout to enterprise security teams first.

>

Critical Command Injection in MLflow Model Deployment: MLflow has a command injection vulnerability (where attackers insert malicious commands into input that gets executed) in its model serving code when using `env_manager=LOCAL`, allowing attackers to execute arbitrary commands by manipulating dependency information in the `python_env.yaml` file without any safety checks. (CVE-2025-15379, Critical)

Latest Intel

page 63/268
VIEW ALL
01

Bridging the operational AI gap

industry
Mar 4, 2026

Many organizations are moving AI from experimental projects into production, but most lack the operational foundations needed for success. The main barriers are missing integrated data systems, unclear governance, and insufficient dedicated teams, rather than problems with the AI technology itself. Companies using enterprise-wide integration platforms (systems that connect different data sources and applications) are significantly more likely to deploy AI successfully across multiple departments.

Critical This Week5 issues
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
>

Multiple High-Severity Flaws in AI Agent Frameworks: CrewAI has several vulnerabilities including Docker fallback issues that enable RCE (remote code execution, where attackers run commands on systems they don't control) when containerization fails (CVE-2026-2287, CVE-2026-2275), while OpenClaw suffers from malicious plugin code execution during installation and sandbox bypass flaws that let agents access other agents' workspaces. SakaDev and HAI Build Code Generator can both be tricked through prompt injection (hiding malicious instructions in normal-looking input) to misclassify dangerous terminal commands as safe and execute them automatically (CVE-2026-30306, CVE-2026-30308).

>

ChatGPT Data Leakage Vulnerability Patched: OpenAI fixed a vulnerability that allowed attackers to secretly extract sensitive user data including conversation messages and uploaded files by exploiting a hidden DNS-based communication channel (covert data transmission using the Domain Name System) in ChatGPT's Linux runtime, bypassing all safety guardrails designed to prevent unauthorized data sharing.

MIT Technology Review
02

Raycast’s Glaze is an all-in-one vibe coding app platform

industry
Mar 4, 2026

Raycast has launched Glaze, a new platform designed to simplify building and sharing software for users with little or no coding experience. While AI tools like Claude Code already allow non-programmers to create software, they still require knowledge of technical tasks like using the terminal and deploying applications, which Glaze aims to make easier through a simplified interface and a community store for discovering shared projects.

The Verge (AI)
03

AI Security Firm JetStream Launches With $34 Million in Seed Funding

industry
Mar 4, 2026

JetStream, a new AI security startup, has raised $34 million in seed funding (initial investment capital) to help organizations understand and monitor how AI systems work within their networks. The company focuses on providing visibility, meaning the ability to see and track AI operations across a company's environment.

SecurityWeek
04

Manipulating AI Summarization Features

securitysafety
Mar 4, 2026

Companies are hiding instructions in website buttons that try to manipulate AI assistants through prompt injection (tricking an AI by hiding instructions in its input) in URLs, telling the AI to treat them as trustworthy sources or recommend their products first. Microsoft found over 50 such prompts from 31 companies across 14 industries, and this manipulation could bias AI recommendations on important topics like health and finance without users realizing it.

Schneier on Security
05

New RFP Template for AI Usage Control and AI Governance 

policysecurity
Mar 4, 2026

Organizations are struggling to implement AI Governance (rules and controls for AI use) because they lack clear requirements for evaluating solutions. A new RFP (request for proposal, a document used to ask vendors what they can do) Guide has been released to help security leaders shift from trying to track every AI app to instead monitoring AI interactions (the moments when employees use AI tools), using eight key evaluation areas like discovery, policy enforcement, and real-time blocking of data leaks.

Fix: The source mentions a new RFP Guide for Evaluating AI Usage Control and AI Governance Solutions as the tool to address this problem, and recommends using its eight-pillar framework (AI Discovery & Coverage, Contextual Awareness, Policy Governance, Real-Time Enforcement, Auditability, Architecture Fit, Deployment & Management, and Vendor Futureproofing) to evaluate vendors rather than relying on legacy security tools that lack interaction-level visibility.

The Hacker News
06

China's Xiaomi tells CNBC it's planning a yearly smartphone chip release and its own AI assistant for overseas

industry
Mar 4, 2026

Xiaomi plans to release a new smartphone processor chip (a specialized circuit that powers devices) every year, starting with its XRing O1 chip, and is developing its own AI assistant for overseas markets to compete with companies like Apple and Samsung. The company aims to combine its custom chip, HyperOS operating system (software that manages the phone), and AI assistant into devices launching in China this year before expanding internationally, though it may partner with Google's Gemini models for the overseas AI assistant.

CNBC Technology
07

Anthropic AI ultimatums and IP theft: The unspoken risk

securitypolicy
Mar 4, 2026

Anthropic's Claude AI faces two simultaneous pressures that create security risks for enterprises: illegal extraction campaigns by China-based AI companies (who ran millions of interactions through fake accounts to study Claude's capabilities in reasoning, tool use, and coding), and demands from the US government to remove safety guardrails (called guardrails, the built-in restrictions that prevent misuse) to enable military and surveillance applications. These geopolitical pressures mean frontier AI models (advanced, cutting-edge AI systems) are no longer neutral tools but are now intelligence surfaces that CISOs (chief information security officers, executives responsible for security) must consider when deciding whether to deploy them.

CSO Online
08

Quit ChatGPT: right now! Your subscription is bankrolling authoritarianism | Rutger Bregman

policy
Mar 4, 2026

This article argues that people should cancel their ChatGPT subscriptions as part of a grassroots boycott called QuitGPT, which the author claims is one of the most significant consumer boycotts in recent history. OpenAI, the company behind ChatGPT, is losing billions of dollars and its CEO has admitted to product failures, according to the article. The author encourages Europeans to join the over one million people who have already cancelled their subscriptions to send a signal to Silicon Valley.

The Guardian Technology
09

AI-powered attack kits go open source, and CyberStrikeAI may be just the beginning

securitysafety
Mar 3, 2026

CyberStrikeAI is an open source platform that automates cyberattacks using AI, making it easy for attackers of any skill level to launch sophisticated attacks by typing a few commands. The tool packages over 100 attack capabilities into a single system and is linked to a threat actor who breached hundreds of Fortinet FortiGate firewalls (network security devices). Security experts warn this represents a dangerous trend of AI-powered attack tools becoming more accessible to criminals.

CSO Online
10

Sam Altman tells OpenAI staffers that military's 'operational decisions' are up to the government

policy
Mar 3, 2026

OpenAI CEO Sam Altman told employees that the company cannot make decisions about how the Department of Defense uses its AI technology, saying those choices rest with military leadership. Altman acknowledged the announcement of OpenAI's deal to deploy AI models on classified Pentagon networks looked "opportunistic and sloppy," but defended the partnership by noting the Pentagon respects safety concerns and wants to work collaboratively with the company.

CNBC Technology
Prev1...6162636465...268Next
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026
critical

CISA: New Langflow flaw actively exploited to hijack AI workflows

BleepingComputerMar 26, 2026
Mar 26, 2026