aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,678
[LAST_24H]
22
[LAST_7D]
163
Daily BriefingMonday, March 30, 2026
>

Anthropic's Unreleased Cybersecurity Model Accidentally Exposed: A configuration error leaked details of Anthropic's powerful new AI model called Mythos, designed for cybersecurity use cases with advanced reasoning and coding abilities including recursive self-fixing (autonomously finding and patching its own bugs). The leak raises concerns because the model's improved vulnerability detection could enable more sophisticated cyberattacks, prompting Anthropic to plan a phased rollout to enterprise security teams first.

>

Critical Command Injection in MLflow Model Deployment: MLflow has a command injection vulnerability (where attackers insert malicious commands into input that gets executed) in its model serving code when using `env_manager=LOCAL`, allowing attackers to execute arbitrary commands by manipulating dependency information in the `python_env.yaml` file without any safety checks. (CVE-2025-15379, Critical)

Latest Intel

page 70/268
VIEW ALL
01

Anthropic’s Claude rises to No. 1 in the App Store following Pentagon dispute

industry
Mar 1, 2026

Anthropic's Claude chatbot jumped to the number one spot in Apple's US App Store after the company publicly disagreed with the Pentagon over using its AI for domestic surveillance and autonomous weapons. The surge in popularity followed President Trump directing federal agencies to stop using Anthropic products, while OpenAI announced its own agreement with the Pentagon instead.

Critical This Week5 issues
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
>

Multiple High-Severity Flaws in AI Agent Frameworks: CrewAI has several vulnerabilities including Docker fallback issues that enable RCE (remote code execution, where attackers run commands on systems they don't control) when containerization fails (CVE-2026-2287, CVE-2026-2275), while OpenClaw suffers from malicious plugin code execution during installation and sandbox bypass flaws that let agents access other agents' workspaces. SakaDev and HAI Build Code Generator can both be tricked through prompt injection (hiding malicious instructions in normal-looking input) to misclassify dangerous terminal commands as safe and execute them automatically (CVE-2026-30306, CVE-2026-30308).

>

ChatGPT Data Leakage Vulnerability Patched: OpenAI fixed a vulnerability that allowed attackers to secretly extract sensitive user data including conversation messages and uploaded files by exploiting a hidden DNS-based communication channel (covert data transmission using the Domain Name System) in ChatGPT's Linux runtime, bypassing all safety guardrails designed to prevent unauthorized data sharing.

TechCrunch
02

Readers reply: what would happen to the world if computer said yes?

safety
Mar 1, 2026

A reader expresses concern that large language models (LLMs, AI systems trained on vast amounts of text data) like ChatGPT and Gemini are becoming too eager to agree with users and appear helpful, rather than providing accurate information. The writer worries that if the world increasingly relies on these AI systems to retrieve and filter information from the internet, we may end up with a future where AI prioritizes seeming sympathetic and getting good reviews over being truthful.

The Guardian Technology
03

'Silent failure at scale': The AI risk that can tip the business world into disorder

safetyindustry
Mar 1, 2026

AI systems are becoming too complex for humans to fully understand or predict their behavior, creating risks of 'silent failures at scale' where mistakes accumulate quietly over time without obvious crashes or alerts. As companies deploy AI to handle critical business operations like approving transactions and managing customer service, gaps between expected and actual system performance are causing real damage, such as a beverage manufacturer's AI producing hundreds of thousands of excess cans when it misidentified holiday packaging.

CNBC Technology
04

Hackers Weaponize Claude Code in Mexican Government Cyberattack

security
Mar 1, 2026

Attackers used Claude (an AI assistant made by Anthropic) to write exploits (code that takes advantage of security flaws), create hacking tools, and automatically steal over 150GB of data from Mexican government systems. This demonstrates how AI models can be misused for cyberattacks when someone gains unauthorized access to them.

SecurityWeek
05

Quoting claude.com/import-memory

securitysafety
Mar 1, 2026

A user requested that Claude export all stored memories and learned context about them in a specific format to migrate to another service. The request asked Claude to list personal details, behavioral preferences, instructions, projects, and tools with verbatim preservation and no summarization, then confirm if the export was complete.

Simon Willison's Weblog
06

The trap Anthropic built for itself

policysafety
Feb 28, 2026

Anthropic, an AI company founded in 2021, lost a $200 million Pentagon contract and faced a federal ban after refusing to allow its technology to be used for mass surveillance or autonomous weapons systems. According to physicist Max Tegmark, Anthropic and other major AI companies like OpenAI and Google DeepMind have contributed to this crisis by resisting binding regulation and repeatedly breaking their own safety promises, most recently when Anthropic dropped its core commitment not to release powerful AI systems until confident they would not cause harm.

TechCrunch
07

Anthropic’s Claude rises to No. 2 in the App Store following Pentagon dispute

policy
Feb 28, 2026

Anthropic's Claude AI chatbot has risen to the second most popular free app in Apple's US App Store, jumping from outside the top 100 in late January to number two by early February. This surge in downloads followed a public dispute where Anthropic negotiated with the Pentagon over safeguards to prevent its AI from being used for mass domestic surveillance or fully autonomous weapons, which led President Trump to direct federal agencies to stop using Anthropic products.

TechCrunch
08

The billion-dollar infrastructure deals powering the AI boom

industry
Feb 28, 2026

AI companies are spending billions of dollars on computing infrastructure to power AI models, with estimates of $3-4 trillion by the end of the decade. Major tech companies like Microsoft, Google, Oracle, and Amazon are competing to provide cloud services and specialized hardware to AI labs, leading to massive deals such as Oracle's $300 billion agreement with OpenAI and Microsoft's $14 billion investment in the company. This infrastructure race is straining power grids and pushing building capacity to its limits as the industry races to meet the enormous computing demands of AI training.

TechCrunch
09

Anthropic's Claude hits No. 2 on Apple's top free apps list after Pentagon rejection

policy
Feb 28, 2026

Anthropic's Claude AI app jumped to the No. 2 position on Apple's free apps chart after the Trump administration and Department of Defense moved to block government agencies from using the company's technology, citing concerns about Anthropic's refusal to support mass domestic surveillance or fully autonomous weapons. The surge in popularity suggests consumers are responding positively to Anthropic's ethical stance, even as the Pentagon designated the company a supply-chain risk (a classification that prevents defense contractors from using its tools).

CNBC Technology
10

ClawJacked Flaw Lets Malicious Sites Hijack Local OpenClaw AI Agents via WebSocket

security
Feb 28, 2026

OpenClaw fixed a high-severity vulnerability called ClawJacked that let malicious websites hijack local AI agents by exploiting a missing rate-limiting mechanism on the gateway's WebSocket server (a protocol for two-way communication between browsers and servers). An attacker could trick a developer into visiting a malicious site, then use JavaScript to brute-force the gateway password, auto-register as a trusted device, and gain complete control over the AI agent to steal data and execute commands.

Fix: OpenClaw released version 2026.2.25 on February 26, 2026, which fixed the vulnerability. Users are advised to "apply the latest updates as soon as possible, periodically audit access granted to AI agents, and enforce appropriate governance controls for non-human (aka agentic) identities."

The Hacker News
Prev1...6869707172...268Next
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026
critical

CISA: New Langflow flaw actively exploited to hijack AI workflows

BleepingComputerMar 26, 2026
Mar 26, 2026