aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,657
[LAST_24H]
7
[LAST_7D]
152
Daily BriefingMonday, March 30, 2026
>

Anthropic's Leaked "Mythos" Model Raises Dual-Use Security Concerns: An unreleased Anthropic AI model called Mythos was accidentally exposed through a configuration error, revealing advanced reasoning and coding abilities specifically aimed at cybersecurity. The model's improved capability to find and exploit software vulnerabilities, plus its ability to autonomously fix its own code problems, could enable both more sophisticated cyberattacks and better defenses.

>

Mistral Secures $830M for European AI Data Center: French AI startup Mistral raised $830 million in debt financing to build a Paris-area data center with thousands of Nvidia GPUs (specialized chips used for AI training) to train its large language models, aiming for 200 MW of European computing capacity by 2027.

Latest Intel

page 45/266
VIEW ALL
01

Google’s Gemini AI is getting a bigger role across Docs, Sheets, and Slides

industry
Mar 10, 2026

Google is expanding its Gemini AI assistant into more of its Workspace apps, including a new chat window in Google Docs that lets users describe documents for AI to create, AI-powered spreadsheet generation, and a Gemini-powered search feature in Drive. The Gemini assistant can pull information from the web, Drive, Gmail, and other sources to help users with their work.

Critical This Week5 issues
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
>

Critical Command Injection in MLflow Model Deployment: MLflow has a command injection vulnerability (where an attacker inserts malicious commands into input that gets executed) in its model serving code when deploying models with `env_manager=LOCAL`. The flaw allows attackers to execute arbitrary commands on deployment systems by inserting malicious content into the `python_env.yaml` file, which MLflow reads and uses in shell commands without validation. (CVE-2025-15379, Critical)

The Verge (AI)
02

The Download: AI’s role in the Iran war, and an escalating legal fight

policyindustry
Mar 10, 2026

This newsletter covers multiple AI and technology developments, including AI's expanding role in military decision-making during the Iran conflict through 'vibe-coded' intelligence dashboards (AI systems that present information in visually appealing but potentially unreliable formats), legal disputes between AI companies and governments, and emerging threats like GPS jamming in the Middle East. The piece also highlights concerns about AI cloning real people's voices without consent, developments in AI agents, and psychological effects of AI companions on users.

MIT Technology Review
03

Sandbar secures $23M Series A for its AI note-taking ring

industry
Mar 10, 2026

Sandbar, a startup founded by former Meta employees, raised $23 million to develop the Stream ring, a wearable device with a microphone that records notes and lets users chat with an AI assistant through a phone app. The ring's microphone is off by default and only activates when users lift their hand to their face, which signals intent for private note-taking rather than recording surrounding conversations.

TechCrunch
04

Trump's war predictions, Pershing Square files for IPO, Anthropic's lawsuit and more in Morning Squawk

policy
Mar 10, 2026

Anthropic, an AI company, filed a lawsuit against the federal government after the Pentagon blacklisted it as a 'supply chain risk' (a security classification typically reserved for foreign adversaries), claiming the move is unlawful and causes irreparable harm. The blacklisting followed Anthropic's disagreement with the Pentagon over how its AI systems could be used. Defense experts worry this precedent could harm U.S. competitiveness by cutting off access to a major American AI vendor.

CNBC Technology
05

Global Cyber Attacks Remain Near Record Highs in February 2026 Despite Ransomware Decline

security
Mar 10, 2026

In February 2026, organizations worldwide faced an average of 2,086 cyber attacks per week, a 9.6% increase from the previous year, indicating that high attack volumes are now a constant threat rather than a temporary spike. While ransomware attacks declined compared to last year, overall attack activity remains near record levels due to automation, expanded digital systems, and security risks from enterprise GenAI (generative AI used by businesses) usage.

Check Point Research
06

Escape Raises $18 Million to Automate Pentesting

industry
Mar 10, 2026

Escape, a company that uses AI agents (software programs that act autonomously to complete tasks) to automate pentesting (simulated security attacks to find vulnerabilities), has raised $18 million in funding. The company plans to use this money to improve its AI capabilities and expand its teams.

SecurityWeek
07

How to Stop AI Data Leaks: A Webinar Guide to Auditing Modern Agentic Workflows

securitysafety
Mar 10, 2026

AI Agents (software programs that automatically perform tasks like sending emails or moving data) create security risks because they have broad access to sensitive information with little oversight, making them targets for hackers who can trick them into leaking company secrets. Traditional security tools were designed to protect human users, not autonomous digital workers, leaving AI agents largely invisible to security teams. The article promotes an upcoming webinar that promises to explain how hackers target these agents and how to secure them without overly restricting their capabilities.

The Hacker News
08

Family of child injured in Canada school shooting sues OpenAI

safetypolicy
Mar 10, 2026

A family is suing OpenAI after their 12-year-old daughter was critically injured in a Canadian school shooting, claiming that OpenAI knew the suspect was planning an attack through ChatGPT conversations but failed to alert authorities. The suspect's account was banned in June 2025 after employees flagged messages about gun violence as indicating imminent harm, but police were never notified, and the suspect later opened a second account to continue planning.

Fix: According to OpenAI's statement, the company has implemented several changes: enlisting mental health and behavioral experts to assess cases, making the criteria for police referral more flexible, strengthening detection systems to prevent evasion of safeguards, and establishing a direct point of contact with Canadian law enforcement to quickly flag cases with potential for real-world violence. OpenAI's CEO also pledged to strengthen protocols on notifying police about potentially harmful interactions.

BBC Technology
09

Oracle earnings will show whether its expensive AI bet is starting to pay off

industry
Mar 10, 2026

Oracle is reporting earnings on Tuesday as investors try to determine whether its massive investment in AI infrastructure is profitable. The company raised $50 billion in financing (debt and equity) to build data centers, mainly to serve OpenAI, and bond investors are watching closely because Oracle had to borrow heavily compared to other major cloud computing companies, raising concerns about its financial health and credit rating.

CNBC Technology
10

Improving instruction hierarchy in frontier LLMs

safetyresearch
Mar 10, 2026

AI systems receive instructions from multiple sources (system policies, developers, users, and online data), and models must learn to prioritize the most trustworthy ones to stay safe. When models treat untrusted instructions as authoritative, they can be tricked into revealing private information, following harmful requests, or falling victim to prompt injection (hidden malicious instructions hidden in input data). OpenAI's solution uses a clear instruction hierarchy (System > developer > user > tool) and trains models with IH-Challenge, a reinforcement learning dataset designed to teach models to follow high-priority instructions even when lower-priority ones conflict with them.

Fix: OpenAI's models are trained on a clear instruction hierarchy where System instructions have highest priority, followed by developer instructions, then user instructions, then tool outputs. The company also created IH-Challenge, a reinforcement learning training dataset that generates conversations with conflicting instructions where high-priority instructions are kept simple and objectively gradable, ensuring models learn to prioritize correctly without resorting to useless shortcuts like over-refusing benign requests.

OpenAI Blog
Prev1...4344454647...266Next
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026
critical

CISA: New Langflow flaw actively exploited to hijack AI workflows

BleepingComputerMar 26, 2026
Mar 26, 2026