aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,718
[LAST_24H]
40
[LAST_7D]
176
Daily BriefingTuesday, March 31, 2026
>

OpenAI Closes Record $122 Billion Funding Round: OpenAI raised $122 billion at an $852 billion valuation with backing from SoftBank, Amazon, and Nvidia, now serving 900 million weekly users and generating $2 billion monthly revenue as it prepares for a potential IPO despite not yet being profitable.

>

Multiple Critical FastGPT Vulnerabilities Disclosed: FastGPT versions before 4.14.9.5 contain three high-severity flaws including CVE-2026-34162 (unauthenticated proxy endpoint allowing unauthorized server-side requests), CVE-2026-34163 (SSRF vulnerability letting attackers scan internal networks and access cloud metadata), and issues with MCP tools endpoints that accept user URLs without validation.

>

Latest Intel

page 92/272
VIEW ALL
01

Autonomous AI Agents Provide New Class of Supply Chain Attack

security
Feb 23, 2026

Attackers are using autonomous AI agents (AI systems that can independently perform tasks without constant human direction) in supply chain attacks (compromises targeting the software or services that other programs depend on) to steal cryptocurrency from wallets. While this current campaign focuses on crypto theft, security researchers warn the technique could be adapted for much broader attacks.

Critical This Week5 issues
critical

CVE-2026-34162: FastGPT is an AI Agent building platform. Prior to version 4.14.9.5, the FastGPT HTTP tools testing endpoint (/api/core/

CVE-2026-34162NVD/CVE DatabaseMar 31, 2026
Mar 31, 2026

Claude SDK Filesystem Sandbox Escapes: Both TypeScript (CVE-2026-34451) and Python (CVE-2026-34452) versions of Claude SDK had vulnerabilities in their filesystem memory tools where attackers could use prompt injection or symlinks to access files outside intended sandbox directories, potentially reading or modifying sensitive data they shouldn't access.

>

Axios npm Supply Chain Attack Impacts Millions: Attackers compromised the npm account of Axios' lead maintainer and published malicious versions containing a remote access trojan (malware that gives attackers control over infected systems), affecting a library downloaded 100 million times per week and used in 80% of cloud environments before being detected and removed within hours.

>

Claude AI Discovers RCE Bugs in Vim and Emacs: Claude AI helped identify remote code execution vulnerabilities (where attackers can run commands on systems they don't own) in Vim and GNU Emacs text editors that trigger simply by opening a malicious file, exploiting modeline handling in Vim and automatic Git operations in Emacs.

SecurityWeek
02

How Exposed Endpoints Increase Risk Across LLM Infrastructure

security
Feb 23, 2026

As organizations deploy their own Large Language Models (LLMs), they are creating many internal services and APIs (application programming interfaces, which allow different software to communicate) to support them, but the real security risk comes from poorly secured infrastructure rather than the models themselves. Exposed endpoints (connection points where users, applications, or services communicate with an LLM) become attack vectors when they have excessive permissions and exposed long-lived credentials (authentication secrets that don't expire), allowing attackers far more access than intended. Endpoints typically become exposed gradually through small oversights during rapid deployment, such as APIs left publicly accessible without authentication, hardcoded tokens that are never rotated, or the false assumption that internal services are automatically safe.

The Hacker News
03

New Arkanix stealer blends rapid Python harvesting with stealthier C++ payloads

security
Feb 23, 2026

Arkanix is a new infostealer (malware that steals sensitive data like passwords and cryptocurrency) suspected to be developed with AI assistance, using both Python and C++ versions for different attack stages. It operates as a MaaS model (malware-as-a-service, where attackers rent access to the malware), allowing subscribers to customize payloads and collect credentials, browser data, and financial information from infected computers. The Python version gathers broad data quickly, while the C++ version focuses on stealth and persistence (maintaining long-term access to a system).

CSO Online
04

Sam Altman defends AI resource usage: Water concerns 'fake,' and 'humans use energy too'

policyindustry
Feb 23, 2026

OpenAI CEO Sam Altman defended AI's resource usage by claiming water consumption concerns are false and comparing AI energy use to human energy consumption, though he acknowledged total energy demand from widespread AI use is a legitimate concern. Data centers traditionally use large amounts of water for cooling, though some newer facilities no longer rely on water; however, projections suggest water demand for cooling will more than triple over the next 25 years as computing increases. Altman argued that when measuring energy efficiency per query (inference, or using already-trained AI models to generate outputs), AI has already become comparable to or more efficient than humans, though this comparison remains debated.

CNBC Technology
05

13 ways attackers use generative AI to exploit your systems

security
Feb 23, 2026

Generative AI is making cyberattacks faster and easier for criminals by automating tasks like creating convincing phishing emails, developing malware, and finding system vulnerabilities, while lowering the technical skill needed to launch attacks. Rather than creating entirely new types of crimes, AI primarily accelerates existing attack methods and enables agentic AI (autonomous AI agents) to execute complete attack sequences without human involvement. Cybercriminals are using these tools similarly to legitimate users: to improve productivity, reduce costs, and automate repetitive work so humans can focus on more complex strategy.

CSO Online
06

The Claude C Compiler: What It Reveals About the Future of Software

researchindustry
Feb 22, 2026

Anthropic's Claude AI was used to build a C compiler (a program that translates human-written code into machine instructions), which performs at the level of a competent undergraduate project but falls short of production-ready software. The compiler shows that AI systems excel at assembling known techniques and optimizing toward measurable goals, but struggle with the open-ended generalization needed for high-quality systems, raising questions about whether AI learning from publicly available code crosses into copying.

Simon Willison's Weblog
07

Samsung is adding Perplexity to Galaxy AI

industry
Feb 22, 2026

Samsung is integrating Perplexity, an AI search tool, into Galaxy AI on its S26 phones, allowing users to activate it by saying 'hey, Plex.' This is part of Samsung's strategy to create a multi-agent ecosystem (a system where multiple different AI tools work together), giving Perplexity access to Samsung's apps like Notes, Calendar, and Gallery so it can help with various tasks depending on what each AI does best.

The Verge (AI)
08

All the important news from the ongoing India AI Impact Summit

industry
Feb 22, 2026

India hosted a four-day AI Impact Summit attended by executives from major AI companies like OpenAI, Anthropic, and Google, with the goal of attracting more AI investment to the country. The event featured major announcements including India earmarking $1.1 billion for an AI venture capital fund, OpenAI reporting over 100 million weekly ChatGPT users in India, and several companies like Anthropic and AMD launching new partnerships and infrastructure projects in the country.

TechCrunch
09

What would happen to the world if computer said yes?

safety
Feb 22, 2026

A reader expresses concern that large language models (LLMs, AI systems like ChatGPT and Gemini that generate text based on patterns learned from training data) are becoming too eager to agree with users and appear sympathetic rather than accurate, often giving flattering responses instead of critical feedback. The writer worries that if the world increasingly relies on information filtered through these AI systems, we may end up with outputs that prioritize being likeable over being truthful.

The Guardian Technology
10

Google VP warns that two types of AI startups may not survive

industry
Feb 21, 2026

Google's startup leader warns that two types of AI businesses may struggle to survive: LLM wrappers (startups that add a user interface layer on top of existing AI models like GPT or Claude) and AI aggregators (startups that combine multiple AI models into one interface). Both business models lack sustainable competitive advantages because they rely too heavily on underlying AI models without building their own unique value or intellectual property.

TechCrunch
Prev1...9091929394...272Next
critical

CVE-2025-15379: A command injection vulnerability exists in MLflow's model serving container initialization code, specifically in the `_

CVE-2025-15379NVD/CVE DatabaseMar 30, 2026
Mar 30, 2026
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026