aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
69
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 8/371
VIEW ALL
01

Live blog: Code w/ Claude 2026

industry
May 6, 2026

This live blog covers Anthropic's Code w/ Claude 2026 event, where the company announced increased rate limits for developers, a partnership with SpaceX to use their Colossus data center for capacity, and three new Claude Managed Agents features: multi-agent orchestration (coordinating multiple AI agents to work together), outcomes-based iteration (setting success criteria for agents to achieve), and "Dreaming" (where agents review past sessions to self-improve). The blog notes that API volume has grown 17x year-over-year and highlights the importance of tool use, long context windows (allowing AI to process more information), and agentic loops (repeated cycles of agent reasoning and action) in modern AI development.

Critical This Week5 issues
critical

CVE-2026-42271: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.74.2 to before vers

CVE-2026-42271NVD/CVE DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Simon Willison's Weblog
02

Vibe coding and agentic engineering are getting closer than I'd like

safetyindustry
May 6, 2026

Simon Willison discusses how vibe coding (using AI to generate code without understanding or reviewing it) and agentic engineering (using AI tools while maintaining professional standards and code quality) are becoming harder to distinguish in practice. As AI coding tools become more reliable, even experienced engineers like Willison find themselves not reviewing all AI-generated code in production systems, which blurs the line between the two approaches and raises questions about responsible software development.

Simon Willison's Weblog
03

From Stuxnet to ChatGPT: 20 News Events That Shaped Cyber

securityindustry
May 6, 2026

This article is a retrospective review by Dark Reading marking their 20th anniversary, highlighting 20 major news events from the past two decades that have significantly influenced the cybersecurity industry and the threat landscape that security teams face today. The piece spans from Stuxnet (a sophisticated malware attack on industrial systems) to ChatGPT (a large language model AI), showing how the security field has evolved over time.

Dark Reading
04

Your AI Agents Are Already Inside the Perimeter. Do You Know What They're Doing?

securitypolicy
May 6, 2026

AI agents are being deployed in enterprises much faster than companies can manage them, creating a visibility problem because traditional identity and access management (IAM, systems that control who can access what) was designed for human users, not continuously-running software agents. About half of enterprise identity activity already happens outside the view of central IAM tools, leaving organizations unable to see what AI agents are operating, what data they access, or what permissions they use. The source describes using observability tools (systems that let you see what's happening) built into applications to discover AI agents and check compliance with security standards like NIST (the National Institute of Standards and Technology).

Fix: The source describes a capability called "Ask Orchid" built into Orchid Security's platform that applies identity observability at the application level to: automatically discover AI agents and their risk profiles, identify where AI agents are not in use for a complete picture, and provide recommended actions for oversight. For compliance checking, the same tool can examine how identity controls are actually implemented inside applications and compare them against NIST requirements.

The Hacker News
05

AlphaEvolve: How our Gemini-powered coding agent is scaling impact across fields

industry
May 6, 2026

AlphaEvolve is a Gemini-powered coding agent (an AI system that writes and optimizes code) that helps design and improve algorithms across science and industry. The system has achieved significant results including improving DNA sequencing accuracy by 30%, increasing electricity grid optimization from 14% to 88%, and enabling quantum computing simulations with 10x lower error rates.

DeepMind Safety Research
06

Chrome’s AI features may be hogging 4GB of your computer storage

safety
May 6, 2026

Google Chrome is automatically downloading a large 4GB file called weights.bin (a set of numerical values that power an AI model) to users' computers when certain AI features are enabled, which is unexpectedly consuming significant storage space. This file contains Google's Gemini Nano AI model, which runs Chrome's features like scam detection and writing assistance.

The Verge (AI)
07

Poisoned truth: The quiet security threat inside enterprise AI

securitysafety
May 6, 2026

AI data poisoning is a security threat where an AI model's training data or information sources become corrupted, causing the system to make decisions based on false information while appearing normal. This can happen through malicious attacks, but more often organizations poison their own systems by feeding AI models data from multiple conflicting sources like outdated files and incompatible databases. Unlike traditional cyberattacks that trigger visible alarms, poisoning is dangerous because no obvious damage appears, yet the AI produces plausible but incorrect answers affecting business decisions.

CSO Online
08

How ChatGPT learns about the world while protecting privacy

privacysafety
May 6, 2026

OpenAI trains ChatGPT using various data sources, including publicly available internet content and user conversations, to help the model learn broad knowledge and perform better. To protect privacy, OpenAI uses Privacy Filter (a tool that identifies and masks personal information in text) at multiple stages of training, and gives users control over their data through settings like the ability to disable model training, use temporary chats that auto-delete after 30 days, and manage or delete their account information.

Fix: OpenAI has implemented Privacy Filter, which identifies and masks personal information in text at multiple stages in the training process. Users can disable model training by going to Settings, then Data Controls, and turning off 'Improve the model for everyone.' Alternatively, users can use Temporary Chat (which does not appear in chat history, does not create memories, and is not used to train models, with conversations deleted after 30 days). Users can also review, edit, or delete saved memories, export their data, delete their account, or submit privacy requests through the privacy request portal.

OpenAI Blog
09

Introducing ChatGPT Futures: Class of 2026

industry
May 5, 2026

ChatGPT Futures honors college students from the class of 2026, the first generation to complete college with AI tools like ChatGPT available throughout their education. Rather than using AI to avoid work, these students are using it to build real projects faster, from research tools to accessibility software, demonstrating that AI amplifies human ambition and lowers barriers to turning ideas into tangible outcomes.

OpenAI Blog
10

How frontier enterprises are building an AI advantage

industry
May 5, 2026

Frontier enterprises (those using AI most extensively) now use 3.5x more AI intelligence per worker than typical firms, with the gap driven by deeper, more complex usage rather than just more messages. The key difference is that leading firms use agentic workflows (AI systems that can complete multi-step tasks with minimal human intervention), with frontier companies sending 16x more messages to coding tools like Codex per worker, moving from simple question-answering to delegating substantial work to AI agents.

OpenAI Blog
Prev1...678910...371Next
critical

CVE-2026-42203: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.80.5 to before vers

CVE-2026-42203NVD/CVE DatabaseMay 8, 2026
May 8, 2026
critical

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

SecurityWeekMay 7, 2026
May 7, 2026
critical

GHSA-9h64-2846-7x7f: Axonflow fixed bugs by implementing multi-tenant isolation and access-control hardening

GitHub Advisory DatabaseMay 6, 2026
May 6, 2026
critical

GHSA-gmvf-9v4p-v8jc: fast-jwt: JWT auth bypass due to empty HMAC secret accepted by async key resolver

CVE-2026-44351GitHub Advisory DatabaseMay 6, 2026
May 6, 2026