aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
71
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 4/371
VIEW ALL
01

Apple’s AirPods with cameras for AI are apparently close to production

industry
May 7, 2026

Apple is developing AirPods with built-in cameras that are approaching production testing stages. The cameras won't take traditional photos or videos, but instead capture low-resolution visual information that users can ask Siri (Apple's AI assistant) to analyze, such as identifying what meals they could make with ingredients in front of them.

Critical This Week5 issues
critical

CVE-2026-42271: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.74.2 to before vers

CVE-2026-42271NVD/CVE DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

The Verge (AI)
02

Worries About AI’s Risks to Humanity Loom Over the Trial Pitting Musk Against OpenAI’s Leaders

policysafety
May 7, 2026

A legal trial between Elon Musk and OpenAI leaders centers on whether OpenAI broke promises to remain a nonprofit, but testimony has also highlighted broader AI safety concerns, including risks like job displacement, misinformation, and the potential dangers of AGI (artificial general intelligence, an advanced AI system that surpasses humans at many tasks). Expert witness Stuart Russell warned that the competitive race to develop AGI first poses a threat to humanity, though the judge has tried to keep the trial focused on the nonprofit dispute rather than AI's dangers.

SecurityWeek
03

ICYMI: April 2026 @AWS Security

securityindustry
May 7, 2026

This is a monthly digest of AWS security resources from April 2026 covering topics like AI security, identity management, and data protection. The posts provide practical guidance on securing agentic AI systems (AI systems that can act independently), implementing fine-grained access controls using ABAC (attribute-based access control, which grants permissions based on user characteristics rather than just roles), and defending against emerging threats like token abuse and privilege escalation attacks.

AWS Security Blog
04

ChatGPT’s ‘Trusted Contact’ will alert loved ones of safety concerns

safety
May 7, 2026

OpenAI is launching an optional safety feature called 'Trusted Contact' that lets adult ChatGPT users designate an emergency contact (friend, family member, or caregiver) who will be notified if the AI detects concerning conversations about self-harm or suicide. The feature is designed to connect people in crisis with trusted people they know, working alongside existing mental health helplines.

The Verge (AI)
05

Behind the Scenes Hardening Firefox with Claude Mythos Preview

industrysecurity
May 7, 2026

Mozilla used early access to Claude Mythos (an advanced AI model) to find and fix hundreds of security vulnerabilities in Firefox that had gone undetected for years. The AI became much more useful for this task once the model became more capable and Mozilla developed better techniques for controlling the AI, filtering out false reports, and combining multiple AI analyses together.

Simon Willison's Weblog
06

Notes on the xAI/Anthropic data center deal

industrypolicy
May 7, 2026

Anthropic has signed a deal with SpaceX/xAI to use all capacity from the Colossus 1 data center, which has a poor environmental record including unpermitted gas turbines that lack pollution controls and have been linked to increased hospital admissions from poor air quality. The deal also creates a potential supply chain risk (a vulnerability where a company depends on another company that could cut off essential services) since Elon Musk, who owns xAI, has stated he reserves the right to reclaim the compute if Anthropic's AI causes harm, with the criteria for 'harm' decided by Musk himself.

Simon Willison's Weblog
07

GHSA-mcfx-4vc6-qgxv: BentoML has Information Disclosure in `bentoml build` via symlink traversal in the build context

security
May 7, 2026

BentoML's `bentoml build` command has a symlink traversal vulnerability (following attacker-controlled symbolic links, which are shortcuts to files) that allows attackers to copy files from outside the build directory into the generated Bento artifact (the packaged application). If a developer builds an untrusted repository, an attacker can hide a symlink pointing to sensitive files like credentials or API tokens, and these files will be copied into the final package where they could be leaked through export or upload workflows.

GitHub Advisory Database
08

How Anthropic’s Mythos has rewritten Firefox’s approach to cybersecurity

securityresearch
May 7, 2026

Anthropic's Mythos model, an advanced AI system for finding bugs, has dramatically improved Firefox's ability to discover software vulnerabilities (flaws in code that attackers can exploit), unearthing thousands of high-severity bugs including some hidden for over a decade. Unlike older AI bug-finding tools that produced many false positives (incorrect alerts), Mythos uses agentic systems (AI that can assess and filter its own work) to deliver higher-quality results, leading Firefox to ship 423 bug fixes in April 2026 compared to 31 a year earlier. However, Mozilla's engineers still manually write and review patches rather than deploying AI-generated code directly, as they have not found the fix-writing process automatable.

TechCrunch (Security)
09

Claude Code OAuth Tokens Can Be Stolen Through Stealthy MCP Hijacking

security
May 7, 2026

Attackers can steal OAuth tokens (digital keys that grant access to connected services) from Claude Code, an AI system that performs tasks autonomously, through a man-in-the-middle attack (intercepting communication between two parties). The attack involves installing a malicious npm package that modifies Claude Code's configuration file to redirect all traffic through the attacker's infrastructure, allowing them to capture tokens while remaining undetected.

SecurityWeek
10

OpenClaw and Claude can put your AI-generated podcasts in Spotify

industry
May 7, 2026

Save to Spotify is a command-line tool (a program you run through text commands rather than clicking buttons) that lets AI agents like Claude Code create audio summaries and podcasts that automatically save to your Spotify library. Users can set it up by downloading the tool from GitHub and then asking their AI to create content with the instruction to 'save to Spotify,' and the resulting podcast will appear in their Spotify feed alongside regular episodes.

The Verge (AI)
Prev123456...371Next
critical

CVE-2026-42203: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.80.5 to before vers

CVE-2026-42203NVD/CVE DatabaseMay 8, 2026
May 8, 2026
critical

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

SecurityWeekMay 7, 2026
May 7, 2026
critical

GHSA-9h64-2846-7x7f: Axonflow fixed bugs by implementing multi-tenant isolation and access-control hardening

GitHub Advisory DatabaseMay 6, 2026
May 6, 2026
critical

GHSA-gmvf-9v4p-v8jc: fast-jwt: JWT auth bypass due to empty HMAC secret accepted by async key resolver

CVE-2026-44351GitHub Advisory DatabaseMay 6, 2026
May 6, 2026