aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 39/371
VIEW ALL
01

llm-openai-via-codex 0.1a0

industry
Apr 23, 2026

This is a brief announcement about llm-openai-via-codex version 0.1a0, a tool that connects OpenAI's services with the llm command-line interface. The post appears to be from Simon Willison's monthly briefing on LLM developments from April 2026.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Simon Willison's Weblog
02

Anthropic’s Mythos breach was humiliating

securitysafety
Apr 23, 2026

Anthropic's Claude Mythos model, which the company claimed was too dangerous to release publicly due to its advanced cybersecurity capabilities, was accessed by unauthorized users since the day the company announced it would share the model with selected companies for testing. The breach undermines Anthropic's reputation as a company focused on AI safety.

The Verge (AI)
03

OpenAI announces GPT-5.5, its latest artificial intelligence model

industry
Apr 23, 2026

OpenAI released GPT-5.5, a new AI model that performs better at coding, using computers, and research with less guidance from users. The model meets OpenAI's "High" cybersecurity risk classification, meaning it could amplify existing pathways to harm, though it does not reach the "Critical" threshold. The company conducted third-party testing and red teaming (adversarial testing where security experts try to break the system) and iterated on cyber safeguards for months before release.

CNBC Technology
04

Enabling trust and learner agency in lifelong learning: A dual-chain, privacy-preserving credential architecture

securityprivacy
Apr 23, 2026

This academic paper proposes a dual-chain, privacy-preserving credential architecture designed to enable trust and learner agency in lifelong learning systems. The work focuses on creating secure credential management that protects learner privacy while maintaining verifiable educational records across multiple institutions and learning contexts.

Elsevier Security Journals
05

OpenAI says its new GPT-5.5 model is more efficient and better at coding

industry
Apr 23, 2026

OpenAI released GPT-5.5, a new AI model designed to be more efficient and better at coding tasks than its predecessor GPT-5.4. The model can handle complex, multi-step tasks by planning its own approach, using available tools, and checking its own work without requiring users to carefully direct every action.

The Verge (AI)
06

The Guardian view on Anthropic’s Claude Mythos: when AI finds every flaw, who controls the internet? | Editorial

securitysafety
Apr 23, 2026

Anthropic created Claude Mythos, an AI model that can autonomously find and exploit zero-day vulnerabilities (previously unknown security flaws that hackers don't yet know about), write code to exploit them, and potentially take over major operating systems and web browsers, but the company chose not to release it publicly due to these risks. To address the threat, Anthropic launched Project Glasswing, partnering with 40 organizations to help them "patch" (fix) vulnerabilities before attackers can exploit them, though all current partners are American companies.

Fix: Anthropic has named 40 organisations as partners under Project Glasswing to help mount a defence by asking them to "patch" vulnerabilities before hackers get a chance to exploit them.

The Guardian Technology
07

GHSA-pfm2-2mhg-8wpx: n8n-MCP Logs Sensitive Request Data on Unauthorized /mcp Requests

security
Apr 23, 2026

n8n-mcp (a tool that connects n8n automation software to external services) was logging sensitive information like bearer tokens and API keys when it received unauthorized requests to its HTTP endpoint, even though it correctly rejected those requests. This happened because the logs captured request metadata before checking authentication, which could expose secrets if logs were shared or stored outside secure boundaries.

Fix: Upgrade to n8n-mcp v2.47.11 or later using 'npx n8n-mcp@latest' for npm or 'docker pull ghcr.io/czlonkowski/n8n-mcp:latest' for Docker. If immediate upgrade is not possible, restrict network access to the HTTP port using a firewall or reverse proxy, or switch to stdio transport mode by setting MCP_MODE=stdio.

GitHub Advisory Database
08

Bad Memories Still Haunt AI Agents

security
Apr 23, 2026

Cisco discovered a serious vulnerability in how Anthropic (an AI company) stores and manages memories, which are pieces of information that AI systems keep between conversations. While Anthropic fixed this particular issue, security experts warn that poorly managed memory files remain a widespread risk to AI systems.

Fix: Anthropic fixed the vulnerability that Cisco found. The source does not provide additional details about the specific fix, version numbers, or other mitigation steps.

Dark Reading
09

THE PEOPLE DO NOT YEARN FOR AUTOMATION

policyindustry
Apr 23, 2026

This article discusses 'software brain,' a way of thinking that sees everything through algorithms and automation, which has been amplified by AI development. Despite widespread enthusiasm from tech executives, polling shows that most Americans—particularly Gen Z—are increasingly skeptical or angry about AI, with only 35 percent excited about it and over 80 percent concerned about potential harms.

The Verge (AI)
10

You’re about to feel the AI money squeeze

industry
Apr 23, 2026

Anthropic, an AI company, has severely restricted OpenClaw, a popular AI agent tool (software that uses AI to perform tasks autonomously), requiring users to pay significantly more to continue using it. The restriction was implemented because Anthropic needed to reduce strain on its systems and increase profitability, as the tool's usage patterns weren't sustainable under their existing subscription model.

The Verge (AI)
Prev1...3738394041...371Next