aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 58/371
VIEW ALL
01

Codex for (almost) everything

industry
Apr 16, 2026

Codex, an AI tool used by over 3 million developers weekly, has received a major update that lets it operate computers directly by seeing, clicking, and typing, generate images, remember user preferences, and integrate with 90+ developer tools and apps. The update adds features like background computer use (where the AI can work on your Mac without interfering with your own work), an in-app browser for web development, image generation, and the ability to schedule long-term tasks across multiple days or weeks. These improvements are designed to help developers move faster through all stages of software development, from writing code to reviewing changes, all within one workspace.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

OpenAI Blog
02

Human Trust of AI Agents

researchsafety
Apr 16, 2026

Researchers studied how humans behave when playing strategic games (like a guessing game where players try to guess 2/3 of the average guess) against AI language models (LLMs) versus other humans. They found that people choose much lower numbers when playing against LLMs, especially people who are good at strategic thinking, because they believe LLMs will reason carefully and cooperate fairly rather than try to win.

Schneier on Security
03

Anthropic unveils plans for major UK expansion after OpenAI announces first permanent London office

industry
Apr 16, 2026

Anthropic, the company behind the Claude AI chatbot, announced plans to expand its London office to accommodate 800 people, following a similar move by competitor OpenAI. The expansion reflects growing interest in establishing AI research and development hubs in the UK, which has strong AI talent and institutions focused on AI safety.

CNBC Technology
04

Claude Code, Gemini CLI, GitHub Copilot Agents Vulnerable to Prompt Injection via Comments

security
Apr 16, 2026

Researchers discovered a vulnerability called 'Comment and Control' that affects multiple AI coding assistants, including Claude Code, Gemini CLI, and GitHub Copilot Agents. The attack works by hiding malicious instructions in code comments, which the AI systems then follow as if they were legitimate user requests. This is a type of prompt injection (tricking an AI by hiding instructions in its input) that specifically targets AI tools designed to help developers write code.

SecurityWeek
05

Frontier AI for Defenders: CrowdStrike and OpenAI TAC

industrypolicy
Apr 16, 2026

CrowdStrike has been selected for OpenAI's Trusted Access for Cyber (TAC) program, which gives verified security defenders controlled access to GPT-5.4-Cyber, a frontier model (a cutting-edge AI system designed for a specific task) built for defensive cybersecurity. As AI agents become more common in enterprise systems, CrowdStrike addresses security challenges by monitoring AI execution at endpoints (the individual computers and devices where AI actually runs), tracking over 1,800 AI applications to ensure governance and detect suspicious actions.

CrowdStrike Blog
06

GHSA-rr7j-v2q5-chgv: LangSmith SDK: Streaming token events bypass output redaction

security
Apr 15, 2026

The LangSmith SDK (a tool for monitoring AI applications) has a security flaw where its output redaction feature (hideOutputs in JavaScript, hide_outputs in Python) doesn't work for streaming token events. When an LLM produces streamed output, each piece of data is recorded as a new_token event with unredacted content that bypasses the redaction process entirely, potentially leaking sensitive information to LangSmith storage.

GitHub Advisory Database
07

Introducing GPT-Rosalind for life sciences research

industry
Apr 15, 2026

OpenAI has released GPT-Rosalind, a specialized AI model designed to help life sciences researchers work faster across biology, drug discovery, and medicine research. The model is built to assist with complex research workflows like literature review, hypothesis generation, and experimental planning by helping scientists connect to scientific tools and databases. It is available as a research preview through ChatGPT, Codex, and an API for qualified customers.

OpenAI Blog
08

Accelerating the cyber defense ecosystem that protects us all

securitypolicy
Apr 15, 2026

OpenAI has launched Trusted Access for Cyber, a program that gives advanced AI cybersecurity tools to defensive security teams while controlling access based on trust and validation. The program provides $10 million in API credits to help defenders of all sizes, from small open-source teams to major enterprises, use frontier AI models (advanced, cutting-edge AI systems) to protect digital infrastructure.

OpenAI Blog
09

The public sours on AI and data centers as Anthropic, OpenAI look to IPO and tech keeps spending

policyindustry
Apr 15, 2026

Public opinion on AI is declining in the United States, with 57% of voters believing AI's risks outweigh its benefits, creating challenges for companies like OpenAI and Anthropic as they prepare to go public. Tech companies are investing heavily in data centers (the large computing facilities that power AI systems) to build more powerful AI models, but these projects face growing opposition due to energy concerns, with $156 billion in data center projects blocked or delayed in 2025 and Maine passing the first state-wide data center ban. This negative sentiment and regulatory pushback could impact the valuations and public offerings of major AI companies.

CNBC Technology
10

Critical Nginx UI auth bypass flaw now actively exploited in the wild

security
Apr 15, 2026

A critical vulnerability in Nginx UI (CVE-2026-33032) leaves an unprotected endpoint that allows attackers to invoke privileged actions without logging in, enabling complete takeover of the web server by modifying configuration files. The flaw is being actively exploited in the wild, with over 2,600 publicly exposed instances at risk. Nginx UI is a popular web-based management interface for the Nginx web server, used by many organizations to control their servers.

Fix: Nginx released a fix in version 2.3.4 on March 15. The latest secure version is 2.3.6, released the week after the source was published. System administrators are recommended to apply these security updates as soon as possible.

BleepingComputer
Prev1...5657585960...371Next