aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSaturday, May 16, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 80/371
VIEW ALL
01

CVE-2026-34371: LibreChat is a ChatGPT clone with additional features. Prior to 0.8.4, LibreChat trusts the name field returned by the e

security
Apr 7, 2026

LibreChat, a ChatGPT alternative with extra features, had a vulnerability in versions before 0.8.4 where it didn't properly validate filenames from its code execution sandbox, allowing attackers to write files anywhere on the server using path traversal (sequences like ../ that navigate to parent directories). Any user able to run code through the sandbox could exploit this to write arbitrary files with the permissions of the LibreChat server.

Fix: This vulnerability is fixed in version 0.8.4.

NVD/CVE Database
02

Cracks in the Bedrock: Escaping the AWS AgentCore Sandbox

securityresearch
Apr 7, 2026

Researchers discovered that AWS Bedrock AgentCore's Code Interpreter sandbox, which is supposed to isolate AI agents from external networks, could be bypassed using DNS tunneling (a technique that hides data inside DNS queries to leak information out of restricted environments). Additionally, they found a critical security flaw where the microVM Metadata Service (a system that provides credentials to running programs) lacked proper authentication, potentially allowing attackers to steal sensitive credentials through SSRF attacks (server-side request forgery, where a program is tricked into making requests on behalf of an attacker).

Fix: AWS introduced internal remediations and outlined several important mitigation strategies for customers. The source notes that users cannot patch the managed environment directly but can leverage platform-level controls AWS provides. However, the specific details of these mitigation strategies and platform-level controls are not fully described in the provided excerpt.

Palo Alto Unit 42
03

Anthropic's Project Glasswing - restricting Claude Mythos to security researchers - sounds necessary to me

securitypolicy
Apr 7, 2026

Anthropic released Claude Mythos, a new AI model with exceptionally strong cybersecurity research abilities, but restricted access to only a small group of preview partners through Project Glasswing instead of releasing it publicly. The model can autonomously develop complex exploits (attacks that chain multiple vulnerabilities together to break into systems), finding thousands of high-severity vulnerabilities in major operating systems and web browsers, which is a major leap forward compared to older models like Claude Opus 4.6.

Simon Willison's Weblog
04

GHSA-8jxr-pr72-r468: Java-SDK has a DNS Rebinding Vulnerability

security
Apr 7, 2026

The java-sdk has a DNS rebinding vulnerability (an attack where a hacker tricks your browser into accessing a private server by manipulating domain name resolution) that allows attackers to make tool calls to local or private MCP (model context protocol, a system for AI agents to interact with tools) servers if you visit a malicious website. This happens because the java-sdk wasn't validating the Origin header (a security check that confirms requests come from trusted sources) before version 1.0.0, violating the MCP specification.

Fix: Users can mitigate this risk by: 1) Running the MCP server behind a reverse proxy (a security layer like Nginx or HAProxy that forwards requests and can validate headers) configured to strictly validate the Host and Origin headers, or 2) Using a framework that inherently enforces strict CORS (cross-origin resource sharing, a browser security feature that controls which websites can access your data) and Origin validation, such as Spring AI.

GitHub Advisory Database
05

GHSA-mh2q-q3fh-2475: OpenTelemetry-Go: multi-value `baggage` header extraction causes excessive allocations (remote dos amplification)

security
Apr 7, 2026

OpenTelemetry-Go has a denial-of-service vulnerability where the library parses multiple `baggage` HTTP headers (a standard for distributed tracing metadata) separately instead of treating them as one combined value. An attacker can send many baggage header lines to force the server to waste CPU and memory on repeated parsing work, even though each individual header stays within size limits, causing high latency and excessive allocations per request.

Fix: The source recommends: "avoid repeated parsing across multi-values by enforcing a global budget and/or normalizing multi-values into a single value before parsing. one mitigation approach is to treat multi-values as a single comma-joined string and cap total parsed bytes (for example 8192 bytes total)." The fix is accepted when allocations and parsing operations stay within 2x of baseline and response latency (p95) stays below 2ms.

GitHub Advisory Database
06

Anthropic limits Mythos AI rollout over fears hackers could use model for cyberattacks

safetysecurity
Apr 7, 2026

Anthropic released Claude Mythos Preview, an advanced AI model that excels at finding security vulnerabilities (weaknesses in software), but is limiting access to a select group of companies through a program called Project Glasswing to prevent attackers from misusing it. The model can identify bugs that were previously hard to detect, including a 27-year-old bug in OpenBSD (an operating system focused on security), and Anthropic is working with U.S. government agencies to manage the risks of this powerful cybersecurity capability.

Fix: Anthropic is limiting access to Claude Mythos Preview by only providing it to a select group of companies, including Apple, Google, Microsoft, Nvidia, and Amazon Web Services, along with over 40 other firms, for defensive security work. Additionally, the company stated it 'has been in ongoing discussions' with U.S. government officials including the Cybersecurity and Infrastructure Security Agency and the Center for AI Standards and Innovation about the model's cyber capabilities.

CNBC Technology
07

Anthropic Teams Up With Its Rivals to Keep AI From Hacking Everything

securitysafety
Apr 7, 2026

Anthropic announced Claude Mythos Preview, a powerful AI model capable of finding software vulnerabilities and developing exploits, alongside Project Glasswing, an industry consortium of over 40 major tech companies that will receive early access to test the model on their systems. The staggered release approach, modeled after coordinated vulnerability disclosure (the practice of giving developers time to patch bugs before public disclosure), aims to help organizations identify and fix security weaknesses before the model becomes widely available in the coming months.

Fix: Anthropic is conducting a staggered release of Mythos Preview beginning with an industry collaboration phase, giving Project Glasswing partners private access to the model so they can 'turn Mythos Preview on their own systems so they can mitigate vulnerabilities and exploit chains that the model develops in simulated attacks.' This approach is based on coordinated vulnerability disclosure practices.

Wired (Security)
08

Anthropic Unveils ‘Claude Mythos’ – A Cybersecurity Breakthrough That Could Also Supercharge Attacks

securityindustry
Apr 7, 2026

Anthropic has developed a new AI model called Claude Mythos as part of Project Glasswing, an initiative aimed at securing critical software before it can be exploited by attackers. The model is framed as both a cybersecurity advance and a potential risk, since advanced AI capabilities could theoretically be misused if they fall into the wrong hands.

SecurityWeek
09

CVE-2026-24175: NVIDIA Triton Inference Server contains a vulnerability where an attacker could cause a server crash by sending a malfor

security
Apr 7, 2026

NVIDIA Triton Inference Server has a vulnerability (CVE-2026-24175) where an attacker can crash the server by sending a malformed request header, potentially causing a denial of service (disruption of normal service). The vulnerability stems from an uncaught exception (an error that the program doesn't handle properly), which allows attackers to exploit this weakness.

NVD/CVE Database
10

CVE-2026-24174: NVIDIA Triton Inference Server contains a vulnerability where an attacker could cause a server crash by sending a malfor

security
Apr 7, 2026

NVIDIA Triton Inference Server has a vulnerability (CVE-2026-24174) where an attacker can crash the server by sending a malformed request (a request with incorrect formatting), causing a denial of service (when a system becomes unavailable to legitimate users). The vulnerability stems from incorrect conversion between numeric types (the software not properly handling different number formats).

NVD/CVE Database
Prev1...7879808182...371Next