aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Browse All

All tracked items across vulnerabilities, news, research, incidents, and regulatory updates.

to
Export CSV
3286 items

CVE-2025-46722: vLLM is an inference and serving engine for large language models (LLMs). In versions starting from 0.7.0 to before 0.9.

mediumvulnerability
security
May 29, 2025
CVE-2025-46722

vLLM (a system for running large language models) versions 0.7.0 through 0.8.x have a bug in how they create hash values (fingerprints) for images. The hashing method only looks at the raw pixel data and ignores important image properties like width and height, so two different-sized images with the same pixels would create identical hash values. This can cause the system to incorrectly reuse cached results or expose data it shouldn't.

Fix: This issue has been patched in version 0.9.0.

NVD/CVE Database

CVE-2025-46570: vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is p

lowvulnerability
security
May 29, 2025
CVE-2025-46570

vLLM, an inference and serving engine for large language models, had a vulnerability in versions before 0.9.0 where timing differences in the PageAttention mechanism (a feature that speeds up processing by reusing matching text chunks) were large enough that attackers could detect and exploit them. This type of attack is called a timing side-channel attack, where an attacker learns information by measuring how long operations take.

CVE-2025-5320: A vulnerability classified as problematic has been found in gradio-app gradio up to 5.29.1. This affects the function is

lowvulnerability
security
May 29, 2025
CVE-2025-5320

A vulnerability (CVE-2025-5320) was found in Gradio, a web framework for building AI demos, affecting versions up to 5.29.1. An attacker could manipulate the localhost_aliases parameter in the CORS Handler (the component that controls which websites can access the application) to gain elevated privileges, though executing this attack is difficult and requires remote access.

Security Spotlight: Securing Cloud & AI Products with Guardrails

infonews
securitysafety

AI Safety Newsletter #56: Google Releases Veo 3

infonews
safetyindustry

CVE-2025-5277: aws-mcp-server MCP server is vulnerable to command injection. An attacker can craft a prompt that once accessed by the M

criticalvulnerability
security
May 28, 2025
CVE-2025-5277

CVE-2025-5277 is a command injection vulnerability (a flaw where an attacker can trick a program into running unwanted commands) in aws-mcp-server, an MCP server (a software tool that helps AI systems interact with AWS cloud services). An attacker can craft a malicious prompt that, when accessed by an MCP client (a program that connects to the server), executes arbitrary commands on the host system, with a critical severity rating of 9.4.

AI ClickFix: Hijacking Computer-Use Agents Using ClickFix

infonews
securitysafety

AI Literacy Programs in Europe – Supporting Article 4 of the EU AI Act

inforegulatory
policy
May 23, 2025

This article describes a curated database of AI literacy training programs across Europe designed to help organizations and professionals comply with Article 4 of the EU AI Act (a regulation requiring organizations to build employee understanding of AI). The programs are selected based on whether they teach what AI is, its risks and benefits, and how to use it responsibly in the workplace.

CVE-2025-3893: While editing pages managed by MegaBIP a user with high privileges is prompted to give a reasoning for performing this a

infovulnerability
security
May 23, 2025
CVE-2025-3893

CVE-2025-3893 is a SQL injection vulnerability (a type of attack where malicious code is inserted into a database query) in MegaBIP that occurs when users with high privileges edit pages and provide reasoning for their actions. The user input is not sanitized (cleaned of potentially harmful code), allowing attackers to manipulate the database. This vulnerability has a CVSS severity score of 8.6 (HIGH), indicating it is serious.

Assessing the Security of 4 Popular AI Reasoning Models

infonews
securitysafety

CVE-2025-47277: vLLM, an inference and serving engine for large language models (LLMs), has an issue in versions 0.6.5 through 0.8.4 tha

criticalvulnerability
security
May 20, 2025
CVE-2025-47277

vLLM versions 0.6.5 through 0.8.4 have a vulnerability when using `PyNcclPipe` (a tool for peer-to-peer communication between multiple computers running the AI model) with the V0 engine. The issue is that a network communication interface called `TCPStore` was listening on all network connections instead of just the private network specified by the `--kv-ip` parameter, potentially exposing the system to unauthorized access.

CVE-2025-46725: Langroid is a Python framework to build large language model (LLM)-powered applications. Prior to version 0.53.15, `Lanc

criticalvulnerability
security
May 20, 2025
CVE-2025-46725

Langroid, a Python framework for building AI applications, has a vulnerability in versions before 0.53.15 where the `LanceDocChatAgent` component uses pandas eval() (a function that executes Python code stored in strings) in an unsafe way, allowing attackers to run malicious commands on the host system. The vulnerability exists in the `compute_from_docs()` function, which processes user queries without proper protection.

CVE-2025-46724: Langroid is a Python framework to build large language model (LLM)-powered applications. Prior to version 0.53.15, `Tabl

criticalvulnerability
security
May 20, 2025
CVE-2025-46724

Langroid, a Python framework for building LLM-powered applications, had a code injection vulnerability (CWE-94, a flaw where untrusted input can be executed as code) in its `TableChatAgent` component before version 0.53.15 because it used `pandas eval()` without proper safeguards. This could allow attackers to run arbitrary code if the application accepted untrusted user input.

AI Safety Newsletter #55: Trump Administration Rescinds AI Diffusion Rule, Allows Chip Sales to Gulf States

infonews
policyindustry

CVE-2025-43714: The ChatGPT system through 2025-03-30 performs inline rendering of SVG documents (instead of, for example, rendering the

mediumvulnerability
security
May 19, 2025
CVE-2025-43714

ChatGPT through March 30, 2025, renders SVG documents (scalable vector graphics, a type of image format) directly in web browsers instead of displaying them as plain text, which allows attackers to inject HTML (the code that structures web pages) and potentially trick users through phishing attacks.

CVE-2025-2099: A vulnerability in the `preprocess_string()` function of the `transformers.testing_utils` module in huggingface/transfor

highvulnerability
security
May 19, 2025
CVE-2025-2099

A vulnerability in the `preprocess_string()` function of the huggingface/transformers library (version v4.48.3) allows a ReDoS attack (regular expression denial of service, where a poorly written pattern causes the computer to do exponential amounts of work). An attacker can send specially crafted input with many newline characters that makes the function use excessive CPU, potentially crashing the application.

CVE-2025-1975: A vulnerability in the Ollama server version 0.5.11 allows a malicious user to cause a Denial of Service (DoS) attack by

mediumvulnerability
security
May 16, 2025
CVE-2025-1975

CVE-2025-1975 is a vulnerability in Ollama server version 0.5.11 that allows an attacker to crash the server through a Denial of Service attack by sending specially crafted requests to the /api/pull endpoint (the function that downloads AI models). The vulnerability stems from improper validation of array index access (CWE-129, which means the program doesn't properly check if it's trying to access memory locations that don't exist), which happens when a malicious user customizes manifest content and spoofs a service.

CVE-2025-4701: A vulnerability, which was classified as problematic, has been found in VITA-MLLM Freeze-Omni up to 20250421. This issue

mediumvulnerability
security
May 15, 2025
CVE-2025-4701

CVE-2025-4701 is a vulnerability in VITA-MLLM Freeze-Omni (versions up to 20250421) where improper input validation in the torch.load function of models/utils.py allows deserialization (converting data back into executable code) of untrusted data through a manipulated file path argument. This vulnerability has a CVSS score (a 0-10 rating of how severe a vulnerability is) of 4.8 (medium severity) and can be exploited locally by users with basic privileges.

Specialized Models Beat Single LLMs for AI Security

infonews
securityresearch

AI Safety Newsletter #54: OpenAI Updates Restructure Plan

inforegulatory
policysafety
Previous97 / 165Next

Fix: Update vLLM to version 0.9.0 or later. The issue has been patched in version 0.9.0.

NVD/CVE Database
NVD/CVE Database
May 28, 2025

This article collection discusses security challenges in AI and cloud systems, particularly focusing on agentic AI (AI systems that can take autonomous actions). Key risks include jailbreaks (tricking AI systems into ignoring safety rules), prompt injection (hidden malicious instructions in AI inputs), and tool misuse by autonomous agents, which require contextual red teaming (security testing designed for specific use cases) rather than generic testing to identify real vulnerabilities.

Protect AI Blog
May 28, 2025

Google released Veo 3, a frontier video generation model (an advanced AI system at the cutting edge of technology) that generates both video and audio with high quality and appears to be a marked improvement over existing systems. The model performs well on human preference benchmarks and may represent the point where video generation becomes genuinely useful rather than just a novelty. Additionally, Google announced several other AI improvements at its I/O 2025 conference, including Gemini 2.5 Pro and enhanced reasoning capabilities, while Anthropic released Claude Opus 4 and Claude Sonnet 4 with frontier-level performance.

CAIS AI Safety Newsletter
NVD/CVE Database
May 24, 2025

ClickFix is a social engineering technique (a method that tricks people rather than exploiting technical vulnerabilities) that adversaries are adapting to attack computer-use agents (AI systems that can control computers by clicking and typing). The attack works by deceiving users into believing something is broken or needs verification, then tricking them into clicking buttons or running commands that compromise their system.

Embrace The Red
EU AI Act Updates

Fix: Version 5.20 of MegaBIP fixes this issue.

NVD/CVE Database
May 21, 2025

This content discusses security challenges in agentic AI (autonomous AI systems that can take actions independently), emphasizing that traditional jailbreak testing (attempts to trick AI into breaking its rules) misses real operational risks like tool misuse and data theft. The material suggests that contextual red teaming (security testing that simulates realistic attack scenarios in specific business environments) is needed to properly assess vulnerabilities in autonomous AI systems.

Protect AI Blog

Fix: Update to vLLM version 0.8.5 or later. According to the source: "As of version 0.8.5, vLLM limits the `TCPStore` socket to the private interface as configured."

NVD/CVE Database

Fix: Upgrade to Langroid version 0.53.15 or later. The fix involves input sanitization (cleaning and filtering user input) to the affected function by default to block common attack vectors, along with added warnings in the project documentation about the risky behavior.

NVD/CVE Database

Fix: Upgrade to Langroid version 0.53.15 or later. According to the source, 'Langroid 0.53.15 sanitizes input to `TableChatAgent` by default to tackle the most common attack vectors, and added several warnings about the risky behavior in the project documentation.'

NVD/CVE Database
May 20, 2025

The Trump Administration cancelled the Biden-era AI Diffusion Rule, which had regulated exports of AI chips and AI models (software trained to perform tasks) to different countries. At the same time, the administration approved major sales of advanced AI chips to the UAE and Saudi Arabia, with deals including up to 500,000 chips per year to the UAE and 18,000 advanced chips to Saudi Arabia.

CAIS AI Safety Newsletter
NVD/CVE Database
NVD/CVE Database
NVD/CVE Database
NVD/CVE Database
May 13, 2025

The article argues that using multiple specialized AI security models (each designed to detect specific threats like prompt injection, toxicity, or PII detection) is more effective than using a single large model for all security tasks. Specialized models offer advantages including faster response times to new threats, easier management, better performance, lower costs, and greater resilience because if one model fails, the others can still provide protection.

Protect AI Blog
May 13, 2025

OpenAI announced a restructured plan in May 2025 that aims to preserve nonprofit control over the company's for-profit operations, replacing a December 2024 proposal that had faced criticism. The new plan would convert OpenAI Global LLC into a public-benefit corporation (PBC, a corporate structure designed to balance profit with charitable purpose) where the nonprofit would retain shareholder status and board appointment power, though critics argue this may not preserve the governance safeguards that existed in the original structure.

CAIS AI Safety Newsletter