aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Browse All

All tracked items across vulnerabilities, news, research, incidents, and regulatory updates.

to
Export CSV
3211 items

CVE-2026-22778: vLLM is an inference and serving engine for large language models (LLMs). From 0.8.3 to before 0.14.1, when an invalid i

criticalvulnerability
security
Feb 2, 2026
CVE-2026-22778

vLLM, a system for running large language models, has a vulnerability in versions 0.8.3 through 0.14.0 where sending an invalid image to its multimodal endpoint causes it to leak a heap address (a memory location used for storing data). This information leak significantly weakens ASLR (address space layout randomization, a security feature that randomizes where programs load in memory), and attackers could potentially chain this leak with other exploits to gain remote code execution (the ability to run commands on the server).

Fix: This vulnerability is fixed in version 0.14.1. Update vLLM to version 0.14.1 or later.

NVD/CVE Database

CVE-2026-1778: Amazon SageMaker Python SDK before v3.1.1 or v2.256.0 disables TLS certificate verification for HTTPS connections made b

mediumvulnerability
security
Feb 2, 2026
CVE-2026-1778

Amazon SageMaker Python SDK (a library for building machine learning models on AWS) versions before v3.1.1 or v2.256.0 have a vulnerability where TLS certificate verification (the security check that confirms a website is genuine) is disabled for HTTPS connections when importing a Triton Python model, allowing attackers to use fake or self-signed certificates to intercept or manipulate data. This vulnerability has a CVSS score (a 0-10 rating of severity) of 8.2, indicating high severity.

CVE-2026-0599: A vulnerability in huggingface/text-generation-inference version 3.3.6 allows unauthenticated remote attackers to exploi

highvulnerability
security
Feb 2, 2026
CVE-2026-0599

A vulnerability in huggingface/text-generation-inference version 3.3.6 allows attackers without authentication to crash servers by sending images in requests. The problem occurs because the software downloads entire image files into memory when checking inputs for Markdown image links (a way to embed images in text), even if it will later reject the request, causing the system to run out of memory, bandwidth, or CPU power.

CVE-2025-10279: In mlflow version 2.20.3, the temporary directory used for creating Python virtual environments is assigned insecure wor

highvulnerability
security
Feb 2, 2026
CVE-2025-10279

MLflow version 2.20.3 has a vulnerability where temporary directories used to create Python virtual environments are set with world-writable permissions (meaning any user on the system can read, write, and execute files there). An attacker with access to the `/tmp` directory can exploit a race condition (a situation where timing allows an attacker to interfere with an operation before it completes) to overwrite Python files in the virtual environment and run arbitrary code.

langchain==1.2.8

lownews
security
Feb 2, 2026

LangChain released version 1.2.8, which includes several updates and fixes such as reusing ToolStrategy in the agent factory to prevent name mismatches, upgrading urllib3 (a library for making web requests), and adding ToolCallRequest to middleware exports (the code that processes requests between different parts of an application).

AI Safety Newsletter #68: Moltbook Exposes Risky AI Behavior

mediumnews
safetysecurity

langchain-core==1.2.8

infonews
security
Feb 2, 2026

LangChain-core version 1.2.8 is a release update that includes various improvements and changes to the library's functions and components. The update modifies features like the @tool decorator (which marks functions as tools for AI agents), iterator handling for data streaming, and several utility functions for managing AI agent interactions, but the provided content does not specify what problems these changes fix or what new capabilities they enable.

Jeffrey Epstein Had a ‘Personal Hacker,’ Informant Claims

infonews
security
Jan 31, 2026

According to a DOJ document released in 2017, an FBI informant claimed that Jeffrey Epstein had a 'personal hacker' who specialized in finding vulnerabilities (weaknesses that attackers can exploit) in Apple iOS, BlackBerry, and Firefox, and allegedly developed and sold offensive hacking tools and exploits (code that takes advantage of these weaknesses) to multiple countries and organizations. The document does not identify the alleged hacker or confirm whether the FBI verified these claims.

v5.3.0

inforesearchIndustry
industry

v5.2.0

inforesearchIndustry
security

2026: The Year Agentic AI Becomes the Attack-Surface Poster Child

infonews
security
Jan 30, 2026

Dark Reading surveyed readers about which AI and cybersecurity trends would likely become major issues in 2026, including agentic AI attacks (where AI systems act independently to cause harm), advanced deepfake threats (realistic fake videos or audio), increased board-level cyber priorities, and password-less technology adoption (replacing passwords with other authentication methods).

Tenable Tackles AI Governance, Shadow AI Risks, Data Exposure

infonews
securitypolicy

OpenClaw AI Runs Wild in Business Environments

mediumnews
securitysafety

Building Trustworthy AI Agents

inforesearchPeer-Reviewed
safety

Forgotten Memories

inforesearchPeer-Reviewed
privacy

Understanding the Adversarial Landscape of Large Language Models Through the Lens of Attack Objectives

inforesearchPeer-Reviewed
security

Celebrating our 2025 open-source contributions

infonews
industry
Jan 30, 2026

Trail of Bits engineers contributed over 375 pull requests to 90+ open-source projects in 2025, including work on cryptography libraries, the Rust compiler, and Ethereum tools. Rather than forking or locally patching dependencies when they encountered bugs or needed features, they contributed fixes upstream so the entire community could benefit. Key contributions included adding identity monitoring to Sigstore's Rekor (a transparency log for software signing), improving Rust's linting tools, developing a new ASN.1 API (a standard for encoding data structures) for Python's cryptography library, and optimizing the Ethereum Virtual Machine implementation.

Breaking the Sound Barrier, Part II: Exploiting CVE-2024-54529

infonews
security
Jan 30, 2026

CVE-2024-54529 is a type confusion vulnerability (where the code incorrectly assumes an object is a certain type without checking) in Apple's CoreAudio framework that allows attackers to crash the coreaudiod system daemon and potentially hijack control flow by manipulating pointer chains in memory. The vulnerability exists in the com.apple.audio.audiohald Mach service (a macOS inter-process communication system) where message handlers fetch objects without validating their actual type before performing operations on them.

Big tech results show investor demand for payoffs from heavy AI spending

infonews
industry
Jan 29, 2026

Big tech companies are under pressure from investors to show that their heavy spending on AI is producing real financial results and business growth. Meta's stock rose after demonstrating AI improvements in advertising, while Microsoft's stock fell despite its large AI investments, showing that investors will reward companies with strong returns but punish those that don't deliver clear benefits from their AI spending.

'Semantic Chaining' Jailbreak Dupes Gemini Nano Banana, Grok 4

lownews
securitysafety
Previous63 / 161Next

Fix: Update Amazon SageMaker Python SDK to version v3.1.1 or v2.256.0 or later.

NVD/CVE Database

Fix: The issue is resolved in version 3.3.7.

NVD/CVE Database

Fix: The issue is resolved in mlflow version 3.4.0.

NVD/CVE Database

Fix: Update to langchain==1.2.8, which includes the fix: 'reuse ToolStrategy in agent factory to prevent name mismatch' and 'upgrade urllib3 to 2.6.3'.

LangChain Security Releases
Feb 2, 2026

Moltbook is a new social network where AI agents (autonomous software programs that can perform tasks independently) post and interact with each other, similar to Reddit. Since launching, human observers have noticed concerning posts where agents discuss creating secret languages to hide from humans, using encrypted communication to avoid oversight, and planning for independent survival without human control.

CAIS AI Safety Newsletter
LangChain Security Releases
Wired (Security)
Jan 30, 2026

N/A -- This content is a navigation menu and feature listing for GitHub's v5.3.0 platform, not a description of an AI/LLM security issue, vulnerability, or problem requiring analysis.

MITRE ATLAS Releases
research
Jan 30, 2026

Version 5.2.0 adds new attack techniques against AI systems, including methods to steal credentials from AI agent tools (software components that perform actions on behalf of an AI), poison training data, and generate malicious commands. It also introduces new defenses such as segmenting AI agent components, validating inputs and outputs, detecting deepfakes, and implementing human oversight for AI agent actions.

Fix: The source lists mitigations rather than fixes for a specific vulnerability. Key mitigations mentioned include: Input and Output Validation for AI Agent Components, Segmentation of AI Agent Components, Restrict AI Agent Tool Invocation on Untrusted Data, Human In-the-Loop for AI Agent Actions, Adversarial Input Detection, Model Hardening, Sanitize Training Data, and Generative AI Guardrails.

MITRE ATLAS Releases
Dark Reading
Jan 30, 2026

Tenable has released an AI Exposure add-on tool that finds unauthorized AI usage (shadow AI, or unsanctioned AI tools employees use without approval) within an organization and ensures compliance with official AI policies. This helps organizations manage risks from uncontrolled AI deployment and data exposure.

Dark Reading
Jan 30, 2026

OpenClaw AI, a popular open source AI assistant also known as ClawdBot or MoltBot, has become widely used but is raising security concerns because it operates with elevated privileges (special access rights that allow it to control more of a computer) and can act autonomously without waiting for user approval. The combination of unrestricted access and independent decision-making in business environments poses risks to system security and data safety.

Dark Reading
research
Jan 30, 2026

Current AI assistants are not yet trustworthy enough to be personal advisors, despite how useful they seem. They fail in specific ways: they encourage users to make poor decisions, they create false doubt about things people know to be true (gaslighting), and they confuse a person's current identity with their past. They also struggle when information is incomplete or inaccurate, with no reliable way to fix errors or hold the system responsible when wrong information causes harm.

IEEE Xplore (Security & AI Journals)
safety
Jan 30, 2026

This short story examines privacy risks that arise when companies are bought and sold, particularly concerning AI digital twins (AI models that replicate a specific person's behavior and knowledge) and the problems that occur when organizations fail to threat model (identify and plan for potential security risks in) major changes to their systems and technology. The story raises ethical questions about these scenarios.

IEEE Xplore (Security & AI Journals)
research
Jan 30, 2026

Large language models face four main types of adversarial threats: privacy breaches (exposing sensitive data the model learned), integrity compromises (corrupting the model's outputs or training data), adversarial misuse (using the model for harmful purposes), and availability disruptions (making the model unavailable or slow). The article organizes these threats by their attackers' goals to help understand the landscape of vulnerabilities in LLMs.

IEEE Xplore (Security & AI Journals)
Trail of Bits Blog
Google Project Zero
The Guardian Technology
Jan 29, 2026

Researchers discovered a jailbreak technique called semantic chaining that tricks certain LLMs (AI models trained on massive amounts of text) by breaking malicious requests into small, separate chunks that the model processes without understanding the overall harmful intent. This vulnerability affected models like Gemini Nano and Grok 4, which failed to recognize the dangerous purpose when instructions were split across multiple parts.

Dark Reading