aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSaturday, May 16, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 81/371
VIEW ALL
01

CVE-2026-24173: NVIDIA Triton Inference Server contains a vulnerability where an attacker could cause a server crash by sending a malfor

security
Apr 7, 2026

NVIDIA Triton Inference Server has a vulnerability (CVE-2026-24173) where an attacker can send a malformed request to crash the server, causing a denial of service (when a service becomes unavailable due to an attack). The vulnerability is related to integer overflow or wraparound (when a number exceeds the maximum value a system can store, causing unexpected behavior).

NVD/CVE Database
02

CVE-2026-24147: NVIDIA Triton Inference Server contains a vulnerability in triton server where an attacker may cause an information disc

security
Apr 7, 2026

CVE-2026-24147 is a vulnerability in NVIDIA Triton Inference Server (a tool that runs AI models) where an attacker can upload a malicious model configuration file to cause information disclosure (exposing sensitive data) or denial of service (making the system unavailable). The vulnerability stems from improper path traversal (a flaw that lets attackers access files outside intended directories) validation when handling uploaded files.

NVD/CVE Database
03

CVE-2026-24146: NVIDIA Triton Inference Server contains a vulnerability where insufficient input validation and a large number of output

security
Apr 7, 2026

NVIDIA Triton Inference Server has a vulnerability where it doesn't properly check user inputs and can crash when given a large number of outputs, potentially causing a denial of service (making the server unavailable to users). The vulnerability stems from excessive memory allocation triggered by malformed input.

NVD/CVE Database
04

XFaceMark: Explainable deep fake watermarking using YOLO, and random MRFO

researchsecurity
Apr 7, 2026

This paper presents XFaceMark, a method that uses YOLO (an object detection system that identifies items in images) and random MRFO (a nature-inspired optimization algorithm) to add watermarks to deepfakes (AI-generated fake videos or images) in a way that can be explained and understood. The approach aims to make deepfakes traceable while allowing researchers to understand how the watermarking process works.

Elsevier Security Journals
05

SBOMs into Agentic AIBOMs: Schema Extensions, Agentic Orchestration and Reproducibility Evaluation

research
Apr 7, 2026

This academic paper discusses extending SBOMs (software bill of materials, which are detailed lists of all components and dependencies in software) to create AIBOMs that can describe agentic AI systems (AI systems that can take independent actions and make decisions). The paper proposes schema extensions, methods for coordinating multiple AI agents, and ways to evaluate whether AI systems produce consistent and reproducible results.

ACM Digital Library (TOPS, DTRAP, CSUR)
06

Anthropic is launching a new AI model for cybersecurity

securityindustry
Apr 7, 2026

Anthropic is launching a new AI model called Claude Mythos Preview as part of Project Glasswing, a cybersecurity partnership with major tech companies like Nvidia, Google, and Microsoft. The model is designed to help large organizations and governments automatically detect vulnerabilities (security weaknesses) in their systems with minimal human involvement. Anthropic is limiting access to launch partners only and not releasing it publicly due to security concerns.

The Verge (AI)
07

Anthropic debuts preview of powerful new AI model Mythos in new cybersecurity initiative

securityindustry
Apr 7, 2026

Anthropic released a preview of Mythos, a powerful new AI model, as part of Project Glasswing, a cybersecurity initiative involving over 40 partner organizations like Amazon, Microsoft, and Apple. The model, which was not specifically trained for cybersecurity but has strong coding and reasoning abilities, has reportedly identified thousands of zero-day vulnerabilities (security flaws unknown to the public and software vendors) in software systems during initial testing. The preview is limited to partner organizations for defensive security work and will not be made generally available to the public.

TechCrunch (Security)
08

Cybersecurity in the Age of Instant Software

securitysafety
Apr 7, 2026

AI is making software creation faster and easier, leading to a future where temporary applications (instant software) might be created and deleted on demand, but this also means AI tools are getting better at both finding and exploiting vulnerabilities (weaknesses in code that attackers can use). While defenders can use the same AI capabilities to patch vulnerabilities and fix security problems, today's AI-generated software tends to contain many security flaws because AI doesn't yet write secure code well.

Schneier on Security
09

Max severity Flowise RCE vulnerability now exploited in attacks

security
Apr 7, 2026

Hackers are actively exploiting CVE-2025-59528, a critical vulnerability in Flowise (an open-source platform for building AI agents and custom LLM applications) that allows arbitrary JavaScript code injection without validation through the CustomMCP node. The flaw was publicly disclosed in September, affects thousands of exposed instances online, and enables attackers to execute commands and access files on vulnerable systems.

Fix: Upgrade to Flowise version 3.1.1 or at least version 3.0.6 as soon as possible. Additionally, consider removing Flowise instances from the public internet if external access is not required.

BleepingComputer
10

The New Rules of Engagement: Matching Agentic Attack Speed

securitypolicy
Apr 7, 2026

Nation-states are using AI agents (autonomous AI systems that can perform tasks without human intervention) to launch cyberattacks at speeds that traditional security responses cannot match. The article argues that cybersecurity defenses cannot rely on small, gradual improvements but must instead undergo fundamental architectural changes to address this new threat level.

SecurityWeek
Prev1...7980818283...371Next