aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 57/371
VIEW ALL
01

DiffMI: Breaking Face Recognition Privacy via Diffusion-Driven Training-Free Model Inversion

securityresearch
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 16, 2026

Researchers developed DiffMI, a new attack that can recover people's facial identities from face recognition systems by reversing the embeddings (compressed numerical representations of faces). Unlike previous attacks, DiffMI doesn't require expensive training on specific targets and can work against unseen faces and new recognition models, achieving success rates between 84-93% against systems designed to resist such attacks.

IEEE Xplore (Security & AI Journals)
02

Authentication With Passports for Deep RF Sensing Model Protection

securityresearch
Apr 16, 2026

```json { "summary": "This paper introduces AuthRF, a security system that protects RF sensing models (AI systems that interpret radio frequency signals from WiFi or radar) by using user-specific digital "passports" embedded in the signal processing pipeline. Valid passports allow the model to work correctly, while invalid or fake ones distort the signal and degrade performance, preventing unauthorized use. The approach is designed to be proactive and work during runtime, addressing limitation

IEEE Xplore (Security & AI Journals)
03

Defending Against Patch-Based and Texture-Based Adversarial Attacks With Spectral Decomposition

researchsecurity
Apr 16, 2026

Adversarial examples (inputs crafted to fool AI systems) are a serious security risk for deep neural networks (AI systems with many layers), especially in physical-world attacks like fooling object detection in surveillance cameras. This research proposes Adversarial Spectrum Defense (ASD), a defense method that uses spectral decomposition (breaking down data into different frequency components) via Discrete Wavelet Transform (a mathematical technique to analyze patterns at multiple scales) to detect and defend against patch-based and texture-based adversarial attacks, and shows it achieves better protection when combined with Adversarial Training (training the AI on attack examples to make it more robust).

Fix: The source proposes Adversarial Spectrum Defense (ASD), which 'leverages spectral decomposition via Discrete Wavelet Transform (DWT) to analyze adversarial patterns across multiple frequency scales' and 'by integrating this spectral analysis with the off-the-shelf Adversarial Training (AT) model, ASD provides a comprehensive defense strategy against both patch-based and texture-based adversarial attacks.' The paper reports that 'ASD+AT achieved state-of-the-art (SOTA) performance against various attacks, outperforming the APs of previous defense methods by 21.73%'.

IEEE Xplore (Security & AI Journals)
04

Canva’s AI 2.0 update goes all in on prompt-powered design tools

industry
Apr 16, 2026

Canva released AI 2.0, a major update that adds prompt-based editing capabilities, allowing users to describe what they want and have the AI assistant create or modify designs accordingly. The update includes a new orchestration layer (a system that coordinates multiple AI models) that lets users access Canva's full toolkit through a single conversational interface instead of separate tools.

The Verge (AI)
05

Making AI operational in constrained public sector environments

industrypolicy
Apr 16, 2026

Public sector organizations face unique challenges deploying AI due to strict data security requirements, limited internet connectivity, and lack of GPU (graphics processing units, specialized computer hardware for running complex AI models) infrastructure. Small language models (SLMs, specialized AI models using billions rather than hundreds of billions of parameters) offer a practical solution because they can run locally on government systems, use less computing power than large language models (LLMs, the biggest AI systems like ChatGPT), and keep sensitive data under government control.

Fix: Use small language models (SLMs) instead of large language models (LLMs) in public sector environments. SLMs can be housed locally for greater security and control, are less computationally demanding, and allow sensitive information to be used effectively while avoiding operational complexity. Implement methods such as smart retrieval, vector search, and verifiable source grounding to build AI systems that meet public sector needs. Store data securely outside the model and access it only when queried, using carefully engineered prompts to retrieve only the most relevant information.

MIT Technology Review
06

Treating enterprise AI as an operating layer

industry
Apr 16, 2026

This article discusses how enterprise organizations can gain competitive advantage in AI by treating it as an operating layer (the combination of software, data capture, feedback loops, and governance that connects AI models to actual business operations) rather than just using AI as an on-demand service. The key difference is that an operating layer allows intelligence to accumulate and improve over time through organizational feedback, whereas calling an API (application programming interface, a way to request services from software) for each task treats AI as stateless and interchangeable. Incumbent organizations have a structural advantage because they already possess proprietary operational data, domain expert workers, and accumulated knowledge that startups must build from scratch.

MIT Technology Review
07

Why having “humans in the loop” in an AI war is an illusion

safetypolicy
Apr 16, 2026

AI systems are now actively controlling weapons in warfare, but the assumption that human oversight provides adequate safeguards is flawed because humans cannot understand how AI systems make decisions (they are "black boxes" where even creators cannot fully interpret their reasoning). The real danger is that humans may approve AI actions without knowing the system's hidden reasoning, creating an "intention gap" between what operators think the AI will do and what it actually does.

Fix: The science of AI must comprise both building highly capable AI technology and understanding how this technology works. Huge advances have been made in developing and building more capable models, but the source text cuts off before completing this section on solutions.

MIT Technology Review
08

[Webinar] Find and Eliminate Orphaned Non-Human Identities in Your Environment

securitypolicy
Apr 16, 2026

In 2024, 68% of cloud breaches were caused by compromised service accounts and forgotten API keys, which are unmanaged non-human identities (automated credentials like tokens and API keys) that attackers can exploit. Organizations have 40 to 50 automated credentials per employee, most remaining active and unmonitored after projects end or employees leave, creating security risks that traditional identity management systems cannot address. The webinar promises to teach how to discover, right-size permissions for, and automatically revoke these 'ghost identities' using a discovery scan, permission framework, lifecycle policy, and cleanup checklist.

Fix: The source describes a framework that includes: (1) running a full discovery scan of every non-human identity in your environment, (2) implementing a framework for right-sizing permissions across service accounts and AI integrations, (3) setting up an automated lifecycle policy so dead credentials get revoked before attackers find them, and (4) using a ready-to-use Identity Cleanup Checklist provided during the webinar session.

The Hacker News
09

Behind the Mythos hype, Glasswing has just one confirmed CVE

securityindustry
Apr 16, 2026

Anthropic's Mythos AI model, released through Project Glasswing (a controlled access program for vetted organizations), has generated significant hype for its offensive security capabilities, but VulnCheck's analysis found only one CVE (common vulnerabilities and exposures, a list of known security flaws) explicitly attributed to the project itself. Despite the limited number of publicly confirmed discoveries, security experts view Mythos as significant because it achieved a 72% exploit success rate (the ability to successfully turn vulnerabilities into working attacks), suggesting that advanced AI exploit development is no longer a specialized skill and this capability will likely spread to other AI models and organizations without the same safety protections.

CSO Online
10

Insurance carriers quietly back away from covering AI outputs

policyindustry
Apr 16, 2026

Major insurance companies are withdrawing or limiting coverage for AI-related mistakes and damages because they cannot understand how AI systems reach their conclusions, a problem called lack of explainability (the inability to see the reasoning behind an AI's output). Some insurers are declining to cover AI errors entirely, while others are significantly raising prices, creating a situation where companies using AI may struggle to find affordable insurance for AI-related risks.

CSO Online
Prev1...5556575859...371Next