aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 65/371
VIEW ALL
01

OpenAI touts Amazon alliance in memo, says Microsoft has 'limited our ability' to reach clients

industry
Apr 13, 2026

OpenAI's new revenue chief sent an internal memo highlighting a partnership with Amazon (a cloud computing company competing with Microsoft) as crucial for reaching enterprise customers, while acknowledging that its existing deal with Microsoft has constrained its ability to serve clients who prefer Amazon's AI platform called Bedrock (a service that provides access to major AI models). The memo reflects OpenAI's struggle to compete with rival Anthropic's Claude model in the enterprise market, where companies are investing heavily in AI.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

CNBC Technology
02

CVE-2026-1462: A vulnerability in the `TFSMLayer` class of the `keras` package, version 3.13.0, allows attacker-controlled TensorFlow S

security
Apr 13, 2026

A vulnerability in keras version 3.13.0 allows attackers to run their own code when a model is loaded, even when `safe_mode=True` (a setting meant to prevent unsafe operations). The problem occurs because the `TFSMLayer` class loads external TensorFlow SavedModels (pre-trained model files) without checking if they're safe, and doesn't properly validate file paths or configuration data.

NVD/CVE Database
03

Transferable Adversarial Attack on Referring Video Object Segmentation

securityresearch
Apr 13, 2026

Referring video object segmentation (RVOS, the task of identifying and outlining objects in videos based on text descriptions) is used in safety-critical applications like autonomous driving, but the deep neural networks that power these systems are vulnerable to adversarial perturbations (tiny, intentional changes to input data designed to fool AI models). This research demonstrates for the first time that RVOS models can be reliably attacked using a method called xM-ICM, which corrupts both visual and text information to mislead the models, and shows this attack works even when attackers have limited information about the system.

IEEE Xplore (Security & AI Journals)
04

HKT-SmartAudit: Distilling Lightweight Models for Smart Contract Auditing

research
Apr 13, 2026

HKT-SmartAudit is a framework that creates smaller, faster AI models specifically trained to find bugs in smart contracts (self-executing code on blockchain networks). The framework uses knowledge distillation (a technique where a large, accurate AI model teaches a smaller model by sharing what it has learned), allowing these lightweight models to detect vulnerabilities effectively while using far less computing power than larger models.

IEEE Xplore (Security & AI Journals)
05

FALCON-Net: Feature Aggregation of Local Patterns for AI-Generated Image Detection

research
Apr 13, 2026

FALCON-Net is a detection system designed to identify AI-generated images by analyzing their technical flaws. The system works by examining two key weaknesses in generated images: the lack of device-specific sensor noise (natural imperfections that real cameras add) and unnatural pixel intensity variations that result from oversimplified generation processes. FALCON-Net combines two analysis modules (one for noise patterns and one for local pixel variations) to reliably distinguish AI-generated images from real ones, even when tested on image generation models it wasn't trained on.

IEEE Xplore (Security & AI Journals)
06

FedNSA: Boosting Secure Aggregation by Assembling Differentially Private Noise Shares

securityprivacy
Apr 13, 2026

Federated learning (FL, where multiple devices train AI models together without sharing raw data) faces privacy risks because adversaries can extract sensitive information from model updates. FedNSA is a new protocol that combines differential privacy (adding mathematical noise to hide individual data patterns), encryption, and multi-party computation (MPC, a technique where multiple parties jointly compute results without revealing their individual inputs) to protect model updates while reducing the communication and computational burden that makes secure aggregation impractical on resource-constrained devices like smartphones.

IEEE Xplore (Security & AI Journals)
07

LitCVit: A Lightweight Self-Supervised Contrastive Vision Transformer for Encrypted Malicious Traffic Detection

research
Apr 13, 2026

LitCVit is a lightweight AI model designed to detect malicious encrypted network traffic (data sent over secure connections) without needing to decrypt it or manually extract features. The model uses self-supervised learning (training where the AI learns patterns from unlabeled data) and vision transformers (a type of neural network architecture) to analyze patterns across multiple data packets and flows (sequences of related network communications) while running much faster than existing approaches, achieving 98% accuracy on test datasets.

IEEE Xplore (Security & AI Journals)
08

HENet: A Heterogeneous Encoding Network for General and Robust Adversarial Example Generation

securityresearch
Apr 13, 2026

This paper presents HENet, a new method for creating adversarial examples (inputs with small, intentional changes designed to fool AI models) that work against different types of neural networks like CNNs (convolutional neural networks, commonly used for image tasks) and Transformers (a newer architecture). The method improves two key challenges: making attacks work across different model architectures and making adversarial examples survive image compression like JPEG, which currently weakens their effectiveness.

IEEE Xplore (Security & AI Journals)
09

Exposing the Ghost in the Transformer: Abnormal Detection for Large Language Models via Hidden State Forensics

securityresearch
Apr 13, 2026

Large language models (LLMs, which are AI systems trained on vast amounts of text) are vulnerable to serious attacks like hallucinations (making up false information), jailbreaks (tricking the AI into ignoring its safety rules), and backdoors (hidden malicious instructions inserted during training). This research proposes a detection method using hidden state forensics (analyzing the internal numerical patterns that flow through the model's layers) to identify abnormal or malicious behavior in real-time, achieving over 95% accuracy with minimal computational cost.

IEEE Xplore (Security & AI Journals)
10

DFREC: DeepFake Identity Recovery Based on Identity-Aware Masked Autoencoder

researchsafety
Apr 13, 2026

DFREC is a new method for identifying the original faces used to create deepfakes (fake videos where one person's face is swapped onto another's body). Unlike existing deepfake detection tools that only identify whether an image is fake, DFREC recovers both the source face (the one being used) and target face (the one being impersonated) from a deepfake image, which helps investigators trace who was involved in creating the fake and reduces risks from deepfake attacks. The system uses three components: one to separate source and target face information, one to reconstruct the source face, and one to reconstruct the target face using a Masked Autoencoder (a type of neural network that learns patterns by hiding parts of input data).

IEEE Xplore (Security & AI Journals)
Prev1...6364656667...371Next