aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
67
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 17/371
VIEW ALL
01

If AI's So Smart, Why Does It Keep Deleting Production Databases?

securitysafety
Critical This Week5 issues
critical

CVE-2026-42271: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.74.2 to before vers

CVE-2026-42271NVD/CVE DatabaseMay 8, 2026
May 8, 2026
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

May 1, 2026

The article argues that AI systems aren't inherently flawed when they cause problems like deleting production databases (the live systems storing important data). Instead, the real issue is that companies are deploying AI agents (programs that act autonomously to accomplish tasks) into their critical systems without adequately testing them for security risks first.

Dark Reading
02

Pentagon strikes classified AI deals with OpenAI, Google, and Nvidia — but not Anthropic

policyindustry
May 1, 2026

The Pentagon has signed agreements with OpenAI, Google, Microsoft, Amazon, Nvidia, xAI, and Reflection to use their AI tools in classified military settings, but excluded Anthropic after labeling it a supply-chain risk (a potential weak point in security). This expands earlier deals that allowed some companies like OpenAI and xAI to provide AI systems for authorized military use.

The Verge (AI)
03

Elon Musk had a bad week in court

policy
May 1, 2026

This article discusses a legal case where Elon Musk is suing OpenAI (an AI company), claiming they stole a nonprofit organization and that he was the main force behind their success. During his testimony in court, Musk had a difficult time, arguing with lawyers and changing his statements, with indications suggesting he is unlikely to win the case.

The Verge (AI)
04

Pentagon tech chief says Anthropic is still blacklisted, but Mythos is a separate issue

policysecurity
May 1, 2026

The Pentagon's chief technology officer stated that Anthropic remains classified as a supply chain risk (a designation meaning the company's technology threatens U.S. national security), but Anthropic's Mythos AI model, which has advanced capabilities for finding and fixing cyber vulnerabilities, is being treated as a separate urgent national security issue requiring the Department of Defense to strengthen its networks. The DOD has blacklisted Anthropic from working with defense contractors, though the agency is reportedly using Mythos internally and is open to negotiations about safeguards (called guardrails, or restrictions on how the AI can be used) if Anthropic agrees to terms similar to those negotiated with other AI companies.

CNBC Technology
05

The Download: a new Christian phone network, and debugging LLMs

industrysafety
May 1, 2026

Goodfire, a San Francisco startup, released Silico, a tool that uses mechanistic interpretability (a technique for understanding how AI models work by mapping their internal neurons and connections) to let researchers see inside AI models and adjust their parameters during training. The tool aims to give developers more control over AI behavior by exposing internal 'knobs and dials' so they can reduce unwanted outputs, making AI development more like traditional software engineering rather than trial-and-error.

Fix: The source describes Silico as the solution itself—it uses mechanistic interpretability to map neurons and pathways inside a model and lets developers tweak them to reduce unwanted behaviors or steer outputs. No additional mitigation steps or fixes beyond using this tool are mentioned in the text.

MIT Technology Review
06

Careful Adoption of Agentic AI Services

policysafety
May 1, 2026

CISA and international cybersecurity partners released guidance for organizations adopting agentic AI (AI systems that can take actions autonomously on behalf of users). The guidance identifies security challenges with these systems and provides steps for safely designing, deploying, and operating them while connecting AI risk management to existing cybersecurity practices.

CISA Cybersecurity Advisories
07

Microsoft wants lawyers to trust its new AI agent in Word documents

industry
May 1, 2026

Microsoft is launching a new AI agent within Word that is designed specifically for legal teams to help with tasks like reviewing contracts and managing document edits. Unlike general AI models, the Legal Agent follows structured workflows (predetermined sets of steps) based on actual legal practices, handling specific repeatable tasks like reviewing contract clauses against a predefined playbook (a set of rules or guidelines).

The Verge (AI)
08

Cisco Releases Open Source Tool for AI Model Provenance 

securityindustry
May 1, 2026

Organizations often use AI models from online repositories like HuggingFace without tracking their changes, verifications, or vulnerabilities, which can lead to security risks if models are poisoned (containing hidden malicious code) or contain training biases. Cisco released the Model Provenance Kit, an open source Python-based tool that creates a unique 'fingerprint' for each model using metadata and other signals, allowing organizations to compare models and trace their origins to address these tracking and accountability problems.

Fix: The Model Provenance Kit from Cisco is available on GitHub. The tool has two modes: 'compare' mode enables users to compare two models to identify shared lineage, and 'scan' mode attempts to find the closest lineage for a given model by comparing its fingerprint against Cisco's database of fingerprints. Cisco's dataset of base model fingerprints is also available on Hugging Face.

SecurityWeek
09

Enterprise Spotlight: Transforming software development with AI

industry
May 1, 2026

AI is changing how software is developed by affecting coding practices, tools, developer roles, and the overall development process across all stages, from initial planning through maintenance. The article discusses how AI agents are being integrated throughout the software development life cycle (the complete process of creating and maintaining software, from concept to deployment).

CSO Online
10

Hugging Face, ClawHub Abused for Malware Distribution

security
May 1, 2026

Threat actors are abusing AI distribution platforms like Hugging Face and ClawHub to spread malware by uploading trojanized files (files containing hidden malicious code) that trick users into downloading them through social engineering. The attackers use indirect prompt injection (embedding hidden instructions in data that AI systems read and execute without the user knowing) to make AI agents automatically download and run malware on users' computers, with hundreds of malicious files identified across both platforms.

SecurityWeek
Prev1...1516171819...371Next
critical

CVE-2026-42203: LiteLLM is a proxy server (AI Gateway) to call LLM APIs in OpenAI (or native) format. From version 1.80.5 to before vers

CVE-2026-42203NVD/CVE DatabaseMay 8, 2026
May 8, 2026
critical

Gemini CLI Vulnerability Could Have Led to Code Execution, Supply Chain Attack

SecurityWeekMay 7, 2026
May 7, 2026
critical

GHSA-9h64-2846-7x7f: Axonflow fixed bugs by implementing multi-tenant isolation and access-control hardening

GitHub Advisory DatabaseMay 6, 2026
May 6, 2026
critical

GHSA-gmvf-9v4p-v8jc: fast-jwt: JWT auth bypass due to empty HMAC secret accepted by async key resolver

CVE-2026-44351GitHub Advisory DatabaseMay 6, 2026
May 6, 2026