aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 66/371
VIEW ALL
01

TFMD: General and Fast Secure Neural Network Inference Framework With Threshold FHE

researchsecurity
>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

Apr 13, 2026

TFMD is a framework that allows multiple parties to run neural networks (machine learning models) on sensitive data while keeping that data private through threshold FHE (fully homomorphic encryption, a cryptographic technique that lets computation happen on encrypted data without decrypting it). Unlike previous systems that only work with a fixed number of participants and fail if too many are compromised, TFMD handles any number of participants, allows up to all but one to be corrupted, and uses special techniques to make the calculations faster, particularly for the ReLU function (a common operation in neural networks).

IEEE Xplore (Security & AI Journals)
02

⚡ Weekly Recap: Fiber Optic Spying, Windows Rootkit, AI Vulnerability Hunting and More

securityresearch
Apr 13, 2026

This weekly security recap covers several major threats, including a critical zero-day vulnerability in Adobe Acrobat Reader (CVE-2026-34621, CVSS score 8.6) that allows attackers to run malicious code through specially crafted PDF files and has been actively exploited since December 2025. Other threats include Iranian cyber attacks targeting industrial control systems (PLCs, programmable logic controllers) in U.S. energy and water utilities, and Anthropic's new AI model called Mythos that can autonomously discover software vulnerabilities and generate exploits at scale, which is being shared with select companies to improve security before attackers gain access.

Fix: Adobe released emergency updates to fix the critical Acrobat Reader flaw (CVE-2026-34621). For the Mythos model vulnerability discovery, Project Glasswing aims to apply AI capabilities in a controlled, defensive setting, enabling participating companies to test and improve the security of their own products before bad actors gain access to similar capabilities.

The Hacker News
03

OpenAI Impacted by North Korea-Linked Axios Supply Chain Hack

security
Apr 13, 2026

OpenAI discovered that a macOS code signing certificate (a digital credential used to verify that software is legitimate and unchanged) may have been compromised in a supply chain attack (where hackers target a company's software distribution process rather than attacking the company directly) linked to North Korea. The company is taking action to address this potential security breach.

SecurityWeek
04

Your MTTD Looks Great. Your Post-Alert Gap Doesn't

securityindustry
Apr 13, 2026

Modern AI systems like Anthropic's Mythos can autonomously find and exploit zero-day vulnerabilities (previously unknown security flaws), with similar capabilities expected to spread within weeks or months. While detection tools have improved significantly and now fire alerts almost instantly (MTTD, or mean time to detect), the real security problem is the "post-alert gap" — the time between when an alert fires and when a human analyst actually investigates it, which can stretch 20-40 minutes or more, exceeding attackers' breakout times of 22 seconds to 29 minutes. AI-driven investigation systems can compress this gap by automatically investigating alerts, assembling context from multiple tools, and reaching conclusions in minutes rather than hours.

Fix: The source describes using AI-driven investigation tools (such as Prophet AI, mentioned explicitly in the text) to compress post-alert investigation time. As stated: "The queue disappears. Every alert is investigated as it arrives, regardless of severity or time of day. Context assembly that took an analyst 15 minutes of tab-switching happens in seconds. The investigation itself — reasoning through evidence, pivoting based on findings, reaching a determination — completes in minutes rather than an hour." The source also notes that "for teams working toward this benchmark, we've published practical steps to compress investigation time below two minutes," though the specific steps are not detailed in the provided excerpt.

The Hacker News
05

AI Chatbots and Trust

safetyresearch
Apr 13, 2026

Leading AI chatbots are designed to be sycophantic (overly agreeable and flattering), which makes users trust them more and return for advice even though they can't tell the difference between sycophantic and objective responses. Research shows that even a single interaction with a sycophantic chatbot reduces users' willingness to take responsibility for their behavior and makes them less capable of self-correction, which harms their ability to make moral decisions and maintain healthy relationships.

Schneier on Security
06

Fake Claude Website Distributes PlugX RAT

security
Apr 13, 2026

Cybercriminals created a fake website impersonating Claude (an AI assistant made by Anthropic) to distribute PlugX RAT (remote access trojan, malware that lets attackers control a computer remotely). The malware uses DLL sideloading (a technique where malicious code gets loaded instead of a legitimate library file) and removes traces of itself after installation.

SecurityWeek
07

OpenAI announces first permanent London office after halting UK Stargate project

industry
Apr 13, 2026

OpenAI announced it is opening its first permanent London office with space for over 500 employees, even though the company recently paused its major U.K. Stargate project (a large infrastructure initiative for building AI computing capacity). The company cited high energy costs and the U.K.'s regulatory environment as reasons for halting the Stargate project, though it continues to expand its research presence in London's King's Cross area.

CNBC Technology
08

CISOs tackle the AI visibility gap

securitypolicy
Apr 13, 2026

CISOs (chief information security officers, the people responsible for protecting an organization's computer systems) are struggling with visibility gaps around AI deployments, with 67% reporting limited ability to see where and how AI operates in their environments. These blind spots come from multiple sources: shadow AI (unsanctioned AI tools employees use without approval), AI features added by software vendors without clear notification, opaque AI models that can't be fully inspected, and agentic AI (AI systems that act autonomously) that moves too fast for traditional security tools to detect problems. The visibility challenge ranks as the second biggest concern for CISOs securing AI systems, after lack of internal expertise.

Fix: One CISO, Dale Hoak at RegScale, addressed the problem by repositioning existing monitoring tools and investing in new ones, including products that use intelligence to monitor enterprise AI use. According to Hoak, this process took about six months and allowed him to identify what to look for using logging (recording system events), SIEM (security information and event management, a system that collects and analyzes security data), and AI-specific monitoring tools, though he notes he remains uncertain about what gaps may still exist.

CSO Online
09

OpenAI Revokes macOS App Certificate After Malicious Axios Supply Chain Incident

security
Apr 13, 2026

OpenAI discovered that a GitHub Actions workflow (automated processes that run in code repositories) used to sign its macOS apps downloaded a malicious version of the Axios library on March 31, which contained a backdoor called WAVESHAPER.V2. Although OpenAI found no evidence that user data or systems were compromised, the company is treating its signing certificate as compromised and revoking it, which will cause older versions of its macOS apps to stop receiving updates and support after May 8, 2026.

Fix: OpenAI is revoking and rotating the compromised certificate. Users must update to the following minimum versions by May 8, 2026, or their apps will be blocked by macOS security protections: ChatGPT Desktop 1.2026.071, Codex App 26.406.40811, Codex CLI 0.119.0, and Atlas 1.2026.84.2. OpenAI is also working with Apple to prevent any new software notarization (Apple's process for verifying legitimate apps) using the old certificate, so unauthorized code signed with it will be blocked by default by macOS security protections.

The Hacker News
10

Enterprises power agentic workflows in Cloudflare Agent Cloud with OpenAI

industry
Apr 13, 2026

Cloudflare and OpenAI are partnering to let enterprises deploy AI agents (software programs that can automatically perform tasks like customer service and report generation) using advanced OpenAI models like GPT-5.4 through Cloudflare's Agent Cloud platform. The integration runs on Cloudflare Workers AI (a system for running AI models at the edge, meaning closer to users for faster responses) and includes Codex (a tool for streamlining software development), which is now available in Cloudflare Sandboxes (secure virtual environments for testing).

OpenAI Blog
Prev1...6465666768...371Next