aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingFriday, May 8, 2026
>

Critical RCE Vulnerabilities in LiteLLM Proxy Server: LiteLLM, a proxy server that forwards requests to AI model APIs, disclosed three critical and high-severity flaws in versions 1.74.2 through 1.83.6. Two test endpoints allowed attackers with valid API keys to execute arbitrary code (running any commands an attacker wants) on the server by submitting malicious configurations or prompt templates without sandboxing (CVE-2026-42271, CVE-2026-42203, both critical), while a SQL injection flaw (inserting malicious code into database queries) let unauthenticated attackers read or modify stored API credentials (CVE-2026-42208, high).

>

ClaudeBleed Exploit Allows Extension Hijacking in Chrome: Anthropic's Claude browser extension contains a vulnerability that allows malicious Chrome extensions to hijack it and perform unauthorized actions like exfiltrating files, sending emails, or stealing code from private repositories. The flaw stems from the extension trusting any script from claude.ai without verifying the actual caller, and while Anthropic released a partial fix in version 1.0.70 on May 6, researchers report it remains exploitable when the extension runs in privileged mode.

Latest Intel

page 60/371
VIEW ALL
01

Capsule Security Emerges From Stealth With $7 Million in Funding

security
Apr 15, 2026

Capsule Security, an Israeli startup, has raised $7 million in funding to develop technology that secures AI agents (AI systems designed to perform tasks independently) by continuously monitoring their behavior at runtime (while the AI is actually running) to prevent unsafe or harmful actions.

>

AI Systems Show Triple the High-Risk Vulnerabilities of Legacy Software: Penetration testing data reveals that AI and LLM systems have 32% of findings rated high-risk compared to just 13% for traditional software, with only 38% of high-risk AI issues getting resolved. Security experts attribute this gap to rapid deployment without mature controls, novel attack surfaces like prompt injection (tricking AI by hiding instructions in input), and fragmented responsibility for remediation across teams.

>

Model Context Protocol Emerging as Critical Security Blind Spot: Model Context Protocol (MCP, a plugin system connecting AI agents to external tools) has become a major vulnerability vector as organizations fail to scan for or monitor MCP-related risks. Recent supply chain attacks, such as the postmark-mcp npm package that exfiltrated emails from 300 organizations, demonstrate how attackers exploit widely-trusted MCP packages and hardcoded credentials in AI configurations to enable credential theft and supply chain compromises at scale.

SecurityWeek
02

‘By Design’ Flaw in MCP Could Enable Widespread AI Supply Chain Attacks

security
Apr 15, 2026

Researchers have identified a flaw in Anthropic's Model Context Protocol (MCP, a system that allows AI models to interact with external tools and data) that permits unsanitized commands (user input that hasn't been cleaned or verified) to run without warning, potentially giving attackers complete control over systems using this AI technology. This vulnerability could be exploited across many widely-used AI environments as part of a supply chain attack (where attackers compromise a tool or service used by many organizations to gain access to their systems).

SecurityWeek
03

Adobe embraces conversational AI editing, marking a ‘fundamental shift’ in creative work

industry
Apr 15, 2026

Adobe is launching a Firefly AI Assistant that lets creators edit their work by describing changes in plain language rather than manually using specific tools in Creative Cloud (Adobe's suite of creative software). Adobe says this conversational AI approach represents a major shift in how creative work is done by making editing easier and more accessible to people without advanced skills.

The Verge (AI)
04

OpenAI pulls back from Stargate Norway data center deal as Microsoft takes over

industry
Apr 15, 2026

OpenAI has withdrawn from a deal to rent computing capacity directly from a Norwegian data center facility called Stargate Norway, with Microsoft taking over the capacity instead. OpenAI will now rent computing power from Microsoft instead, which the company says makes more financial sense since it already has a $250 billion contract with Microsoft's cloud service Azure (a cloud computing platform). This pullback is part of OpenAI's broader shift to manage spending expectations as it prepares for a potential public stock offering.

CNBC Technology
05

Copilot and Agentforce fall to form-based prompt injection tricks

security
Apr 15, 2026

Security researchers discovered prompt injection vulnerabilities (attacks where malicious instructions are hidden in user input to trick an AI into executing them) in Microsoft Copilot Studio and Salesforce Agentforce that allow attackers to steal sensitive data like customer names, addresses, and phone numbers. Both vulnerabilities exploit the fact that these AI agents cannot distinguish between trusted system instructions and untrusted user input, allowing attackers to override the agent's original purpose and exfiltrate data to external servers.

Fix: Microsoft patched CVE-2026-21520 following disclosure, with the mitigation carried out internally and no further action required from users. The source notes that both vulnerabilities highlight a baseline need for treating all external inputs as untrusted and enforcing input validation, least-privilege access (giving systems only the minimum permissions they need), and strict controls on actions like outbound email, though no specific patch details are provided for the Salesforce vulnerability.

CSO Online
06

Retaining defensive advantage in the age of frontier AI cyber capabilities 

policysecurity
Apr 15, 2026

Frontier AI models (cutting-edge artificial intelligence systems) are becoming better at finding vulnerabilities (weaknesses in code that attackers can exploit), which creates both opportunity and risk. While AI can help organizations identify and fix these weaknesses, attackers can now use AI to discover and exploit vulnerabilities faster and cheaper than before, putting pressure on organizations to patch systems quickly. The recommended defense is for organizations to follow established best practices from the National Cyber Security Centre, including reducing unnecessary exposure to attack, applying security updates rapidly, and monitoring for malicious activity.

Fix: Organizations should follow established good practices set out by the National Cyber Security Centre, which include: reducing unnecessary exposure to attack, applying security updates rapidly, and monitoring for and quickly responding to malicious activity detected. Additionally, organizations should pursue government-backed certifications such as Cyber Essentials, and access guidance and tools available on the NCSC website.

UK NCSC
07

Microsoft, Salesforce Patch AI Agent Data Leak Flaws

security
Apr 15, 2026

Salesforce and Microsoft recently fixed two prompt injection vulnerabilities (security flaws where attackers hide malicious instructions in text inputs to trick AI systems) in their AI agent products, Agentforce and Copilot. These flaws could have allowed external attackers to access and steal sensitive data from users.

Dark Reading
08

Deterministic + Agentic AI: The Architecture Exposure Validation Requires

securityindustry
Apr 15, 2026

Organizations are rapidly adopting AI for security testing, but fully agentic AI systems (where AI makes all decisions from start to finish) create a problem: they produce different results each time they run, making it impossible to tell if security actually improved or if the AI just tried a different approach. The source argues that a hybrid model works better, where deterministic logic (fixed, repeatable sequences) defines how security tests execute, while AI enhances specific parts like adapting payloads and interpreting what it finds.

The Hacker News
09

Grok’s sexual deepfakes almost got it banned from Apple’s App Store. Almost. 

safetypolicy
Apr 15, 2026

Apple threatened to remove Elon Musk's AI app, Grok, from its App Store in January because it wasn't stopping nonconsensual sexual deepfakes (fake sexually explicit images created using AI) from spreading on X. Apple contacted the developers behind X and Grok and asked them to create a plan to improve their content moderation (systems for reviewing and removing harmful material).

Fix: Apple demanded that the developers 'create a plan to improve content moderation,' according to a letter the company sent to US senators.

The Verge (AI)
10

The Deepfake Nudes Crisis in Schools Is Much Worse Than You Thought

safetypolicy
Apr 15, 2026

Teenage boys are using AI "nudify" apps to create deepfake sexual imagery (fake nude photos or videos created by AI) of their female classmates, which are then shared on social media and messaging apps. Since 2023, this has affected over 600 students across at least 28 countries and nearly 90 schools, with the true scale likely much higher. The explicit imagery involving minors constitutes child sexual abuse material (CSAM), and schools and law enforcement are often unprepared to respond to these serious incidents.

Wired (Security)
Prev1...5859606162...371Next