aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSaturday, May 16, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 76/371
VIEW ALL
01

Google makes it easy to deepfake yourself

safety
Apr 9, 2026

YouTube Shorts is launching a new AI feature that lets creators make digital clones of themselves, called avatars, that look and sound like them and can be used in videos. The feature adds to YouTube's struggle with managing AI-generated content, including deepfakes (synthetic videos where someone's face or voice is digitally recreated to look authentic), AI slop (low-quality AI-generated content), and impersonation scams.

The Verge (AI)
02

Gemini gets notebooks to help you organize projects

industry
Apr 8, 2026

Google is adding a feature called "notebooks" to Gemini (its AI chatbot) that lets users organize files, past conversations, and custom instructions about specific topics in one place. Gemini can then use this organized information as context (background information the AI considers) when answering questions, similar to ChatGPT's Projects feature from 2024.

The Verge (AI)
03

CyberAgent moves faster with ChatGPT Enterprise and Codex

industry
Apr 8, 2026

CyberAgent, a Japanese internet company, adopted ChatGPT Enterprise and Codex to make AI a foundational technology across their organization rather than just an isolated initiative. The company faced challenges around security concerns and uncertainty about what data could safely be shared with AI tools, which slowed adoption and created inconsistent usage across departments.

Fix: CyberAgent addressed these challenges by adopting ChatGPT Enterprise, which provides enterprise-grade security features, access controls, account management, and visibility into usage that allow employees to confidently use AI. The company also established internal guidelines for handling confidential information to ensure safe and secure use, and provided ongoing training support to build a culture of responsible AI adoption.

OpenAI Blog
04

Anthropic loses appeals court bid to temporarily block Pentagon blacklisting

policy
Apr 8, 2026

A federal appeals court in Washington, D.C. denied Anthropic's request to temporarily block the Department of Defense's blacklisting of the company as a supply chain risk (a designation claiming the company's technology threatens U.S. national security). The ruling means Anthropic is excluded from DOD contracts, though a separate court earlier granted Anthropic an injunction allowing it to continue working with other government agencies while the lawsuit challenging the blacklisting continues.

CNBC Technology
05

OpenAI will allocate IPO shares to retail investors as it preps for debut, CFO says

industry
Apr 8, 2026

OpenAI's CFO announced that the company plans to reserve shares for individual investors when it goes public through an initial public offering (IPO, the first time a private company sells shares to the public). The company saw strong demand from regular retail investors during its recent funding round and wants to ensure broad public participation in ownership, following models used by other companies like Tesla and Block.

CNBC Technology
06

Anthropic keeps latest AI tool out of public’s hands for fear of enabling widespread hacking

securitysafety
Apr 8, 2026

Anthropic has developed an AI model called Claude Mythos that is unusually good at finding software vulnerabilities (security weaknesses in code), and it discovered thousands of these flaws in commonly-used applications that don't yet have fixes available. The company decided not to release Mythos widely to the public because they worry it could enable widespread hacking, and instead partnered with cybersecurity specialists to improve defenses before wider distribution.

The Guardian Technology
07

Cracks in the Bedrock: Agent God Mode

security
Apr 8, 2026

Amazon Bedrock AgentCore's starter toolkit automatically creates overly broad IAM roles (identity and access management policies that control what actions software can perform) that grant a single AI agent excessive permissions across an entire AWS account, enabling an attack called Agent God Mode. If compromised, an attacker could exploit these permissions to access other agents' memories, steal container images, and extract sensitive data. AWS updated its documentation to warn that the default roles are only for development and testing, not production use.

Fix: AWS documentation was updated to include a security warning, stating that the default roles are "designed for development and testing purposes" and are not recommended for production deployment.

Palo Alto Unit 42
08

GHSA-2763-cj5r-c79m: PraisonAI Vulnerable to OS Command Injection

security
Apr 8, 2026

PraisonAI has a critical vulnerability where the `execute_command` function and workflow shell execution pass user-controlled input directly to `subprocess.run()` with `shell=True`, allowing attackers to inject arbitrary shell commands through YAML workflow files, agent configurations, and LLM-generated tool calls by exploiting shell metacharacters like semicolons and pipes.

GitHub Advisory Database
09

GHSA-926x-3r5x-gfhw: LangChain has incomplete f-string validation in prompt templates

security
Apr 8, 2026

LangChain had incomplete validation of f-string templates (a Python feature for inserting variables into text) in some prompt template classes. Attackers who could control the template structure could use attribute access (like `object.field`) or indexing (like `array[0]`) to expose internal data from Python objects being formatted. This issue only affected applications that allow untrusted users to write templates, not those using hardcoded templates or only letting users provide variable values.

Fix: LangChain now applies consistent f-string safety validation across all prompt template classes. The fix rejects templates containing attribute access or indexing syntax (such as `.` or `[]`) and rejects nested replacement fields inside format specifiers (templates with `{` or `}` in the format specification part). This blocks malicious patterns while preserving normal f-string formatting features.

GitHub Advisory Database
10

CVE-2026-5803: A security flaw has been discovered in bigsk1 openai-realtime-ui up to 188ccde27fdf3d8fab8da81f3893468f53b2797c. The aff

security
Apr 8, 2026

A security vulnerability (CVE-2026-5803) was found in bigsk1 openai-realtime-ui that allows attackers to perform SSRF (server-side request forgery, where an attacker tricks a server into making unwanted requests to other systems) through the API Proxy Endpoint in server.js by manipulating a query argument, and this flaw can be exploited remotely. The product uses continuous delivery with rolling releases, so specific affected versions are not documented.

Fix: Install the patch named 54f8f50f43af97c334a881af7b021e84b5b8310f to address this issue.

NVD/CVE Database
Prev1...7475767778...371Next