aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,649
[LAST_24H]
2
[LAST_7D]
160
Daily BriefingSaturday, March 28, 2026
>

OpenAI Shuts Down Sora Video App Over Profitability Concerns: OpenAI discontinued its Sora video-generation app and canceled a $1 billion Disney partnership because the service consumed too many computational resources without generating enough revenue to justify costs as the company prioritizes profitability.

>

Critical Injection Vulnerability in localGPT LLM Tool: CVE-2026-5002 allows remote injection attacks (inserting malicious code into input) through the LLM Prompt Handler in PromtEngineer localGPT's backend. The exploit code is publicly available, and the vendor has not responded to disclosure attempts.

>

Latest Intel

page 11/265
VIEW ALL
01

PyPI warns developers after LiteLLM malware found stealing cloud and CI/CD credentials

security
Mar 25, 2026

Malicious versions of LiteLLM, a popular Python library for working with large language models, were published on PyPI and stole credentials from developer environments before being removed after about two hours. The malware used a three-stage attack to harvest sensitive data like API keys, cloud credentials, and SSH keys (private authentication files), then encrypted and sent them to attacker-controlled servers. This incident is part of a larger supply chain attack (a coordinated effort to compromise widely-used software) called TeamPCP that also affected other developer security tools.

Critical This Week5 issues
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026

Political Deepfakes Gain Influence Despite Public Awareness: AI researchers found that creators use generative AI (technology that creates images or videos from text descriptions) to produce fake media of political figures for propaganda and profit, and these deepfakes shape public perception even when viewers know the content is fake.

>

TikTok's AI Ad Labels Failing in Practice: Major companies like Samsung are posting AI-generated ads on TikTok without the required disclosure labels, preventing users from identifying whether advertisements were created by AI or humans despite platform policies requiring transparency.

Fix: PyPI stated: "Anyone who has installed and run the project should assume any credentials available to the LiteLLM environment may have been exposed, and revoke/rotate them accordingly." The affected versions are 1.82.7 and 1.82.8. Wiz customers can check for exposure via the Wiz Threat Center.

CSO Online
02

Try our new dimensional analysis Claude plugin

securityresearch
Mar 25, 2026

Anthropic released a new Claude plugin that uses dimensional analysis (a technique for tracking units of measurement in code) to find bugs more effectively than traditional LLM-based security tools. Instead of asking an AI to identify vulnerabilities directly, the plugin uses the LLM to annotate code with dimensional types, then mechanically flags mismatches, achieving 93% recall compared to 50% for standard prompts.

Fix: Users can download and install the plugin by running: `claude plugin marketplace add trailofbits/skills` followed by `claude plugin install dimensional-analysis@trailofbits`, then invoke it with `claude /dimensional-analysis`.

Trail of Bits Blog
03

6 key trends reshaping the IAM market

securitypolicy
Mar 25, 2026

The identity and access management (IAM) market, which handles who gets access to systems and data, is growing rapidly and shifting focus from simple password-based login toward treating identity as a core security layer. Organizations are increasingly adopting phishing-resistant authentication methods like passkeys (security keys that replace passwords) and managing non-human identities (service accounts, API keys, and AI agents), which now outnumber human users in most enterprises by about three to one. This shift is driven by the rise of agentic AI (autonomous AI systems that act independently) and stricter regulations requiring continuous verification of who accesses what data.

CSO Online
04

Inside our approach to the Model Spec

safetypolicy
Mar 25, 2026

OpenAI's Model Spec is a formal framework that explicitly defines how AI models should behave across different situations, including how they follow instructions, resolve conflicts, and operate safely. The document is designed to be public and readable so that users, developers, researchers, and policymakers can understand, inspect, and debate intended AI behavior rather than having it hidden inside training processes. The Model Spec is not a claim that current models already behave perfectly, but rather a target for improvement that OpenAI uses to train, evaluate, and iteratively improve model behavior over time.

OpenAI Blog
05

The AI Hype Index: AI goes to war

industrypolicy
Mar 25, 2026

This article summarizes recent developments in AI, including controversies over weaponizing AI models like Claude, major user departures from ChatGPT, and large protests against AI in London. On a lighter note, AI agents (software programs that can act independently to accomplish tasks) are becoming popular online, with companies hiring their creators and developing quirky applications where AI agents appear to develop their own beliefs and philosophies.

MIT Technology Review
06

AI is breaking traditional security models — Here’s where they fail first

securityindustry
Mar 25, 2026

Traditional enterprise security relied on slow, manual processes where vulnerabilities were discovered through periodic scans, then triaged and fixed in a delayed workflow. AI and LLM-based systems are breaking this model by automating triage (the process of sorting and prioritizing findings), delivering vulnerabilities with full context and demanding immediate action, which forces organizations to rethink who is responsible for fixes and how quickly decisions happen. This shift also makes accountability explicit rather than implicit, requiring security teams to transition from handling individual findings to overseeing AI decision-making accuracy and approving exceptions.

CSO Online
07

How Charlotte AI AgentWorks Fuels Security's Agentic Ecosystem

industrysecurity
Mar 25, 2026

Modern cybersecurity operations face attacks that happen in seconds, overwhelming traditional human-centered defenses. CrowdStrike introduced Charlotte AI AgentWorks and Charlotte Agentic SOAR, two interconnected systems that use AI agents (autonomous software that can reason and take actions) to work alongside security analysts, automating routine tasks while keeping humans in control through oversight and guardrails.

CrowdStrike Blog
08

OpenAI ends Disney partnership as it closes Sora video-making app

industry
Mar 25, 2026

OpenAI has shut down Sora, its AI video-generation app (software that creates realistic videos from text descriptions), less than two years after launch, to focus on other projects like robotics and autonomous AI agents. The closure ends both the consumer app and professional platform, though image-making tools in ChatGPT remain unaffected. Disney, which had recently licensed its intellectual property (creative works and characters owned by a company) to Sora in a landmark deal, said it will now explore partnerships with other AI platforms.

BBC Technology
09

Introducing the OpenAI Safety Bug Bounty program

securitysafety
Mar 24, 2026

OpenAI has launched a Safety Bug Bounty program to identify AI abuse and safety risks in its products, complementing its existing Security Bug Bounty program. The new program focuses on issues like prompt injection (tricking an AI by hiding instructions in its input) that hijacks AI agents to perform harmful actions, unauthorized feature access, and proprietary information leaks, even if they don't qualify as traditional security vulnerabilities. Researchers can submit reports on reproducible safety issues that pose plausible and material harm to users.

OpenAI Blog
10

Auto mode for Claude Code

safetysecurity
Mar 24, 2026

Anthropic introduced auto mode for Claude Code, a new permissions system where Claude automatically decides whether to allow actions with safeguards in place. A separate classifier model (Claude Sonnet 4.6) reviews each action before it runs to block requests that go beyond the task scope, target untrusted infrastructure, or appear malicious, using customizable default filters that cover allowed operations like read-only requests and local file work, while blocking risky actions like force-pushing to git repositories or executing external code.

Simon Willison's Weblog
Prev1...910111213...265Next
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026
critical

CISA: New Langflow flaw actively exploited to hijack AI workflows

BleepingComputerMar 26, 2026
Mar 26, 2026
critical

GHSA-mxrg-77hm-89hv: n8n: Prototype Pollution in XML and GSuiteAdmin node parameters lead to RCE

CVE-2026-33696GitHub Advisory DatabaseMar 26, 2026
Mar 26, 2026