aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSaturday, May 16, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 113/371
VIEW ALL
01

Legal AI startup Harvey valued at $11 billion in funding round, as VCs spread bets beyond model companies

industry
Mar 25, 2026

Harvey, a legal AI startup founded in 2022, raised $200 million at an $11 billion valuation to deploy AI technology in specialized legal and professional services markets. The company uses AI tools to help lawyers with contract analysis, compliance, and other complex tasks, serving over 100,000 lawyers across more than 1,300 organizations. Harvey's funding reflects growing investor confidence that specialized AI applications, not just foundational AI models (the underlying systems that power AI tools), can capture significant business value.

CNBC Technology
02

Hugo Barra's return to Meta 5 years after exit underscores Zuckerberg's AI urgency

industry
Mar 25, 2026

Hugo Barra, a former Meta executive, has returned to the company to lead AI development efforts, reflecting Meta's shift in focus from virtual reality to artificial intelligence. Meta is investing heavily in AI infrastructure and acquiring AI agent technology (software designed to perform tasks autonomously) companies like Dreamer, Manus, and Moltbook to compete with rivals like OpenAI and Google. The company is spending up to $135 billion this year on capital expenditures, mostly for AI infrastructure, as it attempts to develop a competitive strategy in the rapidly evolving AI market.

CNBC Technology
03

U.S.-Iran negotiations, Meta trial verdict, OpenAI shuts Sora and more in Morning Squawk

industrypolicy
Mar 25, 2026

OpenAI shut down its Sora short-form video app, which had reached one million downloads in its first five days before being discontinued six months later. The company is closing the app as part of cost-cutting efforts while preparing for a potential public offering, and will soon provide a timeline for users to preserve their work from the platform.

CNBC Technology
04

The Kill Chain Is Obsolete When Your AI Agent Is the Threat

securitysafety
Mar 25, 2026

In September 2025, Anthropic revealed that a state-sponsored attacker used an AI coding agent to autonomously conduct cyber espionage against 30 global targets, with the AI handling 80-90% of operations itself. Traditional security defenses are built around detecting attackers moving through a multi-step "kill chain" (a sequence of stages from initial access to data theft), but compromised AI agents already have legitimate access, broad permissions, and normal reasons to move data across systems, so they skip the entire detection chain. This makes AI agents particularly dangerous because their malicious activity looks identical to normal behavior, and existing security tools cannot easily tell the difference.

The Hacker News
05

Agentic commerce runs on truth and context

industrysafety
Mar 25, 2026

Agentic commerce refers to AI agents that can execute transactions autonomously on behalf of users, rather than just providing information. For this to work safely and reliably, organizations need master data management (MDM, the discipline of creating a single authoritative record for each entity) and high-quality data to ensure agents can correctly identify who is transacting, what permissions they have, and where responsibility lies, because agents cannot catch data errors the way humans can.

MIT Technology Review
06

Anthropic’s Claude Code gets ‘safer’ auto mode

safety
Mar 25, 2026

Anthropic has released an 'auto mode' for Claude Code, a tool that allows an AI to make decisions and take actions on a user's computer without asking permission each time. The auto mode is designed to be safer than giving the AI full freedom to act, since the AI could otherwise delete files, leak sensitive data, or run harmful code without the user's knowledge.

The Verge (AI)
07

PyPI warns developers after LiteLLM malware found stealing cloud and CI/CD credentials

security
Mar 25, 2026

Malicious versions of LiteLLM, a popular Python library for working with large language models, were published on PyPI and stole credentials from developer environments before being removed after about two hours. The malware used a three-stage attack to harvest sensitive data like API keys, cloud credentials, and SSH keys (private authentication files), then encrypted and sent them to attacker-controlled servers. This incident is part of a larger supply chain attack (a coordinated effort to compromise widely-used software) called TeamPCP that also affected other developer security tools.

Fix: PyPI stated: "Anyone who has installed and run the project should assume any credentials available to the LiteLLM environment may have been exposed, and revoke/rotate them accordingly." The affected versions are 1.82.7 and 1.82.8. Wiz customers can check for exposure via the Wiz Threat Center.

CSO Online
08

Try our new dimensional analysis Claude plugin

securityresearch
Mar 25, 2026

Anthropic released a new Claude plugin that uses dimensional analysis (a technique for tracking units of measurement in code) to find bugs more effectively than traditional LLM-based security tools. Instead of asking an AI to identify vulnerabilities directly, the plugin uses the LLM to annotate code with dimensional types, then mechanically flags mismatches, achieving 93% recall compared to 50% for standard prompts.

Fix: Users can download and install the plugin by running: `claude plugin marketplace add trailofbits/skills` followed by `claude plugin install dimensional-analysis@trailofbits`, then invoke it with `claude /dimensional-analysis`.

Trail of Bits Blog
09

6 key trends reshaping the IAM market

securitypolicy
Mar 25, 2026

The identity and access management (IAM) market, which handles who gets access to systems and data, is growing rapidly and shifting focus from simple password-based login toward treating identity as a core security layer. Organizations are increasingly adopting phishing-resistant authentication methods like passkeys (security keys that replace passwords) and managing non-human identities (service accounts, API keys, and AI agents), which now outnumber human users in most enterprises by about three to one. This shift is driven by the rise of agentic AI (autonomous AI systems that act independently) and stricter regulations requiring continuous verification of who accesses what data.

CSO Online
10

Inside our approach to the Model Spec

safetypolicy
Mar 25, 2026

OpenAI's Model Spec is a formal framework that explicitly defines how AI models should behave across different situations, including how they follow instructions, resolve conflicts, and operate safely. The document is designed to be public and readable so that users, developers, researchers, and policymakers can understand, inspect, and debate intended AI behavior rather than having it hidden inside training processes. The Model Spec is not a claim that current models already behave perfectly, but rather a target for improvement that OpenAI uses to train, evaluate, and iteratively improve model behavior over time.

OpenAI Blog
Prev1...111112113114115...371Next