aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSunday, May 17, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 187/371
VIEW ALL
01

Anthropic accuses DeepSeek and other Chinese firms of using Claude to train their AI

securityindustry
Feb 23, 2026

Anthropic accused three Chinese AI companies, DeepSeek, MiniMax, and Moonshot, of misusing its Claude model through large-scale fraudulent activity to train their own AI systems. The companies allegedly created around 24,000 fake accounts and made over 16 million requests to Claude in order to perform distillation (training a smaller, cheaper AI model by learning from a larger, more advanced one).

The Verge (AI)
02

Anthropic accuses Chinese AI labs of mining Claude as US debates AI chip exports

securitypolicy
Feb 23, 2026

Anthropic accused three Chinese AI companies (DeepSeek, Moonshot AI, and MiniMax) of using distillation (a technique where one AI model learns from another by analyzing its outputs) to illegally extract capabilities from Claude by creating over 24,000 fake accounts and generating millions of interactions. This theft targeted Claude's most advanced features like reasoning, tool use, and coding, and raises security concerns because stolen models may lack safeguards against misuse like bioweapon development.

Fix: Anthropic stated it will 'continue to invest in defenses that make distillation attacks harder to execute and easier to identify,' and is calling on 'a coordinated response across the AI industry, cloud providers, and policymakers.' The company also argues that export controls on advanced AI chips to China would limit both direct model training and the scale of such distillation attacks.

TechCrunch
03

IBM is the latest AI casualty. Shares are tanking 11% on Anthropic programming language threat

industry
Feb 23, 2026

IBM's stock fell 11% after Anthropic announced that its Claude AI model can now automate COBOL (a decades-old programming language used in banking and business systems) modernization work, which is a core part of IBM's business. Claude can map dependencies, document workflows, and identify risks in old code much faster than human analysts, potentially making IBM's COBOL-related services less valuable.

CNBC Technology
04

Google’s Cloud AI lead on the three frontiers of model capability

industry
Feb 23, 2026

Michael Gerstenhaber, a Google Cloud VP overseeing Vertex (a platform for deploying enterprise AI), describes how AI models are advancing along three distinct frontiers: raw intelligence (accuracy and capability), response time (latency, or how quickly the model answers), and cost-efficiency (whether a model can run reliably at massive, unpredictable scale). Different use cases prioritize these frontiers differently—for example, code generation prioritizes intelligence even if it takes time, customer support prioritizes speed within a latency budget, and large-scale content moderation prioritizes cost-effectiveness at infinite scale.

TechCrunch
05

Cybersecurity stock selling deepens on AI threat concerns. Why we're not bailing

industry
Feb 23, 2026

This article discusses concerns about AI posing a threat to cybersecurity companies, which has caused their stock prices to decline. However, the piece argues against abandoning investments in these companies despite these concerns.

CNBC Technology
06

OpenAI calls in the consultants for its enterprise push

industry
Feb 23, 2026

OpenAI has announced the 'Frontier Alliance,' a partnership with four major consulting firms (Boston Consulting Group, McKinsey, Accenture, and Capgemini) to help enterprises adopt its AI technologies, particularly OpenAI Frontier, a no-code platform for building and deploying AI agents. The partnership aims to address slow enterprise adoption of AI by helping consultants redesign company strategies and workflows to integrate OpenAI's tools rather than simply adding AI to existing processes.

TechCrunch
07

Guide Labs debuts a new kind of interpretable LLM

researchindustry
Feb 23, 2026

Guide Labs has open-sourced Steerling-8B, an 8 billion parameter LLM designed to be interpretable, meaning its decisions can be traced back to its training data and understood rather than treated as a black box. The model uses a new architecture with a concept layer that buckets data into traceable categories, allowing developers to understand why the model produces specific outputs and control its behavior for applications like blocking copyrighted content or preventing bias in loan evaluations.

TechCrunch
08

Writing about Agentic Engineering Patterns

researchindustry
Feb 23, 2026

A software engineer is creating a collection of documented patterns for agentic engineering, which refers to using coding agents (AI tools that can generate, execute, and iterate on code independently) to help professional developers work faster and better. The project will be published as a series of chapters on a blog, inspired by classic design pattern documentation, with the first two chapters covering how cheap code generation changes software development and how test-first development (TDD) helps agents write better code.

Simon Willison's Weblog
09

Cybersecurity stocks drop for a second day as new Anthropic tool fuels AI disruption fears

industry
Feb 23, 2026

Cybersecurity stock prices fell sharply after Anthropic announced a new AI tool for its Claude model that can scan software code for vulnerabilities and suggest fixes, causing investors to worry that AI might replace traditional cybersecurity services. However, some analysts argue the threat is limited, noting that while AI could improve efficiency in specific tasks like code scanning, it cannot yet replace full end-to-end security platforms (complete systems that handle all stages of protecting against attacks).

CNBC Technology
10

Does Big Tech actually care about fighting AI slop?

safetypolicy
Feb 23, 2026

Instagram's leader Adam Mosseri warned that AI can now convincingly fake almost any content, making it hard for creators to stand out with authentic material. He proposed solving this by having camera manufacturers cryptographically sign images (using math-based codes that prove an image wasn't altered) at the moment they're captured, creating a verifiable record of what's real versus AI-generated.

Fix: Camera manufacturers will cryptographically sign images at capture, creating a chain of custody to establish a trustworthy system for determining what's not AI.

The Verge (AI)
Prev1...185186187188189...371Next