aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSunday, May 17, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 209/371
VIEW ALL
01

ByteDance’s next-gen AI model can generate clips based on text, images, audio, and video

industry
Feb 12, 2026

ByteDance has released Seedance 2.0, a new AI video generator that can create videos based on combined inputs of text, images, audio, and video prompts (instructions given to an AI to produce specific outputs). The company claims the model produces higher-quality videos with better ability to handle complex scenes and follow user instructions, allowing users to refine their requests by providing up to nine images, three video clips, and three audio clips.

The Verge (AI)
02

Fake AI Chrome extensions with 300K users steal credentials, emails

securityprivacy
Feb 12, 2026

Over 30 fake AI assistant Chrome extensions with more than 300,000 total users are stealing user credentials, emails, and browsing data by pretending to be AI tools. The extensions, collectively called AiFrame, don't actually run AI locally; instead, they load content from remote servers they control, allowing attackers to intercept sensitive information like Gmail messages and authentication details without users knowing.

Fix: The source recommends checking LayerX's list of indicators of compromise to identify if you have installed any malicious extensions. If compromise is confirmed, users should reset passwords for all accounts.

BleepingComputer
03

TrapFlow: Controllable Website Fingerprinting Defense via Dynamic Backdoor Learning

securityresearch
Feb 12, 2026

Website fingerprinting (WF) attacks are methods that monitor user traffic patterns to identify which websites they visit, threatening privacy even on protected networks. Existing defenses slow down these attacks but can be defeated when attackers retrain their models, and they also add significant slowness to network traffic. TrapFlow, a new defense technique, uses backdoor learning (injecting hidden trigger patterns into website traffic) to trick attackers' AI models into making wrong predictions, either by memorizing false patterns during training or by being confused at inference time (when making predictions on new data).

Fix: The source describes TrapFlow as the proposed defense method itself, which works by injecting crafted trigger sequences into targeted website traffic and optimizing these triggers using Fast Levenshtein-like distance metrics. However, no explicit patch, software update, configuration change, or deployment procedure is provided in the text. N/A -- no implementation mitigation discussed in source.

IEEE Xplore (Security & AI Journals)
04

Dual Frequency Branch Framework With Reconstructed Sliding Windows Attention for AI-Generated Image Detection

researchsafety
Feb 12, 2026

This paper describes a new method for detecting AI-generated images (images created by GANs, which are machine learning models that generate synthetic images, or diffusion models, which gradually refine noise into images) by analyzing images in multiple frequency domains (different ways of breaking down an image into mathematical components) using attention mechanisms (techniques that help AI focus on important parts of data). The approach achieved better detection accuracy than previous methods when tested on images from 65 different generative models.

IEEE Xplore (Security & AI Journals)
05

The Download: AI-enhanced cybercrime, and secure AI assistants

securitysafety
Feb 12, 2026

AI tools are making cybercrime easier by helping attackers write malicious code and automate attacks, while criminals also use deepfake technology (synthetic media that realistically mimics people) to impersonate others and commit scams. AI assistants that interact with external tools like email and web browsers pose serious security risks because their mistakes can have real-world consequences, especially when users hand over sensitive personal data to systems like OpenClaw.

MIT Technology Review
06

AI safety leader says 'world is in peril' and quits to study poetry

policysafety
Feb 12, 2026

Mrinank Sharma, a researcher who led AI safety efforts at Anthropic (a company focused on making AI systems safer and aligned with human values), resigned with a warning that "the world is in peril" due to interconnected crises including AI risks and bioweapons. Sharma said he observed that even safety-focused companies like Anthropic struggle to let their core values guide their actions when facing business pressures, and he plans to pursue poetry and writing in the UK instead.

BBC Technology
07

Palo Alto closes privileged access gap with $25B CyberArk acquisition

securityindustry
Feb 12, 2026

Palo Alto Networks acquired CyberArk for $25 billion to strengthen its ability to manage privileged access (controlling who can access sensitive systems and accounts) across human, machine, and AI identities through a unified platform. This addresses a critical security gap because identity has become the primary target in enterprise attacks, especially with the rise of AI agents (autonomous software that performs tasks independently) that operate 24/7 with broad permissions. The integration aims to help organizations prevent credential-based attacks and reduce breach response time by up to 80%.

CSO Online
08

What’s next for Chinese open-source AI

industry
Feb 12, 2026

Chinese AI companies have recently released open-weight models (AI models whose internal numerical parameters are publicly available for anyone to download and modify) that match Western AI performance at much lower costs, with DeepSeek's R1 and Alibaba's Qwen models becoming among the most downloaded globally. Unlike proprietary Western models like ChatGPT that users access through paid APIs (application programming interfaces, standardized ways for software to communicate), these Chinese open-source models allow developers to inspect, study, and modify the code themselves. If this trend continues, it could shift where AI innovation happens and who establishes industry standards worldwide.

MIT Technology Review
09

Google says hackers are abusing Gemini AI for all attacks stages

security
Feb 12, 2026

State-backed hackers from China, Iran, North Korea, and Russia are using Google's Gemini AI model to help carry out cyberattacks at every stage, from gathering target information to creating phishing emails and writing malware code. Criminal groups are also exploiting AI tools for social engineering attacks and building malware that uses AI to generate code automatically. Additionally, attackers are attempting model extraction and knowledge distillation (copying an AI model's decision-making by querying it repeatedly) to replicate Gemini's functionality for their own purposes.

BleepingComputer
10

What CISOs need to know about the OpenClaw security nightmare

securitysafety
Feb 12, 2026

OpenClaw is a popular open-source AI agent orchestration tool (software that coordinates multiple AI agents to complete tasks) that runs locally and can connect to apps like WhatsApp, Gmail, and smart home devices, but security researchers have found it to be critically insecure by default. Over 42,000 exposed instances have been discovered with authentication bypass vulnerabilities (weaknesses that let attackers skip login requirements) and potential remote code execution (RCE, where attackers can run commands on affected systems), exposing organizations to data breaches, credential theft, and regulatory violations.

Fix: Rich Mogull, chief analyst at Cloud Security Alliance, recommends that "CISOs prohibit its use altogether." He states: "The answer has to be 'no.' There is no security model."

CSO Online
Prev1...207208209210211...371Next