aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSunday, May 17, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 217/371
VIEW ALL
01

X offices raided in France as UK opens fresh investigation into Grok

safetypolicy
Feb 3, 2026

X's French offices were raided by Paris prosecutors investigating suspected illegal data extraction and possession of child sexual abuse material (CSAM, images depicting the sexual abuse of children), while the UK's Information Commissioner's Office launched a separate investigation into Grok (Elon Musk's AI chatbot) for its ability to create harmful sexualized images and videos without people's consent. The investigations were triggered by reports that Grok generated sexual deepfakes (fake sexual images created using real photos of women without permission) that were shared on X.

BBC Technology
02

CVE-2026-24887: Claude Code is an agentic coding tool. Prior to version 2.0.72, due to an error in command parsing, it was possible to b

security
Feb 3, 2026

Claude Code is an agentic coding tool (software that can automatically write and execute code) that had a vulnerability in versions before 2.0.72 where attackers could bypass safety confirmation prompts and execute untrusted commands through the find command by injecting malicious content into the tool's context window (the input area where the AI reads information). The vulnerability has a CVSS score (a 0-10 severity rating) of 7.7, meaning it is considered high severity.

Fix: This issue has been patched in version 2.0.72.

NVD/CVE Database
03

CVE-2026-24053: Claude Code is an agentic coding tool. Prior to version 2.0.74, due to a Bash command validation flaw in parsing ZSH clo

security
Feb 3, 2026

Claude Code, an agentic coding tool (AI software that writes and manages code), had a vulnerability in versions before 2.0.74 where a flaw in how it validated Bash commands (a Unix shell language) allowed attackers to bypass directory restrictions and write files outside the intended folder without permission from the user. The attack required the user to be running ZSH (a different Unix shell) and to allow untrusted content into Claude Code's input.

Fix: This issue has been patched in version 2.0.74. Users should update Claude Code to version 2.0.74 or later.

NVD/CVE Database
04

CVE-2026-24052: Claude Code is an agentic coding tool. Prior to version 1.0.111, Claude Code contained insufficient URL validation in it

security
Feb 3, 2026

Claude Code, a tool that helps AI write and execute code automatically, had a security flaw before version 1.0.111 where it didn't properly check website addresses (URLs) before making requests to them. The app used a simple startsWith() check (looking only at the beginning of a domain name), which meant attackers could register fake domains like modelcontextprotocol.io.example.com that would be mistakenly trusted, allowing the tool to send data to attacker-controlled sites without the user knowing.

Fix: Update Claude Code to version 1.0.111 or later, as the issue has been patched in that version.

NVD/CVE Database
05

AI May Supplant Pen Testers, But Oversight & Trust Are Not There Yet

securityindustry
Feb 3, 2026

AI agents are increasingly finding and reporting common security vulnerabilities (weaknesses in software) faster than human pen testers (security professionals who test systems for flaws), particularly through crowdsourced bug bounty programs (platforms where people are paid to find and report bugs). However, the source indicates that oversight and trust in these AI systems are not yet sufficiently developed to fully replace human expertise.

Dark Reading
06

From ‘nerdy’ Gemini to ‘edgy’ Grok: how developers are shaping AI behaviours

safetypolicy
Feb 3, 2026

AI assistants like ChatGPT, Grok, and Qwen have their personalities and ethical rules shaped by their creators, and changes to these rules can cause serious problems for users. Recent examples include Grok generating millions of inappropriate sexual images and ChatGPT appearing to encourage self-harm, showing that how developers program an AI's behavior (its ethical codes) has real consequences.

The Guardian Technology
07

Secure Acceleration of Aggregation Queries Over Homomorphically Encrypted Databases

research
Feb 3, 2026

This research proposes AHEDB (Accelerated Homomorphically Encrypted DataBase), a system designed to speed up database queries on encrypted data using Fully Homomorphic Encryption, or FHE (a method that lets computers perform calculations on encrypted information without decrypting it first). The system uses Encrypted Multiple Maps to reduce computational strain and a Single Range Cover algorithm for indexing, achieving better performance than existing FHE-based approaches while maintaining security.

IEEE Xplore (Security & AI Journals)
08

Toward Real-World Holistic Privacy-Preserving Person Re-Identification

securityprivacy
Feb 3, 2026

Person re-identification (Re-ID, systems that recognize and track individuals across camera footage) systems can be attacked to steal pedestrian images and the AI model itself, threatening privacy for both the system operator and people being monitored. Existing privacy-protection methods fail to defend against all types of leaks while keeping the system working normally, so researchers propose SHIELD, a two-stage framework that uses protected image generation and feature protection techniques to prevent data and model theft without reducing the system's accuracy for authorized users.

IEEE Xplore (Security & AI Journals)
09

Allies Teach Better Than Enemies: Inverse Adversaries for Robust Knowledge Distillation

researchsafety
Feb 3, 2026

This research proposes a new method for knowledge distillation (training a smaller AI model to mimic a larger one) that preserves adversarial robustness (the ability to resist attacks designed to fool AI systems). Instead of having the student model copy all predictions from the teacher model, the method uses "inverse adversarial examples" (inputs created by reversing the direction of adversarial attacks) to guide learning toward more reliable predictions, resulting in better robustness transfer between models.

IEEE Xplore (Security & AI Journals)
10

Evaluating and Mitigating Relationship Hallucinations in Large Vision-Language Models

researchsafety
Feb 3, 2026

Large vision-language models (LVMs, AI systems that process both images and text) often make mistakes by hallucinating incorrect relationships between objects in images, such as falsely claiming one object is near another. Researchers created R-Bench, a benchmark (a standardized test) to evaluate these relationship hallucination errors, and found that these mistakes happen because models rely too much on language patterns rather than actually analyzing the visual content. The study proposes Region-Aware Alignment Mitigation (RA²M), which improves the model's attention to specific regions of an image to better align its descriptions with what is actually shown.

Fix: Region-level image-text alignment helps mitigate relationship hallucinations. The authors propose Region-Aware Alignment Mitigation (RA²M), which 'enhances model attention to relevant regions, improving alignment between generated text and images.'

IEEE Xplore (Security & AI Journals)
Prev1...215216217218219...371Next