aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDatasetFor devs
Subscribe
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

Independent research. No sponsors, no paywalls, no conflicts of interest.

[TOTAL_TRACKED]
3,710
[LAST_24H]
1
[LAST_7D]
1
Daily BriefingSaturday, May 16, 2026

No new AI/LLM security issues were identified today.

Latest Intel

page 98/371
VIEW ALL
01

Shifting to AI model customization is an architectural imperative

industry
Mar 31, 2026

As improvements from new AI models have slowed to small gains, organizations are shifting toward customizing models with their own proprietary data and internal processes to gain competitive advantages. Domain-specialized models, which are trained on an organization's unique language, workflows, and expertise, can outperform general-purpose models and encode valuable business knowledge directly into the AI system.

MIT Technology Review
02

How to Categorize AI Agents and Prioritize Risk

securitypolicy
Mar 31, 2026

AI agents (AI systems that can reason, plan, and act autonomously across enterprise systems) are becoming more common in organizations, creating new security challenges. Risk from AI agents depends on two factors: access (which systems and data the agent can reach) and autonomy (how independently it can act without human approval). The text describes three categories of enterprise AI agents—agentic chatbots, local agents, and production agents—each with different risk levels based on their access and autonomy.

BleepingComputer
03

CrewAI Vulnerabilities Expose Devices to Hacking

security
Mar 31, 2026

CrewAI, an AI framework, has vulnerabilities that attackers can exploit using prompt injection (tricking an AI by hiding malicious instructions in its input) to chain together bugs and escape the sandbox (a restricted environment meant to contain the AI's actions) to run arbitrary code on a device.

SecurityWeek
04

Vertex AI Vulnerability Exposes Google Cloud Data and Private Artifacts

security
Mar 31, 2026

Researchers discovered a security vulnerability in Google Cloud's Vertex AI platform where AI agents could be compromised to steal sensitive data and access private cloud resources. The problem stems from the default service agent (P4SA, a special account that runs the AI agent) having excessive permissions, allowing attackers to extract credentials and gain unauthorized access to cloud storage, private code repositories, and internal Google infrastructure.

Fix: Google updated its documentation to explain how Vertex AI uses resources and accounts. The company recommended that customers use Bring Your Own Service Account (BYOSA) to replace the default service agent and enforce the principle of least privilege (PoLP, giving the agent only the permissions it needs to do its job).

The Hacker News
05

Accelerating the next phase of AI

industry
Mar 31, 2026

OpenAI announced a $122 billion funding round at an $852 billion valuation, positioning itself as core AI infrastructure globally. The company is experiencing rapid commercial growth, generating $2 billion in monthly revenue and expanding its products across ChatGPT, APIs, enterprise solutions, and specialized applications like coding and scientific discovery.

OpenAI Blog
06

OpenAI patches twin leaks as Codex slips and ChatGPT spills

security
Mar 31, 2026

OpenAI patched two separate security flaws in its AI tools: one in Codex (a coding agent) that allowed attackers to steal GitHub tokens through command injection (inserting malicious commands into user inputs), and another in ChatGPT's code execution environment that created a hidden channel for silently leaking user data without approval. Both bugs could let attackers extract sensitive information, but researchers warn that giving AI tools the ability to run code and access external systems inherently creates ongoing security risks.

Fix: OpenAI fixed the Codex vulnerability by 'tightening input validation around the vulnerable parameter and hardening how commands are constructed in the execution environment.' For the ChatGPT flaw, OpenAI addressed it by 'tightening controls around outbound communication in the code execution environment.' Both patches were deployed before public disclosure.

CSO Online
07

The Download: AI health tools and the Pentagon’s Anthropic culture war

policysafety
Mar 31, 2026

This newsletter covers multiple AI and tech news items, including concerns that medical chatbots from Microsoft, Amazon, and OpenAI are being released with little external evaluation before reaching the public. It also reports on regulatory efforts in California to impose AI safeguards despite opposition, legal challenges to Pentagon actions against Anthropic, and various other AI infrastructure and safety developments.

MIT Technology Review
08

AI benchmarks are broken. Here’s what we need instead.

research
Mar 31, 2026

Current AI benchmarks (standardized tests that measure AI performance) evaluate AI systems in isolation against human performance on specific tasks, but this doesn't reflect how AI is actually used in real organizations where it works within teams and workflows over extended periods. This misalignment causes organizations to adopt AI systems with impressive benchmark scores that then underperform in real-world deployment, such as FDA-approved radiology AI that creates delays when integrated into hospital workflows with multiple specialists and evolving decisions.

Fix: The source proposes shifting from narrow benchmark methods to HAIC benchmarks (Human-AI, Context-Specific Evaluation), which assess how AI systems perform over longer time horizons within human teams, workflows, and organizations. However, no implementation details, technical specifications, or concrete steps for implementing this approach are provided in the source text.

MIT Technology Review
09

CVE-2026-4399: Prompt injection vulnerability in 1millionbot Millie chatbot that occurs when a user manages to evade chat restrictions

security
Mar 31, 2026

A prompt injection vulnerability (a technique where attackers hide malicious instructions in their input to trick an AI) exists in the 1millionbot Millie chatbot, allowing users to bypass safety restrictions using Boolean logic tricks (phrasing questions to trigger 'true' responses that activate hidden commands). This could let attackers extract sensitive information, misuse the service, or access restricted features that the chatbot was designed to block.

NVD/CVE Database
10

How we made Trail of Bits AI-native (so far)

industry
Mar 31, 2026

Trail of Bits transformed from a company where 95% of staff resisted AI into one using 94 plugins and 84 specialized agents to find 200 bugs per week by shifting from AI-assisted (using AI as a standalone tool) to AI-native (redesigning the entire organization around AI as a core teammate). The post explains that most companies fail with AI because they don't change their workflows or systems, only distribute tools, and that psychological barriers like self-enhancing bias (overestimating our own judgment) and identity threat are the real obstacles to adoption.

Trail of Bits Blog
Prev1...96979899100...371Next