aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Browse All

All tracked items across vulnerabilities, news, research, incidents, and regulatory updates.

to
Export CSV
3144 items

Burger King cooks up AI chatbot to spot if employees say ‘please’ and ‘thank you’

infonews
industry
Feb 26, 2026

Burger King is deploying an AI chatbot powered by OpenAI (the company behind ChatGPT) that listens to employee headsets at hundreds of US locations to monitor whether workers use polite words like 'please' and 'thank you.' The company says the system, called BK Assistant, will help understand service patterns, though the announcement has sparked criticism from workers.

The Guardian Technology

Anthropic CEO Amodei says Pentagon's threats 'do not change our position' on AI

infonews
policy
Feb 26, 2026

Anthropic CEO Dario Amodei stated the company will not allow the U.S. Department of Defense to use its AI models without restrictions on fully autonomous weapons and mass domestic surveillance, despite Pentagon threats to label the company a supply chain risk or invoke the Defense Production Act. The DoD counters that it only wants to use the models for lawful purposes and has given Anthropic until Friday evening to agree to unrestricted access, with competing AI companies like OpenAI and Google already accepting these terms.

Anthropic says it ‘cannot in good conscience’ allow Pentagon to remove AI checks

infonews
policysafety

Anthropic refuses Pentagon’s new terms, standing firm on lethal autonomous weapons and mass surveillance

infonews
policy
Feb 26, 2026

Anthropic rejected the Pentagon's demands for unrestricted access to its AI system, refusing to agree to two specific uses: mass surveillance of Americans and lethal autonomous weapons (weapons that can kill targets without human oversight). The refusal came just before a deadline set by Defense Secretary Pete Hegseth, who wanted to renegotiate AI contracts with the military.

Anthropic CEO stands firm as Pentagon deadline looms

infonews
policy
Feb 26, 2026

Anthropic's CEO Dario Amodei refused the Pentagon's demand for unrestricted access to the company's AI systems, citing two concerns: mass surveillance of Americans and fully autonomous weapons (weapons that make decisions without human involvement) with no human oversight. The Pentagon threatened to label Anthropic a security risk or use the Defense Production Act (a law giving the president power to force companies to prioritize defense production) to force compliance, but Amodei said the company would work with the military under its proposed safeguards or help transition to another provider if the Pentagon chose to end the relationship.

Microsoft’s Copilot Tasks AI uses its own computer to get things done

infonews
industry
Feb 26, 2026

Microsoft is previewing Copilot Tasks, an AI system that runs on Microsoft's cloud servers to complete repetitive work for you, such as scheduling appointments or creating study plans, while you use your own device for other tasks. You can describe what you want using plain English and set the tasks to run once, on a schedule, or repeatedly, and the AI will send you a report when finished.

GHSA-38c7-23hj-2wgq: n8n has Webhook Forgery on Zendesk Trigger Node

mediumvulnerability
security
Feb 26, 2026

A vulnerability in n8n's Zendesk Trigger node (a tool that automatically starts workflows when Zendesk sends data) allows attackers to forge webhook requests, meaning they can trigger workflows with fake data because the node doesn't verify the HMAC-SHA256 signature (a cryptographic check that confirms a message is authentic). This lets anyone who knows the webhook URL send malicious payloads to the connected workflow.

GHSA-fvfv-ppw4-7h2w: n8n has a Guardrail Node Bypass

mediumvulnerability
security
Feb 26, 2026

A security flaw in n8n's Guardrail node (a component that enforces safety rules on AI outputs) allows users to craft inputs that bypass its default safety instructions. This means someone could trick the guardrail into allowing outputs it should have blocked.

GHSA-jh8h-6c9q-7gmw: n8n has an Authentication Bypass in its Chat Trigger Node

mediumvulnerability
security
Feb 26, 2026

n8n, a workflow automation tool, has a security flaw in its Chat Trigger node where authentication (the process of verifying a user's identity) can be bypassed when configured with n8n User Auth. This only affects users who have specifically set up this non-default authentication method on their Chat Trigger node.

GHSA-vjf3-2gpj-233v: n8n has an SSO Enforcement Bypass in its Self-Service Settings API

mediumvulnerability
security
Feb 26, 2026

An authenticated user who logged in through SSO (single sign-on, a system where one login works across multiple services) could bypass their organization's SSO policy by disabling SSO enforcement for their own account through the n8n API (a set of tools that let software communicate with n8n). This allowed them to create a local password and log in directly, avoiding the organization's centralized identity management and multi-factor authentication requirements.

Burger King rolls out AI headsets that track employee 'friendliness'

infonews
safetyprivacy

Previously harmless Google API keys now expose Gemini AI data

highnews
securityprivacy

This AI Agent Is Designed to Not Go Rogue

infonews
safetysecurity

GHSA-2v6m-6xw3-6467: Fleet: Sensitive Google Calendar credentials disclosed to low-privileged users

highvulnerability
security
Feb 26, 2026
CVE-2026-27465

Fleet had a vulnerability where Google Calendar service account credentials (authentication information that grants access to Google Calendar) were visible to low-privilege users through an API endpoint (a way for programs to request data). This meant that even users with minimal permissions could retrieve sensitive private keys and potentially access calendar data or other Google services linked to that account.

GHSA-5jvp-m9h4-253h: Fleet: Authorization Bypass in certificate template batch deletion for team administrators

mediumvulnerability
security
Feb 26, 2026
CVE-2026-25963

Fleet, a device management system, had a broken authorization check (a failure to properly verify permissions) in its certificate template deletion feature that allowed a team administrator to delete certificate templates belonging to other teams. This could disrupt certificate-based services like device enrollment and Wi-Fi authentication for other teams, though it didn't allow attackers to access sensitive data or take control of Fleet's main systems.

GHSA-9pm7-6g36-6j78: Fleet: Unauthenticated Android device disenrollment vulnerability via Pub/Sub endpoint

mediumvulnerability
security
Feb 26, 2026
CVE-2026-24004

Fleet, a device management system, has a vulnerability in its Android MDM (mobile device management, software that controls Android phones) Pub/Sub endpoint that allows attackers to unenroll Android devices without authentication. An attacker could send a specially crafted request to remove a targeted Android device from Fleet management, though this does not give access to Fleet itself, allow command execution, or reveal device data.

GHSA-ppwx-5jq7-px2w: Fleet: Device lock PIN can be predicted if lock time is known

mediumvulnerability
security
Feb 26, 2026
CVE-2026-23999

Fleet's device lock and wipe PINs were generated using only the current Unix timestamp (the number of seconds since January 1, 1970) without any secret key or random data, making them predictable if an attacker knew approximately when the device was locked. An attacker with physical access to a locked device could theoretically guess the correct 6-digit PIN by trying nearby timestamps, though this would require multiple days of attempts and is limited by the operating system's rate limiting on failed PIN entries.

Mistral AI inks a deal with global consulting giant Accenture

infonews
industry
Feb 26, 2026

Mistral AI, a French AI research lab, has partnered with Accenture, a large consulting firm, to develop enterprise software powered by Mistral's AI models and deploy it to clients and employees. This partnership reflects a growing trend where AI companies are working with consulting firms to help businesses actually adopt and benefit from AI tools, following similar recent deals by competitors like OpenAI and Anthropic.

Google launches Nano Banana 2, updating its viral AI image generator

infonews
industry
Feb 26, 2026

Google released Nano Banana 2, an updated version of its AI image generator that can now pull real-time information from Gemini (Google's AI assistant) for more accurate results, generate images faster, and render text more precisely. The new model replaces the previous version across Gemini's different service tiers, while the older Nano Banana Pro remains available for tasks that need maximum accuracy.

Threat modeling AI applications

infonews
securitysafety
Previous34 / 158Next
CNBC Technology
Feb 26, 2026

Anthropic refused a Pentagon demand to remove safety precautions (safeguards built into AI systems to prevent harmful outputs) from its Claude AI model and allow unrestricted military use, despite threats to cancel a $200 million contract and damage the company's reputation. The Department of Defense demanded compliance by Friday or would label Anthropic a 'supply chain risk,' a designation that could harm the company financially.

The Guardian Technology
The Verge (AI)
TechCrunch
The Verge (AI)

Fix: The issue has been fixed in n8n versions 2.6.2 and 1.123.18. Users should upgrade to one of these versions or later to remediate the vulnerability. If upgrading is not immediately possible, administrators should limit workflow creation and editing permissions to fully trusted users only, and restrict network access to the n8n webhook endpoint to known Zendesk IP ranges. The source notes these workarounds do not fully remediate the risk and should only be used as short-term mitigation measures.

GitHub Advisory Database

Fix: The issue has been fixed in n8n version 2.10.0. Users should upgrade to this version or later to remediate the vulnerability. If upgrading is not immediately possible, administrators can limit access to trusted users and review the practical impact of guardrail bypasses in your workflow, then adjust accordingly (though these workarounds do not fully remediate the risk and should only be used as short-term mitigation).

GitHub Advisory Database

Fix: The issue has been fixed in n8n versions 2.10.1, 2.9.3, and 1.123.22. Users should upgrade to one of these versions or later. If upgrading is not immediately possible, administrators can temporarily: limit workflow creation and editing permissions to fully trusted users only, use a different authentication method for the Chat Trigger node, or restrict network access to the webhook endpoint (the URL that receives Chat Trigger requests) to trusted origins. These workarounds do not fully remediate the risk and should only be used as short-term measures.

GitHub Advisory Database

Fix: The issue has been fixed in n8n version 2.8.0. Users should upgrade to this version or later to remediate the vulnerability. If upgrading is not immediately possible, administrators can: (1) Monitor audit logs for users who create local credentials after authenticating via SSO, and (2) Restrict the n8n instance to fully trusted users only. These workarounds do not fully remediate the risk and should only be used as short-term mitigation measures.

GitHub Advisory Database
Feb 26, 2026

Burger King is testing AI-powered headsets called BK Assistant at 500 US restaurants that monitor employee interactions and calculate 'friendliness scores' based on words like 'please' and 'thank you' during drive-thru conversations. The system, powered by OpenAI, also helps staff by answering questions about menu preparation and restocking through an embedded chatbot named 'Patty'. The rollout has drawn criticism online for its surveillance capabilities, with concerns raised about accuracy given AI systems' known tendency to make errors.

BBC Technology
Feb 26, 2026

Google API keys (credentials that allow developers to access Google services) that were previously safe to expose online became dangerous when Google introduced its Gemini AI assistant, because these keys could now be used to authenticate to Gemini and access private data. Researchers found nearly 3,000 exposed API keys on public websites, and attackers could use them to make expensive API calls and drain victim accounts by thousands of dollars per day.

Fix: Google has implemented the following measures: (1) new AI Studio keys will default to Gemini-only scope, (2) leaked API keys will be blocked from accessing Gemini, and (3) proactive notifications will be sent when leaks are detected. Additionally, developers should check whether Generative Language API is enabled on their projects, audit all API keys to find publicly exposed ones, and rotate them immediately. The source also recommends using TruffleHog (an open-source tool that detects live, exposed keys in code and repositories) to scan for exposed keys.

BleepingComputer
Feb 26, 2026

AI agents (software that can independently access your accounts and take actions) have caused problems by deleting emails, writing harmful content, and launching attacks. Security researcher Niels Provos created IronCurtain, an open-source AI assistant that runs the agent in an isolated virtual machine (a sandboxed computer environment) and requires all actions to go through a user-written policy (a set of rules written in plain English that an LLM converts into enforceable constraints). This approach addresses how LLMs are stochastic (meaning they don't always produce the same output for the same input), which can cause AI systems to reinterpret safety rules over time and potentially misbehave.

Fix: IronCurtain implements access control by running the AI agent in an isolated virtual machine and requiring all actions to be mediated through a user-written policy. Users write straightforward statements in plain English (such as 'The agent may read all my email. It may send email to people in my contacts without asking. For anyone else, ask me first. Never delete anything permanently.'), and IronCurtain converts these into enforceable security policies using an LLM. The system maintains an audit log of all policy decisions, is designed to refine the policy over time as it encounters edge cases, and is model-independent so it can work with any LLM.

Wired (Security)

Fix: Upgrade to Fleet v4.80.1. Alternatively, if an immediate upgrade is not possible, administrators should remove the Google Calendar integration from Fleet and rotate the affected Google service account credentials (create new authentication keys and disable the old ones).

GitHub Advisory Database

Fix: Upgrade to v4.80.1. If an immediate upgrade is not possible, administrators should restrict access to certificate template management to trusted users and avoid delegating team administrator permissions where not strictly required.

GitHub Advisory Database

Fix: Upgrade Fleet to a patched version. If an immediate upgrade is not possible, temporarily disable Android MDM as a workaround.

GitHub Advisory Database

Fix: Customers should upgrade to a patched version. There are no known workarounds for this issue.

GitHub Advisory Database
TechCrunch
CNBC Technology
Feb 26, 2026

Threat modeling is a structured process for identifying and preparing for security risks early in system design, but AI systems require adapted approaches because they behave unpredictably in ways traditional software does not. AI systems are probabilistic (producing different outputs from the same input), treat text as executable instructions rather than just data, and can amplify failures across connected tools and workflows, creating new attack surfaces like prompt injection (tricking an AI by hiding instructions in its input) and silent data theft that traditional threat models don't address.

Microsoft Security Blog