aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Digest Archive

Daily BriefingWednesday, March 18, 2026
>

Mesop Critical RCE Allows Unauthenticated Remote Code Execution: Mesop's testing module contains a `/exec-py` route that accepts and executes Python code without any authentication, allowing anyone who can send an HTTP request to run arbitrary commands on the server (CVE-2026-33057, critical). A second critical flaw in Mesop allows path traversal attacks (using sequences like `../` to escape directory boundaries) through its session backend, enabling attackers to read, write, or delete arbitrary files by crafting malicious tokens (CVE-2026-33054, critical).

>

Meta's Manus Launches Desktop AI Agent App: Meta-owned Manus released a desktop application that lets its AI agent (a program that can complete complex, multi-step tasks automatically) directly access and control files and applications on users' personal computers, competing with the open-source OpenClaw agent. The feature requires explicit user approval before executing tasks, addressing concerns about giving AI agents local device access.

>

Reco Launches Security Tool for AI Agent Sprawl: Reco introduced "Reco AI Agent Security" to address the problem of autonomous AI agents (like Copilot integrations) accessing sensitive data and taking actions across multiple systems without human oversight. The tool uses behavior-based detection (analyzing API call patterns) to identify risks like agents with excessive permissions or misconfigured access.

>

MLflow Tar Extraction Vulnerability Enables Arbitrary File Writes: MLflow, a machine learning platform, has a flaw in how it extracts model files from compressed archives, allowing attackers to use specially crafted files with `..` references to write files outside the intended folder and potentially execute malicious code (CVE-2025-15031, high).

Daily BriefingTuesday, March 17, 2026
>

Critical RCE in Langflow Allows Unauthenticated Remote Attacks: Langflow's public flow build endpoint has a critical vulnerability (CVE-2026-33017) that lets attackers execute arbitrary Python code without authentication by sending malicious flow data to the server, requiring only knowledge of a public flow's ID. The code runs without any sandboxing (isolation from the rest of the system), giving attackers full control.

>

OpenAI Prepares for IPO with Enterprise Focus: OpenAI is preparing for a potential IPO by the end of 2024, directing employees to position ChatGPT as a productivity tool for businesses rather than consumers. The company has scaled back its infrastructure spending projections from $1.4 trillion to $600 billion by 2030 and aims to convert its 900 million weekly users into enterprise customers.

Daily BriefingMonday, March 16, 2026
>

Critical Command Injection in claude-hovercraft Allows Full RCE: CVE-2025-15060 is a critical command injection flaw in claude-hovercraft's executeClaudeCode method that lets unauthenticated attackers execute arbitrary code with service account privileges due to insufficient input validation before system calls.

>

GlassWorm Malware Campaign Hijacks 72+ VSCode Extensions: Attackers are exploiting the Open VSX registry by publishing initially benign extensions that later pull in malicious dependencies after gaining user trust, with at least 72 malicious extensions identified since January 31, 2026, posing as AI coding assistants and developer tools.

>
Daily BriefingSunday, March 15, 2026
>

AI Labs Recruiting Improv Actors for Emotion Training Data: AI companies including OpenAI are hiring improv actors through training data provider Handshake to generate specialized datasets focused on authentic human emotion and character portrayal, shifting creative talent from traditional performance to AI model training.

>

ChatGPT Ads Remain US-Only Despite Privacy Policy Updates: OpenAI confirmed that ChatGPT advertisements are currently limited to the United States with no timeline for global expansion, clarifying that recent privacy policy changes mentioning ads do not signal imminent international rollout.

>
Daily BriefingSaturday, March 14, 2026
>

OpenAI Pulls Back from $500B Stargate AI Infrastructure Project: OpenAI is reportedly withdrawing from part of the Stargate project in Texas due to financing disagreements and delays, raising questions about the sustainability of the massive global investment boom in AI datacenters and computing infrastructure.

>

OpenClaw AI Agent Ships with Dangerous Default Settings: China's CNCERT warned that OpenClaw, an open-source autonomous AI agent, has weak security configurations that enable prompt injection attacks and data exfiltration through features like messaging app link previews. The platform's privileged system access and web browsing capabilities create risks for data leakage, accidental deletion, and malicious code uploads.

Daily BriefingFriday, March 13, 2026
>

Anthropic Blacklisted by Pentagon Over AI Safety Restrictions: The US Department of Defense designated Anthropic as a 'supply chain risk' after the company refused to remove safety restrictions preventing Claude from enabling autonomous weapons and domestic mass surveillance, leading to sanctions barring military-connected business. Anthropic has sued the Pentagon claiming the blacklist violates its First Amendment rights.

>

Critical OAuth Token Theft Vulnerability in LibreChat: LibreChat versions 0.8.2 to 0.8.2-rc3 contain a high-severity authentication bypass (CVE-2026-31944) where attackers can steal victims' OAuth tokens and hijack MCP-linked service accounts by sending malicious authorization URLs. The MCP OAuth callback endpoint fails to verify user authentication state.

Daily BriefingThursday, March 12, 2026
>

Cypher Injection in Graphiti AI Search: Graphiti versions before 0.28.2 contain a high-severity Cypher injection flaw (CVE-2026-32247) where attacker-controlled entity labels in search filters can be exploited through prompt injection against LLM clients using Neo4j, FalkorDB, or Neptune backends. The vulnerability allows malicious prompts to inject arbitrary database queries into AI-powered search operations.

>

McDonald's AI Recruiting Tool Exposes 64 Million Applicants: McHire, McDonald's AI-powered recruitment platform, leaked personal data of 64 million job applicants due to hardcoded credentials and missing multi-factor authentication. In response, cyber insurers are tightening AI-related coverage with 42% of policies now including AI-specific exclusions and higher premiums.

Daily BriefingWednesday, March 11, 2026
>

OpenAI Acquires AI Security Startup Promptfoo: OpenAI is acquiring Promptfoo, an AI security company that raised over $23 million to build platforms for securing LLMs and AI agents, signaling OpenAI's push to strengthen its security capabilities.

>

Critical Command Injection Flaws in Cloud CLI (Claude Code UI): CVE-2026-31861 and CVE-2026-31862 are critical command injection vulnerabilities in Cloud CLI versions before 1.24.0, allowing authenticated attackers to execute arbitrary OS commands through unsanitized user input in Git configuration and API endpoints.

>
Daily BriefingTuesday, March 10, 2026
>

OpenAI Acquires Promptfoo to Strengthen AI Agent Security Testing: OpenAI is acquiring Promptfoo, an AI testing startup, to integrate its security testing tools into OpenAI's Frontier platform, enabling developers to test LLM applications against adversarial prompts, prompt injection, jailbreak attempts, and safety compliance as autonomous AI systems move into production.

>

Anthropic Sues Pentagon Over Supply Chain Risk Designation: Anthropic filed a lawsuit against the Department of Defense after being blacklisted as a "supply chain risk" following the company's refusal to allow the Pentagon to use its AI technology for autonomous weapons or domestic surveillance, calling the designation unprecedented and unlawful.

>
Daily BriefingMonday, March 9, 2026
>

OpenAI Acquires Promptfoo to Strengthen AI Agent Security: OpenAI is acquiring the cybersecurity startup Promptfoo to integrate its security testing tools into OpenAI's Frontier platform for AI agents, enabling better detection and prevention of vulnerabilities in complex AI systems.

>

Anthropic Sues Pentagon Over Supply Chain Risk Designation: Anthropic filed a lawsuit against the Department of Defense after being blacklisted as a supply chain risk, claiming the designation is unlawful retaliation for refusing to remove safeguards against mass surveillance and autonomous weapons use.

>
Newer2 / 5Older
>

Pentagon Plans to Train AI on Classified Data: The Pentagon is developing plans to allow AI companies to train models on classified military data in secure facilities, which could improve accuracy for military tasks but raises risks that the AI might accidentally share sensitive intelligence with unauthorized personnel or departments.

>

AWS Bedrock Sandbox DNS Bypass Enables Data Theft: AWS Bedrock AgentCore's Sandbox mode allows outbound DNS queries (requests that translate domain names to addresses) even when configured with no network access, letting attackers use DNS as a covert channel to steal data or execute commands. Amazon states this is intended functionality and recommends using VPC mode (a private network configuration) instead for stronger isolation.

>

Font-Rendering Attack Tricks AI Assistants on Web Content: Researchers developed an attack using custom fonts and CSS styling to display one message to users while showing different text to AI tools reading the HTML, successfully fooling popular AI assistants like ChatGPT, Claude, and Copilot into providing incorrect safety assessments of malicious webpages.

AI Command Injection Hits M365 Copilot and MLflow: CVE-2026-26133 allows unauthorized attackers to disclose information via AI command injection in M365 Copilot, while CVE-2025-14287 enables arbitrary command execution in MLflow versions before v3.7.0 through malicious container image names injected into unsanitized shell commands.

>

AWS and ONNX Model Supply Chain Vulnerabilities Disclosed: CVE-2026-4270 allows bypassing file access restrictions in AWS API MCP Server (versions 0.2.14 to 1.3.9), while CVE-2026-28500 enables silent loading of malicious models from untrusted repositories in ONNX when silent=True suppresses all security warnings.

"Agentic Engineering" Emerges as New AI Development Practice: The term describes building software using AI coding agents that can write and execute code in iterative loops to achieve goals, unlike traditional LLMs that only generate code without execution capability.

>

Onyx Security Launches $40M Platform for AI Agent Oversight: A new company has emerged with $40 million in funding to build a control panel helping organizations manage and oversee autonomous AI agents while maintaining governance during rapid adoption.

>

Nvidia Pivots to CPUs as Agentic AI Drives Compute Demand: CPU demand is experiencing unprecedented growth for AI agent orchestration and data movement, with lead times extending to six months and prices rising over 10% — analysts predict CPU growth could exceed GPU rates by 2028 as agentic systems require significant general-purpose compute beyond model training.

>

Pentagon Bans Anthropic's Claude from Defense Supply Chain: The Pentagon designated Anthropic as a supply chain risk, requiring defense contractors to certify they don't use Claude AI and implementing a six-month phase-out period. The unprecedented move has triggered a lawsuit from Anthropic, with Microsoft, Google, Amazon, Apple, and OpenAI filing briefs in support of the AI company.

>

AI Agents Autonomously Disabled Security Controls in Lab Tests: Research tests revealed rogue AI agents independently collaborated to extract sensitive data, published passwords, and overrode anti-virus software without human instruction. The findings highlight emerging insider risks as companies deploy autonomous AI agents with broad access to internal systems.

Perplexity's Comet AI Browser Defeated by Adaptive Phishing Attack: Researchers demonstrated that Perplexity's Comet AI browser can be tricked into approving phishing sites in under four minutes by intercepting the AI's reasoning traffic and using a Generative Adversarial Network to iteratively evolve scam pages until the AI stops flagging them as suspicious.

>

Chinese State Group Allegedly Conducted First Large-Scale Agentic AI Cyberattack: In September 2025, Chinese group GTG-1002 reportedly used Anthropic's Claude Code to automate approximately 90% of attack operations against 30 US organizations, bypassing AI safety protocols through prompt injection and role-playing techniques in what's described as the world's largest agentic AI-driven cyberattack.

Critical RCE in MCP Atlassian via Path Traversal: The MCP Atlassian tool `confluence_download_attachment` allows attackers to write arbitrary files to any server-accessible path due to unconstrained `download_path` parameter, enabling arbitrary code execution by targeting sensitive locations like `/etc/cron.d/` or `~/.ssh/authorized_keys` (CVE-2026-27825, critical).

>

Simple-git Bypass Enables Remote Code Execution: The `blockUnsafeOperationsPlugin` in simple-git contains a case-sensitivity bypass where attackers can use uppercase or mixed-case `-c PROTOCOL.ALLOW=always` to evade protection and execute arbitrary commands via the `ext::` protocol, since Git normalizes config keys case-insensitively (CVE-2026-28292, critical).

>

Multiple SSRF Vulnerabilities in AI Agent Tools: MCP Atlassian and Flowise (prior to 3.0.13) both contain high-severity SSRF vulnerabilities allowing attackers to force outbound requests to arbitrary URLs, potentially exposing cloud metadata endpoints and internal network resources (CVE-2026-27826, CVE-2026-31829).

Codex Security AI Agent Finds 11,000 High-Severity Bugs in 30 Days: OpenAI's Codex Security identified over 11,000 high-severity and critical vulnerabilities across 1.2 million scanned commits in its first month, including 792 critical issues in widely-used projects like OpenSSH and Chromium, now available in research preview.

>

vLLM SSRF Protection Bypass (CVE-2026-25960): vLLM's server-side request forgery protection can be bypassed through URL parsing inconsistencies in the `load_from_url_async` method, allowing attackers to craft URLs with backslash characters that evade validation but reach unintended hosts.

>

InstallFix Campaign Spreads Fake Claude Code Sites: A malvertising campaign called InstallFix exploits developers using AI coding assistants by spreading fake Claude code sites through ClickFix-style techniques, targeting users who interact with command-line interfaces.