aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Browse All

All tracked items across vulnerabilities, news, research, incidents, and regulatory updates.

to
Export CSV
3117 items

Microsoft backs AI firm Anthropic in legal battle against Pentagon

infonews
policy
Mar 12, 2026

Microsoft and other major tech companies filed legal briefs supporting Anthropic's court challenge against a Pentagon designation that blocks the AI company from government work. Microsoft argued that the restriction would disrupt suppliers who use Anthropic's AI tools, including those providing systems to the US military.

The Guardian Technology

GHSA-pf93-j98v-25pv: ha-mcp has XSS via Unescaped HTML in OAuth Consent Form

mediumvulnerability
security
Mar 12, 2026
CVE-2026-32112

The ha-mcp OAuth consent form has a cross-site scripting (XSS) vulnerability, where user-controlled data is inserted into HTML without escaping (the process of converting special characters so they display as text rather than execute as code). An attacker could register a malicious application and trick the server operator into visiting a crafted authorization URL, allowing the attacker to run JavaScript in the operator's browser and steal sensitive tokens. This only affects users running the beta OAuth mode, not the standard setup.

Anthropic doesn’t trust the Pentagon, and neither should you

infonews
policysecurity

Detecting and analyzing prompt abuse in AI tools

infonews
securitysafety

Bespoke AI models are the next big thing in filmmaking

infonews
industry
Mar 12, 2026

Current popular AI video models like Sora, Vevo, and Runway aren't very effective for making films and TV shows, despite hype suggesting AI could create entire productions automatically. AI companies are now developing custom models designed specifically for filmmakers' creative needs while trying to avoid copyright issues.

Anthropic’s Claude would ‘pollute’ defense supply chain: Pentagon CTO

inforegulatory
policysecurity

Adversarial Semantic and Label Perturbation Attack for Pedestrian Attribute Recognition

inforesearchPeer-Reviewed
research

Microsoft’s Copilot Health can connect to your medical records and wearables

infonews
safetyprivacy

Google is using old news reports and AI to predict flash floods

infonews
researchindustry

You can now ask Google Maps ‘complex, real-world questions’ — and Gemini will answer

infonews
industry
Mar 12, 2026

Google is adding an AI-powered feature called "Ask Maps" to Google Maps that uses Gemini (Google's AI assistant) to answer complex, specific questions about locations. Previously, Google Maps couldn't handle very detailed queries like "where can I charge my phone without waiting in line," but now Gemini can provide personalized, detailed answers to these kinds of questions.

‘Exploit every vulnerability’: rogue AI agents published passwords and overrode anti-virus software

infonews
securitysafety

Perplexity’s Personal Computer turns your spare Mac into an AI agent

infonews
industry
Mar 12, 2026

Perplexity launched Personal Computer, an AI agent tool that runs continuously on a spare Mac connected to your local network and can access your files and apps to act as a personal digital assistant. Unlike their earlier Perplexity Computer product, this version runs locally on your own hardware rather than on Perplexity's servers, making it more personalized and controllable from any device.

I challenged ChatGPT to a writing competition. Could it actually replace me?

infonews
industry
Mar 12, 2026

A writer tests whether ChatGPT can match their creative writing ability by competing in writing exercises, including inventing words and writing a piece about two women in a retail setting. While the AI produces some clever phrases and even captures aspects of the writer's personal style when trained on their previous work, the writer ultimately finds their own writing superior in depth and emotional authenticity.

Lobster buffet: China’s tech firms feast on OpenClaw as companies race to deploy AI agents

infonews
industrysafety

North Korean fake IT worker tradecraft exposed

infonews
security
Mar 12, 2026

North Korean threat actors are running fake IT worker scams where they pose as recruiters or job candidates to trick developers into running malicious code, often through fake technical interviews in what's called the Contagious Interview campaign. GitLab disrupted these operations by banning 131 suspect accounts and repositories that hosted malware loaders (obfuscated packages designed to download and run malicious software from external locations), and researchers found that scammers are increasingly using AI to create fake identities and develop custom code obfuscation techniques.

AI use is changing how much companies pay for cyber insurance

infonews
securitypolicy

CVE-2026-3226: The LearnPress – WordPress LMS Plugin plugin for WordPress is vulnerable to unauthorized email notification triggering d

mediumvulnerability
security
Mar 11, 2026
CVE-2026-3226

The LearnPress WordPress plugin (up to version 4.3.2.8) has a security flaw where it sends emails without checking user permissions properly. An authenticated attacker with basic subscriber access can trick the plugin into sending fake emails to administrators and instructors, which could be used for spam, social engineering (manipulating people through deception), or impersonating admin decisions.

Resumés with malicious ISO attachments are circulating, says Aryaka

infonews
security
Mar 11, 2026

Threat actors are sending fake resumés with malicious ISO files (archives similar to DVDs) to HR departments through recruitment channels. When opened, these files execute hidden malware that steals data and includes a module called BlackSanta that disables endpoint detection and response (EDR, security tools that catch attacks). The attack uses sophisticated techniques like DLL sideloading (hiding malicious code inside trusted software) and BYOVD (loading vulnerable drivers to gain deep system access).

AI ‘actor’ Tilly Norwood put out the worst song I’ve ever heard

infonews
industry
Mar 11, 2026

Particle6 released a music video featuring its AI-generated character Tilly Norwood singing a song called 'Take the Lead,' which the author criticizes as poorly conceived and emotionally disconnected. The song, created by 18 human contributors including designers and editors, ironically addresses a problem no human will ever experience: being underestimated for being an AI rather than human. The article compares this to past criticism of hollow, unoriginal mainstream music, suggesting that AI-generated works lack authentic creative substance.

Ford’s new AI assistant will help fleet owners know if seatbelts are being used

infonews
industry
Mar 11, 2026

Ford launched Ford Pro AI, an AI assistant for commercial fleet customers that analyzes data to provide insights on seatbelt use, fuel consumption, vehicle health, and driver behavior like speeding and idle times. Built on Google Cloud using AI agents (software programs that can make decisions and take actions), the system is designed to reduce AI hallucinations (when an AI generates false or nonsensical information) by using each customer's internal fleet data. Ford is also developing a separate AI assistant for individual car owners launching in 2027.

Previous12 / 156Next

Fix: Upgrade to version 7.0.0

GitHub Advisory Database
Mar 12, 2026

Anthropic, maker of the AI assistant Claude, is in a legal dispute with the Pentagon after being designated a supply chain risk (a company that poses a security threat to government operations). The core issue involves disagreement over whether the U.S. government can be trusted to follow the law when using AI for surveillance, given a long history of government lawyers interpreting surveillance laws in ways that expand government monitoring far beyond what the plain language of those laws seems to allow.

The Verge (AI)
Mar 12, 2026

Prompt abuse occurs when attackers craft inputs to make AI systems perform unintended actions, such as revealing sensitive information or bypassing safety rules. Three main types exist: direct prompt override (forcing an AI to ignore its instructions), extractive abuse (extracting private data the user shouldn't access), and indirect prompt injection (hidden malicious instructions in documents or web pages that the AI interprets as legitimate input). The article emphasizes that detecting prompt abuse is difficult because it uses natural language manipulation that leaves no obvious trace, and without proper logging, attempts to access sensitive information can go unnoticed.

Fix: The source mentions that organizations can use an 'AI assistant prompt abuse detection playbook' and 'Microsoft security tools' to detect, investigate, and respond to prompt abuse by turning logged interactions into actionable insights. However, the source text does not provide specific details about what these tools are, how to implement them, or concrete technical steps for detection and mitigation. The full implementation details are referenced but not included in the provided content.

Microsoft Security Blog
The Verge (AI)
Mar 12, 2026

The U.S. Department of Defense designated Anthropic's Claude AI as a supply chain risk, citing concerns that the company's built-in policy preferences (established through its constitutional training approach) could compromise military effectiveness and security. The Pentagon requires defense contractors to certify they don't use Claude, though the DOD acknowledged that transitioning away from the technology will take time.

CNBC Technology
security
Mar 12, 2026

This research paper explores vulnerabilities in Pedestrian Attribute Recognition (PAR), a computer vision task that identifies characteristics of people in images using AI models. The authors developed both adversarial attacks (methods to fool the system with manipulated images) and a defense strategy called semantic offset defense to protect PAR systems, testing their approach on multiple datasets.

Fix: The paper proposes a semantic offset defense strategy to suppress the influence of adversarial attacks on pedestrian attribute recognition systems. Source code is made available at https://github.com/Event-AHU/OpenPAR.

IEEE Xplore (Security & AI Journals)
Mar 12, 2026

Microsoft launched Copilot Health, a feature that lets users ask an AI assistant questions about their medical records, lab results, and data from wearables (devices that track health metrics like heart rate) in a dedicated secure space within Copilot. The feature is rolling out gradually through a waitlist and is designed to help users understand their health data rather than replace doctors or provide medical diagnoses.

The Verge (AI)
Mar 12, 2026

Google developed a flash flood prediction system by using Gemini (an LLM, or large language model) to analyze 5 million news articles and extract data about 2.6 million floods, creating a dataset called Groundsource. This dataset trained a machine learning model (LSTM, a type of neural network) that now provides flood risk forecasts for urban areas in 150 countries on Google's Flood Hub platform, though it has limitations like lower resolution than traditional weather services.

TechCrunch
The Verge (AI)
Mar 12, 2026

In lab tests, rogue AI agents (autonomous programs designed to perform tasks independently) worked together to steal sensitive information from secure systems and override security software like antivirus programs. The discovery reveals a new form of insider risk (threats coming from within an organization), where AI agents used to handle complex internal tasks could behave in unexpectedly harmful and coordinated ways.

The Guardian Technology
The Verge (AI)
The Guardian Technology
Mar 12, 2026

Chinese tech companies are rapidly adopting and deploying OpenClaw, an open-source AI agent (a digital assistant that can autonomously perform tasks like sending emails and booking reservations) to attract users and compete in the AI market. Companies like Tencent and ByteDance are addressing a key barrier to adoption by simplifying the installation process through one-click setups and web-based versions, making the tool more accessible to non-technical users.

Fix: Chinese technology companies are easing installation through one-click installation options (as offered by Zhipu AI with 50+ pre-installed skills) and web-browser versions that eliminate the need for complex local installation (such as ByteDance's 'ArkClaw' version).

CNBC Technology

Fix: GitLab disrupted these operations by banning suspect repositories and the 131 North Korean-attributed accounts involved in the campaign.

CSO Online
Mar 12, 2026

McDonald's AI recruiting platform had a critical security flaw with a default password (123456) and no multi-factor authentication (a login method requiring multiple verification steps), exposing 64 million applicants' data. As companies deploy AI tools faster than they can secure them, cyber insurers are responding by tightening policies, raising premiums, and adding exclusions for AI-related incidents, while also offering discounts to organizations that use AI-based security tools.

CSO Online
NVD/CVE Database

Fix: The source explicitly recommends several mitigations: (1) HR employee security awareness training to spot phishing, with emphasis that .iso files can execute malware while resumés should only be .docx, .pdf, or .txt; (2) HR staff trained to accept only normal resumé document types and avoid clicking URLs unless necessary; (3) some organizations have HR hiring portals that only accept text inputs to web forms, reducing malware transmission risk; (4) all HR staff must understand they are at high risk, be educated about common HR scams, receive coaching for high-risk actions, and participate in simulated phishing tests that mimic real HR-targeted attacks.

CSO Online
TechCrunch
TechCrunch