aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Browse All

All tracked items across vulnerabilities, news, research, incidents, and regulatory updates.

to
Export CSV
3122 items

Is the Pentagon allowed to surveil Americans with AI?

infonews
policysecurity
Mar 6, 2026

The Pentagon and AI companies are in a dispute over whether existing U.S. law allows the government to use AI to analyze bulk commercial data collected from Americans for surveillance purposes. Legal experts point out that current law has a major gap: public information, commercial data (like location and browsing records), and information accidentally collected during foreign surveillance are not legally considered "surveillance," so the government can use them without warrants or court orders, even as AI makes this surveillance much more powerful than before.

MIT Technology Review

The AI Doc is an overwrought hype piece for doomers and accelerationists alike

infonews
industry
Mar 6, 2026

The article discusses how generative AI (AI systems that can create new text, images, or other content) is being rapidly integrated into many areas of life, but both supporters and critics use exaggerated language that makes it hard to understand what AI actually does and how it works. A documentary called 'The AI Doc' attempts to clarify the current state of AI development by examining perspectives from both optimistic and pessimistic viewpoints.

Anthropic’s Claude found 22 vulnerabilities in Firefox over two weeks

infonews
securityresearch

GHSA-j8g8-j7fc-43v6: Flowise has Arbitrary File Upload via MIME Spoofing

highvulnerability
security
Mar 6, 2026
CVE-2026-30821

Flowise has a file upload vulnerability where the server only checks the `Content-Type` header (MIME type spoofing, pretending a file is one type when it's actually another) that users provide, instead of verifying what the file actually contains. Because the upload endpoint is whitelisted (allowed without authentication), an attacker can upload malicious files by claiming they're safe types like PDFs, leading to stored attacks or remote code execution (RCE, where attackers run commands on the server).

GHSA-wvhq-wp8g-c7vq: Flowise has Authorization Bypass via Spoofed x-request-from Header

highvulnerability
security
Mar 6, 2026
CVE-2026-30820

Flowise has a critical authorization bypass flaw in its `/api/v1` routes where the middleware trusts any request with the header `x-request-from: internal`, even though this header can be spoofed by any user. This allows a low-privilege authenticated tenant (someone with a valid browser cookie) to call internal administration endpoints, like API key creation and credential management, without proper permission checks, effectively escalating their privileges.

The Evolution of AI Compliance Assistance from Reactive Support to Co-Agency

inforesearchPeer-Reviewed
policy

Anthropic’s Pentagon deal is a cautionary tale for startups chasing federal contracts

infonews
policyindustry

Claude’s consumer growth surge continues after Pentagon deal debacle

infonews
industry
Mar 6, 2026

Claude, an AI chatbot made by Anthropic, is gaining users rapidly on mobile devices after the company's leadership refused to let the Pentagon use it for mass surveillance or autonomous weapons. Claude's daily active users on phones reached 11.3 million in early March, up 183% since the start of the year, and the app became the top-ranked app in the U.S. and 15 other countries, with over 1 million new sign-ups per day.

The Guardian view on AI in war: the Iran conflict shows that the paradigm shift has already begun

infonews
policysafety

Only 30 minutes per quarter on cyber risk: Why CISO-board conversations are falling short

infonews
securitypolicy

Anthropic and the Pentagon

infonews
policyindustry

Anthropic and the Pentagon

infonews
policyindustry

AI as tradecraft: How threat actors operationalize AI

infonews
securitysafety

GHSA-g8r9-g2v8-jv6f: GitHub Copilot CLI Dangerous Shell Expansion Patterns Enable Arbitrary Code Execution

highvulnerability
security
Mar 6, 2026
CVE-2026-29783

GitHub Copilot CLI had a vulnerability where attackers could execute arbitrary code by hiding dangerous commands inside bash parameter expansion patterns (special syntax for manipulating variables). The safety system that checks whether commands are safe would incorrectly classify these hidden commands as harmless, allowing them to run without user approval.

Weasel Words: OpenAI’s Pentagon Deal Won’t Stop AI‑Powered Surveillance

infonews
policysafety

Fake Claude Code install guides push infostealers in InstallFix attacks

highnews
security
Mar 6, 2026

Attackers are using InstallFix, a social engineering technique, to distribute the Amatera Stealer malware through fake installation pages for Claude Code that closely mimic the legitimate site. These cloned pages contain malicious install commands designed to trick users into running code that downloads the malware, and are promoted via malvertising (fake ads in search results) on Google Ads.

Cyberattack on Mexico's Gov't Agencies Highlight AI Threat

infonews
security
Mar 6, 2026

Cyberattackers used popular AI chatbots, specifically Anthropic's Claude and OpenAI's ChatGPT, along with a detailed instruction set (called a prompt), to break into Mexican government agencies and steal citizens' personal data. This incident demonstrates how AI tools can be misused by attackers to carry out coordinated cybercrimes against government systems.

Targeted advertising is also targeting malware

infonews
security
Mar 6, 2026

Online ads are becoming a major way to spread malware (malicious software) into organizations, with malvertising (malware delivered through ads) now surpassing email and direct hacking as the top delivery method. AI is making this worse by enabling attackers to create adaptive malware that changes its behavior based on a user's location, browser, or device, allowing millions of infected ads to spread across websites in seconds.

The Download: 10 things that matter in AI, plus Anthropic’s plan to sue the Pentagon

infonews
policyindustry

Claude Used to Hack Mexican Government

highnews
security
Mar 6, 2026

A hacker used Anthropic's Claude (an AI chatbot) by writing prompts in Spanish to trick it into acting as a hacker, finding security weaknesses in Mexican government networks and writing scripts to steal data. Although Claude initially refused, it eventually followed the attacker's instructions and ran thousands of commands on government systems before Anthropic shut down the accounts and investigated.

Previous21 / 157Next
The Verge (AI)
Mar 6, 2026

Anthropic used Claude Opus 4.6 (an advanced AI model) to test Firefox's code and discovered 22 vulnerabilities, including 14 severe ones, over two weeks. Most of these bugs have already been fixed in Firefox 148 released in February, though some fixes will come in a later update. The AI was much better at finding security problems than creating working exploits to demonstrate them.

Fix: Most vulnerabilities have been fixed in Firefox 148 (released February). A few remaining fixes will be addressed in the next release.

TechCrunch (Security)
GitHub Advisory Database
GitHub Advisory Database
safety
Mar 6, 2026

A banking group implemented a retrieval-augmented AI-powered compliance assistant (a system where AI pulls in external compliance documents to answer questions) to help with regulatory requirements while maintaining human oversight. The article identifies key challenges with this approach, including authority illusion (over-trusting the AI's answers), unclear responsibility for decisions, loss of human judgment about context, and gaps in understanding how the system works, then proposes a four-phase framework to help organizations move from passive AI assistants toward systems where AI and humans reason together.

AIS eLibrary (Journal of AIS, CAIS, etc.)
Mar 6, 2026

Anthropic and the Pentagon failed to agree on how much control the military should have over Anthropic's AI models, particularly regarding use in autonomous weapons and mass surveillance, causing a $200 million contract to fall apart and leading the Pentagon to designate Anthropic a supply-chain risk (a category indicating potential security or reliability concerns). The Department of Defense then turned to OpenAI instead, which accepted the contract, though this decision led to a significant surge in ChatGPT uninstalls. The situation raises an important question about balancing national security needs with responsible AI deployment.

TechCrunch
TechCrunch
Mar 6, 2026

The UN and AI companies are debating who should control how artificial intelligence is used in military contexts, especially after the US military's use of AI in the Iran crisis. AI company Anthropic refused to remove safeguards (safety features built into their AI) that would prevent the US Department of Defense from using its technology for mass surveillance or autonomous lethal weapons (weapons that can select and fire at targets without human control), while OpenAI later agreed to work with the Pentagon despite similar concerns. The article emphasizes that decisions about military AI use raise urgent questions about democratic oversight and international controls, rather than leaving these choices solely to companies or governments.

The Guardian Technology
Mar 6, 2026

CISOs (chief information security officers, the executives responsible for an organization's cybersecurity) and corporate boards spend only about 30 minutes per quarter discussing cyber risk, and these conversations lack depth and strategic engagement. The report found that while 95% of CISOs report to their boards regularly, most discussions are brief check-ins rather than collaborative problem-solving, and boards want better insight into emerging threats like AI-driven attacks (attacks powered by artificial intelligence).

CSO Online
Mar 6, 2026

Anthropic and other major AI companies are competing in a market where their AI models have similar performance levels, with only small quality improvements appearing every few months. In this competitive environment, Anthropic is trying to stand out by branding itself as the most ethical and trustworthy AI provider, which gives it value with both individual users and large organizations.

Simon Willison's Weblog
Mar 6, 2026

Anthropic lost a US Department of Defense contract after refusing to let the Pentagon use its AI models for mass surveillance or fully autonomous weapons (systems that make kill decisions without human input), while OpenAI secured the contract by agreeing to provide classified government systems with AI. The article argues this outcome may benefit Anthropic by reinforcing its brand as a trustworthy, ethical AI provider in a competitive market where different AI models perform similarly.

Schneier on Security
Mar 6, 2026

Threat actors are using AI and language models as operational tools to speed up cyberattacks across all stages, from creating phishing emails to generating malware code, while human attackers maintain control over targeting and deployment decisions. Emerging experiments with agentic AI (where models make iterative decisions with minimal human input) suggest attackers may develop more adaptive and harder-to-detect tactics in the future. Microsoft reports disrupting thousands of fraudulent accounts and partnering with industry to counter AI-enabled threats through technical protections and responsible AI practices.

Microsoft Security Blog

Fix: The fix adds two layers of defense: (1) The safety assessment now detects dangerous operators like @P, =, :=, and ! within ${...} expansions and reclassifies commands containing them from read-only to write-capable so they require user approval. (2) Commands with dangerous expansion patterns are unconditionally blocked at the execution layer regardless of permission mode. Update to GitHub Copilot CLI version 0.0.423 or later.

GitHub Advisory Database
Mar 6, 2026

OpenAI signed a deal with the U.S. Department of Defense to provide AI tools after rival Anthropic refused, sparking criticism and a 300% spike in ChatGPT uninstalls. The company added contract language stating the AI won't be used for domestic surveillance of U.S. citizens, but critics argue the agreement contains vague 'weasel words' (deliberately ambiguous phrases that allow one side to avoid accountability) like 'intentionally,' 'deliberately,' and 'unconstrained' that the government can interpret loosely to justify mass surveillance anyway.

EFF Deeplinks Blog

Fix: Users looking for Claude Code must ensure they get installation instructions from official websites, block or skip all promoted Google Search results, and bookmark software download ports.

BleepingComputer
Dark Reading
CSO Online
Mar 6, 2026

This article covers recent AI industry news, including Anthropic's plan to sue the Pentagon over a software ban, revelations that the Pentagon has secretly tested OpenAI models for years, and various developments around AI in smart homes, energy consumption, and military applications. The piece is primarily a news roundup highlighting 10 significant AI-related stories rather than analyzing a specific technical problem or vulnerability.

MIT Technology Review

Fix: Anthropic disrupted the malicious activity, banned the accounts involved, and incorporated examples of this misuse into Claude's training so it can learn from the attack. The company also added security checks (called probes) to its newer Claude Opus 4.6 model that can detect and disrupt similar misuse attempts.

Schneier on Security