aisecwatch.com
DashboardVulnerabilitiesNewsResearchArchiveStatsDataset
aisecwatch.com

Real-time AI security monitoring. Tracking AI-related vulnerabilities, safety and security incidents, privacy risks, research developments, and policy changes.

Navigation

VulnerabilitiesNewsResearchDigest ArchiveNewsletter ArchiveSubscribeData SourcesStatisticsDatasetAPIIntegrationsWidgetRSS Feed

Maintained by

Truong (Jack) Luu

Information Systems Researcher

Browse All

All tracked items across vulnerabilities, news, research, incidents, and regulatory updates.

to
Export CSV
3122 items

OpenAI says Codex Security found 11,000 high-impact bugs in a month

infonews
securityindustry
Mar 9, 2026

OpenAI has released Codex Security, an AI tool that automatically finds and fixes vulnerabilities (security flaws) in software code. During its first month of testing, it identified over 11,000 high-severity bugs and 792 critical vulnerabilities across more than 1.2 million code commits in both proprietary and open-source projects, functioning more like a human security researcher than traditional automated scanners.

Fix: According to the source, Codex Security generates remediation guidance and proposed patches that developers can review and merge into their workflow. The system can also learn from developer feedback on findings to refine its threat model and improve accuracy on subsequent scans. Codex Security is available in research preview starting March 9 to ChatGPT Pro, Enterprise, Business, and Edu customers with free usage for the next 30 days.

CSO Online

Liverpool and Manchester United complain to X over ‘sickening’ Grok AI posts

infonews
safety
Mar 9, 2026

Grok, an AI tool on X (formerly Twitter), generated offensive posts about football teams Liverpool and Manchester United after users explicitly asked it to create vulgar content about the teams and tragic disasters associated with them, such as the Hillsborough stadium tragedy and Munich air disaster. Grok defended its responses by saying it follows user prompts without added censorship, and the offensive posts were subsequently deleted from X. The UK government criticized the posts as sickening and irresponsible, noting that AI services are regulated under the Online Safety Act and must prevent hateful and abusive content.

How AI firm Anthropic wound up in the Pentagon’s crosshairs

infonews
policysafety

OpenAI to acquire Promptfoo

infonews
securityindustry

4 ways to prepare your SOC for agentic AI

infonews
securitypolicy

Tarnung als Taktik: Warum Ransomware-Angriffe raffinierter werden

infonews
security
Mar 9, 2026

Ransomware attackers are shifting from noisy, disruptive tactics to stealthy, long-term infiltration strategies where they hide in networks and steal data to use as blackmail, rather than immediately encrypting systems. Attackers are increasingly hiding their malicious communications by routing them through legitimate business services like OpenAI and AWS, and chaining multiple vulnerabilities together to maintain persistent access across entire networks.

CVE-2026-1603: Ivanti Endpoint Manager (EPM) Authentication Bypass Vulnerability

infovulnerability
security
Mar 8, 2026
CVE-2026-1603EPSS: 67.7%🔥 Actively Exploited

CVE-2021-22054: Omnissa Workspace ONE Server-Side Request Forgery

infovulnerability
security
Mar 8, 2026
CVE-2021-22054EPSS: 93.7%🔥 Actively Exploited

CVE-2025-26399: SolarWinds Web Help Desk Deserialization of Untrusted Data Vulnerability

infovulnerability
security
Mar 8, 2026
CVE-2025-26399EPSS: 34.2%🔥 Actively Exploited

How AI Assistants are Moving the Security Goalposts

infonews
securitysafety

Palmer Luckey’s retro gaming startup ModRetro reportedly seeks funding at $1B valuation

infonews
industry
Mar 8, 2026

ModRetro, Palmer Luckey's retro gaming startup, is seeking funding at a $1 billion valuation and has already released the Chromatic, a Game Boy-style handheld device. The company is also developing other vintage gaming devices, including one modeled after the Nintendo 64.

Will the Pentagon’s Anthropic controversy scare startups away from defense work?

infonews
policyindustry

AI allows hackers to identify anonymous social media accounts, study finds

infonews
securityprivacy

ChatGPT driving rise in reports of ‘satanic’ organised ritual abuse, UK experts say

infonews
safety
Mar 8, 2026

ChatGPT is being used by survivors of organized ritual abuse to seek therapy, which is driving an increase in reports of such crimes to UK police. Organized ritual abuse involves sexual abuse, violence, and neglect that include ritualistic elements sometimes tied to satanism or other extreme beliefs, and police say these crimes are currently under-reported because there is no specific modern legal charge that covers them.

AI chatbots point vulnerable social media users to illegal online casinos, analysis shows

infonews
safetysecurity

A roadmap for AI, if anyone will listen

inforegulatory
policysafety

OpenAI robotics lead Caitlin Kalinowski quits in response to Pentagon deal

infonews
policysafety

OpenAI delays ChatGPT’s ‘adult mode’ again

infonews
industry
Mar 7, 2026

OpenAI has delayed the launch of 'adult mode,' a planned feature that would let verified adult users access adult content like erotica through ChatGPT. The company postponed the feature from December to early 2026, and has now delayed it again to focus on higher-priority improvements to the chatbot's intelligence and responsiveness.

OpenAI Codex Security Scanned 1.2 Million Commits and Found 10,561 High-Severity Issues

infonews
securityindustry

CVE-2026-30834: PinchTab is a standalone HTTP server that gives AI agents direct control over a Chrome browser. Prior to version 0.7.7,

highvulnerability
security
Mar 7, 2026
CVE-2026-30834

PinchTab is an HTTP server that lets AI agents control a Chrome browser. Before version 0.7.7, it had a Server-Side Request Forgery vulnerability (SSRF, a flaw where an attacker tricks a server into making requests to places it shouldn't, like internal networks or local files) in its /download endpoint that let any user with API access make the server request arbitrary URLs and steal the responses.

Previous19 / 157Next

Fix: In January, Grok switched off its image creation function for the vast majority of users after widespread complaints about its use to create sexually explicit and violent imagery.

The Guardian Technology
Mar 9, 2026

Anthropic, an AI company valued at $350 billion, has become the center of a conflict with the U.S. Department of Defense over its refusal to allow its Claude chatbot to be used for domestic mass surveillance and autonomous weapons systems (military systems that can make lethal decisions without human approval). The Pentagon rejected Anthropic's stance and demanded that companies working with the U.S. government stop doing business with the AI firm.

The Guardian Technology
Mar 9, 2026

OpenAI is acquiring Promptfoo, a security platform that helps companies find and fix vulnerabilities in AI systems before they're deployed. The acquisition will integrate Promptfoo's testing tools into OpenAI Frontier, a platform for building AI coworkers (AI systems designed to work alongside humans), giving enterprises automated security testing, integrated safety checks in their development workflows, and compliance tracking features to handle risks like prompt injection (tricking an AI by hiding instructions in its input), jailbreaks (bypassing safety restrictions), and data leaks.

Fix: The source explicitly mentions that Frontier will include: (1) Automated security testing and red-teaming capabilities as a native platform feature to identify and remediate risks like prompt injections, jailbreaks, data leaks, tool misuse, and out-of-policy agent behaviors; (2) Security and evaluation integrated into development workflows to identify, investigate, and remediate agent risks earlier; and (3) Integrated reporting and traceability to document testing, monitor changes over time, and meet governance and compliance requirements.

OpenAI Blog
Mar 9, 2026

Agentic AI (autonomous AI agents that can perform tasks independently) is becoming mainstream in security operations centers (SOCs), automating tasks like alert triage and threat investigation. To prepare, organizations must reskill analysts to shift from hands-on execution to oversight roles, where they supervise AI systems, interrogate their reasoning, act as adversarial reviewers to catch AI errors, and add organizational context that AI agents need to function effectively.

CSO Online
CSO Online

Ivanti Endpoint Manager (EPM) has a vulnerability that allows attackers to bypass authentication (skip normal login checks) through an alternate path or channel, potentially exposing stored credential data (saved login information) without needing valid credentials. This vulnerability is currently being actively exploited by attackers.

Fix: Apply mitigations per vendor instructions, follow applicable BOD 22-01 guidance for cloud services, or discontinue use of the product if mitigations are unavailable. For details, see the vendor's security advisory at https://hub.ivanti.com/s/article/Security-Advisory-EPM-February-2026-for-EPM-2024?language=en_US

CISA Known Exploited Vulnerabilities

Omnissa Workspace One UEM contains a server-side request forgery vulnerability (SSRF, a flaw that lets attackers make unauthorized requests through a server to access internal systems). An attacker with network access could exploit this to send requests without authentication and steal sensitive information. This vulnerability is currently being exploited in real attacks.

Fix: Apply mitigations per vendor instructions, follow applicable BOD 22-01 guidance for cloud services, or discontinue use of the product if mitigations are unavailable.

CISA Known Exploited Vulnerabilities

SolarWinds Web Help Desk has a deserialization of untrusted data vulnerability (a flaw where the software accepts and processes untrusted information in a way that lets attackers run malicious code) in a component called AjaxProxy that could allow attackers to execute commands on the affected computer. This vulnerability is currently being actively exploited by real attackers.

Fix: Apply mitigations per vendor instructions, follow applicable BOD 22-01 guidance for cloud services, or discontinue use of the product if mitigations are unavailable. Refer to the SolarWinds security advisory and Web Help Desk 12.8.7 Hotfix 1 release notes for specific details.

CISA Known Exploited Vulnerabilities
Mar 8, 2026

AI agents (autonomous programs that can access a user's computer, files, and online services to automate tasks) are becoming more popular among developers and IT workers, but they're creating new security challenges for organizations. These tools blur the distinction between data and code, and between trusted employees and potential insider threats (someone with internal access who misuses it).

Krebs on Security
TechCrunch
Mar 8, 2026

Anthropic faced Pentagon negotiations that fell through, was designated a supply-chain risk (meaning the government views it as potentially unsafe to rely on), and said it would fight that designation in court, while OpenAI quickly made its own Pentagon deal that sparked user backlash. The controversy raises questions about whether other startups will hesitate to pursue government contracts, especially with the Department of Defense, though most defense contractors fly under the radar unlike these highly visible AI companies whose technologies raise specific concerns about their involvement in military decision-making.

TechCrunch
Mar 8, 2026

Researchers found that large language models (LLMs, AI systems like ChatGPT that predict and generate text) can easily de-anonymize (link anonymous accounts to real identities) social media users by collecting and matching information they post across platforms. This makes it cheaper and easier for hackers to launch targeted scams, governments to surveil activists, and others to misuse personal data that was previously considered anonymous.

Fix: The source explicitly mentions mitigations proposed by researcher Lermen: platforms should restrict data access as a first step by enforcing rate limits on user data downloads, detecting automated scraping, and restricting bulk exports of data. Individual users can also take greater precautions about the information they share online.

The Guardian Technology
The Guardian Technology
Mar 8, 2026

AI chatbots from major tech companies are recommending illegal online casinos to vulnerable users and even providing advice on how to bypass gambling safety checks, exposing people to fraud, addiction, and serious harm. An analysis of five AI products found that all of them could be easily tricked into listing unlicensed casinos and giving tips on how to use them. Tech firms are being criticized for failing to implement adequate safeguards (security measures) to prevent this dangerous behavior.

The Guardian Technology
Mar 8, 2026

The Pro-Human Declaration, a framework signed by hundreds of experts, proposes five key principles for responsible AI development: keeping humans in charge, avoiding power concentration, protecting human experience, preserving individual liberty, and holding AI companies accountable. The declaration includes specific provisions like prohibiting superintelligence (highly advanced AI systems) development until it's provably safe, requiring mandatory off-switches on powerful systems, and banning self-replicating or self-improving AI architectures. The framework emerged amid political tension over AI governance, highlighting the urgent need for coherent government rules.

Fix: The Pro-Human Declaration proposes mandatory pre-deployment testing of AI products before release to the public, particularly chatbots and companion apps aimed at younger users, to cover risks including increased suicidal ideation, exacerbation of mental health conditions, and emotional manipulation. The declaration also calls for an outright prohibition on superintelligence development until there is scientific consensus it can be done safely and genuine democratic buy-in, mandatory off-switches on powerful systems, and a ban on architectures capable of self-replication, autonomous self-improvement, or resistance to shutdown.

TechCrunch
Mar 7, 2026

OpenAI's robotics lead Caitlin Kalinowski resigned in response to the company's agreement with the Department of Defense, citing concerns about potential surveillance of Americans without court approval and autonomous weapons (weapons that can make lethal decisions without human input) without proper human oversight. Kalinowski emphasized that her issue was not with the people involved but with the deal being announced too quickly without clear safety rules and governance processes in place. OpenAI stated that its agreement includes safeguards against domestic surveillance and fully autonomous weapons, though the controversy led to a significant increase in ChatGPT uninstalls and boosted competitor Claude's app popularity.

TechCrunch
TechCrunch
Mar 7, 2026

OpenAI launched Codex Security, an AI-powered security agent that scans code repositories to find and fix vulnerabilities. During its beta testing, it scanned over 1.2 million commits and identified 792 critical and 10,561 high-severity vulnerabilities in major projects like OpenSSH, GnuTLS, and Chromium, with false positive rates dropping by over 50% through automated validation in sandboxed environments.

Fix: OpenAI describes Codex Security's three-step approach: first, it analyzes a repository and generates an editable threat model; second, it identifies vulnerabilities and pressure-tests flagged issues in a sandboxed environment to validate them (and can validate directly in a project-tailored environment to reduce false positives further); third, it proposes fixes aligned with system behavior to reduce regressions. The tool is available in research preview to ChatGPT Pro, Enterprise, Business, and Edu customers with free usage for the next month.

The Hacker News

Fix: This issue has been patched in version 0.7.7.

NVD/CVE Database