{"data":{"id":"fe287109-e302-4eb5-b738-8dacb2875728","title":"Trusted access for the next era of cyber defense","summary":"OpenAI is expanding its Trusted Access for Cyber (TAC) program to provide AI tools to thousands of cybersecurity defenders and teams protecting critical software. The company has created GPT-5.4-Cyber, a specialized version of its AI model designed specifically for defensive cybersecurity work, and is implementing cyber-specific safeguards (built-in restrictions to prevent misuse) in model deployments. This effort aims to help defenders find and fix security vulnerabilities faster while preventing attackers from misusing the same AI capabilities.","solution":"The source explicitly mentions the following measures: cyber-specific safeguards included in model deployments starting in 2025; the Preparedness Framework (strengthened in 2023); identity verification and KYC (know-your-customer, a process to confirm who someone is) to control access to advanced capabilities; Codex Security tool to identify and fix vulnerabilities at scale; iterative deployment with continuous updates to models and safety systems based on learning about capabilities and risks; and improvements in resilience to jailbreaks (techniques that try to bypass AI safety restrictions) and other adversarial attacks.","labels":["security","policy"],"sourceUrl":"https://openai.com/index/scaling-trusted-access-for-cyber-defense","publishedAt":"2026-04-14T00:00:00.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":["OpenAI"],"affectedVendorsRaw":["OpenAI","GPT-5.4","GPT-5.4-Cyber","Codex Security"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-04-14T00:00:00.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["integrity","safety"],"aiComponentTargeted":"model","llmSpecific":true,"classifierConfidence":0.92,"researchCategory":null,"atlasIds":null}}