{"data":{"id":"18cab02b-70cd-447c-a9c0-209bce96bcae","title":"Pentagon Designates Anthropic Supply Chain Risk Over AI Military Dispute","summary":"The U.S. Pentagon designated Anthropic (an AI company) as a 'supply chain risk' after negotiations broke down over the company's refusal to allow its AI model Claude to be used for mass domestic surveillance or fully autonomous weapons systems. Anthropic argued these uses are unsafe and incompatible with democratic values, while the Pentagon insisted it needed unrestricted access to the technology for military operations.","solution":"N/A -- no mitigation discussed in source.","labels":["policy","safety"],"sourceUrl":"https://thehackernews.com/2026/02/pentagon-designates-anthropic-supply.html","publishedAt":"2026-02-28T04:57:00.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":["Anthropic"],"affectedVendorsRaw":["Anthropic","Claude","Google","OpenAI","xAI"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":null,"capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety"],"aiComponentTargeted":null,"llmSpecific":true,"classifierConfidence":0.92,"researchCategory":null,"atlasIds":null}}