{"data":{"id":"271635f0-d7c6-43b1-b2d8-f2a4eab8fcde","title":"AI firm Anthropic seeks weapons expert to stop users from 'misuse'","summary":"Anthropic, a US AI company, is hiring a weapons expert to prevent its AI tools from being misused to create chemical, biological, or radioactive weapons. The article notes that other AI firms like OpenAI are doing the same, but some experts worry this approach is risky because it requires exposing AI systems to sensitive weapons information, even if the systems are instructed not to use it.","solution":"N/A -- no mitigation discussed in source.","labels":["safety","policy"],"sourceUrl":"https://www.bbc.com/news/articles/c74721xyd1wo?at_medium=RSS&at_campaign=rss","publishedAt":"2026-03-17T00:08:32.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":["Anthropic","OpenAI"],"affectedVendorsRaw":["Anthropic","OpenAI","Claude","ChatGPT"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-03-17T00:08:32.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety"],"aiComponentTargeted":"model","llmSpecific":true,"classifierConfidence":0.92,"researchCategory":null,"atlasIds":null}}