{"data":{"id":"1b511362-89fc-42b3-8d4d-81bdfd21737f","title":"Our commitment to community safety","summary":"OpenAI describes its safety approach for ChatGPT to prevent misuse for violence, threats, or harm. The system is trained to distinguish between harmful requests and legitimate questions about violence for educational or historical reasons, while using detection systems and expert guidance to identify concerning patterns across conversations and take action like revoking access when needed.","solution":"N/A -- no mitigation discussed in source. The text describes OpenAI's existing safety measures (model training, automated detection systems, expert consultation, policy enforcement, and access revocation) but does not present these as solutions to a specific problem or security vulnerability that requires fixing.","labels":["safety","policy"],"sourceUrl":"https://openai.com/index/our-commitment-to-community-safety","publishedAt":"2026-04-28T00:00:00.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":["OpenAI"],"affectedVendorsRaw":["OpenAI","ChatGPT"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-04-28T00:00:00.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety"],"aiComponentTargeted":"model","llmSpecific":true,"classifierConfidence":0.92,"researchCategory":null,"atlasIds":null}}