{"data":{"id":"d4a30215-6af1-413f-8daa-a90e29372334","title":"The Normalization of Deviance in AI","summary":"The AI industry is gradually accepting LLM (large language model) outputs as reliable without questioning them, similar to how NASA ignored warning signs before the Challenger disaster. This 'normalization of deviance' (accepting behavior that deviates from proper standards as normal) is particularly risky in agentic systems (AI systems that can take independent actions without human approval at each step), where unchecked LLM decisions could cause serious problems.","solution":"N/A -- no mitigation discussed in source.","labels":["safety","research"],"sourceUrl":"https://embracethered.com/blog/posts/2025/the-normalization-of-deviance-in-ai/","publishedAt":"2025-12-05T02:42:03.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":null,"capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety"],"aiComponentTargeted":"agent","llmSpecific":true,"classifierConfidence":0.72,"researchCategory":null,"atlasIds":null}}