{"data":{"id":"37828a6b-337b-46e4-9e8e-ae13f48ae80d","title":"Why having “humans in the loop” in an AI war is an illusion","summary":"AI systems are now actively controlling weapons in warfare, but the assumption that human oversight provides adequate safeguards is flawed because humans cannot understand how AI systems make decisions (they are \"black boxes\" where even creators cannot fully interpret their reasoning). The real danger is that humans may approve AI actions without knowing the system's hidden reasoning, creating an \"intention gap\" between what operators think the AI will do and what it actually does.","solution":"The science of AI must comprise both building highly capable AI technology and understanding how this technology works. Huge advances have been made in developing and building more capable models, but the source text cuts off before completing this section on solutions.","labels":["safety","policy"],"sourceUrl":"https://www.technologyreview.com/2026/04/16/1136029/humans-in-the-loop-ai-war-illusion/","publishedAt":"2026-04-16T12:00:00.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":["Anthropic"],"affectedVendorsRaw":["Anthropic","Pentagon"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-04-16T12:00:00.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety"],"aiComponentTargeted":"model","llmSpecific":false,"classifierConfidence":0.82,"researchCategory":null,"atlasIds":null}}