{"data":{"id":"a5dbcc7c-5aa0-4df3-9357-8cb48a51a25c","title":"Designing AI agents to resist prompt injection","summary":"AI agents that browse the web and take actions are vulnerable to prompt injection (instructions hidden in external content to manipulate the AI into unintended actions), which increasingly uses social engineering tactics rather than simple tricks. Rather than trying to perfectly detect malicious inputs (which is as hard as detecting lies), the most effective defense is to design AI systems with built-in limitations on what agents can do, similar to how human customer service agents are restricted to limit damage if they're manipulated.","solution":"N/A -- no mitigation discussed in source.","labels":["security","safety"],"sourceUrl":"https://openai.com/index/designing-agents-to-resist-prompt-injection","publishedAt":"2026-03-11T11:30:00.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["prompt_injection"],"issueType":"news","affectedPackages":null,"affectedVendors":["OpenAI"],"affectedVendorsRaw":["OpenAI"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-03-11T11:30:00.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["integrity","safety"],"aiComponentTargeted":"agent","llmSpecific":true,"classifierConfidence":0.85,"researchCategory":null,"atlasIds":null}}