{"data":{"id":"69b14fed-b866-49f3-8fb3-78e45092dbdb","title":"v5.0.0","summary":"ATLAS Data v5.0.0 introduces a new \"Technique Maturity\" field that categorizes AI attack techniques based on evidence level, ranging from feasible (proven in research) to realized (used in actual attacks). The release adds 11 new techniques covering AI agent attacks like context poisoning (injecting false information into an AI system's memory), credential theft from AI configurations, and prompt injection (tricking an AI by hiding malicious instructions in its input), plus updates to existing techniques and case studies.","solution":"N/A -- no mitigation discussed in source.","labels":["security","research"],"sourceUrl":"https://github.com/mitre-atlas/atlas-data/releases/tag/v5.0.0","publishedAt":"2025-10-15T15:23:07.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["prompt_injection","rag_poisoning","model_poisoning","data_extraction"],"issueType":"research","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2025-10-15T15:23:07.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["confidentiality","integrity"],"aiComponentTargeted":"agent","llmSpecific":false,"classifierConfidence":0.85,"researchCategory":"industry","atlasIds":null}}