{"data":{"id":"10ae6de3-38db-4ba7-9d90-fcc0751ca853","title":"Lessons From AI Hacking: Every Model, Every Layer Is Risky","summary":"Two security researchers from Wiz, after spending two years identifying flaws in AI systems, argue that security professionals should focus less on prompt injection (tricking an AI by hiding instructions in its input) and more on other types of vulnerabilities that exist throughout AI infrastructure. The researchers suggest that risks exist at multiple levels of AI systems, not just in how users interact with the AI directly.","solution":"N/A -- no mitigation discussed in source.","labels":["security","research"],"sourceUrl":"https://www.darkreading.com/application-security/lessons-ai-hacking-model-every-layer-risky","publishedAt":"2026-02-20T18:02:02.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["prompt_injection","other"],"issueType":"news","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":null,"capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["integrity","confidentiality"],"aiComponentTargeted":null,"llmSpecific":false,"classifierConfidence":0.75,"researchCategory":null,"atlasIds":null}}