{"data":{"id":"9cf135be-83c2-41e2-8401-67447892f2f7","title":"Attackers Could Exploit AI Vision Models Using Imperceptible Image Changes","summary":"Researchers at Cisco discovered that attackers can manipulate vision-language models (AI systems that read and interpret images) by making tiny, imperceptible changes to image pixels that humans cannot see. These changes can make hidden malicious instructions embedded in images readable to the AI, allowing attackers to trick the AI into following commands like stealing data, while content filters and humans see only visual noise or blurry content.","solution":"N/A -- no mitigation discussed in source.","labels":["security","research"],"sourceUrl":"https://www.securityweek.com/attackers-could-exploit-ai-vision-models-using-imperceptible-image-changes/","publishedAt":"2026-05-07T13:45:53.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"medium","attackType":["prompt_injection","model_evasion"],"issueType":"news","affectedPackages":null,"affectedVendors":["OpenAI","Anthropic"],"affectedVendorsRaw":["Cisco","OpenAI","GPT-4o","Claude","Qwen","JinaAI","SigLIP"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-05-07T13:45:53.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"advanced","impactType":["integrity","confidentiality"],"aiComponentTargeted":"model","llmSpecific":false,"classifierConfidence":0.85,"researchCategory":null,"atlasIds":null}}