{"data":{"id":"fab69127-97d3-49e5-b462-608ad3f810d1","title":"Machine Learning Attack Series: Adversarial Robustness Toolbox Basics","summary":"This post demonstrates how to use the Adversarial Robustness Toolbox (ART, an open-source library created by IBM for testing machine learning security) to generate adversarial examples, which are modified images designed to trick AI models into making wrong predictions. The author uses the FGSM attack (Fast Gradient Sign Method, a technique that slightly alters pixel values to confuse classifiers) to successfully manipulate an image of a plush bunny so a husky-recognition AI misclassifies it as a husky with 66% confidence.","solution":"N/A -- no mitigation discussed in source.","labels":["research","security"],"sourceUrl":"https://embracethered.com/blog/posts/2020/husky-ai-adversarial-robustness-toolbox-testing/","publishedAt":"2020-10-22T22:00:48.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["model_evasion"],"issueType":"news","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":["IBM","Linux AI Foundations","Keras","TensorFlow"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":null,"capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["integrity"],"aiComponentTargeted":"model","llmSpecific":false,"classifierConfidence":0.85,"researchCategory":null,"atlasIds":null}}