{"data":{"id":"4930df4d-07d9-4439-a7da-38c0bc25b393","title":"Allies Teach Better Than Enemies: Inverse Adversaries for Robust Knowledge Distillation","summary":"This research proposes a new method for knowledge distillation (training a smaller AI model to mimic a larger one) that preserves adversarial robustness (the ability to resist attacks designed to fool AI systems). Instead of having the student model copy all predictions from the teacher model, the method uses \"inverse adversarial examples\" (inputs created by reversing the direction of adversarial attacks) to guide learning toward more reliable predictions, resulting in better robustness transfer between models.","solution":"N/A -- no mitigation discussed in source.","labels":["research","safety"],"sourceUrl":"http://ieeexplore.ieee.org/document/11370752","publishedAt":"2026-02-03T13:17:37.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"research","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-02-03T13:17:37.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"advanced","impactType":["safety"],"aiComponentTargeted":"model","llmSpecific":false,"classifierConfidence":0.92,"researchCategory":"peer_reviewed","atlasIds":null}}