Certified Local Transferability for Evaluating Adversarial Attacks
Summary
Deep neural networks (DNNs, AI models with multiple layers that learn patterns) are vulnerable to adversarial examples, which are inputs slightly modified to trick the model into making wrong predictions. This paper introduces a concept called the certified local transferable region, a mathematically guaranteed area around an input where a single small perturbation (adversarial attack) will fool the model, and proposes a method called RAOS (reverse attack oracle-based search) to measure how large these vulnerable areas are as a way to evaluate how robust neural networks truly are.
Classification
Related Issues
Original source: http://ieeexplore.ieee.org/document/11142670
First tracked: March 16, 2026 at 04:14 PM
Classified by LLM (prompt v3) · confidence: 85%