{"data":{"id":"2552bc9d-e7a0-41be-8e30-6e5066b1be31","title":"Enhancing Adversarial Transferability With Cost-Efficient Landscape Flattening","summary":"This research paper describes a method called CLEF (Cost-efficient LandscapE Flattening) that improves adversarial transferability, which is the ability of adversarial examples (inputs deliberately crafted to fool AI models) to fool different models beyond the one they were designed for. The method works by flattening the input loss landscape (the mathematical surface showing how wrong a model's predictions are) by optimizing adversarial perturbations (small changes added to inputs) at both high-loss and low-loss points. The researchers show their approach can improve how well these adversarial examples transfer across different models while using fewer computations than previous methods.","solution":"N/A -- no mitigation discussed in source.","labels":["research","security"],"sourceUrl":"http://ieeexplore.ieee.org/document/11395656","publishedAt":"2026-02-13T13:16:36.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["model_evasion"],"issueType":"research","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-02-13T13:16:36.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"advanced","impactType":["integrity"],"aiComponentTargeted":"model","llmSpecific":false,"classifierConfidence":0.88,"researchCategory":"peer_reviewed","atlasIds":null}}