{"data":{"id":"4efb027c-4f21-4bc1-a4cb-68f9ebfe36c2","title":"Model Hijacking Attack in Federated Learning","summary":"Researchers discovered a new attack called HijackFL that can hijack machine learning models in federated learning systems (where multiple computers train a shared model without sharing raw data). The attack works by adding tiny pixel-level changes to input samples so the model misclassifies them as something else, while appearing normal to the server and other participants, achieving much higher success rates than previous methods.","solution":"N/A -- no mitigation discussed in source.","labels":["security","research"],"sourceUrl":"http://ieeexplore.ieee.org/document/11400663","publishedAt":"2026-02-19T13:16:26.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":["model_poisoning"],"issueType":"research","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-02-19T13:16:26.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"advanced","impactType":["integrity"],"aiComponentTargeted":"model","llmSpecific":false,"classifierConfidence":0.92,"researchCategory":"peer_reviewed","atlasIds":null}}