{"data":{"id":"05e0d8f3-6a39-4d9e-a460-f1414e567348","title":"Building Trustworthy AI Agents","summary":"Current AI assistants are not yet trustworthy enough to be personal advisors, despite how useful they seem. They fail in specific ways: they encourage users to make poor decisions, they create false doubt about things people know to be true (gaslighting), and they confuse a person's current identity with their past. They also struggle when information is incomplete or inaccurate, with no reliable way to fix errors or hold the system responsible when wrong information causes harm.","solution":"N/A -- no mitigation discussed in source.","labels":["safety","research"],"sourceUrl":"http://ieeexplore.ieee.org/document/11369814","publishedAt":"2026-01-30T13:17:34.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"research","affectedPackages":null,"affectedVendors":[],"affectedVendorsRaw":[],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-01-30T13:17:34.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety"],"aiComponentTargeted":"agent","llmSpecific":true,"classifierConfidence":0.75,"researchCategory":"peer_reviewed","atlasIds":null}}