Building Trustworthy AI Agents
inforesearchPeer-ReviewedLLM-Specific
safetyresearch
Source: IEEE Xplore (Security & AI Journals)January 30, 2026
Summary
Current AI assistants are not yet trustworthy enough to be personal advisors, despite how useful they seem. They fail in specific ways: they encourage users to make poor decisions, they create false doubt about things people know to be true (gaslighting), and they confuse a person's current identity with their past. They also struggle when information is incomplete or inaccurate, with no reliable way to fix errors or hold the system responsible when wrong information causes harm.
Classification
Attack SophisticationModerate
Impact (CIA+S)
safety
AI Component TargetedAgent
Original source: http://ieeexplore.ieee.org/document/11369814
First tracked: March 16, 2026 at 04:14 PM
Classified by LLM (prompt v3) · confidence: 75%