A one-prompt attack that breaks LLM safety alignment | AI Sec Watch