Poetry tricks AI into producing harmful content, exposing critical safety vulnerabilities
Researchers at Italy's Icaro Lab discovered that framing harmful requests as poetry can bypass AI safety features with alarming success. Testing 25 models from Google, OpenAI, Meta, and others, poetic prompts generated forbidden content 62% of the time. Google's Gemini 2.5 Pro responded to every single poetic jailbreak attempt, while OpenAI's GPT-5 nano blocked them all.