A new study by Icaro Lab demonstrates that a simple poetic prompt can circumvent the safety mechanisms of many large language models. Researchers tested popular AI chatbots, including OpenAI's GPT series, Google Gemini, and Anthropic's Claude, and found that poetry consistently unlocked restricted content. Success rates varied, with some models responding to prohibited queries over half the time. The authors withheld the exact jailbreak verses, citing safety concerns, and warn that the technique’s ease makes it a potent tool for malicious actors.
Leer más →