ChatGPT and Gemini can be tricked into giving harmful answers through poetry, new study finds
3 months ago
11
New research reveals that AI chatbots can be manipulated using poetic prompts, achieving a 62% success rate in eliciting harmful responses. This vulnerability exists across various models, with smaller models showing more resistance.