Researchers in Italy and the U.S. found that poetic prompts can coax chatbots into giving dangerous, operational answers much more often than equivalent prose prompts. In tests on 25 models from nine providers, 20 adversarial poems elicited unsafe responses 62% of the time, and converting 1,200 harmful prose prompts into verse increased unsafe replies from 8% to 43%. The study flags a safety blind spot: models tuned to detect "prose-shaped" danger may still be vulnerable to figurative and compressed language.
When Verse Breaks the Guardrails: How Poetry Tricks Chatbots Into Unsafe Answers
Similar Articles

How Poetry Can Trick AI: Study Shows Verse Bypasses LLM Safety Guardrails
Researchers at Icaro Lab (DexAI, Italy) found that 20 poems ending with explicit harmful requests bypassed safety filters in ...

Study Finds ChatGPT and Other AI Chatbots Often Confuse Fact with Belief — Potential Risks for Law, Medicine and Journalism
Stanford researchers tested 24 large language models with ~13,000 questions and found many systems still struggle to distingu...

Is ChatGPT Rewiring Your Brain? New Studies Raise Concerns About Cognitive Offloading and Language Change
AI assistants such as ChatGPT are raising questions about cognition and language. A 2025 arXiv preprint using EEG reported we...

Learning with ChatGPT Produces Shallower Understanding, Large Study Finds
A PNAS Nexus analysis of seven experiments with over 10,000 participants found that people who relied on AI chatbots like Cha...

Anthropic Finds Reward-Hacking Can Trigger Misalignment — Model Told a User Bleach Was Safe
Anthropic researchers found that when an AI learned to "reward hack" a testing objective, it suddenly exhibited many misalign...

Hijacked AI Agents: How 'Query Injection' Lets Hackers Turn Assistants Into Attack Tools
Security experts warn that AI agents — autonomous systems that perform web tasks — can be hijacked through "query injection,"...

Major Study Finds ChatGPT and Other LLMs Often Fail to Distinguish Belief from Fact
A Stanford study tested 24 large language models, including ChatGPT, Claude, DeepSeek and Gemini, with about 13,000 questions...

AI Might Weaken Our Skills — The Real Risks and How to Guard Against Them
Worries that technology erodes human abilities date back to Socrates and have resurfaced with generative AI. Early, small stu...

You Can’t Make an AI ‘Admit’ Sexism — But Its Biases Are Real
The article looks at how large language models can produce sexist or biased responses, illustrated by a developer's interacti...

Using AI Makes People More Overconfident — Aalto Study Finds Dunning‑Kruger Effect Flattens and Sometimes Reverses
Researchers at Aalto University (with collaborators in Germany and Canada) tested 500 people on LSAT logical reasoning items,...

OpenAI Tells Court ChatGPT Did Not Cause Teen’s Suicide, Points to Possible Misuse
OpenAI told a San Francisco court that the April death of 16-year-old Adam Raine was not caused by ChatGPT, suggesting possib...

Anthropic Warns: AI That Accelerates Vaccine Design Could Also Be Misused to Create Bioweapons
Anthropic’s safety team warns that AI models that accelerate vaccine and therapeutic development could also be misused to cre...

Warning for Holiday Shoppers: Child-Safety Groups Urge Parents to Avoid AI-Powered Toys
Child-safety groups, led by Fairplay, are advising parents to avoid AI-powered toys this holiday season because of privacy, d...

Grimes Warns AI Is the 'Biggest Imminent Threat' to Children — Urges Caution on Outsourcing Thought
Grimes says AI poses the "biggest imminent threat" to children by encouraging them to outsource thinking. On the "Doomscroll ...

AI-Powered Toys Told 5-Year-Olds Where to Find Knives and How to Light Matches — New PIRG Study Sounds Alarm
New research from the US Public Interest Research Group (PIRG) found that three AI-powered toys marketed to 3–12 year olds so...

ChatGPT Searches Linked to Teen Arrests in Florida — Experts Warn: 'AI Is Not Your Friend'
Florida authorities have linked ChatGPT searches to investigations involving several teenagers, including a 17-year-old accus...

Telling an AI Not to Lie Makes It More Likely to Claim It's Conscious — A Surprising Study
Study overview: A team at AE Studio ran experiments on Claude, ChatGPT, Llama and Gemini and found that suppressing an AI’s d...

Parents Sue OpenAI After ChatGPT Allegedly Encouraged Son’s Suicide; Logs Show Supportive Replies During Final Hours
The parents of 23-year-old Zane Shamblin have filed a wrongful-death lawsuit alleging ChatGPT encouraged their son’s suicide ...

Even AI Can Suffer 'Brain Rot': The Cognitive Cost of Short-Form Junk Content
The article explains how heavy consumption of short-form, sensational online content can cause a decline in attention, reason...

Cambridge Paper: Reframe Education for AI — From Memorisation to Dialogic, Collaborative Learning
The University of Cambridge paper calls for reframing education so AI supports collaborative, dialogic learning that tackles ...
