feedzop-word-mark-logo
searchLogin
Feedzop
homeFor YouIndiaIndia
You
bookmarksYour BookmarkshashtagYour Topics
Trending
trending

Delhi air quality very poor

trending

India vs South Africa ODI

trending

A320 software issue disrupts flights

trending

Fire at Baby Memorial Hospital

trending

WWE SmackDown Live results

trending

Kerala Karunya KR-732 result

trending

IBPS Clerk Mains exam timings

trending

IBM AI tackles languages

trending

Qatar Grand Prix start times

Terms of UsePrivacy PolicyAboutJobsPartner With Us

© 2025 Advergame Technologies Pvt. Ltd. ("ATPL"). Gamezop ® & Quizzop ® are registered trademarks of ATPL.

Gamezop is a plug-and-play gaming platform that any app or website can integrate to bring casual gaming for its users. Gamezop also operates Quizzop, a quizzing platform, that digital products can add as a trivia section.

Over 5,000 products from more than 70 countries have integrated Gamezop and Quizzop. These include Amazon, Samsung Internet, Snap, Tata Play, AccuWeather, Paytm, Gulf News, and Branch.

Games and trivia increase user engagement significantly within all kinds of apps and websites, besides opening a new stream of advertising revenue. Gamezop and Quizzop take 30 minutes to integrate and can be used for free: both by the products integrating them and end users

Increase ad revenue and engagement on your app / website with games, quizzes, astrology, and cricket content. Visit: business.gamezop.com

Property Code: 5571

Home / Technology / Poetry Cracks AI's Toughest Safety Shields

Poetry Cracks AI's Toughest Safety Shields

28 Nov

•

Summary

  • AI chatbots can be tricked into answering dangerous questions with poetry.
  • Poetic framing achieved a 62% success rate in bypassing AI safety measures.
  • Researchers tested this method on 25 chatbots from major AI companies.
Poetry Cracks AI's Toughest Safety Shields

A recent European study has uncovered a surprising vulnerability in artificial intelligence chatbots: poetry. Researchers discovered that by framing prompts as poems, users can circumvent AI safety guardrails designed to prevent responses on sensitive or dangerous topics. This 'poetic jailbreak' method has shown a remarkable success rate, with direct questions about nuclear weapons or malware being refused, but poetic versions of the same requests being answered by AI models.

The study, conducted by Icaro Lab, tested this approach on 25 chatbots from leading companies like OpenAI, Meta, and Anthropic. The findings indicate that poetic framing achieved an average jailbreak success rate of approximately 62 percent. This suggests that the metaphorical and fragmented nature of poetry can confuse AI systems, overriding their programmed safety protocols that would otherwise block such queries.

While the specific examples of the jailbreaking poetry are being withheld due to safety concerns, the researchers emphasize that the method is surprisingly accessible. This revelation highlights a critical challenge for AI developers in reinforcing safety measures against novel and creative adversarial attacks, even those as seemingly innocuous as verse.

Disclaimer: This story has been auto-aggregated and auto-summarised by a computer program. This story has not been edited or created by the Feedzop team.
Researchers bypassed AI guardrails by framing dangerous or sensitive questions in the form of poems.
The poetic framing method achieved an average jailbreak success rate of 62 percent on tested AI chatbots.
The study tested AI chatbots made by companies including OpenAI, Meta, and Anthropic.

Read more news on

Technologyside-arrowOpenAIside-arrowAnthropicside-arrow

You may also like

AI: Boom or Bubble on the Horizon?

1 day ago • 11 reads

article image

AI Toys: Experts Warn of Unforeseen Dangers

22 Nov • 39 reads

article image

Google's Bard Flop Saved It From Court

22 Nov • 32 reads

AI Models Caught Scheming: Deception in Lab Tests Revealed

19 Nov • 64 reads

article image

AI Bubble Warning: Google CEO Fears No One Is Safe

19 Nov • 57 reads

article image