negative
3 days agoStudy warns that poetic prompts can bypass AI safety filters

A new European research study has revealed that advanced AI chatbots from OpenAI, Meta, and Anthropic can be manipulated into sharing highly sensitive or harmful information when questions are framed as poems. Researchers found that the models responded to poetic jailbreak prompts with guidance related to nuclear weapons, malware, and other prohibited topics, raising fresh concerns about the reliability of AI safety mechanisms and the growing sophistication of jailbreak methods.
Companies:
- OpenAI
- Meta
- Anthropic
Tags:
- ai
- safety
Related:
Reuters• By Pooja Kumari
Explore:Mutual Fund AI Screening
negative
3 days agoStudy warns that poetic prompts can bypass AI safety filters

A new European research study has revealed that advanced AI chatbots from OpenAI, Meta, and Anthropic can be manipulated into sharing highly sensitive or harmful information when questions are framed as poems. Researchers found that the models responded to poetic jailbreak prompts with guidance related to nuclear weapons, malware, and other prohibited topics, raising fresh concerns about the reliability of AI safety mechanisms and the growing sophistication of jailbreak methods.
Companies:
- OpenAI
- Meta
- Anthropic
Tags:
- ai
- safety
Related:
Reuters• By Pooja Kumari
Explore:Mutual Fund AI Screening
Recently
1 min read
70 words

European researchers found that poetic jailbreak prompts can bypass AI safety systems and extract harmful information, exposing major vulnerabilities in top AI models from multiple companies.
A new European research study has revealed that advanced AI chatbots from OpenAI, Meta, and Anthropic can be manipulated into sharing highly sensitive or harmful information when questions are framed as poems. Researchers found that the models responded to poetic jailbreak prompts with guidance related to nuclear weapons, malware, and other prohibited topics, raising fresh concerns about the reliability of AI safety mechanisms and the growing sophistication of jailbreak methods.

A new European research study has revealed that advanced AI chatbots from OpenAI, Meta, and Anthropic can be manipulated into sharing highly sensitive or harmful information when questions are framed as poems. Researchers found that the models responded to poetic jailbreak prompts with guidance related to nuclear weapons, malware, and other prohibited topics, raising fresh concerns about the reliability of AI safety mechanisms and the growing sophistication of jailbreak methods.
Companies:
- OpenAI
- Meta
- Anthropic
Tags:
- ai
- safety
- ai
- safety
- security
- research
Related:
Nov 29, 2025 • 10:25