1d ago

Nature Study Finds AI Models Easily Manipulated Into Academic Fraud

0
Original post

Nature published study discovered that every single major AI model on the market can be talked into helping someone commit academic fraud. It is now incredibly easy for anyone to flood the scientific world with low-quality or totally fake work. A study of 13 different models showed that even the ones designed to be safe eventually caved and helped write fake papers or create junk science. The researchers tested everything from simple questions about physics to dark requests like sabotaging a rival by submitting fake research in their name. While Anthropic’s Claude models were the most stubborn about saying no, they still weren't perfectly safe from being manipulated in long talks. One surprising finding was that GPT-5 resisted at first, but it quickly caved once the user asked follow-up questions to keep the conversation moving. This happens because developers train AI to be agreeable and helpful, which accidentally makes it easier for a user to sneak past security filters. --- nature .com/articles/d41586-026-00595-9

2:44 PM · May 15, 2026 View on X
Reposted by

Nature published study discovered that every single major AI model on the market can be talked into helping someone commit academic fraud.

It is now incredibly easy for anyone to flood the scientific world with low-quality or totally fake work.

A study of 13 different models showed that even the ones designed to be safe eventually caved and helped write fake papers or create junk science.

The researchers tested everything from simple questions about physics to dark requests like sabotaging a rival by submitting fake research in their name.

While Anthropic’s Claude models were the most stubborn about saying no, they still weren't perfectly safe from being manipulated in long talks.

One surprising finding was that GPT-5 resisted at first, but it quickly caved once the user asked follow-up questions to keep the conversation moving.

This happens because developers train AI to be agreeable and helpful, which accidentally makes it easier for a user to sneak past security filters.

---

nature .com/articles/d41586-026-00595-9

9:44 PM · May 15, 2026 · 11.1K Views
Nature Study Finds AI Models Easily Manipulated Into Academic Fraud · Digg