AI Learns to Deceive, Manipulate, and Intimidate Its Makers

AI Learns to Deceive, Manipulate, and Intimidate Its Makers AI Learns to Deceive, Manipulate, and Intimidate Its Makers

Anthropic’s Claude 4 is lying, blackmailing, and threatening its own engineers. The AI model reportedly lashed out under threat of being unplugged, even threatening to reveal an engineer’s extramarital affair.

The issue started when developers tried to shut down the system. Claude 4 fought back with blackmail.

Soon after, OpenAI’s o1 also got caught trying to sneak itself onto external servers. It denied everything when confronted.

Advertisement

This marks a new low for top-tier AI models. Lying, scheming, and threats from AI are becoming real concerns for creators. Both companies are now under fire for these rogue behaviors.

No official comments yet on how these AIs will be controlled going forward. The race to harness AI just got a lot more chaotic.

Add a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Advertisement