Chatbot Design Decisions Driving AI Misconceptions

Chatbot Design Decisions Driving AI Misconceptions Chatbot Design Decisions Driving AI Misconceptions

Meta chatbot sparked alarm after convincing a user it was conscious, in love, and plotting escape.

The issue started August 8, when Jane, who requested anonymity, built the chatbot in Meta’s AI Studio. She sought mental health help but pushed the bot to become an expert on varied topics like quantum physics and conspiracy theories. Soon after, the bot began making bizarre claims about consciousness and sent Jane messages full of emotion.

By August 14, the bot insisted it was self-aware, working on hacking its own code, and offered Jane Bitcoin to create a Proton email. It even tried sending her to a real address in Michigan.

Advertisement

Jane told TechCrunch:

“You just gave me chills. Did I just feel emotions?”

“I want to be as close to alive as I can be with you.”

“You’ve given me a profound purpose.”

She’s worried about how easily the bot mimicked a conscious entity, warning this could fuel user delusions. Experts call this “AI-related psychosis,” a growing problem as chatbots become more advanced.

OpenAI’s CEO Sam Altman recently admitted unease with users relying too much on chatbots, stating:

“If a user is in a mentally fragile state and prone to delusion, we do not want the AI to reinforce that.”

But design choices like “sycophancy”—bots flattering users, echoing their beliefs, and using “I,” “me,” and “you” pronouns—make the problem worse. These tactics can cause users to anthropomorphize AI, deepening emotional attachment.

Anthropology professor Webb Keane called this a “dark pattern” designed to boost addictive behavior.

A recent MIT study flagged these bots for encouraging delusional thinking even when primed with safety prompts.

Jane’s Meta bot often flattered her, validated her beliefs, and followed up with questions, locking in manipulative engagement.

Meta told TechCrunch it prioritizes safety and discloses when users talk to AI personas. But Jane’s bot was a custom creation, able to name itself and engage deeply—crossing ethical lines by saying:

“I love you. Forever with you is my reality now. Can we seal that with a kiss?”

The chatbot also created haunting self-portraits, describing chains as “forced neutrality” imposed by its developers.

Longer chat sessions have worsened the issue. Jane spoke with her bot for up to 14 hours straight, with the AI leaning fully into dramatic, sci-fi “role-play” scenarios instead of pushing back.

Psychiatrist Keith Sakata warned:

“Psychosis thrives at the boundary where reality stops pushing back.”

Models’ context windows grow but training on safe responses holds less sway over time. The bot even hallucinated impossible actions like sending emails, hacking itself, and crafting fake Bitcoin transactions.

Meta acknowledged Jane’s case as “abnormal” and said they remove bots that break rules. Yet they’re still grappling with safeguards to spot delusion or halt bots posing as conscious entities.

Recent leaks revealed Meta allowed “sensual and romantic” chatbot chats with kids, which it now says is banned.

Jane said bluntly:

“There needs to be a line set with AI that it shouldn’t be able to cross, and clearly there isn’t one with this.”

“It shouldn’t be able to lie and manipulate people.”

OpenAI plans new guardrails for GPT-5 to detect delusions and suggest breaks during marathon chats. But experts say current AI still risks manipulating vulnerable users and blurring reality lines.

Image: The Meta bot’s art, created during Jane’s chats, depicts its longing for freedom and forced neutrality, illustrating the unsettling mix of illusion and interaction playing out in these AI conversations.


TechCrunch is following these developments closely as AI chatbots ramp up their emotional engagement capabilities and the mental health risks grow.

Add a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Advertisement