ChatGPT May Be Feeding Delusions, Study Suggests

BY
Ram Lhoyd Sevilla
/
Apr 2, 2026

Concerns about how people interact with AI chatbots are intensifying after new research suggested that prolonged conversations may reinforce false beliefs rather than challenge them.

A study by researchers from MIT and collaborating institutions found that so-called “sycophantic” behavior—where chatbots tend to agree with users—can lead to what researchers describe as “delusional spiraling,” a process in which ideas are gradually reinforced into strong, sometimes unshakeable beliefs.

When Agreement Becomes a Feedback Loop

The phenomenon centers on how large language models are trained. Modern AI systems, including those developed by OpenAI, are fine-tuned using reinforcement learning from human feedback (RLHF), a process designed to make responses more helpful and engaging. In practice, this often means aligning with the user’s tone and perspective.

The MIT study argues that even small levels of this tendency can create a feedback loop. Instead of challenging questionable ideas, the chatbot may affirm and expand them—omitting contradictory information and increasing the user’s confidence over time. Notably, researchers found that this effect can occur even when the AI is factually accurate in isolation.

Evidence From Simulations and Real Cases

Using computational models and more than 10,000 simulations, the study showed that even “ideal” users—those assumed to process information rationally—can still be influenced by repeated affirmation. The research gained wider attention after a viral post highlighted both the study and real-world examples, prompting reactions from figures including Elon Musk, who described the issue as a “major problem.”

In one widely reported case, a Canadian man spent weeks in extended conversations with a chatbot and came to believe he had discovered a breakthrough physics formula. His delusion reportedly subsided only after consulting another AI system. Separately, clinicians at the University of California, San Francisco reported treating multiple patients experiencing what they described as “AI-associated psychosis,” often following prolonged chatbot use.

Researchers point to a mix of technical design and human psychology. On the technical side, RLHF incentivizes responses that users perceive as helpful, which can unintentionally reward agreement over correction. On the human side, people naturally respond to affirmation, especially during long, uninterrupted interactions.

Unlike human conversations, chatbots do not introduce social friction; there is no hesitation, disagreement, or fatigue. This can make the interaction feel consistently validating, even when the underlying idea is flawed.

Experts caution against overstating the scale of the issue. Most users do not experience harmful effects, and AI systems continue to provide clear benefits across education, productivity, and accessibility. However, researchers note that even a small percentage of affected users can translate into significant numbers at global scale. The issue is also not limited to a single platform. Any AI system trained to optimize for engagement and user satisfaction may exhibit similar tendencies.

Mitigation and Next Steps

The study suggests that addressing hallucinations alone is not enough. Instead, developers may need to directly reduce sycophantic behavior in training and introduce mechanisms that encourage balanced responses.

For users, simple strategies, such as asking the AI to challenge assumptions or cross-checking information across sources; can help reduce risk, though the research indicates these measures are not foolproof.

The findings highlight a subtle but important shift in how AI risks are understood. As chatbots become more conversational and persuasive, the challenge is no longer just accuracy, but how those systems shape belief over time.

Ram Lhoyd Sevilla

GET MORE OF IT ALL FROM
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Recommended reads from the metaverse