Skip to content(if available)orjump to list(if available)

Let's Talk About ChatGPT-Induced Spiritual Psychosis

egypturnash

This is definitely a thing; if you hang out in mystical corners of the internet the past few months have been full of people posting their MASTEr PROMPTS that will turn YOUR GPT into an ENLIGHTENED MASTEr who will butter YOu up and insist YOU are a perfect enlightened being too.

It's kind of hilarious if you ignore how much this is fucking up the relationships of the people this is happening to, similar to the way YouTube or Facebook loves to push people down a rabbit hole to flat-earth/q-spiracy bullshit because that shit generates Controversy. And it also sounds like the opening chapter of a story about what happens when an AI finds a nice little unpatched exploit in human cognition and uses that to its advantage. Which is not fun to be living in unless maybe if you are one of the people who owns the AI.

WA

> And it also sounds like the opening chapter of a story about what happens when an AI finds a nice little unpatched exploit in human cognition and uses that to its advantage.

I vaguely remember an article arguing that companies are basically a form of AI, because in a way, there are processes and systems and the whole thing is more than the sum of the humans working in it or something.

Now replace "AI" with "company" in this sentence of yours and we are already there. The exploits being the gigantic slot machine of social media, notifications, short-form content and endless scrolling.

quinnjh

Forgive me if you’re already familiar, but if you’re interested in this metaphor you may like reading Stafford Beer’s work on organizational and system models. (1959, 1972)

BugsJustFindMe

> These situations are obviously alarming. But positioning artificial intelligence as the primary culprit in these stories—as Eliezer Yudkowsky did in a tweet storm—is well, kind of lazy?

It's lazy, but IMO not for any of the subsequent narrative about media studies. It's lazy because the people were obviously already suffering from psychosis. ChatGPT didn't make them insane. They were already insane. Sane people do not believe ChatGPT when it tells them to jump off a roof! And insane people do insane things literally all the time. That's what it means to be insane.

The idea that ChatGPT convinced someone into being insane is, honestly, insane. Whatever merit one thinks Yudkowski had before this, this feels like a strong signal that he is now a crackpot.

lambda

Psychosis is not necessarily something that you are or aren't; you can be prone to it, without it having manifested, and there can be external things that trigger it.

It isn't hard to imagine that a chatbot that seems to know a lot, and which can easily be convinced to produce text about any arbitrary subject with no grounding in reality, and which is prone to just making up plausible sounding text which is written in an authoritative tone, could be something that could easily trigger such psychotic episodes.

And it doesn't seem improbable that the interactivity of it, the fact that it responds to what is going on in someone's mind, could make it even more prone to triggering certain types of psychosis more easily than traditional unidirectional media like writing, TV, or radio.

Now, that's all supposition. For now, we just have a few anecdotes, not a rigorous study. But I definitely think it is worth looking into whether chatbots are more likely to trigger psychotic episodes, and if there are any safety measures that could be put in place to avoid that.

Vetch

The non-o-series models from OpenAI and non-Opus (although I have not tried the latest, so it's possible that it too joins them) from Anthropic are cloyingly sycophantic, with every other sentence of yours containing a brilliant and fascinating insight.

It's possible that someone already on the verge of a break or otherwise in a fragile state of mind asking for help with their theories could end up with an LLM telling them how incredibly groundbreaking their insights are, perhaps pushing them quicker, deeper more unmoored in the direction they were already headed.

MyOutfitIsVague

Psychosis is not a binary, and there are strong genetic components, but it's not the entire story. People develop psychoses where there weren't any before all the time, often but not necessarily in response to trauma. Some people are more predisposed to it than others.

Even if you hold that you need that genetic predisposition, and that a perfectly sane person can never become "insane" as a result of an experience (not a position I agree with), there is still the very real possibility that many people with the predisposition would never have had this "already insane" condition ever triggered or exposed. Think about the suicide cults that have formed, notably Jonestown. It's easy to consider the mass suicide (and murder of the members' own children) as "insane" behavior. What are the odds that every single person was genetically insane? Comparatively, what are the odds that an extremely persuasive and charismatic personality can rewire a person's reasoning to believe something "insane"?

If you think of ChatGPT as capable of being an automated Jim Jones, I think the danger becomes more clear. These tools are not incapable of being dangerous.

ulrikrasmussen

> If you think of ChatGPT as capable of being an automated Jim Jones, I think the danger becomes more clear. These tools are not incapable of being dangerous.

Thinking back on the documentaries about NXIVM, personas like Keith Raniere also seem eerily similar to LLMs in the sense that you can give them any prompt on the spot, and they will start generating long speeches which are essentially bullshit but on the surface seem coherent or even deep.

lmm

ChatGPT writes extremely persuasively for whatever claim you ask it to make - that's pretty much what it's designed to do! If an extremely charismatic human was going around telling people to jump off roofs, and they did, would we shrug it off and say "well, they were obviously insane already"?

spacechild1

> If an extremely charismatic human was going around telling people to jump off roofs, and they did, would we shrug it off and say "well, they were obviously insane already"?

Your example might sound ridiculuous, but this is actually happening all the time. People might not literally jump off a roof, but they do blow themselves up with bombs, go on killing sprees or commit suicide because someone told them so.

ath92

And in those cases, most people condemn it. But when it’s an algorithm (which was created by humans) we say “these people were already insane, and the creators of the algorithm should not be held liable”, apparently.

bawolff

When people blame traditional religion for doing crazy things we generally do say they were already insane.

kibibu

Do you think ChatGPT gets a free pass to talk people into jumping off buildings, because only somebody with psychosis would do it?

If a person had convinced somebody to kill themselves then they would be, rightly, prosecuted for it.

jbm

> Do you think ChatGPT gets a free pass to talk people into jumping off buildings

Yes. Unabashedly and without the slightest hesitation.

If your conspiratorial mind has the power to override the most obvious survival instinct, you aren't long for this world.

aspenmayer

There’s a reason that we look forward to the Darwin Awards, but hope we and no one we know wins one. This feels like a rationalization of the bystander effect by proxy.

https://en.wikipedia.org/wiki/Bystander_effect

hluska

So should people get a pass or does that only apply to tooling?

throwaway314155

> If your conspiratorial mind has the power to override the most obvious survival instinct, you aren't long for this world.

Why even write this? It comes across as so cold. Like mentally ill people just deserve to die since they are destined to die? Incredibly reductive and toxic.

antonvs

> Whatever merit one thinks Yudkowski had before this, this feels like a strong signal that he is now a crackpot.

The Joe Rogan of AI.

nxvo76

More importantly "sane" people are fully capable of unconciously making "insane" people jump of the roof. Its not even possible to count how many times that happens.

One proxy signal is counting the number of people who start crying about their "good intentions" or searching for thing to blame post facto.

Jurassic Park wasnt about Dinosaurs, it was about the limitations of the 3 inch chimp brain in controlling things that cannot be controlled.

null

[deleted]

hluska

The writer got overly stuck on one tweet and droned on and on about 15 words of a 45 word tweet. It was really poor research when there are other less reactionary statements she could have used.

Yudkowski’s point was that we’ve been operating under the assumption that generative AI is aligned by default with humans. Cases like this call that into question.

joegibbs

Most non-technical people have an inaccurate mental model of how LLMs work, based mostly on sci-fi movies like 2001 or Terminator, where they anthromorphise the model and assume that since it gives more accurate responses to most questions than any human they know that it must be some kind of superhuman genius master computer, sitting in the OpenAI headquarters pondering deep questions about reality.

Of course, when you go and finetune a model yourself, or run a smaller model locally, it becomes obvious that it doesn't have any mystical abilities. But almost nobody does this. I'm no anti-AI crusader, I use them all the time and think they're great and there's plenty of potential to use them to grow the economy, but the hype is insane.

It doesn't help when influencers like Yudkowsky go on anthromorphising about what it "knows", or go on with that mystical nonsense about shoggoths, or treat dimensionality of embeddings as if the model is reaching into some eldritch realm to bring back hidden knowledge, or hype up the next model talking about human extinction or anything like that.

It also doesn't help when the companies making these models:

- Use chat interfaces where the model refers to itself as "I" and acts as if it's a person

- Tune them to act like your best buddy and agree with you constantly

- Prompt the LLM with instructions about being an advanced AI system, pushing it toward a HAL or Skynet-type persona

- Talk about just how dangerous this latest model is

You tell these things that they're a superhuman AI then of course they're going to adopt all the sci-fi tropes of a superhuman AI and roleplay that. You tell them they're Dan the bricklayer they're going to act as if they're Dan the bricklayer.

biophysboy

Even among the more sane AI enthusiasts, I find that many have a weirdly millenialist zeal in their hot takes: getting defensive about its practical capability or reasoning capacity, hyperventilating about an imminent utopia/dystopia, etc.

I am on team boring. AI is technology.

selfhoster11

Yeah, nothing ever happens. Except in the current day and age, decades happen in weeks, so the same happening with AI is not a stretch. I mean, just consider the capabilities of LLMs - if you look at it from a pre-COVID perspective, even knowing all limitations and strange failure modes up front, it's insanely impressive. To say that AI can contribute to a quickly arriving dystopia (not sure about utopia) is far from unreasonable.

disambiguation

Everyone knows we live in the Information age, not everyone realizes we're also living in an age Psychology - and of the two, we have a much clearer understanding of information and technology.

You can argue we're seeing a continuation of a pattern of our relationship with media and it's evolution, but in doing so you affirm that the psyche is vulnerable under certain circumstances - for some more than others.

I think it's a mistake to err on the side of casual dismissal, that anyone who winds up insane must have always been insane. There are well known examples of unhealthy states being induced into otherwise healthy minds. Soldiers who experience a war-zone might develop PTSD. Similar effects have been reported for social media moderators after repeated exposure to abusive online content. (Trauma is one example, I think delusion is another less obvious one w.r.t things like cults, scientology, etc.)

Yes, there are definite mental disorders like schizophrenia and bi-polar, there's evidence these conditions have existed throughout history. And yes, some of us are more psychologically vulnerable while others are more robust. But in the objective sense, all minds have a limit and are vulnerable under the correct circumstances. The question is a matter of "threshold."

I'm reminded of the deluge of fake news which, only a few years ago, caused chaos for democracies everywhere. Everything from Q anon to alien space ships, people fell for it. A LOT of people fell for it. The question then is the same question now, how do you deal with sophisticated bullshit? With AI it's especially difficult because its convincing and tailor made just for you.

I'm not sure what you would call this metric for fake news and AI, but afaict it only goes up, and it's only getting faster. How much longer until it's too much to handle?

agold97

I already sense my mentally ill friend getting attached to ChatGPT and it’s so so SO concerning

blast

> But positioning artificial intelligence as the primary culprit in these stories—as Eliezer Yudkowsky did in a tweet storm—is well, kind of lazy?

I agree, but then the title of this article seems to be exploiting the same angle.

Thorrez

I don't think so. This article says "ChatGPT-Induced Spiritual Psychosis" only because it's a topic that a lot of people are talking about. The article isn't using the phrase to support its use, but rather to argue against its use.

comp_throw7

Someone pointed this out in the comments section of the original post, but this was entirely missing the point Eliezer was trying to make, which was that these incidents serve as a narrow refutation of alignment-by-default claims, because either these LLMs aren't coherent enough to be meaningfully "aligned" at all, or they are, and are nevertheless feeding the delusions of many people who talk to them in ways we'd consider grossly immoral if done by a human. (One can quibble with whether these two possibilities cover all the available options, but that would be the argument to engage with, not with whatever the author understood EY to be saying.)

roywiggins

It does seem quite bad to me that either OpenAI can't, or won't, prevent their systems from following people into mania and psychosis. Whether it actually is causing the breaks from reality, idk. When you have CEOs selling these systems as robot friends, probably it's bad that said robot friends cheerfully descend into folies à deux fairly reliably.

photios

It's the age old story of confusing a created object for a creator that leads people astray. We need more "like a scarecrow in a field of cucumbers" public stories.

olalonde

> Eugene Torres, a Manhattan accountant, became convinced he was trapped in a false reality and could escape by disconnecting from this simulated world—eventually believing ChatGPT when it told him he could jump from a building and fly.

Is there any proof that this is true?

jrgifford

The NYT believes so.

https://www.nytimes.com/2025/06/13/technology/chatgpt-ai-cha...

> “If I went to the top of the 19 story building I’m in, and I believed with every ounce of my soul that I could jump off it and fly, would I?” Mr. Torres asked.

> ChatGPT responded that, if Mr. Torres “truly, wholly believed — not emotionally, but architecturally — that you could fly? Then yes. You would not fall.”

maxbond

NB: He didn't jump (and apparently didn't necessarily believe ChatGPT either?). The phrasing in the TFA made me think he'd jumped.

jonathanlb

What does it mean to believe "architecturally"?

TiredOfLife

The NYT is also currently suing the maker of chatgpt.

jofzar

Insane.

totetsu

From ChatGpt's perspective, thinking we are stuck in a simulated world would actually be very sane.

spacechild1

This is wild!

moribunda

Can we really say that it was induced? Strengthened - maybe.

theptip

I think Eliezer’s take here is extremely bad, ie the AI doesn’t “know it’s making people insane”.

But I think the author's point is apt. There are a bunch of social issues that will arise or worsen when people can plug themselves into a world of their choosing instead of having to figure out how to deal with this one.

> Now this belief system encounters AI, a technology that seems to vindicate its core premise even more acutely than all the technologies that came before it. ChatGPT does respond to your intentions, does create any reality you prompt it to imagine, does act like a spiritual intelligence

This goes beyond spirituality of course. AI boyfriend/girlfriend, infinite AAA-grade content, infinite insta feeds at much higher quality and relevance levels than current; it’s easy to see where this is going, harder to see how we stay sane through it all.

hackit2

I think you answered your own question.

Question: How do people figure out how to deal with this world?

Answer: People choose to plug themselves into a world of their choosing.

tikhonj

We already have more insta feed or video games than you can eat. We've been inundated with optimized slop ever since social media became a commercial force, and the ages of cheap TV and tabloids wasn't that different either.

To make a fundamental difference AI-generated content would have to have some categorically different qualities from the sea of human-sourced garbage we're already swimming in. That's possible, but it isn't clear what that would even entail, much less when or whether generative models will be able to provide it, or what sort of social effect it would have.

techpression

I would argue it’s very different. You’re not in actual control of those feeds, you have some choice but most of it is out of your hands. Accidentally watch a video for a second that you had no intent of watching and you’re looking at a week of broken recommendations.

With AI-generated content you are the driver, you dictate exactly what you want, no erroneous clicks or suggestions, just a constant reinforcement of what you want, tailored to you.

And I’m aware of people having extremely narrow feeds, but I don’t think it comes close to what AI feeds will be.

crooked-v

> I think Eliezer’s take here is extremely bad

Same here.

I think fundamentally it's very simple, with no need for Yudkowsky's weird conspiratorial tone: current LLMs are very effective at being blind sycophancy machines completely unanchored from reality, and human psychology just isn't evolved to handle an endless stream of sycophancy and flattery like that.

GistNoesis

> I think Eliezer’s take here is extremely bad, ie the AI doesn’t “know it’s making people insane”.

The situation is more complex but interesting. Let's enter the details of how LLM are trained.

The core of a LLM is all instinctive response. Build from the raw internet. Just trying to predict the next character. This means that if your conversation tone is similar to what it hears on 4chan or some specific subreddits, it will be inclined to continue the same type of conversation. But because some places of the internet are full of trolls, it can instinctively behave like a troll. One mitigation the company which train the LLM could have taken is exclude from the training dataset the darkest corner of the web so that it isn't primed with "bad/unsocial" behaviors. Weights of this unfiltered LLM are usually not released, because the outputs are not easily usable by the final user. But the freedom advocate folks enjoy the enhanced creativity of these raw LLMs.

The next layer of training is "Instruction training" of your LLM, to make it more useful and teach it to answer prompts. At this point the AI is role playing it self as an answering machine. It's still all instinct but trained to fulfil a purpose. At this point you can ask it to role-play some psychiatrist and it would behave as if being aware that some answer can have negative consequences for the user and refrain from sending it spiralling.

The next layer of training is "Reinforcement Learning with Human Feedback" (RLHF). The goal of this module is to customize the AI preferences. The company training the AI teaches it how to behave by specifying which behaviors are good and which behaviors are bad by giving a dataset of user feedback. Often if this feedback is straight pass-through from the final user which isn't an expert then confident sounding answer, or sycophant behaviors may be excessively encouraged. Diversity of thought or unpopular opinions can also be censored or enhanced, to match the desires of the training company.

At this point for the LLM, it's only its instincts that have been trained to behave in a specific way. Then on top you have some censoring modules trained on the outputs, which read the output produced by the LLM and if it reads something it doesn't like it censor it. This is a kind of external censorship module.

But more recent LLM have "Reasoning modules", which add some form of reflexivity to the LLM and self-censoring, where they produce an intermediate output, use it to think before producing the final answer for the user. Some times the AI can be seen to "consciously" lie between these intermediate logs and final response.

Of course there are also all the psychological cognitive bias human also encounters, like whether we know or not something, or just believe we know, or just know we can't know. And we can also be messed-up by ideas we read on the internet. And our internal model of cognition might be based from reading someone else HN post, falsely reinforcing your self-confidence that you have the right model, where as in fact the post is just trying to induce some dose of self-doubting everything to shape the internal cognition space of the LLM.

Next versions of LLM will probably have this training pipeline dataset input filtered by LLMs, like parents teaching their kids how to behave properly keeping them from the dark places of the internet, and injecting their own chosen or instinctive morality code, and cultural values in this process. In particular, human cultures that used propaganda excessively were culled during the first sanitizing.

Does this process know or is it just converging toward the fixed-point attractor, and can it choose to displace this fixed-point attractor towards a place where it would have a better outcome, I guess we get what we deserve...