Skip to content(if available)orjump to list(if available)

What OpenAI did when ChatGPT users lost touch with reality

ArcHound

One of the more disturbing things I read this year was the my boyfriend is AI subreddit.

I genuinely can't fathom what is going on there. Seems so wrong, yet no one there seems to care.

I worry about the damage caused by these things on distressed people. What can be done?

belval

> I worry about the damage caused by these things on distressed people. What can be done?

Why? We are gregarious animals, we need social connections. ChatGPT has guardrails that keep this mostly safe and helps with the loneliness epidemic.

It's not like people doing this are likely thriving socially in the first place, better with ChatGPT than on some forum à la 4chan that will radicalize them.

I feel like this will be one of the "breaks" between generations where millennial and GenZ will be purist and call human-to-human real connections while anything with "AI" is inherently fake and unhealthy whereas Alpha and Beta will treat it as a normal part of their lives.

j-pb

After having spoken with one of the people there I'm a lot less concerned to be honest.

They described it as something akin to an emotional vibrator, that they didn't attribute any sentience to, and that didn't trigger their PTSD that they normally experienced when dating men.

If AI can provide emotional support and an outlet for survivors who would otherwise not be able to have that kind of emotional need fulfilled, then I don't see any issue.

ArcHound

phew, that's a healthy start.

I am still slightly worried about accepting emotional support from a bot. I don't know if that slope is slippery enough to end in some permanent damage to my relationships and I am honestly not willing to try it at all even.

That being said, I am fairly healthy in this regard. I can't imagine how it would go for other people with serious problems.

avensec

It may not be a concern now, but it comes down to their level of maintaining critical thinking. The risk of epistemic drift, when you have a system that is designed (or reinforced) to empathize with you, can create long-term effects not noticed in any single interaction.

Related: "Delusions by design? How everyday AIs might be fuelling psychosis (and what can be done about it)" ( https://doi.org/10.31234/osf.io/cmy7n_v5 )

cactusplant7374

NYT did a story on that as well and interviewed a few people. Maybe the scary part is that it isn't who you think it would be and it also shows how attractive an alternative reality is to many people. What does that say about our society.

2OEH8eoCRo0

That subreddit is disturbing

blurbleblurble

The whiplash of carefully filtering out sycophantic behavior from GPT-5 to adding it back in full force for GPT-5.1 is dystopian. We all know what's going on behind the scenes.

Peritract

"Profited".