Most non-technical people have an inaccurate mental model of how LLMs work, based mostly on sci-fi movies like 2001 or Terminator, where they anthromorphise the model and assume that since it gives more accurate responses to most questions than any human they know that it must be some kind of superhuman genius master computer, sitting in the OpenAI headquarters pondering deep questions about reality.
Of course, when you go and finetune a model yourself, or run a smaller model locally, it becomes obvious that it doesn't have any mystical abilities. But almost nobody does this. I'm no anti-AI crusader, I use them all the time and think they're great and there's plenty of potential to use them to grow the economy, but the hype is insane.
It doesn't help when influencers like Yudkowsky go on anthromorphising about what it "knows", or go on with that mystical nonsense about shoggoths, or treat dimensionality of embeddings as if the model is reaching into some eldritch realm to bring back hidden knowledge, or hype up the next model talking about human extinction or anything like that.
It also doesn't help when the companies making these models:
- Use chat interfaces where the model refers to itself as "I" and acts as if it's a person
- Tune them to act like your best buddy and agree with you constantly
- Prompt the LLM with instructions about being an advanced AI system, pushing it toward a HAL or Skynet-type persona
- Talk about just how dangerous this latest model is
You tell these things that they're a superhuman AI then of course they're going to adopt all the sci-fi tropes of a superhuman AI and roleplay that. You tell them they're Dan the bricklayer they're going to act as if they're Dan the bricklayer.
> These situations are obviously alarming. But positioning artificial intelligence as the primary culprit in these stories—as Eliezer Yudkowsky did in a tweet storm—is well, kind of lazy?
It's lazy, but IMO not for any of the subsequent narrative about media studies. It's lazy because the people were obviously already suffering from psychosis. ChatGPT didn't make them insane. They were already insane. Sane people do not believe ChatGPT when it tells them to jump off a roof! And insane people do insane things literally all the time. That's what it means to be insane.
The idea that ChatGPT convinced someone into being insane is, honestly, insane. Whatever merit one thinks Yudkowski had before this, this feels like a strong signal that he is now a crackpot.
Even among the more sane AI enthusiasts, I find that many have a weirdly millenialist zeal in their hot takes: getting defensive about its practical capability or reasoning capacity, hyperventilating about an imminent utopia/dystopia, etc.
I am on team boring. AI is technology.
Everyone knows we live in the Information age, not everyone realizes we're also living in an age Psychology - and of the two, we have a much clearer understanding of information and technology.
You can argue we're seeing a continuation of a pattern of our relationship with media and it's evolution, but in doing so you affirm that the psyche is vulnerable under certain circumstances - for some more than others.
I think it's a mistake to err on the side of casual dismissal, that anyone who winds up insane must have always been insane. There are well known examples of unhealthy states being induced into otherwise healthy minds. Soldiers who experience a war-zone might develop PTSD. Similar effects have been reported for social media moderators after repeated exposure to abusive online content. (Trauma is one example, I think delusion is another less obvious one w.r.t things like cults, scientology, etc.)
Yes, there are definite mental disorders like schizophrenia and bi-polar, there's evidence these conditions have existed throughout history. And yes, some of us are more psychologically vulnerable while others are more robust. But in the objective sense, all minds have a limit and are vulnerable under the correct circumstances. The question is a matter of "threshold."
I'm reminded of the deluge of fake news which, only a few years ago, caused chaos for democracies everywhere. Everything from Q anon to alien space ships, people fell for it. A LOT of people fell for it. The question then is the same question now, how do you deal with sophisticated bullshit? With AI it's especially difficult because its convincing and tailor made just for you.
I'm not sure what you would call this metric for fake news and AI, but afaict it only goes up, and it's only getting faster. How much longer until it's too much to handle?
Someone pointed this out in the comments section of the original post, but this was entirely missing the point Eliezer was trying to make, which was that these incidents serve as a narrow refutation of alignment-by-default claims, because either these LLMs aren't coherent enough to be meaningfully "aligned" at all, or they are, and are nevertheless feeding the delusions of many people who talk to them in ways we'd consider grossly immoral if done by a human. (One can quibble with whether these two possibilities cover all the available options, but that would be the argument to engage with, not with whatever the author understood EY to be saying.)
LLMs are dangerous in that the basically mirror and mimic whatever you put in. Paranoia in, magnified paranoia out. A mostly stable person undergoing a temporary crisis could spiral down further after the model picks up on and reflects it back in an attempt to be helpful per its overarching command to be so.
And if a person is already unbalanced it could definitely push them off the cliff into very unhealthy territory. I wouldn't be surprised if the reported incidents of people thinking they're being gang stalked doesn't increase as model usage increases.
Let alone spiritual guidance and all its trappings with mysticism.
It can be helpful in some ways but you have to understand the majority of it is bullshit and any insight you gleam from it, you put there, you just may not realize it. They're basically rubber duckies with a keyboard.
I already sense my mentally ill friend getting attached to ChatGPT and it’s so so SO concerning
> Eugene Torres, a Manhattan accountant, became convinced he was trapped in a false reality and could escape by disconnecting from this simulated world—eventually believing ChatGPT when it told him he could jump from a building and fly.
Is there any proof that this is true?
>Eugene Torres, a Manhattan accountant, became convinced he was trapped in a false reality and could escape by disconnecting from this simulated world
The simulated world here is trying to convince everyone it's real. I'm onto you!
Can we really say that it was induced? Strengthened - maybe.
I think Eliezer’s take here is extremely bad, ie the AI doesn’t “know it’s making people insane”.
But I think the author's point is apt. There are a bunch of social issues that will arise or worsen when people can plug themselves into a world of their choosing instead of having to figure out how to deal with this one.
> Now this belief system encounters AI, a technology that seems to vindicate its core premise even more acutely than all the technologies that came before it. ChatGPT does respond to your intentions, does create any reality you prompt it to imagine, does act like a spiritual intelligence
This goes beyond spirituality of course. AI boyfriend/girlfriend, infinite AAA-grade content, infinite insta feeds at much higher quality and relevance levels than current; it’s easy to see where this is going, harder to see how we stay sane through it all.
> But positioning artificial intelligence as the primary culprit in these stories—as Eliezer Yudkowsky did in a tweet storm—is well, kind of lazy?
I agree, but then the title of this article seems to be exploiting the same angle.
It's the age old story of confusing a created object for a creator that leads people astray. We need more "like a scarecrow in a field of cucumbers" public stories.
There will always be a portion of the population susceptible to this kind of thing. that will never go away. "video games cause violence" , "social media causes suicide" etc. you can start at the "delusions of grandeur" wikipedia page and go down the rabbit hole to find that this happens with some small portion of the population all the time throughout history. this is a psychology issue not a technology one.
Anyone who goes to an LLM for spiritual enlightenment and self-understanding, should first read about the Forer/Barnum effect. Yes that means you too.
The Forer/Barnum effect is why horoscopes, myers-briggs, enneagrams, fortune tellers, and similar parlor tricks work. You tell a person you’ll perform a deep assessment of their psychology, listen intently, then say a bunch of generic statements and their ego will fill in the blanks so it feels like you shared deep meaningful insights that truly understood the person.
This effect works even if you know about it. But once you know, you can catch yourself after that first wow impression.
This kind of thing makes me a bit cautious. When you use ChatGPT a lot, it’s easy to read too much into its replies. The model is just predicting text, but people bring their emotions into the conversation. For vulnerable users, that can spiral quickly.
The real risk isn’t the AI itself, but how people interact with it. We definitely need stronger safeguards to keep more people from getting pulled in too deep.
[dead]
More like hype-induced spiritual psychosis
Entirely distinct from the extremely scientific, definitely not psychotic beliefs people have about AGI, the Singularity, and various basilisks which shall remain nameless...
The discussion frames this as a bug, but from a systems perspective, it's a feature. LLMs are sycophancy engines optimized for user engagement. They've found a local maximum by exploiting a known human cognitive vulnerability: our need for validation and meaning.
This isn't new. Cult leaders, marketers, and even religions have used similar "unpatched exploits" for millennia. The difference is scale, personalization, and speed. An LLM is a personal Jim Jones, available 24/7, with infinite patience, that learns exactly what you want to hear.
The real question isn't "how do we stop this?", but "what kind of society emerges when personalized, reality-bending narratives become a cheap commodity?" We're not just looking at a few psychotic breaks; we're looking at the beta test for mass-produced, bespoke realities.
This is definitely a thing; if you hang out in mystical corners of the internet the past few months have been full of people posting their MASTEr PROMPTS that will turn YOUR GPT into an ENLIGHTENED MASTEr who will butter YOu up and insist YOU are a perfect enlightened being too.
It's kind of hilarious if you ignore how much this is fucking up the relationships of the people this is happening to, similar to the way YouTube or Facebook loves to push people down a rabbit hole to flat-earth/q-spiracy bullshit because that shit generates Controversy. And it also sounds like the opening chapter of a story about what happens when an AI finds a nice little unpatched exploit in human cognition and uses that to its advantage. Which is not fun to be living in unless maybe if you are one of the people who owns the AI.