The LLMentalist Effect
142 comments
·February 8, 2025JKCalhoun
everdrive
I think it's fair to argue that part of human intelligence is actually just a statistical matching. The best example which comes to mind is actually grammar. Grammar has a very complex set of rules, however most people cannot really name them or describe them accurately; instead, they just know whether or not a sentence _sounds_ correct. This feels a lot like the same statistical matching performed by LLMs. An individual's reasoning iterates in their mind what words follow each other, and what phrases are likely.
Outside of grammar, you can hear a lot of this when people talk; their sentences wander, and they don't always seem to know ahead of time where their sentences will end up. They start anchored to a thought, and seem to hope that the correct words end up falling into place.
Now, does all thought work like this? Definitely not, and more importantly, there are many other facets of thought which are not present in LLMs. When someone has wandered badly when trying to get a sentence out, they are often also able to introspect and see that they failed to articulate their thought. They can also slow down their speaking, or pause, and plan out ahead of time; in effect, using this same introspection to prevent themselves from speaking poorly in the first place. Of course there's also memory, consciousness, and all sorts of other facets of intelligence.
What I'm on the fence about is whether this point, or your point actually detracts from the author's argument.
mock-possum
Recognizing a pattern and reproducing it is a huge part of the human experience, and that’s the main thing that’s intelligent-like about LLMs. A lot of time they’re lack context / cohesion, and they’re at a disadvantage for not being able to follow normal human social cues to course correct, as you point out.
ianbicking
Yeah, the basic premise is off because LLM responses are regularly tested against ground truth (like running the code they produce), and LLMs don't get to carefully select what requests they fulfill. To the contrary they fulfill requests even when they are objectively incapable of answering correctly, such as incomplete or impossible questions.
I do think there is a degree of mentalist-like behavior that happens, maybe especially because of the RLHF step, where the LLM is encouraged to respond in ways that seem more truthful or compelling than is justified by its ability. We appreciate the LLM bestowing confidence on us, and rank an answer more highly if it gives us that confidence... not unlike the person who goes to a spiritualist wanting to receive comforting news of a loved one who has passed. It's an important attribute of LLMs to be aware of, but not the complete explanation the author is looking for.
pockmarked19
Which aspect of how a human leg works is a truck tire similar to?
Terr_
There's another illusory effect here: Humans are being encouraged to confuse a fictional character with the real-world "author" system.
I can create a mad-libs program which dynamically reassembles stories involving a kind and compassionate Santa Claus, but that does not mean the program shares those qualities. I have not digitally reified the spirit of Christmas, not even if excited human kids contribute some of the words that shape its direction and clap with glee.
P.S.: This "LLM just makes document bigger" framing is also very useful understanding how prompt injection and hallucinations are constant core behaviors, which we just ignore except when they inconvenience us The assistant-bot in the story can be twisted or vanish so abruptly because it's just something in a digital daydream.
bloomingkales
And only to your eyes and those you force your vision onto. The rest of the universe never sees it. You don’t exist to much of the universe (if a tree falls and no one is around to hear it, you understand what I mean).
So you simultaneously exist and don’t exist. Sorry about this, your post took me on this tangent.
GuB-42
"Do LLMs think?" is a false problem outside of the field of philosophy.
The real question that gets billions invested is "Is it useful?".
If the "con artist" solves my problem, that's fine by me. It is like having a mentalist tell me "I see that you are having a leaky faucet and I see your future in a hardware store buying a 25mm gasket and teflon tape...". In the end, I will have my leak fixed and that's what I wanted, who care how it got to it?
lukev
I don't disagree that "is it useful" is the important question.
The amount of money being invested is very clearly disproportionate to the current utility, and much of it is obviously based on the premise that LLMs can (or will soon) be able to think.
edanm
> The amount of money being invested is very clearly disproportionate to the current utility,
I don't think this is so clear. At least, if by "current utility" you also include potential future utility even without any advance in the underlying models.
Most of the money invested in the 2000 bubble was lost, but that didn't mean the utility of the internet was overblown.
lukev
Well, I find myself forced to agree with you there. I think these models are tremendously useful for data processing, and that chasing "reasoning" and the production of artifacts for human consumption are entirely red herrings.
everdrive
I think at best, there's a wide gulf between what LLMs can actually do, and what people _believe_ they can do. This is not to say that LLMs are not useful, but just that people broadly and regularly misunderstand their usefulness. This is not necessarily the fault of LLMs, but it does render their usage and adoption a bit problematic.
seunosewa
I think that people figure out what LLMs are capable of doing well pretty quickly after they start using them. I'd say that their capabilities are underestimated by many people who once tried to use non-reasoning LLMs to do things that reasoning LLMs can do very well today.
yannyu
It has to both be useful and economical. If that answer cost $2 to get you and a search+youtube video would have been just as effective and much cheaper, then it's possible that the new way to get the answer isn't significantly better than the old way.
The graveyards of startups are littered with economically infeasible solutions to problems.
null
cratermoon
So far it seems LLM-based systems are still reaching for a use.
crummy
Does copilot not count?
cratermoon
If copilot is so good why does MS/Github keep pushing it on people who haven't asked for it and don't want it? They're now giving it away for free, just to be able to put it everywhere.
orbital-decay
>LLMs <snip> do not meaningfully share any of the mechanisms that animals or people use to reason or think.
This seems to be a hard assumption the entire post, and many other similar ones, rely upon. But how do you know how people think or reason? How do you know human intelligence is not an illusion? Decades of research were unable to answer this. Now when LLMs are everywhere, suddenly everybody is an expert in human thinking with extremely strong opinions. To my vague intuition (based on understanding of how LLMs work) it's absolutely obvious they do share at least some fundamental mechanisms, regardless of vast low-level architecture/training differences. The entire discussion on whether it's real intelligence or not is based on ill-defined terms like "intelligence", so we can keep going in circles with it.
By the way, OpenAI does nothing of this, see [1]:
>artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work
Neither do others. So the author describes "tech industry" unknown to me.
aaplok
> Decades of research were unable to answer this.
From the article:
> The field of AI research has a reputation for disregarding the value of other fields [...] It’s likely that, being unaware of much of the research in psychology on cognitive biases or how a psychic’s con works, they stumbled into a mechanism and made chatbots that fooled many of the chatbot makers themselves.
Just because cognitive scientists don't know everything about how intelligence works (or on what it is) doesn't mean that they know nothing. There has been a lot of progress in cognitive science, in the last decade in particular on reasoning.
> based on ill-defined terms like "intelligence".
The whole discussion is about "artificial intelligence". Arguably AI researchers ought to have a fairly well defined stance of what "intelligence" means and can't use a trick like "nobody knows what intelligence is" to escape criticism.
orbital-decay
>Just because cognitive scientists don't know everything about how intelligence works (or on what it is) doesn't mean that they know nothing.
Author's claim is pretty strong: that human intelligence and what is called GenAI have no common ground at all. This is untrue at least intuitively for the entire field of ML. "Intuitively" because it cannot be proven or disproven until you know exactly what is human intelligence, or whatever the author means by thinking or reasoning.
>Arguably AI researchers ought to have a fairly well defined stance of what "intelligence" means and can't use a trick like "nobody knows what intelligence is" to escape criticism.
If you don't formalize your definition, the discussion can be easily escaped by either side by simply moving along the abstraction tree. The tiresome stochastic parrot/token predictor argument is about trees while the human intelligence is discussed in terms of a forest. And if you do formalize it, it's possible to discover that human intelligence is not what it seems either. I'm not even starting on the difference between individual intelligence, collective intelligence, and biological evolution, it's not easy to define where one ends and another begins.
AI researchers mainly focus on usefulness (see the definition above). The proof is in the pudding. Philosophical discussions are fine to have, but pretty meaningless at best and designed to support a narrative at worst.
aaplok
>Author's claim is pretty strong: that human intelligence and what is called GenAI have no common ground at all.
I feel that you're being unfair to the author here; the quote you responded to in your GP post alluded to "reason or think", and their argument is that LLMs don't. This is more specific than the sweeping statement you attribute to them.
> AI researchers mainly focus on usefulness (see the definition above).
And usefulness is something the article doesn't touch on, I think? The point of the article is that some users attribute capabilities to LLMs that can be explained with the same mechanisms as the capabilities they attribute to psychics (which are well understood to be nonexistent).
> Philosophical discussions are fine to have, but pretty meaningless at best and designed to support a narrative at worst.
Why this is relevant to AI research is (in my interpretation) that it is known to be hard for humans collectively to evaluate how intelligent an entity really is. We are easily fooled into seeing "intelligence" where it is not. This is something that cognitive scientists have spent a lot of time thinking about, and may perhaps be able to comment on.
For what it's worth I've seen cognitive scientists use AI to do cools stuff. I remember seeing someone who was using AI to show that it is possible to build inference without language (I am speaking from memory here, it was a while ago and I lost the reference sadly so I hope I'm not deforming it too much). She was certainly not claiming that her experiments showed how intelligence worked, but only that they pointed to the fact that language does not have to be a prerequisite for building inferences. Interesting stuff though without the sensationalism that sometimes accompanies the advocacy of LLMs.
BehindBlueEyes
> The field of AI research has a reputation for disregarding the value of other fields [...] It’s likely that, being unaware of much of the research in psychology on cognitive biases or how a psychic’s con works
Is the reputation warranted? Just a US thing? Or maybe the question is "since when did this change"? because in the mid 2000s in france at least, llm research was led by cognitive psychology professors who dabble in programming or had partnerships with a nearby technical university.
aaplok
I am not based in the US, nor in France, so I can't say what the situation is there. I was only pointing out that quote to GP.
My experience isn't much, since I am neither doing AI nor cognitive science but I have seen cognitive scientists do cool stuff with AI as a means to study cognitive science, and I have seen CS researchers involving themselves into the world of AI with varying success.
I would not be as emphatic as the author of the article but I would say that a good portion of AI research lost the focus on what is intelligence and instead just aimed at getting computers to perform various tasks. Which is completely fine until these researchers start claiming that they have produced intelligent machines (a manifestation of the Dunning Kruger effect).
mewpmewp2
As far as I know the best definition of intelligence is "ability to solve problems".
nurettin
The best we've got is "produce some speech that may contain the answer, BUT craft the question in a way that is more likely to generate an answer".
aaplok
best definition according to whom?
There is a lot of work to define intelligence, both in the context of cognitive science and in the context of AI [0].
I haven't spent enough time looking for a good review article, but for example, this article [1] says this: "Intelligence in the strict sense is the ability to know with conscience. Knowing with conscience implies awareness." and contrasts it with this: "Intelligence, in a broad sense, is the ability to process information. This can be applied to plants, machines, cells, etc. It does not imply knowledge." If you are interested in the topic the whole article is interesting and worth reading.
[0] https://scholar.google.com/scholar?hl=en&as_sdt=0%2C5&as_ylo...
[1] https://www.cell.com/heliyon/fulltext/S2405-8440(21)00373-X
swaraj
You should try the arc agi puzzles yourself, and then tell me you think these things aren't intelligent
https://arcprize.org/blog/openai-o1-results-arc-prize
I wouldn't say it's full agi or anything yet, but these things can definitely think in a very broad sense of the word
gessha
[Back in the 1980s]
You should try to play chess yourself, and then tell me you think these things aren't intelligent.
jhbadger
While I agree that we should be skeptical about the reasoning capabilities of LLMs, comparing them to chess programs misses the point. Chess programs were specifically created to play chess. That's all they could do. They couldn't generalize and play other board games, even related games like Shogi and Xiangqi, the Japanese and Chinese versions of chess. LLMs are amazing at being able to do things they never were programmed to do simply by accident.
gessha
Are they though? They’ve been shown to generalize poorly to tasks where you switch up some of the content.
dosinga
This feels rather forced. The article seems to claim both that LLMs don't actually work, it is all an illusion and that of course the LLMs know everything, they stole all our work from the last 20 years by scraping the internet and underpaying people to produce content. If it was a con, it wouldn't have to do that. Or in other words, if you had a psychic who actually memorized all biographies of all people ever, they wouldn't need their cons
pona-a
Why would it have to be one or the other? Yes, it's been proven LLMs do create world models, how good they are is a separate matter. There still could be goal misalignment, especially when it comes to RLHF.
If the model has in its internal world model knowledge it likely does not know how to solve a coding question, but the RLHF stage has reviewers rate refusals lower, it would in turn force its hand when it comes to tricks it knows it can pull based on its model of human reviewers. It can only implement the surface level boilerplate and pass that off as a solution, write its code in APL to obfuscate its lack of understanding, or keep misinterpreting the problem into a simpler one.
A psychic that read on ten thousand biographies might start to recall them, or he might interpolate the blanks with a generous dose of BS, or more likely do both in equal measure.
dosinga
Thanks. That is a good point. The RLHF phase indeed might "force" the LLM to adopt con artist tricks and probably does.
pama
This is from 2023 and is clearly dated. It is mildly interesting to notice how quickly things changed since then. Nowadays models can solve original math puzzles much of the time and it is harder to argue they cannot reason when we have access to R1, o1, and o3-mini.
Terr_
> Nowadays models can solve original math puzzles much of the time
Isn't that usually by not even trying, and delegating the work regular programs?
bbor
In what way is your mathematical talent truly you, but a python tool called by an LLM-centric agent not truly that agent?
Terr_
For starters, it means you should not take the success of the math and ascribe it to an advance in the LLM, or whatever phrase is actually being used to describe the the new fancy target of hype and investment.
An LLM is at best, a possible future component of the speculative future being sold today.
How might future generations visualize this? I'm imagining some ancient Greeks, who have invented an inefficient reciprocating pump, which they declare is a heart and that means they've basically built a person. (At the time, many believed the brain was just there to cool the blood.) Look! The fluid being pumped can move a lever: It's waving to us.
null
null
tmnvdb
I'm amazed people are upvoting this piece which does not grapple with any of the real issues in a serious way. I guess some folks just really want AI to go away and are longing to hear that it is just all newfangled nonsense from the city slickers!
grayhatter
how come you elected not to enumerate *any* of the real issues you expected?
ImPostingOnHN
looks like nobody asked
if you wish for GP to do that, ask them to do that
EagnaIonat
I was hoping it was talking about how it can resonate with users using those techniques. Or some experiments to prove the point. But it is not even that.
There is nothing of substance in this and it feels like the author has a grudge against LLMs.
thinkingemote
Interesting. Resonate with people is what a mentalist does. More generically a good bedside manner is what a good doctor does and it's the same thing. Many times we like being cold read and comforted and comfortable!
Agreed on the experiments. What would they look like? Can a chat bot give the same info without any bedside manner?
manmal
Well they have a book to sell, at the bottom of the article.
bbor
Absolutely -- there is a whole cottage industry of people who know that LLMs are big news and decide to chime in, without ever considering that maybe this is a scientific discussion, not a cultural/"take-driven" one. Bluesky in particular is probably their biggest source of income!
You can spot them easily, because instead of critiquing some specific thing and sticking to it, they can't resist throwing in "obviously, LLMs are all 100% useless and anyone who says otherwise is a Tech Bro" somewhere. Like:
completely unknown processes that have no parallel in the biological world.
c'mon... Anyone who knows a tiny bit about ML knows that both of those claims are just absurdly off base.s1mplicissimus
[flagged]
karmakaze
AlphaGo also doesn't reason. That doesn't mean it can't do things that humans do by reasoning. It doesn't make sense to make these comparisons. It's like saying that planes don't really fly because they aren't flapping their wings.
Edit: Don't conflate mechanisms with capabilities.
psytrancefan
At this point I think it is because of a type of religious sentimentality about the sanctity of human reasoning.
To take your analogy even further, it is like asking when is the plane going to improve enough that it can really fly by flapping it's wings.
throwaway87543
Some people seem to want to believe that true thought is spiritual in nature. They will never accept that something physical and made by man could do it. They would stop believing humans are intelligent if given conclusive proof of how the brain works.
viach
> 1 The tech industry has accidentally invented the initial stages a completely new kind of mind, based on completely unknown principles...
> 2) The intelligence illusion is in the mind of the user and not in the LLM itself.
3) The intelligence of the users is illusion either?
scandox
You're right! AI makes us ask really important questions about our own intelligence. I think it will lead to greater recognition that we are first and foremost animals: creatures of intention and action. I think we've put way too much emphasis on our intellectual dimension in the last few hundred years. To the point that some people started to believe that was what we are.
viach
Yup. And the real danger of AI is not that it enslaves humans but in that it will bring great disillusionment and existential crisis.
Someone should write a blog post about this to warn humanity.
ImPostingOnHN
the real danger isn't that it enslaves humans, but that it makes most humans useless to those with capital and robots and AI
at that point, capital can tell most humans to just go away and die, and can use their technology to protect themselves in the meantime
Earw0rm
Perhaps we're confusing intelligence with awareness.
What LLMs seem to emulate surprisingly well is something like a person's internal monologue, which is part of but not the whole of our mind.
It's as if it has the ability to talk to itself extremely quickly and while plugged directly into ~all of the written information humanity has ever produced, and what we see is the output of that hidden, verbally-reasoned conversation.
Something like that could be called intelligent, in terms of its ability to manipulate symbols and rearrange information, without having even a flicker of awareness, and entirely lacking the ability to synthesise new knowledge based on an intuitive or systemic understanding of a domain, as opposed to a complete verbal description of said domain.
Or to put it another way - it can be intelligent in terms of its utility, without possessing even an ounce of conscious awareness or understanding.
> 1 The tech industry has accidentally invented the initial stages a completely new kind of mind, based on completely unknown principles...
> 2) The intelligence illusion is in the mind of the user and not in the LLM itself.
I've felt as though there is something in between. Maybe:
3) The tech industry invented the initial stages a kind of mind that, though misses the mark, is approaching something not too dissimilar to how an aspect of human intelligence works.
> By using validation statements, … the chatbot and the psychic both give the impression of being able to make extremely specific answers, but those answers are in fact statistically generic.
"Mr. Geller, can you write some Python code for me to convert a 1-bit .bmp file to a hexadecimal string?"
Sorry, even if you think the underlying mechanisms have some sort of analog there's real value in LLM's, not so psychics doing "cold readings".