top | item 45802029

The Case That A.I. Is Thinking

278 points| ascertain | 4 months ago |newyorker.com

https://archive.ph/fPLJH

1011 comments

order
[+] tkz1312|4 months ago|reply
Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

Consciousness or self awareness is of course a different question, and ones whose answer seems less clear right now.

Knee jerk dismissing the evidence in front of your eyes because you find it unbelievable that we can achieve true reasoning via scaled matrix multiplication is understandable, but also betrays a lack of imagination and flexibility of thought. The world is full of bizarre wonders and this is just one more to add to the list.

[+] keiferski|4 months ago|reply
I don’t see how being critical of this is a knee jerk response.

Thinking, like intelligence and many other words designating complex things, isn’t a simple topic. The word and concept developed in a world where it referred to human beings, and in a lesser sense, to animals.

To simply disregard that entire conceptual history and say, “well it’s doing a thing that looks like thinking, ergo it’s thinking” is the lazy move. What’s really needed is an analysis of what thinking actually means, as a word. Unfortunately everyone is loathe to argue about definitions, even when that is fundamentally what this is all about.

Until that conceptual clarification happens, you can expect endless messy debates with no real resolution.

“For every complex problem there is an answer that is clear, simple, and wrong.” - H. L. Mencken

[+] layer8|4 months ago|reply
Sometimes after a night’s sleep, we wake up with an insight on a topic or a solution to a problem we encountered the day before. Did we “think” in our sleep to come up with the insight or solution? For all we know, it’s an unconscious process. Would we call it “thinking”?

The term “thinking” is rather ill-defined, too bound to how we perceive our own wakeful thinking.

When conversing with LLMs, I never get the feeling that they have a solid grasp on the conversation. When you dig into topics, there is always a little too much vagueness, a slight but clear lack of coherence, continuity and awareness, a prevalence of cookie-cutter verbiage. It feels like a mind that isn’t fully “there” — and maybe not at all.

I would agree that LLMs reason (well, the reasoning models). But “thinking”? I don’t know. There is something missing.

[+] notepad0x90|4 months ago|reply
I don't get why you would say that. it's just auto-completing. It cannot reason. It won't solve an original problem for which it has no prior context to "complete" an approximated solution with. you can give it more context and more data,but you're just helping it complete better. it does not derive an original state machine or algorithm to solve problems for which there are no obvious solutions. it instead approximates a guess (hallucination).

Consciousness and self-awareness are a distraction.

Consider that for the exact same prompt and instructions, small variations in wording or spelling change its output significantly. If it thought and reasoned, it would know to ignore those and focus on the variables and input at hand to produce deterministic and consistent output. However, it only computes in terms of tokens, so when a token changes, the probability of what a correct response would look like changes, so it adapts.

It does not actually add 1+2 when you ask it to do so. it does not distinguish 1 from 2 as discrete units in an addition operation. but it uses descriptions of the operation to approximate a result. and even for something so simple, some phrasings and wordings might not result in 3 as a result.

[+] simulator5g|4 months ago|reply
Having seen photocopiers so many times produce coherent, sensible, and valid chains of words on a page, I am at this point in absolutely no doubt that they are thinking.
[+] geon|4 months ago|reply
Having seen LLMs so many times produce incoherent, nonsensical and invalid chains of reasoning...

LLMs are little more than RNGs. They are the tea leaves and you read whatever you want into them.

[+] burnte|4 months ago|reply
The first principle is that you must not fool yourself, and you are the easiest person to fool. - Richard P. Feynman

They're not thinking, we're just really good at seeing patterns and reading into things. Remember, we never evolved with non-living things that could "talk", we're not psychologically prepared for this level of mimicry yet. We're still at the stage of Photography when people didn't know about double exposures or forced perspective, etc.

[+] marcus_holmes|4 months ago|reply
Yes, I've seen the same things.

But; they don't learn. You can add stuff to their context, but they never get better at doing things, don't really understand feedback. An LLM given a task a thousand times will produce similar results a thousand times; it won't get better at it, or even quicker at it.

And you can't ask them to explain their thinking. If they are thinking, and I agree they might, they don't have any awareness of that process (like we do).

I think if we crack both of those then we'd be a lot closer to something I can recognise as actually thinking.

[+] ben_w|4 months ago|reply
> Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

While I'm not willing to rule *out* the idea that they're "thinking" (nor "conscious" etc.), the obvious counter-argument here is all the records we have of humans doing thinking, where the records themselves are not doing the thinking that went into creating those records.

And I'm saying this as someone whose cached response to "it's just matrix multiplication it can't think/be conscious/be intelligent" is that, so far as we can measure all of reality, everything in the universe including ourselves can be expressed as matrix multiplication.

Falsification, not verification. What would be measurably different if the null hypothesis was wrong?

[+] satisfice|4 months ago|reply
I think you are the one dismissing evidence. The valid chains of reasoning you speak of (assuming you are talking about text you see in a “thinking model” as it is preparing its answer) are narratives, not the actual reasoning that leads to the answer you get.

I don’t know what LLMs are doing, but only a little experimentation with getting it to describe its own process shows that it CAN’T describe its own process.

You can call what a TI calculator does “thinking” if you want. But what people are interested in is human-like thinking. We have no reason to believe that the “thinking” of LLMs is human-like.

[+] josefx|4 months ago|reply
Counterpoint: The seahorse emoji. The output repeats the same simple pattern of giving a bad result and correcting it with another bad result until it runs out of attempts. There is no reasoning, no diagnosis, just the same error over and over again within a single session.
[+] techblueberry|4 months ago|reply
Isn’t anthropomorphizing LLMs rather than understanding their unique presence in the world a “ lack of imagination and flexibility of thought”? It’s not that I can’t imagine applying the concept “thinking” to the output on the screen, I just don’t think it’s an accurate description.
[+] didibus|4 months ago|reply
I guess it depends if you definite thinking thinking as chaining coherent reasoning sentences together 90-some% of the time.

But if you define thinking as the mechanism and process we mentally undergo and follow mentally... I don't think we have any clue if that's the same. Do we also just vector-map attention tokens and predict the next with a softmax? I doubt, and I don't think we have any proof that we do.

[+] ph4rsikal|4 months ago|reply
It might appear so, but then you could validate it with a simple test. If the LLM would play a 4x4 Tic Tac Toe game, would the agent select the winning move 100% of all time or block a losing move 100% of the time? If these systems were capable of proper reasoning, then they would find the right choice in these obvious but constantly changing scenarios without being specifically trained for it.

[1] https://jdsemrau.substack.com/p/nemotron-vs-qwen-game-theory...

[+] johnnienaked|4 months ago|reply
If you understand how they operate and you are reasonable and unbiased there is no way you could consider it thinking
[+] noiv|4 months ago|reply
Different PoV: You have a local bug and ask the digital hive mind for a solution, but someone already solved the issue and their solution was incorporated... LLMs are just very effficient at compressing billions of solutions into a few GB.

Try to ask something no one ever came up with a solution so far.

[+] conartist6|4 months ago|reply
Yeah but if I assign it a long job to process I would also say that an x86 CPU is "thinking" about a problem for me.

What we really mean in both cases is "computing," no?

[+] ForHackernews|4 months ago|reply
But all those times the same system produces irrational gibberish don't count? GPT-5 will commonly make mistakes no thinking human could ever make.

Human: I'm trying to get my wolf, sheep and cabbage across the river in this boat, but the wolf keeps eating the sheep or the sheep eats the cabbage

Bot: You should put the sheep in the boat and take it across — if we delve into the biology of Canis lupus we discover that wolves don't eat cabbage!

H: Ok, so that worked great so far, the sheep is on one side and the wolf/cabbage is on the other.

B: Now, Option 1 is to bring the wolf across, or Option 2 you can bring the cabbage. I recommend (2) taking the cabbage as cabbages are smaller and easier to transport in a boat.

H: But then the sheep eats the cabbage, right? Remember that?

B: Exactly, that's sharp thinking. If you put the sheep and the cabbage together on the same side of the river, the sheep is sure to devour the cabbage. We need to not just separate sheep from cabbages — we need to separate cabbages from sheep! :rocketship:

[+] intended|4 months ago|reply
what sound does a falling tree make if no one is listening?

I’ve asked LLMs to write code for me in fields I have little background knowledge, and then had to debug the whole thing after essentially having to learn the language and field.

On the other hand, for things I am well versed in, I can debug the output and avoid entire swathes of failed states, by having a clear prompt.

Its why I now insist that any discussion on GenAI projects also have the speaker mention the level of seniority they have ( proxy for S/W eng experience), Their familiarity with the language, the project itself (level of complexity) - more so than the output.

I also guarantee - that most people have VERY weak express knowledge of how their brains actually work, but deep inherent reflexes and intuitions.

[+] raincole|4 months ago|reply
I'd represent the same idea but in a different way:

I don't know what the exact definition of "thinking" is. But if a definition of thinking rejects the possibility of that current LLMs think, I'd consider that definition useless.

[+] NoMoreNicksLeft|4 months ago|reply
>Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

If one could write a quadrillion-line python script of nothing but if/elif/else statements nested 1 million blocks deep that seemingly parsed your questions and produced seemingly coherent, sensible, valid "chains of reasoning"... would that software be thinking?

And if you don't like the answer, how is the LLM fundamentally different from the software I describe?

>Knee jerk dismissing the evidence in front of your eyes because

There is no evidence here. On the very remote possibility that LLMs are at some level doing what humans are doing, I would then feel really pathetic that humans are as non-sapient as the LLMs. The same way that there is a hole in your vision because of a defective retina, there is a hole in your cognition that blinds you to how cognition works. Because of this, you and all the other humans are stumbling around in the dark, trying to invent intelligence by accident, rather than just introspecting and writing it out from scratch. While our species might someday eventually brute force AGI, it would be many thousands of years before we get there.

[+] dmz73|4 months ago|reply
Having seen LLMs so many time produce incoherent, nonsense, invalid answers to even simplest of questions I cannot agree with categorization of "thinking" or "intelligence" that applies to these models. LLMs do not understand what they "know" or what they output. All they "know" is that based on training data this is most likely what they should output + some intentional randomization to make it seem more "human like". This also makes it seem like they create new and previously unseen outputs but that could be achieved with simple dictionary and random number generator and no-one would call that thinking or intelligent as it is obvious that it isn't. LLMs are better at obfuscating this fact by producing more sensible output than just random words. LLMs can still be useful but they are a dead-end as far as "true" AI goes. They can and will get better but they will never be intelligent or think in the sense that most humans would agree those terms apply. Some other form of hardware/software combination might get closer to AI or even achieve full AI and even sentience but that will not happen with LLMs and current hardware and software.
[+] ryanackley|4 months ago|reply
I think we can call it "thinking" but it's dangerous to anthropomorphize LLMs. The media and AI companies have an agenda when doing so.
[+] outworlder|4 months ago|reply
They may not be "thinking" in the way you and I think, and instead just finding the correct output from a really incredibly large search space.

> Knee jerk dismissing the evidence in front of your eyes

Anthropomorphizing isn't any better.

That also dismisses the negative evidence, where they output completely _stupid_ things and make mind boggling mistakes that no human with a functioning brain would do. It's clear that there's some "thinking" analog, but there are pieces missing.

I like to say that LLMs are like if we took the part of our brain responsible for language and told it to solve complex problems, without all the other brain parts, no neocortex, etc. Maybe it can do that, but it's just as likely that it is going to produce a bunch of nonsense. And it won't be able to tell those apart without the other brain areas to cross check.

[+] jimbohn|4 months ago|reply
It's reinforcement learning applied to text, at a huge scale. So I'd still say that they are not thinking, but they are still useful. The question of the century IMO is if RL can magically solve all our issues when scaled enough.
[+] IAmGraydon|4 months ago|reply
>Knee jerk dismissing the evidence in front of your eyes because you find it unbelievable that we can achieve true reasoning via scaled matrix multiplication is understandable, but also betrays a lack of imagination and flexibility of thought.

You go ahead with your imagination. To us unimaginative folks, it betrays a lack of understanding of how LLMs actually work and shows that a lot of people still cannot grasp that it’s actually an extremely elaborate illusion of thinking.

[+] mft_|4 months ago|reply
Personal take: LLMs are probably part of the answer (to AGI?) but are hugely handicapped by their current architecture: the only time that long-term memories are formed is during training, and everything after that (once they're being interacted with) sits only in their context window, which is the equivalent of fungible, fallible, lossy short-term memory. [0] I suspect that many things they currently struggle with can be traced back to this.

Overcome this fundamental limitation and we'll have created introspection and self-learning. However, it's hard to predict whether this will allow them to make novel, intuitive leaps of discovery?

[0] It's an imperfect analogy, but we're expecting perfection from creations which are similarly handicapped as Leonard Shelby in the film Memento.

[+] almosthere|4 months ago|reply
Well, I think because we know how the code is written, in the sense that humans quite literally wrote the code for it - it's definitely not thinking, and it is literally doing what we asked, based on the data we gave it. It is specifically executing code we thought of. The output of course, we had no flying idea it would work this well.

But it is not sentient. It has no idea of a self or anything like that. If it makes people believe that it does, it is because we have written so much lore about it in the training data.

[+] everdrive|4 months ago|reply
This is merely a debate about what it means to "think." We didn't really previously need to disambiguate thinking / intelligence / consciousness / sentience / ego / identity / etc.

Now, we do. Partly because of this we don't have really well defined ways to define these terms and think about. Can a handheld calculator think? Certainly, depending on how we define "think."

[+] monkeycantype|4 months ago|reply
Until we have a testable, falsifiable thesis of how consciousness forms in meat, it is rash to exclude that consciousness could arise from linear algebra. Our study of the brain has revealed an enormous amount about how our anatomy processes information, but nothing of substance on the relationship between matter and consciousness. The software and data of an operating LLM is not purely abstract, it has a physical embodiment as circuits and electrons. Until we understand how matter is connected to consciousness, we also cannot know whether the arrangements and movements of electrons meet the criteria for forming consciousness.
[+] ale|4 months ago|reply
This reads like 2022 hype. It's like people stil do not understand that there's a correlation between exaggerating AI's alleged world-threatening capabilities and AI companies' market share value – and guess who's doing the hyping.
[+] b00ty4breakfast|4 months ago|reply
all this "AI IS THINKING/CONSCIOUS/WHATEVER" but nobody seems worried of that implication that, if that is even remotely true, we are creating a new slave market. This either implies that these people don't actually believes any of this boostering rhetoric and are just cynically trying to cash in or that the technical milieu is in a profoundly disturbing place ethically.

To be clear, I don't believe that current AI tech is ever going to be conscious or win a nobel prize or whatever, but if we follow the logical conclusions to this fanciful rhetoric, the outlook is bleak.

[+] sbdaman|4 months ago|reply
I've shared this on YN before but I'm a big fan of this piece by Kenneth Taylor (well, an essay pieced together from his lectures).

The Robots Are Coming

https://www.bostonreview.net/articles/kenneth-taylor-robots-...

"However exactly you divide up the AI landscape, it is important to distinguish what I call AI-as-engineering from what I call AI-as-cognitive-science. AI-as-engineering isn’t particularly concerned with mimicking the precise way in which the human mind-brain does distinctively human things. The strategy of engineering machines that do things that are in some sense intelligent, even if they do what they do in their own way, is a perfectly fine way to pursue artificial intelligence. AI-as-cognitive science, on the other hand, takes as its primary goal that of understanding and perhaps reverse engineering the human mind.

[...]

One reason for my own skepticism is the fact that in recent years the AI landscape has come to be progressively more dominated by AI of the newfangled 'deep learning' variety [...] But if it’s really AI-as-cognitive science that you are interested in, it’s important not to lose sight of the fact that it may take a bit more than our cool new deep learning hammer to build a humanlike mind.

[...]

If I am right that there are many mysteries about the human mind that currently dominant approaches to AI are ill-equipped to help us solve, then to the extent that such approaches continue to dominate AI into the future, we are very unlikely to be inundated anytime soon with a race of thinking robots—at least not if we mean by “thinking” that peculiar thing that we humans do, done in precisely the way that we humans do it."

[+] Philadelphia|4 months ago|reply
People have a very poor conception of what is easy to find on the internet. The author is impressed by the story about Chat GPT telling his friend how to enable the sprinkler system for his kids. But I decided to try just googling it — “how do i start up a children's park sprinkler system that is shut off” — and got a Youtube video that shows the same thing, plus a lot of posts with step by step directions. No AI needed. Certainly no evidence of advanced thinking.
[+] ivraatiems|4 months ago|reply
The author searches for a midpoint between "AIs are useless and do not actually think" and "AIs think like humans," but to me it seems almost trivially true that both are possible.

What I mean by that is that I think there is a good chance that LLMs are similar to a subsystem of human thinking. They are great at pattern recognition and prediction, which is a huge part of cognition. What they are not is conscious, or possessed of subjective experience in any measurable way.

LLMs are like the part of your brain that sees something and maps it into a concept for you. I recently watched a video on the creation of AlexNet [0], one of the first wildly successful image-processing models. One of the impressive things about it is how it moves up the hierarchy from very basic patterns in images to more abstract ones (e. g. these two images' pixels might not be at all the same, but they both eventually map to a pattern for 'elephant').

It's perfectly reasonable to imagine that our brains do something similar. You see a cat, in some context, and your brain maps it to the concept of 'cat', so you know, 'that's a cat'. What's missing is a) self-motivated, goal-directed action based on that knowledge, and b) a broader context for the world where these concepts not only map to each other, but feed into a sense of self and world and its distinctions whereby one can say: "I am here, and looking at a cat."

It's possible those latter two parts can be solved, or approximated, by an LLM, but I am skeptical. I think LLMs represent a huge leap in technology which is simultaneously cooler than anyone would have imagined a decade ago, and less impressive than pretty much everyone wants you to believe when it comes to how much money we should pour into the companies that make them.

[0] https://www.youtube.com/watch?v=UZDiGooFs54

[+] cs702|4 months ago|reply
Many people who object to the idea that current-generation AI is thinking do so only because they believe AI is not "conscious"... but there is no known law in the universe requiring that intelligence and consciousness must always go together. With apologies to René Descartes[a], intelligence and consciousness are different.

Intelligence can be verified and quantified, for example, with tests of common sense and other knowledge.[b] Consciousness, on the other hand, is notoriously difficult if not impossible to verify, let alone quantify. I'd say AI is getting more intelligent, and more reliable, in fits and starts, but it's not necessarily becoming conscious.

---

[a] https://en.wikipedia.org/wiki/Cogito%2C_ergo_sum

[b] For example, see https://arxiv.org/abs/2510.18212

[+] adamzwasserman|4 months ago|reply
I've written a full response to Somers' piece: The Case That A.I. Is Thinking: What The New Yorker Missed: https://emusings.substack.com/p/the-case-that-ai-is-thinking...

The core argument: When you apply the same techniques (transformers, gradient descent, next-token prediction) to domains other than language, they fail to produce anything resembling "understanding." Vision had a 50+ year head start but LLMs leapfrogged it in 3 years. That timeline gap is the smoking gun.

The magic isn't in the neural architecture. It's in language itself—which exhibits fractal structure and self-similarity across scales. LLMs navigate a pre-existing map with extraordinary regularity. They never touch the territory.

[+] cyrusradfar|4 months ago|reply
I think the challenge with many of these conversations is that they assume consciousness emerges through purely mechanical means.

The “brain as a computer” metaphor has been useful in limited contexts—especially for modeling memory or signal processing; but, I don’t think it helps us move forward when talking about consciousness itself.

Penrose and Hameroff’s quantum consciousness hypothesis, while still very speculative, is interesting precisely because it suggests that consciousness may arise from phenomena beyond classical computation. If that turns out to be true, it would also mean today’s machines—no matter how advanced—aren’t on a path to genuine consciousness.

That said, AI doesn’t need to think to be transformative.

Steam engines weren’t conscious either, yet they reshaped civilization.

Likewise, AI and robotics can bring enormous value without ever approaching human-level awareness.

We can hold both ideas at once: that machines may never be conscious, and still profoundly useful.

[+] adamzwasserman|4 months ago|reply
The article misses three critical points:

1. Conflates consciousness with "thinking" - LLMs may process information effectively without being conscious, but the article treats these as the same phenomenon

2. Ignores the cerebellum cases - We have documented cases of humans leading normal lives with little to no brain beyond a cerebellum, which contradicts simplistic "brain = deep learning" equivalences

3. Most damning: When you apply these exact same techniques to anything OTHER than language, the results are mediocre. Video generation still can't figure out basic physics (glass bouncing instead of shattering, ropes defying physics). Computer vision has been worked on since the 1960s - far longer than LLMs - yet it's nowhere near achieving what looks like "understanding."

The timeline is the smoking gun: vision had decades of head start, yet LLMs leapfrogged it in just a few years. That strongly suggests the "magic" is in language itself (which has been proven to be fractal and already heavily compressed/structured by human cognition) - NOT in the neural architecture. We're not teaching machines to think.

We're teaching them to navigate a pre-existing map that was already built.

[+] djoldman|4 months ago|reply
TFA is a part of what seems like a never-ending series about concepts that lack a useful definition.

"Thinking" and "intelligence" have no testable definition or specification, therefore it's a complete waste of time to suppose that AI is thinking or intelligent.

[+] yalogin|4 months ago|reply
I don't see how we make the jump from current LLMs to AGI. May be it's my limited understanding of the research but current LLMs seem to not have any properties that indicate AGI. Would love to get thoughts from someone that understands it
[+] JonChesterfield|4 months ago|reply
The real question is not whether machines think but whether men do.
[+] j1mr10rd4n|4 months ago|reply
Geoffrey Hinton's recent lecture at the Royal Institute[1] is a fascinating watch. His assertion that human use of language being exactly analogous to neural networks with back-propagation really made me think about what LLMs might be able to do, and indeed, what happens in me when I "think". A common objection to LLM "intelligence" is that "they don't know anything". But in turn... what do biological intelligences "know"?

For example, I "know" how to do things like write constructs that make complex collections of programmable switches behave in certain ways, but what do I really "understand"?

I've been "taught" things about quantum mechanics, electrons, semiconductors, transistors, integrated circuits, instruction sets, symbolic logic, state machines, assembly, compilers, high-level-languages, code modules, editors and formatting. I've "learned" more along the way by trial and error. But have I in effect ended up with anything other than an internalised store of concepts and interconnections? (c.f. features and weights).

Richard Sutton takes a different view in an interview with Dwarkesh Patel[2] and asserts that "learning" must include goals and reward functions but his argument seemed less concrete and possibly just a semantic re-labelling.

[1] https://www.youtube.com/watch?v=IkdziSLYzHw [2] https://www.youtube.com/watch?v=21EYKqUsPfg

[+] mrob|4 months ago|reply
I don't believe LLMs can be conscious during inference because LLM inference is just repeated evaluation of a deterministic [0] pure function. It takes a list of tokens and outputs a set of token probabilities. Any randomness is part of the sampler that selects a token based on the generated probabilities, not the LLM itself.

There is no internal state that persists between tokens [1], so there can be no continuity of consciousness. If it's "alive" in some way it's effectively killed after each token and replaced by a new lifeform. I don't see how consciousness can exist without possibility of change over time. The input tokens (context) can't be enough to give it consciousness because it has no way of knowing if they were generated by itself or by a third party. The sampler mechanism guarantees this: it's always possible that an unlikely token could have been selected by the sampler, so to detect "thought tampering" it would have to simulate itself evaluating all possible partial contexts. Even this takes unreasonable amounts of compute, but it's actually worse because the introspection process would also affect the probabilities generated, so it would have to simulate itself simulating itself, and so on recursively without bound.

It's conceivable that LLMs are conscious during training, but in that case the final weights are effectively its dead body, and inference is like Luigi Galvani poking the frog's legs with electrodes and watching them twitch.

[0] Assuming no race conditions in parallel implementations. llama.cpp is deterministic.

[1] Excluding caching, which is only a speed optimization and doesn't affect results.

[+] lbrandy|4 months ago|reply
I have no idea how you can assert what is necessary/sufficient for consciousness in this way. Your comment reads like you believe you understand consciousness far more than I believe anyone actually does.
[+] jdauriemma|4 months ago|reply
I don't think the author is saying that LLMs are conscious or alive.
[+] dagss|4 months ago|reply
Thinking != consciousness