parpfish
4 days ago
My take on stochastic parrots is the similar to the authors concluding section.
This debate isn’t about the computations underlying cognition, it’s about wanting to feel special.
The contention that “it’s a stochastic parrot” usually implied “it’s merely a stochastic parrot, and we know that we must be so much more than that, so obviously this thing falls short”.
But… there never was any compelling proof that we’re anything more than stochastic parrots. Moreover, those folks would say that any explanation of cognition falls short because they can always move the goal posts to make sure that humans are special.
jfengel
4 days ago
The way I see it, the argument comes down to an assertion that as impressive as these technologies are, they are a local maximum that will always have some limitation that keeps it from feeling truly human (to us).
That's an assertion that has not been proven one way or the other. It's certainly true that progress has leveled off after an extraordinary climb, and most people would say it's still not a fully general intelligence yet. But we don't know if the next step requires incremental work or if it requires a radically different approach.
So it's just taking a stance on an unproven assertion rather than defining anything fundamental.
moritonal
4 days ago
True, and worth appreciating how Humans hit some pretty blunt local maxima that machines have long since suppassed such as land speed or operations per second.
AndrewDucker
3 days ago
Do you really just complete the next token when speaking? You don't plan ahead, you don't form concepts and then translate them into speech? You don't make models, work with them, carefully consider their interactions, and then work out ways to communicate them?
Because, with a reasonable understanding of how LLMS work, the way that they produce text is nothing like the way that my mind works.
ToValueFunfetti
3 days ago
LLMs very probably don't work the same way humans do, but they do appear to both plan ahead and translate internal concepts into speech: https://www.anthropic.com/news/tracing-thoughts-language-mod...
Suppafly
3 days ago
>Do you really just complete the next token when speaking? You don't plan ahead, you don't form concepts and then translate them into speech? You don't make models, work with them, carefully consider their interactions, and then work out ways to communicate them?
A good amount of people spend most of their lives operating that way.
alganet
4 days ago
A machine that can, unattended, move the goal post to make itself look special by inventing new forms of expression then would beat "those folks" for good.
Not the best human behavior, but certainly human behavior.
LLMs are built with the intention of mimicry. It's no surprise they look like mimicry. If we show a new trick they can't do, and their makers train it to mimic that, can we be blamed for calling that mimicry... mimicry?
golly_ned
4 days ago
The promise of LLMs isn’t that they appear to be intelligent through mimicry, but that they do understand.
gizmo686
4 days ago
How do LLMs promise to "understand". Broadly speaking, AI/ML can be divided into two groups: mimicary, which is given a corpus of assumed good data and attempts to generalize it, and reinforcement learning where the AI is given a high level fitness function and set loose to optimize it.
The current generation of LLM falls pretty heavy in the mimicary family of AI.
smokel
4 days ago
This is a false dichotomy.
You introduce two categories, state that LLMs are not part of one category, and then conclude that it must be in the other. In reality, the distinction between the two classes is not so clear.
The transformer architecture is quite something, and the number of layers and nodes involved in a typical LLM is staggering. This goes way beyond linear regression.
habinero
3 days ago
In my experience, it's a phrase use to mock the weirdos who anthropomorphize and fetishize a statistics model into some fantasy "intelligence", when it has none by design.
It's less "humans are special" and more "this ain't it, chief".
floydnoel
3 days ago
well said. as i saw put recently, "LLMs are calculators for text" and nobody seems to get confused whether calculators are sentient
user
4 days ago
MichaelZuo
4 days ago
Well it’s even more dismal in reality.
Gather enough parrots on a stage and at least one can theoretically utter a series of seemingly meaningfuly word-like sounds that is legitimately novel, that has never been uttered before.
But I doubt any randomly picked HN user will actually accomplish that in fact before say age 40. Most people just don’t ever get enough meaningful speaking opportunities to make that statistically likely. There’s just too many tens of billions of people that have already existed and uttered words.
anon373839
4 days ago
That not reality, it’s theory.
MichaelZuo
4 days ago
Can you write down the actual argument?
It seems to be plausible, to me, given enough parrots.
thfuran
4 days ago
Novel utterances happen all the damn time. See https://venturebeat.com/business/15-of-all-google-searches-a... for tangential evidence.
Edit: actually that looks like it's just an offhand mention of Google's initial report, but I don't really feel like spending more time tracking down details to rebut so silly a claim.
MichaelZuo
4 days ago
Unique gibberish and spelling errors also count as a “unique search” so I don’t see how it relates.
Do you have an argument that makes sense?
user
3 days ago
anon373839
3 days ago
This is embarrassing, but I hastily misread your comment as saying something it didn’t say. So just disregard my comment altogether!
goatlover
4 days ago
We're conscious animals who communicate because we navigate social spaces, not because we're completing the next token. I wonder about hackers who think they're nothing more than the latest tech.
int_19h
4 days ago
You postulate it as if these two are mutually exclusive, but it's not at all clear why we can't be "completing the next token" to communicate in order to navigate social spaces. This last part is just where our "training" (as species) comes from, it doesn't really say anything about the mechanism.
goatlover
4 days ago
Because what's motivating our language is a variety of needs, emotions and experiences as social animals. As such we have goals and desires. We're not sitting there waiting to be prompted for some output.
int_19h
3 days ago
You constantly have input from all your senses, which is effectively your "prompt". If you stick a human into a sensory deprivation tank for long enough, very weird things happen.
parpfish
4 days ago
How do you know we’re not just completing the next token?
It seems eminently plausible that the way cognition works is to take in current context and select the most appropriate next action/token. In fact, it’s hard to think of a form of cognition that isnt “given past/context, predict next thing”
throwup238
4 days ago
Philosophers have been arguing a parallel point for centuries. Does intelligence require some sort of (ostensibly human-ish) qualia or does “if it quacks like a duck, it is a duck” apply?
I think it's better to look at large language models in the context of Wittgenstein. Humans are more than next token predictors because we participate in “language games” through which we experimentally build up a mental model for what each word means. LLMs learn to “rule follow” via a huge corpus of human text but there’s no actual intelligence there (in a Wittgensteinian analysis) because there’s no “participation” beyond RLHF (in which humans are playing the language games for the machine). There’s a lot to unpack there but that’s the gist of my opinion.
Until we get some rigorous definitions for intelligence or at least break it up into many different facets, I think pie in the sky philosophy is the best we can work with.
giardini
4 days ago
Trivially, because any two of rarely produce the same "next token"!
An ensemble of LLMs trained identically would generate the same next token(s) forever. But we don't - we generate different sequences.
We are not LLMs.
goatlover
4 days ago
If you ignore everything that makes us human to make some sort of analogy between brain activity and LLMs. Let us not forget they are tools we made to serve our goals.
bluefirebrand
4 days ago
> How do you know we’re not just completing the next token
Because we (humans) weren't born into a world with computers, internet, airplanes, satellites, etc
"Complete next token" means that everything is already in the data set. It can remix things in interesting ways, sure. But that isn't the same as creating something new
Edit: I would love to hear someone's idea about how you could "parrot" your way into landing people on the moon without any novel discovery or invention
aeonik
4 days ago
Everything is made out of just Protons, Neutrons, Electrons, along with some fields that allow interaction. (and Muons, Neutrinos, and a few others)
Everything that is physical is nothing but remixes and recombinations of a very small set of tokens.
bluefirebrand
4 days ago
> Everything that is physical is nothing but remixes and recombinations of a very small set of tokens.
We're not talking about "physical" with LLMs, we're talking about knowledge and creativity and reasoning, which are metaphysical.
The sum total of human knowledge cannot possibly be purely composed of remixes and recombinations, there has to be some baseline that humans invented for there to even be something to remix!
aeonik
3 days ago
All of that is rooted in physics though.
Lnowledge and creativity absolutely are physical things. It's clear from brain injury studies that there are very localized and specific functions to this creativity.
Drugs also clearly have a very physical affect on these attributes.
goatlover
4 days ago
You're conflating symbolic descriptions for the physical stuff itself.
aeonik
3 days ago
You're right to flag the distinction between symbols and substance, but I think you're misapplying it here.
I'm not conflating symbolic systems with the physical substrate: they're obviously different levels of abstraction. What I am saying is that symbolic reasoning, language, creativity, and knowledge all emerge from the same underlying physical processes. They're not magic. They're not floating in some Platonic realm. They’re instantiated in real, measurable patterns, whether in neurons or silicon.
You can't have metaphysics without physics. And we have solid evidence, from neuroscience, from pharmacology, from evolutionary biology, that the brain's symbolic output is fundamentally a physical phenomenon. Injuries, chemicals, electrical stimulation, they all modulate “metaphysical” experience in completely physical ways.
Emergence matters here. Yes, atoms aren’t thoughts, but enough atoms arranged the right way do start behaving like a thinking system. That’s the whole point of complex systems theory, chaos theory, and even early AI work like Hofstadter and Dennett. I recommend "Gödel, Escher, Bach", or Melanie Mitchell's "Complexity: A Guided Tour", if you're curious.
If you're arguing there's something else, some kind of unphysical or non-emergent component to knowledge or creativity, I'd honestly love to hear more, because that's a bold claim. But waving away the physical substrate as irrelevant doesn’t hold up under scrutiny.
triceratops
4 days ago
Everyone's computing the next token. Intelligence is computing the right token.
imtringued
4 days ago
Except when I encounter people like you, they are mostly interested in saying that LLMs are the be all end all of intelligence only to walk their statements back every time a new LLM innovation comes out that proves them wrong.
Since humans are just stochastic parrots, we don't need to add features or change anything about LLMs. Innovation is for the weak and stupid. We can just scale LLMs by doing nothing except adding more data, parameters and training time.
The status quo bias is unreal. I don't even know what purpose it serves other than discouraging technological progress. The people claiming to champion LLMs by denying the differences between humans and LLMs are their biggest enemies.
gsf_emergency_2
4 days ago
>moving the goalposts
I don't know whether to acknowledge that Kurt Goedel was a very special parrot
Because "moving the goalposts" (exceedingly intricately, one must add) was indeed a specialty of his (or Cantor's).
Otherwise, imho, humans are superior to parrots because we can kinda derive "understanding" in the face of paradox.. (also Einstein,Lobachevsky, special parrots?)
YMMV, I personally haven't tried to get fresh in this specific way with a you-get-whadya-pay-for chatbot..
(Though some say the average person is also an user of tools like Occam's razor, so maybe we are all stochastic crows?)