frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Open in hackernews

Chomsky on what ChatGPT is good for (2023)

https://chomsky.info/20230503-2/
96•mef•6h ago

Comments

schoen•6h ago
(2023)
BryanLegend•6h ago
Yean, a lot has happened in two years
asveikau•5h ago
From what I've heard, Chomsky had a stroke which impacted his language. You will, unfortunately, not hear a recent opinion from him on current developments.
throwaway314155•31m ago
Geez, talk about irony. That's terrible.
msh•6h ago
He should just surrender and give chatgpt whatever land it wants.
bigyabai•6h ago
Manufactured intelligence to modulate a world of manufactured consent!

I agree with the rest of these comments though, listening to Chomsky wax about the topic-du-jour is a bit like trying to take lecture notes from the Swedish Chef.

scrubs•6h ago
>"bit like trying to take lecture notes from the Swedish Chef."

I'll be liberally borrowing, and using that simile! It's hilarious. Bork, bork, bork!

The best thing is you can be right, and the other side can't take offense. It's the Muppets after all. It's brilliant!

klipt•6h ago
Dude is 96, so he definitely has a different perspective than most, for better or worse.
whattheheckheck•6h ago
3.35 hrs Chomsky interview on ML Street Talk https://youtu.be/axuGfh4UR9Q
next_xibalba•6h ago
Chomsky is always saying that LLMs and such can only imitate, not understand language. But I wonder if there is a degree of sophistication at which he would concede these machines exceed "imitation". If his point is that LLMs arrive at language in a way different than humans... great. But I'm not sure how he can argue that some kind of extremely sophisticated understanding of natural language is not embedded in these models in a way that, at this point, exceeds the average human. In all fairness, this was written in 2023, but given his longstanding stubbornness on this topic, I doubt it would make a difference.
icedrift•6h ago
From what I've read/watched of Chomsky he's holding out for something that truly cannot be distinguished from human no matter how hard you tried.
dghlsakjg•5h ago
Isn’t that just a Turing test?

I’m perfectly willing to bet that there are LLMs that can pass a Turing test, even against a mind like Chomsky.

mattnewton•5h ago
I think what would "convince" Chomsky is more akin to the explainability research currently in it's infancy, producing something akin to a branch of information theory for language and thought.

Chomsky talks about how the current approach can't tell you about what humans are doing, only approximate it; the example he has given in the past is taking thousands of hours of footage of falling leaves and then training a model to make new leaf falling footage versus producing a model of gravity, gas mechanics for the air currents, and air resistance model of leaves. The later representation is distilled down into something that tells you about what is happening at the end of some scientific inquiry, and the former is a opaque simulation for engineering purposes if all you wanted was more leaf falling footage.

So I interpret Chomsky as meaning "Look, these things can be great for an engineering purpose but I am unsatisfied in them for scientific research because they do not explain language to me" and mostly pushing back against people implying that the field he dedicated much of his life to is obsolete because it isn't being used for engineering new systems anymore, which was never his goal.

ramoz•5h ago
It's always good to humble the ivory tower.
xigency•4h ago
That's not quite a valid point considering the article's conclusion: sowing dissent in the sciences allows companies to more easily package and sell carcinogens like asbestos, lead paint, and tobacco products.

I understand his diction is a bit impenetrable but I believe the intention is to promote literacy and specificity, not just to be a smarty-pants.

flornt•5h ago
I guess it's because LLM does not understand the meaning as you understand what you read or thought. LLMs are machines that modulate hierarchical positions, ordering the placement of a-signifying sign without a clue of the meaning of what they ordered (that's why machine can hallucinate :they don't have a sense of what they express)
mrandish•6h ago
[Edit to remove: It was not clear that this was someone else's intro re-posted on Chomsky's site]
kweingar•6h ago
This is an interview published in Common Dreams, rehosted at Chomsky's site. Those are the interviewer's words, not Chomsky's.
mrandish•6h ago
Okay, I didn't know that. I'll delete my comment.
skydhash•6h ago
Quite a nice overview. For almost any specific measure, you can find something that is better than human at that point. And now LLMs architecture have made possible for computers to produce complete and internally consistent paragraphs of text, by rehashing all the digital data that can be found on the internet.

But what we're good as using all of our capabilities to transform the world around us according to an internal model that is partially shared between individuals. And we have complete control over that internal model, diverging from reality and converging towards it on whims.

So we can't produce and manipulate text faster, but rarely the end game is to produce and manipulate text. Mostly it's about sharing ideas and facts (aka internal models) and the control is ultimately what matters. It can help us, just like a calculator can help us solve an equation.

EDIT

After learning to draw, I have that internal model that I switch to whenever I want to sketch something. It's like a special mode of observation, where you no longer simply see, but pickup a lot of extra details according to all the drawing rules you internalized. There's not a lot, they're just intrinsically connected with each other. The difficult part is hand-eye coordination and analyzing the divergences between what you see and the internal model.

I think that's why a lot of artists are disgusted with AI generators. There's no internal models. Trying to extract one from a generated picture is a futile exercice. Same with generated texts. Alterations from the common understanding follows no patterns.

randmeerkat•5h ago
> It can help us, just like a calculator can help us solve an equation.

A calculator is consistent and doesn’t “hallucinate” answers to equations. An LLM puts an untrustworthy filter between the truth and the person. Google was revolutionary because it increased access to information. LLMs only obscure that access, while pretending to be something more.

skydhash•5h ago
I'm not a native english speaker, so I've used for an essay where they told us to target a certain word count. I was close, but the verbiage to get to that word count doesn't come naturally to me. So I used Germini and tell it to rewrite the text targeting that word count (my only prompt). Then I reviewed the answer, rewriting where it strayed from the points I was making.

Also I used it for a few programming tasks I was pretty sure was in the datasets (how to draw charts with python and manipulate pandas frame). I know the domain, but wasn't in the mood to analyse the docs to get the implementation information. But the information I was seeking was just a few lines of sample code. In my experience, anything longer is pretty inconsistent and worthless explanations.

petermcneeley•6h ago
I recently saw a new LLM that was fooled by "20 pounds of bricks vs 20 feathers". These are not reasoning machines.
dghlsakjg•5h ago
I recently had a computer tell me that 0.1 + 0.2 != 0.3. It must not be a math capable machine.

Perhaps it is more important to know the limitations of tools rather than dismiss their utility entirely due to the existence of limitations.

blamestross•5h ago
A computer isn't a math capable machine.

> Perhaps it is more important to know the limitations of tools rather than dismiss their utility entirely due to the existence of limitations.

Well, yes. And "reasoning" is only something LLMs do coincidentally, to their function as sequence continuation engines. Like performing accurate math on rationale numbers, it can happen if you put in a lot of work and accept a LOT of expensive computation. Even then there exists computations that just are not reasonable or feasible.

Reminding folks to dismiss the massive propaganda engine pushing this bubble isn't "dismissing their utility entirely".

These are not reasoning machines. Treating them like they are will get you hurt eventually.

dghlsakjg•3h ago
My point is that computers, when used properly, can absolutely do math. And LLMs, when used properly, can absolutely explain the reasoning behind why a pound of bricks and a pound of feathers weigh the same.

Can they reason? Maybe, depending on your definition of reasoning.

An example: which weighs more a pound of bricks and 453.59 grams of feathers? Explain your reasoning.

LLM: The pound of bricks weighs slightly more.

*Reasoning:*

* *1 pound* is officially defined as *0.45359237 kilograms*, which is *453.59237 grams*. * You have *453.59 grams* of feathers.

So, the pound of bricks (453.59237 grams) weighs a tiny fraction more than the 453.59 grams of feathers. For most practical purposes, they'd be considered the same, but technically, the bricks are heavier by 0.00237 grams. /llm

It is both correct and the reasoning is sound. Do I understand that the machine is a pattern following machine, yes! Is there an argument to be made that humans are also that? Probably. Chomsky himself argued in favor of a universal grammar, after all.

I’m steel manning this a bit, but the point is that LLMs are capable of doing some things which are indistinguishable from human reasoning in terms of results. Does the process matter in all cases?

HDThoreaun•5h ago
Tons of people fall for this too. Are they not reasoning? LLMs can also be bad reasoning machines.
petermcneeley•5h ago
I dont have much use for a bad reasoning machine.
the8472•5h ago
I could retort with another gotcha argument, but instead of doing that perhaps we can do better than that?

An attempt: They are bad reasoning machines that already are useful in a few domains and they're improving faster than evolutionary speeds. So even if they're not useful today in a domain relevant to you there's a significant possibility they might be in a few months. AlphaEvolve would have been scifi a decade ago.

"It's like if a squirrel started playing chess and instead of "holy shit this squirrel can play chess!" most people responded with "But his elo rating sucks""

HDThoreaun•5h ago
I can think of tons of uses for a bad reasoning machine as long as it’s cheap enough.
fzzzy•5h ago
20 feathers?
lostmsu•4h ago
Yes, Claude 4 Sonnet just said they both weight 20 pounds. UPD. and so did Gemini 2.5 Flash. And MS Copilot in "Think deeper" mode.
StrandedKitty•5h ago
Surely it just reasoned that you made a typo and "autocorrected" your riddle. Isn't this what a human would do? Though to be fair, a human would ask you again to make sure they heard you correctly. But it would be kind of annoying if you had to verify every typo when using an LLM.
downboots•5h ago
But are you aware of the weight comparison of a gallon of water vs a gallon of butane ?
petermcneeley•4h ago
No im not. A gallon is a measure of volume? This is a USA unit.
johnfn•6h ago
> It’s as if a biologist were to say: “I have a great new theory of organisms. It lists many that exist and many that can’t possibly exist, and I can tell you nothing about the distinction.”

> Again, we’d laugh. Or should.

Should we? This reminds me acutely of imaginary numbers. They are a great theory of numbers that can list many numbers that do 'exist' and many that can't possibly 'exist'. And we did laugh when imaginary numbers were first introduced - the name itself was intended as a derogatory term for the concept. But who's laughing now?

bubblyworld•5h ago
In the case of complex numbers mathematicians understand the distinction extremely well, so I'm not sure it's a perfect analogy.
kelsey978126•5h ago
This is the point where i realized he has no clue what he is saying. Theres so many creatures that once existed that can never again exist on earth due to the changes that the planet has gone through over millions, billions of years. The oxygen rich atmosphere that supported the dinosaurs for instance. If we had some kind of system that can put together proper working DNA for all the creatures that ever actually existed on this planet, some half of them would be completely nonviable if introduced to the ecosystem today. He is failing to see that there is an incredible understanding of systems that we are producing with this work, but he is a very old man from a very different time and contrarianism is often the only way to look smart or reasoned when you have no clue whats actually going on, so I am not shocked by his take.
chongli•5h ago
Imaginary numbers are not relevant at all. There’s nothing whatsoever to do with the everyday use of the word imaginary. They could just as easily have been called “vertical numbers” and real numbers called “horizontal numbers” in order to more clearly illustrate their geometric interpretation in the complex plane.

The term “imaginary number” was coined by Rene Descartes as a derogatory and the ill intent behind his term has stuck ever since. I suspect his purpose was theological rather than mathematical and we are all the worse for it.

netcan•5h ago
Insect behaviour. Flight of birds. Turtle navigation. A footballer crossing the field to intercept a football.

This is what Chomsky always wanted ai to be... especially language ai. Clever solutions to complex problems. Simple once you know how they work. Elegant.

I sympathize. I'm a curious human. We like elegant, simple revelations that reveal how out complex world is really simple once you know it's secrets. This aesthetic has also been productive.

And yet... maybe some things are complicated. Maybe LLMs do teach us something about language... that language is complicated.

So sure. You can certainly critique "ai blogosphere" for exuberance and big speculative claims. That part is true. Otoh... linguistics is one of the areas that ai based research may turn up some new insights.

Overall... what wins is what is most productive.

submeta•5h ago
Chomsky’s notion is: LLMs can only imitate, not understand language. But what exactly is understanding? What if our „understanding“ is just unlocking another level in a model? Unlocking a new form of generation?
smokel•5h ago
Understanding is probably not much more than making abstractions into simpler terms until you are left with something one can relate to by intuition or social consensus.
HPsquared•5h ago
Transforming, in other words.
egberts1•5h ago
Just because it can transform doesn't mean that the logic still remains correct.

I found this out when attempting to transform wiki pages into blog-specific-speak, repeatedly.

npteljes•3h ago
I have trouble with the notion "understanding". I get the usefulness of the word, but I don't think that we are capable to actually understand. I also think that we are not even able to test for understanding - a good imitation is as good as understanding. Also, understanding has limits. In school, they often say on class that you should forget whatever you have been taught so far, because this new layer of knowledge that they are about to teach you. Was the previous knowledge not "understanding" then? Is the new one "understanding"?

If we define "understanding" like "useful", as in, not an innate attribute, but something in relation to a goal, then again, a good imitation, or a rudimentary model can get very far. ChatGPT "understood" a lot of things I have thrown at it, be that algorithms, nutrition, basic calculations, transformation between text formats, where I'm stuck in my personal development journey, or how to politely address people in the email I'm about to write.

>What if our „understanding“ is just unlocking another level in a model?

I believe that it is - that understanding is basically an illusion. Impressions are made up from perceptions and thinking, and extrapolated over the unknown. And just look how far that got us!

roughly•2h ago
> But what exactly is understanding?

He alludes to quite a bit here - impossible languages, intrinsic rules that don’t actually express in the language, etc - that leads me to believe there’s a pretty specific sense by which he means “understanding,” and I’d expect there’s a decent literature in linguistics covering what he’s referring to. If it’s a topic of interest to you, chasing down some of those leads might be a good start.

(I’ll note as several others have here too that most of his language seems to be using specific linguistics terms of art - “language” for “human language” is a big tell, as is the focus on understanding the mechanisms of language and how humans understand and generate languages - I’m not sure the critique here is specifically around LLMs, but more around their ability to teach us things about how humans understand language.)

dinfinity•2h ago
> But what exactly is understanding?

I would say that it is to what extent your mental model of a certain system is able to make accurate predictions of that system's behavior.

thasso•5h ago
> The world’s preeminent linguist Noam Chomsky, and one of the most esteemed public intellectuals of all time, whose intellectual stature has been compared to that of Galileo, Newton, and Descartes, tackles these nagging questions in the interview that follows.

By whom?

bigyabai•5h ago
People who particularly agreed with Chomsky's inherently politicized beliefs, presumably.
mattw1•5h ago
In all seriousness tho, not much of anything he says is taken seriously in an academic sense any more. Univeral Grammar, Minimalism, etc. He's a very petty dude. The reason he doesn't engage with GPT is because it suggests that linguistic learning is unlike a theory he spent his whole life [unsuccessfully] promoting, but he's such a haughty know-it-all, that I guess dummies take that for intelligence? It strikes me as not dissimilar to Trump in a way, where arrogance is conflated with strength, intelligence, etc. Fake it til you make it, or like, forever, I guess.
apelapan•3h ago
The comparison to Trump seems very unfair. I'm not in the academy and didn't know the current standing of his work, but he was certainly a big name that popped up everywhere (as a theorists in the field, not as a general celebrity) when I took an introduction to linguistics 20+ years ago.

As this is Hacker News, it is worth mentioning that he developed the concept of context-free grammars. That is something many of us encounter on a regular basis.

No matter what personality flaws he might have and how misguided some of his political ideas might be, he is one of the big thinkers of the 20th century. Very much unlike Trump.

thomasahle•5h ago
Google Scholar
TiredOfLife•5h ago
People who work for russia
hatsunearu•5h ago
That is unbelievable that someone could glaze someone this hard
paulsutter•5h ago
"Expert in (now-)ancient arts draws strange conclusion using questionable logic" is the most generous description I can muster.

Quoting Chomsky:

> These considerations bring up a minor problem with the current LLM enthusiasm: its total absurdity, as in the hypothetical cases where we recognize it at once. But there are much more serious problems than absurdity.

> One is that the LLM systems are designed in such a way that they cannot tell us anything about language, learning, or other aspects of cognition, a matter of principle, irremediable... The reason is elementary: The systems work just as well with impossible languages that infants cannot acquire as with those they acquire quickly and virtually reflexively.

Response from o3:

LLMs do surface real linguistic structure:

• Hidden syntax: Attention heads in GPT-style models line up with dependency trees and phrase boundaries—even though no parser labels were ever provided. Researchers have used these heads to recover grammars for dozens of languages.

• Typology signals: In multilingual models, languages that share word-order or morphology cluster together in embedding space, letting linguists spot family relationships and outliers automatically.

• Limits shown by contrast tests: When you feed them “impossible” languages (e.g., mirror-order or random-agreement versions of English), perplexity explodes and structure heads disappear—evidence that the models do encode natural-language constraints.

• Psycholinguistic fit: The probability spikes LLMs assign to next-words predict human reading-time slow-downs (garden-paths, agreement attraction, etc.) almost as well as classic hand-built models.

These empirical hooks are already informing syntax, acquisition, and typology research—hardly “nothing to say about language.”

foobarqux•5h ago
> LLMs do surface real linguistic structure...

It's completely irrelevant because the point he's making is that LLMs operate differently from human languages as evidenced by the fact that they can learn language structures that humans cannot learn. Put another way, I'm sure you can point out an infinitude of similarities between human language faculty and LLMs but it's the critical differences that make LLMs not useful models of human language ability.

> When you feed them “impossible” languages (e.g., mirror-order or random-agreement versions of English), perplexity explodes and structure heads disappear—evidence that the models do encode natural-language constraints.

This is confused. You can pre-train an LLM on English or an impossible language and they do equally well. On the other hand humans can't do that, ergo LLMs aren't useful models of human language because they lack this critical distinctive feature.

rahimnathwani•4h ago
Is that true? This paper claims it is not.

https://arxiv.org/abs/2401.06416

foobarqux•4h ago
Yes it's true, you can read my response to one of the authors @canjobear describing the problem with that paper in the comment linked below. But to summarize: in order to show what they want to show they have to take the simple, interesting languages based on linear order that Moro showed a human cannot learn and show that LLMs also can't learn them and they don't do that.

The reason the Moro languages are of interest are that they are computationally simple so it's a puzzle why humans can't learn them (and no surprise that LLMs can). The authors of the paper miss the point and show irrelevant things like there exist complicated languages that both humans and LLMs can't learn.

https://news.ycombinator.com/item?id=42290482

paulsutter•4h ago
> You can pre-train an LLM on English or an impossible language and they do equally well

It's impressive that LLMs can learn languages that humans cannot. In what frame is this a negative?

Separately, "impossible language" is a pretty clear misnomer. If an LLM can learn it, it's possible.

foobarqux•3h ago
The latter. Moro showed that you can construct simple language rules, in particular linear rules, like the third word of every sentence modifies the noun, that humans have a hard time learning (specifically they use different parts of their brain in MRI scans and take longer to process than control languages) and are different from conventional human language structure (which hierarchical structure dependent, i.e. roughly that words are interpreted according to their position in a parse tree not their linear order).

That's what "impossible language" means in this context, not something like computationally impossible or random.

paulsutter•3h ago
Ok then .. what makes that a negative? You're describing a human limitation and a strength of LLMs
foobarqux•3h ago
It's not a negative, it's just not what humans do, which is Chomsky's (a person studying what humans do) point.

As I said in another comment this whole dispute would be put to bed if people understood that they don't care about what humans do (and that Chomsky does).

paulsutter•3h ago
Suggestion for you then, in your first response you would have been clearer to say "The reason Chomsky seems like such a retard here, is because he clings to irrelevant nonsense"

It's completely unremarkable that humans are unable to learn certain languages, and soon it will be unremarkable when humans have no cognitive edge over machines.

Response: Science? "Ancient Linguistics" would more accurately describe Chomsky's field of study and its utility

foobarqux•3h ago
> Suggestion for you then, in your first response you would have been clearer to say "The reason Chomsky seems like such a retard here, is because he clings to irrelevant nonsense"

If science is irrelevant to you it's you who should have recognized this before spouting off.

retskrad•5h ago
It’s time to stop writing in this elitist jargon. If you’re communicating and few people understands you, then you’re a bad communicator. I read the whole thing and thought: wait, was there a new thought or interesting observation here? What did we actually learn?
mmooss•4h ago
Where do you see 'elitist jargon'? That didn't even cross my mind.
titzer•5h ago
All this interview proves is that Chomsky has fallen far, far behind how AI systems work today and is retreating to scoff at all the progress machine learning has achieved. Machine learning has given rise to AI now. It can't explain itself from principles or its architecture. But you couldn't explain your brain from principles or its architecture, you'd need all of neuroscience to do it. Because the brain is digital and (probably) does not reason like our brains do, it somehow falls short?

While there's some things in this I find myself nodding along to in this, I can't help but feel it's an a really old take that is super vague and hand-wavy. The truth is that all of the progress on machine learning is absolutely science. We understand extremely well how to make neural networks learn efficiently; it's why the data leads anywhere at all. Backpropagation and gradient descent are extraordinarily powerful. Not to mention all the "just engineering" of making chips crunch incredible amounts of numbers.

Chomsky is extremely ungenerous to the progress and also pretty flippant about what this stuff can do.

I think we should probably stop listening to Chomsky; he hasn't said anything here that he hasn't already say a thousand times for decades.

l5870uoo9y•5h ago
Perhaps it should be mentioned that he is 96 years old.
bicx•5h ago
Wow, he is, isn’t he. I hope I’m this coherent when I’m 96.
lxgr•5h ago
> [...] I can't help but feel it's an a really old take [...]

To be fair the article is from two years ago, which when talking about LLMs in this age arguably does count as "old", maybe even "really old".

lostmsu•4h ago
I think GPT-2 (2019) was already strong enough argument for possibility of modeling knowledge and language that Chomsky rejected.
gf000•4h ago
Though given that LLMs fundamentally can't know whether they know something or not (without a later pass of fine-tuning on what they should know) is a pretty good argument against them being good knowledge bases.
tgv•5h ago
> But you couldn't explain your brain from principles or its architecture, you'd need all of neuroscience to do it

That's not a good argument. Neuroscience was constructed by (other) brains. The brain is trying to explain itself.

> The truth is that all of the progress on machine learning is absolutely science.

But not much if you're interested in finding out how our brain works, or how language works. One of the interesting outcomes of LLMs is that there apparently is a way to represent complex ideas and their linguistic connection in a (rather large) unstructured state, but it comes without thorough explanation or relation to the human brain.

> Chomsky is [...] pretty flippant about what this stuff can do.

True, that's his style, being belligerently verbose, but others have been pretty much fawning and drooling over a stochastic parrot with a very good memory, mostly with dollar signs in their eyes.

sswatson•4h ago
> but others have been pretty much fawning…

This is not relevant. An observer who deceives for purposes of “balancing” other perceived deceptions is as untrustworthy and objectionable as one who deceives for other reasons.

dereify•5h ago
> I think we should probably stop listening to Chomsky

He's not able to communicate anymore, so we get that one for free.

cj•5h ago
> Not to mention all the "just engineering" of making chips crunch incredible amounts of numbers.

Are LLM's still the same black box as they were described as a couple years ago? Are their inner workings at least slightly better understood than in the past?

Running tens of thousands of chips crunching a bajillion numbers a second sounds fun, but that's not automatically "engineering". You can have the same chips crunching numbers with the same intensity just to run an algorithm to run a large prime number. Chips crunching numbers isn't automatically engineering IMO. More like a side effect of engineering? Or a tool you use to run the thing you built?

What happens when we build something that works, but we don't actually know how? We learn about it through trial and error, rather than foundational logic about the technology.

Sorta reminds me of the human brain, psychology, and how some people think psychology isn't science. The brain is a black box kind of like a LLM? Some people will think it's still science, others will have less respect.

This perspective might be off base. It's under the assumption that we all agree LLM's are a poorly understood black box and no one really knows how they truly work. I could be completely wrong on that, would love for someone else to weigh in.

Separately, I don't know the author, but agreed it reads more like a pop sci book. Although I only hope to write as coherently as that when I'm 96 y/o.

ogogmad•5h ago
> Running tens of thousands of chips crunching a bajillion numbers a second sounds fun, but that's not automatically "engineering".

Not if some properties are unexpectedly emergent. Then it is science. For instance, why should a generic statistical model be able to learn how to fill in blanks in text using a finite number of samples? And why should a generic blank-filler be able to produce a coherent chat bot that can even help you write code?

Some have even claimed that statistical modelling shouldn't able to produce coherent speech, because it would need impossible amounts of data, or the optimisation problem might be too hard, or because of Goedel's incompleteness theorem somehow implying that human-level intelligence is uncomputable, etc. The fact that we have a talking robot means that those people were wrong. That should count as a scientific breakthrough.

Shorel•4h ago
> because it would need impossible amounts of data

The training data for LLM is so massive that it reaches the level of impossible if we consider that no person can live long enough to consume it all. Or even a small percent of it.

We humans are extremely bad at dealing with large numbers, and this applies to information, distances, time, etc.

ogogmad•4h ago
Your final remark sounds condescending. Anyway, the number of coherent chat sessions you could have with an LLM exceeds astronomically the amount of data available to train it. How is that even possible?
dwaltrip•3h ago
The current AI training method doesn't count because a human couldn't do it? What?
jbentley1•5h ago
"I think we should probably stop listening to Chomsky"

I've been saying this my whole life, glad it's finally catching on

to-too-two•4h ago
Why? He's made significant contributions to political discourse and science.
foobarqux•4h ago
> The truth is that all of the progress on machine learning is absolutely science

It is not science, which is the study of the natural world. You are using the word "science" as an honorific, meaning something like "useful technical work that I think is impressive".

The reason you are so confused is that you can't distinguish studying the natural world from engineering.

rxtexit•2h ago
It really shouldn't be hard to understand that a titan of a field has forgot more than what an arm chair enthusiast knows.

I remember having thoughts like this until I listened to him talk on a podcast for 3 hours about chatGPT.

What was most obvious is Chomsky really knows linguistics and I don't.

"What Kind of Creatures Are We?" is good place to start.

We should take having Chomsky still around to comment on LLMs as one of the greatest intellectual gifts.

Much before listening to his thoughts on LLMs was me projecting my disdain for his politics.

prpl•5h ago
Reminds me of SUSY, string theory, the standard model, and beyond that, string theory etc…

What is elegant as a model is not always what works, and working towards a clean model to explain everything from a model that works is fraught, hard work.

I don’t think anyone alive will realize true “AGI”, but it won’t matter. You don’t need it, the same way particle physics doesn’t need elegance

newAccount2025•5h ago
I tried to read it, but gave up when he started dismissing perspectives with vague abstractions. What is an “impossible language?” What does it have to do with anything? Maybe there are interesting ideas here, but they are beyond me.
Smaug123•5h ago
From some Googling and use of Claude (and from summaries of the suggestively titled "Impossible Languages" by Moro linked from https://en.wikipedia.org/wiki/Universal_grammar ), it looks like he's referring to languages which violate the laws which constrain the languages humans are innately capable of learning. But it's very unclear why "machine M is capable of learning more complex languages than humans" implies anything about the linguistic competence or the intelligence of machine M.
foobarqux•5h ago
It doesn't, it just says that LLMs are not useful models of the human language faculty.
specialist•4h ago
This is where I'm stuck.

For other commentators, as I understand it, Chomsky's talking about well-defined grammar and language and production systems. Think Hofstadter's Godel Escher Bach. Not "folk" understanding of language.

I have no understanding or intuition, or even a finger nail grasp, for how an LLM generates, seemingly emulating, "sentences", as though created with a generative grammar.

Is any one comparing and contrasting these two different techniques? Being noob, I wouldn't even know where to start looking.

I've gleaned that someone(s) are using LLM/GPT to emit abstract syntax trees (vs a mere stream of tokens), to serve as input for formal grammars (eg programming source code). That sounds awesome. And something I might some day sorta understand.

I've also gleaned that, given sufficient computing power, training data for future LLMs will have tokenized words (vs just character sequences). Which would bring the two strategies closer...? I have no idea.

(Am noob, so forgive my poor use of terminology. And poor understanding of the tech, too.)

foobarqux•4h ago
I don't really understand your question but if a deep neural network predicts the weather we don't have any problem accepting that the deep neural network is not an explanatory model of the weather (the weather is not a neural net). The same is true of predicting language tokens.
fc417fc802•31m ago
> is not an explanatory model of the weather (the weather is not a neural net)

I don't follow. Aren't those entirely separate things? The most accurate models of anything necessarily account for the underlying mechanisms. Perhaps I don't understand what you mean by "explanatory"?

Specifically in the case of deep neural networks, we would generally suppose that it had learned to model the underlying reality. In effect it is learning the rules of a sufficiently accurate simulation.

foobarqux•16m ago
> The most accurate models of anything necessarily account for the underlying mechanisms

But they don't necessarily convey understanding to humans. Prediction is not explanation.

There is a difference between Einstein's General Theory of Relativity and a deep neural network that predicts gravity. The latter is virtually useless for understanding gravity (that's even if makes better predictions).

> Specifically in the case of deep neural networks, we would generally suppose that it had learned to model the underlying reality. In effect it is learning the rules of a sufficiently accurate simulation.

No, they just fit surface statistics, not underlying reality. Many physics phenomena were predicted using theories before they were observed, they would not be in the training data even though they were part of the underlying reality.

cmiles74•2h ago
Firstly, can't speak for Chomsky.

In this article he is very focused on science and works hard to delineate science (research? deriving new facts?) from engineering (clearly product oriented). In his opinion ChatGPT falls on the engineering side of this line: it's a product of engineering, OpenAI is concentrating on marketing. For sure there was much science involved but the thing we have access to is a product.

IMHO Chomsky is asking: while ChatGPT is a fascinating product, what is it teaching us about language? How is it advancing our knowledge of language? I think Chomsky is saying "not much."

Someone else mentioned embeddings and the relationship between words that they reveal. Indeed, this could be a worthy area of further research. You'd think it would be a real boon when comparing languages. Unfortunately the interviewer didn't ask Chomsky about this.

Orangeair•5h ago
[2023]
lanfeust6•5h ago
I'm noticing that leftists overwhelmingly toe the same line on AI skepticism, which suggests to me an ideological motivation.
Supermancho•5h ago
> AI skepticism

Isn't AI optimism an ideological motivation? It's a spectrum, not a mental model.

lanfeust6•4h ago
Whether one expects AI to be powerful or weak should have nothing to do with political slant, but here it seems to inform the opinion. It begs the question: what do they want to be true? The enemy is both too strong and too weak.

They're firmly on one extreme end of the spectrum. I feel as though I'm somewhere in between.

mattw1•5h ago
I have experienced this too. It's definitely part of the religion but I'm not sure why tbh. Maybe they equate it with like tech is bad mkay, which, looking at who leads a lot of the tech companies, is somewhat understandable, altho very myopic.
santoshalper•5h ago
I see this as much more of a hackers vs. corporations ideological split. Which imperfectly maps to leftism vs conservatism.

The perception on the left is that once again, corporations are foisting products on us that nobody wants, with no concern for safety, privacy, or respect for creators.

For better or worse, the age of garage-tech is mostly dead and Tech has become synonymous with corporatism. This is especially true with GenAI, where the resources to construct a frontier model (or anything remotely close to it) are far outside what a hacker can afford.

mattw1•5h ago
That makes sense, and there's definitely an element of truth to that position. The trouble is, the response is to dissociate with the technology, which is really not a tenable position if you intend to have a meaningful part in like... anything in the future. What I see-- and this is just my personal experience-- is that leftists tend to want to pretend it isn't happening, or that it won't matter. When it fact nothing matters more.

The deepest of deep ironies: I talk to people all the time talking about ushering in an age of post-capitalism and ignoring AI. When I personally can't see how the AI of the next decade and capitalism can coexist, the latter being based on human labor and all. Like, AI is going to be the reason what you want is going to happen, so why ignore it?

lanfeust6•4h ago
> I see this as much more of a hackers vs. corporations ideological split.

That framing may be true within tech circles, not the broader political divide. "Hackers" aren't collectively discounting and ignoring AI tools regardless of their enthusiasm for open-source.

Safety-ism is also most popular among those see useful potential in AI, and a generous enough timeline for AGI.

santoshalper•5h ago
Or an ideological alignment of values. Generative AI is strongly associated with large corporations that are untrusted (to put it generously) by those on the left.

An equivalent observation might be that the only people who seem really, really excited about current AI products are grifters who want to make money selling it. Which looks a lot like Blockchain to many.

rxtexit•2h ago
Then you obviously didn't listen to a word Chomsky has said on the subject.

I was quite dismissive of him on LLMs until I realized the utter hubris and stupidity of dismissing Chomsky on language.

I think it was someone asking if he was familiar with the Wittgenstein Blue and Brown books and of course because he as already an assistant professor at MIT when they came out.

I still chuckle at my own intellectual arrogance and stupidity when thinking about how I was dismissive of Chomsky on language. I barely know anything and I was being dismissive of one of unquestionable titans and historic figures of a field.

EasyMark•2h ago
I think viewing the world as either leftist or right wing is rather limiting philosophy and way to go through life. Most people are a lot more complicated than that.
0xDEAFBEAD•5h ago
I confess my opinion of Noam Chomsky dropped a lot from reading this interview. The way he set up a "Tom Jones" strawman and kept dismissing positions using language like "we'd laugh", "total absurdity", etc. was really disappointing. I always assumed that academics were only like that on reddit, and in real life they actually made a serious effort at rigorous argument, avoiding logical fallacies and the like. Yet here is Chomsky addressing a lay audience that has no linguistics background, and instead of even attempting to summarize the arguments for his position, he simply asserts that opposing views are risible with little supporting argument. I expected much more from a big-name scholar.

"The first principle is that you must not fool yourself, and you are the easiest person to fool."

foobarqux•5h ago
"Tom Jones" isn't a strawman, Chomsky is addressing an actual argument in a published paper from Steven Piantadosi. He's using a pseudonym to be polite and not call him out by name.

> instead of even attempting to summarize the arguments for his position..

He makes a very clear, simple argument, accessible to any layperson who can read. If you are studying insects what you are interested in is how insects do it not what other mechanisms you can come up with to "beat" insects. This isn't complicated.

lostmsu•4h ago
That's understandable but irrelevant. Only a few people have major interest in how humans think exactly. But nearly everyone is hang on the question if the LLMs could think better.
foobarqux•4h ago
It's not irrelevant, it's the core of the disconnect: The problem is that everyone is arguing as if they passionately care about how humans work when, as you say, they don't care at all.

People should just recognize, as you have done, that they don't actually care about how the human language faculty works. It's baffling that they instead choose to make absurd arguments to defend fields they don't care one way or another about.

When Chomsky says that LLMs aren't how the human faculty works it would be so easy to tell the truth and say "I don't care how the human language faculty works" and everyone can go focus on the things they are interested in, just as it would be easy for a GPS designer to say "I don't care how insect navigation works".

There is no problem as long as you don't pretend to be caring about (this aspect of) science.

LudwigNagasena•4h ago
Is it polite to deprive readers of context necessary to understand what the speaker is talking about? I was also very confused by that part and I had no idea whom or what he was talking about or why he even started taking about that.

I searched for an actual paper by that guy because you’ve mentioned his real name. I found “Modern language models refute Chomsky’s approach to language”. After reading it seems even more true that Chomsky’s Tom Jones is a strawman.

foobarqux•3h ago
> After reading it seems even more true that Chomsky's Tom Jones is a strawman.

Lol. It's clear you are not interested in having any kind of rational discussion on the topic and are driven by some kind of zealotry when you claim to have read a technical 40 page paper (with an additional 18 pages of citations) in 30 minutes.

Even if by some miraculous feat you had read it you haven't made a single actual argument or addressed any of the points made by Chomsky.

LudwigNagasena•2h ago
It’s certainly not a dense paper with careful nuanced derivations that you have to ponder to grasp. It’s a light read you can skim especially if you aren’t interested in LLM Trump improv and you are familiar with the general thought behind connectionism, construction grammar, other modern linguistic theories and, of course, universal grammar. The debate is as old as UG, but now with a new LLM flavor.

I don’t know which argument you expect from me. I read it and found nothing similar to “Stop wasting your time; naval vessels do it all the time.” So I concluded it’s a strawman. Being against a particular controversial approach in linguistics doesn’t mean being against science.

foobarqux•1h ago
> I read it and found nothing similar to “Stop wasting your time; naval vessels do it all the time.”

You implied in the previous paragraph that you didn't in fact read it and you only "skimmed" it. Maybe that's why you "found nothing similar to 'stop wasting your time; naval vessels do it all the time". But even in skimming the paper it's incomprehensible how you could miss it: At least the first 23 pages of the draft version I have just describe how well LLMs perform and completely ignores the relevant question of how human language works. (It doesn't get any better after the first 23 pages). So presumably you just don't know what an analogy is and are literally searching for the term "naval vessels".

Here's just one example demonstrating that Piantodosi does in fact claim what Chomsky says he does: Piantodosi writes "The success of large language models is a failure for generative theories because it goes against virtually all of the principles these theories have espoused." Rewriting that statement using Chomsky's analogy illustrates how idiotic the original statement is: "The success of naval vessels is a failure for insect navigation theories because it goes against all of the principles these theories have espoused".

LudwigNagasena•22m ago
There is a difference between supporting one research paradigm over another and rejecting science altogether to focus on engineering. The first quote and the context around it implies the latter.

The success of naval vessels show it’s possible to navigate without innate star and wind comprehension, so maybe we should think of that inner stuff as phlogiston. (Yeah, this analogy isn’t as nice but it’s quite hard to translate the nuance of linguistic debate into nautical terms.)

0xDEAFBEAD•1h ago
>The systems work just as well with impossible languages that infants cannot acquire as with those they acquire quickly and virtually reflexively.

Where is the research on impossible language that infants can't acquire? A good popsci article would give me leads here.

Even assuming Chomsky's claim is true, all it shows is that LLMs aren't an exact match for human language learning. But even an inexact model can still be a useful research tool.

>That’s highly unlikely for reasons long understood, but it’s not relevant to our concerns here, so we can put it aside. Plainly there is a biological endowment for the human faculty of language. The merest truism.

Again, a good popsci article would actually support these claims instead of simply asserting them and implying that anyone who disagrees is a simpleton.

I agree with Chomsky that the postmodern critique of science sucks, and I agree that AI is a threat to the human race.

foobarqux•35m ago
> Where is the research on impossible language that infants can't acquire? A good popsci article would give me leads here.

It's not infants, it's adults but Moro "Secrets of Words" is a book that describes the experiments and is aimed at lay people.

> Even assuming Chomsky's claim is true, all it shows is that LLMs aren't an exact match for human language learning. But even an inexact model can still be a useful research tool.

If it is it needs to be shown, not assumed. Just as you wouldn't by default assume that GPS navigation tells you about insect navigation (though it might somehow).

> Again, a good popsci article would actually support these claims instead of simply asserting them and implying that anyone who disagrees is a simpleton.

He justifies the statement in the previous sentence (which you don't quote) where he says that it is self-evident by virtue of the fact that something exists at the beginning (i.e. it's not empty space). That's the "merest truism". No popsci article is going to help understand that if you don't already.

asmeurer•5h ago
It's amusing that he argues (correctly) that "there is no Great Chain of Being with humans at the top," but then claims that LLMs cannot tell us anything about language because they can learn "impossible languages" that infants cannot learn. Isn't that an anthropomorphic argument, saying that what a language is inherently defined by human cognition?
tgv•5h ago
When Chomsky says "language," he means "natural/human language," not e.g. /[ab]*/ or prime numbers.
foobarqux•5h ago
Yes, studying human language is actually inherently defined by what humans do, just -- as he points out, if you could understand the article -- studying insect navigation is defined by what insects do and not what navigation systems human could design.
LudwigNagasena•5h ago
That was a weird ride. He was asked whether AI will outsmart humans and went on a rant about philosophy of science seemingly trying to defend the importance of his research and culminated with some culture war commentary about postmodernism.
lucisferre•5h ago
"The desert ants in my backyard have minuscule brains, but far exceed human navigational capacities, in principle, not just performance. There is no Great Chain of Being with humans at the top."

This quote brought to mind the very different technological development path of the spider species in Adrian Tchaikovsky's Children of Time. They used pheromones to 'program' a race of ants to do computation.

lostmsu•4h ago
I don't know what he's talking about. Humans clearly outperform ants in navigation. Especially if you allow arbitrary markings on the terrain.

Sounds like "ineffable nature" mumbo-jumbo.

tpm•2h ago
Arbitrary markings on the terrain? Why not GPS, satellite photo etc? All of those are human inventions and we can navigate much better and in a broader set of environments than ants thanks to them.
A4ET8a8uTh0_v2•5h ago
It is unfortunate opinion, because I personally hold Chomsky in fairly high regard and give most of his thoughts I am familiar with a reasonable amount of consideration if only because he could, I suppose in the olden days now, articulate his points well and make you question your own thought process. This no longer seems to be the case though as I found the linked article somewhat difficult to follow. I suppose age can get to anyone.

Not that I am an LLM zealot. Frankly, some of the clear trajectory it puts humans on makes me question our futures in this timeline. But even if I am not a zealot, but merely an amused, but bored middle class rube, the serious issues with it ( privacy, detailed personal profiling that surpasses existing systems, energy use, and actual power of those who wield it ), I can see it being implemented everywhere with a mix of glee and annoyance.

I know for a fact it will break things and break things hard and it will be people, who know how things actually work that will need to fix those.

I will be very honest though. I think Chomsky is stuck in his internal model of the world and unable to shake it off. Even his arguments fall flat, because they don't fit the domain well. It seems like they should given that he practically made his name on syntax theory ( which suggests his thoughts should translate well into it ) and yet.. they don't.

I have a minor pet theory on this, but I am still working on putting it into some coherent words.

calibas•4h ago
The fact that we have figured out how to translate language into something a computer can "understand" should thrill linguists. Taking a word (token) and abstracting it's "meaning" as a 1,000-dimension vector seems like something that should revolutionize the field of linguistics. A whole new tool for analyzing and understanding the underlying patterns of all language!

And there's a fact here that's very hard to dispute, this method works. I can give a computer instructions and it "understands" them in a way that wasn't possible before LLMs. The main debate now is over the semantics of words like "understanding" and whether or not an LLM is conscious in the same way as a human being (it isn't).

krackers•1h ago
Restricted to linguistics, LLM's supposed lack of understanding should be a non-sequitur. If the question is whether LLMs have formed a coherent ability to parse human languages, the answer is obviously yes. In fact not just human languages, as seen with multimodality the same transformer architecture seems to work well to model and generate anything with inherent structure.

I'm surprised that he doesn't mention "universal grammar" once in that essay. Maybe it so happens that humans do have some innate "universal grammar" wired in by instinct but it's clearly not _necessary_ to be able to parse things. You don't need to set up some explicit language rules or generative structure, enough data and the model learns to produce it. I wonder if anyone has gone back and tried to see if you can extract out some explicit generative rules from the learned representation though.

Since the "universal grammar" hypothesis isn't really falsifiable, at best you can hope for some generalized equivalent that's isomorphic to the platonic representation hypothesis and claim that all human language is aligned in some given latent representation, and that our brains have been optimized to be able to work in this subspace. That's at least a testable assumption, by trying to reverse engineer the geometry of the space LLMs have learned.

oysterville•4h ago
Two year old interview should be labeled as such.