Yeah, not all humans do it. It's too energy expensive, biological efficiency wins.
As of ML.. Maybe next time, when someone figures out how to combine deductive with inductive, in zillion small steps, with falsifying built-in.. (instead of confronting them 100% one against 100% the other)
The problem with the AI discourse is that the language games are all mixed up and confused. We're not just talking about capability, we're talking about significance too.
The author defines American style intelligence as "the ability to adapt to new situations, and learn from experience".
Then argues that the current type of machine-learning driven AI is American style-intelligent because it is inductive, which is not what was supposedly (?) being argued for.
Of course current AI/ML models cannot adapt to new situations and learn from experience, outside the scope of its context window, without a retraining or fine-tuning step.
Intelligence, in the real world, is the ability to reason about logic. If 1 + 1 is 2, and 1 + 2 is 3, then 1 + 3 must be 4. This is deterministic, and it is why LLMs are not intelligent and can never be intelligent no matter how much better they get at superficially copying the form of output of intelligence. Probabilistic prediction is inherently incompatible with deterministic deduction. We're years into being told AGI is here (for whatever squirmy value of AGI the hype huckster wants to shill), and yet LLMs, as expected, still cannot do basic arithmetic that a child could do without being special-cased to invoke a tool call. How is it that we can go about ignoring reality for so long?
The calculations are internal but they happen due to the orchestration of specific parts of the brain. That is to ask, why can't we consider our brains to be using their own internal tools?
I certainly don't think about multiplying two-digit numbers in my head in the same manner as when playing a Dm to a G7 chord that begs to resolve to a C!
It's not much different than ChatGPT being trained write a to Python script.
A notable difference is that it's much more efficient to teach something new to a 5-year old than fine-tune or retrain an LLM.
It appears LLMs got RHLFed into generating suitable Python scripts after the issue was exposed, which is an efficient way of getting better answers, but feels rather like handing the child really struggling with their arithmetic a calculator...
The key thing is modeling. You must model a situation in a useful way in order to apply logic to it. And then there is intention, which guides the process.
> With recent advances in AI, it becomes ever harder for proponents of intelligence-as-understanding to continue asserting that those tools have no clue and “just” perform statistical next-token prediction.
??????? No, that is still exactly what they do. The article then lists a bunch of examples in which this in trivially exactly what is happening.
> “The cat chased the . . .” (multiple connections are plausible, so how is that not understanding probability?)
It doesn't need to "understand" probability. "The cat chased the mouse" shows up in the distribution 10 times. "The cat chased the bird" shows up in the distribution 5 times. Absent any other context, with the simplest possible model, it now has a probability of 2/3 for the mouse and 1/3 for the bird. You can make the probability calculations as complex as you want, but how could you possibly trout this out as an example that an LLM completing this sentence isn't a matter of trivial statistical prediction? Academia needs an asteroid, holy hell.
[I originally edited this into my post, but two people had replied by then, so I've split it off into its own comment.]
Don't get me wrong, I agree that general LLMs are still pretty bad at basic mathematical reasoning. I just tested Claude with a basic question about prime factoring [0] and there seems to be little improvement in this sort of question in the last 3 years. But there's a chance that they will become good enough that any errors they make are practically imperceptible to humans, at which point we have to question whether what's happening in our own heads is substantially different.
It's worth noting that at this point you appear to be siding with the "inductive" meaning of intelligence from the article, i.e. intelligence is about (or is demonstrated by) achieving certain behaviours.
Also, there are fairly basic calculations where many humans will err indefinitely unless explicitly shown otherwise, e.g. probabilistic reasoning such as the Monty Hall problem.
[0] https://claude.ai/share/e22e43c3-7751-405d-ba00-319f1a85c9ad
Prove that humans do it.
Even in your own mind, "5 + 25" and "7 + 27" take two different routes based on different pattern-matched heuristics. You don't have a single "universal addition algorithm". You have a bag of pattern-matching features, and maybe a slow unreliable fallback path, if you have learned any. Which is why "5000+5000" is trivial but "4523+5438" isn't.
The problem with your "theory" is a common one: it's utter bullshit. It says "here's what I think is true" but stops short of "why", let alone "why it fits reality better than the alternatives" or "here's what this theory allows you to do". It's like claiming that "all matter is phlogiston". It was entertaining when people were doing that in Ancient Greece, but by now, it's just worthless noise.
> but stops short of "why",
No, I rather explicitly stated how I think reasoning regulation happens as a result of the body's attempt to conserve energy in environments where complex reasoning does not meaningfully improve chances of survival, and allocates more energy in environments where complex reasoning does meaningfully improve chances of survival. That is the "why".
> "here's what this theory allows you to do"
What it allows us to do is achieve deterministic results after observing new conditions we've never seen before and reasoning about how those conditions interact to produce a final result. Our ability to build computers and spaceships is the direct result of reasoning applied acrossing many, many steps. LLMs cannot get even the first step correct, of correctly deducing Z from X + Y, and without that there is a fundamental difference in the capability to solve problems that can never be resolved. They can be special-cased to use a calculator to solve the first step, but this still leaves them incapable of solving any new logical exercise they are not pre-emptively programmed to handle, which makes them no different than any other software.
If human mind was based on logic, 20+80 and 54+39 would be equally as fast. As they are for a calculator. The calculator has a simple, fast, broadly applicable addition primitive based on pure binary logic. Humans don't.
In case of addition, LLMs work the same way humans do. They have "happy paths" of pattern-matching, and a slow unreliable "fallback path". They just have better "happy path" coverage than almost all humans do.
And your attempts at making your "theory" look any better than phlogiston are rather pathetic. I'm not asking you "what logic can do". I'm asking you "what your theory can do". Nothing of value.
"Intelligence in AI" lacks any existential dynamic, our LLMs are literally linguistic mirrors of human literature and activity tracks. They are not intelligent, but for the most part we can imagine they are, while maintaining sharp critical analysis because they are idiot savants in the truest sense.
For example, we all have an internal physics model in our heads that's build up through our continuous interaction with our environment. That acts as our shared context. That's why if I tell you to bring me a cup of tea, I have a reasonable expectation that you understand what I requested and can execute this action intelligently. You have a conception of a table, of a cup, of tea, and critically our conception is similar enough that we can both be reasonably sure we understand each other.
Incidentally, when humans end up talking about abstract topics, they often run into exact same problem as LLMs, where the context is missing and we can be talking past each other.
The key problem with LLMs is that they currently lack this reinforcement loop. The system merely strings tokens together in a statistically likely fashion, but it doesn't really have a model of the domain it's working in to anchor them to.
In my opinion, stuff like agentic coding or embodiment with robotics moves us towards genuine intelligence. Here we have AI systems that have to interact with the world, and they get feedback on when they do things wrong, so they can adjust their behavior based on that.
Concrete statistician: I can learn the problem in its full complexity, unlike the dumdum below me.
Abstract thinker: I understand it, because I can reduce its dimensity to a small number of parameters.
CS: I can predict this because I have statistics about its past behavior.
AT: I told you so.
CS: You couldn't possibly know this, because it has never happened before. You suffer from the hindsight bias.
AT: But I told you.
CS: It has never happened, you couldn't possibly have statistics of when such things occur. You were just lucky.
CS: I'm smart, I can be taught anything
AT: You are stupid because you need to be taught everything.
War (or another sort of mass death or other kind of suffering) emerges.
The revolutionary war: CS America vs AT british empire.
The french revolution: CS revolutionaries vs AT aristocracy.
American Civil war: CS North, vs AT South
WWII: AT Nazis, vs CS jews.
Probably many more.
But the essay is a huge letdown. The European vs. American framing obscures more than it illuminates. The two concepts of intelligence are not really analyzed at all — one could come up with interpretations under which they’re perfectly compatible with each other. The dismissal of Marx and Freud, two of the deepest thinkers in history, is embarrassing, saying a lot more about the author than about those thinkers.
(For anyone who hasn't read much Freud, here's a very short essay that may surprise you with its rigor: https://www.marxists.org/reference/subject/philosophy/works/...)
The whole purpose is not to inform or provoke thought, but for the whole thing to exude prestige and exclusivity, like an article you'd find in a magazine in a high-end private clinics waiting room.
“LLMs only predict what a human would say, rather than predicting the actual consequences of an action or engaging with the real world. This is the core deficiency: intelligence requires not just mimicking patterns, but acting, observing real outcomes, and adjusting behavior based on those outcomes — a cycle Sutton sees as central to reinforcement learning.” [1]
An LLM itself is a form of crystallized intelligence, but it does not learn and adapt without a human driver, and that to me is a key component of intelligent behavior.
[1] https://medium.com/@sulbha.jindal/richard-suttons-challenge-...
1) Defining "intelligence" as ability to "understand" isn't actually defining it at all, unless you have a rigorous definition of what it means to understand. It's basically just punting the definition from one loosely defined concept to another.
2) The word "intelligence", in common usage, is only loosely defined, and heavily overloaded, and you'll get 10 different definitions if you ask 10 different people. It's too late to change this, since the meaning of words comes from how they are used. If you want to know the various ways the word is used then look in a dictionary. These are literally the meanings of the word. If you want something more precise then you are not looking for the meaning of the word, but rather trying to redefine it.
3) When we talk about "intelligence" with regards to AI, or AGI, it seems that what people really want to do is to define a new word, something like "hard-intelligence", something rigorously defined, that would let us definitively say whether, or to what degree, an "intelligent" system (animal or machine) has this property or not.
Of course to be useful, this new word "hard-intelligence" needs to be aligned with what people generally mean by "intelligence", and presumably in the future the one of the dictionary senses of "intelligence" will be hard-intelligence.
I think the most useful definition of this new word "hard-intelligence" is going to be a functional one - a capability (not mechanism) of a system, that can be objectively tested for, even with a black box system. However, since the definition should also align with that of "intelligence", which historically refers to an animal/human capability, then it seems useful to also consider where does this animal capability come from, so that our definition can encompass that in most fundamental way possible.
So, with that all said, here's how I would define "hard-intelligence", and why I would define it this way. This post is already getting too long, so I'll keep it brief.
The motivating animal-based consideration for my definition is evolution, and what is the capability that animals evolved to possess intelligence (to varying degrees) have that other animals do not, and what survival benefit does this bring that compensates for the huge cost of large brains in animals with advanced intelligence?
I consider the essence of evolved animal intelligence to be prediction, which means that the animal is not restricted to reacting to the present, but also can plan for the predicted future, which obviously has massive survival benefit - being able to predict where the food and water will be, how the predator is going to behave, etc, etc.
The mechanics of how functional prediction has evolved in different animals varies from something like a fly, whose hard-coded instincts help it avoid predicted swats (that looming visual input predicts I'm about to be swatted by the cow's tail, so I better move), all the way to up to species like ourselves where we can learn predictive signals, outcomes, and adaptive behaviors, rather than these being hard coded. It is widely accepted that our cortex (and equivalent in birds) is basically a prediction machine, which has evolved under selection pressure of developing this super-power of being able to see into the future.
So, my definition of "hard-intelligence" is degree of ability to use, and learn from, past experience to successfully predict the future.
That's it.
There are of course some predictive patterns, and outcomes, that are simple to learn and recognize, and others that are harder, so this is a matter of degree and domain, etc, but at the end of the day it's an objective measure that can be tested for - given the same experiential history to learn from, can different systems correctly predict the continuations of new inputs that follow a similar pattern.
This definition obviously captures the evolutionary super-power of predicting the future, which is at least one of the things that intelligent animals can do, but my assertion, on which the utility of this definition of "hard-intelligence" is based, is that prediction is in fact the underlying mechanism of everything that we consider as "intelligent" behavior. For example, reasoning and planning is nothing more than predicting the outcomes of a sequence of hypothetical what-ifs.
tl/dr - "intelligence" is too fuzzy of a concept to be useful. We need to define a new rigorously defined word to discuss and measure machine (and animal) intelligence. I have suggested a definition.
Every thing I used to come to HN for (learning, curiosity, tech, etc) has been mostly replaced by an artificial version of the thing. I still get tricked daily by titles I think are referring a real thing, but turn out to be about AI.
There is an interiority to our thought life. At least I know there is for myself, because I know what it's like to experience the world as me. I assume that other humans have this same kind of interiority as me, because they are humans like me. And then animals to greater or lesser extent based on how similarly they behave or sense the world around them to humans.
But if there is an "interiority" for LLMs, it must be very very different to humans. The reasoning of an LLM springs into existence for every prompt, then goes away entirely for the next prompt, starting over again from scratch.
Yes this is an over simplification. The LLM has been trained with all kinds of knowledge about the world that persists between invocations. But if the floating point numbers are just sitting there on a disk or other storage medium, it doesn't seem possible that it could be experiencing anything until called into use again.
And the strangeness of the LLM having a completely transformed personality and biases based solely on a few sentences in a prompt. "You are a character in the Lord of the Rings..."
I think this is the sense in which many people argue that an LLM is not "intelligent". It's really an argument that an LLM does not experience the world anything like the way a human being does.
And it turned out that he was, as the Slavs say,—and having no better guidance Le Académie Française I am afraid I have to say—a robot. And I loved him just as much as any supposed human I have met.
But does a robot have intelligence? Ah, but we are getting ahead of ourselves.
LLMs are more likely to judge programs (correctly or incorrectly) as being semantically equivalent when they are syntactically similar, even though syntactically similar programs can actually do drastically different things. In fact LLMs are generally pretty bad at program equivalence, suggesting they don't really "understand" what programs are doing, even for a fairly mechanical definition of "understand".
https://arxiv.org/pdf/2502.12466
While this is a point in time study and I'm sure all these tools will evolve, this matches my intuition for how LLMs behave and the kinds of mistakes they make.
By comparison the approach in this article seems narrow and doesn't explain a whole lot, and more importantly doesn't give us any hypotheses we can actually test against these systems.
barishnamazov•2w ago
Day 1: Fed. (Inductive confidence rises)
Day 100: Fed. (Inductive confidence is near 100%)
Day 250: The farmer comes at 9 AM... and cuts its throat. Happy thanksgiving.
The Turkey was an LLM. It predicted the future based entirely on the distribution of the past. It had no "understanding" of the purpose of the farmer.
This is why Meyer's "American/Inductive" view is dangerous for critical software. An LLM coding agent is the Inductive Turkey example. It writes perfect code for 1000 days because the tasks match the training data. On Day 1001, you ask for something slightly out of distribution, and it confidently deletes your production database because it added a piece of code that cleans your tables.
Humans are inductive machines, for the most part, too. The difference is that, fortunately, fine-tuning them is extremely easy.
usgroup•2w ago
data: T T T T T T F
rule1: for all i: T
rule2: for i < 7: T else F
p-e-w•2w ago
usgroup•2w ago
wasabi991011•2w ago
Yes, "equally likely" is the minimal information prior which makes it best suited when you have no additional information. But it's not unlikely that have some sort of context you can use to decide on a better prior.
usgroup•2w ago
p-e-w•2w ago
But we already know that LLMs can do much better than that. See the famous “grokking” paper[1], which demonstrates that with sufficient training, a transformer can learn a deep generalization of its training data that isn’t just a probabilistic interpolation or extrapolation from previous inputs.
Many of the supposed “fundamental limitations” of LLMs have already been disproven in research. And this is a standard transformer architecture; it doesn’t even require any theoretical innovation.
[1] https://arxiv.org/abs/2301.02679
encyclopedism•2w ago
LLM's are known properties in that they are an algorithm! Humans are not. PLEASE at the very least grant that the jury is STILL out on what humans actually are in terms of their intelligence, that is after all what neuroscience is still figuring out.
barishnamazov•2w ago
Not that humans can't make these mistakes (in fact, I have nuked my home directory myself before), but I don't think it's a specific problem some guardrails can solve currently. I'm looking for innovations (either model-wise or engineering-wise) that'd do better than letting an agent run code until a goal is seemingly achieved.
glemion43•2w ago
Security is my only concern and for that we have a team doing only this but that's also just a question of time.
Whatever LLMs ca do today doesn't matter. It matters how fast it progresses and we will see if we still use LLMs in 5 years or agi or some kind of world models.
bdbdbdb•2w ago
"Humans aren't perfect"
This argument always comes up. The existence of stupid / careless / illiterate people in the workplace doesn't excuse spending trillions on computer systems which use more energy than entire countries and are yet unreliable
glemion43•2w ago
If you have 1% of them and they cost you 50-100k per year than replacing them with computers make plenty of sense.
barishnamazov•2w ago
I'm not sure what you're referring to. I didn't say anything about capabilities of people. If anything, I defend people :-)
> And yes guard rails can be added easily.
Do you mean models can be prevented to do dumb things? I'm not too sure about that, unless a strict software architecture is engineered by humans where LLMs simply write code and implement features. Not everything is web development where we can simply lock filesystems and prod database changes. Software is very complex across the industry.
glemion43•2w ago
People whom you have to always handhold and were code review is fundamental.
You can write tests, pr gates etc.
It's still a scale in what you can let them do unsupervised vs controlling them more closely but already better than real people I know. Because they are also a lot faster.
barishnamazov•2w ago
Low quality engineering is always visible to outside. Low quality engineers using LLMs won't get any better.
glemion43•2w ago
I had big hopes for one person until he copy pasted verbatim LLM responses to me...
naveen99•2w ago
myth_drannon•2w ago
marci•2w ago
Not
"Train from scratch"
mirekrusin•2w ago
aleph_minus_one•2w ago
If this was true, educating people fast for most jobs would be a really easy and solved problem. On the other hand in March 2018, Y Combinator put exactly this into its list of Requests for Startups, which gives strong evidence that this is a rather hard, unsolved problem:
> https://web.archive.org/web/20200220224549/https://www.ycomb...
armchairhacker•2w ago
“‘r’s in strawberry” and other LLM tricks remind me of brain teasers like “finished files” (https://sharpbrains.com/blog/2006/09/10/brain-exercise-brain...). Show an average human this brain teaser and they’ll probably fall for it the first time.
But never a second; the human learned from one instance, effectively forever, without even trying. ChatGPT had to be retrained and to not fall for the “r”’s trick, which cost much more than one prompt, and (unless OpenAI are hiding a breakthrough, or I really don’t understand modern LLMs) required much more than one iteration.
That seems to be the one thing that prevents LLMs from mimicking humans, more noticeable and harder to work around than anything else. An LLM can beat a Turing test where it only must generate a few sentences. No LLM can imitate human conversation over a few years (probably not even a few days), because it would start forgetting much more.
graemep•2w ago
At the school level: teachers are trained, buildings are built, parents rely on kids being at school so they can go out to work....
At higher levels and in training it might be easier to change things, but IMO it is school level education that is the most important for most people and the one that can be improved the most (and the request for startups reflects that).
I can think of lots of ways things can be done better. I have done quite a lot of them as a home educating parent. As far as I can see my government (in the UK) is determined to do the exact opposite of the direction I think we should go in.
sdenton4•2w ago
While I have no doubt you had good results home schooling, you will almost certainly run into difficulty scaling your results.
graemep•2w ago
1. Kids need far fewer hours of one on one than classroom teaching
2. There is much greater proportion of self teaching, especially as kids get older.
I estimate adult time required per child is similar to schools with small class sizes, and it requires somewhat less skilled adults.
Nevermark•2w ago
Which is still a problem of educating humans. Just moved up the chain one step. Educators are often very hard to educate.
Even mathematics isn't immune to this. Calculus is pervasively taught with prematurely truncated algebra of differentials. Which means for second order derivatives and beyond, the "fraction" notation does not actually describe ratios, when this does not need to be the case.
But when will textbooks remove this unnecessary and complicating disconnect between algebra and calculus? There is no significant movement to do so.
Educators and textbook writers are as difficult to educate as anyone else.
funkyfiddler69•2w ago
Because of millions of years of generational iterations, by which I mean recursive teaching, learning and observing, the outcomes of which all involved generations perceive, assimilate and adapt to in some (multi-) culture- and sub-culture driven way that is semi-objectively intertwined with local needs, struggles, personal desires and supply and demand. All that creates a marvelous self-correcting, time-travelling OODA loop. []
Machines are being finetuned by 2 1/2 generations abiding by exactly one culture.
Give it time, boy! (effort put into/in over time)
[] https://en.wikipedia.org/wiki/OODA_loop