Interview with the same two people from a year ago: https://www.dwarkesh.com/p/sholto-douglas-trenton-bricken
Trenton and Sholto are very much “talking their book”. They’re doing it well, but it’s highly filtered and partial chat.
source: know a podcast episode which got removed because an OpenAI employee used “black box” to refer to NNs.
First, the claim that one of the main things holding LLMs back is a lack of expert feedback. To me, that just means the models are guessing—because they don’t have knowledge like humans do, they rely on pattern-matching, not understanding. If the user doesn’t know the answer, the LLM can’t help. That’s not just a minor limitation—it’s foundational. Framing it as a feedback issue is a way of sidestepping the deeper problem.
Second, the speculation about Claude winning a Pulitzer or a Nobel Prize. I get the underlying point—they're wondering whether LLMs are better at creative or scientific work. But couching it in terms of prestigious awards just adds to the hype. Why not just say “creative vs. scientific tasks”? Framing it as “what will it win first?” cheapens what those prizes represent and makes the model seem far more capable than it actually is.
Third, one of them claims a friend at a drug company says they’re about to release a drug discovered by AI. But when pressed for details, it turns out to be pure hearsay. It’s a textbook example of the kind of vague hype that surrounds LLMs—bold claims with no real substance when you dig.
That said, I appreciate that the host, despite being quite friendly with the guests, actually pushes back and holds them to their claims. That's not very common in AI discussions, and I respect it.
> If the user doesn’t know the answer, the LLM can’t help. That’s not just a minor limitation—it’s foundational.
doesn't mean the LLM is not useful. My favorite use case for LLMs is them doing something I know 100% how to solve myself. However, they do it much faster and I can 100% understand their solution and give feedback in case it is wrong. No surprises. I much prefer this as it allows me to work on multiple tasks in parallel.
An accountant knows how to do math but they use the math in Excel to do it for them. They 100% know how to do it themselves manually, but Excel does it far faster.
I don't think this comparison makes sense - humans do not have knowledge as a unique thing differentiating us from LLMs. "Knowledge" is just the ability to produce relevant/"correct" results in some interaction, and some of us have acquired experiences that involve dealing with a particular subject and other knowledgable people within that subject for years or even decades, granting us (very fallible!) "knowledge" within that specific area.
Humans hallucinate answers, even within their area of knowledge. Our memory is imperfect, so we're creating an answer based on bits and pieces, while anything missing is subconsciously glossed over and filled in with something plausible. It's just that the more experiences you have in an area, the better your hallucination. Those weird moments when your parents assure you of some fact that challenge the limits of just how wrong something can be would then be an example of hallucination in the extreme polar opposite end.
(Note that I am not implying that the human brain works like an LLM, but rather just challenge the concept of "knowledge" being fundamenetally different from LLM behavior.)
Confabulate.
It's probably helpful to reach for epistemology to make sense of knowledge as a concept. While something like "true justified belief" is far from a perfect definition of knowledge, it a least is at least correct to a good approximation.
If we "know in advance", why wouldn't the AI?
If the AI doesn't "know in advance", why would we?
When I'm asked a question, the response and thoughts around it emerge only at that point, and from thin air. I certainly don't trawl an internal, constant knowledge bank of retrievable facts, and the time of day or preexisting conversation may affect whether or not I successfully "know" something in a particular conversation, despite having learnt it. At the same time, some things I "know" will just be wrong - the memory distorted or mixed with another, the whole thing imagined, or even just a failed recollection. And no, I certainly don't know if it has gone wrong unless it happens to result in e.g., contradictions I notice myself.
These kinds of justifications for why AI doesn't "know" always imply that the process of human thought emergence is well established and understood and easily differentiated, which just isn't the case. There most certainly are more differences than similarities between AI and the human mind, but I dont think the concept of knowledge is necessarily one of them.
https://www.futurehouse.org/research-announcements/demonstra...
Code: https://github.com/Future-House/robin (NOTE: "will be available")
No. It means that the models don't have access to the physical world to run their own experiments. It's really hard to get smarter without feedback. Try it!
You could have just posted this comment, then not gotten any pushback! And you would have been in the same pit of confabulation. But now feedback lets you correct your misapprehension. Cool init?
Trenton please if you are listening give us an explanation for this!
Kind of funny that it was introduced randomly on Reddit a couple of years ago instead of in a journal or something[0]. But I believe it's widely implemented and used now.
[0] https://www.reddit.com/r/LocalLLaMA/comments/17vonjo/your_se...
Learning that an earlier Opus was highly committed to animal welfare and they’re not sure why; other versions and models have not been
Hearing that their mechanistic interpretability groups recently beat an internal red team quickly, using Claude’s help
Speculation on de-risking AI growth for nation states that aren’t building frontier models
Real real profound comment near the end noting that even if we get no more algorithmic improvements ever, current tech is sufficient to replace most white collar jobs, given enough data.
That last one is a super profound and interesting point to me. We’re at the inflection point where white collar jobs will at the very least be implemented by this tech and overseen by humans, and the current tech and economics make it desirable to do so for companies that pay for a lot of white collar work.
Feels like it’s time to buckle up. Not AI2027 buckle up, but still time to buckle up.
I caught that too, very interesting.
Many people treat their pets as part of their family. Many people make statements online like: "I like dogs / cats more than most people". So it's not surprising that this sentiment would be picked up in training data. I'd wager animals on average are thought of more highly on a site like Reddit than humans are. i.e. the average sentiment of a post on cats is more positive than a post about a human. Cats, in the training data, may be perceived as "more important" or "better" than humans.
It is surprising Opus holds on to this tighter than other models.
I don't know how I feel about it. Writing basic literature reviews is tedious, on the other hand, it is approaching the line where I can feel the temptation to let it work while I sip a cocktail.
At what point will I be replaced by the open science literature we made?
If you are at all interested in the current challenges being grappled on in this space, this does a great job of illuminating some of them. Many many interesting passages in here and the text transcript has links to relevant papers when their topics are brought up. Really like that aspect and would love to see that done a lot more often.
"the AI labs spent a few years quietly scaling up supervised learning, where the best-case outcome was obvious: an excellent simulator of human text
now they are scaling up reinforcement learning, which is something fundamentally different. and no one knows what happens next"
I tend to believe this. AlphaGo and AlphaZero, which were both trained with RL at scale, led to strategies that have never been seen before. They were also highly specialized neural networks for a very specific task, which is quite different from LLMs, which are quite general in their capabilities. Scaling RL on LLMs could lead to models that have very unpredictable behaviors and properties on a variety of tasks.
This is all going to sound rather hyperbolic - but I think we're living in quite unprecedented times, and I am starting to believe Kurzweil's vision of the Singularity. The next 10-20 years are going to be very unpredictable. I don't quite know what the answer will be, but I believe scaling mechanistic interpretability will probably yield some breakthroughs into how these models approach problems.
cratermoon•8mo ago
TheNewsIsHere•8mo ago
anonzzzies•8mo ago
JKCalhoun•8mo ago
simonw•8mo ago
(And, to be fair, I try to always use the terms "reasoning" and "thinking" in scare-quotes when I'm writing about LLMs. But honestly I mostly do that to avoid tedious arguments about how "they're not actually thinking"!)
cratermoon•8mo ago
arghwhat•8mo ago
No one else knows how that works, so it would most certainly be worth sharing, especially what the discriminator for whether something "thinks" would be.
While avoiding being "laughed out of the room", of course.
esafak•8mo ago
See also: Levels of AGI for Operationalizing Progress on the Path to AGI (https://arxiv.org/abs/2311.02462)
bigyabai•8mo ago
We can peer very easily into your train of thought here. You won't produce evidence to logically justify your stance, you can't take a position of authority on the subject, and basically rely purely on pathos to sell a fear of LLMs. Aristotle would call your rhetoric AI-generated if he lived to see the modern age.
loudmax•8mo ago
Obviously, we shouldn't anthropomorphize too much here, and even the most powerful LLMs don't "understand" or "reason" or "think" the way humans do. But whatever they're doing, it's at least analogous to what we do. These concepts are genuinely useful for making better use of these tools.
simonw•8mo ago
emp17344•8mo ago
The fact that these models have devoured the contents of the entire internet and still aren’t AGI is an indication that LLMs won’t develop a capacity for cognition.
nullstyle•8mo ago
"We" didn't have the expectation, but upon interacting with LLMs, we can see and feel and hear that thought is taking place, and have realized that maybe the special place we held for human thought isn't so special as we had hoped.
> The fact that these models have devoured the contents of the entire internet and still aren’t AGI is an indication that LLMs won’t develop a capacity for cognition.
Why do you assume that training on the internet would lead to AGI?
AstroBen•8mo ago
uh.. no we can't?
nullstyle•8mo ago
An example for the dense: A friend of mine used his conversations with ChatGPT to get through a tough spot in his marriage. The distinction between what happened in ChatGPT's internal processing and what would have happened in the internal processing of the brain of a family therapist is a meaningless distinction.
My friend was able to benefit from the thoughtwork of ChatGPT.
AstroBen•8mo ago
nullstyle•8mo ago
simonw•8mo ago
That looks like thinking. If you want to argue it's not thinking you are absolutely welcome to, but you need to say more than just "no we can't?".
613style•8mo ago
At some point the cry will change to, "It only looks like thinking from the outside!" And some time after that: "It only looks conscious from the outside..."
emp17344•8mo ago
https://arxiv.org/abs/2505.13775
Just because these models are good at outputting text that appears to be a series of thoughts, doesn’t mean that LLMs are thinking.
AstroBen•8mo ago
'thinking' here needs a clearer definition. Even for us humans the language is just an after the fact attempt at describing what went on, or is going on inside our heads
Consider: https://www.unsw.edu.au/newsroom/news/2019/03/our-brains-rev...
ustad•8mo ago
Before it makes a move, a little LED labeled ‘thinking’ blinks for a while.
Looks like thinking, too.
emp17344•8mo ago
Again, why should we believe that LLMs are capable of thought, other than the fact that they vaguely “seem” to communicate with conscious intent?
nullstyle•8mo ago
I believe this to be the case because the appearance of thoughtfulness is a smooth function, not a line in the sand. Why are you bringing consciousness into this discussion? Consciousness isn't thoughtfulness.
> Again, why should we believe that LLMs are capable of thought, other than the fact that they vaguely “seem” to communicate with conscious intent?
I define "to think" as the act of applying context to a model. What do you define it as? Given my definition, llm inference seems plainly to be an act of thought.
nullstyle•8mo ago
Answer my question, please. edit: let me rephrase to be more clear with what i'm asking. Why are you assuming that consuming the entire internet would lead to either "AGI is here" or "AGI will never be here"? New techniques for developing better intelligence are emerging all the time.
whilenot-dev•8mo ago
[0]: https://openai.com/index/planning-for-agi-and-beyond/
nullstyle•8mo ago
whilenot-dev•8mo ago
[0]: https://redditinc.com/blog/reddit-and-oai-partner
WhitneyLand•8mo ago
emp17344•8mo ago
keybored•8mo ago
whilenot-dev•8mo ago
To quote N. Chomsky when he made an analogy: "When you have a theory, there are two questions you'd need to ask: (1) Why are things this way? (2) Why are things not that way?"[1]
I get it, this post-truth century is difficult to navigate. Engineering achievements get picked up through a mix of technical excitement and venture boredom, its fallacies are stylized by growth hackers as this unsolvable paradox that seems to fit just the right terminology for any marketing purposes, only to be piped through one hype cycle that just paints another picture of the doomsday of capitalism with black and white colors. /rant
[0]: https://en.wikipedia.org/wiki/Argument_from_ignorance
[1]: https://www.youtube.com/watch?v=axuGfh4UR9Q&t=9438s
simonw•8mo ago
That doesn't mean I think they can think myself. I just get frustrated at the quality of discussion every time this topic comes up.
keybored•8mo ago
> That doesn't mean I think they can think myself. I just get frustrated at the quality of discussion every time this topic comes up.
You’re arguing for the potential of a ghost in a machine based on people not being able to answer hard philosophical questions behind human behavior and properties.[1] That’s a kind of Vulgar Theism quality of argumentation.
[1] Not exclusively human. I’m sure killer whales can think.
whilenot-dev•8mo ago
keybored•8mo ago
Now: If you can’t come up with a good definition on the spot I don’t see why my graph dump can’t do it as well
whytaka•8mo ago
If I were to ask you a question and you were to blurt out an answer by reflex or by unguided stream of consciousness, I can accuse you of not thinking. The kind of thinking I'm referring to you here is one where you take pause to let go of prejudices and consider alternatives before answering.
I'd say that LLMs are at best simulating reflexive streams of consciousness. Even with chain-of-thought, it never pauses to actually "think".
But maybe even our own pauses are just internal chains of thought. Look at me be a reflexive stream of consciousness.
sebzim4500•8mo ago
keybored•8mo ago
monkaiju•8mo ago
icedchai•8mo ago