frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

A new bridge links the math of infinity to computer science

https://www.quantamagazine.org/a-new-bridge-links-the-strange-math-of-infinity-to-computer-scienc...
75•digital55•3h ago•10 comments

Google Antigravity exfiltrates data via indirect prompt injection attack

https://www.promptarmor.com/resources/google-antigravity-exfiltrates-data
486•jjmaxwell4•4h ago•139 comments

Show HN: We built an open source, zero webhooks payment processor

https://github.com/flowglad/flowglad
176•agreeahmed•5h ago•117 comments

How to repurpose your old phone into a web server

https://far.computer/how-to/
136•louismerlin•3d ago•61 comments

Ilya Sutskever: We're moving from the age of scaling to the age of research

https://www.dwarkesh.com/p/ilya-sutskever-2
113•piotrgrabowski•5h ago•88 comments

Unifying our mobile and desktop domains

https://techblog.wikimedia.org/2025/11/21/unifying-mobile-and-desktop-domains/
30•todsacerdoti•5h ago•10 comments

ZoomInfo CEO blocks researcher after documenting pre-consent biometric tracking

https://github.com/clark-prog/blackout-public
54•SignalDr•2h ago•7 comments

FLUX.2: Frontier Visual Intelligence

https://bfl.ai/blog/flux-2
200•meetpateltech•7h ago•63 comments

Launch HN: Onyx (YC W24) – Open-source chat UI

156•Weves•8h ago•110 comments

Trillions spent and big software projects are still failing

https://spectrum.ieee.org/it-management-software-failures
257•pseudolus•10h ago•235 comments

Jakarta is now the biggest city in the world

https://www.axios.com/2025/11/24/jakarta-tokyo-worlds-biggest-city-population
183•skx001•16h ago•118 comments

Constant-time support coming to LLVM: Protecting cryptographic code

https://blog.trailofbits.com/2025/11/25/constant-time-support-coming-to-llvm-protecting-cryptogra...
24•ahlCVA•9h ago•8 comments

The 101 of analog signal filtering (2024)

https://lcamtuf.substack.com/p/the-101-of-analog-signal-filtering
106•harperlee•4d ago•8 comments

Python is not a great language for data science

https://blog.genesmindsmachines.com/p/python-is-not-a-great-language-for
88•speckx•6h ago•82 comments

Human brains are preconfigured with instructions for understanding the world

https://news.ucsc.edu/2025/11/sharf-preconfigured-brain/
405•XzetaU8•16h ago•275 comments

Bad UX World Cup 2025

https://badux.lol/
109•CharlesW•4h ago•28 comments

Unison 1.0

https://www.unison-lang.org/unison-1-0/
150•pchiusano•3h ago•41 comments

Stop Putting Your Passwords into Random Websites (Yes, Seriously, You Are the PR

https://labs.watchtowr.com/stop-putting-your-passwords-into-random-websites-yes-seriously-you-are...
5•Deeg9rie9usi•1h ago•0 comments

Inflatable Space Stations

https://worksinprogress.co/issue/inflatable-space-stations/
51•bensouthwood•4d ago•18 comments

Orion 1.0

https://blog.kagi.com/orion
323•STRiDEX•6h ago•184 comments

Making Crash Bandicoot (2011)

https://all-things-andy-gavin.com/video-games/making-crash/
181•davikr•10h ago•26 comments

The fall of Labubus and the mush of modern internet trends

https://www.michigandaily.com/arts/digital-culture/the-fall-of-labubus-and-the-mush-of-modern-int...
6•gnabgib•1d ago•0 comments

This blog is now hosted on a GPS/LTE modem (2021)

https://blog.nns.ee/2021/04/01/modem-blog
42•xx_ns•3h ago•5 comments

Most Stable Raspberry Pi? Better NTP with Thermal Management

https://austinsnerdythings.com/2025/11/24/worlds-most-stable-raspberry-pi-81-better-ntp-with-ther...
277•todsacerdoti•16h ago•81 comments

Ozempic does not slow Alzheimer's, study finds

https://www.semafor.com/article/11/25/2025/ozempic-does-not-slow-alzheimers-study-finds
118•danso•6h ago•65 comments

PRC elites voice AI-skepticism

https://jamestown.org/prc-elites-voice-ai-skepticism/
125•JumpCrisscross•1d ago•68 comments

LPLB: An early research stage MoE load balancer based on linear programming

https://github.com/deepseek-ai/LPLB
27•simonpure•6d ago•0 comments

Roblox is a problem but it's a symptom of something worse

https://www.platformer.news/roblox-ceo-interview-backlash-analysis/
213•FiddlerClamp•6h ago•289 comments

US banks scramble to assess data theft after hackers breach financial tech firm

https://techcrunch.com/2025/11/24/us-banks-scramble-to-assess-data-theft-after-hackers-breach-fin...
94•indigodaddy•5h ago•20 comments

Claude Advanced Tool Use

https://www.anthropic.com/engineering/advanced-tool-use
643•lebovic•1d ago•255 comments
Open in hackernews

Ilya Sutskever: We're moving from the age of scaling to the age of research

https://www.dwarkesh.com/p/ilya-sutskever-2
112•piotrgrabowski•5h ago

Comments

andy_ppp•1h ago
So is the translation endless scaling has stopped being as effective?
jsheard•1h ago
The translation is that SSI says that SSIs strategy is the way forward so could investors please stop giving OpenAI money and give SSI that money instead. SSI does not have anything to show yet, nor do they intend to show anything until they have created an actual Machine God, but SSI says they can pull it off so it's all good to go ahead and wire the GDP of Norway directly to Ilya.
gessha•53m ago
It’s a snake oil salesman’s world.
aunty_helen•30m ago
If we take AGI as a certainty, ie we think we can achieve AGI using silicon, then Ilya is one of the best bets you can take if you are looking to invest in this space. He has a history and he's motivated to continue working on this problem.

If you think that AGI is not possible to achieve, then you probably wouldn't be giving anyone money in this space.

shwaj•55m ago
Are you asking whether the whole podcast can be boiled down to that translation, or whether you can infer/translate that from the title?

If the former, no. If the latter, sure, approximately.

Animats•36m ago
It's stopped being cost-effective. Another order of magnitude of data centers? Not happening.

The business question is, what if AI works about as well as it does now for the next decade or so? No worse, maybe a little better in spots. What does the industry look like? NVidia and TSMC are telling us that price/performance isn't improving through at least 2030. Hardware is not going to save us in the near term. Major improvement has to come from better approaches.

Sutskever: "I think stalling out will look like…it will all look very similar among all the different companies. It could be something like this. I’m not sure because I think even with stalling out, I think these companies could make a stupendous revenue. Maybe not profits because they will need to work hard to differentiate each other from themselves, but revenue definitely."

Somebody didn't get the memo that the age of free money at zero interest rates is over.

The "age of research" thing reminds me too much of mid-1980s AI at Stanford, when everybody was stuck, but they weren't willing to admit it. They were hoping, against hope, that someone would come up with a breakthrough that would make it work before the house of cards fell apart.

Except this time everything costs many orders of magnitude more to research. It's not like Sutskever is proposing that everybody should go back to academia and quietly try to come up with a new idea to get things un-stuck. They want to spend SSI's market cap of $32 billion on some vague ideas involving "generalization". Timescale? "5 to 20 years".

This is a strange way to do corporate R&D when you're kind of stuck. Lots of little and medium sized projects seem more promising, along the lines of Google X. The discussion here seems to lean in the direction of one big bet.

You have to admire them for thinking big. And even if the whole thing goes bust, they probably get to keep the house and the really nice microphone holder.

energy123•12m ago
The ideas likely aren't vague at all given who is speaking. I'd bet they're extremely specific. Just not transparently shared with the public because it's intellectual property.
Quothling•36m ago
Not really, but there is a finite amount of data to train models on. I found it rather interesting to hear him talk about how Gemini has been better at getting results out of the data than their competition, and how this is the first insights into a new way of dealing with how they train models on the same data to get different results.

I think the title is an interesting thing, because the scaling isn't about compute. At least as I understand it, what they're running out of is data, and one of the ways they deal with this, or may deal with this, is to have LLM's running concurrently and in competition. So you'll have thousands of models competing against eachother to solve challenges through different approaches. Which to me would suggest that the need for hardware scaling isn't about to stop.

imiric•22m ago
The translation to me is: this cow has run out of milk. Now we actually need to deliver value, or the party stops.
gizmodo59•1h ago
Even as criticism targets major model providers, his inability to answer clearly about revenue & dismissing it as a future concern reveals a great deal about today's market. It's remarkable how effortlessly he, Mira, and others secure billions, confident they can thrive in such an intensely competitive field.

Without a moat defined by massive user bases, computing resources, or data, any breakthrough your researchers achieve quickly becomes fair game for replication. May be there will be new class of products, may be there is a big lock-in these companies can come up with. No one really knows!

markus_zhang•1h ago
TBH if you truly believe you are in the frontier of AI you probably don’t need to care too much about those numbers.

Yes corporations need those numbers, but those few humans are way more valuable than any numbers out there.

Of course, only when others believe that they are in the frontier too.

SilverElfin•56m ago
Mira was a PM who somehow was at the right place at the right time. She isn’t actually an AI expert. Ilya however, is. I find him to be more credible and deserving in terms of research investment. That said, I agree that revenue is important and he will need a good partner (another company maybe) to turn ideas into revenue at some point. But maybe the big players like Google will just acquire them on no revenue to get access to the best research, which they can then turn into revenue.
fragmede•48m ago
That’s kind of a shitty way to put it. Mira wasn’t a PM at OpenAI. She was CTO and before that VP of Engineering. Prior to OpenAI she was an engineer at Tesla on the Model X and Leap Motion. You’re right that she’s not a published ML researcher like Ilya, but "right place, right time" undersells leading the team that shipped ChatGPT, DALL-E, and GPT-4.
impossiblefork•55m ago
I think software patents in AI are a possibility. The transformer was patented after all, with way it was bypassed being the decoder-only models.

Secrecy is also possible, and I'm sure there's a whole lot of that.

outside1234•55m ago
He has no answer for it so the only thing he can do is deflect and turn on the $2T reality distortion field.
signatoremo•18m ago
Nobody knows the answer. He would be lying if he gave any number. His startup is able to secure funding solely based on his credential. The investors know very well but they hope for a big payday.

Do you think OpenAI could project their revenue in 2022, before ChatGPT came out?

luke5441•53m ago
He's just doing research with some grant money? Why would you ask a researcher for a path to profitability?

I just hope the people funding his company are aware that they gave some grant money to some researchers.

jonny_eh•46m ago
Exactly, as far as anyone outside of the deal participants knows, Ilya hasn't made any promises with respect to revenue.
singiamtel•22m ago
Is it a grant? My understanding is that they're raising money as a startup

https://www.reuters.com/technology/artificial-intelligence/o...

alyxya•48m ago
They have a moat defined by being well known in the AI industry, so they have credibility and it wouldn't be hard for anything they make to gain traction. Some unknown player who replicates it, even if it was just as good as what SSI does, will struggle a lot more with gaining attention.
baxtr•36m ago
Being well known doesn’t qualify as a moat.
mrandish•22m ago
Agreed. But it can be a significant growth boost. Senior partners at high-profile VCs will meet with them. Early key hires they are trying to recruit will be favorably influenced by their reputation. The media will probably cover whatever they launch, accelerating early user adoption. Of course, the product still has to generate meaningful value - but all these 'buffs' do make several early startup challenges significantly easier to overcome. (Source: someone who did multiple tech startups without those buffs and ultimately reached success. Spending 50% of founder time for six months to raise first funding is a significant burden (working through junior partners and early skepticism) vs 20% of founder time for three weeks.)
mrandish•36m ago
> confident they can thrive in such an intensely competitive field.

I agree these AI startups are extremely unlikely to achieve meaningful returns for their investors. However, based on recent valley history, it's likely high-profile 'hot startup' founders who are this well-known will do very well financially regardless - and that enables them to not lose sleep over whether their startup becomes a unicorn or not.

They are almost certainly already multi-millionaires (not counting ill-liquid startup equity) just from private placements, signing bonuses and banking very high salaries+bonus for several years. They may not emerge from the wreckage with hundreds of millions in personal net worth but the chances are very good they'll probably be well into the tens of millions.

newyankee•34m ago
Sometimes I wonder who the rational individuals at the other end of these deals are and what makes them so confident. I always assume they have something that general public cannot deduce from public statements
yen223•28m ago
This looks like the classic VC model:

1. Most AI ventures will fail

2. The ones that succeed will be incredibly large. Larger than anything we've seen before

3. No investor wants to be the schmuck who didn't bet on the winners, so they bet on everything.

wrs•26m ago
"Rational [citation needed] individuals at the other end of these deals"

Your assumption is questionable. This is the biggest FOMO party in history.

Nextgrid•23m ago
If the whole market goes to bet at the roulette, you go bet as well.

Best case scenario you win. Worst case scenario you’re no worse off than anyone else.

From that perspective I think it makes sense.

The issue is that investment is still chasing the oversized returns of the startup economy during ZIRP, all while the real world is coasting off what’s been built already.

There will be one day where all the real stuff starts crumbling at which point it will become rational to invest in real-world things again instead of speculation.

(writing this while playing at the roulette in a casino. Best case I get the entertainment value of winning and some money on the side, worst case my initial bet wouldn’t make a difference in my life at all. Investors are the same, but they’re playing with billions instead of hundreds)

SilverElfin•58m ago
How did Dwarkesh manage to build a brand that can attract famous people to his podcast? He didn’t have prior fame from something else in research or business, right? Curious if anyone knows his growth strategy to get here.
piker•53m ago
Seems like he’s Lex without the Rogan association so hardcore liberal folks can listen without having to buy morality offsets. He’s good, and he’s filling a void in an established underserved genre is my take.
camillomiller•48m ago
Fridman is a morally broken grifter, who just built a persona and a brand on proven lies, claiming an association with MIT that was de facto non-existent. Not wanting to give the guy recognition is not a matter of being liberal or conservative, but just interested in truthfulness.
wahnfrieden•42m ago
Patel takes anticommunism to such an extreme that he speculates whether naziism is preferable, that Hitler should have the war against Soviets, that the US should have collaborated with Hitler to defeat communism, and that the enduring spread of naziism would have been a good tradeoff to make.
chermi•36m ago
Where does he say this?
wahnfrieden•22m ago
the Sarah Paine interviews
pxc•6m ago
[delayed]
fragmede•43m ago
Tell me more about these morality offsets I can buy! I got a bunch of friends that listen to Joe Rogan, so I listen to him to know what they're talking about, but I've been doing so without these offsets, so my morality's been taking hits. Please help me before I make a human trafficking app for Andrew Tate!
just-the-wrk•38m ago
I think its important to include that Lex is laundromat for whatever the guest is trying to sell. Dwarkesh does an impressive amount of background and speaks with experts about their expertise.
pxc•29m ago
> I think its important to include that Lex is laundromat for whatever the guest is trying to sell.

This is also Rogan's chief problem as a podcaster, isn't it?

bugglebeetle•25m ago
His recent conversation with Sutton suggests otherwise. Friedman is a vapid charlatan par excellence. Dwarkesh suffers from a different problem, where, by rubbing shoulders with experts, he has come to the mistaken belief that he possesses expertise, absent the actual humility and actual work that would entail.
FergusArgyll•38m ago
He's the best interviewer I ever found, try listening to his first couple episodes - they're from his dorm or something. If you can think of a similar style and originality in questioning I'd love a suggestion!
chermi•37m ago
People are impressed by his interviews because he puts a lot of effort into researching the topic before the interview. This is a positive feedback loop.
just-the-wrk•37m ago
He does deep research on topics and invites people who recognize his efforts and want to engage with an informed audience.
l5870uoo9y•31m ago
Overnight success takes years (he has been doing the podcast for 5 years).
polishdude20•4m ago
Maybe he's an Industry plant
oytis•58m ago
Ages just keep flying by
scotty79•52m ago
Translation: Free lunch of getting results just by throwing money at the problem is over. Now for the first time in years we actually need to think what we are doing and firgure out why things that work, do work.

Somehow, despite being vastly overpaid I think AI researchers will turn out to be deeply inadequate for the task. As they have been during the last few AI winters.

eats_indigo•51m ago
did he just say locomotion came from squirrels
jonny_eh•44m ago
timestamp?
FergusArgyll•40m ago
I think he was referencing something Richard Sutton said (iirc); along the lines of "If we can get to the intelligence of a squirrel, we're most of the way there"
Animats•27m ago
I've been saying that for decades now. My point was that if you could get squirrel-level common sense, defined as not doing anything really bad in the next thirty seconds while making some progress on a task, you were almost there. Then you can back-seat drive the low-level system with something goal-oriented.

I once said that to Rod Brooks, when he was giving a talk at Stanford, back when he had insect-level robots and was working on Cog, a talking head. I asked why the next step was to reach for human-level AI, not mouse-level AI. Insect to human seemed too big a jump. He said "Because I don't want to go down in history as the creator of the world's greatest robot mouse".

He did go down in history as the creator of the robot vacuum cleaner, the Roomba.

alyxya•51m ago
The impactful innovations in AI these days aren't really from scaling models to be larger. It's more concrete to show higher benchmark scores, and this implies higher intelligence, but this higher intelligence doesn't necessarily translate to all users feeling like the model has significantly improved for their use case. Models sometimes still struggle with simple questions like counting letters in a word, and most people don't have a use case of a model needing phd level research ability.

Research now matters more than scaling when research can fix limitations that scaling alone can't. I'd also argue that we're in the age of product where the integration of product and models play a major role in what they can do combined.

TheBlight•47m ago
"Scaling" is going to eventually apply to the ability to run more and higher fidelity simulations such that AI can run experiments and gather data about the world as fast and as accurately as possible. Pre-training is mostly dead. The corresponding compute spend will be orders of magnitude higher.
alyxya•38m ago
That's true, I expect more inference time scaling and hybrid inference/training time scaling when there's continual learning rather than scaling model size or pretraining compute.
TheBlight•32m ago
Simulation scaling will be the most insane though. Simulating "everything" at the quantum level is impossible and the vast majority of new learning won't require anything near that. But answers to the hardest questions will require as close to it as possible so it will be tried. Millions upon millions of times. It's hard to imagine.
pron•33m ago
> this implies higher intelligence

Not necessarily. The problem is that we can't precisely define intelligence (or, at least, haven't so far), and we certainly can't (yet?) measure it directly. And so what we have are certain tests whose scores, we believe, are correlated with that vague thing we call intelligence in humans. Except these test scores can correlate with intelligence (whatever it is) in humans and at the same time correlate with something that's not intelligence in machines. So a high score may well imply high intellignce in humans but not in machines (e.g. perhaps because machine models may overfit more than a human brain does, and so an intelligence test designed for humans doesn't necessarily measure the same thing we think of when we say "intelligence" when applied to a machine).

This is like the following situation: Imagine we have some type of signal, and the only process we know produces that type of signal is process A. Process A always produces signals that contain a maximal frequency of X Hz. We devise a test for classifying signals of that type that is based on sampling them at a frequency of 2X Hz. Then we discover some process B that produces a similar type of signal, and we apply the same test to classify its signals in a similar way. Only, process B can produce signals containing a maximal frequency of 10X Hz and so our test is not suitable for classifying the signals produced by process B (we'll need a different test that samples at 20X Hz).

alyxya•25m ago
Fair, I think it would be more appropriate to say higher capacity.
pron•16m ago
Ok, but the point of a test of this kind is to generalise its result. I.e. the whole point of an intelligence test is that we believe that a human getting a high score on such a test is more likely to do some useful things not on the test better than a human with a low score. But if the problem is that the test results - as you said - don't generalise as we expect them, then the tests are not very meaningful to begin with. If we don't know what to expect from a machine with a high test score when it comes to doing things not on the test, then the only "capacity" we're measuring is the capacity to do well on such tests, and that's not very useful.
pessimizer•25m ago
> most people don't have a use case of a model needing phd level research ability.

Models also struggle at not fabricating references or entire branches of science.

edit: "needing phd level research ability [to create]"?

nutjob2•14m ago
> this implies higher intelligence

Models aren't intelligent, the intelligence is latent in the text (etc) that the model ingests. There is no concrete definition of intelligence, only that humans have it (in varying degrees).

The best you can really state is that a model extracts/reveals/harnesses more intelligence from its training data.

dragonwriter•11m ago
> There is no concrete definition of intelligence

Note that if this is true (and it is!) all the other statements about intelligence and where it is and isn’t found in the post (and elsewhere) are meaningless.

darkmighty•8m ago
There is no concrete definition of a chair either.
Herring•40m ago
https://metr.org/blog/2025-03-19-measuring-ai-ability-to-com...

He’s wrong we still scaling, boys.

epistasis•32m ago
That blog post is eight months old. That feels like pretty old news in the age of AI. Has it held since then?
conception•26m ago
It looks like it’s been updated as it has codex 5.1 max on it
rockinghigh•31m ago
You should read the transcript. He's including 2025 in the age of scaling.

> Maybe here’s another way to put it. Up until 2020, from 2012 to 2020, it was the age of research. Now, from 2020 to 2025, it was the age of scaling—maybe plus or minus, let’s add error bars to those years—because people say, “This is amazing. You’ve got to scale more. Keep scaling.” The one word: scaling.

> But now the scale is so big. Is the belief really, “Oh, it’s so big, but if you had 100x more, everything would be so different?” It would be different, for sure. But is the belief that if you just 100x the scale, everything would be transformed? I don’t think that’s true. So it’s back to the age of research again, just with big computers.

Herring•24m ago
Nope, Epoch.ai thinks we have enough to scale till 2030 at least. https://epoch.ai/blog/can-ai-scaling-continue-through-2030

^

/_\

***

jmkni•38m ago
This reveals a new source of frustration, I can't watch this in work, and I don't want to read and AI generated summary so...?
cheeseblubber•32m ago
There is a transcript of the entire conversation if you scroll down a little
delichon•38m ago
If the scaling reaches the point at which the AI can do the research at all better than natural intelligence, then scaling and research amount to the same thing, for the validity of the bitter lesson. Ilya's commitment to this path is a statement that he doesn't think we're all that close to parity.
itissid•37m ago
All coding agents are geared towards optimizing one metric, more or less, getting people to put out more tokens — or $$$.

If these agents moved towards a policy where $$$ were charged for project completion + lower ongoing code maintenance cost, moving large projects forward, _somewhat_ similar to how IT consultants charge, this would be a much better world.

Right now we have chaos monkey called AI and the poor human is doing all the cleanup. Not to mention an effing manager telling me you now "have" AI push 50 Features instead of 5 in this cycle.

kace91•24m ago
>this would be a much better world.

Would it?

We’d close one of the few remaining social elevators, displace higher educated people by the millions and accumulate even more wealth at the top of the chain.

If LLMs manage similar results to engineers and everyone gets free unlimited engineering, we’re in for the mother of all crashes.

On the other hand, if LLMs don’t succeed we’re in for a bubble bust.

wrs•30m ago
"The idea that we’d be investing 1% of GDP in AI, I feel like it would have felt like a bigger deal, whereas right now it just feels...[normal]."

Wow. No. Like so many other crazy things that are happening right now, unless you're inside the requisite reality distortion field, I assure you it does not feel normal. It feels like being stuck on Calvin's toboggan, headed for the cliff.

johnxie•29m ago
I don’t think he meant scaling is done. It still helps, just not in the clean way it used to. You make the model bigger and the odd failures don’t really disappear. They drift, forget, lose the shape of what they’re doing. So “age of research” feels more like an admission that the next jump won’t come from size alone.
energy123•24m ago
It still does help in the clean way it used to. The problem is that the physical world is providing more constraints like lack of power and chips. Three years ago there was scaling headroom created by the gaming industry and other precursor activities.
kmmlng•15m ago
The scaling laws are also power laws, meaning that most of the big gains happen early in the curve, and improvements become more expensive the further you go along.
_giorgio_•24m ago
Scaling is not over, there's no wall.

Oriol Vinyals VP of Gemini research

https://x.com/OriolVinyalsML/status/1990854455802343680?t=oC...

neonate•21m ago
https://xcancel.com/OriolVinyalsML/status/199085445580234368...?
JohnnyMarcone•8m ago
He didn't say it's over, just that continued scaling won't be transformational.
lvl155•24m ago
You have LLMs but you also need to model actual intelligence, not its derivative. Reasoning models are not it.
xeckr•23m ago
He is, of course, incentivised to say that.
malfist•10m ago
Researcher says it's time to fund research. News at 11
londons_explore•16m ago
> These models somehow just generalize dramatically worse than people. It's a very fundamental thing

My guess is we'll discover that biological intelligence is 'learning' not just from your experience, but that of thousands of ancestors.

There are a few weak pointers in that direction. Eg. A father who experiences a specific fear can pass that fear to grandchildren through sperm alone. [1].

I believe this is at least part of the reason humans appear to perform so well with so little training data compared to machines.

[1]: https://www.nature.com/articles/nn.3594

l5870uoo9y•15m ago
> These models somehow just generalize dramatically worse than people.

The whole mess surrounding Grok's ridiculous overestimation of Elon's abilities in comparison to other world stars, did not so much show Grok's sycophancy or bias towards Elon, as it showed that Grok fundamentally cannot compare (generalize) or has a deeper understanding of what the generated text is about. Calling for more research and less scaling is essentially saying; we don't know where to go from here. Seems reasonable.

radicaldreamer•11m ago
I think the problem with that is that Grok has likely been prompted to do that in the system prompt or some prompts that get added for questions about Elon. That doesn't reflect on the actual reasoning or generalization abilities of the underlying model most likely.
asolove•6m ago
Yes it does.

Today on X, people are having fun baiting Grok into saying that Elon Musk is the world’s best drinker of human piss.

If you hired a paid PR sycophant human, even of moderate intelligence, it would know not to generalize from “say nice things about Elon” to “say he’s the best at drinking piss”.

l5870uoo9y•3m ago
You can also give AI models Nobel-prize winning world literature and ask why this is bad and they will tear apart the text, without ever thinking "wait this is some of the best writing produced by man".
Havoc•6m ago
There’s no way that wasn’t specifically prompted.
tmp10423288442•12m ago
He's talking his book. Doesn't mean he's wrong, but Dwarkesh is now big enough that you should assume every big name there is talking their book.
alexnewman•10m ago
A lot more of human intelligence is hard coded