frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

GPT-5.3-Codex System Card [pdf]

https://cdn.openai.com/pdf/23eca107-a9b1-4d2c-b156-7deb4fbc697c/GPT-5-3-Codex-System-Card-02.pdf
1•tosh•4m ago•0 comments

Atlas: Manage your database schema as code

https://github.com/ariga/atlas
1•quectophoton•7m ago•0 comments

Geist Pixel

https://vercel.com/blog/introducing-geist-pixel
1•helloplanets•10m ago•0 comments

Show HN: MCP to get latest dependency package and tool versions

https://github.com/MShekow/package-version-check-mcp
1•mshekow•18m ago•0 comments

The better you get at something, the harder it becomes to do

https://seekingtrust.substack.com/p/improving-at-writing-made-me-almost
2•FinnLobsien•19m ago•0 comments

Show HN: WP Float – Archive WordPress blogs to free static hosting

https://wpfloat.netlify.app/
1•zizoulegrande•21m ago•0 comments

Show HN: I Hacked My Family's Meal Planning with an App

https://mealjar.app
1•melvinzammit•21m ago•0 comments

Sony BMG copy protection rootkit scandal

https://en.wikipedia.org/wiki/Sony_BMG_copy_protection_rootkit_scandal
1•basilikum•24m ago•0 comments

The Future of Systems

https://novlabs.ai/mission/
2•tekbog•24m ago•1 comments

NASA now allowing astronauts to bring their smartphones on space missions

https://twitter.com/NASAAdmin/status/2019259382962307393
2•gbugniot•29m ago•0 comments

Claude Code Is the Inflection Point

https://newsletter.semianalysis.com/p/claude-code-is-the-inflection-point
3•throwaw12•31m ago•1 comments

Show HN: MicroClaw – Agentic AI Assistant for Telegram, Built in Rust

https://github.com/microclaw/microclaw
1•everettjf•31m ago•2 comments

Show HN: Omni-BLAS – 4x faster matrix multiplication via Monte Carlo sampling

https://github.com/AleatorAI/OMNI-BLAS
1•LowSpecEng•31m ago•1 comments

The AI-Ready Software Developer: Conclusion – Same Game, Different Dice

https://codemanship.wordpress.com/2026/01/05/the-ai-ready-software-developer-conclusion-same-game...
1•lifeisstillgood•34m ago•0 comments

AI Agent Automates Google Stock Analysis from Financial Reports

https://pardusai.org/view/54c6646b9e273bbe103b76256a91a7f30da624062a8a6eeb16febfe403efd078
1•JasonHEIN•37m ago•0 comments

Voxtral Realtime 4B Pure C Implementation

https://github.com/antirez/voxtral.c
2•andreabat•39m ago•1 comments

I Was Trapped in Chinese Mafia Crypto Slavery [video]

https://www.youtube.com/watch?v=zOcNaWmmn0A
2•mgh2•45m ago•0 comments

U.S. CBP Reported Employee Arrests (FY2020 – FYTD)

https://www.cbp.gov/newsroom/stats/reported-employee-arrests
1•ludicrousdispla•47m ago•0 comments

Show HN: I built a free UCP checker – see if AI agents can find your store

https://ucphub.ai/ucp-store-check/
2•vladeta•52m ago•1 comments

Show HN: SVGV – A Real-Time Vector Video Format for Budget Hardware

https://github.com/thealidev/VectorVision-SVGV
1•thealidev•54m ago•0 comments

Study of 150 developers shows AI generated code no harder to maintain long term

https://www.youtube.com/watch?v=b9EbCb5A408
1•lifeisstillgood•54m ago•0 comments

Spotify now requires premium accounts for developer mode API access

https://www.neowin.net/news/spotify-now-requires-premium-accounts-for-developer-mode-api-access/
1•bundie•57m ago•0 comments

When Albert Einstein Moved to Princeton

https://twitter.com/Math_files/status/2020017485815456224
1•keepamovin•58m ago•0 comments

Agents.md as a Dark Signal

https://joshmock.com/post/2026-agents-md-as-a-dark-signal/
2•birdculture•1h ago•0 comments

System time, clocks, and their syncing in macOS

https://eclecticlight.co/2025/05/21/system-time-clocks-and-their-syncing-in-macos/
1•fanf2•1h ago•0 comments

McCLIM and 7GUIs – Part 1: The Counter

https://turtleware.eu/posts/McCLIM-and-7GUIs---Part-1-The-Counter.html
2•ramenbytes•1h ago•0 comments

So whats the next word, then? Almost-no-math intro to transformer models

https://matthias-kainer.de/blog/posts/so-whats-the-next-word-then-/
1•oesimania•1h ago•0 comments

Ed Zitron: The Hater's Guide to Microsoft

https://bsky.app/profile/edzitron.com/post/3me7ibeym2c2n
2•vintagedave•1h ago•1 comments

UK infants ill after drinking contaminated baby formula of Nestle and Danone

https://www.bbc.com/news/articles/c931rxnwn3lo
1•__natty__•1h ago•0 comments

Show HN: Android-based audio player for seniors – Homer Audio Player

https://homeraudioplayer.app
3•cinusek•1h ago•2 comments
Open in hackernews

Interview with Geoffrey Hinton

https://www.ft.com/content/31feb335-4945-475e-baaa-3b880d9cf8ce
41•cs702•5mo ago

Comments

bigyabai•5mo ago
Can we get a second opinion from someone that actually specializes in LLMs?
pajamasam•5mo ago
Oh, you mean like one of the "Godfathers of Deep Learning"[1]?

[1]: https://www.techtimes.com/articles/240511/20190329/godfather...

bigyabai•5mo ago
No, I mean like an expert on LLMs. Hinton has said in his own words that he doesn't understand how LLMs or the transformer architecture work, I don't want his 2 cents.

I'd prefer a second opinion from someone with credentials that aren't cosmetically related to the source material.

cs702•5mo ago
Maybe. At the end of the OP, Hinton qualifies his guesses with:

> "We don’t know what is going to happen, we have no idea, and people who tell you what is going to happen are just being silly," he adds. "We are at a point in history where something amazing is happening, and it may be amazingly good, and it may be amazingly bad. We can make guesses, but things aren’t going to stay like they are."

I agree with Hinton: We only know that things will change, not how they will change. We can only make guesses.

Anyone claiming to know with certainty is full of baloney.

nradov•5mo ago
This is the same clown who said years ago that we should stop training radiologists because AI would automate their jobs away. And now demand for radiologists is higher than ever. Don't waste time with his predictions.
dang•5mo ago
Please make your substantive points without swipes or name-calling. This is in the site guidelines: https://news.ycombinator.com/newsguidelines.html.

Your comment would be fine without that first bit.

pseudolus•5mo ago
https://archive.ph/gzBV7
ggm•5mo ago
I am surprised by the direct quotes where Hinton says "by any measure AI is intelligent" because I think these words will come back to haunt him, much as prognostications from Chomsky dogged his heels.

By narrow measures of outcome, AI synthesises answers which meet needs in questioners. I think intelligence includes aspects of behaviour (such a word) of a system which go beyond simply providing answers. I don't think AI can do this, yet if ever.

Uehreka•5mo ago
> I think intelligence

That very phrasing belies the problem with the word: There is no consensus on what intelligence is, what a clear test for it would be or whether such a test could even exist. There are only people on the internet with personal theories and opinions.

So when people say AI is not intelligent, my next questions are whether rocks, trees, flies, dogs, dolphins, humans and “all humans” are intelligent. The person will answer yes/no immediately in a tone that makes it sound like what they’re saying must be obvious, and yet their answers frequently do not agree with each other. We do not have a consensus definition of intelligence that can be used to include some things and exclude others.

ggm•5mo ago
Exemplifies why I think Hinton was on dangerous ground. He is normally far more cautious in his use of language.
peterashford•5mo ago
We have a model for what intelligence is - what humans do. If we produce a human-like AI I think we'll agree it's intelligent.

The fact that there are degrees of intelligence (dogs > flies) isn't that big of an issue, imo. It's the logically night is day argument - just because we can't point to a clear cut off point between these concepts, doesn't mean they aren't distinct concepts. So it follows with intelligence. It doesn't require consensus, just the same way that "is it night now?" doesn't require consensus

ggm•5mo ago
> I think we'll agree it's intelligent.

If there's one thing I've found never came true for me, it's almost any sentence of substantive opinion about "philosophy" which starts with "I think we'll agree"

And I do think this AI/AGI question is a philosophy question.

I don't know if you'll agree with that.

Whilst your analogy has strong elements of "consensus not required" I am less sure that applies right now, to what we think about AI/AGI. I think consensus is pretty .. important, and also, absent.

peterashford•4mo ago
As a person with a degree in Philosophy, I'm going to disagree with you. Discussions about meaning often rely a great deal on what people would usually expect words to mean.

I do think issues of AI/AGI are often Philosophical in nature and I was not claiming that consensus is not required. What I'm arguing against is the requirement for rigorous definitions for words that exist in natural language. Language is messy and often ambiguous - especially around AI.

If we want to have rigorous terms to discuss the state of the art - which I believe we do - then we should define terms for that purpose. i.e: AI needs it's own jargon.

shkkmo•5mo ago
> We have a model for what intelligence is - what humans do.

At what point does a human become intelligent? Is a 12 cell embryon intelligent? Is a newborn intelligent? Is a 1 year old intelligent?

> It's the logically night is day argument - just because we can't point to a clear cut off point

Um...what? There may be more than one of them, but precise definitions exist for the transitions between day and night. I think that is a very poor analogy to intelligence.

There are not just degrees of intelligence but different kinds. It is easier for us to understand and evaluate intelligence that is more similar ours and it becomes increasingly harder the more alien it becomes.

Given that, I don't see how you could reject that assertion that LLMs have some kind of intelligence.

Asking

peterashford•4mo ago
It wasn't an analogy for Intelligence it was an illustration of a class of definition
imiric•5mo ago
It's frustrating to read this type of response whenever this topic is raised. It does nothing but derail the conversation into absurdism.

Yes, we don't have clear definitions of intelligence, just like we don't for life, and many other fundamental concepts. And yet it's possible to discuss these topics within specific contexts based on a generally and colloquially shared definition. As long as we're willing to talk about this in good faith with the intention to arrive at some interesting conclusions, and not try to "win" an argument.

So, given this, it is safe to assert that we haven't invented artificial intelligence. We have invented something that mimics it very well, which will be useful to us in many domains, but calling this intelligence is a marketing tactic promoted by people who have something to gain from that narrative.

ggm•5mo ago
I am doing this, because normally Hinton is my go-to for cautious, useful input to a debate. When he makes this kind of sweeping statement, my hackles get up. The rest of the article had nothing I didn't expect. I did NOT expect him to make such a sweeping assertion.

They're useful. They're not intelligent. He invited the reproach.

Uehreka•5mo ago
> It does nothing but derail the conversation into absurdism.

The conversation (about whether AI is “intelligent”) was already absurd, I’m just pointing it out ;)

The more important conversation is about whether AI is useful, dangerous, and/or worth it. If AI is competent enough at a task to replace a human for 1/10 the cost, it doesn’t really matter if it “has a mortal soul” or “responds to sensory stimuli” or “can modify its weights in real time”, we need to be talking about what that job loss means for society.

That’s my main frustration: that the “is it intelligent” debate devolves into pointless unsettleable philosophical questions and sucks up all the oxygen, and the actual things of consequence go undiscussed.

throw310822•5mo ago
The annoying thing is that we already had an operative definition of intelligence that worked perfectly well for seventy years. It's the Turing test. We've only become dissatisfied with it because we don't like the fact that machines pass it.
imiric•5mo ago
The Turing test was never meant to measure intelligence, let alone define it. It is an "imitation game" that measures the ability of machines to mimic intelligent behavior enough to fool humans into believing they're interacting with another human, and a thought experiment about the practical implications of that. Machines have arguably been able to do this for decades.

This is interesting in its own right, and has propelled the computing industry since it was proposed, but it's not a measurement of intelligence. The reality is that we don't have a good measurement of intelligence, and struggle to define it to begin with.

throw310822•5mo ago
> The Turing test was never meant to measure intelligence, let alone define it.

Original proposal:

"I propose to consider the question, "Can machines think?" This should begin with definitions of the meaning of the terms "machine" and "think." The definitions might be framed so as to reflect so far as possible the normal use of the words, but this attitude is dangerous [...] Instead of attempting such a definition I shall replace the question by another, which is closely related to it and is expressed in relatively unambiguous words."

Clearly Turing is saying "we cannot precisely define what thinking means, so let's instead check if we can tell apart a human and a machine when they communicate freely through a terminal". It's not about fooling humans (what would be the point of it?) but about replacing the ambiguous question "can they think" with an operative definition that can be tested unambiguously. What Turing is saying is that a machine that passes the test is "as good as if it were thinking".

> Machines have arguably been able to do this for decades.

Absolutely not and it's surprisingly uninformed to claim so.

lostmsu•4mo ago
I'm think half of the debate is due to machines not passing Turing test vs HN crowd.
nerpderp82•5mo ago
His unhinged attacks on Chomsky are comical at this point.
loughnane•5mo ago
I think this is true for most capital equipment, at least until governments step in to take the edge off.

Whether AI has a more powerful effect than it's predecessors remains to be seen. It could.

laweijfmvo•5mo ago
I have zero faith in the US stepping in at this point
SonOfKyuss•5mo ago
Oh they are stepping in. It’s just that they are stepping in to prevent any regulations
nextworddev•5mo ago
Reverse regulation
morkalork•5mo ago
And not just in the US either, they've been threatening Europe with tariffs and sanctions over attempting to regulate American tech companies in Europe.
loughnane•5mo ago
Agreed. I think when the history of this time is written the failure of the government to spread around the gains of capitalism and free trade will be seen as what led to the end of a political era.
bdangubic•5mo ago
yes, the governments (especially of the united states) love nothing more than to take shit from rich and give it to the poor
roflulz•5mo ago
name a country that is better that isn't a petrostate?
bdangubic•5mo ago
germany.

there is a ton more but you asked for one :)

more_corn•5mo ago
Obvs
htrp•5mo ago
Hinton is definitely making a full turn towards luddite.
ks2048•5mo ago
I view him as a scientist - which is orthogonal to the benefits of technology.
jpalawaga•5mo ago
It's certainly possible. It helps remove leverage from people trying to make a living. If your labour isn't needed, then what?

An alternative view could be, this is just the same as every other technological innovation.

gobdovan•5mo ago
Hinton is too speculative and inconsistent for me. A reporter outside the AI field even called him out for saying with confidence that only blue collar work will survive AI by pointing out a few years back he said with the same confidence that only creative work will survive.

I can't but compare his takes with Stuart Russell takes, which are so well grounded, coherent and easily presented. I often revisit Stuart Russell discussion with Steven Pinker on AI for the clarity he brings to the topic.

treyfitty•5mo ago
Eh, idk who Hinton is, but I’d cut him some slack for making both statements- I could imagine a case where “creatives” can semantically be understood as “new blue collar.” Musicians, dancers, photographers… are not blue color manufacturing employees, but they are fiscally more similar than their white collar counterparts. It’s possible he used inconsistent terms because he really means “low-wage employees who are far away from the monetary benefit creation decisions,” but that’s a mouthful
glitchc•5mo ago
If you don't know who Geoffrey Hinton is, I suggest you make a trip to Wikipedia post haste. Our modern LLM renaissance wouldn't exist without him.
nurettin•5mo ago
Ehhh it sounds like he's a poster boy who rode on the success of others (LeCun, Deepmind) and says whatever the current popular opinion is until proven wrong and shows no hint of predictive capability.
glitchc•5mo ago
Say what? Show some respect, son!

Hinton published the seminal paper on backpropagation. He also invented Boltzmann machines, unsupervised learning and mixture of ecperts models. He championed machine learning for 20 years even though there was zero funding for it through the 80s and 90s. He was Yann LeCun's PhD adviser. That means Yann LeCun didn't know ass from tea kettle until Hinton introduced him to machine learning.

Know perchance a fellow by the name of Ilya Sutskever? ChatGPT ring any bells? Also a student of Hinton's. The list is very long.

nurettin•5mo ago
Frankly, this all sounds like hero worship and the language is very cringe.

I know the backprop paper. I've read it in the early 2000s. And I remember Hinton as a co-author. Same with Boltzmann machines. Co-author. "Advisor to that great guy", "Teacher of this great guy", "Nobel price together with that guy" <- all of this leads me to the above conclusion. YMMV

qwertytyyuu•5mo ago
Come on there is space for theatrics on hacker news
polotics•5mo ago
just one example of the halo effect: having been instrumental in the development of an important technology doesn't magically make one an expert in the economic impact of that technology, as economy is a completely different field of study
sriram_malhar•5mo ago

    Frankly, this all sounds like hero worship and the language is very cringe.

"Frankly, I just want to be a contrarian"
nurettin•5mo ago
"Hell hath no fury like a conformist scorned"
glitchc•5mo ago
I'm not a fanboy, far from it. I'm not affiliated with the lab or his work. I'm not even a big fan of machine learning. But Hinton's contributions to the field cannot be understated. He single-handedly kept it alive for two decades amid a massive lack of funding. Anyone who has worked in research will attest that this is an incredible feat.

People on Hacker News seems to idolize the lone genius who somehow pulled himself up from his bootstraps. That person does not exist. The truth is that great minds are made, moulded into shape. That the best people behind our AI technology emerged from his lab is no coincidence. Those trash-talking Hinton on this forum are unlikely to achieve 1/100th of what he has accomplished.

“The housecat may mock the tiger,” said the master, “but doing so will not make his purr into a roar.” [1]

[1] http://www.catb.org/esr/writings/unix-koans/end-user.html

nurettin•4mo ago
Ah, someone who enjoys defending his tigers against cats. Well, enjoy your low hanging fruits, tiger. Someone "so knowledgeable" and totally not a poster boy making bold claims that don't come true at all will continue not making sense to me. But don't get me wrong, I know this is an old man shakes fist at cloud kind of situation. Who cares what doesn't make sense to me anyway? My academic circles don't, and they do fine. I continue to contribute what little I can. But my field is more SR than NN.
pizzalife•5mo ago
“Show some respect?”

Do these historical accolades give him a blank check to be wrong in the present?

eloisant•5mo ago
re-read the comment he was responding to.

"sounds like he's a poster boy who rode on the success of others"

The person who wrote that didn't even bother checking who Hinton was before pulling that sentence out of their ass.

gobdovan•5mo ago
Hinton is the guy from the article. He is a big figure in AI research.

For context: he once argued AI could handle complex tasks but not drawing or music. Then when Stable Diffusion appeared, he flipped to "AI is creative." Now he's saying carpentry will be the last job to be automated, so people should learn that.

The pattern is sweeping, premature claims about what AI can or can't do that don't age well. His economic framing is similarly simplified to the point of being either trivial or misleading.

SanjayMehta•5mo ago
Carpentry is already partially automated. I’ve worked on cutting algorithms to minimise waste. There are a number of startups which will go from a 3D interior design to manufacturing. Think of customised Ikea.
chubot•5mo ago
I guess it's worth reminding people that in 2016, Geoff Hinton said some pretty arrogant things that turned out to be totally wrong:

Let me start by saying a few things that seem obvious. I think if you work as a radiologist, you're like the coyote that’s already over the edge of the cliff but hasn’t yet looked down

It’s just completely obvious that within five years deep learning is going to do better than radiologists.… It might be 10 years, but we’ve got plenty of radiologists already.”

https://www.youtube.com/watch?v=2HMPRXstSvQ

This article has some good perspective:

https://newrepublic.com/article/187203/ai-radiology-geoffrey...

His words were consequential. The late 2010s were filled with articles that professed the end of radiology; I know at least a few people who chose alternative careers because of these predictions.

---

According to US News, radiology is the 7th best paying job in 2025, and the demand is rising:

https://money.usnews.com/careers/best-jobs/rankings/best-pay...

https://radiologybusiness.com/topics/healthcare-management/h...

I asked AI about radiologists in 2025, and it came up with this article:

https://medicushcs.com/resources/the-radiologist-shortage-ad...

The Radiologist Shortage: Rising Demand, Limited Supply, Strategic Response

(Ironically, this article feels spammy to me -- AI is probably being too credulous about what's written on the web!)

---

I read Cade Metz's book about Hinton and the tech transfer from universities to big tech ... I can respect him for persisting in his line of research for 20-30 years, while others saying he was barking up the wrong tree

But maybe this late life vindication led to a chip on his shoulder

The way he phrased this is remarkably confident and arrogant, and not like the behavior of respected scientist (now with a Nobel Prize) ... It's almost like Twitter-speak that made its way into real life, and he's obviously not from the generation that grew up with Twitter

giardini•5mo ago
I wouldn't be too hard on Hinton. Researchers in image processing, geophysics and medicine have been saying the same thing since at least the early 1980's. There was always something coming that was just over the next hill that would take the human out of the loop. That special something always evaporated with time. I suppose it did keep funding coming in.
eloisant•5mo ago
The bottom line is that predicting the future is hard. I'm always skeptical of people who pretend they can.

Of course, because you have different people all predicting a different future, some of them are bound to get it right. That doesn't mean the same person will be right again.

gobdovan•5mo ago
Yeah, even forgot about that... I suppose that the same kind of confidence made him stick with neural nets for so long too, despite mainstream AI thinking it's a dead end. But that's the thing in academia, bold claims get encouraged, as ideas still get you the credit, even if they prove useful decades later and not in the way you imagined.
meowface•5mo ago
To be fair we haven't hit 2026 yet so his prediction might still turn out to be somewhat accurate. But yeah, probably not.
dinfinity•5mo ago
I think the key insight is that AI is (undoubtedly going to be) better at analysis and diagnosis than radiologists, but isn't yet widely deployed because:

1. The medical world doesn't accept new technologies easily. Humans get a much higher pass on bad performance than technology and especially than new technology. Things need to be extensively tested and certified, so adoption is slow.

2. AI is legally very different than a radiologist. The liability structure is completely different, which matters a lot in an environment that deals with life or death decisions.

3. Image analysis is not language analysis and generation. This specific machine learning part is not the bit of machine learning that has advanced enormously in the past two years. General knowledge of the world doesn't help that much when the task is to look at pixels and determine whether it's cancer or not. Now this can be improved by integrating the image analysis with all the other possibly relevant information (case history etc.) and diagnosing the case via that route.

chubot•5mo ago
Well maybe, but none of that implies that there will be fewer radiologists will be employed, or that people studying radiology now are fools.

The overwhelming likely thing is that radiologist jobs will change, just like programming jobs will change.

e.g. see my comment on: Did Google and Stack Overflow "replace" programmers?

https://news.ycombinator.com/item?id=43013363

That is, I do not think programmers will be "replaced". The job will just be different; people will come to rely on LLMs for their jobs, like they rely on search engines.

Likewise, you can probably hire fewer doctors now because Google appeared in ~2000, but nobody talked about them being "replaced". There is NOT less demand for doctors.

---

It also reminds me of the prediction around self-driving cars, which is 13+ years ago at this point:

https://news.ycombinator.com/item?id=45149270

I believe Hacker News mostly fell for the hype in ~2012-2016. And even though the predictions turned out to be comically wrong, many people are still attached to them

https://en.wikipedia.org/wiki/List_of_predictions_for_autono...

i.e. I don't think Hinton will be proven "right" with ANY amount of time. The whole framing is just off.

It's not humans xor AI, it's humans + AI. And the world is not static

dinfinity•4mo ago
The world is indeed not static. That it hasn't happened yet doesn't mean it won't.

Predictions about self driving were off, but far from "comically wrong". Waymo's operations are proof of that.

And to conclude things based on the state of the replacement of programmers after only 2-3 years of ChatGPT being a thing is folly.

The reality is that AI has far fewer limitations and legacy cruft than humans to deal with. Don't get me wrong, I like humans, but our performance is very close to the peak of what it could ever be. That of AI not so much. Remember that AI has been evolving for less than 100 years and it is already where it is today. That took us/biology orders of magnitude more time.

The only real question is how fast it will replace (which) human labor.

ionwake•5mo ago
the guy invented AI who cares if he is a couple years wrong with a prediction that will come to fruition, jeez. I would call the geezer confident not arrogant. Who is this? Some burned post doc? No offence but you are randomly laying into the guy as if this is Oprah
theologic•5mo ago
Hinton's citations is 5x what Russell has. There's a good reason why he's won both the Turing and the Nobel Prize. He is just an incredible researcher. I would make the argument that sometimes when you're an incredibly bright, talented person in terms of understanding problems that many other people simply are incapable of following, you're not the right person to be setting expectations how fast a product may ramp into mainstream society.

Russell is much more measured in his statements and much more policy driven.

In my mind you need to listen to both and try to figure out where they're coming from.

moralestapia•5mo ago
>Hinton's citations is 5x what Russell has.

Ugh, Scientism at its best (worst?). Do you also back up Watson statements about race? I'm sure you don't, as that's not part of your training.

Accomplished researchers can say dumb things too, it happens all the time.

aswegs8•5mo ago
Experts are often worse than laymen in predicting macro developments because they are narrowed by their focus and biased towards it. Tetlock's Superforecasting is a great book on this.

Some condensed source I found on the topic:

https://www.ing.com/Newsroom/News/The-more-famous-an-expert-...

theologic•5mo ago
I would like to confirm that I've been highly influenced by Tetlock, and I appreciate his work. It was an integral part of a strategy in a business group I was running.

With that stated, if you can see beyond his vitriol, Nassim Taleb has some valid comments toward Tetlock's methodologies. I love Taleb, but hate his tendency to try and shock people. However, he does raise valid concerns about fat tails.

danaris•5mo ago
And yet, if an accomplished researcher says something and has ample sourcing to back it up, it's worth paying attention to, even if only to be able to effectively refute it.

Calling it "scientism" to care about these things as a way of dismissing the argument out of hand is anti-intellectualism at its worst.

moralestapia•5mo ago
>Hinton's citations is 5x what Russell has. There's a good reason why he's won both the Turing and the Nobel Prize.

Those are not arguments, that's scientism.

I upvoted you anyway, as you're somehow trying.

theologic•5mo ago
Although I refer to this above, I do like your comment, and I would like to add the historical context for scientism. It was originally proposed by Hayek, who is one of the fundamental pillars of Austrian economics economics. A school which I am a massive fan, and has heavily influenced my thoughts.

He was basically concerned that a group of rulers were trying to use science to dominate everybody else. I agree with Hayek and his concerns.

The problem is that it has now morphed into a term that can basically mean many things. While a bit of a rat hole, wikipedia does have a great treatment of the term and its journey.

When you use a term like this, which by it's very nature is pejorative, and you use it without regard for its definition, it is not only anti-intellectualism, but also poor communication. It turns into communication that is set up to foster divisions rather than learning.

theologic•5mo ago
I write this knowing that it is long, and probably written mainly for myself. However, I find that every once in a while, somebody PMs me and says, "wow, that was great," which strokes my ego and makes me want to post. I hope that somebody does read this—in light of the fact that short answers are often not satisfying.

moralestapia's reply contains a series of fallacies and thinking errors.

1. Straw Man Fallacy and Misrepresentation

moralestapia distorted my statement that Hinton is a well-regarded researcher into the idea that I am suggesting any statement he makes about the AI ramp is correct. This is actually the opposite of what I wrote. Despite Hinton's groundbreaking work in AI, I said he is not good at understanding commercial ramps, as talented people often cannot conceive that others cannot follow their vision.

moralestapia created a classic straw man argument: attacking a misrepresented version of the other’s position.

2. Red Herring and Totally Irrelevant Reference

moralestapia brings in Watson's racist viewpoints that got him ostracized.

Watson’s controversial statements about race have no connection to the topic at hand or to the argument being made and serve only to distract.

3a. Anchoring (Behavioral Economics) and Innuendo Effect

This next point comes from my fascination with Kahneman and Tversky—understanding their framework is incredibly important in how we relate. But more than that, I think they basically pull back the curtain and allow us to see what we do to hide the truth.

The first bit of information we’re exposed to disproportionately influences how we interpret subsequent information or people—that’s anchoring. By introducing Watson’s racist statements in response to my comments, he “anchored” the discussion with something emotionally and morally loaded.

The phrase “Do you also back up Watson's statements about race? I’m sure you don’t, as that’s not part of your training.” superficially acknowledges that I am most likely not defending racist statements. However, just raising the hypothetical (“do you also defend X?”) puts the idea into the discussion.

Research in psychology and behavioral economics confirms that once an accusation or association is mentioned, it becomes part of the mental framework—making “I know you are not” less effective at removing the implication than never introducing it in the first place.

In other words, moralestapia does a great job of saying, "look, there seems to be somebody who could possibly support racist thought," by specifically saying, "I know you don't have this viewpoint."

3b. Guilt by Association and Ad Hominem (Implied)

This is slightly more subtle: if we unpack it, moralestapia is suggesting that if I did support Hinton's commercial ramp (which I don't), then somehow I must also accept all of another person’s viewpoints. As referenced above, moralestapia introduces the notion that I may (although he says he doubts it) be a racist.

4. Hasty Generalization

moralestapia's blanket statement “Accomplished researchers can say dumb things too, it happens all the time” is true in a trivial sense, but it’s being used to generalize and dismiss Hinton’s credibility out of hand, regardless of the context or specifics. This is a generalization that bypasses engagement with the actual substance of Hinton’s expertise or my original point.

I’m a bit less picky here because you could argue that this was exactly my point: although Hinton is clearly brilliant, he can't always appreciate that others may not be able to follow or implement his vision.

But specifically, moralestapia doesn't seem to understand or appreciate Hinton. I’ll try to explain how to listen to a brilliant mind and not dismiss him with statements like "can say dumb things too." I think moralestapia is cutting himself off from an enormous amount of learning by making such dismissals or by accusing others of “Scientism.” In this case, I would also submit that moralestapia does not understand the historic context of how this was derived by Hayek. As somebody that leans heavily toward Austrian School of Economics and Hayek, it's funny of being accused of a term that I would say that I agree with as per Hayek's definition.

So, let’s discuss Hinton.

It’s already been cited in this thread that Hinton has lost credibility because he predicted in 2016 that radiologists would be replaced.

But I’d hope that most people can recognize that Hinton is very capable of making a technical judgment, just not necessarily a marketing or commercial one. That’s the root of my comment: listen to him, but understand his background.

From a technical perspective, using a tensor-based approach and all the tools we have today should, in theory, allow us to replace radiologists with AI. There’s nothing technically preventing this if a business case could be made for the investment. From this standpoint, Hinton was—and is—100% correct.

So, why was he so off?

While it’s not a rigorous argument, I’ll just mention that 1 out of every 5 dollars of our GDP is spent on medicine—basically double any other country—and that number is growing faster than inflation or GDP growth.

I’ll propose, without full argument, that a quick review shows the field is flooded with rent-seeking, as defined by Krueger & Tullock.

As an investor in the medical field, you quickly realize that breaking into medicine is a minefield of regulation, bureaucracy, and entrenched interests.

Navigating the labyrinth of government approvals and encountering medical practitioners—who often seem intent on protecting their own positions—can open your eyes to what rent-seeking really looks like. Why do we pay U.S. doctors 200–300% more than Scandinavian doctors? This is not a rigorous argument, but it should highlight that something looks wrong from 50,000 ft.

Penetration into this field is not a technical problem; it’s a political one. Hinton is a talented researcher who finished his career at a massive company called Google, shielded from market realities. That doesn’t mean we should dismiss his ability to perceive the capabilities of the technology. However, to get a valid time frame, we need to process his output.

Finally, do we say that "well I'll just listen to Stuart Russell." While I think you should listen to him, you need to recognize his impact on the technical community is much less. More than this, the specifics of what has redefined AI, which I would like to call a tensor based approach, is NOT were he made his mark in AI. On the flip side of this, Hinton is all about applying Tensors, and is the reason why he is so widely cited and recognized.

Freedom2•5mo ago
> too speculative and inconsistent for me.

I wonder if he's a HN commenter as well, in that case.

I do appreciate your mention of Stuart Russell however. I've recently been watching a few of his talks and have found them very insightful.

xnx•5mo ago
Thank you for your contributions Geoffrey. Enjoy your retirement. You do not need to be a "thought leader".
arduanika•5mo ago
But wait, what if he says a thing that agrees with all my pre-existing assumptions? Can't he be a "thought leader" then?
deadbabe•5mo ago
More of a thought shepherd. A thought leader is a person who comes up with radical new ideas that start trending as other people subscribe to it.
p1dda•5mo ago
Exactly, his contributions lies in the past.
protocolture•5mo ago
Considering that the general trend has been that a few people get rich and everyone else gets poorer, this is a nothing statement.
arduanika•5mo ago
Fake news, based on confounding intra-country data with international data.
pizza•5mo ago
Isn't the difference between "capital -> labor -> capital" and "capital -> AI -> capital" (which is basically just "capital -> capital -> capital") that it's about the elimination of labor through its financialization? Not just that the poor get poorer, but from the POV of the rich that they don't even exist. (Conditional of course upon AI actually really being that much more productive than people and not dependent on them.)
SubiculumCode•5mo ago
Well, when AI replaces us as consumers, the economy will truly leave us behind
js8•5mo ago
Capitalism already replaces many citizens with consumers, leaving some behind. Citizens have a say in how money are spent, not just how much money they get. Consumers might get nominal choice in how much money they get, but the choice of how to spend is either forced (e.g. rent or healthcare) or is meaningless (e.g. which brand of car to drive instead of having public transport).
eloisant•5mo ago
Don't worry, capitalism knows how to make a business out of poor people. Just look at Walmart, and credit cards.
bitmasher9•5mo ago
AI is just capital, you said it yourself. AI just adds more capital to the capital + labor = more capital equation. The more value capital brings in comparison to labor, the more value capital takes from the result compared to labor.

The issue is the increasing imbalance of capital being overvalued compared to labor, and how that has a negative impact on most individuals.

DalasNoin•5mo ago
I guess pizza is saying that labor (in the sense of human labor/employment) is coming to an end. If AI + robotics just drives the economy capital self multiplies with out labor.
bitmasher9•5mo ago
Is labor actually going to zero?
some_guy_nobel•5mo ago
Really makes you take a step back and wonder why Luigi was so celebrated. /s
visarga•5mo ago
You should also consider that people have AI too, everyone has AI, AI makes no difference in competition. Like having Linux, or Google Search - immensely useful but not a differential. So "capital -> AI -> capital" would just be the inferior option. Like building a product with minimal coding and just putting together open source packages, that anyone could replicate in a few hours.
dns_snek•5mo ago
People don't "have" AI. People are being provided AI as a service by the corporations who invested billions into training it. If they ever make a leap where it can actually replace the workforce, it won't be provided for cheap like it is right now, and all the current open models will be obsolete.
jimbohn•5mo ago
I think that this kind of automation will lead to assets outweighting "real" work way more than they already do now, and I hope I am wrong.
pkrecker•5mo ago
At least for the US (and I suspect for most other countries), that is incorrect. When adjusted for inflation both median wealth and median wages have increased in the last 30 years.
imiric•5mo ago
It's not. It directly contradicts the false utopia that tech bros have been selling us of AI solving world hunger, curing all diseases, bringing prosperity to everyone, and similar nonsense.

A statement like this from someone influential is important to break that narrative, despite the HN crowd finding it obvious.

Gimpei•5mo ago
This is not true. Look at median household income in FRED. It is unequivocally up for everyone across the board.

Inequality has increased but it’s no longer clear that it’s as severe an increase as Piketty and Saez once argued. So, yes, things could certainly be much better. The US could, for example, benefit from a more progressive taxation and a stronger social safety net. But at the same time, we aren’t all headed to hell.

protocolture•5mo ago
Household income isnt really representative while cost of living is increasing faster than income.

If I make 100 tokens and that buys me 100 food, thats better than making 1000 tokens that buys me 1 food.

bgwalter•5mo ago
A surprising number of people are in favor of (or pay lip service to) some kind of VAT on robots and "AI":

https://en.wikipedia.org/wiki/Robot_tax

Make it 50% of the sales price like with cigarettes, since "AI" makes people dumber.

pseudocomposer•5mo ago
This is true of everything anyone invents who happens to be unlucky enough to live under capitalism.

Rich, greedy people ruin everything.

briandw•5mo ago
The average person in the capitalist west is so ridiculously wealthy compared people that lived pre Industrial Revolution, or in any communist or socialist country.
pseudocomposer•5mo ago
The US is the only capitalist country in the west, and our life expectancy is decreasing. Is paper wealth worth a short, miserable life where you work all the time?

Moreover, most of the rest of the world’s poverty exists so that a few greedy pigs here can be even more wealthy. We have the CIA and the one-party system that controls it to thank for that.

johnrob•5mo ago
Businesses fall into 2 categories:

B2C (sell to people)

B2B (sell to B2C companies)

If the “C” is broke, it seems like there won’t by any rich people. In other words, if the masses are poor and jobless, who is sending money to the rich?

Genego•5mo ago
There will always be plenty of money to go around no matter what situation we end up with. It will end up being much more centralized than things are today. But there won't ever be a situation where everyone is going to be poor, even if the masses are.
deadfoxygrandpa•5mo ago
what about fat government contracts
xhrpost•5mo ago
The rich class will have to shrink. Think about poor countries like North Korea, there's still a rich/powerful class, it's just a lot smaller.
seanmcdirmid•5mo ago
North Korea is actually not that poor, well, many African countries are poorer. For the relatively high education level of their population, they are severely underdeveloped, but they would catch up quickly if the Kim’s disappeared somehow.
9rx•5mo ago
> if the masses are poor and jobless, who is sending money to the rich?

What would they need it for? Remember, money is just the accounting of debt. Under the old world model, workers offer a loan to the rich — meaning that they do work for the rich and at some point in the future the rich have to pay that work back with something of equal value [e.g. food, gadgets, etc.].

But if in the new world the rich have AI to do the work they want done, the jobless masses can simply be cut out of the picture. No need to be indebted to them in the first place.

seanmcdirmid•5mo ago
We can summarize this as “wealth can exist without money”, all you need is to accumulate lots of land, resources, and labor (robot or otherwise), and you can be wealthy also.
TomasBM•5mo ago
You can think even simpler.

As one individual, you don't really owe anything to anyone. The only time you owe something is in social terms, when you borrow it in your name, or promise reward for work. And even then, people try to get out of paying things back, but in most cases, the courts, the police or the payees themselves get them to do it anyway.

If you own some land, and suddenly, you can get work on it done without giving almost anything in return (except electrical power), you don't owe anything to anyone. And if you can defend that land effectively, you don't physically need anyone else.

This concept of the social contract, where some abstract group of rich owes something to an abstract group of workers, is actually just a series of consequences that happened to a bunch of individuals when debts weren't paid. But if you're rich, the consequences are no longer an issue, and you're not motivated by some other thing (morals or empathy, for example), the social contract breaks down in your favor.

It's a good thing to remind oneself that social contracts don't maintain themselves, we need to maintain them.

9rx•5mo ago
> This concept of the social contract, where some abstract group of rich owes something to an abstract group of workers, is actually just a series of consequences that happened to a bunch of individuals when debts weren't paid.

The debt to the workers almost never goes unpaid. The workers quickly call the debt to get food and shelter in return.

More often the workers fail to repay their debts to the rich. This is how you get entities like Berkshire Hathaway or Apple sitting on mountains of money. That money is the symbol of the loans that were extended to the workers, with the workers not being able to offer equivalent value in return.

Even among the rich, holding money is unusual, though. They usually like to call the debt for something of real value (e.g. land) as well.

TomasBM•5mo ago
I think we're referring to different debts.

Your original comment said:

   > at some point in the future the rich have to pay that work back with something of equal value [e.g. food, gadgets, etc.].
And this is true in the abstract sense that some rich need to provide the goods, or the money that workers can exchange for said goods, or they'll face consequences. In other words, they better pay their debt, because history has shown that bad things happen - to specific rich individuals or across the society - when debts to workers aren't paid. This is one large part of the social contract.

So, I agree with you there. Then, when you said:

   > But if in the new world the rich have AI to do the work they want done, the jobless masses can simply be cut out of the picture.
To this, I provide the caveat that any rich individual could at any point in history cut out any working individual (let alone non-workers), as long as they could get away with it. The only reason most didn't get away with it was because no one else could do the work, and/or they couldn't protect themselves against eventual retaliation. But most did get away with providing what we today consider to be unfair compensation.

So, if any rich individual gets that equation back in their favor, no one will uphold the social contract (i.e., balance the equation) unless the rest of us do it.

For this, you don't need AGI or robots that can do anything a human can, or a large pool of non-workers. You just need reliable autonomous weapons to coerce most work, or enough automation to get work done when coercion fails; humans can do the rest.

   > The workers quickly call the debt to get food and shelter in return.
This is different from my point. As far as that one rich individual is concerned, their debt to the workers stops when he gives them the money, regardless if it's US dollars or Disney dollars. It's up to the state or another rich individual to accept the money for the goods & services. So, that's the other debt, i.e., that goods will be provided for the money.

I'm not as worried about that other debt, where some money becomes worthless, or the state stops enforcing its currency, or people stop accepting money for goods. I'm more worried about the former debt, where someone can renege or just pay much less, because rich and state interests align. So, even if the state takes over with UBI, and you continue working, you'll just have less and less overall, and there'll be no real leverage against it.

9rx•5mo ago
> As far as that one rich individual is concerned, their debt to the workers stops when he gives them the money

Only if it was the employer that originally extended the debt and it is the worker repaying the debt with their labor. Otherwise the employer still has a debt outstanding. The IOUs are distributed, so, practically speaking, that debt may be to another entity (e.g. a bank), but it really makes no difference how many other people get involved. No matter how you slice it, the math is all the same.

deadbabe•5mo ago
Just because you fail to get rich with AI

does not mean you cannot get rich by some other means

starchild3001•5mo ago
"I guess it’s nice that he’s become more optimistic here. He usually just talks about how it’ll probably kill us all." (From Reddit)
eloisant•5mo ago
This time I can understand his point, and agree with most of it.

I was really surprised to hear a scientist like him, who knows how the tech works, to go full scare of a Skynet AI.

starchild3001•4mo ago
He's a very smart and imaginative guy. The most imaginative people sometimes aren't the best at predicting the future (top 1), but rather good at predicting possibilities (top N).
Kapura•5mo ago
yeah man this is just capitalism. everybody should read piketty.
nick49488171•5mo ago
Isn't that the result of every innovation ever, except for complete revolution?
kristoffer•5mo ago
No. Poverty has significantly declined during the 20th century ..
umeshunni•5mo ago
It's the opposite.

Most revolutions (bolsheviks, cuba, iran, arab spring etc) have made people significantly poorer, while most innovations have made people significantly richer (railroads, electricity, first and second agricultural revolutions, manufacturing)

ks2048•5mo ago
I would be interested in some quantitative answer to this - although I imagine it would be hard to define. You skipped lots of revolutions (American, French) and of course there are non economic “costs” (the Terror of France, etc) as well as the problem of short-vs-long term effects.
mallowdram•5mo ago
Hinton never notices that information isn't psychological, his generation failed at resolving the conduit metaphor, ie making the arbitrary in any ways specific (unless it's about diagnostic targets). Of course Hinton is a doomer, he's bullish on his own tech's explosive growth, which, as semantic dilution and/or delusion, is really just a huge bust, generally. Hinton will be seen as another version of Tucker, not Tesla.
js8•5mo ago
When Hinton was asked, what's the solution, he simply said, "socialism". I am a socialist too. But if you believe his thesis (that AI will concentrate power in the hands of few), I think there is hope. Let me outline a positive vision of AI.

I assume it will be technologically possible to run "medium LLM" (for the lac of a better name) on your phone. A medium LLM is something that knows a limited vocabulary (say slightly larger than simple English), and perhaps doesn't remember the capital of France; but it can reason well with the limited vocabulary. So it can answer what's the capital of France by reading Wikipedia, and likewise, it can work with complicated words using their definitions in terms of simpler words.

Now, everybody would run an AI like that on their phone. It would help people solve real world problems of navigating the world and talking to others. Most importantly, it would help people unite by surpassing the Dunbar number. If you (with help of your phone AI) can keep track of 15 million contacts rather than 150, it is life changing and increases trust-building by orders of magnitude. And soon, machines will be able to do that for us.

Socialists have always emphasized education and communication, for everyone, because these are the true constituents of sovereignty and emancipation. We have, I believe, technological means (widely available universal computation and telecommunication) to provide a kind of mind extension for the mind that will surpass our neocortex and will allow everyone to engage with much larger number of people. Lot more human cooperation will result.

I think activists should look into building and embracing such app - a decentralized communication frontend agent, which would let you build trust with large number of people, without much effort, and help you by coordinating with them (really, learning from them their skills and their struggles). We don't need social media giants to do this for us in a centralized way.

So I posit this techno-anarchism as an antidote against techno-feudalism.

tamimio•5mo ago
http://archive.is/bxNly
nirui•5mo ago
Not sure if I'm allowed to quote from the protected article, but the full talk point was:

> “What’s actually going to happen is rich people are going to use AI to replace workers,” he says. “It’s going to create massive unemployment and a huge rise in profits. It will make a few people much richer and most people poorer. That’s not AI’s fault, that is the capitalist system.”

It's kind curious how that would happen.

In the old days, if you want to maintain a monopoly, you can try drain out the talent pool so no one else can hire the best people to do the work that you're doing, and you can also try patent wall to delay your competitors from launching their product.

But if a worker can be replaced by an AI, it could also mean that the competitiveness of the work is significantly reduced, to the point that theoretically everybody can do it. The only way (i guess) to remain the monopoly is then tightening control on the AI while optimize the process to kill off all potential competitors etc. It's all Red Ocean policies (https://www.wallstreetprep.com/knowledge/red-ocean-strategy/).

"Massive unemployment" maybe, but I don't think "huge rise in profits" is guaranteed.

fragmede•5mo ago
At a high level, if profits are revenue minus costs, and cost goes down, and revenue stays the same, then profit goes up. If you fire all your employees and pay a tenth of that to an AI company instead, if we handwave that it's possible (it's not, currently), then there's your huge rise in profit.
nirui•5mo ago
It might work for a short while. But my point is, what if other companies also wants to join the fun? What could stop those companies from replicating the successor and eventually drive them out of the market?

Back then, you have people, which is hard to duplicate and thus can act as a barrier of entry. But AI is just a program, which can be copied with ease, and runs on maybe expensive but standardized hardware.

staticman2•5mo ago
Hinton said on X “I left so that I could talk about the dangers of AI without considering how this impacts Google,"

Now he is claiming he left to watch Netflix?