frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

An open replacement for the IBM 3174 Establishment Controller

https://github.com/lowobservable/oec
1•bri3d•1m ago•0 comments

The P in PGP isn't for pain: encrypting emails in the browser

https://ckardaris.github.io/blog/2026/02/07/encrypted-email.html
1•ckardaris•3m ago•0 comments

Show HN: Mirror Parliament where users vote on top of politicians and draft laws

https://github.com/fokdelafons/lustra
1•fokdelafons•4m ago•1 comments

Ask HN: Opus 4.6 ignoring instructions, how to use 4.5 in Claude Code instead?

1•Chance-Device•5m ago•0 comments

We Mourn Our Craft

https://nolanlawson.com/2026/02/07/we-mourn-our-craft/
1•ColinWright•8m ago•0 comments

Jim Fan calls pixels the ultimate motor controller

https://robotsandstartups.substack.com/p/humanoids-platform-urdf-kitchen-nvidias
1•robotlaunch•11m ago•0 comments

Exploring a Modern SMTPE 2110 Broadcast Truck with My Dad

https://www.jeffgeerling.com/blog/2026/exploring-a-modern-smpte-2110-broadcast-truck-with-my-dad/
1•HotGarbage•12m ago•0 comments

AI UX Playground: Real-world examples of AI interaction design

https://www.aiuxplayground.com/
1•javiercr•12m ago•0 comments

The Field Guide to Design Futures

https://designfutures.guide/
1•andyjohnson0•13m ago•0 comments

The Other Leverage in Software and AI

https://tomtunguz.com/the-other-leverage-in-software-and-ai/
1•gmays•15m ago•0 comments

AUR malware scanner written in Rust

https://github.com/Sohimaster/traur
3•sohimaster•17m ago•1 comments

Free FFmpeg API [video]

https://www.youtube.com/watch?v=6RAuSVa4MLI
3•harshalone•17m ago•1 comments

Are AI agents ready for the workplace? A new benchmark raises doubts

https://techcrunch.com/2026/01/22/are-ai-agents-ready-for-the-workplace-a-new-benchmark-raises-do...
2•PaulHoule•22m ago•0 comments

Show HN: AI Watermark and Stego Scanner

https://ulrischa.github.io/AIWatermarkDetector/
1•ulrischa•23m ago•0 comments

Clarity vs. complexity: the invisible work of subtraction

https://www.alexscamp.com/p/clarity-vs-complexity-the-invisible
1•dovhyi•24m ago•0 comments

Solid-State Freezer Needs No Refrigerants

https://spectrum.ieee.org/subzero-elastocaloric-cooling
2•Brajeshwar•24m ago•0 comments

Ask HN: Will LLMs/AI Decrease Human Intelligence and Make Expertise a Commodity?

1•mc-0•25m ago•1 comments

From Zero to Hero: A Brief Introduction to Spring Boot

https://jcob-sikorski.github.io/me/writing/from-zero-to-hello-world-spring-boot
1•jcob_sikorski•25m ago•1 comments

NSA detected phone call between foreign intelligence and person close to Trump

https://www.theguardian.com/us-news/2026/feb/07/nsa-foreign-intelligence-trump-whistleblower
9•c420•26m ago•1 comments

How to Fake a Robotics Result

https://itcanthink.substack.com/p/how-to-fake-a-robotics-result
1•ai_critic•26m ago•0 comments

It's time for the world to boycott the US

https://www.aljazeera.com/opinions/2026/2/5/its-time-for-the-world-to-boycott-the-us
3•HotGarbage•27m ago•0 comments

Show HN: Semantic Search for terminal commands in the Browser (No Back end)

https://jslambda.github.io/tldr-vsearch/
1•jslambda•27m ago•1 comments

The AI CEO Experiment

https://yukicapital.com/blog/the-ai-ceo-experiment/
2•romainsimon•28m ago•0 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
5•surprisetalk•32m ago•1 comments

MS-DOS game copy protection and cracks

https://www.dosdays.co.uk/topics/game_cracks.php
4•TheCraiggers•33m ago•0 comments

Updates on GNU/Hurd progress [video]

https://fosdem.org/2026/schedule/event/7FZXHF-updates_on_gnuhurd_progress_rump_drivers_64bit_smp_...
2•birdculture•34m ago•0 comments

Epstein took a photo of his 2015 dinner with Zuckerberg and Musk

https://xcancel.com/search?f=tweets&q=davenewworld_2%2Fstatus%2F2020128223850316274
14•doener•34m ago•2 comments

MyFlames: View MySQL execution plans as interactive FlameGraphs and BarCharts

https://github.com/vgrippa/myflames
1•tanelpoder•35m ago•0 comments

Show HN: LLM of Babel

https://clairefro.github.io/llm-of-babel/
1•marjipan200•36m ago•0 comments

A modern iperf3 alternative with a live TUI, multi-client server, QUIC support

https://github.com/lance0/xfr
3•tanelpoder•37m ago•0 comments
Open in hackernews

Richard Sutton – Father of Reinforced Learning thinks LLMs are a dead end

https://www.dwarkesh.com/p/richard-sutton
12•RyeCombinator•4mo ago

Comments

YeGoblynQueenne•4mo ago
Sutton's alternative to LLMs is RL obviously, I mean duh. He says an alternative theory for the foundation of intelligence is "sensation, action, reward", that animals do this throughout their lives and that intelligence is about figuring out what actions to take to increase the rewards.

Well I have a problem with that, with all respect to Richard Sutton who is one of the AI gods. I don't think his Skinnerian behaviourist paradigm is realistic, I don't think "sensation, action, reward" works in physical reality, in the real world: because in the real world there are situations where pursuing your goals does not increase your reward.

Here's an example of what I mean. Imagine the "reward" that an animal will get from not falling down a cliff and dying. If the animal falls down the cliff and dies, reward is probably negative (maybe even infinitely negative: it's game over, man). But if the animal doesn't fall down the cliff and die, what is the reward?

There's no reward. If there was any reward for not falling down a cliff and dying, then all animals would ever do would be to sit around not falling down cliffs and dying, and just increasing their reward for free. That wouldn't lead to the development of intelligence very fast.

You can try to argue that an animal will obtain a positive reward from just not dying, but that doesn't work: for RL to enforce some behaviour P, it is P that has to be rewarded, not just being alive, in general. Deep RL systems don't learn to play chess by refusing to play.

For RL to work, agents must constantly maximise their reward, not just increase it or just avoid it going negative-infinite. And you just cannot do that in the physical world because there are situations where doing the wrong thing kills you and doing the right thing does not increase your reward.

Digital RL agents can avoid this kind of zero-gains scenario because they can afford to act randomly until they hit a reward, so e.g. an RL chess player can afford to play at random until it figures out how to play. But that doesn't work in the real world, where acting at random has a very high chance of killing an animal. Imagine an animal that randomly jumps off cliffs: game over, man. In the real world if you chase reward without already knowing where it comes from, you better have a very large number of lives [1].

So reward is not all you need. There may be cases where animals use a reward system to guide their behaviours, just like there are cases where humans learn by imitation, but in the general case they don't. It doesn't work. RL doesn't work in the real world and it's not how animals developed intelligence.

__________________

[1] Support for the theory that all animals are descended from cats?