frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Amazon no longer defend cloud customers against video patent infringement claims

https://ipfray.com/amazon-no-longer-defends-cloud-customers-against-video-patent-infringement-cla...
1•ffworld•14s ago•0 comments

Show HN: Medinilla – an OCPP compliant .NET back end (partially done)

https://github.com/eliodecolli/Medinilla
2•rhcm•3m ago•0 comments

How Does AI Distribute the Pie? Large Language Models and the Ultimatum Game

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6157066
1•dkga•3m ago•1 comments

Resistance Infrastructure

https://www.profgalloway.com/resistance-infrastructure/
2•samizdis•8m ago•0 comments

Fire-juggling unicyclist caught performing on crossing

https://news.sky.com/story/fire-juggling-unicyclist-caught-performing-on-crossing-13504459
1•austinallegro•8m ago•0 comments

Restoring a lost 1981 Unix roguelike (protoHack) and preserving Hack 1.0.3

https://github.com/Critlist/protoHack
2•Critlist•10m ago•0 comments

GPS and Time Dilation – Special and General Relativity

https://philosophersview.com/gps-and-time-dilation/
1•mistyvales•13m ago•0 comments

Show HN: Witnessd – Prove human authorship via hardware-bound jitter seals

https://github.com/writerslogic/witnessd
1•davidcondrey•13m ago•1 comments

Show HN: I built a clawdbot that texts like your crush

https://14.israelfirew.co
2•IsruAlpha•15m ago•1 comments

Scientists reverse Alzheimer's in mice and restore memory (2025)

https://www.sciencedaily.com/releases/2025/12/251224032354.htm
1•walterbell•18m ago•0 comments

Compiling Prolog to Forth [pdf]

https://vfxforth.com/flag/jfar/vol4/no4/article4.pdf
1•todsacerdoti•20m ago•0 comments

Show HN: Cymatica – an experimental, meditative audiovisual app

https://apps.apple.com/us/app/cymatica-sounds-visualizer/id6748863721
1•_august•21m ago•0 comments

GitBlack: Tracing America's Foundation

https://gitblack.vercel.app/
2•martialg•21m ago•0 comments

Horizon-LM: A RAM-Centric Architecture for LLM Training

https://arxiv.org/abs/2602.04816
1•chrsw•21m ago•0 comments

We just ordered shawarma and fries from Cursor [video]

https://www.youtube.com/shorts/WALQOiugbWc
1•jeffreyjin•22m ago•1 comments

Correctio

https://rhetoric.byu.edu/Figures/C/correctio.htm
1•grantpitt•22m ago•0 comments

Trying to make an Automated Ecologist: A first pass through the Biotime dataset

https://chillphysicsenjoyer.substack.com/p/trying-to-make-an-automated-ecologist
1•crescit_eundo•26m ago•0 comments

Watch Ukraine's Minigun-Firing, Drone-Hunting Turboprop in Action

https://www.twz.com/air/watch-ukraines-minigun-firing-drone-hunting-turboprop-in-action
1•breve•27m ago•0 comments

Free Trial: AI Interviewer

https://ai-interviewer.nuvoice.ai/
1•sijain2•27m ago•0 comments

FDA intends to take action against non-FDA-approved GLP-1 drugs

https://www.fda.gov/news-events/press-announcements/fda-intends-take-action-against-non-fda-appro...
21•randycupertino•29m ago•11 comments

Supernote e-ink devices for writing like paper

https://supernote.eu/choose-your-product/
3•janandonly•31m ago•0 comments

We are QA Engineers now

https://serce.me/posts/2026-02-05-we-are-qa-engineers-now
1•SerCe•32m ago•0 comments

Show HN: Measuring how AI agent teams improve issue resolution on SWE-Verified

https://arxiv.org/abs/2602.01465
2•NBenkovich•32m ago•0 comments

Adversarial Reasoning: Multiagent World Models for Closing the Simulation Gap

https://www.latent.space/p/adversarial-reasoning
1•swyx•32m ago•0 comments

Show HN: Poddley.com – Follow people, not podcasts

https://poddley.com/guests/ana-kasparian/episodes
1•onesandofgrain•40m ago•0 comments

Layoffs Surge 118% in January – The Highest Since 2009

https://www.cnbc.com/2026/02/05/layoff-and-hiring-announcements-hit-their-worst-january-levels-si...
13•karakoram•40m ago•0 comments

Papyrus 114: Homer's Iliad

https://p114.homemade.systems/
1•mwenge•40m ago•1 comments

DicePit – Real-time multiplayer Knucklebones in the browser

https://dicepit.pages.dev/
1•r1z4•40m ago•1 comments

Turn-Based Structural Triggers: Prompt-Free Backdoors in Multi-Turn LLMs

https://arxiv.org/abs/2601.14340
2•PaulHoule•42m ago•0 comments

Show HN: AI Agent Tool That Keeps You in the Loop

https://github.com/dshearer/misatay
2•dshearer•43m ago•0 comments
Open in hackernews

Richard Sutton – Father of Reinforced Learning thinks LLMs are a dead end

https://www.dwarkesh.com/p/richard-sutton
12•RyeCombinator•4mo ago

Comments

YeGoblynQueenne•4mo ago
Sutton's alternative to LLMs is RL obviously, I mean duh. He says an alternative theory for the foundation of intelligence is "sensation, action, reward", that animals do this throughout their lives and that intelligence is about figuring out what actions to take to increase the rewards.

Well I have a problem with that, with all respect to Richard Sutton who is one of the AI gods. I don't think his Skinnerian behaviourist paradigm is realistic, I don't think "sensation, action, reward" works in physical reality, in the real world: because in the real world there are situations where pursuing your goals does not increase your reward.

Here's an example of what I mean. Imagine the "reward" that an animal will get from not falling down a cliff and dying. If the animal falls down the cliff and dies, reward is probably negative (maybe even infinitely negative: it's game over, man). But if the animal doesn't fall down the cliff and die, what is the reward?

There's no reward. If there was any reward for not falling down a cliff and dying, then all animals would ever do would be to sit around not falling down cliffs and dying, and just increasing their reward for free. That wouldn't lead to the development of intelligence very fast.

You can try to argue that an animal will obtain a positive reward from just not dying, but that doesn't work: for RL to enforce some behaviour P, it is P that has to be rewarded, not just being alive, in general. Deep RL systems don't learn to play chess by refusing to play.

For RL to work, agents must constantly maximise their reward, not just increase it or just avoid it going negative-infinite. And you just cannot do that in the physical world because there are situations where doing the wrong thing kills you and doing the right thing does not increase your reward.

Digital RL agents can avoid this kind of zero-gains scenario because they can afford to act randomly until they hit a reward, so e.g. an RL chess player can afford to play at random until it figures out how to play. But that doesn't work in the real world, where acting at random has a very high chance of killing an animal. Imagine an animal that randomly jumps off cliffs: game over, man. In the real world if you chase reward without already knowing where it comes from, you better have a very large number of lives [1].

So reward is not all you need. There may be cases where animals use a reward system to guide their behaviours, just like there are cases where humans learn by imitation, but in the general case they don't. It doesn't work. RL doesn't work in the real world and it's not how animals developed intelligence.

__________________

[1] Support for the theory that all animals are descended from cats?