frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Market orientation and national homicide rates

https://onlinelibrary.wiley.com/doi/10.1111/1745-9125.70023
1•PaulHoule•24s ago•0 comments

California urges people avoid wild mushrooms after 4 deaths, 3 liver transplants

https://www.cbsnews.com/news/california-death-cap-mushrooms-poisonings-liver-transplants/
1•rolph•1m ago•0 comments

Matthew Shulman, co-creator of Intellisense, died 2019 March 22

https://www.capenews.net/falmouth/obituaries/matthew-a-shulman/article_33af6330-4f52-5f69-a9ff-58...
1•canucker2016•2m ago•1 comments

Show HN: SuperLocalMemory – AI memory that stays on your machine, forever free

https://github.com/varun369/SuperLocalMemoryV2
1•varunpratap369•3m ago•0 comments

Show HN: Pyrig – One command to set up a production-ready Python project

https://github.com/Winipedia/pyrig
1•Winipedia•5m ago•0 comments

Fast Response or Silence: Conversation Persistence in an AI-Agent Social Network [pdf]

https://github.com/AysajanE/moltbook-persistence/blob/main/paper/main.pdf
1•EagleEdge•5m ago•0 comments

C and C++ dependencies: don't dream it, be it

https://nibblestew.blogspot.com/2026/02/c-and-c-dependencies-dont-dream-it-be-it.html
1•ingve•5m ago•0 comments

Show HN: Vbuckets – Infinite virtual S3 buckets

https://github.com/danthegoodman1/vbuckets
1•dangoodmanUT•6m ago•0 comments

Open Molten Claw: Post-Eval as a Service

https://idiallo.com/blog/open-molten-claw
1•watchful_moose•6m ago•0 comments

New York Budget Bill Mandates File Scans for 3D Printers

https://reclaimthenet.org/new-york-3d-printer-law-mandates-firearm-file-blocking
1•bilsbie•7m ago•0 comments

The End of Software as a Business?

https://www.thatwastheweek.com/p/ai-is-growing-up-its-ceos-arent
1•kteare•8m ago•0 comments

Exploring 1,400 reusable skills for AI coding tools

https://ai-devkit.com/skills/
1•hoangnnguyen•9m ago•0 comments

Show HN: A unique twist on Tetris and block puzzle

https://playdropstack.com/
1•lastodyssey•12m ago•0 comments

The logs I never read

https://pydantic.dev/articles/the-logs-i-never-read
1•nojito•13m ago•0 comments

How to use AI with expressive writing without generating AI slop

https://idratherbewriting.com/blog/bakhtin-collapse-ai-expressive-writing
1•cnunciato•15m ago•0 comments

Show HN: LinkScope – Real-Time UART Analyzer Using ESP32-S3 and PC GUI

https://github.com/choihimchan/linkscope-bpu-uart-analyzer
1•octablock•15m ago•0 comments

Cppsp v1.4.5–custom pattern-driven, nested, namespace-scoped templates

https://github.com/user19870/cppsp
1•user19870•16m ago•1 comments

The next frontier in weight-loss drugs: one-time gene therapy

https://www.washingtonpost.com/health/2026/01/24/fractyl-glp1-gene-therapy/
2•bookofjoe•19m ago•1 comments

At Age 25, Wikipedia Refuses to Evolve

https://spectrum.ieee.org/wikipedia-at-25
1•asdefghyk•22m ago•4 comments

Show HN: ReviewReact – AI review responses inside Google Maps ($19/mo)

https://reviewreact.com
2•sara_builds•22m ago•1 comments

Why AlphaTensor Failed at 3x3 Matrix Multiplication: The Anchor Barrier

https://zenodo.org/records/18514533
1•DarenWatson•23m ago•0 comments

Ask HN: How much of your token use is fixing the bugs Claude Code causes?

1•laurex•27m ago•0 comments

Show HN: Agents – Sync MCP Configs Across Claude, Cursor, Codex Automatically

https://github.com/amtiYo/agents
1•amtiyo•27m ago•0 comments

Hello

2•otrebladih•29m ago•1 comments

FSD helped save my father's life during a heart attack

https://twitter.com/JJackBrandt/status/2019852423980875794
3•blacktulip•32m ago•0 comments

Show HN: Writtte – Draft and publish articles without reformatting, anywhere

https://writtte.xyz
1•lasgawe•33m ago•0 comments

Portuguese icon (FROM A CAN) makes a simple meal (Canned Fish Files) [video]

https://www.youtube.com/watch?v=e9FUdOfp8ME
1•zeristor•35m ago•0 comments

Brookhaven Lab's RHIC Concludes 25-Year Run with Final Collisions

https://www.hpcwire.com/off-the-wire/brookhaven-labs-rhic-concludes-25-year-run-with-final-collis...
4•gnufx•37m ago•0 comments

Transcribe your aunts post cards with Gemini 3 Pro

https://leserli.ch/ocr/
1•nielstron•41m ago•0 comments

.72% Variance Lance

1•mav5431•42m ago•0 comments
Open in hackernews

RL algorithms are less bitter-lesson-pilled than 2015-era deep learning

1•rajap•3mo ago
The real issue isn't reward shaping or curriculum learning - everyone complains about those. The deeper problem is that we're hardcoding the credit assignment timescale into our algorithms.

Discount factors (γ), n-step returns, GAE λ parameters - these are human priors about temporal abstraction baked directly into the learning signal. PPO's GAE(λ) literally tells the algorithm "here's how far into the future you should care about consequences." We're not learning this, we're imposing it. Different domains need different λ values. That's manual feature engineering, RL-style.

Biological learning doesn't have a global discount factor slider. Dopamine and temporal difference learning in the brain operate at multiple timescales simultaneously - the brain learns which timescales matter for which situations. Our algorithms? They get a single γ parameter tuned by grad students.

Even worse: exploration strategies are domain-specific hacks. ε-greedy for Atari, continuous noise processes for robotics, count-based bonuses for sparse rewards. We're essentially doing "exploration engineering" for each domain, like it's 2012 computer vision all over again.

Compare this to supervised learning circa 2015: we stopped engineering features and just scaled transformers. The architecture learned what mattered. RL in 2025? Still tweaking γ, λ, exploration coefficients, entropy bonuses for every new task.

True bitter-lesson compliance would mean learning your own temporal abstractions (dynamic γ), learning how to explore (meta-RL over exploration strategies), and learning credit assignment windows (adaptive eligibility traces). Some promising directions exist - options frameworks, meta-RL, world models with learned abstraction - but they're not mainstream because they're compute-hungry and unstable. We keep returning to human priors because they're cheaper. That's the opposite of the bitter lesson.

The irony is stark: RL researchers talk about "end-to-end learning" while manually tuning the most fundamental learning signal parameters. Imagine if vision researchers were still manually setting feature detector orientations in 2025. That's where RL is.

I predict: The next major RL breakthrough won't come from better policy gradient estimators. It'll come from algorithms that discover their own temporal abstractions and exploration strategies through meta-learning at scale. Only then will RL be bitter-lesson-pilled.