frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

A delightful Mac app to vibe code beautiful iOS apps

https://milq.ai/hacker-news
1•jdjuwadi•2m ago•1 comments

Show HN: Gemini Station – A local Chrome extension to organize AI chats

https://github.com/rajeshkumarblr/gemini_station
1•rajeshkumar_dev•2m ago•0 comments

Welfare states build financial markets through social policy design

https://theloop.ecpr.eu/its-not-finance-its-your-pensions/
2•kome•6m ago•0 comments

Market orientation and national homicide rates

https://onlinelibrary.wiley.com/doi/10.1111/1745-9125.70023
3•PaulHoule•6m ago•0 comments

California urges people avoid wild mushrooms after 4 deaths, 3 liver transplants

https://www.cbsnews.com/news/california-death-cap-mushrooms-poisonings-liver-transplants/
1•rolph•7m ago•0 comments

Matthew Shulman, co-creator of Intellisense, died 2019 March 22

https://www.capenews.net/falmouth/obituaries/matthew-a-shulman/article_33af6330-4f52-5f69-a9ff-58...
3•canucker2016•8m ago•1 comments

Show HN: SuperLocalMemory – AI memory that stays on your machine, forever free

https://github.com/varun369/SuperLocalMemoryV2
1•varunpratap369•9m ago•0 comments

Show HN: Pyrig – One command to set up a production-ready Python project

https://github.com/Winipedia/pyrig
1•Winipedia•11m ago•0 comments

Fast Response or Silence: Conversation Persistence in an AI-Agent Social Network [pdf]

https://github.com/AysajanE/moltbook-persistence/blob/main/paper/main.pdf
1•EagleEdge•11m ago•0 comments

C and C++ dependencies: don't dream it, be it

https://nibblestew.blogspot.com/2026/02/c-and-c-dependencies-dont-dream-it-be-it.html
1•ingve•12m ago•0 comments

Show HN: Vbuckets – Infinite virtual S3 buckets

https://github.com/danthegoodman1/vbuckets
1•dangoodmanUT•12m ago•0 comments

Open Molten Claw: Post-Eval as a Service

https://idiallo.com/blog/open-molten-claw
1•watchful_moose•13m ago•0 comments

New York Budget Bill Mandates File Scans for 3D Printers

https://reclaimthenet.org/new-york-3d-printer-law-mandates-firearm-file-blocking
2•bilsbie•14m ago•1 comments

The End of Software as a Business?

https://www.thatwastheweek.com/p/ai-is-growing-up-its-ceos-arent
1•kteare•15m ago•0 comments

Exploring 1,400 reusable skills for AI coding tools

https://ai-devkit.com/skills/
1•hoangnnguyen•15m ago•0 comments

Show HN: A unique twist on Tetris and block puzzle

https://playdropstack.com/
1•lastodyssey•19m ago•0 comments

The logs I never read

https://pydantic.dev/articles/the-logs-i-never-read
1•nojito•20m ago•0 comments

How to use AI with expressive writing without generating AI slop

https://idratherbewriting.com/blog/bakhtin-collapse-ai-expressive-writing
1•cnunciato•21m ago•0 comments

Show HN: LinkScope – Real-Time UART Analyzer Using ESP32-S3 and PC GUI

https://github.com/choihimchan/linkscope-bpu-uart-analyzer
1•octablock•21m ago•0 comments

Cppsp v1.4.5–custom pattern-driven, nested, namespace-scoped templates

https://github.com/user19870/cppsp
1•user19870•22m ago•1 comments

The next frontier in weight-loss drugs: one-time gene therapy

https://www.washingtonpost.com/health/2026/01/24/fractyl-glp1-gene-therapy/
2•bookofjoe•25m ago•1 comments

At Age 25, Wikipedia Refuses to Evolve

https://spectrum.ieee.org/wikipedia-at-25
2•asdefghyk•28m ago•4 comments

Show HN: ReviewReact – AI review responses inside Google Maps ($19/mo)

https://reviewreact.com
2•sara_builds•28m ago•1 comments

Why AlphaTensor Failed at 3x3 Matrix Multiplication: The Anchor Barrier

https://zenodo.org/records/18514533
1•DarenWatson•30m ago•0 comments

Ask HN: How much of your token use is fixing the bugs Claude Code causes?

1•laurex•33m ago•0 comments

Show HN: Agents – Sync MCP Configs Across Claude, Cursor, Codex Automatically

https://github.com/amtiYo/agents
1•amtiyo•34m ago•0 comments

Hello

2•otrebladih•35m ago•1 comments

FSD helped save my father's life during a heart attack

https://twitter.com/JJackBrandt/status/2019852423980875794
3•blacktulip•38m ago•0 comments

Show HN: Writtte – Draft and publish articles without reformatting, anywhere

https://writtte.xyz
1•lasgawe•40m ago•0 comments

Portuguese icon (FROM A CAN) makes a simple meal (Canned Fish Files) [video]

https://www.youtube.com/watch?v=e9FUdOfp8ME
1•zeristor•42m ago•0 comments
Open in hackernews

Recursive Language Models (RLMs)

https://alexzhang13.github.io/blog/2025/rlm/
135•talhof8•3mo ago

Comments

jgbuddy•3mo ago
This is old news! Agent-loops are not a model architechture
laughingcurve•3mo ago
Everything old is new again when you are in academia
hodgehog11•3mo ago
This feels primarily like an issue with machine learning, at least among mathematical subdisciplines. As new people continue to be drawn into the field, they rarely bother to read what has come even a few years prior (nevermind a few decades prior).
layer8•3mo ago
Loops aren’t recursion?
antonvs•3mo ago
Loops and recursion are fundamentally equivalent.

See e.g. https://textbooks.cs.ksu.edu/cc210/16-recursion/08-recursion...

layer8•3mo ago
Only if you have indexable memory that you can use as a stack, which in the context of LMs isn’t a given.

As another example, a finite-state-machine language can have loops, but it can’t recurse unless there is external memory it has access to in a way that it can serve as a stack. Regular expressions also fall into that pattern; they can loop, but they can’t recurse. For that you need a pushdown automaton: https://en.wikipedia.org/wiki/Pushdown_automaton.

adastra22•3mo ago
I’m confused over your definition of model architecture.
halfmatthalfcat•3mo ago
It broke new ground!
ayazhan•3mo ago
https://arxiv.org/abs/2510.04871 another recursive based model
yorwba•3mo ago
It's a completely different kind of recursion for a completely different (non-language) task.
foolswisdom•3mo ago
I actually came here expecting this to be a language model application of that recursive reasoning paper.
d0mine•3mo ago
> TRM obtains 45% test-accuracy on ARC-AGI-1 and 8% on ARC-AGI-2, higher than most LLMs (e.g., Deepseek R1, o3-mini, Gemini 2.5 Pro) with less than 0.01% of the parameters.
gdiamos•3mo ago
Recursion is so popular in computing that this term “recursive language model” is heavily overloaded

It was even before the rise of LLMs

The authors may want to consider a more specific name

quibit•3mo ago
> Lastly, in our experiments we only consider a recursive depth of 1 — i.e. the root LM can only call LMs, not other RLMs. It is a relatively easy change to allow the REPL environment to call RLMs instead of LMs, but we felt that for most modern “long context” benchmarks, a recursive depth of 1 was sufficient to handle most problems. However, for future work and investigation into RLMs, enabling larger recursive depth will naturally lead to stronger and more interesting systems.

It feels a little disingenuous to call it a Recursive Language Model when the recursive depth of the study was only 1.

yandie•3mo ago
This isn't just context optimization. Not much different from agent-to-agent workflow IMO.
cs702•3mo ago
Briefly, an RLM wraps an existing language model (LM) together with an environment that can dynamically manipulate the prompt that will be fed into the LM.

The authors use as an environment a Python REPL that itself can call other instances of the LM. The prompt is programmatically manipulated as a Python variable on the REPL.

The motivation is for the LM to use Python commands, including commands that call other LM instances, to figure out how best to modify the context at inference time.

The results from early testing look impressive at a first glance: An RLM wrapping GPT-5-mini outperforms GPT-5 by a wide margin on long-context tasks, at significant lower cost.

I've added this to my reading list.

NitpickLawyer•3mo ago
A comparison to dSPY would be nice. cmd+f in the provided link doesn't bring any results tho...
cs702•3mo ago
An RLM is like a language model using DSPy plus all of Python to manipulate its prompt.
integricho•3mo ago
Sounds like unforgivable overhead for very questionable benefits, this whole LLM space is an overengineered slop, and everyone is jumping in building layers on top of layers of slop.
nowittyusername•3mo ago
My existing project is very similar to this with some other goodies. I agree with the author that focus on systems versus LLM's is the proper next move. Orchestrating systems that manage multiple different llms and other scripts together can accomplish a lot more then a simple ping pong type of behavior. Though I suspect most people who work with agentic solutions are already quite aware of this. What most in that space haven't cracked yet is the dynamic self modifying and improving system, that should be the ultimate goal for these types of systems.
ipnon•3mo ago
Hopefully this can solve the problem of Claude needing to compact itself every 10 minutes, blocking execution. It would be better if it was always compacting in the background. But that requires perhaps more compute than is realistic.
wild_egg•3mo ago
Tell it to use subagents more. I often say something like "you're banned from taking direct actions, use subagents for everything" and it can run easily for 60-90 minutes before a compaction.
rancar2•3mo ago
For that issue, try Codex until Claude catches up to your style.
behnamoh•3mo ago
in today's news: MIT researchers found out about AI agents and rebranded it as RLM for karma.
rf15•3mo ago
or: found out about RNNs with extra steps.
fizx•3mo ago
I read the article, and I'm struggling to see what ideas it brings beyond CodeAct (tool use is python) or the "task" tool in Claude code (spinning off sub-agents to preserve context).
nathanwh•3mo ago
This reminded me of ViperGPT[1] from a couple of years ago, which is similar but specific to vision language models. Both of them have a root llm which given a query produces a python program to decompose the query into separate steps, with the generated python program calling a sub model. One difference is this model has a mutable environment in the notebook, but I'm not sure how much of a meaningful difference that is.

[1] https://viper.cs.columbia.edu/static/viper_paper.pdf

UltraSane•3mo ago
Extending this so that the Root LLM can choose the best option from many other LLMs seems pretty powerful.
ttul•3mo ago
This is what Codex is doing. The LM has been trained to work well with the kinds of tools that a solid developer would use to navigate and search around a code repository and then to reason about what it finds. It’s also really competent at breaking down a task into steps. But I think the real magic - watching this thing for at least 40 of the last 50 working hours - is how it uses command line tools to dig through code quickly and accurately.

It’s not relying on the LM context much. You can generally code away for an hour before you run out of context and have to run a compression step or just start fresh.

lukebechtel•3mo ago
this doesn't appear to bring anything new to the table.

please correct me if I'm wrong..this is just subagent architecture?

sophia_james•3mo ago
I’m not sure if I understood this correctly:

1.Recursion is used to break down the large context and dispatch to different LLM calls to get the useful context.

2.This may lead to longer test-time execution on large contexts (even with parallelism in deep recursion), and the monetary cost may increase rapidly.

I think it’s a different idea from using RAG or manually maintaining a context window

correct me if I'm wrong

pontusrehula•3mo ago
If you would setup an RLM, would you set a higher temperature for the root LLM calls and a lower temperature for LLM calls deeper in the recursion?
patcon•3mo ago
Just wanted to say that I really like this question. Very thought-provoking :)

EDIT: makes me think of many computation systems in various substrates, and how they work. Focus vs distraction/creativity. ADHD workers in hierarchies of capitalism, purpose of breadth vs depth of exploration at various levels of the stack, who's at the "top" and why, etc etc

Weaver_zhu•3mo ago
IMO the author is a little over-claiming this work by naming 'recursive'. Quote from this blog:

> Lastly, in our experiments we only consider a recursive depth of 1 — i.e. the root LM can only call LMs, not other RLMs.

> but we felt that for most modern “long context” benchmarks, a recursive depth of 1 was sufficient to handle most problems.

I don't think a size 2 call stack algorithm should be regarded as 'recursive'.

vrighter•3mo ago
a model calls another (not self) model, which in turn returns without calling anything else. What you'ce discovered is called a function call.

It simply hopes two drunks are more coherent than one drunk.