frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

We don't need continual learning for AGI. What top labs are currently doing

4•kok14•3h ago
Many people think that we won't reach AGI or even ASI if LLM's don't have something called "continual learning". Basically, continual learning is the ability for an AI to learn on the job, update its neural weights in real-time, and get smarter without forgetting everything else (catastrophic forgetting). This is what we do everyday, without much effort.

What's interesting now, is if you look at what the top labs are doing, they’ve stopped trying to solve the underlying math of real-time weight updates. Instead, they’re simply brute-forcing it. It is exactly why, in the past ~ 3 months or so, there has been a step-function increase in how good the models have gotten.

Long story short, the gist of it is, if you combine:

very long context windows

reliable summarization

structured external documentation,

you can approximate a lot of what people mean by continual learning.

How it works is, the model does a task and absorbs a massive amount of situational detail. Then, before it “hands off” to the next instance of itself, it writes two things: short “memories” (always carried forward in the prompt/context) and long-form documentation (stored externally, retrieved only when needed). The next run starts with these notes, so it doesn't need to start from scratch.

Through this clever reinforcement learning (RL) loop, they train this behaviour directly, without any exotic new theory.

They treat memory-writing as an RL objective: after a run, have the model write memories/docs, then spin up new instances on the same, similar, and dissimilar tasks while feeding those memories back in. How this is done, is by scoring performance across the sequence, and applying an explicit penalty for memory length so you don’t get infinite “notes” that eventually blow the context window.

Over many iterations, you reward models that (a) write high-signal memories, (b) retrieve the right docs at the right time, and (c) edit/compress stale notes instead of mindlessly accumulating them.

This is pretty crazy. Because when you combine the current release cadence of frontier labs where each new model is trained and shipped after major post-training / scaling improvements, even if your deployed instance never updates its weights in real-time, it can still “get smarter” when the next version ships AND it can inherit all the accumulated memories/docs from its predecessor.

This is a new force multiplier, another scaling paradigm, and likely what the top labs are doing right now (source: TBA).

Ignoring any black swan level event (unknown, unknowns), you get a plausible 2026 trajectory:

We’re going to see more and more improvements, in an accelerated timeline. The top labs ARE, in effect, using continual learning (a really good approximation of it), and they are directly training this approximation, so it rapidly gets better and better.

Don't believe me? Look at what both OpenAi(https://openai.com/index/introducing-openai-frontier/) and Anthropic(https://resources.anthropic.com/2026-agentic-coding-trends-report) have mentioned as their core things they are focusing on. It's exactly why governments & corporations are bullish on this; there is no wall....

Comments

uaas•35m ago
Would a human end a long-winded text with…?

Aura-State: Formally Verified LLM State Machine Compiler

19•rohanmunshi08•3d ago•5 comments

We don't need continual learning for AGI. What top labs are currently doing

4•kok14•3h ago•1 comments

Ask HN: Who wants to be hired? (March 2026)

124•whoishiring•2d ago•367 comments

An offline map using OruxMaps(satellite,routing,3D terrain,GPS and POI)

2•supergoogler•4h ago•0 comments

Ask HN: Has anyone noticed the fear-driven prompt suggestions that GPT5.3 makes?

10•cedarscarlett•11h ago•6 comments

Ask HN: Who is hiring? (March 2026)

243•whoishiring•2d ago•357 comments

Tell HN: Digital Ocean has run out of GPU droplets

14•nathannaveen•20h ago•2 comments

Ask HN: If your project is free, what are you building and why keep it free?

8•LeanVibe•20h ago•17 comments

Stathat Is Shutting Down

6•jervant•13h ago•2 comments

Ask HN: Maintainers, do LLM-only users often clutter your issues/PRs?

7•lucrbvi•19h ago•9 comments

Ask HN: Anyone have experience making physical toys that you've sold?

2•general_reveal•14h ago•1 comments

Ask HN: How are you all staying sane?

146•throwaway53463•3d ago•150 comments

I lost my ability to learn anything new because of AI and I need your opinions

18•dokdev•1d ago•27 comments

Ask HN: What sources like HN do you consume?

54•DavidHaerer•3d ago•36 comments

Ask HN: What's your experience self-hosting in 2026?

25•rustcore•1d ago•9 comments

Tell HN: I got Claude Max for my open source project

9•asim•1d ago•5 comments

Ask HN: Why has ChatGPT disabled links to websites?

6•krschacht•22h ago•4 comments

Ask HN: What will OpenAI employees do now who have signed notdividedorg petition

17•Imustaskforhelp•1d ago•16 comments

Ask HN: What Online LLM / Chat do you use?

12•ddxv•2d ago•18 comments

Ask HN: How many of you hold an amateur radio license in your country?

3•ToddWBurgess•1d ago•3 comments

Ask HN: What prompt do you use to get Claude to consistently render LaTeX?

5•TobyTheCamel•1d ago•6 comments

Whats Up with Claude Lately?

19•mech422•2d ago•14 comments

Ask HN: Would engineers be interested in a technical prep consultant?

8•TechPrepper•2d ago•9 comments

Ask HN: How is Claude agent experience in Xcode 26.3?

8•malshe•1d ago•2 comments

Ask HN: What is the "Control Plane" for local AI agents?

7•denis4inet•1d ago•8 comments

Ask HN: Who still works async and has a 'no meetings' work policy in 2026?

7•RaulOnRails•2d ago•5 comments

Ask HN: How do you find contracting/freelance roles without recruiters nowadays?

3•Gooblebrai•16h ago•2 comments

Facebook Appears to Be Down

13•Molitor5901•1d ago•16 comments

Ask HN: How Do Emergency Alerts on Phone Work?

6•rishikeshs•2d ago•4 comments

Tell HN: GitHub Having Issues

52•Sytten•1d ago•27 comments