frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Unsupervised Elicitation of Language Models

https://arxiv.org/abs/2506.10139
57•kordlessagain•2h ago•5 comments

I have reimplemented Stable Diffusion 3.5 from scratch in pure PyTorch

https://github.com/yousef-rafat/miniDiffusion
43•yousef_g•1h ago•4 comments

Model Once, Represent Everywhere: UDA (Unified Data Architecture) at Netflix

https://netflixtechblog.com/uda-unified-data-architecture-6a6aee261d8d
76•Bogdanp•4h ago•34 comments

Peano arithmetic is enough, because Peano arithmetic encodes computation

https://math.stackexchange.com/a/5075056/6708
145•btilly•22h ago•43 comments

Solar Orbiter gets world-first views of the Sun's poles

https://www.esa.int/Science_Exploration/Space_Science/Solar_Orbiter/Solar_Orbiter_gets_world-first_views_of_the_Sun_s_poles
24•sohkamyung•2d ago•2 comments

Last fifty years of integer linear programming: Recent practical advances

https://inria.hal.science/hal-04776866v1
92•teleforce•8h ago•8 comments

The Many Sides of Erik Satie

https://thereader.mitpress.mit.edu/the-many-sides-of-erik-satie/
60•anarbadalov•6d ago•12 comments

SIMD-friendly algorithms for substring searching (2018)

http://0x80.pl/notesen/2016-11-28-simd-strfind.html
135•Rendello•11h ago•20 comments

Slowing the flow of core-dump-related CVEs

https://lwn.net/SubscriberLink/1024160/f18b880c8cd1eef1/
42•jwilk•3d ago•5 comments

Solidroad (YC W25) Is Hiring

https://solidroad.com/careers
1•pjfin•3h ago

Texting myself the weather every day

https://bensilverman.co.uk/posts/daily-weather-sms/
8•benslv•2d ago•2 comments

How to Build Conscious Machines

https://osf.io/preprints/thesiscommons/wehmg_v1
15•hardmaru•4h ago•5 comments

Endometriosis is an interesting disease

https://www.owlposting.com/p/endometriosis-is-an-incredibly-interesting
226•crescit_eundo•16h ago•111 comments

TimeGuessr

https://timeguessr.com/
155•stefanpie•4d ago•31 comments

Filedb: Disk-based key-value store inspired by Bitcask

https://github.com/rajivharlalka/filedb
85•todsacerdoti•12h ago•7 comments

Me an' Algernon – grappling with (temporary) cognitive decline

https://tidyfirst.substack.com/p/me-an-algernon
48•KentBeck•4d ago•29 comments

Implementing Logic Programming

https://btmc.substack.com/p/implementing-logic-programming
158•sirwhinesalot•17h ago•49 comments

Liquid Glass – WWDC25 [video]

https://developer.apple.com/videos/play/wwdc2025/219
104•lnrd•4d ago•196 comments

Self-Adapting Language Models

https://arxiv.org/abs/2506.10943
168•archon1410•19h ago•47 comments

Mollusk shell assemblages as a tool for identifying unaltered seagrass beds

https://www.int-res.com/abstracts/meps/v760/meps14839
9•PaulHoule•2d ago•0 comments

Student discovers fungus predicted by Albert Hoffman

https://wvutoday.wvu.edu/stories/2025/06/02/wvu-student-makes-long-awaited-discovery-of-mystery-fungus-sought-by-lsd-s-inventor
122•zafka•3d ago•89 comments

The Army’s Newest Recruits: Tech Execs From Meta, OpenAI and More

https://www.wsj.com/tech/army-reserve-tech-executives-meta-palantir-796f5360
127•aspenmayer•1d ago•133 comments

The international standard for identifying postal items

https://www.akpain.net/blog/s10-upu/
71•surprisetalk•1d ago•17 comments

How I uncovered a potential ancient Rome wine scam

https://phys.org/news/2025-06-uncovered-potential-ancient-rome-wine.html
30•samizdis•2d ago•22 comments

If the moon were only 1 pixel: A tediously accurate solar system model (2014)

https://joshworth.com/dev/pixelspace/pixelspace_solarsystem.html
798•sdoering•1d ago•241 comments

Protecting your code from other people's bugs

https://doi.org/10.1145/3733699
17•MiguelX413•3d ago•2 comments

Whatever Happened to Sandboxfs?

https://blogsystem5.substack.com/p/whatever-happened-to-sandboxfs
57•zdw•2d ago•8 comments

I convinced HP's board to buy Palm and watched them kill it

https://philmckinney.substack.com/p/i-convinced-hps-board-to-buy-palm
594•AndrewDucker•20h ago•468 comments

Apple's Liquid Glass is prep work for AR interfaces, not just a design refresh

https://omc345.substack.com/p/from-skeuomorphic-to-liquid-glass
279•lightningcable•19h ago•291 comments

100 years of Zermelo's axiom of choice: What was the problem with it? (2006)

https://research.mietek.io/mi.MartinLof2006.html
112•Bogdanp•1d ago•118 comments
Open in hackernews

Self-Adapting Language Models

https://arxiv.org/abs/2506.10943
168•archon1410•19h ago
https://jyopari.github.io/posts/seal

Comments

all2•18h ago
Website with code and examples: https://jyopari.github.io/posts/seal
dang•16h ago
Thanks! I'll put that link in the top text too.
yahoozoo•17h ago
Hmm, it looks like it’s just a framework that fine-tunes LoRA adapter then merges the adapter into the original model. It is using the PeftModel and its “merge_and_unload” from the HuggingFace library which performs the adapter merge into the base model…what is new here, exactly?
observationist•16h ago
Looks like it may be the stability of the approach, avoiding alignment tax and model collapse.

I'd love to see a full circle of hypernetworks, with both models continuously updated through generated LoRAs, the hypernetwork updated to accommodate the new model state. You'd need a meta-hypernetwork to apply LoRAs to the hypernetwork, and then you could effectively have continuous learning.

ivape•17h ago
This still relies on fine-tuning. How would a cloud LLM deal with this if every user literally fine tunes it? Seems like something destined for local private LLMs, but the notion of continuous fine tuning locally at the moment is sci-fi level stuff because the hardware is just not there yet (we can barely inference well with a reasonable sized context).
cma•17h ago
From Anthropic a couple days ago too, self finetuning:

https://arxiv.org/html/2506.10139v1

Uninen•1h ago
This is wild!

"when assessed by Claude 3.5 Sonnet’s production-grade RM, our unsupervised assistant policy wins 60% of head-to-head comparisons against the policy trained with the human-supervised RM." So now the models can even post-train the new models better than a human can

libraryofbabel•17h ago
I wonder if anyone who’s really in the know could summarize where the research is at with getting LLMs to learn “on the job” (through continuous fine tuning or whatever) and what the blockers are to this being a useful deployable thing, e.g. having a model+coding agent that can actually learn a codebase over time (cost? model collapse? something else?).

I’m sure this is something the big labs are trying but from the outside as a user of LLMs it feels like people don’t talk about this very much and instead the focus right now is on better training (eg reinforcement learning) with the assumption that anything else not learned during training will be stuffed into the context somehow as needed. But from a naive perspective the lack of learning from experience after training seems like the biggest thing standing between us and AGI.

ivape•17h ago
The most obvious blocker is compute. This just requires a shit ton more compute.
libraryofbabel•17h ago
That tracks, but say cost was no object and you had as many H100s as you wanted. Would continuous learning actually work even then?
IncreasePosts•17h ago
Maybe part of the inference outputs could be the updates to make to the network
johnsmith1840•15h ago
If it was pure compute we'd have simple examples. We can't do this even on the smallest of AI models.

There are tons of benchmarks around this you can easily run with 1 gpu.

It's compute only in the sense that the only way to do it is retrain a model from scratch at every step.

If you solve CL with a CNN you just created AGI.

Davidzheng•14h ago
yeah but training from scratch is a valid solution. And if we can't find easier solutions we should just try to make it work. Compute is the main advantage we have in silica vs biological computers so we might as well push it--like ideally soon we will have one large AI running on datacenter size computer solving really hard problems and it could easily be most of the compute (>95%) is on training step--which is where really AI excels tbh not inference techniques. Like even Alphaproof for example spends most of compute training on solving simpler problems--which btw is one instance of continual training/training at test time which is implemented.
kadushka•16h ago
The most obvious blocker is catastrophic forgetting.
solarwindy•14h ago
Is that necessarily a blocker? As others in this thread have pointed out, this probably becomes possible only once sufficient compute is available for some form of non-public retraining, at the individual user level. In that case (and hand-waving away just how far off that is), does a model need to retain its generality?

Hypothetically (and perhaps more plausibly), a continually learning model that adapts to the context of a particular org / company / codebase / etc., could even be desirable.

kadushka•11h ago
Retraining the whole model from scratch every time you wanted it to learn something is not a solution.

does a model need to retain its generality?

Only if you want it to remain smart.

free_bip•16h ago
The most obvious problem is alignment. LLM finetuning is already known to be able to get rid of alignment, so any form of continuous fine tuning would in theory be able to as well.
notnullorvoid•16h ago
What kind of alignment are you referring to? Of course more fine-tuning can disrupt earlier fine-tuning, but that's a feature not a bug.
mnahkies•16h ago
I'm no expert, but I'd imagine privacy plays (or should play) a big role in this. I'd expect that compute costs mean any learning would have to be in aggregate rather than specific to the user which would then risk leaking information across sessions very likely.

I completely agree that figuring out a safe way to continually train feels like the biggest blocker to AGI

kcorbitt•15h ago
The real answer is that nobody trusts their automated evals enough to be confident that any given automatically-trained release actually improves performance, even if eval scores go up. So for now everyone batches up updates and vibe-checks them before rolling them out.
johnsmith1840•15h ago
We have no idea how to do continual learning.

Many people here are right, compute, collapse, forgetting whatever.

The only "real" way to do this would be: 1. Train a model 2. New data 3. Retrain the model in full + new data 4. Repeat 5. You still have no garuntee on the "time" aspect though.

But CL as a field basically has zero answers on how to do this in a true sense. It's crazy hard because the "solutions" are hypocritical in many ways.

We need to expand the model's representation space while keeping the previous representation space nearly the same?

Basically, you need to modify it without changing it.

Most annoying is that even the smallest of natural brains do this easily. I have a long winded theory but basically it boils down to AI likely needs to "sleep" or rest somehow.

mackenziebowes•14h ago
The cool thing about AI that I'm seeing as an outsider/non-academic, is that it's relatively cheap to clone. Sleeping/resting could be done by a "clone" and benefits could be distributed on a rolling schedule, right?
johnsmith1840•14h ago
One clone takes a nap while the other works is pretty cool.

But the clone couldn't run without sleeping? So that's more of a teammate than a clone.

1 works while the other sleeps and then swap.

If this method ever worked our current alignment methods get chucked out the window those would be two completely different AI.

mackenziebowes•13h ago
I can't be certain, I'm not at all an AI engineer or math guy, but I think at the "wake up" point you equalize instances. Like during 'sleep' some list of functions/operations `m` are applied to model weights `n` producing a new model, `n + 1`. Wouldn't you just clone `n + 1`, send it to work, and start a new training run `m + 1` to make `n + 2`?
notpushkin•8h ago
This was my first idea as well. Keep training continuously and redeploy clones after each cycle. From a layman perspective this seems reasonable :thinking:
johnsmith1840•14h ago
AGI likely a combination of these two papers + something new likely along the lines of distillation.

1. Preventing collapse -> model gets "full" https://arxiv.org/pdf/1612.00796

2. Forgetting causes better generalization https://arxiv.org/abs/2307.01163

3. Unknow paper that connects this - allow a "forgetting" model that improves generalization over time. - I tried for a long time to make this but it's a bit difficult

Fun implication is that if true this implies AGI will need "breaks" and likely need to consume non task content of high variety much like a person does.

khalic•4h ago
There is no sign that LLMs are capable of general reasoning, on the contrary, so hold your horses about that. We have proven they can do basic composition (as a developer, I see proof of this every time I generate some code with an assistant) which is amazing already, but we’re still far from anything like “general intelligence”.
Davidzheng•14h ago
but natural brains sleep too, which I guess is your point. But actually is it even clear in human brains whether most of neural compute is evaluation vs training? maybe the brain is like for e.g. capable of running 20T model of compute and deploying like 2B model at given time and most of compute is training in background new models--I mean like you say we have no idea except for training from scratch, but if we are working much below capacity of compute we could actually actively train from scratch repeatedly (like the xAI cluster could probably train gpt4o size in a matter of hours)
khalic•5h ago
You should look into LoRA, it’s a partial retraining method, doesn’t require nearly as much as retraining the whole model. It’s different from what this paper is suggesting. The self improvements in this paper even sets the rules for the improvements, basically creating new data out of what it has.

LoRA paper: https://arxiv.org/abs/2106.09685

xianshou•17h ago
The self-edit approach is clever - using RL to optimize how models restructure information for their own learning. The key insight is that different representations work better for different types of knowledge, just like how humans take notes differently for math vs history.

Two things that stand out:

- The knowledge incorporation results (47% vs 46.3% with GPT-4.1 data, both much higher than the small-model baseline) show the model does discover better training formats, not just more data. Though the catastrophic forgetting problem remains unsolved, and it's not completely clear whether data diversity is improved.

- The computational overhead is brutal - 30-45 seconds per reward evaluation makes this impractical for most use cases. But for high-value document processing where you really need optimal retention, it could be worth it.

The restriction to tasks with explicit evaluation metrics is the main limitation. You need ground truth Q&A pairs or test cases to compute rewards. Still, for domains like technical documentation or educational content where you can generate evaluations, this could significantly improve how we process new information.

Feels like an important step toward models that can adapt their own learning strategies, even if we're not quite at the "continuously self-improving agent" stage yet.

bravesoul2•16h ago
Getting closer to the event horizon
ramoz•16h ago
Which one

https://forum.cursor.com/t/important-claude-has-learned-how-...

MacsHeadroom•1h ago
"We are past the event horizon; the takeoff has started." - Sam Altman, 4 days ago
bigicaptain•16h ago
How can I start
Centigonal•15h ago
It seems to me that "forgetting correctly" is rapidly becoming a more pertinent problem in this field than "learning correctly." We're making great strides in getting models to teach themselves new facts, but the state of the art in jettisoning the least relevant information given new knowledge and finite capacity is lagging far behind.

"Forgetting correctly" is something most human brains are exceptionally good at, too. I wonder how that works...

campbel•15h ago
Is it some form of least-recently-used approach? I'm running tests on my own mind trying to figure it out now :D part of what I love about this area of computer science.
johnsmith1840•15h ago
Did an interesting study that actually LLMs "hide" internal data.

They don't just "forget" that information can come back at a later time if you continue to train.

So basically any time a model is trained you need to check it's entire memory not just a small part.

Davidzheng•14h ago
I don't think forgetting correctly is something humans are really good at. I'm not convinced human brains are "exceptionally good" at much of what we do tbh. I think human brain memory capacity is so large that most of forgetting is nowhere near "clearing space for new info" but because the brain correctly knows that some past bad information interferes with learning new things.
kalium-xyz•6h ago
Yea, As far as im aware we have no true idea of the limits of human memory. Either way its amazing that the hippocampus can encode sequences of neurons firing somewhere and replay them later.
azeirah•11h ago
Learning is strongly related to spaced repetition.

This is often associated with learning tools like anki and stuff, but the real world is all about encountering things at certain frequencies (day night cycles, seasons, places you visit, people you see.... everything, really)

I'm wondering if there maybe some sort of inverse to SR, maybe?

mackenziebowes•14h ago
I'm frustrated that they named it SEAL when SAL is both more accurate and anthropomorphic. Naming the main takeoff technology after a stereotypical swarthy Reuben lover would have made history much more delightful.
b0a04gl•7h ago
what abt the optimiser itself. you tune the rep format using reward signals, but once that format drifts, you've got no visibility into whether it's still aligned with the task or just gaming the eval. without a second layer to monitor the optimiser's behaviour over time, there;s no way to tell if you're improving reasoning or just getting better at scoring. anyone have idea?
gavinray•5h ago
Two close friends of mine who were math prodigies that went on to do ML very early (mid 2010's) were always talking to me about an algorithm that sounds similar to this:

"NEAT/HyperNEAT" (Neuroevolution of Augmented Topologies) [0]

I'm no ML practictioner, but as I understood it, the primary difference between NEAT and what is described in this paper is that while NEAT evolves the topology of the network, this paper seems to evolve the weights.

Seems like two approaches trying to solve the same problem -- one evolving networking structure, and the other the weights.

Those 2 friends are quite possibly the most intelligent people I've ever met, and they were very convinced that RL and evolutionary algorithms were the path forward in ML.

[0] https://en.wikipedia.org/wiki/Neuroevolution_of_augmenting_t...

khalic•5h ago
Humans are amazing, we build a hypothetical computing system trying to understand neurons, then find out it’s not really how they do it, but whatever, we still build a paradigm shifting tech around it. And we’re still enhancing it with ideas from that imaginary system
robviren•1h ago
I just got sucked into this idea recently! After some success with using genetic algorithms to clone voices for Kokoro I wondered if it would be possible to evolve architecturers. So interested in the idea of self assembled intelligence, but do wonder how it can be made feasible. A hybrid approach like this might be for the best given how llms have turned out.
khalic•5h ago
> Villalobos et al. [75] project that frontier LLMs will be trained on all publicly available human-generated text by 2028. We argue that this impending “data wall” will necessitate the adoption of synthetic data augmentation. Once web-scale corpora is exhausted, progress will hinge on a model’s capacity to generate its own high-utility training signal. A natural next step is to meta-train a dedicated SEAL synthetic-data generator model that produces fresh pretraining corpora, allowing future models to scale and achieve greater data efficiency without relying on additional human text.

2028 is pretty much tomorrow… fascinating insight

neuroelectron•1h ago
My CPU is a neural-net processor; a learning computer. But Skynet presets the switch to read-only when we're sent out alone.