frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Ask HN: What do founders do to avoid getting blocked by competitor patents?

1•wasiurrehman•1m ago•0 comments

Model Adjacent – a new stab at products in the AI era

https://mercurialsolo.substack.com/p/model-adjacent
1•mercurialsolo•1m ago•0 comments

Finishing My ZX Spectrum Emulator with Gemini 3 Pro – Bitwrangler.uk

https://bitwrangler.uk/2025/12/29/finishing-my-zx-spectrum-emulator-with-gemini-3-pro/
1•pricechild•4m ago•0 comments

Show HN: A practical guide to building Solana USDC payments in React

https://ulomira.com/books/fast-low-fee-crypto-payments
1•fullstackragab•4m ago•0 comments

George H. Butler and the Limits of Being Right

https://secretaryrofdefenserock.substack.com/p/george-h-butler-and-the-limits-of
1•barry-cotter•5m ago•0 comments

Pushed by GenAI and Front End Upgrades, Ethernet Switching Hits New Highs

https://www.nextplatform.com/2026/01/08/pushed-by-genai-and-front-end-upgrades-ethernet-switching...
1•rbanffy•7m ago•0 comments

NASA orders "controlled medical evacuation" from the International Space Station

https://arstechnica.com/space/2026/01/in-a-first-nasa-orders-astronauts-home-after-unspecified-me...
2•rbanffy•9m ago•0 comments

Slopes in AABB Collision Systems

https://andreyor.st/posts/2026-01-09-slopes-in-aabb-collision-systems/
1•ibobev•11m ago•0 comments

Digging into the LLM-as-a-Judge Results

https://www.gilesthomas.com/2026/01/20260109-llm-from-scratch-30-digging-into-llm-as-a-judge
1•ibobev•12m ago•0 comments

EU considers designating WhatsApp as large platform

https://www.reuters.com/technology/eu-considers-designating-whatsapp-very-large-platform-spokespe...
1•giuliomagnifico•14m ago•0 comments

Claude Code sessions should be encrypted

https://yoav.blog/2026/01/09/claude-code-sessions-should-be-encrypted/
3•yoavfr•15m ago•1 comments

Dedicated vs. VPS for WordPress with a $200 budget

https://wpshell.com/lesson/benchmarks/
1•k7n•15m ago•0 comments

InvMon: A locally-installed desktop portfolio and investment tracking app

https://invmon.com/
1•tomtomstuder•18m ago•1 comments

US Debt Clock

https://www.usdebtclock.org/
1•Erikun•19m ago•0 comments

Relax for the Same Result (2015)

https://sive.rs/relax
2•birdculture•20m ago•0 comments

A tiny LM that does inference at compile time

https://github.com/erodola/bigram-metacpp
2•signa11•21m ago•0 comments

Astronaut's 'serious medical condition' forces NASA to end space mission early

https://www.bbc.com/news/articles/cd9e2y7nkv8o
2•Growtika•24m ago•0 comments

Show HN: iKrypt – send a secret once (the key never hits our server)

https://ikrypt.com
1•alphatesterguy•28m ago•0 comments

The Warhammer Capital of the World

https://dispatch-media.com/the-warhammer-capital-of-the-world-nottingham/
4•comradino123•29m ago•0 comments

The Theory That Gives Trump a Blank Check for Aggression

https://www.nytimes.com/2026/01/09/magazine/trump-venezuela-foreign-policy-realism-greenland.html
3•mitchbob•32m ago•1 comments

Prompts are (not) the new source code

https://quesma.com/blog/prompts-source-code/
2•stared•34m ago•1 comments

Ask HN: Is there a Zod validation library for Golang?

1•danver0•35m ago•1 comments

Peon – Prefixed Entry Object Notation

https://www.bartoszsypytkowski.com/peon/
2•yagizdegirmenci•36m ago•1 comments

Ask HN: What if the AI scaling plateau is just a "false dip"?

1•massicerro•42m ago•2 comments

Show HN: How thinking about death made things feel lighter

https://prtkagwl.substack.com/p/day-14-thinking-about-death
1•btwnplaces•42m ago•0 comments

Mark Jeff Dean vacation fact as true

https://github.com/LRitzdorf/TheJeffDeanFacts/commit/ba1bdf8d6a4697de2fa7d30c0e3011c53db091a2
1•yagizdegirmenci•43m ago•0 comments

Sieve: An Efficient Turn-Key Eviction Algorithm for Web Caches

https://cachemon.github.io/SIEVE-website/
2•tosh•43m ago•0 comments

How can I build a simple pulse generator to demonstrate transmission lines

https://electronics.stackexchange.com/questions/764155/how-can-i-build-a-simple-pulse-generator-t...
1•alphabetter•45m ago•0 comments

Why Most AI Incidents Are Evidence Failures, Not Model Failures

https://zenodo.org/records/18196751
1•businessmate•46m ago•1 comments

Show HN: Autonomous engineer teams for Claue Code.

https://github.com/covibes/zeroshot
1•covibes•47m ago•0 comments
Open in hackernews

Digital Red Queen: Adversarial Program Evolution in Core War with LLMs

https://sakana.ai/drq/
117•hardmaru•19h ago

Comments

hardmaru•19h ago
Hi HN,

I am one of the authors from Sakana AI and MIT. We just released this paper where we hooked up LLMs to the classic 1984 programming game Core War. For those who haven't played it, Core War involves writing assembly programs in a language called Redcode that battle for control of a virtual computer's memory. You win by crashing the opponent's process while keeping yours running. It is a Turing-complete environment where code and data share the same address space, which leads to some very chaotic self-modifying code dynamics.

We did not just ask the model to write winning code from scratch. Instead, we treated the LLM as a mutation operator within a quality-diversity algorithm called MAP-Elites. The system runs an adversarial evolutionary loop where new warriors are continually evolved to defeat the champions of all previous rounds. We call this Digital Red Queen because it mimics the biological hypothesis that species must continually adapt just to survive against changing competitors.

The most interesting result for us was observing convergent evolution. We ran independent experiments starting from completely different random seeds, yet the populations consistently gravitated toward similar behavioral phenotypes, specifically regarding memory coverage and thread spawning. It mirrors how biological species independently evolve similar traits like eyes to solve similar problems. We also found that this training loop produced generalist warriors that were robust even against human-written strategies they had never encountered during training.

We think Core War is an under-utilized sandbox for studying these kinds of adversarial dynamics. It lets us simulate how automated systems might eventually compete for computational resources in the real world, but in a totally isolated environment. The simulation code and the prompts we used are open source on GitHub.

Other info other than the blog link:

Paper (website): https://pub.sakana.ai/drq/

Arxiv: https://arxiv.org/abs/2601.03335

Code: https://github.com/SakanaAI/drq

NitpickLawyer•16h ago
> adversarial evolutionary loop where new warriors are continually evolved to defeat the champions of all previous rounds.

Interesting. So you're including past generation champions in the "fights"? That would intuitively model a different kind of evolution than just "current factors"-driven evolution.

> We also found that this training loop produced generalist warriors that were robust even against human-written strategies they had never encountered during training.

Nice. Curious, did you do any ablations for the "all previous champions" vs. "current gen champions"?

aldebaran1•14h ago
Very interesting paper, thank you. It makes me wonder what other game substrates could form the basis for adversarial/evolutionary strategy optimization for LLMs, and whether these observations replicate across games.

Since LLMs are text based, a text-based game might be interesting. Something like Nomic?

Or a "meme warfare" game where each agent tries to prompt-inject its adversaries into saying a forbidden codeword, and can modify its own system prompt to attempt to prevent that from happening to itself.

GuB-42•16h ago
Using evolution in the context of Core War is not a new idea by far, it is even referenced in the paper.

Examples here: https://corewar.co.uk/evolving.htm

The difference here is that instead of using a typical genetic algorithm written in a programming language, it uses LLM prompts to do the same thing.

I wonder if the authors tried some of the existing "evolvers" to compare to what the LLM gave out.

api•16h ago
See also:

https://en.wikipedia.org/wiki/Tierra_(computer_simulation)

https://avida-ed.msu.edu

https://github.com/adamierymenko/nanopond

Lots of evolving bug corewar-style systems around.

I think the interesting thing with this one is they're having LLMs create evolving agents instead of blind evolution or some similar ML system.

Ieghaehia9•14h ago
That in turn makes me wonder:

Given fixed opposition, finding a warrior that performs the best is an optimization problem. Maybe, for very small core sizes like a nano core, it would be possible to find the optimum directly by SAT or SMT instead of using evolution? Or would it be impractical even for those core sizes?

slickytail•11h ago
I think it would, for all practical purposes, be impossible to determine an optimal warrior, even at very small core sizes. Not only is the search space huge but the evaluation function can take unbounded time to resolve. We should consider the halting problem embedded inside the optimization target as a clue to the problem's difficulty.
dgacmu•10h ago
Oh man, that's funny to see one of my grad school class projects in that list. Takes me back. :-)

From that experience: The LLM is likely to do drastically better. Most of the prior work, mine included, took a genetic algorithm approach, but an LLM is more likely to make coherent multi-instruction modifications.

It's a shame they didn't compare against some of the standard core wars benchmarks as a way to facilitate comparisons to prior work, though. Makes it hard to say that they're better for sure. https://corewar.co.uk/bench.htm

jacquesm•10h ago
I'm not sure if that will hold up. The LLM is not going to do anything random and that is actually a powerful component that makes original output possible.
kyralis•7h ago
I wonder if a combination would be useful. Use an actual GA to do the mutation, and then let an LLM "fix" each mutated child.
jacquesm•2h ago
Could be. But the interesting thing is that all you can do here is optimize. Random chance is - like attention ;) - all you need.
pkhuong•13h ago
How does the output fare on competitive hills like https://sal.discontinuity.info/hill.php?key=94t ?

AFAIK, the best results so far for fully computer-generated warriors have been on the nano and tiny format (https://sal.discontinuity.info/hill.php?key=nano, https://sal.discontinuity.info/hill.php?key=tiny), with much shorter warriors (at most 5 or 20 instructions).

JKCalhoun•10h ago
What a lovely period of time that was—when "Computer Recreations" ran monthly in Scientific American. I read the column every month and was fascinated to learn about Eliza, Core Wars, Conway's Life, Wa-Tor, etc. It was a time when you coded simply for the fun of it—to explore, learn.

I know you can still do that today, but… something has changed. I don't know what it is. (Maybe I changed.)

Anyway, I was unable to track down PDF versions of the original articles, but, for the curious and newcomers to Core Wars, they're transcribed here:

https://corewar.co.uk/dewdney/

idiotsecant•6h ago
Computers are no longer something fresh and new. They are firmly in the realm of stuff that exists and has Rules. The frontier is dead.
rao-v•8h ago
The idea of what LLMs could do in CoreWars has been hanging around in the back of my head for a while now. So happy to see someone explore it systematically