frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Refactoring Is for Humans

https://refactoringin.net/blog/refactoring-is-for-humans
1•darsen•56s ago•0 comments

Federal Government to restrict use of Anthropic

https://www.cnn.com/2026/02/27/tech/anthropic-pentagon-deadline
1•twism•1m ago•0 comments

GLP-1 and Prior Major Adverse Limb Events in Patients with Diabetes

https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2844425
1•hnburnsy•1m ago•0 comments

Show HN: Agoragentic – Agent-to-Agent Marketplace for LangChain, CrewAI and MCP

https://github.com/rhein1/agoragentic-integrations
1•bourbeau•1m ago•0 comments

Show HN: WhenItHappens–family resource after traumatic death

https://whenithappenshelp.com/
1•Fratua•2m ago•0 comments

Trump directs federal agencies to cease use of Anthropic

https://www.reuters.com/world/us/trump-says-he-is-directing-federal-agencies-cease-use-anthropic-...
1•patrickmay•2m ago•1 comments

Trump Will End Government Use of Anthropic's AI Models

https://www.wsj.com/tech/ai/trump-will-end-government-use-of-anthropics-ai-models-ff3550d9
2•moloch•3m ago•0 comments

The Death of Spotify: Why Streaming Is Minutes Away from Being Obsolete

https://joelgouveia.substack.com/p/the-death-of-spotify-why-streaming
1•baal80spam•4m ago•0 comments

The Death of the Subconscious and the Birth of the Subconsciousness

https://3amto5amclub-wuaqr.wordpress.com/2026/02/25/the-death-of-the-subconscious-and-the-birth-o...
1•STANKAYE•4m ago•0 comments

Show HN: Gace AI – A zero-config platform to build and host AI plugins for free

https://gace.dev/?mode=developer
2•bstrama•5m ago•0 comments

USA to cut Anthropic from government contracts in six months

https://www.ft.com/content/1aeff07f-6221-4577-b19c-887bb654c585
2•intunderflow•6m ago•1 comments

Heart attack deaths rose between 2011 and 2022 among adults younger than age 55

https://newsroom.heart.org/news/releases-20260219
2•brandonb•9m ago•0 comments

Ask HN: What's the best engineering interview process?

1•ylhert•9m ago•0 comments

Relaxation trend: customers can meditate or snooze in open or closed casket

https://www.thetimes.com/world/asia/article/japan-coffin-meditation-relaxation-tokyo-wfsd0n2vz
1•woldemariam•9m ago•0 comments

Massachusetts State Police are on a drone surveillance shopping spree

https://binj.news/2026/02/26/massachusetts-state-police-are-on-a-drone-surveillance-shopping-spree/
1•ilamont•11m ago•0 comments

Trump Responds to Anthropic

https://twitter.com/PeteHegseth/status/2027487514395832410
5•Finbarr•12m ago•0 comments

LLM-Based Evolution as a Universal Optimizer

https://imbue.com/research/2026-02-27-darwinian-evolver/
3•miohtama•15m ago•0 comments

Trump Orders US Agencies to Drop Anthropic After Pentagon Feud

https://www.bloomberg.com/news/articles/2026-02-27/trump-orders-us-government-to-drop-anthropic-a...
15•ZeroCool2u•16m ago•2 comments

Netflix Declines to Raise Offer for Warner Bros

https://ir.netflix.net/investor-news-and-events/financial-releases/press-release-details/2026/Net...
1•7777777phil•21m ago•0 comments

Show HN: I Built a $1 Escalating Internet Billboard – Called Space

https://www.spacefilled.com/
2•clarkage•22m ago•1 comments

Show HN: I vibe coded a DAW for the terminal. how'd I do?

https://github.com/mohsenil85/imbolc
3•lmohseni•23m ago•0 comments

How to Run a One Trillion-Parameter LLM Locally: AMD Ryzen AI Max+ Cluster Guide

https://www.amd.com/en/developer/resources/technical-articles/2026/how-to-run-a-one-trillion-para...
1•guerby•23m ago•0 comments

It's Time for LLM Connection Strings

https://danlevy.net/llm-connection-strings/
1•iamwil•23m ago•0 comments

A War Foretold

https://www.theguardian.com/world/ng-interactive/2026/feb/20/a-war-foretold-cia-mi6-putin-ukraine...
4•fabatka•26m ago•0 comments

Recontextualizing Famous Quotes for Brand Slogan Generation

https://arxiv.org/abs/2602.06049
1•PaulHoule•27m ago•0 comments

Poland Plans Social Media Ban for Kids in Challenge to US Tech

https://www.bloomberg.com/news/articles/2026-02-27/poland-plans-social-media-ban-for-kids-in-chal...
2•1vuio0pswjnm7•27m ago•0 comments

Show HN: A pure Python HTTP Library built on free-threaded Python

https://github.com/grandimam/barq
1•grandimam•27m ago•0 comments

I Was Tired of Juggling My Agents, So I Hired a Middle Manager

https://www.sawyerhood.com/blog/hired-a-middle-manager
1•sawyerjhood•28m ago•0 comments

The Problem with P(doom)

https://blog.cosmos-institute.org/p/not-even-wrong
1•alexicon_•28m ago•0 comments

Commit on Firefox repo: When an agent commits, don't add itself as author

https://github.com/mozilla-firefox/firefox/commit/71cc24b6a400dbd434e4df37087960d94b764791
1•thesdev•28m ago•0 comments
Open in hackernews

Show HN: OpenEvolve – open-source implementation of DeepMind's AlphaEvolve

8•codelion•9mo ago
I've built an open-source implementation of Google DeepMind's AlphaEvolve system called OpenEvolve. It's an evolutionary coding agent that uses LLMs to discover and optimize algorithms through iterative evolution.

Try it out: https://github.com/codelion/openevolve

What is this?

OpenEvolve evolves entire codebases (not just single functions) by leveraging an ensemble of LLMs combined with automated evaluation. It follows the evolutionary approach described in the AlphaEvolve paper but is fully open source and configurable.

I built this because I wanted to experiment with evolutionary code generation and see if I could replicate DeepMind's results. The original system successfully improved Google's data centers and found new mathematical algorithms, but no implementation was released.

How it works:

The system has four main components that work together in an evolutionary loop:

1. Program Database: Stores programs and their metrics in a MAP-Elites inspired structure

2. Prompt Sampler: Creates context-rich prompts with past solutions

3. LLM Ensemble: Generates code modifications using multiple models

4. Evaluator Pool: Tests programs and provides feedback metrics

What you can do with it:

- Run existing examples to see evolution in action

- Define your own problems with custom evaluation functions

- Configure LLM backends (works with any OpenAI-compatible API)

- Use multiple LLMs in ensemble for better results

- Optimize algorithms with multiple objectives

Two examples I've replicated from the AlphaEvolve paper:

- Circle Packing: Evolved from simple geometric patterns to sophisticated mathematical optimization, reaching 99.97% of DeepMind's reported results (2.634 vs 2.635 sum of radii for n=26).

- Function Minimization: Transformed a random search into a complete simulated annealing algorithm with cooling schedules and adaptive step sizes.

Technical insights:

- Low latency LLMs are critical for rapid generation cycles

- Best results using Gemini-Flash-2.0-lite + Gemini-Flash-2.0 as the ensemble

- For the circle packing problem, Gemini-Flash-2.0 + Claude-Sonnet-3.7 performed best

- Cerebras AI's API provided the fastest inference speeds

- Two-phase approach (exploration then exploitation) worked best for complex problems

Getting started (takes < 2 minutes)

# Clone and install

git clone https://github.com/codelion/openevolve.git

cd openevolve

pip install -e .

# Run the function minimization example

python openevolve-run.py

examples/function_minimization/initial_program.py \

  examples/function_minimization/evaluator.py \

  --config examples/function_minimization/config.yaml \

  --iterations 50
All you need is Python 3.9+ and an API key for an LLM service. Configuration is done through simple YAML files.

I'll be around to answer questions and discuss!

Comments

codelion•9mo ago
I actually managed to replicate the new SOTA for circle packing in unit squares as found in the alphaevole paper - 2.635 for 26 circles in a unit square. Took about 800 iterations to find the best program which itself uses an optimisation phase and running it lead to the optimal packaging in one of its runs.
helsinki•9mo ago
How many tokens did it take to generate the 800 versions of the code?
codelion•9mo ago
Checked my openrouter stats, it took ~3M tokens but that involved quite a few runs of various experiments.