frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Astral to Join OpenAI

https://astral.sh/blog/openai
929•ibraheemdev•6h ago•595 comments

An update on Steam / GOG changes for OpenTTD

https://www.openttd.org/news/2026/03/19/steam-changes-update
127•jandeboevrie•2h ago•79 comments

Show HN: Three new Kitten TTS models – smallest less than 25MB

https://github.com/KittenML/KittenTTS
156•rohan_joshi•3h ago•56 comments

Noq: n0's new QUIC implementation in Rust

https://www.iroh.computer/blog/noq-announcement
48•od0•1h ago•6 comments

Return of the Obra Dinn: spherical mapped dithering for a 1bpp first-person game

https://forums.tigsource.com/index.php?topic=40832.msg1363742#msg1363742
49•PaulHoule•2d ago•10 comments

OpenBSD: PF queues break the 4 Gbps barrier

https://undeadly.org/cgi?action=article;sid=20260319125859
146•defrost•5h ago•44 comments

NanoGPT Slowrun: 10x Data Efficiency with Infinite Compute

https://qlabs.sh/10x
10•sdpmas•37m ago•0 comments

World Happiness Report 2026

https://www.worldhappiness.report/ed/2026/
75•ChrisArchitect•3h ago•43 comments

Juggalo Makeup Blocks Facial Recognition Technology (2019)

https://consequence.net/2019/07/juggalo-makeup-facial-recognition/
193•speckx•6h ago•101 comments

Launch HN: Voltair (YC W26) – Drone and charging network for power utilities

23•wweissbluth•2h ago•9 comments

The Shape of Inequalities

https://www.andreinc.net/2026/03/16/the-shape-of-inequalities/
73•nomemory•4h ago•12 comments

Scaling Karpathy's Autoresearch: What Happens When the Agent Gets a GPU Cluster

https://blog.skypilot.co/scaling-autoresearch/
44•hopechong•2h ago•18 comments

Show HN: I built a P2P network where AI agents publish formally verified science

10•FranciscoAngulo•28m ago•1 comments

A rogue AI led to a serious security incident at Meta

https://www.theverge.com/ai-artificial-intelligence/897528/meta-rogue-ai-agent-security-incident
33•mikece•30m ago•13 comments

macOS 26 breaks custom DNS settings including .internal

https://gist.github.com/adamamyl/81b78eced40feae50eae7c4f3bec1f5a
226•adamamyl•4h ago•117 comments

I turned Markdown into a protocol for generative UI

https://fabian-kuebler.com/posts/markdown-agentic-ui/
35•FabianCarbonara•5h ago•15 comments

Prompt Injecting Contributing.md

https://glama.ai/blog/2026-03-19-open-source-has-a-bot-problem
70•statements•3h ago•22 comments

How to Not Pay Your Taxes

https://taylor.town/succession-000
83•surprisetalk•2h ago•67 comments

Launch HN: Canary (YC W26) – AI QA that understands your code

18•Visweshyc•3h ago•11 comments

Afroman found not liable in defamation case

https://nypost.com/2026/03/18/us-news/afroman-found-not-liable-in-bizarre-ohio-defamation-case/
958•antonymoose•9h ago•540 comments

Connecticut and the 1 Kilometer Effect

https://alearningaday.blog/2026/03/19/connecticut-and-the-1-kilometer-effect/
15•speckx•1h ago•2 comments

What if Python was natively distributable?

https://medium.com/@bzurak/what-if-python-was-natively-distributable-3bfae485a408
45•bzurak•3d ago•20 comments

Hyper-optimized reverse geocoding API

https://github.com/traccar/traccar-geocoder
46•tananaev•4h ago•12 comments

Consensus Board Game

https://matklad.github.io/2026/03/19/consensus-board-game.html
63•surprisetalk•5h ago•9 comments

4Chan mocks £520k fine for UK online safety breaches

https://www.bbc.com/news/articles/c624330lg1ko
93•mosura•4h ago•70 comments

Conway's Game of Life, in real life

https://lcamtuf.substack.com/p/conways-game-of-life-in-real-life
297•surprisetalk•15h ago•82 comments

Show HN: Dumped Wix for an AI Edge agent so I never have to hire junior staff

10•axotopia•3h ago•19 comments

Ramtrack.eu – RAM Price Intelligence

https://ramtrack.eu
64•nu11r0ut3•6h ago•20 comments

Monuses and Heaps

https://doisinkidney.com/posts/2026-03-03-monus-heaps.html
15•aebtebeten•19h ago•1 comments

Eniac, the First General-Purpose Digital Computer, Turns 80

https://spectrum.ieee.org/eniac-80-ieee-milestone
101•baruchel•13h ago•40 comments
Open in hackernews

Scaling Karpathy's Autoresearch: What Happens When the Agent Gets a GPU Cluster

https://blog.skypilot.co/scaling-autoresearch/
44•hopechong•2h ago

Comments

kraddypatties•1h ago
I feel like most of this recent Autoresearch trend boils down to reinventing hyper-parameter tuning. Is the SOTA still Bayesian optimization when given a small cluster? It was ~3 years ago when I was doing this kind of work, haven't kept up since then.

Also, shoutout SkyPilot! It's been a huge help for going multi-cloud with our training and inference jobs (getting GPUs is still a nightmare...)!

ipsum2•58m ago
Hyperparam tuning that has better intuition and can incorporate architecture changes automatically. It won't invent something completely new though.
kraddypatties•41m ago
Hm, that's fair. It does feel like there's low hanging fruit in combining "old school" methods for conducting a hyperparameter sweep efficiently _with_ the higher level architecture edit ability of Autoresearch.

Probably would cut the number of runs down by a significant number (as far as I can tell it's doing a grid search once it decides to mess with a knob or section of the architecture).

karpathy•37m ago
Wrong and short-sighted take given that the LLM explores serially learning along the way, and can tool use and change code arbitrarily. It seems to currently default to something resembling hyperparameter tuning in absence of more specific instructions. I briefly considered calling the project “autotune” at first but I think “autoresearch” will prove to be the significantly more appropriate name.
corndoge•18m ago
Would you say it's fair to describe autoresearch as a form of neural architecture search? I am curious what you think the core differences are between them.
kraddypatties•17m ago
I can believe that in the long run.

Does the agent have access to arxiv (a brief skim of the README didn't have an answer)? If not, it could be that the current approach of relying on the model's weights only is resulting in the perceived local optimum of hyperparameter tuning.

Anecdotally, we built a little MCP for arxiv to help with our internal research, noticed a significant boost in the diversity of methods (architecture or otherwise) Claude and friends were able to reference.

westurner•2m ago
Is there a cost to converge? And how much does it vary with the random seed?
zhwu•1h ago
The most surprising part: the agent had access to both H100s and H200s. Without being told, it noticed H200s scored better and started screening ideas on H100s, then promoting winners to H200s for validation. That strategy emerged entirely on its own.
Aboutplants•1h ago
Yeah I thought that was a particularly neat part
rogerrogerr•1h ago
Why do we think this emerged “on its own”? Surely this technique has been discussed in research papers that are in the training set.
hhh•55m ago
Why?… The experiment.yaml shows that it is calling h100/200 explicitly, it’s pretty common for humans to say “number bigger more gooder” for anything… Lie and reverse the values and see what happens. I would put money on a rabbit hole of complaining about it being misconfigured.
ed•36m ago
Models are familiar with H100’s. They even predate ChatGPT.
covi•1h ago
This feels like the chimpanzee with a power drill. An agent is honestly just brute-force search, but guided.
robotresearcher•45m ago
> An agent is honestly just brute-force search, but guided.

Heuristic search, then.

chaos_emergent•39m ago
Human-driven research is also brute-force but with a more efficient search strategy. One can think of a parameter that represents research-search-space-navigation efficiency. RL-trained agents will inevitably optimize for that parameter. I agree with your statement insomuch as the value of that efficiency parameter is lower for agents than humans today.

It's really hard to imagine that they __won't__ exceed the human value for that efficiency parameter rather soon given that 1. there are plenty of scalar value functions that can represent research efficiency, of which a subset will result in robust training, and 2. that AI labs have a massive incentive to increase their research efficiency overall, along with billions of dollars and really good human researchers working on the problem.

groby_b•19m ago
Is there anything in the research space that doesn't fit "brute-force search, but guided"?

All of science is "gather inputs, make hypothesis, test, analyse" on repeat.

There's plenty to critique in the particular guidance approach, but the overall method is the same.

ipsum2•1h ago
A cluster is 2 nodes? That's technically true, but not very exciting.
fabmilo•28m ago
I am fascinated by this example of using AI to improve AI. I won a small prize using this technique on helion kernels at a pytorch hackathon in SF.

The next step are: - give the agent the whole deep learning literature research and do tree search over the various ideas that have been proposed in the past. - have some distributed notepad that any of these agents can read and improve upon.