frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

I ruined my vacation by reverse engineering WSC

https://blog.es3n1n.eu/posts/how-i-ruined-my-vacation/
56•todsacerdoti•2h ago•9 comments

Plain Vanilla Web

https://plainvanillaweb.com/index.html
913•andrewrn•14h ago•450 comments

The Academic Pipeline Stall: Why Industry Must Stand for Academia

https://www.sigarch.org/the-academic-pipeline-stall-why-industry-must-stand-for-academia/
63•MaysonL•3h ago•36 comments

Continuous Thought Machines

https://pub.sakana.ai/ctm/
91•hardmaru•4h ago•6 comments

Intellect-2 Release: The First 32B Model Trained Through Globally Distributed RL

https://www.primeintellect.ai/blog/intellect-2-release
73•Philpax•4h ago•23 comments

Making PyPI's test suite 81% faster – The Trail of Bits Blog

https://blog.trailofbits.com/2025/05/01/making-pypis-test-suite-81-faster/
15•rbanffy•3d ago•1 comments

Car companies are in a billion-dollar software war

https://insideevs.com/features/759153/car-companies-software-companies/
286•rntn•12h ago•472 comments

Why Bell Labs Worked

https://1517.substack.com/p/why-bell-labs-worked
165•areoform•9h ago•127 comments

Absolute Zero Reasoner

https://andrewzh112.github.io/absolute-zero-reasoner/
52•jonbaer•4d ago•7 comments

Ask HN: Cursor or Windsurf?

48•skarat•1h ago•40 comments

High-school shop students attract skilled-trades job offers

https://www.wsj.com/lifestyle/careers/skilled-trades-high-school-recruitment-fd9f8257
175•lxm•15h ago•266 comments

I hacked my clock to control my focus

https://www.paepper.com/blog/posts/how-i-hacked-my-clock-to-control-my-focus.md/
60•rcarmo•7h ago•23 comments

Writing an LLM from scratch, part 13 – attention heads are dumb

https://www.gilesthomas.com/2025/05/llm-from-scratch-13-taking-stock-part-1-attention-heads-are-dumb
248•gpjt•3d ago•45 comments

Scraperr – A Self Hosted Webscraper

https://github.com/jaypyles/Scraperr
157•jpyles•12h ago•52 comments

Title of work deciphered in sealed Herculaneum scroll via digital unwrapping

https://www.finebooksmagazine.com/fine-books-news/title-work-deciphered-sealed-herculaneum-scroll-digital-unwrapping
202•namanyayg•16h ago•82 comments

LSP client in Clojure in 200 lines of code

https://vlaaad.github.io/lsp-client-in-200-lines-of-code
123•vlaaad•12h ago•14 comments

Burrito Now, Pay Later

https://enterprisevalue.substack.com/p/burrito-now-pay-later
126•gwintrob•10h ago•190 comments

Show HN: Codigo – The Programming Language Repository

https://codigolangs.com
15•adamjhf•1d ago•4 comments

One-Click RCE in Asus's Preinstalled Driver Software

https://mrbruh.com/asusdriverhub/
444•MrBruh•1d ago•211 comments

3D printing in vivo for non-surgical implants and drug delivery

https://www.science.org/doi/10.1126/science.adt0293
14•Phreaker00•1d ago•5 comments

The most valuable commodity in the world is friction

https://kyla.substack.com/p/the-most-valuable-commodity-in-the
194•walterbell•3d ago•86 comments

Hill or High Water

https://royalsociety.org/blog/2025/05/hill-or-high-water/
28•benbreen•3d ago•0 comments

Avoiding AI is hard – but our freedom to opt out must be protected

https://theconversation.com/avoiding-ai-is-hard-but-our-freedom-to-opt-out-must-be-protected-255873
111•gnabgib•6h ago•71 comments

In-Memory Ferroelectric Differentiator

https://www.nature.com/articles/s41467-025-58359-4
20•PaulHoule•3d ago•1 comments

ToyDB rewritten: a distributed SQL database in Rust, for education

https://github.com/erikgrinaker/toydb
62•erikgrinaker•10h ago•5 comments

Monitoring my Minecraft server with OpenTelemetry and Prometheus

https://www.dash0.com/blog/monitoring-minecraft-with-opentelemetry
66•mmanciop•3d ago•28 comments

I built a native Windows Todo app in pure C (278 KB, no frameworks)

https://github.com/Efeckc17/simple-todo-c
305•toxi360•14h ago•163 comments

The Epochalypse Project

https://epochalypse-project.org/
177•maxeda•20h ago•77 comments

Synder (YC S21) Is Hiring

https://www.ycombinator.com/companies/synder/jobs/2Wnbc1f-business-development-representative
1•michaelastreiko•13h ago

An online exhibition of pretty software bugs

https://glitchgallery.org/
83•tobr•13h ago•1 comments
Open in hackernews

Absolute Zero: Reinforced Self-Play Reasoning with Zero Data

https://arxiv.org/abs/2505.03335
76•leodriesch•23h ago

Comments

mentalgear•21h ago
"Despite using zero human-curated data, AZR achieves state-of-the-art results on diverse coding and math reasoning benchmarks, even outperforming models trained on large in-domain datasets. This demonstrates the potential for sophisticated reasoning skills to emerge purely through self-play without domain-specific supervision."
wiz21c•14h ago
> "sophisticated reasoning skills"

Does it mean that it uses the data it has to the maximum possible level to produce new reasoning (that add to those produced by less algorithms). IOW, are we still in the realm of: with a given data set, A.I. can produce up to N reasoning capabilities and consequently, can't produce more than that ? IOW, reasoning is bound by knowledge ? And therefore, maybe we could just start from a data/knowledge set in which we add some randomness and self play until some form of reasoning emerge ?

MoonGhost•1h ago
Up to N at a time probably. Then move on using them. The problem is the longer the chain, the more likely it will deviate from the reality. It will include non-obvious atomic decisions and wrong assumptions. This will make the whole thing unstable. I.e. without strict human supervision it likely will start producing crap. Probably some self double checks can help, but still. On the other hand humans aren't that smart either...
a2128•20h ago
To be clear, this is not a model trained on zero data, this is a pretrained model (Qwen 2.5 trained on 18 trillion tokens) finetuned using self-generated data grounded by a Python interpreter
scotty79•18h ago
I think at this point the initial process of exposing the empty model to all the available domain data in bulk is no longer interesting to many people. It's an obvious first step so it's barely mentioned anymore. What's currently worked on is what you do afterwards to get a useful tool in the end.
ethan_smith•15h ago
The breakthrough here is eliminating the need for human-labeled reasoning data while still achieving SOTA results, which has been a major bottleneck in developing reasoning capabilities.
macrolime•19h ago
Pretty sure OpenAI and/or DeepMind have already been doing something very similar for a while already, just without publishing it.
FieryTransition•18h ago
Agreed, it's a pretty obvious solution to the problems once you are immersed in the problem space. I think it's much harder to setup an efficient training pipeline for this which does every single little detail in the pipeline correctly while being efficient.
squillion•19h ago
Warning: abuse of this technique may cause the model to go blind.
ogogmad•16h ago
Is this a joke about wanking?
QuadmasterXLII•18h ago
For everyone who says “modern incentives forbid publishing negative results,” let this stand as a counterexample!
fotcorn•17h ago
Why do you think it's a negative result? The table on page 9 shows great results.
ogogmad•17h ago
I think it's a pun. AlphaZero? AlphaNegative.
andy_ppp•15h ago
-273°C isn’t it?
Waterluvian•17h ago
Related to this: has anyone seen a model respond with “oh wait I was wrong…” when you follow-up with a “can you explain why this answer is right?”

I still find that my uses of GPT and others still struggle with a sort of tunnel vision.

gitroom•16h ago
sometimes i feel like the whole self-play thing is kinda the obvious path now but still nuts seeing it actually work better than huge data dumps. you ever wonder how much of progress is just crazy good pipelines versus actual breakthroughs?
nullc•7h ago
Be nice to see some of these run on languages the pretrained model is a little less good at than Python and JS.