frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

PickleBerry: A TUI based file manager written as a shell script

https://github.com/ekahPruthvi/pickle_berry
1•thunderbong•9m ago•0 comments

The Academic Pipeline Stall: Why Industry Must Stand for Academia

https://www.sigarch.org/the-academic-pipeline-stall-why-industry-must-stand-for-academia/
1•MaysonL•14m ago•0 comments

Tech debt can't be solved as a roadmap item

https://software.rajivprab.com/2025/05/10/tech-debt-cant-be-solved-as-a-roadmap-item/
1•pabs3•20m ago•0 comments

Surgical Knee Denervation for Treatment of Pain Caused by Primary Osteoarthritis

https://lww.com/_layouts/1033/OAKS.Journals/Error/JavaScript.html
1•bookofjoe•26m ago•0 comments

Migrating to Codeberg

https://guix.gnu.org/blog/2025/migrating-to-codeberg/
2•todsacerdoti•26m ago•0 comments

Continuous Thought Machines

https://pub.sakana.ai/ctm/
2•hardmaru•27m ago•0 comments

Larry Weed's Internal Medicine Grand Rounds (1971) [video]

https://www.youtube.com/watch?v=qMsPXSMTpFI
1•akkartik•28m ago•0 comments

US Copyright Office Has Thoughts on AI. Big Tech May Not Like It

https://www.businessinsider.com/ai-training-copyright-laws-big-tech-fair-use-openai-meta-2025-5
3•walterbell•31m ago•0 comments

Show HN: I Built Remind Me AI. It's Like Unlimited GPT Tasks. Try the Demo

https://app.arcade.software/share/PrR3lG51cRjJkk6vEz3R
1•ShawnBuilds•43m ago•0 comments

How to Think Like Leonardo da Vinci

https://www.artofmanliness.com/character/advice/think-like-da-vinci/
3•kamphey•46m ago•1 comments

Bios Boot to D

https://theartofmachinery.com/2017/01/24/boot_to_d.html
1•teleforce•48m ago•1 comments

Network Evaluation Service

https://github.com/hendemic/network-eval-service
1•gregsadetsky•51m ago•0 comments

Ask HN: Anyone Go to Stripe Sessions?

1•kamphey•58m ago•0 comments

DIY solar launches in US – no installers or permits needed

https://www.zdnet.com/home-and-office/energy/ecoflows-new-backyard-solar-energy-system-starts-at-599-no-installation-crews-or-permits-needed/
1•jakestein•59m ago•1 comments

A conversation with Jony Ive [video]

https://www.youtube.com/watch?v=wLb9g_8r-mE
1•kamphey•1h ago•0 comments

Developers as Suppliers

https://daringfireball.net/linked/2025/05/10/benedict-evans-apple-developers-as-suppliers
4•smugma•1h ago•0 comments

Intellect-2 Release: The First 32B Model Trained Through Globally Distributed RL

https://www.primeintellect.ai/blog/intellect-2-release
8•Philpax•1h ago•3 comments

This Linux trick will blow your mind [video]

https://www.youtube.com/watch?v=248Q1kusVy0
2•levont•1h ago•0 comments

Show HN: Parsie – A Google Sheets Add-On to Extract Data from Any Documents

https://workspace.google.com/marketplace/app/advanced_ocr_extract_text_numbers_and_ta/386593592681
2•alicele27•1h ago•0 comments

Python lib generates its code on-the-fly based on usage

https://github.com/cofob/autogenlib
1•klntsky•1h ago•0 comments

LLM Botnet: Are companies using botnets to scrape content?

2•flyriver•1h ago•1 comments

Tesla employees ask Elon Musk to resign, confirm demand problem

https://electrek.co/2025/05/11/tesla-employees-ask-elon-musk-resign-confirm-massive-demand-problem-get-fired/
11•travisgriggs•1h ago•2 comments

Human Behavior

https://www.humanbehavior.co/
1•handfuloflight•1h ago•0 comments

Dance Music Is Booming Again. What's Different This Time? A Lot

https://www.nytimes.com/2025/05/11/arts/music/dance-music-boom-nightlife.html
2•hgv•1h ago•0 comments

Sleep duration, sampling time, and physical exercise alter cardiovascular risk

https://biomarkerres.biomedcentral.com/articles/10.1186/s40364-025-00776-0
2•gnabgib•1h ago•0 comments

Custom SIM card in Tesla Model 3 2024, Tesla Model Y 2025 and Cybertruck

https://olegkutkov.me/2025/05/12/custom-sim-card-in-tesla-model-3-2024-tesla-model-y-2025-and-cybertruck/
45•LorenDB•1h ago•19 comments

Show HN: Converter for half-with and full-width string in Golang

https://github.com/suwakei/go-zhconv
1•suwakei•1h ago•0 comments

Making complex text understandable: Minimally-lossy simplification with Gemini

https://research.google/blog/making-complex-text-understandable-minimally-lossy-text-simplification-with-gemini/
1•m-hodges•1h ago•0 comments

Solved: What Is Microcode and Why You Need to Be Aware of It

https://www.urtech.ca/2025/02/solved-what-is-microcode-why-you-need-to-be-aware-of-it/
1•rolph•1h ago•0 comments

Ask HN: Is There a "Hacker News" for Economics?

3•preciousoo•1h ago•2 comments
Open in hackernews

Absolute Zero: Reinforced Self-Play Reasoning with Zero Data

https://arxiv.org/abs/2505.03335
74•leodriesch•19h ago

Comments

mentalgear•17h ago
"Despite using zero human-curated data, AZR achieves state-of-the-art results on diverse coding and math reasoning benchmarks, even outperforming models trained on large in-domain datasets. This demonstrates the potential for sophisticated reasoning skills to emerge purely through self-play without domain-specific supervision."
wiz21c•11h ago
> "sophisticated reasoning skills"

Does it mean that it uses the data it has to the maximum possible level to produce new reasoning (that add to those produced by less algorithms). IOW, are we still in the realm of: with a given data set, A.I. can produce up to N reasoning capabilities and consequently, can't produce more than that ? IOW, reasoning is bound by knowledge ? And therefore, maybe we could just start from a data/knowledge set in which we add some randomness and self play until some form of reasoning emerge ?

a2128•16h ago
To be clear, this is not a model trained on zero data, this is a pretrained model (Qwen 2.5 trained on 18 trillion tokens) finetuned using self-generated data grounded by a Python interpreter
scotty79•14h ago
I think at this point the initial process of exposing the empty model to all the available domain data in bulk is no longer interesting to many people. It's an obvious first step so it's barely mentioned anymore. What's currently worked on is what you do afterwards to get a useful tool in the end.
ethan_smith•11h ago
The breakthrough here is eliminating the need for human-labeled reasoning data while still achieving SOTA results, which has been a major bottleneck in developing reasoning capabilities.
macrolime•16h ago
Pretty sure OpenAI and/or DeepMind have already been doing something very similar for a while already, just without publishing it.
FieryTransition•15h ago
Agreed, it's a pretty obvious solution to the problems once you are immersed in the problem space. I think it's much harder to setup an efficient training pipeline for this which does every single little detail in the pipeline correctly while being efficient.
squillion•15h ago
Warning: abuse of this technique may cause the model to go blind.
ogogmad•13h ago
Is this a joke about wanking?
QuadmasterXLII•14h ago
For everyone who says “modern incentives forbid publishing negative results,” let this stand as a counterexample!
fotcorn•14h ago
Why do you think it's a negative result? The table on page 9 shows great results.
ogogmad•13h ago
I think it's a pun. AlphaZero? AlphaNegative.
andy_ppp•12h ago
-273°C isn’t it?
Waterluvian•14h ago
Related to this: has anyone seen a model respond with “oh wait I was wrong…” when you follow-up with a “can you explain why this answer is right?”

I still find that my uses of GPT and others still struggle with a sort of tunnel vision.

gitroom•12h ago
sometimes i feel like the whole self-play thing is kinda the obvious path now but still nuts seeing it actually work better than huge data dumps. you ever wonder how much of progress is just crazy good pipelines versus actual breakthroughs?
nullc•3h ago
Be nice to see some of these run on languages the pretrained model is a little less good at than Python and JS.