frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

R9: Plan 9 in Rust

https://github.com/r9os/r9
1•classichasclass•25s ago•0 comments

Are we witnessing the rise of AIDEs?

1•devnullchild•2m ago•1 comments

Dev Who Launched Game into GamePass Warns Against Launching Games into GamePass

https://www.gamespot.com/articles/dev-who-launched-game-into-game-pass-warns-against-launching-games-into-game-pass/1100-6532306/
2•haunter•4m ago•0 comments

China's Chokehold on Rare Earth Metal Samarium Threatens West's Militaries

https://www.nytimes.com/2025/06/09/business/china-rare-earth-samarium-fighter-jets.html
1•bookofjoe•4m ago•1 comments

Aurora DSQL and the Circle of Life

https://marc-bowes.com/dsql-circle-of-life.html
1•stpn•5m ago•0 comments

Patch Tuesday, June 2025 Edition

https://krebsonsecurity.com/2025/06/patch-tuesday-june-2025-edition/
2•todsacerdoti•9m ago•0 comments

Daily Workflow

https://drcuis.github.io/TheCuisBook/Daily-Workflow.html
2•todsacerdoti•10m ago•0 comments

Security Is a Negotiation Problem

https://securityis.substack.com/p/security-is-a-negotiation-problem
1•noleary•14m ago•0 comments

Cross-User Context Leak on LLM

https://twitter.com/AbrahamsAg50246/status/1932546713681866833
2•Aghmat•16m ago•1 comments

How to shrink .pck size in Godot (and other tips for HTML5 mobile export)

https://jacobfilipp.com/godot/
1•surprisetalk•19m ago•0 comments

Using a wicking rope to drain standing water on a flat roof (2024)

https://www.reddit.com/r/Roofing/comments/1ds9zf8/using_a_wicking_rope_to_drain_standing_water_on_a/
2•surprisetalk•20m ago•0 comments

Elder care and AI agents via accounting and software development

https://solresol.substack.com/p/elder-care-and-ai-agents-via-accounting
1•solresol•21m ago•0 comments

OpenAI signs surprise deal with Google Cloud despite fierce AI rivalry

https://arstechnica.com/ai/2025/06/openai-signs-surprise-deal-with-google-cloud-despite-fierce-ai-rivalry/
3•belter•22m ago•0 comments

Snout's Origin: From Luna's Anxiety to an Idea Worth Chasing

https://www.thesnoutapp.com/unleashing-snout
1•TheSnoutApp•23m ago•0 comments

Building a Debugger

https://nostarch.com/building-a-debugger
2•sohkamyung•27m ago•0 comments

Pirate Site Visits Dip to 216B a Year, but Manga Piracy Is Booming

https://torrentfreak.com/pirate-site-visits-dip-to-216-billion-a-year-but-manga-piracy-is-booming-250610/
3•t-3•28m ago•0 comments

Improving Science and Restoring Trust in Public Health

https://www.hubermanlab.com/episode/improving-science-restoring-trust-in-public-health-dr-jay-bhattacharya
1•nradov•29m ago•0 comments

Show HN: I built Settle.nz a searchable, daily-updated housing auction database

https://www.settle.nz/
1•mdotk•30m ago•0 comments

A Tale of Two Claudes: What it gets right, what it gets wrong

https://steveklabnik.com/writing/a-tale-of-two-claudes/
1•goranmoomin•30m ago•0 comments

Talk to Your Slides: Efficient Slide Editing Agent with Large Language Models

https://arxiv.org/abs/2505.11604
1•PaulHoule•31m ago•0 comments

Cutlass: Fcpxml Templates – Automate final cut pro with code

https://andrewarrow.dev/cutlass/
1•andrewfromx•32m ago•0 comments

Basics of PPP

https://gekk.info/articles/ppp.html
1•ogurechny•32m ago•0 comments

Tim Owens Jazz and Broadcast Collection Digitized by Grammy Museum Grant

https://library.unt.edu/news/2025/06-06-tim-owens-collection/
2•gnabgib•34m ago•0 comments

"A Logic-Based Cipher I Designed Broke Every AI. Can You Solve It?"

1•Writer1530•34m ago•0 comments

Fine-Tuning LLMs Is a Waste of Time

https://codinginterviewsmadesimple.substack.com/p/fine-tuning-llms-is-a-huge-waste
3•j-wang•38m ago•1 comments

AI Water and Energy Converter – understand the impact of personal AI usage

https://context.supply/
1•dameis•39m ago•1 comments

Easy place to read ancient philosophy

https://readphilosophy.org
2•bridelamb•40m ago•0 comments

The Weaponization of Waymo

https://www.bloodinthemachine.com/p/the-weaponization-of-waymo
4•SLHamlet•43m ago•0 comments

Designing NoSQL Data: When to Normalize and When to Denormalize

https://constantinos.dev/posts/designing-nosql-data-when-to-normalize-and-when-to-denormalize/
1•eerenio•43m ago•0 comments

Made a website dedicated to programming languages I'll never learn

2•backteria•45m ago•0 comments
Open in hackernews

Reinforcement Pre-Training

https://arxiv.org/abs/2506.08007
53•frozenseven•18h ago

Comments

hzia•13h ago
This is very exciting! Existing data will become a lot more valuable and it brings it one step closer to how we learn as humans!

The downside is that this is going to be extremely expensive, so the data set to conduct RL will need to be curated.

watsonmusic•7h ago
cannot wait seeing how it goes beyond the current llm training pipeline
nsagent•5h ago
It's clear that you're either one of the authors or a friend of theirs. You created this account 8 months ago to comment on another paper [1] that was released by the same authors.

[1]: https://news.ycombinator.com/item?id=41776324

dgshsg•11h ago
I notice that you can do this recursively to arbitrary depth. The cost is terrible though.
watsonmusic•7h ago
it could be adaptive. only high-value tokens were allocated with more compute
babelfish•8h ago
So marginally better (and occasionally worse) performance for an order of magnitude larger training costs…?
watsonmusic•7h ago
14b model performs comparably with 32b size. the improvement is huge
85392_school•7h ago
are we only comparing them in terms of text completion accuracy? does it also improve perf on benchmarks?
watsonmusic•7h ago
A new scaling paradigm finally comes out!
beauzero•7h ago
Interesting
NotAnOtter•6h ago
I'm interested how an innovation like this affects the business prospects.

Let's assume this is a paradigm shift on the scale of Transformers / `Attention is all you need`. Companies build out new models and pump another $100 Billion through it. And then a year from now, another innovation comes out. Same circus. And again.

No one wants to be left behind but trying to keep up will sink smaller companies.

curious_cat_163•6h ago
I am not sure why this ought to require "pump another $100 Billion". Could you elaborate?

Yes, the more recent generation of GPUs optimize for attention math. But they are still fairly "general-purpose" accelerators as well. So when I see papers like this (interesting idea, btw!), my mental model for costs suggests that the CapEx to buy up the GPUs and build out the data centers would get re-used for this and 100s of other ideas and experiments.

And then the hope is that the best ideas will occupy more of the available capacity...

gessha•5h ago
Sir, this is an arxiv paper
NotAnOtter•4h ago
So true, just like this one: https://arxiv.org/abs/1706.03762
Imnimo•5h ago
This is an interesting way of squeezing extra feedback from raw text, but I'm a little skeptical that it's the best way to spend training flops. It feels like most "next tokens" are pretty low information (even after filtering for entropy like they do). Does it make sense to spend a bunch of compute on a reasoning trace on them? Maybe if you're harshly data limited, but not compute limited?
rafaelero•4h ago
This should be used for high entropy tokens during pre-training.
ntonozzi•3h ago
Is there any work related to using some kind of soft tokens for reasoning? It seems so inefficient to try to encode so much information down into a single token for the next pass of the model, when you could output a large vector for each forward pass, and have a drastically larger working memory/scratchpad, and have much higher bandwidth for the models to pass information forward to the next token call. If a single token has 17 bits of information, a vector of 1024 floats could have 32,768 bits of information.