frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

JavaScript Trademark Update

https://deno.com/blog/deno-v-oracle4
675•thebeardisred•12h ago•226 comments

Solving `Passport Application` with Haskell

https://jameshaydon.github.io/passport/
144•jameshh•8h ago•44 comments

MCP: An (Accidentally) Universal Plugin System

https://worksonmymachine.substack.com/p/mcp-an-accidentally-universal-plugin
576•Stwerner•17h ago•253 comments

Engineered Addictions

https://masonyarbrough.substack.com/p/engineered-addictions
514•echollama•17h ago•312 comments

The Death of the Middle-Class Musician

https://thewalrus.ca/the-death-of-the-middle-class-musician/
118•pseudolus•9h ago•167 comments

Improving River Simulation

https://undiscoveredworlds.blogspot.com/2025/04/improving-river-simulation.html
11•Hooke•3d ago•0 comments

We ran a Unix-like OS Xv6 on our home-built CPU with a home-built C compiler (2020)

https://fuel.edby.coffee/posts/how-we-ported-xv6-os-to-a-home-built-cpu-with-a-home-built-c-compiler/
247•AlexeyBrin•19h ago•24 comments

Life of an inference request (vLLM V1): How LLMs are served efficiently at scale

https://www.ubicloud.com/blog/life-of-an-inference-request-vllm-v1
131•samaysharma•13h ago•14 comments

BusyBeaver(6) Is Quite Large

https://scottaaronson.blog/?p=8972
217•bdr•14h ago•154 comments

The Perils of 'Design Thinking'

https://www.theatlantic.com/books/archive/2025/06/invention-of-design-maggie-gram-book-review/683302/
11•Petiver•2d ago•4 comments

Blackwell: Nvidia's GPU

https://chipsandcheese.com/p/blackwell-nvidias-massive-gpu
44•pella•7h ago•13 comments

Community Is Motivation on Tap

https://alanwu.xyz/posts/community/
65•lunw•4d ago•23 comments

Show HN: SVG Lined Tile Generator

https://adpreese.github.io/svg-lined-tiles/
14•adpreese•3d ago•1 comments

Group of investors represented by YouTuber Perifractic buys Commodore

https://www.amiga-news.de/en/news/AN-2025-06-00123-EN.html
53•erickhill•9h ago•12 comments

2025 ARRL Field Day

https://www.arrl.org/field-day
102•rookderby•12h ago•30 comments

Brave creates new TLD on the blockchain

https://brave.com/blog/brave-tld/
6•meander_water•2h ago•0 comments

Show HN: I'm an airline pilot – I built interactive graphs/globes of my flights

https://jameshard.ing/pilot
1441•jamesharding•1d ago•192 comments

Universal pre-training by iterated random computation

https://arxiv.org/abs/2506.20057
23•liamdgray•6h ago•5 comments

An Indoor Beehive in My Bedroom Wall

https://www.keepingbackyardbees.com/an-indoor-beehive-zbwz1810zsau/
88•gscott•14h ago•36 comments

It's Known as 'The List'–and It's a Secret File of AI Geniuses

https://www.wsj.com/tech/meta-ai-recruiting-mark-zuckerberg-openai-018ed7fc
11•pretext•54m ago•4 comments

Tennis Scorigami

https://www.tennis-scorigami.com/
33•jlarks32•2d ago•3 comments

The European wood pigeon helped me appreciate its omnipresent city cousins

https://www.nytimes.com/2025/06/24/magazine/pigeons-city-nature.html
13•Thevet•3d ago•1 comments

Is being bilingual good for your brain?

https://www.economist.com/science-and-technology/2025/06/27/is-being-bilingual-good-for-your-brain
81•Anon84•14h ago•83 comments

Show HN: AGL a toy language that compiles to Go

https://github.com/alaingilbert/agl
63•alain_gilbert•3d ago•12 comments

Sirius: A GPU-native SQL engine

https://github.com/sirius-db/sirius
104•qianli_cs•17h ago•14 comments

Refurb weekend: Gremlin Blasto arcade board

http://oldvcr.blogspot.com/2025/06/refurb-weekend-gremlin-blasto-arcade.html
32•todsacerdoti•7h ago•2 comments

Finding Peter Putnam

https://nautil.us/finding-peter-putnam-1218035/
83•dnetesn•21h ago•63 comments

Gradient Descent Visualiser

https://uclaacm.github.io/gradient-descent-visualiser/
23•hamid914•3d ago•3 comments

Parsing JSON in Forty Lines of Awk

https://akr.am/blog/posts/parsing-json-in-forty-lines-of-awk
95•thefilmore•16h ago•42 comments

Memory Safe Languages: Reducing Vulnerabilities in Modern Software Development [pdf]

https://media.defense.gov/2025/Jun/23/2003742198/-1/-1/0/CSI_MEMORY_SAFE_LANGUAGES_REDUCING_VULNERABILITIES_IN_MODERN_SOFTWARE_DEVELOPMENT.PDF
76•todsacerdoti•13h ago•15 comments
Open in hackernews

Universal pre-training by iterated random computation

https://arxiv.org/abs/2506.20057
23•liamdgray•6h ago

Comments

liamdgray•6h ago
Abstract: "We investigate the use of randomly generated data for the sake of pre-training a model. We justify this approach theoretically from the perspective of algorithmic complexity, building on recent research that shows that sequence models can be trained to approximate Solomonoff induction. We derive similar, but complementary theoretical results. We show empirically that synthetically generated data can be used to pre-train a model before the data is seen. We replicate earlier results that models trained this way show zero-shot in-context learning across a variety of datasets, and that this performance improves with scale. We extend earlier results to real-world data, and show that finetuning a model after pre-training offers faster convergence and better generalization."
bionhoward•4h ago
This is a cool concept, but for comparison, I can’t help but wish there was more comparison between the treatment group and a control group that doesn’t see any universal pretraining data.

It’s good to compare various model sizes and evaluation tasks and random data generators. I just think the paper would more effectively prove its point if it could show models of same sizes which see this random data can learn better from evaluation data later on.

Could even take the initial checkpoint of the model before universal pretraining against the pretrained checkpoint. If the method works, the one that did UP will win.

Maybe I’m way off, I’ll admit I only skimmed it so far. Seems promising, just wishing for some controls.

yorwba•1h ago
In figures 2, 4, and 6, the top left end of the training curves represents models that have not seen any pretraining data. In figure 5, they're represented by dashed curves.
visarga•48m ago
Results are modest, maybe 20-30% fewer training steps to reach target performance. This won't solve the problem of organic data exhaustion. We need 100x more data.

They didn't test against actual language model pretraining, only tested against a random init.

- A: Pre-trained on their synthetic LSTM data -> fine-tuned on Wikipedia

- B: Pre-trained on different natural language corpus -> fine-tuned on Wikipedia

- C: Random initialization -> fine-tuned on Wikipedia

They only test A vs C, not A vs B.

WithinReason•7m ago
This paper addresses the problem of running out of data. You can't do B when you ran out of data so it's irrelevant.