frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

extipy – Debug your Python script with a Jupyter notebook

https://github.com/ebanner/extipy
1•meken•1m ago•1 comments

What the 'Panama Playlists' Exposed About Spotify User Privacy

https://www.nytimes.com/2025/08/24/technology/spotify-panama-playlists-privacy.html
1•reaperducer•4m ago•0 comments

States target wealthy homeowners with new property taxes, sparking backlash

https://seekingalpha.com/news/4490703-states-target-wealthy-homeowners-with-new-property-taxes-sp...
2•donsupreme•5m ago•0 comments

Device Can Read the Pages of a Book Without Opening It

https://www.wbur.org/news/2016/09/13/read-a-book-without-opening-it
1•andrewstuart•8m ago•0 comments

Big Tech Power Rankings – September 1, 2025

https://www.powerrankings.tech/
1•meshugaas•8m ago•0 comments

Space investing goes mainstream as VCs ditch the rocket science requirements

https://techcrunch.com/2025/09/01/space-investing-goes-mainstream-as-vcs-ditch-the-rocket-science...
1•DocFeind•9m ago•0 comments

All you need is Make

https://github.com/avkcode/SRE/tree/main/vault
1•GitPopTarts•10m ago•0 comments

EU

1•klknazik•11m ago•0 comments

Show HN: Restaurant AI Host, Always On

https://heep.ai/
1•mathis-ve•11m ago•0 comments

Nimony: Design Principles

https://nim-lang.org/araq/nimony.html
1•Bogdanp•11m ago•0 comments

How OnlyFans Piracy Is Ruining the Internet for Everyone

https://www.404media.co/how-onlyfans-piracy-is-ruining-the-internet-for-everyone/
2•mikhael•11m ago•0 comments

A day in the life of a vibe coder

https://medium.com/@BryMei/a-day-in-the-life-of-a-vibe-coder-a335fbb7033b
1•bry_guy•13m ago•0 comments

Ask HN: Learning Code Is Dead?

1•demirbey05•16m ago•2 comments

Show HN: TurboTable – Your AI Knowledge Worker

https://turbotable.ai
1•murshudoff•17m ago•0 comments

Dragon Bravo megafire shows the growing wildfire threat to water systems

https://theconversation.com/grand-canyons-dragon-bravo-megafire-shows-the-growing-wildfire-threat...
2•PaulHoule•21m ago•0 comments

How to Exit Vim

https://github.com/hakluke/how-to-exit-vim
1•mirawelner•22m ago•1 comments

Phyphox: Physical Phone Experiments

https://phyphox.org/
1•thunderbong•25m ago•0 comments

Raspberry Pi 5 support (OpenBSD)

https://marc.info/?l=openbsd-cvs&m=175675287220070&w=2
2•brynet•26m ago•0 comments

Massimo

https://blog.platformatic.dev/introducing-massimo
1•feross•26m ago•0 comments

NativePHP – Build Native PHP Apps

https://nativephp.com
1•TheFreim•27m ago•0 comments

Functional Source License (FSL)

https://fsl.software/
2•jotaen•29m ago•0 comments

KPH Crypto Enigma CW Message 2025 [video]

https://www.youtube.com/watch?v=DJQZffcHbf8
1•austinallegro•37m ago•0 comments

Ask HN: Tools for Crossword Puzzle Generation?

2•ekorbia•41m ago•0 comments

Faster linking times with 1.90.0 stable on Linux using the LLD linker

https://blog.rust-lang.org/2025/09/01/rust-lld-on-1.90.0-stable/
1•HieronymusBosch•46m ago•0 comments

China Is Building a Brain-Computer Interface Industry

https://www.wired.com/story/china-is-getting-serious-about-brain-computer-interfaces/
3•01-_-•47m ago•0 comments

The cryptographic sponge and duplex constructions

https://keccak.team/sponge_duplex.html
1•fanf2•47m ago•0 comments

Finnish City Inaugurates 1 MW/100 MWh Sand Battery

https://cleantechnica.com/2025/08/30/finnish-city-inaugurates-1-mw-100-mwh-sand-battery/
3•worik•51m ago•0 comments

1967 was the Summer of Love; we will remember 2025 as the Summer of Vibes

https://chiefmartec.com/2025/08/1967-was-the-summer-of-love-we-will-remember-2025-as-the-summer-o...
2•manveerc•55m ago•0 comments

What's on the Other Side of Earth – Simple Visualization Tool

https://GuessWhereYouAre.com/guide/maps/distance-tool.html
2•FreeGuessr•57m ago•1 comments

The "it" in AI models is the dataset

https://nonint.com/2023/06/10/the-it-in-ai-models-is-the-dataset/
3•jxmorris12•59m ago•1 comments
Open in hackernews

SparseLoCo: Communication-Efficient LLM Training

https://arxiv.org/abs/2508.15706
11•synapz_org•2h ago

Comments

synapz_org•2h ago
Paper: https://arxiv.org/abs/2508.15706 Code: https://github.com/tplr-ai/SparseLoCo

Templar AI has developed SparseLoCo, a distributed training algorithm that achieves extreme compression ratios (1-3% sparsity + 2-bit quantization) while outperforming existing methods like DiLoCo and DeMo on both loss and communication efficiency.

The Core Problem

Training LLMs across data centers or over the internet is bottlenecked by communication: as model scale grows, each synchronization can require transferring hundreds of gigabytes of pseudo-gradients. DiLoCo reduces the frequency of synchronizations, but the communication remains dense and large. This makes distributed training impractical for many scenarios, especially internet-scale collaboration.

Technical Approach

Our key insight: The infrequent communication of DiLoCo can be aggressively compressed via TOP-k sparsification while improving performance.

Algorithm highlights:

* Replace global momentum with per-replica error feedback * Apply TOP-k magnitude compression (1-3% density) + 2-bit quantization to pseudo-gradients * Maintain infrequent communication (H=15-250 steps) like DiLoCo * Use chunked TOP-k for better parallelism and reduced index overhead

Results

Communication reduction: With >97× compression, SparseLoCo outperforms DiLoCo across all benchmarks. Sparse aggregation appears to provide regularization benefits beyond just compression.

Communication infrequency: Consistently outperforms DiLoCo across communication frequency ∈ {15, 30, 50, 100, 250} on 512M parameter models.

Real deployment: Currently running on Bittensor with a 70B model and 20 participants in the gather operation (out of many more total participants): 70 seconds communication with <500Mbps bandwidth. Our previous successful deployment of a medium sized (200B token) run of an 8B parameter model and 20 gather participants achieved communication average of 12 seconds vs 4.5 minutes compute time.

Key Technical Contributions

1. Local momentum approximation: Show that DiLoCo's global outer momentum can be well-approximated by local accumulators (>90% cosine similarity)

2. Error feedback as momentum: Demonstrate that TOP-k + error feedback naturally provides similar benefits to outer momentum

3. Sparse aggregation benefits: Find that sparse aggregation actually improves performance over dense methods—likely due to emphasis on high-saliency components

4. Extreme quantization: Error feedback enables 2-bit quantization without additional accumulators or performance drops

Implementation Details

* Chunked TOP-k (4096 elements/chunk) reduces index transmission overhead * Custom index compression: 8.9, 6.6, 5.6 bits per value for different sparsity levels * Drop-in replacement for DiLoCo all-reduce operations * Compatible with existing distributed training frameworks

Limitations & Future Work

* Tested on 512M parameter models (though deployed on 8-70B) * Chunk size optimization could be further explored * Random-k performs significantly worse than TOP-k

This work makes distributed training viable over commodity internet connections and opens possibilities for global AI training collaborations that were previously bandwidth-prohibited.