frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

X said it would give $1M to a user who had previously shared racist posts

https://www.nbcnews.com/tech/internet/x-pays-1-million-prize-creator-history-racist-posts-rcna257768
1•doener•59s ago•0 comments

155M US land parcel boundaries

https://www.kaggle.com/datasets/landrecordsus/us-parcel-layer
2•tjwebbnorfolk•5m ago•0 comments

Private Inference

https://confer.to/blog/2026/01/private-inference/
1•jbegley•8m ago•0 comments

Font Rendering from First Principles

https://mccloskeybr.com/articles/font_rendering.html
1•krapp•11m ago•0 comments

Show HN: Seedance 2.0 AI video generator for creators and ecommerce

https://seedance-2.net
1•dallen97•15m ago•0 comments

Wally: A fun, reliable voice assistant in the shape of a penguin

https://github.com/JLW-7/Wally
1•PaulHoule•17m ago•0 comments

Rewriting Pycparser with the Help of an LLM

https://eli.thegreenplace.net/2026/rewriting-pycparser-with-the-help-of-an-llm/
1•y1n0•18m ago•0 comments

Lobsters Vibecoding Challenge

https://gist.github.com/MostAwesomeDude/bb8cbfd005a33f5dd262d1f20a63a693
1•tolerance•18m ago•0 comments

E-Commerce vs. Social Commerce

https://moondala.one/
1•HamoodBahzar•19m ago•1 comments

Avoiding Modern C++ – Anton Mikhailov [video]

https://www.youtube.com/watch?v=ShSGHb65f3M
2•linkdd•20m ago•0 comments

Show HN: AegisMind–AI system with 12 brain regions modeled on human neuroscience

https://www.aegismind.app
2•aegismind_app•24m ago•1 comments

Zig – Package Management Workflow Enhancements

https://ziglang.org/devlog/2026/#2026-02-06
1•Retro_Dev•26m ago•0 comments

AI-powered text correction for macOS

https://taipo.app/
1•neuling•30m ago•1 comments

AppSecMaster – Learn Application Security with hands on challenges

https://www.appsecmaster.net/en
1•aqeisi•30m ago•1 comments

Fibonacci Number Certificates

https://www.johndcook.com/blog/2026/02/05/fibonacci-certificate/
1•y1n0•32m ago•0 comments

AI Overviews are killing the web search, and there's nothing we can do about it

https://www.neowin.net/editorials/ai-overviews-are-killing-the-web-search-and-theres-nothing-we-c...
3•bundie•37m ago•1 comments

City skylines need an upgrade in the face of climate stress

https://theconversation.com/city-skylines-need-an-upgrade-in-the-face-of-climate-stress-267763
3•gnabgib•38m ago•0 comments

1979: The Model World of Robert Symes [video]

https://www.youtube.com/watch?v=HmDxmxhrGDc
1•xqcgrek2•42m ago•0 comments

Satellites Have a Lot of Room

https://www.johndcook.com/blog/2026/02/02/satellites-have-a-lot-of-room/
2•y1n0•43m ago•0 comments

1980s Farm Crisis

https://en.wikipedia.org/wiki/1980s_farm_crisis
4•calebhwin•43m ago•1 comments

Show HN: FSID - Identifier for files and directories (like ISBN for Books)

https://github.com/skorotkiewicz/fsid
1•modinfo•49m ago•0 comments

Show HN: Holy Grail: Open-Source Autonomous Development Agent

https://github.com/dakotalock/holygrailopensource
1•Moriarty2026•56m ago•1 comments

Show HN: Minecraft Creeper meets 90s Tamagotchi

https://github.com/danielbrendel/krepagotchi-game
1•foxiel•1h ago•1 comments

Show HN: Termiteam – Control center for multiple AI agent terminals

https://github.com/NetanelBaruch/termiteam
1•Netanelbaruch•1h ago•0 comments

The only U.S. particle collider shuts down

https://www.sciencenews.org/article/particle-collider-shuts-down-brookhaven
2•rolph•1h ago•1 comments

Ask HN: Why do purchased B2B email lists still have such poor deliverability?

1•solarisos•1h ago•3 comments

Show HN: Remotion directory (videos and prompts)

https://www.remotion.directory/
1•rokbenko•1h ago•0 comments

Portable C Compiler

https://en.wikipedia.org/wiki/Portable_C_Compiler
2•guerrilla•1h ago•0 comments

Show HN: Kokki – A "Dual-Core" System Prompt to Reduce LLM Hallucinations

1•Ginsabo•1h ago•0 comments

Software Engineering Transformation 2026

https://mfranc.com/blog/ai-2026/
1•michal-franc•1h ago•0 comments
Open in hackernews

The "setup tax" on AWS H100s is killing iterative research

3•miyamotomusashi•1mo ago
I've been benchmarking the cost economics of fine tuning 70B parameter models on AWS H100 instances versus distributed consumer hardware (RTX 4090s over WAN).

The common assumption is that consumer swarms are too slow due to latency. But my modeling suggests we are ignoring the "setup tax" of the cloud.

The Data:

- Cloud (AWS): For short, iterative runs (1-2 hours), you pay for nearly 45 minutes of dead time per session just setting up environments and downloading 140GB+ weights.

- Swarm (WAN): While inference/training speed is slower (1.6x wall clock time due to network latency), the environment is persistent.

The Trade off: The math shows that for iterative research, the swarm architecture becomes ~ 57% cheaper overall, even accounting for the slower speed. You are trading latency to bypass the startup overhead and the VRAM wall.

I'm trying to validate if this trade off makes sense for real world workflows. For those finetuning 70B+ models: Is time your #1 bottleneck, or would you accept a 1.6x slowdown to cut compute costs by half ?

Comments

aikitty•1mo ago
Really interesting point about the setup tax. I hadn’t thought about how much the ephemeral nature of cloud instances kills you on iterative workflows.

Have you looked at gpu marketplaces like io.net that offer much cheaper instances than AWS. You get both benefits: no setup tax between runs and lower costs. The trade off is you may be paying during idle time between experiments. But if you’re iterating frequently the math should still work out heavily in your favor.

Curious if you’ve modelled that vs your distributed swarm approach. It might be an easier path to cost and time savings without having to architect the distributed setup yourself.

miyamotomusashi•1mo ago
This is a great point. I've benchmarked io.net and vast.ai extensively. You are right that they solve the setup tax (persistent instances) and the cost (cheaper hourly). But they hit a different hard limit: The VRAM Wall.

The Problem: To run a 70B model, you need around 140GB of VRAM.

On io.net/Vast: You can't find a single cheap consumer card with that memory (RTX 4090s cap at 24GB ). You are forced to rent expensive enterprise chips (A100s) or manually orchestrate a multi-node cluster yourself, which brings the DevOps headache.

On the Swarm: We handle that multi-node orchestration automatically. We stitch together 6x cheap 4090s to create one "Virtual GPU" with enough VRAM.

So if your model fits on one card, io.net wins. If it doesn't (like 70B+ models), that's where the swarm architecture becomes necessary.