frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Toroidal Logit Bias – Reduce LLM hallucinations 40% with no fine-tuning

https://github.com/Paraxiom/topological-coherence
1•slye514•1m ago•1 comments

Top AI models fail at >96% of tasks

https://www.zdnet.com/article/ai-failed-test-on-remote-freelance-jobs/
3•codexon•1m ago•1 comments

The Science of the Perfect Second (2023)

https://harpers.org/archive/2023/04/the-science-of-the-perfect-second/
1•NaOH•2m ago•0 comments

Bob Beck (OpenBSD) on why vi should stay vi (2006)

https://marc.info/?l=openbsd-misc&m=115820462402673&w=2
2•birdculture•5m ago•0 comments

Show HN: Glimpsh – exploring gaze input inside the terminal

https://github.com/dchrty/glimpsh
1•dochrty•6m ago•0 comments

The Optima-l Situation: A deep dive into the classic humanist sans-serif

https://micahblachman.beehiiv.com/p/the-optima-l-situation
1•subdomain•6m ago•0 comments

Barn Owls Know When to Wait

https://blog.typeobject.com/posts/2026-barn-owls-know-when-to-wait/
1•fintler•7m ago•0 comments

Implementing TCP Echo Server in Rust [video]

https://www.youtube.com/watch?v=qjOBZ_Xzuio
1•sheerluck•7m ago•0 comments

LicGen – Offline License Generator (CLI and Web UI)

1•tejavvo•10m ago•0 comments

Service Degradation in West US Region

https://azure.status.microsoft/en-gb/status?gsid=5616bb85-f380-4a04-85ed-95674eec3d87&utm_source=...
2•_____k•10m ago•0 comments

The Janitor on Mars

https://www.newyorker.com/magazine/1998/10/26/the-janitor-on-mars
1•evo_9•12m ago•0 comments

Bringing Polars to .NET

https://github.com/ErrorLSC/Polars.NET
3•CurtHagenlocher•14m ago•0 comments

Adventures in Guix Packaging

https://nemin.hu/guix-packaging.html
1•todsacerdoti•15m ago•0 comments

Show HN: We had 20 Claude terminals open, so we built Orcha

1•buildingwdavid•15m ago•0 comments

Your Best Thinking Is Wasted on the Wrong Decisions

https://www.iankduncan.com/engineering/2026-02-07-your-best-thinking-is-wasted-on-the-wrong-decis...
1•iand675•15m ago•0 comments

Warcraftcn/UI – UI component library inspired by classic Warcraft III aesthetics

https://www.warcraftcn.com/
1•vyrotek•17m ago•0 comments

Trump Vodka Becomes Available for Pre-Orders

https://www.forbes.com/sites/kirkogunrinde/2025/12/01/trump-vodka-becomes-available-for-pre-order...
1•stopbulying•18m ago•0 comments

Velocity of Money

https://en.wikipedia.org/wiki/Velocity_of_money
1•gurjeet•20m ago•0 comments

Stop building automations. Start running your business

https://www.fluxtopus.com/automate-your-business
1•valboa•25m ago•1 comments

You can't QA your way to the frontier

https://www.scorecard.io/blog/you-cant-qa-your-way-to-the-frontier
1•gk1•26m ago•0 comments

Show HN: PalettePoint – AI color palette generator from text or images

https://palettepoint.com
1•latentio•26m ago•0 comments

Robust and Interactable World Models in Computer Vision [video]

https://www.youtube.com/watch?v=9B4kkaGOozA
2•Anon84•30m ago•0 comments

Nestlé couldn't crack Japan's coffee market.Then they hired a child psychologist

https://twitter.com/BigBrainMkting/status/2019792335509541220
1•rmason•32m ago•1 comments

Notes for February 2-7

https://taoofmac.com/space/notes/2026/02/07/2000
2•rcarmo•33m ago•0 comments

Study confirms experience beats youthful enthusiasm

https://www.theregister.com/2026/02/07/boomers_vs_zoomers_workplace/
2•Willingham•40m ago•0 comments

The Big Hunger by Walter J Miller, Jr. (1952)

https://lauriepenny.substack.com/p/the-big-hunger
2•shervinafshar•41m ago•0 comments

The Genus Amanita

https://www.mushroomexpert.com/amanita.html
1•rolph•46m ago•0 comments

We have broken SHA-1 in practice

https://shattered.io/
10•mooreds•47m ago•4 comments

Ask HN: Was my first management job bad, or is this what management is like?

1•Buttons840•48m ago•0 comments

Ask HN: How to Reduce Time Spent Crimping?

2•pinkmuffinere•49m ago•1 comments
Open in hackernews

The "setup tax" on AWS H100s is killing iterative research

3•miyamotomusashi•1mo ago
I've been benchmarking the cost economics of fine tuning 70B parameter models on AWS H100 instances versus distributed consumer hardware (RTX 4090s over WAN).

The common assumption is that consumer swarms are too slow due to latency. But my modeling suggests we are ignoring the "setup tax" of the cloud.

The Data:

- Cloud (AWS): For short, iterative runs (1-2 hours), you pay for nearly 45 minutes of dead time per session just setting up environments and downloading 140GB+ weights.

- Swarm (WAN): While inference/training speed is slower (1.6x wall clock time due to network latency), the environment is persistent.

The Trade off: The math shows that for iterative research, the swarm architecture becomes ~ 57% cheaper overall, even accounting for the slower speed. You are trading latency to bypass the startup overhead and the VRAM wall.

I'm trying to validate if this trade off makes sense for real world workflows. For those finetuning 70B+ models: Is time your #1 bottleneck, or would you accept a 1.6x slowdown to cut compute costs by half ?

Comments

aikitty•1mo ago
Really interesting point about the setup tax. I hadn’t thought about how much the ephemeral nature of cloud instances kills you on iterative workflows.

Have you looked at gpu marketplaces like io.net that offer much cheaper instances than AWS. You get both benefits: no setup tax between runs and lower costs. The trade off is you may be paying during idle time between experiments. But if you’re iterating frequently the math should still work out heavily in your favor.

Curious if you’ve modelled that vs your distributed swarm approach. It might be an easier path to cost and time savings without having to architect the distributed setup yourself.

miyamotomusashi•1mo ago
This is a great point. I've benchmarked io.net and vast.ai extensively. You are right that they solve the setup tax (persistent instances) and the cost (cheaper hourly). But they hit a different hard limit: The VRAM Wall.

The Problem: To run a 70B model, you need around 140GB of VRAM.

On io.net/Vast: You can't find a single cheap consumer card with that memory (RTX 4090s cap at 24GB ). You are forced to rent expensive enterprise chips (A100s) or manually orchestrate a multi-node cluster yourself, which brings the DevOps headache.

On the Swarm: We handle that multi-node orchestration automatically. We stitch together 6x cheap 4090s to create one "Virtual GPU" with enough VRAM.

So if your model fits on one card, io.net wins. If it doesn't (like 70B+ models), that's where the swarm architecture becomes necessary.