frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Embryo Selection in 2025

https://www.sebjenseb.net/p/embryo-selection-in-2025
1•paulpauper•2m ago•0 comments

AI #134: If Anyone Reads It

https://thezvi.substack.com/p/ai-134-if-anyone-reads-it
1•paulpauper•3m ago•0 comments

What is AI market worth?

https://www.stephendiehl.com/posts/ai_marketcap/
2•ibobev•6m ago•0 comments

Dev Culture Is Dying the Curious Developer Is Gone

https://dayvster.com/blog/dev-culture-is-dying-the-curious-developer-is-gone/
2•ibobev•6m ago•0 comments

The Tinkerings of Robert Noyce

https://web.stanford.edu/class/e145/2007_fall/materials/noyce.html
2•jdcampolargo•9m ago•0 comments

FFI Overhead

https://github.com/dyu/ffi-overhead
2•steve-chavez•11m ago•0 comments

Feathers fly in dispute over Ambani zoo's pursuit of rare parrot

https://www.reuters.com/sustainability/boards-policy-regulation/feathers-fly-dispute-over-ambani-...
2•petethomas•13m ago•0 comments

Rewrites and Rollouts

https://www.lux.camera/rewrites-and-rollouts/
1•tosh•13m ago•0 comments

Rungis: The Market and the City – A day at Europe's largest fresh food market

https://www.vittlesmagazine.com/p/rungis-the-market-and-the-city
1•speckx•15m ago•0 comments

Google unveils masterplan for letting AI shop on your behalf

https://www.theregister.com/2025/09/16/google_unveils_masterplan_for_letting/
1•porterde•16m ago•1 comments

How the most elderly country is fighting heat in a deadly double crisis

https://www.cnn.com/2025/09/19/asia/japan-climate-heat-elderly-crisis-intl-hnk-dst
1•rawgabbit•16m ago•0 comments

EVs Have Gotten Too Powerful

https://www.wired.com/story/evs-have-gotten-too-powerful/
2•FromTheArchives•16m ago•0 comments

Plastic Recycling Is Mostly Fictional. Trump's EPA Approves

https://jacobin.com/2025/09/trump-epa-plastic-recycling-deregulation/
2•PaulHoule•18m ago•0 comments

Show HN: RocketQA – Write Tests in English (Gherkin), Run with Playwright

https://rocketqa.ai
2•refactormonkey•19m ago•0 comments

iPhone 17 Teardowns Confirm SIM and ESIM-Only Battery Capacities

https://www.macrumors.com/2025/09/19/iphone-17-and-17-pro-sim-esim-battery-capacities/
1•tosh•19m ago•0 comments

Hard Drives Are Making an AI Comeback

https://www.wsj.com/tech/ai/hard-drives-are-making-an-ai-comeback-yes-hard-drives-cc6e461f
1•bookofjoe•20m ago•1 comments

Walmart's lax vetting helped Marketplace boom, but came with fakes and frauds

https://www.cnbc.com/2025/09/19/walmart-marketplace-fakes-scams-investigation.html
3•throwoutway•20m ago•1 comments

Instrumenting the Node.js event loop with eBPF

https://coroot.com/blog/instrumenting-the-node-js-event-loop-with-ebpf/
3•openWrangler•21m ago•0 comments

Getting Google Maps to leak data

https://silliest.website:3/blog/google-maps-dates/
2•speckx•22m ago•0 comments

AI Creates a Counterfeit of Meaning

https://sujato.github.io/meaningless.ai/
1•throwawyci•24m ago•0 comments

Show HN: Extremely simple cluster-compute software

https://docs.burla.dev
2•pancakeguy•25m ago•0 comments

Unveiling Silicon Art: Dieshots of Microchip Masterpieces

http://dieshot.com/
1•limoce•26m ago•0 comments

Jazz Guitarist Stanley Jordan Wrote APL for Music

https://dl.acm.org/doi/pdf/10.1145/75144.75174
2•Bogdanp•26m ago•0 comments

Trump Says U.S. and China Approve TikTok Deal After Call with Xi

https://www.wsj.com/world/china/tiktok-ban-deal-trump-xi-call-f592d6f7
4•jaredwiener•28m ago•0 comments

Alternative World Map Projections

https://en.wikipedia.org/wiki/List_of_map_projections
3•pchangr•28m ago•1 comments

Tomtit – simple CLI task runner with a lot of plugins

https://github.com/melezhik/Sparrow6/blob/master/posts/TomtitIntro.md
1•melezhik•30m ago•1 comments

Huawei unveils Atlas 950 SuperCluster – promises 1 ZettaFLOPS FP4 performance

https://www.tomshardware.com/tech-industry/artificial-intelligence/huawei-unveils-atlas-950-super...
1•buyucu•31m ago•0 comments

Intel x Nvidia: Hammer Lake leaks with "large and powerful" (Nvidia) iGPU

https://www.notebookcheck.net/Intel-x-Nvidia-Hammer-Lake-leaks-with-large-and-powerful-iGPU-as-In...
1•cowboyscott•32m ago•0 comments

Scar Programming Language

https://github.com/scar-lang/scar
2•thunderbong•37m ago•2 comments

Kernel: Introduce Multikernel Architecture Support

https://lwn.net/ml/all/20250918222607.186488-1-xiyou.wangcong@gmail.com/
4•ahlCVA•39m ago•0 comments
Open in hackernews

I regret building this $3000 Pi AI cluster

https://www.jeffgeerling.com/blog/2025/i-regret-building-3000-pi-ai-cluster
149•speckx•1h ago

Comments

elzbardico•1h ago
Frankly, always thought about Pi Clusters as a nerd indulgence, something to play, not to do serious work.
devmor•1h ago
After a few years of experience with them I agree for the most part. They are great for individual projects and even as individual servers for certain loads, but once you start clustering them you will probably get better results from a purpose built computer in the same price range as multiple pis.
randomgermanguy•1h ago
I think the only exception is specifically for studying network/communciation-topologies. I've seen a couple clusters (ca. 10-50 Pi's) in universities for both research and teaching.
NitpickLawyer•1h ago
It reminds me of the Beowulf clusters of the 90s-2000s, that were all the rage at some point, then slowly lost ground... I remember many friends tinkering with some variant of those, we had one in Uni, and there were even some linux distros dedicated to the concept.
gary_0•1h ago
Oh yeah, the "imagine a Beowulf cluster of these" Slashdot meme! I miss those days. At least the "can it run Doom?" meme is still alive and kicking.
unregistereddev•20m ago
Ditto! It reminded me of the time in college when I built a Beowulf cluster from recently-retired Pentium II desktops.

Was it fast? No. But that wasn't the point. I was learning about distributed computing.

fidotron•1h ago
If Pi Clusters were actually cost competitive for performance there would be data centres full of them.
phoronixrly•1h ago
If they were cost competitive for ... anything at all really...
wltr•56m ago
Well I have a Pi as a home server, and it’s very energy efficient, while doing what I want. Since I don’t need latest and greatest (I don’t see any difference with a modern PC for my use case), it’s very competitive for me. No need for any cooling is bonus.
ACCount37•55m ago
Prototyping and low volume.

They're good for long as the development costs dominate the total costs.

jacobr1•50m ago
They are competitive for hobbyist use cases. Limited home servers, or embedded applications that overlap with arduino.
shermantanktop•1h ago
Like the joke about the economists not picking up the $20 bill on the ground?

Faith in the perfect efficiency of the free market only works out over the long term. In the short term we have a lot of habits that serve as heuristics for doing a good job most of the time.

IAmBroom•58m ago
> Faith in the perfect efficiency of the free market only works out over the long term

... and even then it doesn't always prove true.

infecto•55m ago
Sure but for commodities, like server hardware, we can say it’s usually directionally correct. If there are no pi cloud offerings, there is probably a good economic reason for it.
ThrowawayR2•36m ago
There's been so much investigation into alternative architectures for datacenters and cloud providers, including FAANG resorting to designing their own ARM processors and accelerator chips (e.g. AWS Graviton, Google TPUs) and having them fabbed, that that comes off not as clever cynicism but silly cynicism.
dbg31415•1h ago
Was it cost effective? Meh.

Was it a learning experience?

More importantly, did you have some fun? Just a little? (=

phoronixrly•1h ago
> Was it a learning experience?

Also no. The guy's a youtuber

On the other hand, will this make him 100+k views? Yes. It's bait - the perfect combo to attract both the AI crowd and the 'homelab' enthusiasts (of which the bulk are yet to find any use for their raspberry devices)...

aprdm•1h ago
He is not a YouTuber. And even if he was - what's the problem ?

Jeff has many useful OSS software used by many companies around the world daily - including mine. What have you created ?

phoronixrly•57m ago
> What have you created ?

Nothing that is not AGPL-licensed, so you and your company haven't taken advantage of it.

I am not sure how this relates to my comment though.

aprdm•49m ago
Why lie ? https://github.com/geerlingguy/ansible-role-docker/blob/mast...
vel0city•56m ago
He may also be a good OSS contributor and writer, but he is also a Youtuber. Over 500 videos posted, 175M views, nearly a million subscribers.

Not that its a problem, I don't see why it would inherently be a negative thing. Dude seems to make some good content across a lot of different mediums. Cheers to Jeff.

IAmBroom•53m ago
He absolute is a YouTuber.

https://www.jeffgeerling.com/projects

And the inference is that he is doing this for clicks, i.e. clickbait. The very title is disingenuous.

Your attack on the poster above you is childish.

op00to•49m ago
"He is not a YouTuber"... what?

https://www.youtube.com/c/JeffGeerling

"978K subscribers 527 videos"

Jeff's had a pattern of embellishing controversies, misrepresenting what people say, and using his platform to create narratives that benefit his content's engagement. This is yet another example of farming outrage to get clicks. I don't understand why people drool over his content so much.

aprdm•12m ago
I guess I met Jeff's work on Ansible for DevOps: Server and configuration management for humans which is roughly 10 years old.

I then used many of his ansible playbooks on my day to day job, which paid my bills and made my career progress.

I don't check youtube so I didn't know that he was an "youtuber", I do know his other side and how mucH I have leveraged his content/code in my career

Coffeewine•1h ago
It's a pretty rough headline, clearly the author had fun performing the test and constructing the thing.

I would be pretty regretful of just the first sentence in the article, though:

> I ordered a set of 10 Compute Blades in April 2023 (two years ago), and they just arrived a few weeks ago.

That's rough.

geerlingguy•30m ago
That's the biggest regret; but I've backed 6 Kickstarter projects over the years. Median time to deliver is 1 year.

Somehow I've actually gotten every item I backed shipped at some point (which is unexpected).

Hardware startups are _hard_, and after interacting with a number of them (usually one or two people with a neat idea in an underserved market), it seems like more than half fail before delivering their first retail product. Some at least make it through delivering prototypes/crowdfunded boards, but they're already in complete disarray by the end of the shipping/logistics nightmares.

bravetraveler•1h ago
Wow that's a lot of scratch for... scratch. Pays for itself, I'm sure: effective bait :)

'Worth it any more'? At this size, never. A Pi is a Pi is a Pi!

A few are fine for toying around; beyond that, hah. Price:perf is rough, does not improve with multiplication [of units, cost, or complexity].

xnx•1h ago
Fun project. Was the author hoping for cost effective performance?!

I assumed this was a novelty, like building a RAID array out of floppy drives.

leptons•1h ago
A lot of people don't understand the performance limits of the Raspberry Pi. It's a great little platform for some things, but it isn't really fit for half the use cases I've seen.
Our_Benefactors•57m ago
This was my impression as well, the bit about GPU incompatibility with llama.cpp made me think he was in over his head.
LTL_FTC•1h ago
The author is a YouTuber and projects like these pay for themselves with the views they garner. Even the title is designed for engagement.
lumost•1h ago
I don’t really get why anyone would be buying ai compute unless A) to your goal is to rent out the compute B) no vendor can rent you enough compute when you need it C) you have an exotic funding arrangement that makes compute capex cheap and opex expensive.

Unless you can keep your compute at 70% average utilization for 5 years - you will never save money purchasing your hardware compared to renting it.

2OEH8eoCRo0•1h ago
I don't get why anyone would hack on and have fun with unique hardware either /s
seanw444•1h ago
It's also not always just about fun or cost effectiveness. Taking the infrastructure into your own hands is a nice way to know that you're not being taken advantage of, and you only have yourself to rely on to make the thing work. Freedom and self-reliance, in short.
HenryMulligan•1h ago
Data privacy and security don't matter? My secondhand RTX 3060 would buy a lot of cloud credits, but I don't want tons of highly personal data sent to the cloud. I can't imagine how it would be for healthcare and finance, at least if they properly shepherded their data.
tern•43m ago
For most people, no, privacy does not matter in this sense, and "security" would only be a relevant term if there was a pre-existing adversarial situation
justinrubek•1h ago
At some point, the work has to actually be done rather than shuffling the details off to someone else.
causal•50m ago
1) Data proximity (if you have a lot of data, egress fees add up)

2) Hardware optimization (the exact GPU you want may not always be available for some providers)

3) Not subject to price changes

4) Not subject to sudden Terms of Use changes

5) Know exactly who is responsible if something isn't working.

6) Sense of pride and accomplishment + Heating in the winter

horsawlarway•43m ago
There are an absolutely stunning number of ways to lose a whole bunch of money very quickly if you're not careful renting compute.

$3,000 is well under many "oopsie billsies" from cloud providers.

And that's outside of the whole "I own it" side of the conversation, where things like latency, control, flexibility, & privacy are all compelling reasons to be willing to spend slightly more.

I still run quite a number of LLM services locally on hardware I bought mid-covid (right around 3k for a dual RTX3090 + 124gb system ram machine).

It's not that much more than you'd spend if you're building a gaming machine anyways, and the nifty thing about hardware I own is that it usually doesn't stop working at the 5 year mark. I have desktops from pre-2008 still running in my basement. 5 year amortization might have the cloud win, but the cloud stops winning long before most hardware dies. Just be careful about watts.

Personally - I don't think pi clusters really make much sense. I love them individually for certain things, and with a management plane like k8s, they're useful little devices to have around. But I definitely wouldn't plan to get good performance from 10 of them in a box. Much better off spending roughly the same money for a single large machine unless you're intentionally trying to learn.

a2128•37m ago
Why do people buy gaming PCs when it's much cheaper to use streaming platforms? I think the two cases share practically the same parallels in terms of reliability, availability, restrictions, flexibility, sovereignty, privacy, etc.

But also when it comes to Vast/RunPod it can be annoying and genuinely become more expensive if you have to rent 2x the number of hours because you constantly have to upload and download data, checkpoints, continuous storage costs, transfer data to another server because the GPU is no longer available, etc. It's just less of a headache if you have an always available GPU with a hard drive plugged into the machine and that's it

bunderbunder•1h ago
Reminds me a bit of one of my favorite NormConf sessions, "Just use one big machine for model training and inference." https://youtu.be/9BXMWDXiugg?si=4MnGtOSwx45KQqoP

Or the oldie-but-goodie paper "Scalability! But at what COST?": https://www.usenix.org/system/files/conference/hotos15/hotos...

Long story short, performance considerations with parallelism go way beyond Amdahl's Law, because supporting scale-out also introduces a bunch of additional work that simply doesn't exist in a single node implementation. (And, for that matter, multithreading also introduces work that doesn't exist for a sequential implementation.) And the real deep down black art secret to computing performance is that the fastest operations are the ones you don't perform.

Aurornis•1h ago
I thought the conclusion should have been obvious: A cluster of Raspberry Pi units is an expensive nerd indulgence for fun, not an actual pathway to high performance compute. I don’t know if anyone building a Pi cluster actually goes into it thinking it’s going to be a cost effective endeavor, do they? Maybe this is just YouTube-style headline writing spilling over to the blog for the clicks.

If your goal is to play with or learn on a cluster of Linux machines, the cost effective way to do it is to buy a desktop consumer CPU, install a hypervisor, and create a lot of VMs. It’s not as satisfying as plugging cables into different Raspberry Pi units and connecting them all together if that’s your thing, but once you’re in the terminal the desktop CPU, RAM, and flexibility of the system will be appreciated.

bunderbunder•1h ago
The cost effective way to do it is in the cloud. Because there's a very good chance you'll learn everything you intended to learn and then get bored with it long before your cloud compute bill reaches the price of a desktop with even fairly modest specs for this purpose.
aprdm•1h ago
That really depends on what you want to learn and how deep. If you're automating things before the hypervisor comes online or there's an OS running (e.g: working on datacenter automation, bare metal as a service) you will have many gaps
Almondsetat•1h ago
I can get a Xeon E5-2690V4 with 28 threads and 64GB of RAM for about $150. If you need cores and memory to make a lot of VMs you can do it extremely cheaply
nine_k•59m ago
It will probably consume $150 worth of electricity in less than a month, even sitting idle :-\
blobbers•53m ago
The internet says 100W idle, so maybe more like $40-50 electricity, depending on where you live could be cheaper could be more expensive.

Makes me wonder if I should unplug more stuff when on vacation.

yjftsjthsd-h•42m ago
> Makes me wonder if I should unplug more stuff when on vacation.

What's the margin on unplugging vs just powering off?

dijit•32m ago
By "off" you mean, functionally disabled but with whatever auto-update system in the background with all the radios on for "smart home" reasons - or, "off"?
nine_k•40m ago
I was surprised to find out that my apartment pulls 80-100W when everything is seemingly down during the night. A tiny light here and there, several displays in sleep mode, a desktop idling (mere 15W, but), a laptop charging, several phones charging, etc, the fridge switches on for a short moment. The many small amounts add up to something considerable.
rogerrogerr•26m ago
100W over a month (rule of thumb 730 hours) is 73kWh. Which is $7.30 at my $0.10/kWh rate, or less than $25 at (what Google told me is) Cali’s average $0.30/kWh.
mercutio2•10m ago
Your googling gave results that were likely accurate for California 4-5 years ago. My average cost per kWh is about 60 cents.

Rates have gone up enormously because the cost of wildfires is falling on ratepayers, not the utility owners.

Regulated monopolies are pretty great, aren’t they? Heads I win, tales you lose.

titanomachy•24m ago
100W continuous at 12¢/kWh (US average) is only ~$9 / month. Is your electricity 5x more expensive than the US average?
mercutio2•15m ago
Not OP, but my California TOU rates are between a 40 and 70 cents per kWh.

Still only $50/month, not $150, but I very much care about 100W loads doing no work.

RussianCow•11m ago
The US average hasn't been that low in a few years; according to [0] it's 17.47¢/kWh, and significantly higher in some parts of the country (40+ in Hawaii). And the US has low energy costs relative to most of the rest of the world, so a 3-5x multiplier over that for other countries isn't unreasonable. Plus, energy prices are currently rising and will likely continue to do so over the next few years.

$50/month for 100W continuous usage isn't totally mad, and that could climb even higher over the rest of the decade.

sebastiansm•56m ago
On Aliexpress those Xeon+mobo+ram kits are really cheap.
kbenson•18m ago
Source? That seems like something I would want to take advantage if at the moment...
kllrnohj•14m ago
Note the E5-2690V4 is a 10 year old CPU, they are talking about used servers. You can find those on ebay or whatever as well as stores specializing in that. Depending on where you live, you might even find them free as they are often considered literal ewaste by the companies decommissioning them.

It also means it performs like a 10 year old server CPU, so those 28 threads are not exactly worth a lot. The geekbench results, for whatever value those are worth, are very mediocre in the context of anything remotely modern: https://browser.geekbench.com/processors/intel-xeon-e5-2690-...

Like a modern 12-thread 9600x runs absolute circles around it https://browser.geekbench.com/processors/amd-ryzen-5-9600x

montebicyclelo•58m ago
Yeah... Looks like can get about $1/hr for 10 small VMs, ($0.10 per VM).

So for $3000, that's 3000 hours, or 125 days, (if just wastefully leave them on all the time, instead of turning them on when needed).

Say you wanted to play around for a couple of hours, that's like.. $3.

(That's assuming there's no bonus for joining / free tier, too.)

verdverm•57m ago
You can rent a beefy vm with an H100 for $1.50 / hr

I regularly rent this for a few hours at a time for learning and prototyping

Y_Y•45m ago
You can rent a beefy Ferrari 296 down by the beach for $500/hr

I regularly rent this for a few hours at a time for learning and posing

verdverm•7m ago
I'll take the H1/200s over a vehicle any day of the week
wongarsu•16m ago
The VMs quickly get expensive if you leave them running though.

The desktop equivalent of your 10 T3 Micro instances is about $600 if you buy new. For example a Lenovo ThinkCentre M75q Gen 2 Tiny 11JN009QGE has 8x3.2GHz processor with hyperthreading. That's 16 virtual cores compared to the 20 vcpus of the T3 instances, but with much faster cores. And 16GB RAM allows you to match the 1GB per instance.

If you don't have anything and feel generous throw in another $200 for a good monitor and keyboard plus mouse. But you can get a used crap monitor for $20. I'd give you one for free just to be rid of it.

That's a total of $800, or 33 days of forgetting to shut down the 10 VMs. Maybe half that if you buy used.

Granted not everyone has $800 or even $400 to drop on hobby projects, renting VMs often does make sense

nsxwolf•55m ago
That isn’t fun. I have a TI-99/4A in my office hooked up to a raspberry pi so it can use the internet. Why? Because it’s fun. I like to touch and see the things even though it’s all so silly.
bakugo•38m ago
It heavily depends on the use case. For these AI setups, you're completely correct, because the people who talk about how amazing it is to run a <100B model at home almost never actually end up using it for anything real (mostly because these small models aren't actually very good) and are doing it purely for the novelty.

But if you're someone like me who intends to actively use the hardware for real-world purposes, the cloud often simply can't compete on price. At home, I have a mini PC with a 5600G, 32GB of RAM, and a few TBs of NVME storage. The entire thing cost less than $600 a few years ago, and consumes around 20W of power on average.

Even on the cheapest cloud providers available, an equivalent setup would exceed that price in less than half a year. SSD storage in particular is disproportionately expensive on the cloud. For small VMs that don't need much storage, it does make sense, but as soon as you scale up, cloud prices quickly start ballooning.

newsclues•31m ago
Text and reference books are free at the library.

You don’t need hardware to learn. Sure it helps but you can learn from a book and pen and paper exercises.

trenchpilgrim•27m ago
I disagree. Most of what I've learned about systems comes from debugging the weird issues that only happen on real systems, especially real hardware. The book knowledge is like, 20-30% of it.
titanomachy•21m ago
Agreed, I don't think I'd hire a datacenter engineer whose experience consisted of reading books and doing "pen and paper exercises".
glitchc•1h ago
I did some calculations on this. Procuring a Mac Studio with the latest Mx Ultra processor and maxing out the memory seems to be the most cost effective way to break into 100b+ parameter model space.
Palomides•1h ago
even a single new mac mini will beat this cluster on any metric, including cost
randomgermanguy•59m ago
Depends on how heavy one wants to go with the quants (for Q6-Q4 the AMD Ryzen AI MAX chips seem better/cheaper way to get started).

Also the Mac Studio is a bit hampered by its low compute-power, meaning you really can't use a 100b+ dense model, only MoE feasibly without getting multi minute prompt-processing times (assuming 500+ tokens etc.)

GeekyBear•11m ago
Given the RAM limitations of the first gen Ryzen AI MAX, you have no choice but to go heavy on the quantization of the larger LLMs on that hardware.
the8472•48m ago
You could try getting a DGX Thor devkit with 128GB unified memory. Cheaper than the 96GB mac studio and more FLOPs.
eesmith•42m ago
Geerling links to last month's essay on a Frameboard cluster, at https://www.jeffgeerling.com/blog/2025/i-clustered-four-fram... . In it he writes 'An M3 Ultra Mac Studio with 512 gigs of RAM will set you back just under $10,000, and it's way faster, at 16 tokens per second.' for 671B parameters, that is, that M3 is at least 3x the performance of the other three systems.
GeekyBear•34m ago
Now that we know that Apple has added tensor units to the GPU cores the M5 series of chips will be using, I might be asking myself if I couldn't wait a bit.
teleforce•23m ago
Not quite, as it stands now the most cost effective way is most likely framework desktop or similar system for example HP G1a laptop/PC [1],[2].

[1] The Framework Desktop is a beast:

https://news.ycombinator.com/item?id=44841262

[2] HP ZBook Ultra:

https://www.hp.com/us-en/workstations/zbook-ultra.html

llm_nerd•21m ago
The next generation M5 should bring the matmul functionality seen on the A19 Pro to the desktop SoC's GPU -- "tensor" cores, in essence -- and will dramatically improve the running of most AI models on those machine.

Right now the Macs are viable purely because you can get massive amounts of unified memory. Be pretty great when they have the massive matrix FMA performance to complement it.

moduspol•1h ago
Also cost effective is to buy used rack mount servers from Amazon. They may be out of warranty but you get a lot more horsepower for your buck, and now your VMs don’t have to be small.
Aurornis•1h ago
Putting a retired datacenter rack mount server in your house is a great way to learn how unbearably loud a real rack mount datacenter server is.
tempest_•47m ago
ahah and pricey power wise.

Currently the cloud providers are dumping second gen xeon scalables and those things are pigs when it comes to power use.

Sound wise its like someone running a hair dryer at full speed all the time and it can be louder under load.

Tsiklon•36m ago
To quote @swiftonsecurity - https://x.com/swiftonsecurity/status/1650223598903382016 ;

> DO NOT TAKE HOME THE FREE 1U SERVER YOU DO NOT WANT THAT ANYWHERE A CLOSET DOOR WILL NOT STOP ITS BANSHEE WAIL TO THE DARK LORD AN UNHOLY CONDUIT TO THE DEPTHS OF INSOMNIA BINDING DARKNESS TO EVEN THE DAY

Y_Y•42m ago
If you're following this path, make sure to use the finest traditional server rack that money can buy: https://www.ikea.com/ie/en/p/lack-side-table-white-30449908/
allanrbo•39m ago
No, again, just run VMs on your desktop/laptop. The software doesn't know or care if it's a rack mounted machine.
vlovich123•56m ago
I’d say it’s inconclusive. For traditional compute it wins on power and cost (it’ll always lose on space). The inference is noted to not be able to use the GPU due to llama.cpp’s vulkan backend AND that clustering software in llama.cpp is bad. I’d say it’s probably still going to be worse for AI but it’s inconclusive where it could be due to the software immaturity (ie not worth it today but could be with better software)
llm_nerd•43m ago
If you assume that the author did this to have content for his blog and his YouTube channel, it makes much more sense. Going back to the well with a "I regret" entry allows for extra exploiting of a pretty dubious venture.

YouTube is absolute jam packed full of people pitching home "lab" sort of AI buildouts that are just catastrophically ill-advised, but it yields content that seems to be a big draw. For instance Alex Ziskind's content. I worry that people are actually dumping thousands to have poor performing ultra-quantized local AIs that will have zero comparative value.

philipwhiuk•42m ago
I doubt anyone does this seriously.
nerdsniper•38m ago
I sure hope no one does this seriously expecting to save some money. I enjoy the videos on "catastrophically ill-advised" build-outs. My primary curiosities that get satisfied by them are:

1) How much worse / more expensive are they than a conventional solution?

2) What kinds of weird esoteric issues pop up and how do they get solved (e.g. the resizable BAR issue for GPU's attached to RPi's PCIe slot)

ww520•35m ago
Now. Imagine a Beawulf of these...
TZubiri•15m ago
Fun fact, a raspberry pi does not have a built in Real Time Clock with its own battery, so it relies on network clocks to keep the time.

Another fun fact, the network module of the pi is actually connected to the USB bus, so there's some overhead as well as a throughput limitation.

Fun fact, the Pi does not have a power button, relying on software to shut down cleanly. If you lose access to the machine, it's not possible to avoid corrupted states on the disk.

Despite all of this, if you want to self host some website, the raspberry pi is still an amazingly cost effective choice, from anywhere between 2 to 20000 monthly users, one pi will be overprovisioned. And you can even get an absolutely overkill redundant pi as a failover, but still a single pi can reach 365 days of uptime with no problem, and as long as you don't reboot or lose power or lose internet, you can achieve more than a couple of nines of reliability.

wccrawford•15m ago
Geerling's titles have been increasingly click-bait for a while now. It's pretty sad, because I like his content, but hate the click-bait BS.
aprdm•1h ago
Love Jeff's ansible roles/playbooks and his cluster building ! Quite interesting, I should reserve some time to play with a Pi cluster and ansible, sounds fun
noelwelsh•1h ago
I'd love to understand the economics here. $3000 purely for fun seems like a lot. $3000 for promotion of a channel? consulting? seems reasonable.
philipwhiuk•35m ago
Jeff has a million YouTube subscribers, gets $2000 a month from Patreon and has 200 GitHub sponsors.

The economics of spending $3,000 on a video probably work out fine.

geerlingguy•22m ago
It's definitely a stretch for my per-video budget, but I did want to have a 'maxed out' Pi cluster for future testing as well.

A lot of people (here, Reddit, elsewhere) speculate about how good/bad a certain platform or idea is. Since I have the means to actually test how good or bad something is, I try to justify the hardware costs for it.

Similar to testing various graphics cards on Pis, I've probably spent a good $10,000 on those projects over the past few years, but now I have a version of every major GPU from the past 3 generations to test on, not only on Pi, but other Arm platforms like Ampere and Snapdragon.

Which is fun, but also educational; I've learned a lot about inference, GPU memory access, cache coherency, the PCIe bus...

So a lot of intangibles, many of which never make it directly into a blog post or video. (Similar story with my time experiments).

markx2•1h ago
> "But if you're on the blog, you're probably not the type to sit through a video anyway. So moving on..."

Thank you!

AlfredBarnes•1h ago
My pi's are just an easy onramp for me to have a functional NAS, PIHole, and webcam security.

Not at all the best, but they were cheap. If i WANTED the best or reliable, i'd actually buy real products.

hn_throw_250915•57m ago
I read through it and it’s amusing but along with the title being something I’d receive in email from a newsletter mailing list I’ve never subscribed to (hoping it has an unsubscribe link at the bottom), there’s nothing really of hacker curiosity here to keep me hooked. It’s shallow and appeals to some LCD “I did the thing with the stuff and the results will shock you because of how obvious they are now click here” mentality. Vainposting at its most average. The Mac restoration video was somewhat easier to sit through if only because the picture quality beats out a handful of other YT videos doing the exactly same thing as I’m holding back a jaw grating wince of watching someone butchering a board with poor knowledge of soldering iron practice, so YMMV? Back to hackaday for me I think. I’m not here to read submarine resumes of people applying to work at Linus Tech Tips.
paxys•54m ago
This is just the evolution of clickbait titles. The only thing missing is a thumbnail of an AI generated raspberry pi cluster with a massive arrow pointing to it and the words "not worth it!!"
Joker_vD•51m ago
There also needs to be a Face Screaming in Fear Emoji plastered on the other side of it.
imtringued•54m ago
Oh come on Jeff, you forgot to buy GPUs for your AI cluster. Such a beginner mistake.

All you needed to do is buy 4x xtx 7900 used on ebay and build a four node raspberry pi cluster using the external GPU setup you've come up with in one of your previous blog posts [0].

[0] https://www.jeffgeerling.com/blog/2024/use-external-gpu-on-r...

geerlingguy•18m ago
More on that soon... ;)
cosarara•51m ago
> Compared to the $8,000 Framework Cluster I benchmarked last month, this cluster is about 4 times faster:

Slower. 4 times slower.

teleforce•34m ago
That's definitely a typo because I've to read the sentence 3 times from the article still cannot make a sense until I saw the figure.

TL;DR, just buy one framework desktop and it's better than the Pi AI cluster of the OP in every single performance metrics including cost, performance, efficiency, headache, etc.

geerlingguy•28m ago
Oops, fixed the typo! Thanks.

And regarding efficiency, in CPU-bound tasks, the Pi cluster is slightly more efficient. (Even A76 cores on a 16nm node still do well there, depending on the code being run).

drillsteps5•47m ago
If he was building compute device for LLM inference specifically it would help to check in advance what that would entail. Like GPU requirement. Which putting bunch of RPis in the cluster doesn't help one bit.

Maybe I'm missing something.

deadbabe•46m ago
I really don’t understand the hype over raspberry Pi.

It’s an overrated, overhyped little computer. Like ok it’s small I guess but why is it the default that everyone wants to build something new on? Because it’s cheap? Whatever happened to buy once, cry once? Why not just build an actual powerful rig? For your NAS? For your firewalls? For security cameras? For your local AI agents?

theultdev•38m ago
I use mine for a plex server.

I don't need to transcode + I need something I can leave on that draws little power.

I have a powerful rig, but the one time I get to turn it off is when I'd need the media server lol.

There's a lot of scenarios where power usage comes into play.

These clusters don't make much sense to me though.

deadbabe•35m ago
That’s insane, drawing very little power from an always on server is a solved problem.
geerlingguy•19m ago
What's your idea of very little power, though?

I know for many who run SBCs (RK3588, Pi, etc.), very little is 1-2W idle, which is almost nothing (and doesn't even need a heatsink if you can stand some throttling from time to time).

Most of the Intel Mini PCs (which are about the same price, with a little more performance) idle at 4-6W, or more.

jonatron•33m ago
In the category of SBC's, it's pretty much the only one that has good software support, not outdated images made with a bunch of kernel patches for a specific kernel version.
hendersoon•39m ago
I mean, obviously it isn't practical, he got a couple of videos out of it.
nromiun•36m ago
There is a reason all the big supercomputers have started using GPUs in the last decade. They are much more efficient. If you want 32bit parallel performance just buy some consumer GPUs and hook them up. If you need 64bit buy some prosumer GPUs like the RTX 6000 Pro and you are done.

Nobody is really building CPU clusters these days.

deater•36m ago
as someone who has built various raspberry pi clusters over the years (I even got an academic paper out of one) the big shame is that as far as I know it's still virtually impossible to use the fairly powerful GPUs they have for GPGPU work
zamadatix•35m ago
The article focuses on compute performance but I wonder if that was ever the bottleneck considering the memory bandwidth involved.
amelius•33m ago
Ok, what are the back-of-the-envelope computations that he should have done before starting to build this?
geerlingguy•28m ago
Pi memory bandwidth is less than 10 GB/sec, so AI use cases will be extremely limited. Network I/O is maximum of 1 Gbps (or more if you do some unholy thing with M.2 NICs), so that also limits maximum networked performance.

But still can be decent for HPC learning, CI testing, or isolated multi-node smaller-app performance.

bearjaws•32m ago
Am I the only one who looks at both the Pi Cluster and the Framework PC and wonders how they are both slower and less cost effective than a MacBook Pro M4 Max? 88 token/s on a 2.3b model is not exactly great, most likely you will want a 32 or 70b model.
Drblessing•18m ago
The bee-link AI max+ is the best value AI pc right now.
stirfish•5m ago
I came here to ask about these. You like yours?
pluto_modadic•13m ago
His stances on a woman's body mean I don't care much what he spends money on.