frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Mini PC for local LLMs in 2026

https://terminalbytes.com/best-mini-pc-for-local-llm-2026/
27•charlieirish•1h ago

Comments

znpy•1h ago
As somebody that has a vague interest in running local LLMs… they day i decide to burn cash on hardware I might as well go all-in a get either a 128gb mac studio or an nvidia dgx spark (or some other equivalent gb10-based system).

The 64gb mac mini is also interesting, if anything because it is very likely to hold most of its value when reselling.

I’m keeping an eye on the next apple hardware refreshes, particularly for mac minis and mac studios.

amelius•1h ago
The models are good enough now, so I'm waiting for the day they start selling inference ASICs with 100x the token output speed. See Taalas demo.
adityamwagh•58m ago
Taalas is a nice concept, but I don’t want to use the same model forever!
amelius•49m ago
Just buy a new one every few years, just like your phone and laptop. And sell the old one.
2ndorderthought•1h ago
I just use my gaming pc. So I can play games or code with assistance for fun. It's awesome because it's mine and technically I can do whatever I want with it. Having a decent computer around and lower end laptops is pretty budget friendly.
walthamstow•1h ago
The 14inch Macbook Pros with 64GB are really good value considering it's a much more complicated machine than the Mini.
edot•49m ago
I am in a similar boat to you, but I can’t make the money math work. Local LLMs obviously have a privacy benefit but DeepSeek V4 Flash (which you’ll struggle to get running on any single Mac - you’d need at least 128gb RAM) is $0.14$/mtok input $0.28/mtok output on the API. You’d have to be just absolutely burning tokens to ever make this make sense.

Mac Studio M4 Max with 128gb at $3,699 (if you can find it) would equate to 10 million tokens a day of mixed input-output for over 5 years to break even. At which point that hardware is outdated compared to the SOTA models that will probably still be cheap on hosted platforms.

pjmlp•1h ago
Currently NVidia's mini PC, or the version licensed to Asus, is one of the few that I can actually buy with Linux pre-installed with a fully OEM supported version.

One would expect that by now buying desktop class computers on shops with a Linux experience would be rather common.

Geekcom devices that it advertises as Linux ready, are actually sold with Windows pre-installed.

I guess they mean WSL ready.

Neywiny•1h ago
I would guess they mean it's ready for you to install Linux on it
pjmlp•54m ago
Yeah, ignoring the whole fragmentation that keeps happening on the desktop stack, The Year of Desktop Linux will never happen if only computer nerds get to build such systems, as it has always been.

Instead normies get The Year of Linux kernel deployed with all kinds of consumer devices, and The Year of Linux VMs on retail.

alexktz•1h ago
Could we post articles that are obviously written by an LLM with a flair?
aalam•1h ago
"Here's the part that nobody talks about"

"Two gotchas before you click buy"

I really think there could be a score for entropy in playfulness that should differentiate LLM output

bachmeier•1h ago
"Local inference is rarely cheaper if you’re being honest with yourself about how much you actually use it."

Sorry, but this is not even close to "being honest", it's bad math. That calculation assumes you do nothing with the computer other than local inference.

hdgvhicv•58m ago
Doesnt that calculation assume you value your privacy and owmership at zero too?
spwa4•31m ago
Huh, you make me curious. Let's actually do that calculation. Let's say you do actually do 24/7/365 AI use. Let's say by some miracle you can do 60 t/s on Qwen 3.6 27b, and let's say this PC cost $3000 (you should be able to do this on a DGX spark, and one of the non-Nvidia models, e.g. the Dell one. $3000 would be a good price, but not totally out of the question). And, of course, let's say these prices remain stable.

So that gets you 1_892_160_000 tokens per year at full blast.

If you go the openrouter, eh, route, you'd get charged $2 per million tokens (anywhere from $2 to $3.6 per million tokens). So the value you'd get from your machine at 100% utilization is 1892 * $2 = $3784 up to 1892 * $3.6 = $6800)

So yeah, not counting electricity and your time the machine "is worth it".

[1] https://openrouter.ai/qwen/qwen3.6-27b/providers

bluechair•1h ago
“What’s the memory bandwidth (GB/s) of the device holding the model weights?”

Isn’t the recommended option going to be dog slow at 256 GB/s.

croes•1h ago
> 128GB Ryzen AI MAX+ 395, listed at $2,099.

Wasn‘t that a discounted price?

cowmix•57m ago
I got mine almost exactly a year ago - $1699 direct from GMKTEK. To think it retails for 2X that, a year later, blows my mind.
dannyw•1h ago
> The 256 GB/s number is real, but for context, an Apple M5 Ultra hits ~800 GB/s on its unified memory

The M5 Ultra has not been even announced.

This article appears to be predominately or entirely LLM-produced with little to no human review, and contains numerous material and misinforming errors.

It also omits serious contenders that's worth at least comparing, like the DGX Spark.

woadwarrior01•55m ago
It appears to be an LLM-generated affiliate link farm.
jcgrillo•59m ago
I got a well used HP Z840 with 256GB ECC DDR4 and twin Xeons ca. 2014. Then I slapped 2 AMD V640 32GB passively cooled GPUs in it with some 3D printed fan shrouds and 2 1U 15k rpm fans each. They just fit! I needed to order a quad 8pin power cable, the standard configuration has 3 6pin cables--but there's unused pins on the GPU power rail, and there are aftermarket suppliers.

72 Xeon cores

256GB ECC DDR4

64GB VRAM

$2200 total

I run it on a 20A 240V outlet to make sure the power supply can deliver enough watts, but so far it's working pretty well. The eWaste LLM rig is probably not as good value for money as a new machine, but it gets the job done cheaper (for now).

EDIT: IIRC this approach gets me more VRAM bandwidth than Strix Halo at the cost of less addressable GBs (but a lot more total system RAM), but I figured with CPU offloading that might make up for it?

ALSO EDIT: Note you can get a 128GB Strix Halo motherboard minus power supply, fans, case, etc from Framework for $2200.. that could work if you have some parts lying around.

lkey•56m ago
This article was authored by AI. It contains hallucinated info from compilations of random reddit threads.
visarga•53m ago
Yes, I too think it's authored by AI, but can you indicate where it is wrong?
mark_l_watson•55m ago
I bought a 32G MacMini over two years ago and it has been great for experimenting with local models, and now is even useful for local coding (at a slow speed!) with models supporting large context sizes.

With the current extreme RAM shortage I deeply regret not buying a 64G MacMini a few months ago.

I bet a zillion people feel the same way.

pjmlp•35m ago
Which is why the Mac Pro was actually relevant.

Those of us on PC land can at least extend them, or exchange the GPU, even if pricey.

Apple has lost the server and workstation market by their own decisions.

visarga•54m ago
Good research, but man do I feel the LLM vibe shining through. That sustained information density...
jcgrillo•43m ago
Look closer, it really isn't good research
jmyeet•53m ago
There's some mention of Apple silicon here but it's worth expanding upon. Macs have a unified memory architecture. So if you have a Mac with 64GB of memory then the GPU can use all of that. This is potentially quite useful but Apple silicon in general is limited by memory bandwidth. For comparison, a 5090 is 1792GB/s. Here are some examples:

- GMKTek EVO-X2: 120GB/s reads, 212GB/s writes

- NVidia DGX Spark 273GB/s

- Mac Mini M4 120GB/s but only $600+

- Mac Mini w/ M4 Pro 273GB/s ($2199 for 64GB)

- Mac Studio M4 Max 410GB/s ($3500 for 128GB)

- Mac Studio M3 Ultra 819GB/s ($5500 for 96GB)

- Macbook Pro 16" with M5 Pro 64GB 307GB/s ($3300)

- Macbook Pro 16" with M5 Max 128GB 460GB/s ($5399)

Sadly, Apple discontinued the 512GB Mac Studio. Mac Studios are a little long in the tooth now and due for an upgrade this year. I suspect that prices will be a lot higher given the RAM prices but we'll see.

Upcoming Blender Development Fund and AI Policies

https://www.blender.org/news/upcoming-blender-development-fund-and-ai-policies/
1•sensanaty•55s ago•0 comments

The Annoying Usefulness ofa Emacs [video]

https://www.youtube.com/watch?v=DMbrNhx2zWQ
1•susam•1m ago•0 comments

The Sky Tonight

https://theskylive.com/guide
1•susam•2m ago•0 comments

New US phone network for Christians to block porn and gender-related content

https://www.technologyreview.com/2026/05/01/1136739/a-new-t-mobile-network-for-christians-aims-to...
2•thinkingemote•5m ago•0 comments

Making Your Writing Work Harder for You

https://training.kalzumeus.com/newsletters/archive/content-marketing-strategy
2•eigenBasis•7m ago•0 comments

Show HN: TradingAgents without the API bill – run multi agents in Claude Code

https://github.com/lucemia/trading-agents-plugin
1•lucemia51•12m ago•0 comments

Stop Supplying. Start Owning

https://allensthoughts.com/2026/05/01/stop-supplying-start-owning/
2•herbertl•13m ago•0 comments

Uber wants to turn its drivers into a sensor grid for AV companies

https://techcrunch.com/2026/05/01/uber-wants-to-turn-its-millions-of-drivers-into-a-sensor-grid-f...
2•nickvec•13m ago•0 comments

Zugzwang

https://en.wikipedia.org/wiki/Zugzwang
5•Qem•17m ago•0 comments

If Claude writes the code, what makes me still a developer?

https://betweentheprompts.com/if-claude-writes-the-code/
2•scastiel•20m ago•0 comments

Santa Cruz restaurant changes logo after flurry of negative reviews for AI art

https://www.sfgate.com/food/article/santa-cruz-restaurant-ai-21955920.php
3•randycupertino•21m ago•0 comments

LLMs consistently pick resumes they generate over ones by humans or other models

https://arxiv.org/abs/2509.00462
34•laurex•24m ago•10 comments

Domination: A contrarian view of AI risk (2024)

https://matthewbutterick.com/chron/domination.html
2•vermilingua•33m ago•0 comments

I moved my blog from Jekyll to Emacs Lisp

https://martinsos.com/posts/my-blog-in-elisp
2•Martinsos•35m ago•1 comments

The History of Lipstick

https://www.saturdayeveningpost.com/2026/04/common-threads-the-history-of-lipstick/
2•ohjeez•35m ago•0 comments

Alberta allows windfall oil and gas payments to select ranchers – on public land

https://thenarwhal.ca/alberta-grazing-oil/
3•Teever•37m ago•0 comments

US blockade costs Iran $4.8B, US Navy acting 'sort of like pirates,' Trump says

https://www.jpost.com/middle-east/iran-news/article-894867
2•Levitating•39m ago•3 comments

A preliminary model to establish a digital twin for coffee roasting

https://www.nature.com/articles/s41598-026-43923-9?fromPaywallRec=false
3•bookofjoe•39m ago•0 comments

Show HN: RegularMonk – a web app that helps me use my phone less

https://www.regularmonk.com/hello
1•amit9968•40m ago•0 comments

Apple Faces Lawsuits over AirTag Stalking After Class Action Denied

https://www.macrumors.com/2026/05/01/airtag-stalking-lawsuits-apple/
1•mgh2•40m ago•1 comments

Make Common Sense Common Again

https://nik.art/make-common-sense-common-again/
1•herbertl•42m ago•0 comments

Stackless coroutines for gamedev in ~200 lines of C++

https://vittorioromeo.com/index/blog/sfex_coroutine.html
2•tzury•42m ago•0 comments

Proudly Pathetic

https://craigatallahfrost.com/post/2025/08/17/proudly-pathetic/
1•herbertl•44m ago•0 comments

NASA to increase CLPS contract to support surge of lunar lander missions

https://spacenews.com/nasa-to-increase-value-of-clps-contract-to-support-surge-of-lunar-lander-mi...
3•rbanffy•45m ago•0 comments

America's Expanding Domestic Surveillance

https://www.wsj.com/articles/americas-expanding-domestic-surveillance-08b73187
7•Brajeshwar•50m ago•0 comments

The Fake Hawaii CTO Who Fooled Everyone

https://dallasexpress.com/national/from-vegas-stages-to-official-warnings-the-fake-hawaii-cto-who...
1•greenchair•50m ago•0 comments

Apple Stores Targeted in $16.2M Counterfeit Device Scheme

https://pasadenanow.com/main/pasadena-apple-store-among-locations-targeted-in-16-2-million-counte...
1•kid64•50m ago•0 comments

Docker vs. Podman: Which Containerization Tool Is Right for You – DataCamp

https://www.datacamp.com/blog/docker-vs-podman
1•abdelhousni•54m ago•1 comments

Ask HN:Do people configure Claude Code to use other models

https://openrouter.ai/apps/claude-code
1•ripvanwinkle•54m ago•6 comments

LibreLocal 2026 – Global Meetups Across Six Continents

https://tux.re/forum/viewtopic.php?t=217
1•tuxyz•54m ago•0 comments