frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The Deletion of Docker.io/Bitnami

https://community.broadcom.com/tanzu/blogs/beltran-rueda-borrego/2025/08/18/how-to-prepare-for-th...
63•zdkaster•1h ago•20 comments

Altered states of consciousness induced by breathwork accompanied by music

https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0329411
225•gnabgib•5h ago•86 comments

Bookmarks.txt is a concept of keeping URLs in plain text files

https://github.com/soulim/bookmarks.txt
48•secwang•3h ago•28 comments

Canaries in the Coal Mine? Recent Employment Effects of AI [pdf]

https://digitaleconomy.stanford.edu/wp-content/uploads/2025/08/Canaries_BrynjolfssonChandarChen.pdf
41•p1esk•3h ago•27 comments

Yamanot.es: A music box of train station melodies from the JR Yamanote Line

https://yamanot.es/
192•zdw•9h ago•56 comments

Sci-Hub has been blocked in India

https://sci-hub.se/sci-hub-blocked-india
57•the-mitr•1h ago•9 comments

Malicious versions of Nx and some supporting plugins were published

https://github.com/nrwl/nx/security/advisories/GHSA-cxm3-wv7p-598c
349•longcat•1d ago•382 comments

Nvidia DGX Spark

https://www.nvidia.com/en-us/products/workstations/dgx-spark/
86•janandonly•3d ago•88 comments

Toyota is recycling old EV batteries to help power Mazda's production line

https://www.thedrive.com/news/toyota-is-recycling-old-ev-batteries-to-help-power-mazdas-productio...
228•computerliker•4d ago•104 comments

Launch HN: Bitrig (YC S25) – Build Swift apps on your iPhone

124•kylemacomber•14h ago•89 comments

Unexpected productivity boost of Rust

https://lubeno.dev/blog/rusts-productivity-curve
352•bkolobara•14h ago•318 comments

Will Bardenwerper on Baseball's Betrayal of Its Minor League Roots

https://lithub.com/will-bardenwerper-on-baseballs-betrayal-of-its-minor-league-roots/
12•PaulHoule•2d ago•1 comments

Google has eliminated 35% of managers overseeing small teams in past year

https://www.cnbc.com/2025/08/27/google-executive-says-company-has-cut-a-third-of-its-managers.html
381•frays•8h ago•168 comments

VIM Master

https://github.com/renzorlive/vimmaster
253•Fluffyrnz•14h ago•87 comments

Researchers find evidence of ChatGPT buzzwords turning up in everyday speech

https://news.fsu.edu/news/education-society/2025/08/26/on-screen-and-now-irl-fsu-researchers-find...
140•giuliomagnifico•8h ago•219 comments

Show HN: Meetup.com and eventribe alternative to small groups

https://github.com/polaroi8d/cactoide
76•orbanlevi•9h ago•36 comments

The GitHub website is slow on Safari

https://github.com/orgs/community/discussions/170758
331•talboren•20h ago•246 comments

GMP damaging Zen 5 CPUs?

https://gmplib.org/gmp-zen5
179•sequin•13h ago•146 comments

What is this? The case for continually questioning our online experience

https://systems-souls-society.com/what-is-this-the-case-for-continually-questioning-our-online-ex...
3•Gigamouse•2d ago•0 comments

The Therac-25 Incident (2021)

https://thedailywtf.com/articles/the-therac-25-incident
418•lemper•23h ago•253 comments

Certificates for Onion Services

https://onionservices.torproject.org/research/proposals/usability/certificates/
8•keepamovin•3h ago•0 comments

On the screen, Libyans learned about everything but themselves (2021)

https://newlinesmag.com/argument/on-the-screen-libyans-learned-about-everything-but-themselves/
19•thomassmith65•2d ago•1 comments

Object-oriented design patterns in C and kernel development

https://oshub.org/projects/retros-32/posts/object-oriented-design-patterns-in-osdev
214•joexbayer•1d ago•140 comments

Beginning 1 September, we will need to geoblock Mississippi IPs

https://dw-news.dreamwidth.org/44429.html
186•AndrewDucker•10h ago•221 comments

Areal, Are.na's new typeface

https://www.are.na/editorial/introducing-areal-are-nas-new-typeface
120•g0xA52A2A•2d ago•79 comments

About Containers and VMs

https://linuxcontainers.org/incus/docs/main/explanation/containers_and_vms/
65•Bogdanp•2d ago•44 comments

A failure of security systems at PayPal is causing concern for German banks

https://www.nordbayern.de/news-in-english/paypal-security-systems-down-german-banks-block-payment...
228•tietjens•12h ago•160 comments

Implementing Forth in Go and C

https://eli.thegreenplace.net/2025/implementing-forth-in-go-and-c/
144•Bogdanp•16h ago•20 comments

Using information theory to solve Mastermind

https://www.goranssongaspar.com/mastermind
99•SchwKatze•4d ago•32 comments

Lago – Open-Source Usage Based Billing – Is Hiring in Sales, Eng, Ops (EU, US)

https://www.ycombinator.com/companies/lago/jobs
1•AnhTho_FR•13h ago
Open in hackernews

Nvidia DGX Spark

https://www.nvidia.com/en-us/products/workstations/dgx-spark/
86•janandonly•3d ago

Comments

oracel•5h ago
Can it run Crysis?
brookst•5h ago
It can write Crysis.
mynegation•5h ago
It’s gonna run Crysis for my wallet alright
bigyabai•5h ago
Bit of a non-sequitur, but it seems like it's been playable via box86 for quite a while now: https://youtu.be/NcBJG3z8kF0
lvl155•5h ago
Is this worth getting vs AMD?
zxexz•5h ago
What are you trying to do?
DoctorOetker•5h ago
suppose 1/3rd of memory is used to host a teacher network, and 2/3rds of memory is used to host a student network, how long would knowledge distillation typically take?
nightski•5h ago
Am I missing something or does the comparably priced (technically cheaper) Jetson Thor have double the PFLOPs of the Spark with the same memory capacity and similar bandwidth?
modeless•5h ago
Also Thor is actually getting sent out to robotics companies already. Did anyone outside Nvidia get a DGX Spark yet?
Apes•5h ago
My understanding is the DGX Spark is optimized for training / fine tuning and the Jetson Thor is optimized for running inference.

Architecturally, the DGX Spark has a far better cache setup to feed the GPU, and offers NVLINK support.

AlotOfReading•4h ago
There's a lot of segmentation going on in the Blackwell generation from what I'm told.
Y_Y•5h ago
It's a bit disingenuous to claim 1 PFLOPs without making clear that's for FP4 (with "structured sparsity"?)
csunoser•5h ago
It does say `Experience up to 1 petaFLOP of AI performance at FP4 precision with the NVIDIA Grace Blackwell architecture.` in the features section.

But yeah, this should have been further up.

godelski•4h ago
If you scroll down a little and see the chip icon, where it says "NVIDIA GB10 Superchip " it also says "Experience up to 1 petaFLOP of AI performance at FP4 precision with the NVIDIA Grace Blackwell architecture."

Further down, in the exploded view it says "Blackwell GPU 1PetaFLOP FP4 AI Compute"

Then further down in the spec chart they get less specific again with "Tensor Performance^1 1 PFLOP" and "^1" says "1 Theoretical FP4 TOPS using the sparsity feature."

Also, if you click "Reserve Now" the second line below that redundant "Reserve Now" button says "1 PFLOPS of FP4 AI performance"

I mean I'll give you that they could be more clear and that it's not cool to just hype up on FP4 performance, but they aren't exactly hiding the context like they did during GTC. I wouldn't call this "disingenuous"

Y_Y•3h ago
Even if that "sparsity feature" is that two or of every four adjacent values in your areay be zeros, and that performance halves if not doing this?

I think lots of children are going to be very disappointed running their blas benchmarks on Christmas morning and seeing barely tens of teraflops.

(For reference see how the still optimistic numbers are for the H200 when you use realistic datatypes.

https://nvdam.widen.net/s/nb5zzzsjdf/hpc-datasheet-sc23-h200... )

MBCook•5h ago
I’m not in this space, so I don’t know what’s normal, but I guess I’m a little surprised to see only 10 gig Ethernet for high speed connectivity.

Yeah, it’s miles better than WiFi. But if there was something I’d think maybe benefit from Thunderbolt this would’ve been it.

The ability to transfer large models or datasets that way just seems like it would be much faster and a real win for some customers.

coder543•5h ago
This thing has a ConnectX-7, which gives it 2 x 200 Gbps networking. The 10 gig port is far from the fastest network interface on the Spark.
MBCook•3h ago
But can you hook that up to a normal PC?
coder543•3h ago
You were complaining about speed. Yes, a PC can have the same ports, and then you get much faster speeds than Thunderbolt can provide.

Why would you ever want a DGX Spark to talk to a “normal PC” at 40+ Gbps speeds anyways? The normal PC has nothing that interesting to share with it.

But, yes, the DGX Spark does have four USB4 ports which support 40Gbps each, the same as Thunderbolt 4. I still don’t see any use case for connecting one of those to a normal PC.

renewiltord•41m ago
Yes. Just buy the Mellanox card. We had a bunch of ConnectX 5 hooked up through SFP. Needs cooling but fast.
x2tyfi•4h ago
You’re almost always going to bottleneck on your home internet or upstream ISP, rather than this local interface. That being said, you aren’t going to be waiting too long either way, depending on download speed. Deepseek R1 is 671GB. Multiply by 8 to get into bits: 5368Gb At full 10gbps (which, again, you probably won’t get): 5368Gb / 10gbps = 537 seconds to download 537s / 60 = 8.95 minutes. Call it 10m with overhead.
rightisleft•2h ago
I think I interviewed you the other day and you didn’t get the job…
wiredpancake•1h ago
What?
ls612•5h ago
Is this the $3500 one?
x2tyfi•4h ago
That was their Digits box.
jrgifford•4h ago
Digits is no more, it’s DGX. Source: signup link in the press release (https://nvidianews.nvidia.com/news/nvidia-puts-grace-blackwe...) now goes to DGX Spark preorder (https://www.nvidia.com/en-us/products/workstations/dgx-spark...)
wmf•4h ago
It's the same thing. They renamed it from Digits to Spark.
gardnr•3h ago
There are cheaper ASUS and MSI versions "coming soon" with the same chip and less storage / memory.
canucker2016•2h ago
on the reserve now page for USA, there's:

  ASUS Ascent GX10 - 1TB $2,999

  MSI EdgeXpert MS-C931 - 4TB $3,999
the 1TB/4TB seems to be the size of the included NVMe SSD.

the reserve now page also lists

  NVIDIA DGX Spark Bundle
  2 NVIDIA DGX Spark Units - 4TB with Connecting Cable $8,049
The DGX Spark specs lists an NVIDIA ConnectX-7 Smart NIC which is rated at 200Gbe to connect to another DGX Spark, for about double the amount of memory for models.
dirtyhand•5h ago
I was considering getting an RTX 5090 to run inference on some LLM models, but now I’m wondering if it’s worth paying an extra $2K for this option instead
BoorishBears•4h ago
No. These are practically useless for AI.

Their prompt processing speeds are absolutely abysmal: if you're trying to tinker from time to time, a GPU like a 5090 or renting GPUs is a much better option.

If you're just trying to prep for impending mainstream AI applications, few will be targeting this form factor: it's both too strong compared to mainstream hardware, and way too weak compared to dedicated AI-focused accelerators.

-

I'll admit I'm taking a less nuanced take than some would prefer, but I'm also trying to be direct: this is not ever going to be a better option than a 5090.

aurareturn•4h ago

  Their prompt processing speeds are absolutely abysmal
They are not. This is Blackwell with Tensor cores. Bandwidth is the problem here.
BoorishBears•4h ago
They're abysmal compared to anything dedicated at any reasonable batch size because of both bandwidth and compute, not sure why you're wording this like it disagrees with what I said.

I've run inference workloads on a GH200 which is an entire H100 attached to an ARM processor and the moment offloading is involved speeds tank to Mac Mini-like speeds, which is similarly mostly a toy when it comes to AI.

aurareturn•4h ago
Again, prompt processing isn't the major problem here. It's bandwidth. 256GB/s bandwidth (maybe ~210 in real world) limits the tokens per second well before prompt processing.

Not entirely sure how your ARM statement matters here. This is unified memory.

BoorishBears•3h ago
"GH200 which is an entire H100 attached to an ARM processor"

What do you think attached means here? It's unified memory.

I'm telling you empirically, at any none single batch size, an H100 with unified memory prompt processing dips to double digits despite having significantly more compute and bandwidth, the moment a model is large enough to require offloading: aka require dipping into the unified portion of it's memory.

Even at bs=1 the performance is pretty absymal, I don't have the benchmarks handy anymore but not even 50% of the performance before offloading.

You can't debate me on this, this is the reality. That is why you can rent a GH200 for the same price of as an H100 right now, if not cheaper: nobody wants them.

-

"NoT EntIrelY SuRe"... sometimes I forget how absolutely exhausting it is trying to speak to a jackass who thinks they know so much better than you that they can't even begin to try to process what you're saying before replying.

And there's something about the HN-fake-niceness that's attached to it that really gets my fucking goat.

Y_Y•3h ago
I like the cut of your jib and your experience matches mine, but without real numbers this is all just piss in the wind (as far as online discussions go).
BoorishBears•2h ago
You're right, it's unfortunate I didn't keep the benchmarks around: I benchmark a lot of configurations and providers for my site and have a script I typically run that produces graphs for various batch sizes (https://ibb.co/0RZ78hMc)

The performance with offloading was just so bad I didn't even bother proceeding to the benchmark (without offloading you get typical H100 speeds)

Apes•4h ago
RTX 5090 is about as good as it gets for home use. Its inference speeds are extremely fast.

The limiting factor is going to be the VRAM on the 5090, but nvidia intentionally makes trying to break the 32GB barrier extremely painful - they want companies to buy their $20,000 GPUs to run inference for larger models.

apitman•3h ago
If you want to run small models fast get the 5090. If you want to run large models slow get the Spark. If you want to run small models slow get a used MI50. If you want to run large models fast get a lot more money.
skhameneh•3h ago
RTX 5090 for running smaller models.

Then the RTX Pro 6000 for running a little bit larger models (96gb VRAM, but only ~15-20% more perf than 5090).

Some suggest Apple Silicon only for running larger models on a budget because of the unified memory, but the performance won't compare.

senectus1•5h ago
Power consumption : TBD

?? this seems more than a little disingenuous...

canucker2016•2h ago
from https://www.servethehome.com/this-is-the-asus-ascent-gx10-a-...

  ASUS and NVIDIA told us that their GB10 platforms are expected to use up to 170W.
[edit] the PSU is 240W so that'd place an upper limit on power draw, unless they upgrade it.
wewewedxfgdf•4h ago
It'll be stunted in some way - Nvidia always holds back some crucial feature that you need, to push you up to the next highest priced product line.
agnokapathetic•4h ago
it uses LPDDR5x instead of the datacenter variant’s HBM3e.
syntaxing•4h ago
While a completely different price point, I have a Jetson Orin Nano. Some people forget the kernels are more or less set in stone for product like these. I could rebuild my own Jetpack kernel but it’s not that straight forward to update something like CUDA or any other module. Unless you’re a business where your product relies on this hardware, I find it hard to buy this for consumer applications.
coredog64•3h ago
Came in here to say the same thing. Have bought 3 Nvidia dev boards and never again as you quickly get left behind. You're then stuck compiling everything from scratch.
larodi•1h ago
My experience with Jetson Nano was that it had to have its Ubuntu debloatred first (with 3rd party script) before we could get their NN something library to run the image recognition, designated to run on this device.

These seem to be highly experimental boards, even though are super powerful for their form factor.

cherioo•4h ago
The mainstream options seem to be

Ryzen AI Max 395+, ~120 tops (fp8?), 128GB RAM, $1999

Nvidia DGX Spark, ~1000 tops fp4, 128GB RAM, $3999

Mac Studio max spec, ~120 tflops (fp16?), 512GB RAM, 3x bandwidth, $9499

DGX Spark appears to potentially offer the most token per second, but less useful/value as everyday pc.

aurareturn•4h ago

  Mac Studio max spec, ~120 tflops (fp16?), 384GB RAM, 3x bandwidth, $9499
512GB.

DGX has 256GB/s bandwidth so it wouldn't offer the most tokens/s.

rz2k•4h ago
Perhaps they are referring to default GPU allocation that is 75% of the unified memory, but it is trivial to increase it.
jauntywundrkind•4h ago
The GPU memory allocation refers to how capacity is alloted, not bandwidth. Sounds like the same 256-bit/quad-channel 8000MHz lpddr5 you can get today with Strix Halo.
rz2k•3h ago
384GB is 75% of 512GB. The M3 Ultra bandwidth is over 800GB/s, though potentially less in practice.

Using an M3 Ultra I think the performance is pretty remarkable for inference and concerns about prompt processing being slow in particular are greatly exaggerated.

Maybe the advantage of the DGX Spark will be for training or fine tuning.

echelon•4h ago
tokens/s/$ then.
jauntywundrkind•4h ago
NVidia Spark is $4000. Or, will be, supposedly whenever it comes out.

Also notably, Strix Halo and DGX Spark are both ~275GBps memory bandwidth. Not always but in many machine learning cases it feels like that's going to be the limiting factor.

UncleOxidant•1h ago
> Ryzen AI Max 395+, ~120 tops (fp8?), 128GB RAM, $1999

Just got my Framework PC last week. It's easy to setup to run LLMs locally - you have to use Fedora 42, though, because it has the latest drivers. It was super easy to get qwen3-coder-30b (8 bit quant) running in LMStudio at 36 tok/sec.

hasperdi•8m ago
Hi could you share if you get a decent coding performance (quality wise) with this setup? IE. Is it good enough to replace say Claude Code?
rjzzleep•51m ago
Maybe the real value of the DGX spark is to work on Switch 2 emulation. ARM + Nvidia GPU. Start with Switch 2 emulation on this machine and then optimize for others. (Yeah, I know, kind of expensive toy).
ComplexSystems•4h ago
The RAM bandwidth is so slow on this that you can barely train or do inference or do anything on it. I think the only use case they have in mind for this is fine tuning pretrained models.
wmf•4h ago
It's the same as Strix Halo and M4 Max that people are going gaga about, so either everyone is wrong or it's fine.
7thpower•4h ago
The other ones are not framed as an “AI Supercomputer on your desk”, but instead are framed as powerful computers that can also handle AI workloads.
aurareturn•4h ago
M4 max has more than double the bandwidth.

Strix Halo has the same and I agree it’s overrated.

Rohansi•3h ago
I would expect/hope that DGX would be able to make better use of its bandwidth than the M4 Max. Will need to wait and see benchmarks.
gardnr•3h ago
Memory Bandwidth:

Nvidia DGX: 273 GB/s

M4 Max: (up to) 546 GB/s

M3 Ultra: 819 GB/s

RTX 5090: ~1.8 TB/s

RTX PRO 6000 Blackwell: ~1.8 TB/s

hereme888•4h ago
FP4-sparse (TFLOPS) | Price | $/TF4s

5090: 3352 | 1999 | 0.60

Thor: 2070 | 3499 | 1.69

Spark: 1000 | 3999 | 4.00

____________

FP8-dense (TFLOPS) | Price | $/TF8d (4090s have no FP4)

4090 : 661 | 1599 | 2.42

4090 Laptop: 343 | vary | -

____________

Geekbench 6 (compute score) | Price | $/100k

4090: 317800 | 1599 | 503

5090: 387800 | 1999 | 516

M4 Max: 180700 | 1999 | 1106

M3 Ultra: 259700 | 3999 | 1540

____________

Apple NPU TOPS (not GPU-comparable)

M4 Max: 38

M3 Ultra: 36

aurareturn•4h ago
It's not good value when you put it like that. It doesn't have a lot of compute and bandwidth. What it has is the ability to run DGX software for CUDA devs I guess. Not a great inference machine either.
Y_Y•3h ago
You are doing god's work.

In fact you're also doing the work Nvidia should have done when they put together their (imho) ridiculously imprecise spec sheet.

conradev•3h ago
where does an RTX Pro 6000 Blackwell fall in this? I feel like that’s the next step up in performance (and about the same price as two Sparks)
qingcharles•2h ago
I thought the 6000 was slightly lower throughput than 5090, but obviously has a shitload more RAM.
skhameneh•1h ago
It's more throughput, but way less value and there's still no NVLink on the 6000. Something like ~4x the price, ~20% more performance, 3x the VRAM.

There's two models that go by 6000, the RTX Pro 6000 (Blackwell) is the one that's currently relevant.

scosman•3h ago
How does the process management comparison work for GPU vs full systems?
nodesocket•3h ago
Once the updated Mac Studio with M4/M5 Ultra comes out, pretty much going to make the DGX irrelevant right?
wmf•3h ago
Ultras are pretty expensive.
nodesocket•2h ago
I mean the spark is $3,999 and current M3 Max 28-Core CPU 60-Core GPU is the same price. I would expect the refreshed studio will stay around the same price.
KingOfCoders•1h ago
In Germany the 96gb version is 5000 EUR and the 256gb version is 7000 EUR (no 128gb available as far as I can see).
canucker2016•3h ago
5090: 32GB RAM (newegg & amazon lowest price seems to be +300 more)

4090: 24GB RAM

Thor & Spark: 128GB RAM (probably at least 96GB usable by the GPU if they behave similar to the AMD Strix Halo APU)

garyfirestorm•3h ago
What did I miss? This was revealed in May - I don’t see anything new in that link since it was revealed.
wmf•3h ago
Not much. There was a presentation yesterday but it's mostly what we already knew: https://www.servethehome.com/nvidia-outlines-gb10-soc-archit...
eadwu•3h ago
Most people are missing the point. LLMs are not the be all end all of AI.

Even if you were to say memory bandwidth was the problem, there is no consumer grade GPU that can run any SoTA LLM, no matter what you'd have to settle for a more mediocre model.

Outside of LLMs, 256 GB/s is not as much of an issue and many people have dealt with less bandwidth for real world use cases.

gardnr•3h ago
What other use cases would use 128GB VRAM but not require higher throughput to run at acceptable speeds?
sorrythanks•3h ago
NVIDIA DGX Spark - 4TB

$3,999

maz1b•2h ago
Dunno, doesn't seem that good to me. Granted, I recognize the pace of advancement, but fwiw at present time.. yeah.

I'd rather just get an M3 Ultra. Have an M2 Ultra on the desk, and an M3 Ultra sitting on the desk waiting to be opened. Might need to sell it and shell out the cash for the max ram option. Pricey, but seems worthwhile.

DrNosferatu•2h ago
Now we need a threeway benchmark between this DGX Spark, a maxed out AMD Strix* and the Mac 512GB.
maddynator•2h ago
So Raspberry Pi With GPU?
KingOfCoders•1h ago
I think it depends on your model size

   Fits into 32gb: 5090
   Fits into 64gb - 96gb: Mac Studio
   Fits into 128gb: for now 395+ $/token/s, 
     Mac Studio if you don't care about $ 
     but don't have unlimited money for Hxxx
This could be great for models that fit 128gb and you want best $/token/s (if it is faster than a 395+).
monster_truck•1h ago
Paper launch. The people I know there who I have asked about it haven't even seen one yet
fh973•48m ago
Ordered one in spring. Delivery time was pushed from July to September. Apparently they had a bug in the HDMI output.
wtallis•10m ago
That's eerily similar to what happened to Qualcomm's failed Snapdragon X Elite dev kit. That one eventually shipped in small quantities with a Type-C to HDMI dongle in the box to make up for the built-in HDMI port going missing. Then Qualcomm cancelled the whole project and refunded everyone, including people who had already received their hardware.
numpad0•38m ago
Do anyone know why official pages don't mention FP16 performance(250 TFLOPS)?