Their MI300s already beat them, 400s coming soon.
Know any LLMs that are implemented in CUDA?
Show me one single CUDA kernel on Llama's source code.
Note: previous testing I did was on a single (8x) MI300X node, currently I'm doing testing on just a single MI300X GPU, so not quite apples-to-apples, multi-GPU/multi-node training is still a question mark, just a single data point.
OTOH, by emphasizing datacenter hardware, they can cover a relatively small portfolio and maximize access to it via cloud providers.
As much as I'd love to see an entry-level MI350-A workstation, that's not something that will likely happen.
And as of 2025, there is a Python CUDA JIT DSL as well.
Also, even if not the very latest version, the fact that CUDA SDK works on any consumer laptop with NVidia hardware, anyone can slowly get into CUDA, even if their hardware isn't that great.
They should also be dropping free AMD GPUs off helicopters, as Nvidia did a decade or so ago, in order to build up an academic userbase. Academia is getting totally squeezed by industry when it comes to AI compute. We're mostly running on hardware that's 2 or 3 generations out of date. If AMD came with a well supported GPU that cost half what an A100 sells for, voila you'd have cohort after cohort of grad students training models on AMD and then taking that know-how into industry.
Which SDKs do they offer that can do neural network inference and/or training? I'm just asking because I looked into this a while ago and felt a bit overwhelmed by the number of options. It feels like AMD is trying many things at the same time, and I’m not sure where they’re going with all of it.
Plus their consumer card support is questionable to say the least. I really wish it was a viable alternative, but swapping to CUDA really saved me some headaches and a ton or time.
Having to run MiOpen benchmarks for HIP can take forever.
NVIDIA has a moat for smaller systems, but that is not true for clusters.
As long as you have a team to work with the hardware you have, performance beats mindshare.
halJordan•14h ago
AMD deserves exactly zero of the credulity this writer heaps onto them. They just spent four months not supporting their rdna4 lineup in rocm after launch. AMD is functionally capable of day120 support. None of the benchmarks disambiguated where the performance is coming from. 100% they are lying on some level, representing their fp4 performance against fp 8/16.
pclmulqdq•14h ago
caycep•14h ago
fooblaster•14h ago
fc417fc802•13h ago
zombiwoof•11h ago
tormeh•4h ago
viewtransform•13h ago
"25 complimentary GPU hours (approximately $50 US of credit for a single MI300X GPU instance), available for 10 days. If you need additional hours, we've made it easy to request additional credits."
archerx•12h ago
zombiwoof•11h ago
archerx•3h ago
booder1•12h ago
They should care about the availability of their hardware so large customers don't have to find and fix their bugs. Let consumers do that...
echelon•12h ago
Makes it a little hard to develop for without consumer GPU support...
stingraycharles•12h ago
cma•12h ago
wmf•11h ago
selectodude•11h ago
jiggawatts•10h ago
stingraycharles•11h ago
This is why it’s so important AMD gets their act together quickly, as the benefits of these kind of things are measured in years, not months.
danielheath•9h ago
moffkalast•9h ago
lhl•8h ago
Flash Attention: academia, 2y behind for AMD support
bitsandbytes: academia, 2y behind for AMD support
Marlin: academia, no AMD support
FlashInfer: acadedmia/startup, no AMD
ThunderKittens: academia, no AMD support
DeepGEMM, DeepEP, FlashMLA: ofc, nothing from China supports AMD
Without the long tail AMD will continue to always be in a position where they have to scramble to try to add second tier support years later themselves, while Nvidia continues to get all the latest and greatest for free.
This is just off the top of my head on the LLM side where I'm focused on, btw. Whenever I look at image/video it's even more grim.
jimmySixDOF•8h ago
pjmlp•8m ago
littlestymaar•8h ago
I mean of they want to stay at a fraction of the market value and profit of their direct competitor, good for them.
dummydummy1234•5h ago
It's Nvidia, AMD, and maybe Intel.
shmerl•12h ago
Different architectures was probably a big reason for the above issue.
fooker•10h ago
pjmlp•9m ago
jchw•12h ago
It's baffling that AMD is the same company that makes both Ryzen and Radeon, but the year-to-date for Radeon has been very good, aside from the official ROCm support for RDNA4 taking far too long. I wouldn't get overly optimistic; even if AMD finally committed hard to ROCm and Radeon it doesn't mean they'll be able to compete effectively against NVIDIA, but the consumer showing wasn't so bad so far with the 9070 XT and FSR4, so I'm cautiously optimistic they've decided to try to miss some opportunities to miss opportunities. Let's see how long these promises last... Maybe longer than a Threadripper socket, if we're lucky :)
[1]: https://www.phoronix.com/news/AMD-ROCm-H2-2025
roenxi•11h ago
I dunno; I suppose they can execute on server parts. But regardless, a good plan here is to let someone else go first and report back.
jchw•5h ago
zombiwoof•11h ago
AMD is a marketing company now
ethbr1•7h ago
You mean Ryan Smith of late AnandTech fame?
https://www.anandtech.com/author/85/