The summary is that it's a cache attached to the memory controllers, rather than the CPUs, so it doesn't have to worry about cache coherency so much. This could be useful for shared memory parallelism.
Because it is last level before the actual ram chips, no coherency involved.
The Super Mushroom power-up.
The main advantage of a memory attached cache is that it's cheaper than a regular cache, and can even be put on a seperate die, allowing you to have much more of it.
AMDs previous memory fabric from the early 2000s was called "Hyper Transport", which has a confusing overlap with Intel's Hyper Threading, but I think AMD actually bet intel to the name by a few years.
I think AMD might be ditching it for the next gen in exchange for growing the L2's, because the lower latency is beneficial to ray tracing.
[0]: https://substackcdn.com/image/fetch/$s_!LHJo!,f_auto,q_auto:...
andrewstuart•9h ago
Why would AMD not have focused everything it possibly has on demonstrating and documenting and fixing and showing and smoothing the path for AI on their systems?
Why does AMD come across as so generally clueless when it comes to giving developers what they want, compared to Nvidia?
AMD should do whatever it takes to avoid these sort of situations:
https://youtu.be/cF4fx4T3Voc?si=wVmYmWVIya4DQ8Ut
typpilol•9h ago
Just general compatibility between Nvidia and AMD for stuff that was built for Nvidia originally?
Or do you mean something else?
cakealert•8h ago
Some UX-oriented tooling has sort of solved this problem and will run on AMD: LM Studio
pella•9h ago
andrewstuart•8h ago
YuukiRey•8h ago
Not exactly a gigantic mental leap.
spockz•7h ago
storus•1h ago
ondra•1h ago
vid•36m ago
lmm•8h ago
DeepYogurt•8h ago
AnthonyMouse•5h ago
The customers of hardware companies generally don't want to get proprietary software from them, because everybody knows that if they do, the hardware company will try to use it as a lock-in strategy. So if you make something which is proprietary but not amazing, nobody wants to touch it.
There are two ways around this.
The first is that you embrace open source. This is what Intel has traditionally done and it works really well when your hardware is good enough that it's what people will choose when the software is a commodity. It also means you don't have to do all the work yourself because if you're not trying to lock people in then all the community work that normally goes to trying to reverse engineer proprietary nonsense instead goes into making sure that the open source software that runs on your hardware is better than the proprietary software that runs on your competitor's.
The second is that you spend enough money on lock-in software that people are willing to use it. This works temporarily, because it takes a certain amount of time for competitors and the community to make a decent alternative, but what usually happens after that is that you have a problem because you were expecting there to be a moat and then ten thousand people showed up to each throw in a log or a bag of wet cement. Before too long the moat is filled in and you can't it back because it was premised on your thing working and their thing not, so once their thing works, that's the part that isn't under your control. And at that point the customers have a choice and the customers don't like you.
The problem AMD has is that they were kinda sorta trying to do both in GPUs. They'd make some things open source but also keep trying to hide the inner workings of the firmware from the public, which is what people need in order to allow third parties to make great software for them. But the second strategy was never going to work for AMD because a decade ago they didn't have the resources to even try and now Nvidia is the incumbent and the underdog can't execute a lock-in strategy. But the open source thing works fine here and indeed gets everyone on their side and then it's them and the whole world against Nvidia instead of just them against Nvidia. Which they're gradually starting to figure out.
rjsw•1h ago
z3ratul163071•5h ago
i know, i know we have s...fest sw layers on other chips like the ones from qualcomm, broadcom etc.
ur-whale•3h ago
They are absolutely not.
Just less bad than AMD, which is an extremely low bar.
naasking•39m ago
Holy exaggeration Batman!
oblio•2m ago
sidkshatriya•8h ago
I have some theories. Firstly, Nvidia was smart enough to have a unified compute GPU architecture across all its architectures -- consumer and commercial. AMD has this awkward split between CDNA and RDNA. So while AMD is scrambling to get CDNA competitive, RDNA is not getting as much attention as it should. I'm pretty sure its ROCm stack has all kinds of hacks trying to get things working across consumer Radeon devices (which internally are probably not well suited/tuned for compute anyways). AMD is hamstrung by its consumer hardware for now in the AI space.
Secondly, AMD is trying to be "compatible" to Nvidia (via HIP). Sadly this is the same thing that AMD did with Intel in the past. Being compatible is really a bad idea when the market leader (Nvidia) is not interested in standardising and actively pursues optimisations and extensions. AMD will always play catch up.
TL;DR AMD made some bad bets on what the hardware would look like in the future and never thought software was critical like nvidia.
AMD now realizes that software is critical and what future hardware should look like. However it is difficult to catch up with Nvidia, the most valuable company in the world with almost limitless resources to invest in further improving its hardware and software. Even while AMD improves, it will continue to look bad in comparison to Nvidia as state of art keeps getting pushed forward.
positron26•8h ago
The 7,484+ companies who stand to benefit do not have a good way to split the bill and dogpile a problem that is nearly impossible to progress on without lots of partners adding their perspective via a breadth of use cases. This is why I'm building https://prizeforge.com.
Nvidia didn't do it alone. Industry should not expect or wait on AMD to do it alone. Waiting just means lighting money on fire right now. In return for support, industry can demand more open technology be used across AMD's stack, making overall competition better in response for making AMD competitive.
z3ratul163071•5h ago
JonChesterfield•7h ago
Another is that people unsportingly write things in cuda.
It'll be a "just works" thing eventually, even if you need software from outside AMD to get it running well.
rbanffy•2h ago
Whether we like it or not, CUDA is the de-facto standard for these things. I wonder how much effort would it take for a company the size of AMD to dedicate a couple million dollars a year to track CUDA as closely as feasible. A couple million dollars is a rounding error for a leading silicon maker.
JonChesterfield•1h ago
aaryamanv•7h ago
dontlaugh•6h ago
green7ea•6h ago
Under Linux, getting LM Studio to work using the Vulkan backend was trivial. Llama.cpp was a bit more involved. ROCm worked surprisingly well with Arch — I would credit the package maintainers. The only hard part was sorting out Python packaging for PyTorch (use local packages with system's ROCm).
I wouldn't say it's perfect but it's definitely not as bad as it used to be. I think the biggest downside is the difference in environment when you use this as a dev machine and then run the models on NVIDIA hardware for prod.
ctas•2h ago
green7ea•1h ago
edit: just to be clear, I can't train anything competitive with even the smallest LLMs.
drcongo•3h ago