One big issue with HBM is the amount of idle power it consumes. A single MI355 is ~230W, just idle.
But to answer - memory is progressing very slowly. DDR4 to DDR5 was not even a meaningful jump. Even PCIe SSDs are slowly catching up to it which is both funny and sad.
As for the usecase - I use my memory as a cache for everything. Every system in the last 15-20 years I used I maxed out memory on, I never cared much about speed of my storage, because after loading everything into RAM, the system and apps feel a lot more responsive. The difference on older systems with HDDs were especially noticeable, but even on an SSDs, things have not improved much due to latencies. Of course using any webapp connecting to the network will negate any benefits of this, but it makes a difference with desktop apps. These days I even have enough memory to be able to run local test VMs so I don't need to use server resources.
https://news.ycombinator.com/item?id=46302002
https://morethanmoore.substack.com/p/solving-the-problems-of...
mschuster91•1mo ago
> Mian Quddus, chairman of the JEDEC Board of Directors, said: “JEDEC members are actively shaping the standards that will define next generation modules for use in AI data centers, driving the future of innovation in infrastructure and performance.”
It's nice to see that there still is progress to be made given that a lot of modern semiconductor technology is at the edge of what plain physics and chemistry allow... but hell I can't say I'm happy that it, like with low-latency/high bandwidth communications and HFT, it will again be only the uber rich that can enjoy the new and fancy stuff for years. It's not like you can afford an average decent mid/upper range GPU these days thanks to the AI bros.
KronisLV•1mo ago
I mean, Nvidia was greedy even before then and AMD just did “Nvidia - 50 USD” or thereabout.
Intel Arc tried shaking up the entry level (retailers spit on that MSRP though) but sadly didn’t make that big of a splash despite the daily experience being okay (I have the B580). Who knows, maybe their B770 will provide an okay mid range experience that doesn’t feel like being robbed.
Over here, to get an Nvidia 5060 Ti 16 GB I'd have to pay over 500 EUR which is fucking bullshit, so I don’t.
consp•1mo ago
The bad part is everyone wants to be on the AI money circle line train (see the various money flow images available) and thus everything caters for that. At this point i'd rather have nvidia and amd quit the gpu business and focus on "ai" only, that way a new competitor can enter the business and cater the the niche applications like consumer gpus.
fodkodrasz•1mo ago
Nvidia is expected to sell GPU intellectual property at a bargain to the entry-level segment, making it unprofitable for Intel to develop a competitive product range. This way, Intel would lack both the competence and the infrastructure internally to eventually break Nvidia’s market share in the higher segments.
kadoban•1mo ago
The Intel Arc B60 probably would have made a splash if they had actually produced any of the damn things. 24GB vram for low prices would have been huge for the AI crowd, and there was a lot of excitement and then Intel just didn't offer them for sale.
The company is too screwed up to take advantage of any opportunities.
ksec•1mo ago
mschuster91•1mo ago
Yeah, maybe in a decade. And the "benefits" will be a metric shit ton of job losses plus a crash that will make 2000's dotcom plus 2007ff real estate/euro combined look harmless...
ksec•1mo ago
You are getting 3nm and 2nm along with GAA later this year precisely because of AI.
loeg•1mo ago
MrBuddyCasino•1mo ago