frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Cook County Assessor's Office Open Source AVM (Automated Valuation Model)

https://github.com/ccao-data
1•larsiusprime•3m ago•0 comments

YouTube views are down (don't panic)

https://www.jeffgeerling.com/blog/2025/youtube-views-are-down-dont-panic
1•mikece•3m ago•0 comments

Serving up delicious cocktails in plain text

https://asciibar.com/
2•basilikum•7m ago•0 comments

Nick Clegg and Tim Berners-Lee: the battle for the soul of the Internet

https://observer.co.uk/culture/books/article/nick-clegg-and-tim-berners-lee-the-battle-for-the-so...
1•Anon84•10m ago•0 comments

Elon Musk's $1T Tesla pay package: Here are the major caveats

https://www.fastcompany.com/91399114/elon-musk-1-trillion-tesla-pay-package-here-are-the-major-ca...
1•makerdiety•11m ago•0 comments

European public service channel ARTE's YouTube channel likely hijacked [video]

https://www.youtube.com/watch?v=sTCOAqxg_WU
1•mastazi•18m ago•1 comments

Show HN: Frontier – draw with the world live on an infinite grid of pixels

https://frontier.place
1•rmjmdr•23m ago•0 comments

Show HN: Percentage Increase Calculator:calculate changes between two values

https://percentage-increase-calculator.net/
1•jumpdong•23m ago•0 comments

Why Ken Loach's "Kes" Features Film's Best Football Scene (2023)

https://www.top10films.co.uk/22409-kes-film-review/
1•mellosouls•24m ago•1 comments

The three stages of religious decline around the world

https://www.nature.com/articles/s41467-025-62452-z
2•toomuchtodo•28m ago•0 comments

Computers Are for Girls – Datagubbe.se

https://datagubbe.se/girls/
6•mastazi•29m ago•0 comments

Show HN: NotePulse - Simple Notion Analytics

https://notepulse.net/
1•snam23•29m ago•0 comments

Share a tool for migrating Git repositories to SHA256

https://github.com/antgroup/hugescm/releases/tag/v0.19.0
2•fcharlie•33m ago•1 comments

The End of the American City

https://kevinerdmann.substack.com/p/the-end-of-the-american-city
5•paulpauper•43m ago•0 comments

The Case Against Social Media Is Stronger Than You Think

https://arachnemag.substack.com/p/the-case-against-social-media-is
4•paulpauper•43m ago•1 comments

The brompton-ness of it all

https://backofmind.substack.com/p/the-brompton-ness-of-it-all
3•paulpauper•44m ago•0 comments

Cloud Storage Is Full Email Scam: What It Is and How to Protect Yourself

https://malwaretips.com/blogs/your-cloud-storage-is-full-email-scam/
4•rolph•46m ago•1 comments

What It Costs to Be a Sorority Girl

https://www.theatlantic.com/ideas/archive/2025/09/sorority-rush-coaches-recruitment-sec/684094/
2•tkgally•49m ago•0 comments

Ready Prek Go

https://hugovk.dev/blog/2025/ready-prek-go/
1•Bogdanp•50m ago•0 comments

Internet slowdowns could last months after Red Sea cable damage

https://www.thenationalnews.com/future/technology/2025/09/07/middle-east-internet-slowdowns-could...
2•geox•52m ago•1 comments

Harvey Mudd Miniature Machine

https://www.cs.hmc.edu/~cs5grad/cs5/hmmm/documentation/documentation.html
1•nill0•53m ago•0 comments

peekaping: selfhosted uptime monitoring similar to UptimeKuma (Go/React)

https://github.com/0xfurai/peekaping
2•luckman212•57m ago•0 comments

Show HN: Convert Google Sheets to Code

https://sheet2code.com/
2•joeemison•58m ago•0 comments

John Law at the front lines of San Francisco counterculture

https://www.sfgate.com/sf-culture/article/burning-man-co-founder-least-interesting-thing-20826295...
1•gmays•1h ago•0 comments

Bob Stein and Voyager (2021)

https://www.filfre.net/2021/06/bob-stein-and-voyager/
3•doppp•1h ago•0 comments

The Limits of Logic [pdf]

https://jefelino.github.io/limits-of-logic.pdf
3•nill0•1h ago•0 comments

Go Is the Language of Vibe Coding

https://sinwoobang.notion.site/Why-Golang-Matters-in-the-Vibe-Coding-Era-2656746440e28017af0cc623...
3•maitrouble•1h ago•1 comments

Is that a deprecation? Or is it just removed?

https://huonw.github.io/blog/2025/09/deprecating-vs-removing/
1•dbaupp•1h ago•0 comments

Show HN: Dog Age Calculator

https://dogcatagedcalculator.com/
1•yangyiming•1h ago•0 comments

Loverse: Japan's New AI Dating App Where Every Match Is Virtual

https://www.tokyoweekender.com/entertainment/tech-trends/japanese-ai-dating-app/
2•techdar42•1h ago•1 comments
Open in hackernews

I am giving up on Intel and have bought an AMD Ryzen 9950X3D

https://michael.stapelberg.ch/posts/2025-09-07-bye-intel-hi-amd-9950x3d/
164•secure•19h ago

Comments

danieldk•19h ago
I feel like both Intel and AMD are not doing great in the desktop CPU stability department. I made a machine with a Ryzen 9900X a while back and it had the issue that it would freeze when idling. A few years before I had a 5950X that would regularly crash under load (luckily it was a prebuilt, so it was ultimately fixed).

When you do not have a bunch of components ready to swap out it is also really hard to debug these issues. Sometimes it’s something completely different like the PSU. After the last issues, I decided to buy a prebuilt (ThinkStation) with on-site service. The cooling is a bit worse, etc., but if issues come up, I don’t have to spend a lot of time debugging them.

Random other comment: when comparing CPUs, a sad observation was that even a passively cooled M4 is faster than a lot of desktop CPUs (typically single-threaded, sometimes also multi-threaded).

johnisgood•18h ago
This does not fill me with much hope. What am I even ought to buy at this point then, I wonder. I have a ~13 years old Intel CPU which lacks AVX2 (and I need it by now) and I thought of buying a new desktop (items separately, of course), but that is crazy to me that it freezes because of the CPU going idle. It was never an issue in my case. I guess I can only hope it is not going to be a problem once I completed building my PC. :|

On what metric am I ought to buy a CPU these days? Should I care about reviews? I am fine with a middle-end CPU, for what it is worth, and I thought of AMD Ryzen 7 5700 or AMD Ryzen 5 5600GT or anything with a similar price tag. They might even be lower-end by now?

hhh•18h ago
Just buy an AMD CPU. One person’s experience isn’t the world. Nobody in my circle has had an issue with any chip from AMD in recent time (10 years).

Intel is just bad at the moment and not even worth touching.

tester756•18h ago
This is funny because recently my AMD Ryzen 7 5700X3D died and I've decided that my next CPU will be Intel

https://news.ycombinator.com/item?id=45043269

homebrewer•18h ago
This Intel?

https://youtu.be/OVdmK1UGzGs

https://youtu.be/oAE4NWoyMZk

tester756•3h ago
That's been year ago, my AMD CPU died very recently
_zoltan_•18h ago
that's a pretty bad decision $/perf wise I'd wager. AMD decidedly owns the desktop space, and deserves so.
tester756•3h ago
wdym?

https://www.cpubenchmark.net/cpu_value_alltime.html

CPUs like Intel Core Ultra 7 265K are pretty close to top Ryzens

Panzer04•1h ago
Intel CPUs are decidedly better value when multicore performance is a concern. At the top end they trade blows.

If your workload is pointer-chasing intel's new CPUs aren't great though, and the X3D chips are possibly a good pick (if the workload fits in cache) which is why they get a lot of hype from reviewers who benchmark games and judge the score 90% based on that performance.

johnisgood•18h ago
That is what I thought, thanks.
danieldk•18h ago
I agree that Intel is bad at the moment (especially with the 13th and 14th gen self-destruct issues). But unfortunately I also know plenty of people with issues with AMD systems.

And it's no bad power quality on mains as someone suggested (it's excellent here) or 'in the air' (whatever that means) if it happens very quickly after buying.

I would guess that a lot of it comes from bad firmware/mainboards, etc. like the recent issue with ASRock mainboards destroying Ryzen 9000-series GPUs: https://www.techspot.com/news/108120-asrock-confirms-ryzen-9... Anyone who uses Linux and has dealt with bad ACPI bugs, etc. knows that a lot of these mainboards probably have crap firmware.

I should also say that I had a Ryzen 3700X and 5900X many years back and two laptops with a Ryzen CPU and they have been awesome.

ahofmann•18h ago
I wouldn't be so hopeless. Intel and AMD CPUs are used in millions of builds and most of them just work.
danieldk•18h ago
However, the vast majority of PCs out there are not hobbyist builds but Dell/Lenovo/HP/etc. [1] with far fewer possible configurations (and much more testing as a byproduct). I am not saying these machines never have issues, but a high failure rate would not be acceptable to their business customers.

[1] Well, most non-servers are probably laptops today, but the same reasoning applies.

giveita•17h ago
If you value your time a dell laptop with extended warranty and accidental damage where they replace shit and send people out to fix shit is well worth it. It costs but you can be a dumb user and call "IT" when you need a fix and thats a nice feeling IMO!
PartiallyTyped•18h ago
3 of my last 4 machines have been AMD x NVDA and I have been very happy. The intel x NVDA machine has been my least stable one.
homebrewer•18h ago
It's either bad luck, bad power quality from the mains, or something in the air in that particular area. I know plenty of people running AM5 builds, have done so myself for the last couple of years, and there were no problems with any of them apart from the usual amdgpu bugs in latest kernels (which are "normal" since I'm running mainline kernels — it's easy to solve by just sticking to lts, and it has seemingly improved anyway since 6.15).
scns•15h ago
> I thought of AMD Ryzen 7 5700

Definetly not that one if you plan to pair with a dedicated GPU! The 5700X has twice the L3 cache. All Ryzen 5000 with a GPU have only 16MB, 5700 has the GPU deactivated.

encom•18h ago
>M4 is faster than a lot of desktop CPUs

Yea, but unfortunately it comes attached to a Mac.

An issue I've encountered often with motherboards, is that they have brain damaged default settings, that run CPU's out of spec. You really have to go through it all with a fine toothed comb and make sure everything is set to conservative stock manufacturer recommended settings. And my stupid MSI board resets everything (every single BIOS setting) to MSI defaults when you upgrade its BIOS.

homebrewer•18h ago
Also be careful with overclocking, because the usual advice of "just running EXPO/XMP" often results in motherboards setting voltages on very sensitive components to more than 30% over their stock values, and this is somehow considered normal.

It looks completely bonkers to me. I overclocked my system to ~95% of what it is able to do with almost default voltages, using bumps of 1-3% over stock, which (AFAIK) is within acceptable tolerances, but it requires hours and hours of tinkering and stability testing.

Most users just set automatic overclocking, have their motherboards push voltages to insane levels, and then act surprised when their CPUs start bugging out within a couple of years.

Shocking!

electroglyph•18h ago
yah, the default overclocking stuff is pretty aggressive these days
ahartmetz•17h ago
Since you mention EXPO/XMP, which are about RAM overclocking: RAM has the least trouble with overvoltage. Raising some of the various CPU voltages is a problem, which RAM overclocking may also do.
eptcyka•17h ago
The heat is starting to become an issue for DDR5 with higher voltage.
danieldk•17h ago
Unfortunately, some separately purchasable hardware components seem to be optimized completely for gamers these days (overclocking mainboards, RGB on GPUs, etc.).

I'd rather run everything at 90% and get very big power savings and still have pretty stellar performance. I do this with my ThinkStation with Core Ultra 265K now - I set the P-State maximum performance percentage to 90%. Under load it runs almost 20 degrees Celsius cooler. Single core is 8% slower, multicore 4.9%. Well worth the trade-off for me.

(Yes, I know that there are exceptions.)

mschuster91•13h ago
> Unfortunately, some separately purchasable hardware components seem to be optimized completely for gamers these days

It turned out during the shitcoin craze and then AI craze that hardcore gamers, aka boomers with a lot of time and retirement money on their hands and early millennials working in big tech building giant-ass man caves, are a sizeable demographic with very deep pockets.

The wide masses however, they gotta live with the scraps that remain after the AI bros and hardcore gamers have had their pick.

hedora•7m ago
I’ve had multiple systems that crash or corrupt data when underclocked, so running at 90% might not be what you want.

You can always play with the CPU governor / disable high power states. That should be well-tested.

danieldk•17h ago
Also relevant in this context:

https://www.pugetsystems.com/blog/2024/08/02/puget-systems-p...

to;dr: they heavily customize BIOS settings, since many BIOSes run CPUs out-of-spec by default. With these customizations there was not much of a difference in failure rate between AMD and Intel at that point in time (even when including Intel 13th and 14th gen).

claudex•18h ago
>And my stupid MSI board resets everything (every single BIOS setting) to MSI defaults when you upgrade its BIOS.

I had the same issue with my MSI board, next one won't be a MSI.

techpression•17h ago
My ASUS and Gigabyte did the same too. I think vendors are being lazy and don’t want to write migration code
izacus•16h ago
Did what exactly? All the ASUS and Gigabytes I've seen had PBO (which I guess you're talking about) disabled by default.
techpression•15h ago
Reset all the settings with BIOS updates
danieldk•17h ago
Yea, but unfortunately it comes attached to a Mac.

Yeah. If Asahi worked on newer Macs and Apple Silicon Macs supported eGPU (yes I know, big ifs), the choice would be simple. I had NixOS on my Mac Studio M1 Ultra for a while and it was pretty glorious.

philistine•12h ago
I'd bet you don't care that it's attached to a Mac. I bet you don't want to switch OS. Which is understandable. In a couple of years, when Microsoft finally offers Windows as an ARM purchase, Linux is finally full-fledged done implementing support, and Apple resuscitates Boot Camp, I think a lot of people like you will look at Macs like the Mac Mini differently.
timmytokyo•11h ago
Just get used to the extortionate prices on things like memory and storage. Who wouldn't want to pay $200 to go from a 16GB to a 24GB configuration? Who wouldn't want to pay $600 more for the 2TB storage option? And forget about upgrading after you buy the computer.
bob1029•16h ago
I've got a 5950x that I can reliably crater with a very specific .NET 8 console app when it would otherwise be stable 24/7/365, even under some pretty crazy workloads like Unity.

I think a lot of it boils down to load profile and power delivery. My 2500VA double conversion UPS seems to have difficulty keeping up with the volatility in load when running that console app. I can tell because its fans ramp up and my lights on the same circuit begin to flicker very perceptibly. It also creates audible PWM noise in the PC which is crazy to me because up til recently I've only ever heard that from a heavily loaded GPU.

bell-cot•15h ago
I'm sure there are spec's for how fast a PS should be able to ramp up in response to spikes in demand, how a motherboard should handle sudden load changes, etc.

But if your UPS (or just the electrical outlet you're plugged into) can't cope - dunno if I'd describe that as cratering your CPU.

heelix•13h ago
I wonder if cooling/power is really the key here. I've got a 5950x that ended up getting the water loop I'd intended for my next threadripper - only to find they were not selling the blasted things to anyone but a few companies. With the cooling sized for almost twice what the 5950x could put out, it has been a very stable machine for some crazy workloads. That old dog will likely keep the setup when a zen 5 TR gets swapped in.

For a long time, my Achille's heel was my Bride's vacuum. Her Dyson pulled enough amps that the UPS would start singing and trigger the auto shutdown sequence for the half rack. Took way too long to figure out as I was usually not around when she did it.

486sx33•2h ago
My 5950 didn’t like liquid cooling and lives very well with air cooling :)
api•15h ago
The M series chips aren’t the absolute fastest in raw speed, though they are toward the top of the list, but they destroy x86 lineage chips on performance per watt.

I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.

I’ve read people here and there arguing that instruction sets don’t matter, that it’s all the same past the decoder anyway. I don’t buy it. The superior energy efficiency of ARM chips is so obvious I find it impossible to believe it’s not due to the ISA since not much else is that different and now they’re often made on the same TSMC fabs.

adithyassekhar•13h ago
I'd rather it spin the fan all the time to improve longevity but that's just me.

One of the many reasons why snapdragon windows laptops failed was both amd and Intel (lunar lake) was able to reach the claimed efficiency of those chips. I still think modern x86 can match arm ones in efficiency if someone bothered to tune the os and scheduler for most common activities. M series was based on their phone chips which were designed from the ground up to run on a battery all these years. AMD/Intel just don't see an incentive to do that nor do Microsoft.

ac29•11h ago
> I have an M1 Max, a few revisions old, and the only thing I can do to spin up the fans is run local LLMs or play Minecraft with the kids on a giant ultra wide monitor at full frame rate. Giant Rust builds and similar will barely turn on the fan. Normal stuff like browsing and using apps doesn’t even get it warm.

This anecdote perfectly describes my few generation old Intel laptop too. The fans turn on maybe once a month. I dont think its as power efficient as an M-series Apple CPU, but total system power is definitely under 10W during normal usage (including screen, wifi, etc).

dagmx•3h ago
Afaik they are the fastest cores in raw speed. They’re just not available in very high core offerings so eventually fall behind when parallelism wins.
etempleton•14h ago
My experience is similar. Modern enthusiast CPUs and hardware compatibility is going backwards. I have a 5900x that randomly crashes on idle, but not under load. My 285K has so far been rock solid and generally feels snappier. I feel like both Intel and AMD are really trying to push the envelope to look good on benchmarks and this is the end result.
seec•6h ago
Your comment about the passively cooled M4 is misleading. Sure, in single thread, it will be definitely faster. In multithread unless you are going for low end or older CPUs it's basically a lie. A 10 Core M4 will score around a 14TH gen mobile i5. It will consume much less power but the argument is on performance, so that's beside the point.

And if we are talking about a passively cooled M4 (MacBook Air basically) it will quite heavily throttle relatively quickly, you lose at the very least 30%.

So, let's not misrepresent things, Apple CPUs are very power efficient but they are not magic, if you hit them hard, they still need good cooling. Plenty of people have had the experience with their M4 Max, discovering that actually, if they did use the laptop as a workstation, it will generate a good amount of fan noise, there is no other way around.

Apple stuff is good because most people actually have bursty workload (especially graphic design, video editing and some audio stuff) but if you hammer it for hours on end, it's not that good and the power efficiency point becomes a bit moot.

protocolture•2h ago
>I made a machine with a Ryzen 9900X a while back and it had the issue that it would freeze when idling

I also have this issue.

c0balt•1h ago
If you are on Linux, there are long time known problems with low power cpu states. These states can be entered by your CPU when under low/no load.

A common approach is to go into the BIOS/UEFI settings and check that c6 is disabled. To verify and/or temporarily turn c6 off, see https://github.com/r4m0n/ZenStates-Linux

hedora•10m ago
It’s also worth checking all the autodetected stuff that can be overclocked, like ram speed. That stuff can be wrong, and then you get crazy but similar bugs in linux and windows.
kristopolous•1h ago
M3 ultra is the more capable chip by quite a bit. For instance: 80 GPU cores versus 10 in the m4.

Twice the memory bandwidth, twice the CPU core count... It's really wacky how they've decided to name things

whyoh•19h ago
It's crazy how unreliable CPUs have become in the last 5 years or so, both AMD and Intel. And it seems they're all running at their limit from the factory, whereas 10-20 years ago they usually had ample headroom for overclocking.
stavros•18h ago
That's good, isn't it? I don't want the factory leaving performance on the table.
bell-cot•18h ago
Depends on your priorities. That "performance on the table" might also be called "engineering safety factor for stability".
stavros•18h ago
Given that there used to be plenty of room to overclock the cores while still keeping them stable, I think it was more "performance on the table".
formerly_proven•18h ago
You could also get the idea that vendors sometimes make strange decisions which increase neither performance nor reliability.

For example, various brands of motherboards are / were known to basically blow up AMD CPUs when using AMP/XMP, with the root cause being that they jacked an uncore rail way up. Many people claimed they did this to improve stability, but overclockers now that that rail has a sweet spot for stability and they went way beyond it (so much so that the actual silicon failed and burned a hole in itself with some low-ish probability).

makeitdouble•17h ago
TBF using more conservative energy profiles will bring stability and safety. To that effect in Windows the default profile effectively debuffs the CPU and most people will be fine that way.
techpression•17h ago
The 7800X3D is amazing here, runs extremely cool and stable, you can push it far above its defaults and it still won’t get to 80C even with air cooling. Mine was running between 60-70 under load with PBO set to high. Unfortunately it seems its successor is not that great :/
williamDafoe•16h ago
The 7000 series of CPUs is NOT known for running cool, unlike the AMD 5000 series (which are basically server CPUs repurposed for desktop usage). In the 7000 series, AMD decided to just increase the power of each CPU and that's where most of the performance gains are coming from - but power consumption is 40-50% higher than with similar 5000-series CPUs.
scns•15h ago
When you use EcoMode with them you only lose ~5% performance, but are still ~30% ahead of the corresponding 5000-series CPU. You can reduce PPT/TDP even further while still ahead.

https://www.computerbase.de/artikel/prozessoren/amd-ryzen-79...

techpression•15h ago
I specifically singled out the 7800X3D though, it runs incredibly cool and at a very low power draw for the performance you get.
mldbk•11h ago
> You know, I'm something of a CPU engineer myself :D

Actually almost everything what you wrote is not true, and commenter above already sent you some links.

7800X3D is the GOAT, very power efficient and cool.

Numerlor•2h ago
The only reason the 7800x3d is power efficient is because it simply can't use much power, and so it runs at a better spot of the efficiency curve. Most of the CPUs won't use more than ~88w without doing manual overclocking (not pbo). Compare that to e.g. a 7600x that's 2 cores fewer on the same architecture and will happily pull over 130w.

And even if could push it higher, they run very hot compared to other CPUs at the same power usage as a combination of AMD's very thick IHS, the compute chiplets being small/power dense and 7000 series X3D cache being on top of the compute chiplet unlike 9000 series that has it on the bottom.

The 9800x3d limited in the same way will be both mildly more power efficient from faster cores and run cooler because of the cache location. The only reason it's hotter is that it's allowed to use significantly more power, usually up to 150w stock, for which you'd have to remove the IHS on the 7800X3D if you didn't want to see magic smoke

hu3•24m ago
Same for 9800X3D here, which is basically the same CPU. Watercooled. Silent. Stupidly fast.
mrheosuper•2m ago
we have unstable "code" generator, so unstable CPU would be natural.
eptcyka•18h ago
This is rather late, to be quite fair.
discardable_dan•18h ago
My thoughts exactly: he figured out in 2025 what the rest of us knew in 2022.
positron26•18h ago
One of my work computers died and I hadn't checked the CPU market in years. Rode home that night in a taxi with a Ryzen 1700x completely stoked that AMD was back in the game.

If anyone thinks competition isn't good for the market or that also-rans don't have enough of an effect, just take note. Intel is a cautionary tale. I do agree we would have gotten where we are faster with more viable competitors.

M4 is neat. I won't be shocked if x86 finally gives up the ghost as Intel decides playing in Risc V or ARM space is their only hope to get back into an up-cycle. AMD has wanted to do heterogeneous stuff for years. Risc V might be the way.

One thing I'm finding is that compilers are actually leaving a ton on the table for AMD chips, so I think this is an area where AMD and all of the users, from SMEs on down, can benefit tremendously from cooperatively financing the necessary software to make it happen.

andsoitis•18h ago
> I would say that 25 to 28 degrees celsius are normal temperatures for computers.

An ideal ambient (room) temperature for running a computer is 15-25 celcius (60-77 Fahrenheit)

Source: https://www.techtarget.com/searchdatacenter/definition/ambie...

johnisgood•18h ago

  coretemp-isa-0000
  Adapter: ISA adapter
  Package id 0:  +40.0°C  (high = +80.0°C, crit = +100.0°C)
  Core 0:        +38.0°C  (high = +80.0°C, crit = +100.0°C)
  Core 1:        +39.0°C  (high = +80.0°C, crit = +100.0°C)
Are they saying this is bad? This Intel CPU has been at it for over a decade. There was a fan issue for half a year and would go up to 80 C for... half a year. Still works perfectly fine but it is outdated, it lacks instruction sets that I need, and it has two cores only, and 1 thread per core.

Maybe today's CPUs would not be able to handle it, I am not sure. One would expect these things to only improve, but seems like this is not the case.

Edit: I misread it, oops! Disregard this comment.

Rohansi•18h ago
That is your CPU temperature, not ambient (room) temperature.
johnisgood•18h ago
Oh, I misread. My bad!
trueismywork•18h ago
And that is an impossibility in most of the world today and it will be even more like that going forward.
nl•18h ago
Much of the world (for better or worse) uses airconditioning in places they commonly use desktop computers.
em-bee•17h ago
no they don't. in some countries in europe (maybe in all of them?), installing airconditioning is frowned upon because it is considered a waste of energy. if you want government subsidies for replacing your heating system with a more energy efficient one you are not allowed to have airconditioning. and in the rest of the world only people/countries well of, that don't consider their energy usage, do it. airconditioning is luxury.

using to much airconditioning is also not comfortable. i used to live in singapore. we used to joke that singapore has two seasons: indoors and outdoors. because the airconditioning is powered so high that you had to bring jacket to wear inside. i'd frequently freeze after entering a building. i don't know why they do it, because it doesn't make sense. when i did turn on airconditioning at home i'd go barely below 30. just a few degrees cooler than the outside so it feels more comfortable without making the transition to hard.

trueismywork•3h ago
Europe not having some kind of air conditioning (even if heat pump based) is stupidity in my opinion. It lowers productivity
trueismywork•3h ago
No they dont. They don't have the money. I remember my childhood when gaming in summer holidays in India, my PC would run at full tilt because my room was at 36C (and outside was 48C).
imtringued•17h ago
So you're saying that if you go even 3 degrees Celsius over that temperature range you should expect your CPU to fry itself? Even when the CPU throttled itself to exactly 100°C?
andsoitis•17h ago
> So you're saying that if you go even 3 degrees Celsius over that temperature range you should expect your CPU to fry itself? Even when the CPU throttled itself to exactly 100°C?

It is actually 2.9999, precisely.

crinkly•18h ago
I’ve given up both.
Thorrez•18h ago
What do you use?
aurareturn•18h ago
I’ve given up on both and use Apple Silicon only. AMD and Intel are simply too power hungry for how slow they are and can’t optimize for power like Apple can.
mr_windfrog•18h ago
That's pretty amazing, I've never heard of that before .-_-!
lostlogin•18h ago
I just replaced a headless nuc 9 with a headless M4.

Nuc 9 averaged 65-70W power usage, while the m4 is averaging 6.6W.

The Mac is vastly more performant.

maciejw•18h ago
I also switched to cheapest Mac Mini M4 Pro this year (after 20+ years of using Intel CPUs). MacOS has its quirks, but it provides ZSH and it "just works" (unlike manjaro I used in parallel with Windows). I especially like the preview tool - it has useful pdf and photo editing options.

The hardware is impressive - tiny, metal box, always silent, basic speaker built-in and it can be left always on with minimal power consumption.

Drive size for basic models is limited (512gb) - I solved it by moving photos to NAS. I don't use it for gaming, except Hello Kitty Island Adventure. I would say it's a very competitive choice for a desktop PC in 2025 overall.

mrlonglong•18h ago
To make linking go quicker, use mold.

Pass -fuse=mold when building.

positron26•18h ago
Do beware when doing WASM with mold. I shipped a broken WASM binary that Firefox could run just fine but Chrome would not.
Jnr•18h ago
I have not had any issues with Intel or AMD CPUs but I have so many issues with AMD APUs, I would steer clear of them. In my experience with different models, they have many graphics issues, broken video transcoding and overall extremely unstable. If you need decent integrated graphics then Intel is the only real option.
sellmesoap•18h ago
They make a lot of apus for gaming handhelds, I think they do well in that segment. I've had a handful of desktop and laptop apus with no complaints. Even an APU with ecc support, they've all worked without a hitch. I haven't tried transcoding anything on them mind you.
Jnr•18h ago
Yes, I have Steam Deck and it works great. But I also have 2400G and 5700G and both of those have graphics issues (tested with different recommended RAM sets).
vkazanov•17h ago
My laptop's AMD is great (Ryzen AI 7 PRO 360 w/ Radeon 880M). Gaming, GPI work, battery, small LLMs - all just work on my Ubuntu.

Don't know about transcoding though.

imiric•17h ago
I've had the same experience with an 8600G on Linux. Very frequent graphics driver crashes and KDE/Wayland freezes, on old and new kernels alike. I've been submitting error reports for months, and the issues still persist. The RAM passes MemTest, and the system otherwise works fine, but the graphics issues are very annoying. It's not like I'm gaming or doing anything intensive either; it happens during plain desktop usage.

Yet I also use a 7840U in a gaming handheld running Windows, and haven't had any issues there at all. So I think this is related to AMD Linux drivers and/or Wayland. In contrast, my old laptop with an NVIDIA GPU and Xorg has given me zero issues for about a decade now.

So I've decided to just avoid AMD on Linux on my next machine. Intel's upcoming Panther Lake and Nova Lake CPUs seem promising, and their integrated graphics have consistently been improving. I don't think AMD's dominance will continue for much longer.

ychompinator•18h ago
No desktop CPU I’ve ever used has remained stable at 100 degrees. My 14900k crashes almost immediately at that temp. 3 hours at 100 degrees is obscene.
clait•18h ago
Yeah, i can’t believe they think it’s fine. I would’ve shutdown my PC and rethought my cooling setup the first time it hit 100C tbh
swinglock•17h ago
Then all your desktop CPUs were defective.

Besides AMD CPUs of the early 2000s going up in smokes without working cooling, they all throttle before they become temporarily or permanently unstable. Otherwise they are bad.

I've never had a desktop part fail due to max temperatures, but I don't think I've owned one that advertises nor allows itself to reach or remain at 100c or higher.

If someone sells a CPU that's specified to work at 100 or 110 degrees and it doesn't then it's either defective or fraudulent, no excuses.

sys_64738•13h ago
100C will trigger PROCHOT state in Intel leading to CPU throttling. The CPU will eventually shutdown.
arp242•4h ago
285k: https://www.intel.com/content/www/us/en/products/sku/241060/...

Max Operating Temperature: 105 °C

14900k: https://www.intel.com/content/www/us/en/products/sku/236773/...

Max Operating Temperature: 100 °C

Different CPUs, different specs.

And any CPU from the last decade will just throttle down if it gets too hot. That's how the entire "Turbo" thing works: go as fast as we can until it gets too hot, after which it throttles down.

energy123•18h ago
I can't comment on the quality question, but for memory bandwidth sensitive use cases, Intel desktop is superior.
ttyyzz•17h ago
I'm not convinced, what would be the use case?
energy123•16h ago
Data science where you need to keep ~50GB of data in RAM and do intensive things with it (e.g. loop over it repeatedly with numba). You can't get use out of more than 4 cores because memory bandwidth is the only limitation. The data is too big for AMD's cache to be a factor.

Threadripper is built for this. But I am talking about the consumer options if you are on a budget. Intel has significantly more memory bandwidth than AMD in the consumer end. I don't have the numbers on hand, but someone at /r/localllama did a comparison a while ago.

Marsymars•12m ago
Framework Desktop?
shmerl•18h ago
Well, worryingly there were reports of AMD X3D CPUs burning out too. I hope it will be sorted out.
baobabKoodaa•18h ago
Why is the author showing a chart of room temperatures? CPU temperature is what matters here. Expecting a CPU to be stable at 100C is just asking for problems. Issue probably could have been avoided by making improvements to case airflow.
formerly_proven•18h ago
Strange, laptop CPUs and their thermal solutions are designed in concert to stay at Tjmax when under sustained load and throttle appropriately to maintain maximum temperature (~ power ~ performance).
ACCount37•17h ago
And those mobile devices have much more conservative limits, and much more aggressive throttling behavior.

Smartphones have no active cooling and are fully dependent on thermal throttling for survival, but they can start throttling at as low as 50C easily. Laptops with underspecced cooling systems generally try their best to avoid crossing into triple digits - a lot of them max out at 85C to 95C, even under extreme loads.

dezgeg•14h ago
For handhelds the temperature of the device's case is one factor as well when deciding the thermal limits (so you don't burn the user's hands) - less of a problem on laptops.
Jolter•18h ago
I would expect the CPU to start throttling at high temperatures in order to avoid damage. Allegedly, it never did, and instead died. Do you think that’s acceptable in 2025?
FeepingCreature•17h ago
No but it's also important to realize that this CPU was running at an insane temperature that should never happen in normal operation. I have a laptop with an undersized fan and if I max out all my cores with full load, I barely cross 80. 100 is mental. It doesn't matter if the manufacturer set the peak temperature wrong, a computer whose cpu reaches 100 degrees celsius is simply built incorrectly.

If nothing else, it very clearly indicates that you can boost your performance significantly by sorting out your cooling because your cpu will be stuck permanently emergency throttling.

izacus•16h ago
I somehow doubt that, are you looking at the same temperature? I haven't seen a laptop that would have thermal stop under 95 for a long time and any gaming laptop will run at 95 under load for package temps.
FeepingCreature•16h ago
i7 8550u. Google confirms it stabilizes at 80-85C.

That said, there's a difference between a laptop cpu turbo boosting to 90 for a few minutes and a desktop cpu, which are usually cooler anyway, running at 100 sustained for three hours.

ACCount37•17h ago
Thermal throttling originated as a safety feature. The early implementations were basically a "thermal fuse" in function, and cut all power to the system to prevent catastrophic hardware damage. Only later did the more sophisticated versions that do things like "cut down clocks to prevent temps from rising further" appear.

On desktop PCs, thermal throttling is often set up as "just a safety feature" to this very day. Which means: the system does NOT expect to stay at the edge of its thermal limit. I would not trust thermal throttling with keeping a system running safely at a continuous 100C on die.

100C is already a "danger zone", with elevated error rates and faster circuit degradation - and there are only this many thermal sensors a die has. Some under-sensored hotspots may be running a few degrees higher than that. Which may not be enough to kill the die outright - but more than enough to put those hotspots into a "fuck around" zone of increased instability and massively accelerated degradation.

If you're relying on thermal throttling to balance your system's performance, as laptops and smartphones often do, then you seriously need to dial in better temperature thresholds. 100C is way too spicy.

baobabKoodaa•6h ago
What does room temperature have to do with any of this? Yes, you can lower your CPU temperature by lowering your room temperature. But you can also lower your CPU temperature by a variety of other means; particularly by improving case airflow. CPU temperature is the interesting metric here, not room temperature.
swinglock•17h ago
The text clearly explains all of this.
baobabKoodaa•7h ago
No it does not. Which part of the text do you feel explains this?
chmod775•1h ago
First off, there's a chart for CPU temperature at the very top and they do talk about it:

> I also double-checked if the CPU temperature of about 100 degrees celsius is too high, but no: [..] Intel specifies a maximum of 110 degrees. So, running at “only” 100 degrees for a few hours should be fine.

Secondly, the article reads:

> Tom’s Hardware recently reported that “Intel Raptor Lake crashes are increasing with rising temperatures in record European heat wave”, which prompted some folks to blame Europe’s general lack of Air Conditioning.

> But in this case, I actually did air-condition the room about half-way through the job (at about 16:00), when I noticed the room was getting hot. Here’s the temperature graph:

> [GRAPH]

> I would say that 25 to 28 degrees celsius are normal temperatures for computers.

So apparently a Tom's Hardware article connected a recent heat wave with crashing computers containing Intel CPUs. They brought that up to rule it out by presenting a graph showing reasonable room temperatures.

I hope this helps.

perching_aix•17h ago
> Expecting a CPU to be stable at 100C is just asking for problems.

No. High performance gaming laptops will routinely do this for hours on end for years.

If it can't take it, it shouldn't allow it.

bell-cot•14h ago
I've not looked at the specifics here - but "stable at X degrees, Y% duty cycle, for Z" years is just another engineering spec.

Intel's basic 285K spec's - https://www.intel.com/content/www/us/en/products/sku/241060/... - say "Max Operating Temperature 105 °C".

So, yes - running the CPU that close to its maximum is really not asking for stability, nor longevity.

No reason to doubt your assertion about gaming laptops - but chip binning is a thing, and the manufacturers of those laptops have every reason to pay Intel a premium for CPU's which test to better values of X, Y, and Z.

formerly_proven•18h ago
> Looking at my energy meter statistics, I usually ended up at about 9.x kWh per day for a two-person household, cooking with induction.

> After switching my PC from Intel to AMD, I end up at 10-11 kWh per day.

It's kind of impressive to increase household electricity consumption by 10% by just switching one CPU.

usr1106•16h ago
I guess the author runs it at high load for long times, not only for the benchmarks to write this blog post. And less than 10 kWh is a low starting point, many households would be much higher.
don-bright•15h ago
The amount of power this is using is roughly the same as it takes my car to do my short commute to work
willtemperley•17h ago
Gaming seems to be the final stronghold of x86 and I imagine that will shrink. Clearly games are able to run well on RISC architectures despite decades of x86 optimisation in game engines. Long term, an architecture that consumes more power and is tightly locked down by licensing looks unsustainable compared to royalty-free RISC alternatives. The instability, presumably because Intel are overclocking their own chips to look OK on benchmarks will not help.
smallpipe•17h ago
x86 hasn't been CISC in 3 decades anywhere but in the frontend. An architecture doesn't consume power, a design does. I'm all for shitting on intel, but getting the facts right wouldn't hurt.
uncircle•15h ago
X86 isn’t CISC, sure, but it isn’t a RISC architecture either.
immibis•6h ago
The traditional CISC and RISC division broke down the moment processors started doing more than one thing at a time.

A RISC architecture was actually one with simple control flow and a CISC architecture was one with complex control flow, usually with microcode. This distinction isn't applicable to CPUs past the year 1996 or so, because it doesn't make sense to speak of a CPU having global control flow.

arp242•3h ago
Do RISC architectures still exist? ARM has gained tons of stuff and isn't really "RISC" any more either.

Maybe RISC-V? It's right there in the name, but I haven't really looked at it. However, there are no RISC-V chips that have anywhere near the performance x86 or ARM has, so it remains to be seen if RISC-V can be competitive with x86 or ARM for these types of things.

RISC is one of those things that sounds nice and elegant in principle, but works out rather less well in practice.

willtemperley•12h ago
You’re contradicting yourself. The whole reason x86 burns more power is that the CISC front end can’t be avoided.
wqaatwt•12h ago
Are there any mid-high end RISC-V chips that have comparable performance per watt to x86?
williamDafoe•16h ago
"Stronghold" is a joke phrase, is it not? Intel had ZERO progress in integrated graphics from 2013-2020. ZERO. That's the reason why "it works so well" - because they NEVER improved the performance or architecture! Sure, they diddled with the number of CU's, but as far as graphics architecture, they never changed it, and it was POOR to begin with (couldn't do 1080p esports very well ...)
willtemperley•12h ago
Are x86 consoles a joke?
nofriend•2h ago
x86 is the cpu architecture. i don't believe gp was talking about intels igpu solution at all.
wqaatwt•12h ago
> Clearly games are able to run well on RISC architectures

Theoretically that’s likely true. But is there any empirical evidence?

Even underclocked Intel desktop chips are massively faster.

willtemperley•12h ago
Mobile gaming has run well on RISC for a long time and more recently Macs have shown gaming potential.
wqaatwt•11h ago
Oh sorry. Based on the tone and general fervor in your comment I somehow read it as RISC-V instead of simply RISC (which as other say seems like a mostly meaningless label these days).

Yes, ARM is certainly competitive. But I don’t know how much is that down to Apple being good at making chips instead of the architecture itself.

Qualcomm of course makes decent chips but it’s not like they are that much ahead of x86 on laptops.

Even in Apple’s case, if you only care about raw CPU power instead of performance per watt M series is not that great compared to AMD/Intel.

KronisLV•17h ago
> I also double-checked if the CPU temperature of about 100 degrees celsius is too high, but no: this Tom’s Hardware article shows even higher temperatures, and Intel specifies a maximum of 110 degrees. So, running at “only” 100 degrees for a few hours should be fine.

I'd say that even crashing at max temperatures is still completely unreasonable! You should be able to run at 100C or whatever the max temperature is for a week non-stop if you well damn please. If you can't, then the value has been chosen wrong by the manufacturers. If the CPU can't handle that, the clock rates should just be dialed back accordingly to maintain stability.

It's odd to hear about Core Ultra CPUs failing like that, though - I thought that they were supposed to be more power efficient than the 13th and 14th gen, all while not having their stability issues.

That said, I currently have a Ryzen 7 5800X, OCed with PBO to hit 5 GHz with negative CO offsets per core set. There's also an AIO with two fans and the side panel is off because the case I have is horrible. While gaming the temps usually don't reach past like 82C but Prime95 or anything else that's computationally intensive can make the CPU hit and flatten out at 90C. So odd to have modern desktop class CPUs still bump into thermal limits like that. That's with a pretty decent ambient temperature between 21C to 26C (summer).

williamDafoe•16h ago
Just FYI Google runs their data centers at 85 degrees F (about 30 degrees C). I think Google probably knows more about how to run Intel CPUs for longest life and lowest cost per CPU cycle. After all they are the #5 computer maker on earth. What Intel is doing and what they are recommending is the act of a desperate corporation incapable of designing energy-efficient CPUs, incapable of progressing their performance in MIPS per Watt of power. This is a sign of a failed corporation.
KronisLV•14h ago
> Just FYI Google runs their data centers at 85 degrees F (about 30 degrees C). I think Google probably knows more about how to run Intel CPUs for longest life and lowest cost per CPU cycle. After all they are the #5 computer maker on earth.

Servers and running things at scale are way different from consumer use cases and the cooling solutions you'll find in the typical desktop tower, esp. considering the average budget and tolerance for noise. Regardless, on a desktop chip, even if you hit tJMax, it shouldn't lead to instability as in the post above, nor should the chips fail.

If they do, then that value was chosen wrong by the manufacturer. The chips should also be clocking back to maintain safe operating temps. Essentially, squeeze out whatever performance is available with a given cooling solution: be it passive (I have some low TDP AM4 chips with passive Alpine radiator blocks), air coolers or AIOs or a custom liquid loop.

> What Intel is doing and what they are recommending is the act of a desperate corporation incapable of designing energy-efficient CPUs, incapable of progressing their performance in MIPS per Watt of power.

I don't disagree with this entirely, but the story is increasingly similar with AMD as well - most consumer chip manufacturers are pushing the chips harder and harder out of the factory, so they can compete on benchmarks. That's why you hear about people limiting the power envelope to 80-90% of stock and dropping close to 10 degrees C in temperatures, similarly you hear about the difficulties of pushing chips all that far past stock in overclocking, because they're already pushed harder than the prior generations.

To sum up: Intel should be less delusional in how far they can push the silicon, take the L and compete against AMD on the pricing, instead of charging an arm and a leg for chips that will burn up. What they were doing with the Arc GPUs compared to the competitors was actually a step in the right direction.

Panzer04•1h ago
Google runs datacenters hot because it's probably cheaper than over-cooling them with AC.

Chips are happy to run at high temperatures, that's not an issue. It's just a tradeoff of expense and performance.

c0l0•17h ago
I realize this has not much to do with CPU choice per se, but I'm still gonna leave this recommendation here for people who like to build PCs to get stuff done with :) Since I've been able to afford it and the market has had them available, I've been buying desktop systems with proper ECC support.

I've been chasing flimsy but very annoying stability problems (some, of course, due to overclocking during my younger years, when it still had a tangible payoff) enough times on systems I had built that taking this one BIG potential cause out of the equation is worth the few dozens of extra bucks I have to spend on ECC-capable gear many times over.

Trying to validate an ECC-less platform's stability is surprisingly hard, because memtest and friends just aren't very reliably detecting more subtle problems. PRIME95, y-cruncher and linpack (in increasing order of effectiveness) are better than specialzied memory testing software in my experience, but they are not perfect, either.

Most AMD CPUs (but not their APUs with potent iGPUs - there, you will have to buy the "PRO" variants) these days have full support for ECC UDIMMs. If your mainboard vendor also plays ball - annoyingly, only a minority of them enables ECC support in their firmware, so always check for that before buying! - there's not much that can prevent you from having that stability enhancement and reassuring peace of mind.

Quoth DJB (around the very start of this millenium): https://cr.yp.to/hardware/ecc.html :)

swinglock•17h ago
Excellent point. It's a shame and a travesty that data integrity is still mostly locked away inside servers, leaving most other computing devices effectively toys, the early prototype demo thing but then never finished and sold forever at inflated prices.

I wish AMD would make ECC a properly advertised feature with clear motherboard support. At least DDR5 has some level of ECC.

kevin_thibedeau•2h ago
> At least DDR5 has some level of ECC.

That is mostly to assist manufacturers in selling marginal chips with a few bad bits scattered around. It's really a step backwards in reliability.

dijit•17h ago
> only a minority of them enables ECC support in their firmware, so always check for that before buying!

This is the annoying part.

That AMD permits ECC is a truly fantastic situation, but if it's supported by the motherboard is often unlikely and worse: it's not advertised even when it's available.

I have an ASUS PRIME TRX40 PRO and the tech specs say that it can run ECC and non-ECC but not if ECC will be available to the operating system, merely that the DIMMS will work.

It's much more hit and miss in reality than it should be, though this motherboard was a pricey one: one can't use price as a proxy for features.

c0l0•17h ago
Usually, if a vendor's spec sheet for a (SOHO/consumer-grade) motherboard mentions ECC-UDIMM explicitly in its memory compatibility section, and (but this is a more recent development afaict) DOES NOT specify something like "operating in non-ECC mode only" at the same time, then you will have proper ECC (and therefore EDAC and RAS) support in Linux, if the kernel version you have can already deal with ECC on your platform in general.

I would assume your particular motherboard to operate with proper SECDED+-level ECC if you have capable, compatible DIMM, enable ECC mode in the firmware, and boot an OS kernel that can make sense of it all.

consp•17h ago
Isn't it mostly an ease of mind thing? I've never seen a ECC error on my home server which has plenty of memory in use and runs longer than my desktop. Maybe it's more common with higher clocked, near the limit, desktop PC's.

Also: DDR5 has some false ecc marketing due to the memory standard having an error correction scheme build in. Don't fall for it.

c0l0•17h ago
I see a particular ECC error at least weekly on my home desktop system, because one of my DIMMs doesn't like the (out of spec) clock rate that I make it operate at. Looks like this:

    94 2025-08-26 01:49:40 +0200 error: Corrected error, no action required., CPU 2, bank Unified Memory Controller (bank=18), mcg mcgstatus=0, mci CECC, memory_channel=1,csrow=0, mcgcap=0x0000011c, status=0x9c2040000000011b, addr=0x36e701dc0, misc=0xd01a000101000000, walltime=0x68aea758, cpuid=0x00a50f00, bank=0x00000012
    95 2025-09-01 09:41:50 +0200 error: Corrected error, no action required., CPU 2, bank Unified Memory Controller (bank=18), mcg mcgstatus=0, mci CECC, memory_channel=1,csrow=0, mcgcap=0x0000011c, status=0x9c2040000000011b, addr=0x36e701dc0, misc=0xd01a000101000000, walltime=0x68b80667, cpuid=0x00a50f00, bank=0x00000012
(this is `sudo ras-mc-ctl --errors` output)

It's always the same address, and always a Corrected Error (obviously, otherwise my kernel would panic). However, operating my system's memory at this clock and latency boosts x265 encoding performance (just one of the benchmarks I picked when trying to figure out how to handle this particular tradeoff) by about 12%. That is an improvement I am willing to stomach the extra risk of effectively overclocking the memory module beyond its comformt zone for, given that I can fully mitigate it by virtue of properly working ECC.

Hendrikto•16h ago
Running your RAM so far out of spec that it breaks down regularly, where do you take the confidence that ECC will still work correctly?

Also: Could you not have just bought slightly faste RAM, given the premium for ECC?

c0l0•14h ago
"Breaks down" is a strong choice of words for a single, corrected bit error. ECC works as designed, and demonstrates that it does by detecting this re-occurring error. I take the confidence mostly from experience ;)

And no, as ECC UDIMM for the speed (3600MHz) I run mine at simply does not exist - it is outside of what JEDEC ratified for the DDR4 spec.

adithyassekhar•13h ago
JEDEC rated DDR4 at only 2400mhz right? And anything higher is technically over clocking?
c0l0•13h ago
JEDEC specifies rates up to 3200MT/s, what's officially referred to as DDR4-3200 (PC4-25600).
dijit•12h ago
JEDEC has a few frequencies they support: https://www.jedec.org/standards-documents/docs/jesd79-4a

DDR4-1600 (PC4-12800)

DDR4-1866 (PC4-14900)

DDR4-2133 (PC4-17000)

DDR4-2400 (PC4-19200)

DDR4-2666 (PC4-21300)

DDR4-2933 (PC4-23466)

DDR4-3200 (PC4-25600) (the highest supported in the DDR4 generation)

What's *NOT* supported are some enthusiast ones that typically require more than 1.2v for example: 3600 MT/s, 4000 MT/s & 4266 MT/s

kderbe•9h ago
I would loosen the memory timings a bit and see if that resolves the ECC errors. x265 performance shouldn't fall since it generally benefits more from memory clock rate than latency.

Also, could you share some relevant info about your processor, mainboard, and UEFI? I see many internet commenters question whether their ECC is working (or ask if a particular setup would work), and far fewer that report a successful ECC consumer desktop build. So it would be nice to know some specific product combinations that really work.

immibis•14h ago
I saw a corrected memory error logged every few hours when my current machine was new. It seems to have gone away now, so either some burn-in effect, or ECC accidentally got switched off and all my data is now corrupted. Threadripper 7000 series, 4x64GB DDR5.

Edit: it's probably because I switched it to "energy efficiency mode" instead of "performance mode" because it would occasionally lock up in performance mode. Presumably with the same root cause.

BikiniPrince•43m ago
I pick up old serves for my garage system. With edac it is a dream to isolate the fault and be instantly aware. It also lets you determine the severity of the issue. Dimms can run for years with just the one error or overnight explode into streams of corrections. I keep spares so it’s fairly easy to isolate any faults. It’s just how do you want to spend your time?
hedora•42m ago
You have to go pretty far down the rabbit hole to make sure you’ve actually got ECC with [LP]DDR5

Some vendors use hamming codes with “holes” in them, and you need the CPU to also run ECC (or at least error detection) between ram and the cache hierarchy.

Those things are optional in the spec, because we can’t have nice things.

Scramblejams•32m ago
I run a handful of servers and I have a couple that pop ECC errors every year or three, so YMMV.
Ekaros•17h ago
Seems like failure in choosing cooling solutions. These high-end chips have obscene cooling needs. My guess would be using something that was not designed for TDP in question.

Sufficient cooler, with sufficient airflow is always needed.

ttyyzz•17h ago
Agree. Also, use good thermal paste. 100 °C is not safe or sustainable long term. Unfortunately, I think the manufacturer's specifications regarding the maximum temperature are misleading. With proper cooling, however, you'll be well within that limit.
uniqueuid•17h ago
For what it's worth, I have an i9-13900K paired with the largest air cooler available at the time (a be quiet! Dark Rock 5 IIRC), and it's incapable of sufficiently cooling that CPU.

The 13900k draws more than 200W initially and thermal throttles after a minute at most, even in an air conditioned room.

I don't think that thermal problems should be pushed to end user to this degree.

SomeoneOnTheWeb•16h ago
This means your system doesn't have enough airflow if it throttles this quickly.

But I agree this should not be a problem in the first place.

michaelt•16h ago
The "Dark Rock 5" marketing materials say it provides a 210 W TDP [1] and marketers seldom under-sell their products' capabilities.

So if your CPU is drawing "more than 200W" you're pretty much at the limits of your cooler.

[1] https://www.bequiet.com/en/cpucooler/5110

lmm•1h ago
Feels like CPU manufacturers should be at least slapping a big warning on if they're selling a CPU that draws more power than any available cooler can dissipate.
onli•17h ago
No, those processors clock or shut down if too hot. In no circumstances should they fail because of insufficient cooling. Even without airflow etc.
williamDafoe•16h ago
A badly optimized CPU will take excessive amounts of power. The "failure in choosing cooling solutions" excuse is just the pot calling the kettle black.
hawshemi•17h ago
And welcome to USB slow speeds and issues...
amelius•17h ago
USB issues are driving me nuts. Please, someone show me the path to serenity.
black_puppydog•16h ago
Layoutparser looks really neat! Glad the author led with this. :D
fmajid•15h ago
I generally prefer AMD Zen5 to Intel due to AVX512 not being gimped by crippled E-cores that really don't belong on a desktop system, SMT (hyperthreading) that actually works and using TSMC processes, but they've also had their issues recently:

https://www.theregister.com/2025/08/29/amd_ryzen_twice_fails...

enronmusk•14h ago
If OP's CPU cooler (Noctua NH-D15 G2) wasn't able to cool down his CPU below 100C, he must have been (intentionally or unintentionally with Asus multi core enhancement) overclocked his CPU. Or he didn't apply thermal paste properly or didn't remove the cooler plastic sticker?

I have followed his blog for years and hold him in high respect so I am surprised he has done that and expected stability at 100C regardless of what Intel claim is okay.

Not to mention that you rapidly hit diminishing returns pass 200W with current gen Intel CPUs, although he mentions caring able idle power usage. Why go from 150W to 300W for a 20% performance increase?

rurban•11h ago
I've burned two of those already, watercooled. The H100 traffic was too much for them
unsnap_biceps•9h ago
I also recently swapped to amd and my biggest surprise was how awful their platform is. With intel, getting sensor data just worked without anything special. With amd, it looks like each platform, perhaps model, requires special support. My mobo is a used godlike, and just has zero sensor support.
protocolture•2h ago
Hmm arent the top end Ryzens dying too? This could be a funny future blog post.
scotty79•2h ago
Last Intel desktop CPU that I bought was Pentium 133Mhz. It was also my first PC. Never again ratio of performance to price in my preferred price range favored Intel.