frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Cloudflare Global Network experiencing issues

https://www.cloudflarestatus.com/?t=1
2078•imdsm•6h ago•1349 comments

Gemini 3 for developers: New reasoning, agentic capabilities

https://blog.google/technology/developers/gemini-3-developers/
312•janpio•2h ago•96 comments

Gemini 3 Pro Preview Live in AI Studio

https://aistudio.google.com/prompts/new_chat?model=gemini-3-pro-preview
398•preek•3h ago•166 comments

Pebble, Rebble, and a Path Forward

https://ericmigi.com/blog/pebble-rebble-and-a-path-forward/
60•phoronixrly•46m ago•7 comments

A Day at Hetzner Online in the Falkenstein Data Center

https://www.igorslab.de/en/a-day-at-hetzner-online-in-the-falkenstein-data-center-insights-into-s...
69•speckx•1h ago•18 comments

Gemini 3

https://blog.google/products/gemini/gemini-3/
315•meetpateltech•2h ago•90 comments

5 Things to Try with Gemini 3 Pro in Gemini CLI

https://developers.googleblog.com/en/5-things-to-try-with-gemini-3-pro-in-gemini-cli/
66•keithba•2h ago•21 comments

Strix Halo's Memory Subsystem: Tackling iGPU Challenges

https://chipsandcheese.com/p/strix-halos-memory-subsystem-tackling
23•PaulHoule•1h ago•8 comments

Solving a Million-Step LLM Task with Zero Errors

https://arxiv.org/abs/2511.09030
29•Anon84•1h ago•7 comments

Nearly all UK drivers say headlights are too bright

https://www.bbc.com/news/articles/c1j8ewy1p86o
453•YeGoblynQueenne•4h ago•438 comments

How Quake.exe got its TCP/IP stack

https://fabiensanglard.net/quake_chunnel/index.html
362•billiob•9h ago•74 comments

Do Not Put Your Site Behind Cloudflare If You Don't Need To

https://huijzer.xyz/posts/123/do-not-put-your-site-behind-cloudflare-if-you-dont
323•huijzer•5h ago•242 comments

Show HN: Optimizing LiteLLM with Rust – When Expectations Meet Reality

https://github.com/neul-labs/fast-litellm
16•ticktockten•1h ago•3 comments

Google Antigravity

https://antigravity.google/
168•Fysi•2h ago•123 comments

Google Brings Gemini 3 AI Model to Search and AI Mode

https://blog.google/products/search/gemini-3-search-ai-mode/
60•CrypticShift•2h ago•4 comments

Short Little Difficult Books

https://countercraft.substack.com/p/short-little-difficult-books
89•crescit_eundo•3h ago•42 comments

Google Antigravity, a New Era in AI-Assisted Software Development

https://antigravity.google/blog/introducing-google-antigravity
176•meetpateltech•2h ago•130 comments

The Miracle of Wörgl

https://scf.green/story-of-worgl-and-others/
98•simonebrunozzi•7h ago•54 comments

Gemini 3 Pro Model Card

https://pixeldrain.com/u/hwgaNKeH
393•Topfi•6h ago•260 comments

Mathematics and Computation (2019) [pdf]

https://www.math.ias.edu/files/Book-online-Aug0619.pdf
44•nill0•5h ago•9 comments

Beauty in/of mathematics: tessellations and their formulas

https://www.tandfonline.com/doi/full/10.1080/00036811.2025.2510472
14•QueensGambit•5d ago•0 comments

Ruby 4.0.0 Preview2 Released

https://www.ruby-lang.org/en/news/2025/11/17/ruby-4-0-0-preview2-released/
149•pansa2•4h ago•51 comments

Looking for Hidden Gems in Scientific Literature

https://elicit.com/blog/literature-based-discovery
10•ravenical•5d ago•1 comments

GoSign Desktop RCE flaws affecting users in Italy

https://www.ush.it/2025/11/14/multiple-vulnerabilities-gosign-desktop-remote-code-execution/
45•ascii•5h ago•18 comments

How many video games include a marriage proposal? At least one

https://32bits.substack.com/p/under-the-microscope-ncaa-basketball
307•bbayles•5d ago•74 comments

I've Wanted to Play That 'Killer Shark' Arcade Game Briefly Seen in 'Jaws'

https://www.remindmagazine.com/article/15694/jaws-arcade-video-game-killer-shark-atari-sega-elect...
22•speckx•4d ago•8 comments

Langfuse (YC W23) Hiring OSS Support Engineers in Berlin and SF

https://jobs.ashbyhq.com/langfuse/5ff18d4d-9066-4c67-8ecc-ffc0e295fee6
1•clemo_ra•11h ago

The Uselessness of "Fast" and "Slow" in Programming

https://jerf.org/iri/post/2025/the_uselessness_of_fast/
100•zdw•6d ago•52 comments

Azure hit by 15 Tbps DDoS attack using 500k IP addresses

https://www.bleepingcomputer.com/news/microsoft/microsoft-aisuru-botnet-used-500-000-ips-in-15-tb...
454•speckx•1d ago•287 comments

The surprising benefits of giving up

https://nautil.us/the-surprising-benefits-of-giving-up-1248362/
174•jnord•13h ago•139 comments
Open in hackernews

Understanding the Go Scheduler

https://nghiant3223.github.io/2025/04/15/go-scheduler.html
180•gnabgib•6mo ago

Comments

90s_dev•6mo ago
I heard that the scheduler is a huge obstacle to many potential optimizations, is that true?
NAHWheatCracker•6mo ago
In some ways, yes. If you want to optimize at that level you ought to use another language.

I'm not a low level optimization guy, but I've had occasions where I wanted control over which threads my goroutines are running on or prioritizing important goroutines. It's a trade off for making things less complex, which is standard for Go.

I suppose there's always hope that the Go developers can change things.

silisili•6mo ago
You can kinda work around this though. runtime package has a LockOSThread that pins a goroutine to its current thread and prevents others from using it.

If you model it in a way where you have one goroutine per os thread that receives and does work, it gets you close. But in many cases that means rearching the entire code base, as it's not a style I typically reach for.

naikrovek•6mo ago
That sounds a lot like just using another language.
silisili•6mo ago
It's really not that bad. If you have a codebase in Go you can speed up, it's fine.

That said, if you're greenfielding and see this as a limitation to begin with, picking another language is probably the right way.

jerf•5mo ago
If you need it here or there, no. I've got a use case where I need a single locked thread for a particular syscall's functionality. It's not like it leaks out into the rest of the program and everything else has to change to accomodate it.

If you need it pervasively, Go may not be the correct choice. Then again, the list of languages that is not a correct choice in that case is quite long. That's a minority case. An important one, but a minority one.

jasonthorsness•6mo ago
It's always a sign of good design when something as complex as the scheduler described "just works" with the simple abstraction of the goroutine. What a great article.

"1/61 of the time, check the global run queue." Stuff like this is a little odd; I would have thought this would be a variable dependent on the number of physical cores.

01HNNWZ0MV43FF•6mo ago
That's so funny. I just saw `61` in the Tokio code with a comment "copied this from Go"
__turbobrew__•6mo ago
Make sure you set GOMAXPROCS when the runtime is cgroup limited.

I once profiled a slow go program running on a node with 168 cores, but cpu.max was 2 cores for the cgroup. The runtime defaults to set GOMAXPROCS to the number of visible cores which was 168 in this case. Over half the runtime was the scheduler bouncing goroutines between 168 processes despite cpu.max being 2 CPU.

The JRE is smart enough to figure out if it is running in a resource limited cgroup and make sane decisions based upon that, but golang has no such thing.

xyzzy_plugh•6mo ago
Relevant proposal to make GOMAXPROCS cgroup-aware: https://github.com/golang/go/issues/73193
robinhoodexe•6mo ago
Looks like it was just merged btw.
yencabulator•6mo ago
This should be automatic these days (for the basic scenarios).

https://github.com/golang/go/blob/a1a151496503cafa5e4c672e0e...

jasonthorsness•6mo ago
uh isn't that change 3 hours old?
yencabulator•6mo ago
Oh heh yes it is. I just remembered the original discussion from 2019 (https://github.com/golang/go/issues/33803) and grepped the source tree for cgroup to see if that got done or not, but didn't check when it got done.

As said in 2019, import https://github.com/uber-go/automaxprocs to get the functionality ASAP.

jasonthorsness•6mo ago
super-weird coincidence but welcome, I have been waiting for this for a long time!
williamdclt•6mo ago
I honestly can’t count on my fingers and toes how many times something very precisely relevant to me was brought up or sorted out hours-to-days before I looked it up. And more often than once, by people I personally knew!

Always a weird feeling, it’s a small world

formerly_proven•6mo ago
This is probably going to save quadrillions of CPU cycles by making an untold number of deployed Go applications a bit more CPU efficient. Since Go is the "lingua franca" of containers, many ops people assume the Go runtime is container-aware - it's not (well not in any released version, yet).

If they'd now also make the GC respect memory cgroup limits (i.e. automatic GOMEMLIMIT), we'd probably be freeing up a couple petabytes of memory across the globe.

Java has been doing these things for a while, even OpenJDK 8 has had those patches since probably before covid.

mappu•6mo ago
GOMEMLIMIT is not as easy, you may have other processes in the same container/cgroup also using memory.
kunley•6mo ago
As long as I admit respecting cgroup's setting is a good thing, I am not sure it's really quadrillions.

Or is it? Need calculations

formerly_proven•6mo ago
I would've expected it to be either way too much or way too little, but after doing the math it could be sorta in the right ballpark, at least cosmically speaking.

Let's go with three quadrillion (which is apparently 10^15), let's assume a server CPU does 3 GHz (10^9), that's 10^6, a day is about 100k seconds, so ~ten days. But of course we're only saving cycles. I've seen throughput increase by about 50% when setting GOMAXPROCS on bigger machines, but in most of those cases we're looking at containers with fractional cores. On the other hand, there are many containers. So...

kunley•5mo ago
Nice reasoning, thanks.

Hey, but what did you have in mind with regard to bigger machines? I think we're talking here about lowering GOMAXPROCS to have in effect less context switching of the OS threads. While it can bring some good result, a gut feeling is that it'd be hardly 50% faster overall, is your scenario the same then?

01HNNWZ0MV43FF•6mo ago
Trying to see if Rust and Tokio have the same problem. I don't know enough about cgroups to be sure. Tokio at this line [1] ends up delegating to `std::thread::available_parallelism` [2] which says

> It may overcount the amount of parallelism available when limited by a process-wide affinity mask or cgroup quotas and sched_getaffinity() or cgroup fs can’t be queried, e.g. due to sandboxing.

[1] https://docs.rs/tokio/1.45.0/src/tokio/loom/std/mod.rs.html#...

[2] https://doc.rust-lang.org/stable/std/thread/fn.available_par...

nvarsj•6mo ago
Probably not?

The fundamental issue comes down to background GC and CPU quotas in cgroups.

If your number of worker threads is too high, GC will eat up all the quota.

kortex•6mo ago
Fantastic writeup! Visualizations are great, the writeup is thorough but readable.
weiwenhao•6mo ago
Your write-up is so detailed that I even feel like I could implement a complete golang scheduler myself
davidw•6mo ago
I'd be interested in seeing a comparison of this and the BEAM/Erlang/Elixir scheduler by someone paying attention to the details.