frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Sping – An HTTP/TCP latency tool that's easy on the eye

https://dseltzer.gitlab.io/sping/docs/
36•zorlack•3h ago•3 comments

The two versions of Parquet

https://www.jeronimo.dev/the-two-versions-of-parquet/
131•tanelpoder•3d ago•31 comments

Busy beaver hunters reach numbers that overwhelm ordinary math

https://www.quantamagazine.org/busy-beaver-hunters-reach-numbers-that-overwhelm-ordinary-math-202...
34•defrost•2d ago•4 comments

Ghrc.io appears to be malicious

https://bmitch.net/blog/2025-08-22-ghrc-appears-malicious/
243•todsacerdoti•3h ago•30 comments

Is 4chan the perfect Pirate Bay poster child to justify wider UK site-blocking?

https://torrentfreak.com/uk-govt-finds-ideal-pirate-bay-poster-boy-to-sell-blocking-of-non-pirate...
172•gloxkiqcza•10h ago•128 comments

Prison isn’t set up for today’s tech so we have to do legal work the old way

https://prisonjournalismproject.org/2025/08/19/prisons-outdated-technology-hurts-our-chances-at-f...
69•danso•3h ago•37 comments

We put a coding agent in a while loop

https://github.com/repomirrorhq/repomirror/blob/main/repomirror.md
127•sfarshid•10h ago•95 comments

Making games in Go: 3 months without LLMs vs. 3 days with LLMs

https://marianogappa.github.io/software/2025/08/24/i-made-two-card-games-in-go/
260•maloga•12h ago•181 comments

My ZIP isn't your ZIP: Identifying and exploiting semantic gaps between parsers

https://www.usenix.org/conference/usenixsecurity25/presentation/you
41•layer8•3d ago•14 comments

A Brilliant and Nearby One-off Fast Radio Burst Localized to 13 pc Precision

https://iopscience.iop.org/article/10.3847/2041-8213/adf62f
48•gnabgib•7h ago•7 comments

Y Combinator files brief supporting Epic Games, says store fees stifle startups

https://www.macrumors.com/2025/08/21/y-combinator-epic-games-amicus-brief/
80•greenburger•3d ago•63 comments

Trees on city streets cope with drought by drinking from leaky pipes

https://www.newscientist.com/article/2487804-trees-on-city-streets-cope-with-drought-by-drinking-...
151•bookofjoe•2d ago•80 comments

How to check if your Apple Silicon Mac is booting securely

https://eclecticlight.co/2025/08/21/how-to-check-if-your-apple-silicon-mac-is-booting-securely/
40•shorden•3h ago•9 comments

Burner Phone 101

https://rebeccawilliams.info/burner-phone-101/
285•CharlesW•4d ago•108 comments

How many paths of length K are there between A and B? (2021)

https://horace.io/walks
16•jxmorris12•7h ago•2 comments

Everything I know about good API design

https://www.seangoedecke.com/good-api-design/
199•ahamez•8h ago•77 comments

Cloudflare incident on August 21, 2025

https://blog.cloudflare.com/cloudflare-incident-on-august-21-2025/
143•achalshah•2d ago•29 comments

Halt and Catch Fire Syllabus (2021)

https://bits.ashleyblewer.com/halt-and-catch-fire-syllabus/
105•Kye•6h ago•27 comments

Show HN: Clearcam – Add AI object detection to your IP CCTV cameras

https://github.com/roryclear/clearcam
164•roryclear•15h ago•47 comments

Using acetaminophen during pregnancy may increase childrens autism and ADHD risk

https://hsph.harvard.edu/news/using-acetaminophen-during-pregnancy-may-increase-childrens-autism-...
10•spchampion2•3h ago•0 comments

GNU cross-tools: musl-cross 313.3M

https://github.com/cross-tools/musl-cross
17•1vuio0pswjnm7•4h ago•2 comments

Iterative DFS with stack-based graph traversal (2024)

https://dwf.dev/blog/2024/09/23/2024/dfs-iterative-stack-based
27•cpp_frog•3d ago•2 comments

NASA's Juno mission leaves legacy of science at Jupiter

https://www.scientificamerican.com/article/how-nasas-juno-probe-changed-everything-we-know-about-...
65•apress•3d ago•27 comments

Stepanov's biggest blunder? The curious case of adjacent difference

https://mmapped.blog/posts/43-stepanovs-biggest-blunder
39•signa11•3d ago•8 comments

Comet AI browser can get prompt injected from any site, drain your bank account

https://twitter.com/zack_overflow/status/1959308058200551721
491•helloplanets•12h ago•173 comments

Bash Strict Mode

http://redsymbol.net/articles/unofficial-bash-strict-mode/
4•dcminter•2d ago•2 comments

OS Yamato lets your data fade away

https://github.com/osyamato/os-yamato
17•tsuyoshi_k•3d ago•13 comments

Claim: GPT-5-pro can prove new interesting mathematics

https://twitter.com/SebastienBubeck/status/1958198661139009862
118•marcuschong•4d ago•79 comments

Show HN: I Built a XSLT Blog Framework

https://vgr.land/content/posts/20250821.xml
31•vgr-land•9h ago•11 comments

Will at centre of legal battle over Shakespeare’s home unearthed after 150 years

https://www.theguardian.com/culture/2025/aug/21/will-at-centre-of-legal-battle-over-shakespeares-...
43•forthelose•1d ago•14 comments
Open in hackernews

Cloudflare incident on August 21, 2025

https://blog.cloudflare.com/cloudflare-incident-on-august-21-2025/
143•achalshah•2d ago

Comments

iqfareez•2d ago
Wild that one tenant’s cache-hit traffic could tip over Cloudflare’s interconnect capacity
immibis•2d ago
You'd be surprised how low the capacity of a lot of internet links is. 10Gbps is common on smaller networks - let me rephrase that, a small to medium ISP might only have 10Gbps to each of most of their peering partners. Normally, traffic is distributed, going to different places, coming from different places, and each link is partially utilized. But unusual patterns can fill up one specific link.

10Gbps is old technology now and any real ISP can probably afford 40 or 100 - for hundreds of dollars per link. But they're going to deploy that on their most utilized links first, and only if their peering partner can also afford it and exchanges enough traffic to justify it. So the smallest connections are typically going to be 10. (Lower than 10 is too small to justify a point-to-point peering at all).

If you have 10Gbps fiber at home, you could congest one of these links all by yourself.

Now this is Cloudflare talking to aws-east-1, so they should have shitloads of capacity there, probably at least 8x100 or more. But considering that AWS is the kind of environment where you can spin up 800 servers for a few hours to perform a massively parallel task, it's not surprising that someone did eventually create 800Gbps of traffic to the same place, or however much they have. Actually it's surprising it doesn't happen more often. Perhaps that's because AWS charges an arm and a leg for data transfer - 800Gbps is $5-$9 per second.

aianus•8h ago
Downloading cached data from Cloudflare to AWS is free to the person doing the downloading if they use Internet gateway
transitionnel•4h ago
Future proofing inevitable things should be something to talk about more.

For instance, people will be scraping at a "growing" rate as they figure out how everything AI works. We might as well figure out some standard seeded data packages for training that ~all sources/sectors agree to make available as public torrents to reduce this type of problem.

[I realize this ask is currently idealistic, but it's an anchor point to negotiate from.]

themafia•7h ago
That's what started the incident.

It was prolonged by the fact that Cloudflare didn't react correctly to withdrawn BGP routes to a major peer, that the secondary routes had reduced capacity due to unaddressed problems, and basic nuisance rate limiting had to be done manually.

It seems like they just build huge peering pipes and basically just hope for the best. They've maybe gotten so used to this working that they'll let degraded "secondary" links persist for much longer than they should. It's the typical "Swiss Cheese" style of failure.

vlovich123•3h ago
Wasn’t the problem exacerbated precisely by withdrawing a BGP link because all the same traffic is then forced over a smaller number of physical links?
miyuru•2d ago
AWS us-east-1 is now taking down other providers.
inemesitaffia•2d ago
Didn't even notice
yaboi3•2d ago
Anyone want to tell Cloudflare that BGP advertisements at AWS are automated and their congested network directly cause BGP withdrawals as the automated system detected congestion and decreased traffic to remediate it?
grumple•8h ago
It wouldn't surprise me if the BGP routes in the DCI PNI were manually configured, since this is probably one of the most direct and important connections. I would be surprised if Cloudflare didn't have firsthand knowledge of what happened with AWS during this incident.

I think the withdrawal approach by AWS would normally work, as this action should desaturate the connections. Just really unfortunate that this caused routing through a link that was at half capacity.

__float•1h ago
The way I read the blog post, it seems they're very aware of that.

I imagine Cloudflare and AWS were on a Chime bridge while this all went down, they both have a lot at stake here.

erulabs•8h ago
It’s gonna turn out it was one guy on one machine calling “pnpm install” on a fast machine with a 100gbps uplink.
cluckindan•7h ago
Can we stop with the 2015 jokes already?
chatmasta•6h ago
I’ve actually had an npm install that failed on my ISP but succeeded with Cloudflare VPN and the OP comment was more or less the explanation.
BoorishBears•1h ago
In 2015 it would have been "npm install"

(Thanks Rauch.)

__turbobrew__•7h ago
> This system will allot network resources on a per-customer basis, creating a budget that, once exceeded, will prevent a customer's traffic from degrading the service for anyone else on the platform

How would this work practically? If a single client is overflowing the edge router queues you are kindof screwed already? Even if you dropped all packets from that client you would need to still process the packets to figure out what client they belong to before dropping the packets?

I guess you could somehow do some shuffle sharding where a single client belongs to a few IP prefixes and when that client misbehaves you withdraw those prefixes using BGP to essentially black hole the network routes for that client. If the shuffle sharding is done right only the problem client will have issues as other clients on the same prefixes will be sharded to other prefixes.

jeffbee•7h ago
Perhaps they drop the client's flows on the host side.
__turbobrew__•7h ago
I don’t understand? The issue is that a client/customer outside of cloudflares control DOSed one of their network links. Cloudflare has no control on the client side to implement rate limiting?
fusl•7h ago
I think you misunderstand the flow of traffic here. The data flow, initiated by requests coming from AWS us-east-1, was Cloudflare towards AWS, not the other way around. Cloudflare can easily control where and how their egress traffic gets to the destination (as long as there are multiple paths towards the target) as well as rate limit that traffic to sane levels.
__turbobrew__•6h ago
Ah I see now. Yes in that case they could just reply with 429 codes or just not reply at all.
everfrustrated•7h ago
I think you're overthinking this. Just having a per (cloudflare) customer rate limit would go a long long way.
milofeynman•1h ago
It's load shedding, but it's weighted towards people abusing their quota usually over some rolling weighted average. The benefit is that they are dropped immediately at the edge rather than holding sockets open or using compute/resources. It usually takes 30s-1m to kick in.
senderista•5h ago
There was definitely a recurring pattern at AWS where a single customer would trigger latent bugs/undercapacity resulting in outages. Postmortems would often recommend developing per-customer observability and mitigation.
md224•4h ago
I'm having trouble understanding the second diagram in the article. I can make sense of a directed graph, but this one has thin horizontal lines with arrows leaving them in both directions. These lines look like dividers, not nodes, so I'm not sure how to interpret it.
pm90•4h ago
Only real long term mitigation is to move to another aws region; us-east-1 seems to suffer from all kinds of scaling challenges.
bastawhiz•4h ago
There's nothing to suggest the link between Cloudflare and any other AWS region has more capacity or that there aren't more disruptive Cloudflare customers using those regions.
o11c•3h ago
But there is absolutely something to suggest "if you only support one region for some tasks, you're going to have problems that other people don't have."
Hilift•4h ago
> The incident was a result of a surge of traffic from a single customer that overloaded Cloudflare's links with AWS us-east-1. It was a network congestion event, not an attack or a BGP hijack.

And no one knew a single thing about it until the incident. That is the current network management state of the art, let Cloudflare deal.

wferrell•3h ago
I wonder which customer triggered this…