frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Atlas Shrugged 2: One Hour Later

https://www.angryflower.com/348.html
1•aebtebeten•1m ago•0 comments

Show HN: Star Trek: The Next Generation Episode Guessing Game

https://tng.episodle.com/
1•francisofascii•1m ago•0 comments

A practical 2026 roadmap for modern AI search and RAG systems

https://nemorize.com/roadmaps/2026-modern-ai-search-rag-roadmap
1•reverseblade2•2m ago•0 comments

Who is going to trust the USA again?

https://www.taxresearch.org.uk/Blog/2026/01/08/who-is-going-to-trust-the-usa-again/
1•treadump•3m ago•0 comments

AI Is Eating SaaS: Building an IP Geolocation API in Two Hours

https://vpetersson.com/2026/01/09/ai-is-eating-saas-building-an-ip-geolocation-api-in-two-hours.html
1•ingve•4m ago•0 comments

X UK revenues drop nearly 60% in a year as content concerns spook advertisers

https://www.theguardian.com/technology/2026/jan/09/x-uk-revenues-drop-nearly-60-in-a-year-as-adve...
1•mindracer•5m ago•0 comments

A slim robot under $10k that can do laundry

https://www.engadget.com/home/smart-home/switchbot-came-to-ces-with-a-laundry-robot-you-might-act...
1•MattSayar•5m ago•0 comments

Blockchains and Australian Coffee Cards

https://rubenerd.com/blockchains-and-australian-coffee-cards/
1•speckx•7m ago•0 comments

Buried Talents: How Elites betray their own potential

https://edankrolewicz.substack.com/p/buried-talents
1•ikjasdlk2234•8m ago•0 comments

Ask HN: How are you monetizing ChatGPT / MCP apps today?

1•ssorokin•9m ago•0 comments

Show HN: CloudyWithAChanceOfLatency: testing the water for a net monitoring app

https://cloudywithachanceoflatency.net
1•rixed•9m ago•0 comments

Show HN: CAML-Lint – a linter for narrative/quest JSON used in game design

https://github.com/dkoepsell/CAML-lint/tree/main
1•KoeppyLoco•15m ago•0 comments

People who come off slimming jabs regain weight four times faster than dieters

https://www.bbc.com/news/articles/c050ljnrv2qo
2•breve•15m ago•0 comments

We built a list for Attack Surface Management

https://github.com/Escape-Technologies/awesome-attack-surface-management
2•Gwendal-M•15m ago•0 comments

A single system to test ideas, content, and campaigns before you risk budget

https://blpg.vect.pro
2•yevdduwi•15m ago•1 comments

Schenker Element 16 a semi-modular laptop with a repairable, customizable design

https://liliputing.com/schenker-element-16-is-a-semi-modular-laptop-with-a-repairable-customizabl...
3•7777777phil•18m ago•0 comments

Life Happens at 1x Speed

https://terriblesoftware.org/2026/01/08/life-happens-at-1x-speed/
2•xngbuilds•19m ago•0 comments

How to AI-proof your job

https://www.ft.com/content/5e2593a3-e834-4822-bbc8-7cb27086af24
2•merksittich•19m ago•0 comments

Waveshare releases 7, 8, and 10" ESP32-P4 tablet with SD card, mic, and speaker

https://www.waveshare.com/esp32-p4-wifi6-touch-lcd-7-8-10.1.htm
2•journal•20m ago•1 comments

Ed Feulner, Ed Meese and the Heritage Foundation's Exodus

https://www.wsj.com/opinion/ed-feulner-ed-meese-and-the-heritage-foundations-exodus-8ab6ae02
3•7777777phil•21m ago•0 comments

Why 4 GPUs trained slower than 1 GPU on budget clouds

https://cortwave.github.io/posts/multi-gpu/
2•cortwave•21m ago•0 comments

Modernized Go Fix

https://antonz.org/accepted/modernized-go-fix/
2•blenderob•22m ago•0 comments

Show HN: Clean HTML for Semantic Extraction

https://page-replica.github.io/pure-html-for-rag/demo/
3•nirvanist•22m ago•1 comments

SanDisk to double price of 3D NAND for enterprise SSDs in Q1 2026

https://www.tomshardware.com/pc-components/ssds/sandisk-to-double-price-of-3d-nand-for-enterprise...
2•speckx•24m ago•0 comments

A poker game written in PicoLisp for the Sensor Watch

https://thegeez.net/2026/01/05/watch_bird_poker_picolisp.html
1•fogus•24m ago•0 comments

USA TODAY mapped the potential consequences of a strike on US missile silos

https://www.usatoday.com/story/news/politics/2026/01/06/nuclear-sponge-project-methodology/874001...
2•perihelions•25m ago•0 comments

Kagi releases alpha version of Orion for Linux

https://help.kagi.com/orion/misc/linux-status.html
4•HelloUsername•27m ago•0 comments

"If Starmer is successful in banning X in Britain, I will move forward in . . ."

https://twitter.com/RepLuna/status/2009460496668426449
4•chrisjj•27m ago•4 comments

A Deep Dive into the Linux Kernel Processes and Syscall [pdf]

https://lass.cs.umass.edu/~shenoy/courses/spring20/lectures/Lec09.pdf
3•7777777phil•27m ago•0 comments

Display Size

1•kilvar•28m ago•0 comments
Open in hackernews

Cancelled 2x Cursor Ultra plans, here's why

8•throwawayround•18h ago
Posting this because it took me way too long to figure out what was going on, and I wish I had seen a post like this earlier.

I just canceled two Cursor Ultra plans. My usage went from a steady ~$60–100/month to $500+ in a few days, projecting ~$1,600/month. Support told me this was “expected.”

I did not suddenly start doing 10x more work.

Cursor shows a 200k context window and says content is summarized to stay within limits. Pricing is shown as $ per million tokens. Based on that, I monitored my call count and thought I was being careful.

What I did not realise: - Cursor builds a very large hidden prompt state: conversation history, tool traces, agent state, extended reasoning, codebase context. - That state is prompt-cached. - On every call, the entire cached prefix is replayed. - Anthropic bills cache read tokens for every replay. - Cache reads are billed even if that content is later summarised or truncated before inference.

So the UI says “max 200k context”, but billing says otherwise

Concrete example from my usage:

MAX mode: off Actual user input: ~4k tokens Cache read tokens: ~21 million Total tokens billed: ~22 million Cost for one call: about $12

Claude never attended to 21M tokens. I still paid for them.

This was not just Opus. It happened with Sonnet too.

Support explained that this is exactly how the API is billed so there wasn't an error and I should just use these models more carefully as they could consume a lot of tokens when they are thinking. But there is a limit to that and what I was charged was way high. There is ZERO transparency about how the cache is used. And the cache breakpoints are decided by Cursor so I don't think it's fair to throw the ball to Anthropic here.

The dangerous part is that cost becomes decoupled from anything you can see or reason about as a user. You think you are operating inside a 200k window, but you are paying for a much larger hidden history being replayed over and over.

I am not claiming a bug in Anthropic’s API. This is a product transparency issue. If a tool can silently turn a few hundred dollars of usage into four figures because of hidden caching behaviour, users need much better visibility and controls. Support suggested spend controls but I am actually complaining about how my pre-paid package was consumed.

If you use Cursor with long-running chats, agents, or large codebases, check your cache read tokens carefully. The UI will not warn you. The only thing you will see is a few days into your subscription "Your are projected to run out of your usage allowance in a few days"

I canceled and moved on, giving Claude Code a shot until this is fixed. Posting so others do not find out the hard way.

Comments

techblueberry•18h ago
I'm curious about the people who are like "Look at my workflow, I run 10x agents at a time", are they just running 10k per month bills to their employer? In which case it's like "Neat, cool story bro"
throwawayround•18h ago
I keep tech bros flexing their coding agent bills every day. "I've spent 10k$ on Cursor today AMA"
bigyabai•18h ago
> I should just use these models more carefully as they could consume a lot of tokens when they are thinking.

Support is right. You cannot ever use a pay-as-you-go service unconsciously. This isn't an all-you-can-eat buffet, they're Tracfone minutes and you'll spend them accordingly.

If you can't square that mentally, maybe the big expensive models aren't a good match for your workflow.

throwawayround•18h ago
If you read my post carefully you can see that the problem is not the cost of the models or the way I use them but something that is not just hidden from the Cursor UI but misleading. It keeps displaying a 200k context limit and also uses summaries to manage the context. You can only see that it pulled 21M tokens worth of data from Anthropic cache for a 4000 token prompt if you export your usage as a csv and analyse.
bigyabai•17h ago
> It keeps displaying a 200k context limit

Some radiometers cap-out at 3.6 roentgen because they assumed that nobody would need a larger figure. That doesn't mean that 3.6 roentgen is the hard-limit on radiation, though.

Again - any PAYG product requires very explicit discipline to avoid getting smacked with a shady bill. AWS, Vercel or Claude.

throwawayround•17h ago
I think we’re talking past each other a bit.

This isn’t really about PAYG discipline or being surprised by a big bill. I had a spend cap. The problem is that, because of how Cursor implements caching, that prepaid usage got consumed in a way that was impossible to see or reason about ahead of time. I didn’t “overspend,” I just got far less actual work out of the same budget than I reasonably should have.

Also, the 200k context cap is a real cap. Claude does not attend to more than that. Cursor even summarizes aggressively to stay within it. The confusing part is that billing is driven by something else entirely: replaying cached prompt state that the UI implies is no longer relevant.

So from a user’s perspective, you’re operating inside a 200k window, MAX is off, prompts are small, and things look normal. Yet each call can still replay tens of millions of cached tokens that you never see and can’t inspect without exporting CSVs after the fact.

That’s very different from “you used more than you thought.” It’s more like the primary cost driver is invisible and contradicts the mental model the product presents. If the UI showed cache size or warned that a call would replay ~20M cached tokens, I’d agree this is just normal PAYG behavior. But it doesn’t.

That’s the issue I’m pointing out.