frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Ask HN: Claude Opus performance affected by time of day?

23•scaredreally•9h ago•27 comments

Ask HN: Share your personal website

895•susam•2d ago•2319 comments

Tell HN: The way I do simple data management for new prototypes

7•AndreyK1984•14h ago•5 comments

Ask HN: How can we solve the loneliness epidemic?

748•publicdebates•1d ago•1184 comments

Ask HN: How are you doing RAG locally?

391•tmaly•2d ago•151 comments

Ask HN: How have you or your firm made money with LLMs?

7•bwestergard•9h ago•7 comments

Tell HN: YouTube gave my username switzerland to a half government organization

18•faebi•14h ago•5 comments

Ask HN: What did you find out or explore today?

212•blahaj•2d ago•396 comments

At the phase 'build a startup cause I can't get hired, and maybe I'll get hired'

7•danver0•3h ago•1 comments

Ask HN: Those who quit tech, moved back home, what do you do?

12•akudha•5h ago•5 comments

Ask HN: Browser extension vs. native app for structured form filling?

2•livrasand•3h ago•0 comments

Ask HN: One IP, multiple unrealistic locations worldwide hitting my website

41•nacho-daddy•1d ago•24 comments

Ask HN: Who's Using DuckDB in Production?

3•yakkomajuri•5h ago•4 comments

Ask HN: Local media server, receive and send audio?

2•thedangler•6h ago•1 comments

Ask HN: When "Two-Factor Authentication" (2FA) Aren't Two

2•s3131212•6h ago•1 comments

Ask HN: Have you ever tried low-code tools for your work?

3•andre_fernandes•13h ago•1 comments

Ask HN: LLM Poisoning Resources

4•totallygeeky•7h ago•0 comments

Ask HN: Tips for better image generation? I need help

2•gweets•7h ago•1 comments

Why is nobody using this? Full-duplex voice streaming with Gemini Live in React

3•loffloff•7h ago•0 comments

Ask HN: What are your best purchases under $100?

76•krishadi•1d ago•217 comments

Tell HN: HP Ultra G1a Bios Freezing Issue

2•BizarroLand•8h ago•0 comments

Ask HN: Iran's 120h internet shutdown, phones back. How to stay resilient?

113•us321•3d ago•99 comments

Ask HN: Is sending a lot of requests but respecting rate limits DOSing?

2•SpyCoder77•8h ago•0 comments

Ask HN: Analogy of AI IDEs for code vs. "AI IDEs" for personal health data

2•nemath•8h ago•0 comments

Ask HN: How do you safely give LLMs SSH/DB access?

80•nico•2d ago•105 comments

Ask HN: AI music covers in 2026?

16•sexy_seedbox•1d ago•9 comments

Ask HN: What are you working on? (January 2026)

256•david927•5d ago•874 comments

Tell HN: Execution is cheap, ideas matter again

14•keepamovin•1d ago•5 comments

Ask HN: How to make spamming us uncomfortable for LinkedIn and friends?

12•zx8080•1d ago•7 comments

Ask HN: Is token-based pricing making AI harder to use in production?

2•Barathkanna•10h ago•5 comments
Open in hackernews

Ask HN: Claude Opus performance affected by time of day?

23•scaredreally•9h ago
I am a big fan of Claude Opus as it has been very good at understanding feature requests and generally staying consistent with my codebase (completely written from scratch using Opus).

I've noticed recently that when I am using Opus at night (Eastern US), I am seeing it go down extreme rabbit holes on the same types of requests I am putting through on a regular basis. It is more likely to undertake refactors that break the code and then iterates on those errors in a sort of spiral. A request that would normally take 3-4 minutes will turn into a 10 minute adventure before I revert the changes, call out the mistake, and try again. It will happily admit the mistake, but the pattern seems to be consistent.

I haven't performed a like for like test and that would be interesting, but has anyone else noticed the same?

Comments

bayarearefugee•5h ago
I mostly use Gemini, so I can't speak for Claude, but Gemini definitely has variable quality at different times, though I've never bothered to try to find a specific time-of-day pattern to it.

The most reliable time to see it fall apart is when Google makes a public announcement that is likely to cause a sudden influx of people using it.

And there are multiple levels of failure, first you start seeing iffy responses of obvious lesser quality than usual and then if things get really bad you start seeing just random errors where Gemini will suddenly lose all of its context (even on a new chat) or just start failing at the UI level by not bothering to finish answers, etc.

The sort of obvious likely reason for this is when the models are under high load they probably engage in a type of dynamic load balancing where they fall back to lighter models or limit the amount of time/resources allowed for any particular prompt.

kevinsync•5h ago
I suspect they might transparently fall back too; Opus 4.5 has been really reasonable lately, except right after it launched, and also surrounding any service interruptions / problems reported on status.claude.ai -- once those issues resolve, for a few hours the results feel very "Sonnet", and it starts making a lot more mistakes. When that happens, I'll usually just pause Claude and prompt Codex and Gemini with the same issue to see what comes out of the black hole.. then a bit later, Claude mysteriously regains its wits.

I just assume it went to the bar, got wasted, and needed time to sober up!

scaredreally•5h ago
Precisely. Once I point out the fact that it is doing this, it seems to produce better results for a bit before going back to the same.

I jokingly (and not so) thought that it was trained on data that made it think it should be tired at the end of the day.

But it is happening daily and at night.

woleium•4h ago
I find it helps to tell it to take some stimulants
stavros•4h ago
I didn't believe such conspiracy theories, until one day I noticed Sonnet 4.5 (which I had been using for weeks to great success) perform much worse, very visibly so. A few hours later, Opus 4.5 was released.

Now I don't know what to think.

astrange•2h ago
They don't ever fall back to cheaper models silently.

What Anthropic does do is poke the model to tell you to go to bed if you use it too long ("long conversation reminder") which distracts it from actually answering.

Sometimes they do have associations with things like the day of the year and might be lazier some months than others.

janalsncm•5h ago
It’s possible that they could be using fallback models during peak load times (west coast mid day). I assume your traffic would be routed to an east coast data center though. But secretly routing traffic to a worse model is a bit shady so I’d want some concrete numbers to quantify worse performance.
dcre•4h ago
To be clear, the company has very directly denied doing this.
denysvitali•3h ago
They did yes, but should we trust them?

I remember clearly this problem happening in the past, despite their claims. I initially thought it was an elaborate hoax, but it turned out to be factually true in my case.

dcre•2h ago
I tend to think it would be very hard and very risky for large, successful companies to systematically lie about these things without getting caught, and the people who would be doing the lying in this case are not professional liars, they’re engineers who generally seem trustworthy. So yes, if there is a degradation, I think bugs are much more likely than systematic lying.
causal•5h ago
I've had the same suspicion for various providers - if I had time and motivation I would put together a private benchmark that runs hourly and chart performance over time. If anyone wants to do that I'll upvote your Show HN :)
fhk•1h ago
Hold my beer
oncallthrow•5h ago
For what it’s worth, Anthropic very strongly claim that they don’t degrade model performance by time of day [1]. I have no reason to doubt that, imo Anthropic are about as ethical as LLM companies get.

[1] https://www.anthropic.com/engineering/a-postmortem-of-three-...

joshribakoff•4h ago
Banning paying users with no warning doesn’t seem super ethical. Probably not unethical, either, but I would not frame them as “the most ethical”
phist_mcgee•4h ago
I'd say they're about as good as the average billion dollar American tech company when it comes to ethics.
Madmallard•1h ago
Really bizarre to even put ethical anywhere near any AI company, even as a function of comparison. These companies are driving society into the ground.
hagbard_c•5h ago
Simple, the model is tired after a long day of working so it starts making mistakes. Give it some rest and it is ready to serve again.
anonzzzies•5h ago
Many people 'notice' it (on reddit); I notice it too, but it is hard to prove. I tried the same prompt on the same code every 4 hours for 48 hours, the behaviour was slightly different but not worse or much different in time. But then I just work on my normal code, think wtf is it doing now??? look at the time and see it is US day time and stop.

People put forward many theories for this (weaker model routing; be it a different model, Sonnet or Haiku or lower quantized Opus seem the most popular), Anthropic says it is all not happening.

RickS•4h ago
I've certainly noticed some variance from opus. there are times it gets stuck and loops on dumb stuff that would have been frustrating from sonnet 3.5, let alone something as good as opus 4.5 when it's locked in. But it's not obviously correlated with time, I've hit those snags at odd hours, and gotten great perf during peak times. It might just be somewhat variable, or a shitty context.

Now GPT4.1 was another story last year, I remember cooking at 4am pacific and feeling the whole thing slam to a halt as the US east coast came online.

UncleEntity•4h ago
>> ...or a shitty context

This is my guess, sometimes it churns through things without a care in the world and other times is seem to be intentionally annoying to eat up the token quota without doing anything productive.

Kind of have to see which mode it's in before turning it loose unsupervised and keep an eye on it just in case it decides to get stupid and/or lazy.

joshribakoff•4h ago
Yep, i have long felt like i randomly get sonnet results despite opus billing. I try to work odd hours and notice better results.
jgbuddy•4h ago
It seems clear that, rather than throttling, anthropic serves lower quality versions of their models during peak usage to keep up with demand. They refuse to admit it, and it's hard to prove, but these threads consistently happen ~3 months after every single model release.
killingtime74•4h ago
Are you using the API or a subscription?
storus•4h ago
I had something similar with GPT, like a clockwork every day after like 1pm it started producing total garbage. Not sure if our account was A/B tested or they just routed us to some brutal quantization of GPT, or even a completely different model.
DANmode•1h ago
Always Be Collecting (accounts)
botacode•2h ago
My limited understanding here is that usage loads impact model outputs to make them less deterministic (and likely degrading in quality). See: https://thinkingmachines.ai/blog/defeating-nondeterminism-in...
schmookeeg•1h ago
I do think Claude does jiggery pokery with its model quality but I have had Clod appear any time of day.

What i find IS tied to time of day is my own fatigue, my own ability to detect garbage tier code and footguns, and my patience is short so if I am going to start cussing at Clod, it is almost always after 4 when I am trying to close out my day.