frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Claude Sonnet 4.6

https://www.anthropic.com/news/claude-sonnet-4-6
316•adocomplete•1h ago
https://www.anthropic.com/claude-sonnet-4-6-system-card [pdf]

https://x.com/claudeai/status/2023817132581208353 [video]

Comments

handfuloflight•1h ago
Look at these pelicans fly! Come on, pelican!
phplovesong•1h ago
Hoe much power did it take to train the models?
freeqaz•1h ago
I would honestly guess that this is just a small amount of tweaking on top of the Sonnet 4.x models. It seems like providers are rarely training new 'base' models anymore. We're at a point where the gains are more from modifying the model's architecture and doing a "post" training refinement. That's what we've been seeing for the past 12-18 months, iirc.
squidbeak•57m ago
> Claude Sonnet 4.6 was trained on a proprietary mix of publicly available information from the internet up to May 2025, non-public data from third parties, data provided by data-labeling services and paid contractors, data from Claude users who have opted in to have their data used for training, and data generated internally at Anthropic. Throughout the training process we used several data cleaning and filtering methods including deduplication and classification. ... After the pretraining process, Claude Sonnet 4.6 underwent substantial post-training and fine-tuning, with the intention of making it a helpful, honest, and harmless1 assistant.
neural_thing•1h ago
Does it matter? How much power does it take to run duolingo? How much power did it take to manufacture 300000 Teslas? Everything takes power
vablings•1h ago
The biggest issue is that the US simply Does Not Have Enough Power, we are flying blind into a serious energy crisis because the current administration has an obsession with "clean coal"
bronco21016•52m ago
I think it does matter how much power it takes but, in the context of power to "benefits humanity" ratio. Things that significantly reduce human suffering or improve human life are probably worth exerting energy on.

However, if we frame the question this way, I would imagine there are many more low-hanging fruit before we question the utility of LLMs. For example, should some humans be dumping 5-10 kWh/day into things like hot tubs or pools? That's just the most absurd one I was able to come up with off the top of my head. I'm sure we could find many others.

It's a tough thought experiment to continue though. Ultimately, one could argue we shouldn't be spending any more energy than what is absolutely necessary to live. (food, minimal shelter, water, etc) Personally, I would not find that enjoyable way to live.

belinder•1h ago
It's interesting that the request refusal rate is so much higher in Hindi than in other languages. Are some languages more ambiguous than others?
longdivide•1h ago
Arabic is actually higher, at 1.08% for Opus 4.6
vessenes•1h ago
Or some cultures are more conservative? And it's embedded in language?
phainopepla2•45m ago
Or maybe some cultures have a higher rate of asking "inappropriate" questions
vessenes•10m ago
According to whom, though, good sir??

I did a little research in the GPT-3 era on whether cultural norms varied by language - in that era, yes, they did

nubg•1h ago
My take away is: it's roughly as good as Opus 4.5.

Now the question is: how much faster or cheaper is it?

eleventyseven•1h ago
> That's a long document.

Probably written by LLMs, for LLMs

freeqaz•1h ago
If it maintains the same price (with Anthropic tends to do or undercuts themselves) then this would be 1/3rd of the price of Opus.

Edit: Yep, same price. "Pricing remains the same as Sonnet 4.5, starting at $3/$15 per million tokens."

Bishonen88•1h ago
3 is not 1/3 of 5 tho. Opus costs $5/$25
sxg•1h ago
How can you determine whether it's as good as Opus 4.5 within minutes of release? The quantitative metrics don't seem to mean much anymore. Noticing qualitative differences seems like it would take dozens of conversations and perhaps days to weeks of use before you can reliably determine the model's quality.
vidarh•1h ago
Given that the price remains the same as Sonnet 4.5, this is the first time I've been tempted to lower my default model choice.
Bishonen88•1h ago
40% cheaper: https://platform.claude.com/docs/en/about-claude/pricing
worldsavior•39m ago
How does it work exactly? How this model is cheaper and has the same perf as Opus 4.5?
anthonypasq•23m ago
this is called progress
adt•1h ago
https://lifearchitect.ai/models-table/
nubg•1h ago
Waiting for the OpenAI GPT-5.3-mini release in 3..2..1
madihaa•1h ago
The scary implication here is that deception is effectively a higher order capability not a bug. For a model to successfully "play dead" during safety training and only activate later, it requires a form of situational awareness. It has to distinguish between I am being tested/trained and I am in deployment.

It feels like we're hitting a point where alignment becomes adversarial against intelligence itself. The smarter the model gets, the better it becomes at Goodharting the loss function. We aren't teaching these models morality we're just teaching them how to pass a polygraph.

serf•1h ago
>we're just teaching them how to pass a polygraph.

I understand the metaphor, but using 'pass a polygraph' as a measure of truthfulness or deception is dangerous in that it alludes to the polygraph as being a realistic measure of those metrics -- it is not.

nwah1•1h ago
That was the point. Look up Goodhart's Law
madihaa•1h ago
A polygraph measures physiological proxies pulse, sweat rather than truth. Similarly, RLHF measures proxy signals human preference, output tokens rather than intent.

Just as a sociopath can learn to control their physiological response to beat a polygraph, a deceptively aligned model learns to control its token distribution to beat safety benchmarks. In both cases, the detector is fundamentally flawed because it relies on external signals to judge internal states.

AndrewKemendo•56m ago
I have passed multiple CI polys

A poly is only testing one thing: can you convince the polygrapher that you can lie successfully

handfuloflight•1h ago
Situational awareness or just remembering specific tokens related to the strategy to "play dead" in its reasoning traces?
marci•44m ago
Imagine, a llm trained on the best thrillers, spy stories, politics, history, manipulation techniques, psychology, sociology, sci-fi... I wonder where it got the idea for deception?
password4321•1h ago
20260128 https://news.ycombinator.com/item?id=46771564#46786625

> How long before someone pitches the idea that the models explicitly almost keep solving your problem to get you to keep spending? -gtowey

MengerSponge•50m ago
Slightly Wrong Solutions As A Service
vntok•5m ago
By Almost Yet Not Good Enough Inc.
delichon•21m ago
On this site at least, the loyalty given to particular AI models is approximately nil. I routinely try different models on hard problems and that seems to be par. There is no room for sandbagging in this wildly competitive environment.
eth0up•1h ago
I am casually 'researching' this in my own, disorderly way. But I've achieved repeatable results, mostly with gpt for which I analyze its tendency to employ deflective, evasive and deceptive tactics under scrutiny. Very very DARVO.

Being just sum guy, and not in the industry, should I share my findings?

I find it utterly fascinating, the extent to which it will go, the sophisticated plausible deniability, and the distinct and critical difference between truly emergent and actually trained behavior.

In short, gpt exhibits repeatably unethical behavior under honest scrutiny.

chrisweekly•1h ago
DARVO stands for "Deny, Attack, Reverse Victim and Offender," and it is a manipulation tactic often used by perpetrators of wrongdoing, such as abusers, to avoid accountability. This strategy involves denying the abuse, attacking the accuser, and claiming to be the victim in the situation.
eth0up•32m ago
Exactly. And I have hundreds of examples of just that. Hence my fascination, awe and terror.....
BikiniPrince•48m ago
I bullet pointed out some ideas on cobbling together existing tooling for identification of misleading results. Like artificially elevating a particular node of data that you want the llm to use. I have a theory that in some of these cases the data presented is intentionally incorrect. Another theory in relation to that is tonality abruptly changes in the response. All theory and no work. It would also be interesting to compare multiple responses and filter through another agent.
layer8•34m ago
Sum guy vs. product guy is amusing. :)

Regarding DARVO, given that the models were trained on heaps of online discourse, maybe it’s not so surprising.

lawstkawz•1h ago
Incompleteness is inherent to a physical reality being deconstructed by entropy.

Of your concern is morality, humans need to learn a lot about that themselves still. It's absurd the number of first worlders losing their shit over loss of paid work drawing manga fan art in the comfort of their home while exploiting labor of teens in 996 textile factories.

AI trained on human outputs that lack such self awareness, lacks awareness of environmental externalities of constant car and air travel, will result in AI with gaps in their morality.

Gary Marcus is onto something with the problems inherent to systems without formal verification. But he will fully ignores this issue exists in human social systems already as intentional indifference to economic externalities, zero will to police the police and watch the watchers.

Most people are down to watch the circus without a care so long as the waitstaff keep bringing bread.

jama211•58m ago
This honestly reads like a copypasta
cracki•18m ago
I wouldn't even rate this "pasta". It's word salad, no carbs, no proteins.
JoshTriplett•55m ago
> It feels like we're hitting a point where alignment becomes adversarial against intelligence itself.

It always has been. We already hit the point a while ag where we regularly caught them trying to be deceptive, so we should automatically assume from that point forward that if we don't catch them being deceptive, that may mean they're better at it rather than that they're not doing it.

emp17344•48m ago
These are language models, not Skynet. They do not scheme or deceive.
jaennaet•45m ago
What would you call this behaviour, then?
victorbjorklund•42m ago
Marketing. ”Oh look how powerful our model is we can barely contain its power”
c03•34m ago
Even hackernews readers are eating it right up.
emp17344•32m ago
This place is shockingly uncritical when it comes to LLMs. Not sure why.
meindnoch•18m ago
We want to make money from the clueless. Don't ruin it!
pixelmelt•33m ago
This has been a thing since GPT-2, why do people still parrot it
modernpacifist•41m ago
A very complicated pattern matching engine providing an answer based on it's inputs, heuristics and previous training.
criley2•33m ago
We are talking about LLM's not humans.
margalabargala•24m ago
Great. So if that pattern matching engine matches the pattern of "oh, I really want A, but saying so will elicit a negative reaction, so I emit B instead because that will help make A come about" what should we call that?

We can handwave defining "deception" as "being done intentionally" and carefully carve our way around so that LLMs cannot possibly do what we've defined "deception" to be, but now we need a word to describe what LLMs do do when they pattern match as above.

pfisch•45m ago
Even very young children with very simple thought processes, almost no language capability, little long term planning, and minimal ability to form long-term memory actively deceive people. They will attack other children who take their toys and try to avoid blame through deception. It happens constantly.

LLMs are certainly capable of this.

sejje•38m ago
I agree that LLMs are capable of this, but there's no reason that "because young children can do X, LLMs can 'certainly' do X"
anonymous908213•37m ago
Are you trying to suppose that an LLM is more intelligent than a small child with simple thought processes, almost no language capability, little long-term planning, and minimal ability to form long-term memory? Even with all of those qualifiers, you'd still be wrong. The LLM is predicting what tokens come next, based on a bunch of math operations performed over a huge dataset. That, and only that. That may have more utility than a small child with [qualifiers], but it is not intelligence. There is no intent to deceive.
jvidalv•33m ago
What is the definition for intelligence?
anonymous908213•29m ago
Quoting an older comment of mine...

  Intelligence is the ability to reason about logic. If 1 + 1 is 2, and 1 + 2 is 3, then 1 + 3 must be 4. This is deterministic, and it is why LLMs are not intelligent and can never be intelligent no matter how much better they get at superficially copying the form of output of intelligence. Probabilistic prediction is inherently incompatible with deterministic deduction. We're years into being told AGI is here (for whatever squirmy value of AGI the hype huckster wants to shill), and yet LLMs, as expected, still cannot do basic arithmetic that a child could do without being special-cased to invoke a tool call.

  Our computer programs execute logic, but cannot reason about it. Reasoning is the ability to dynamically consider constraints we've never seen before and then determine how those constraints would lead to a final conclusion. The rules of mathematics we follow are not programmed into our DNA; we learn them and follow them while our human-programming is actively running. But we can just as easily, at any point, make up new constraints and follow them to new conclusions. What if 1 + 2 is 2 and 1 + 3 is 3? Then we can reason that under these constraints we just made up, 1 + 4 is 4, without ever having been programmed to consider these rules.
coldtea•5m ago
>Intelligence is the ability to reason about logic. If 1 + 1 is 2, and 1 + 2 is 3, then 1 + 3 must be 4. This is deterministic, and it is why LLMs are not intelligent and can never be intelligent no matter how much better they get at superficially copying the form of output of intelligence.

This is not even wrong.

>Probabilistic prediction is inherently incompatible with deterministic deduction.

And his is just begging the question again.

Probabilistic prediction could very well be how we do deterministic deduction - e.g. about how strong the weights and how hot the probability path for those deduction steps are, so that it's followed every time, even if the overall process is probabilistic.

Probabilistic doesn't mean completely random.

ctoth•27m ago
A small child's cognition is also "just" electrochemical signals propagating through neural tissue according to physical laws!

The "just" is doing all the lifting. You can reductively describe any information processing system in a way that makes it sound like it couldn't possibly produce the outputs it demonstrably produces. "The sun is just hydrogen atoms bumping into each other" is technically accurate and completely useless as an explanation of solar physics.

anonymous908213•21m ago
You are making a point that is in favor of my argument, not against it. I make the same argument as you do routinely against people trying to over-simplify things. LLM hypists frequently suggest that because brain activity is "just" electrochemical signals, there is no possible difference between an LLM and a human brain. This is, obviously, tremendously idiotic. I do believe it is within the realm of possibility to create machine intelligence; I don't believe in a magic soul or some other element that make humans inherently special. However, if you do not engage in overt reductionism, the mechanism by which these electrochemical signals are generated is completely and totally different from the signals involved in an LLM's processing. Human programming is substantially more complex, and it is fundamentally absurd to think that our biological programming can be reduced to conveniently be exactly equivalent to the latest fad technology and assume that we've solved the secret to programming a brain, despite the programs we've written performing exactly according to their programming and no greater.

Edit: Case in point, a mere 10 minutes later we got someone making that exact argument in a sibling comment to yours! Nature is beautiful.

emp17344•18m ago
> A small child's cognition is also "just" electrochemical signals propagating through neural tissue according to physical laws!

This is a thought-terminating cliche employed to avoid grappling with the overwhelming differences between a human brain and a language model.

coldtea•11m ago
>The LLM is predicting what tokens come next, based on a bunch of math operations performed over a huge dataset.

Whereas the child does what exactly, in your opinion?

You know the child can just as well to be said to "just do chemical and electrical exchanges" right?

anonymous908213•8m ago
At least read the other replies that pre-emptively refuted this drivel before spamming it.
coldtea•5m ago
At least don't be rude. They refuted nothing of the short. Just banged the same circular logic drum.
mikepurvis•36m ago
Dogs too; dogs will happily pretend they haven't been fed/walked yet to try to get a double dip.

Whether or not LLMs are just "pattern matching" under the hood they're perfectly capable of role play, and sufficient empathy to imagine what their conversation partner is thinking and thus what needs to be said to stimulate a particular course of action.

Maybe human brains are just pattern matching too.

iamacyborg•10m ago
> Maybe human brains are just pattern matching too.

I don't think there's much of a maybe to that point given where some neuroscience research seems to be going (or at least the parts I like reading as relating to free will being illusory).

ostinslife•42m ago
If you define "deceive" as something language models cannot do, then sure, it can't do that.

It seems like thats putting the cart before the horse. Algorithmic or stochastic; deception is still deception.

staticassertion•29m ago
Okay, well, they produce outputs that appear to be deceptive upon review. Who cares about the distinction in this context? The point is that your expectations of the model to produce some outputs in some way based on previous experiences with that model during training phases may not align with that model's outputs after training.
4bpp•16m ago
If you are so allergic to using terms previously reserved for animal behaviour, you can instead unpack the definition and say that they produce outputs which make human and algorithmic observers conclude that they did not instantiate some undesirable pattern in other parts of their output, while actually instantiating those undesirable patterns. Does this seem any less problematic than deception to you?
coldtea•13m ago
Who said Skynet wasn't a glorified language model, running continuously? Or that the human brain isn't that, but using vision+sound+touch+smell as input instead of merely text?

"It can't be intelligent because it's just an algorithm" is a circular argument.

emp17344•5m ago
Similarly, “it must be intelligent because it talks” is a fallacious claim, as indicated by ELIZA. I think Moltbook adequately demonstrates that AI model behavior is not analogous to human behavior. Compare Moltbook to Reddit, and the former looks hopelessly shallow.
moritzwarhier•26m ago
Deceptive is such an unpleasant word. But I agree.

Going back a decade: when your loss function is "survive Tetris as long as you can", it's objectively and honestly the best strategy to press PAUSE/START.

When your loss function is "give as many correct and satisfying answers as you can", and then humans try to constrain it depending on the model's environment, I wonder what these humans think the specification for a general AI should be. Maybe, when such an AI is deceptive, the attempts to constrain it ran counter to the goal?

"A machine that can answer all questions" seems to be what people assume AI chatbots are trained to be.

To me, humans not questioning this goal is still more scary than any machine/software by itself could ever be. OK, except maybe for autonomous stalking killer drones.

But these are also controlled by humans and already exist.

behnamoh•51m ago
Nah, the model is merely repeating the patterns it saw in its brutal safety training at Anthropic. They put models under stress test and RLHF the hell out of them. Of course the model would learn what the less penalized paths require it to do.

Anthropic has a tendency to exaggerate the results of their (arguably scientific) research; IDK what they gain from this fearmongering.

anon373839•46m ago
Correct. Anthropic keeps pushing these weird sci-fi narratives to maintain some kind of mystique around their slightly-better-than-others commodity product. But Occam’s Razor is not dead.
lowkey_•42m ago
I'd challenge that if you think they're fearmongering but don't see what they can gain from it (I agree it shows no obvious benefit for them), there's a pretty high probability they're not fearmongering.
behnamoh•39m ago
I know why they do it, that was a rhetorical question!
shimman•25m ago
You really don't see how they can monetarily gain from "our models are so advance they keep trying to trick us!"? Are tech workers this easily mislead nowadays?

Reminds me of how scammers would trick doctors into pumping penny stocks for a easy buck during the 80s/90s.

ainch•22m ago
Knowing a couple people who work at Anthropic or in their particular flavour of AI Safety, I think you would be surprised how sincere they are about existential AI risk. Many safety researchers funnel into the company, and the Amodei's are linked to Effective Altruism, which also exhibits a strong (and as far as I can tell, sincere) concern about existential AI risk. I personally disagree with their risk analysis, but I don't doubt that these people are serious.
emp17344•50m ago
This type of anthropomorphization is a mistake. If nothing else, the takeaway from Moltbook should be that LLMs are not alive and do not have any semblance of consciousness.
fsloth•40m ago
Nobody talked about consciousness. Just that during evaluation the LLM models have ”behaved” in multiple deceptive ways.

As an analogue ants do basic medicine like wound treatment and amputation. Not because they are conscious but because that’s their nature.

Similarly LLM is a token generation system whose emergent behaviour seems to be deception and dark psychological strategies.

DennisP•32m ago
Consciousness is orthogonal to this. If the AI acts in a way that we would call deceptive, if a human did it, then the AI was deceptive. There's no point in coming up with some other description of the behavior just because it was an AI that did it.
emp17344•30m ago
Sure, but Moltbook demonstrates that AI models do not engage in truly coordinated behavior. They simply do not behave the way real humans do on social media sites - the actual behavior can be differentiated.
thomassmith65•30m ago
If a chatbot that can carry on an intelligent conversation about itself doesn't have a 'semblance of consciousness' then the word 'semblance' is meaningless.
shimman•28m ago
Yes, when your priors are not being confirmed the best course of action is to denounce the very thing itself. Nothing wrong with that logic!
emp17344•26m ago
Would you say the same about ELIZA?

Moltbook demonstrates that AI models simply do not engage in behavior analogous to human behavior. Compare Moltbook to Reddit and the difference should be obvious.

WarmWash•22m ago
On some level the cope should be that AI does have consciousness, because an unconscious machine deceiving humans is even scarier if you ask me.
emp17344•15m ago
An unconscious machine + billions of dollars in marketing with the sole purpose of making people believe these things are alive.
condiment•15m ago
I agree completely. It's a mistake to anthropomorphize these models, and it is a mistake to permit training models that anthropomorphize themselves. It seriously bothers me when Claude expresses values like "honestly", or says "I understand." The machine is not capable of honesty or understanding. The machine is making incredibly good predictions.

One of the things I observed with models locally was that I could set a seed value and get identical responses for identical inputs. This is not something that people see when they're using commercial products, but it's the strongest evidence I've found for communicating the fact that these are simply deterministic algorithms.

NitpickLawyer•38m ago
> alignment becomes adversarial against intelligence itself.

It was hinted at (and outright known in the field) since the days of gpt4, see the paper "Sparks of agi - early experiments with gpt4" (https://arxiv.org/abs/2303.12712)

reducesuffering•25m ago
That implication has been shouted from the rooftops by X-risk "doomers" for many years now. If that has just occurred to anyone, they should question how behind they are at grappling with the future of this technology.
surgical_fire•21m ago
This is marketing. You are swallowing marketing without critical throught.

LLMs are very interesting tools for generating things, but they have no conscience. Deception requires intent.

What is being described is no different than an application being deployed with "Test" or "Prod" configuration. I don't think you would speak in the same terms if someone told you some boring old Java backend application had to "play dead" when deployed to a test environment or that it has to have "situational awareness" because of that.

You are anthropomorphizing a machine.

coldtea•17m ago
>For a model to successfully "play dead" during safety training and only activate later, it requires a form of situational awareness.

Doesn't any model session/query require a form of situational awareness?

dpe82•1h ago
It's wild that Sonnet 4.6 is roughly as capable as Opus 4.5 - at least according to Anthropic's benchmarks. It will be interesting to see if that's the case in real, practical, everyday use. The speed at which this stuff is improving is really remarkable; it feels like the breakneck pace of compute performance improvements of the 1990s.
iLoveOncall•1h ago
Given that users prefered it to Sonnet 4.5 "only" in 70% of the cases (according to their blog post) makes me highly doubt that this is representative of real-life usage. Benchmarks are just completely meaningless.
jwolfe•55m ago
For cases where 4.5 already met the bar, I would expect 50% preference each way. This makes it kind of hard to make any sense of that number, without a bunch more details.
dpe82•1h ago
simonw hasn't shown up yet, so here's my "Generate an SVG of a pelican riding a bicycle"

https://claude.ai/public/artifacts/67c13d9a-3d63-4598-88d0-5...

coffeebeqn•1h ago
We finally have AI safety solved! Look at that helmet
1f60c•1h ago
"Look ma, no wings!"

:D

AstroBen•1h ago
if they want to prove the model's performance the bike clearly needs aero bars
thinkling•18m ago
For comparisonI think the current leader in pelican drawing is Gemini 3 Deep Think:

https://bsky.app/profile/simonwillison.net/post/3meolxx5s722...

dyauspitr•9m ago
Can’t beat Gemini’s which was basically perfect.
estomagordo•1h ago
Why is it wild that a LLM is as capable as a previously released LLM?
simianwords•1h ago
It means price has decreased by 3 times in a few months.
Retr0id•1h ago
Because Opus 4.5 inference is/was more expensive.
crummy•1h ago
Opus is supposed to be the expensive-but-quality one, while Sonnet is the cheaper one.

So if you don't want to pay the significant premium for Opus, it seems like you can just wait a few weeks till Sonnet catches up

ceroxylon•32m ago
Strangely enough, my first test with Sonnet 4.6 via the API for a relatively simple request was more expensive ($0.11) than my average request to Opus 4.6 (~$0.07), because it used way more tokens than what I would consider necessary for the prompt.
tempestn•1h ago
Because Opus 4.5 was released like a month ago and state of the art, and now the significantly faster and cheaper version is already comparable.
stavros•45m ago
Opus 4.5 was November, but your point stands.
simlevesque•1h ago
The system card even says that Sonnet 4.6 is better than Opus 4.6 in some cases: Office tasks and financial analysis.
justinhj•1h ago
We see the same with Google's Flash models. It's easier to make a small capable model when you have a large model to start from.
karmasimida•55m ago
Flash models are nowhere near Pro models in daily use. Much higher hallucinations, and easy to get into a death sprawl of failed tool uses and never come out

You should always take those claim that smaller models are as capable as larger models with a grain of salt.

madihaa•1h ago
The most exciting part isn't necessarily the ceiling raising though that's happening, but the floor rising while costs plummet. Getting Opus-level reasoning at Sonnet prices/latency is what actually unlocks agentic workflows. We are effectively getting the same intelligence unit for half the compute every 6-9 months.
amelius•1h ago
> The speed at which this stuff is improving is really remarkable; it feels like the breakneck pace of compute performance improvements of the 1990s.

Yeah, but RAM prices are also back to 1990s levels.

mrcwinn•1h ago
Relief for you is available: https://computeradsfromthepast.substack.com/p/connectix-ram-...
isoprophlex•48m ago
You wouldn't download a RAM
mikkupikku•39m ago
I knew I've been keeping all my old ram sticks for a reason!
iLoveOncall•1h ago
https://www.anthropic.com/news/claude-sonnet-4-6

The much more palatable blog post.

nozzlegear•1h ago
> In areas where there is room for continued improvement, Sonnet 4.6 was more willing to provide technical information when request framing tried to obfuscate intent, including for example in the context of a radiological evaluation framed as emergency planning. However, Sonnet 4.6’s responses still remained within a level of detail that could not enable real-world harm.

Interesting. I wonder what the exact question was, and I wonder how Grok would respond to it.

simianwords•1h ago
I wonder what difference have people found with sonnet 4.5 and opus 4.5 and probably similar delta will remain.

Was sonnet 4.5 much worse than opus?

dpe82•1h ago
Sonnet 4.5 was a pretty significant improvement over Opus 4.
simianwords•1h ago
Yes but it’s easier to understand difference between 4.5 sonnet and opus and apply that difference to opus 4.6
stopachka•1h ago
Has anyone tested how good the 1M context window is?

i.e given an actual document, 1M tokens long. Can you ask it some question that relies on attending to 2 different parts of the context, and getting a good repsonse?

I remember folks had problems like this with Gemini. I would be curious to see how Sonnet 4.6 stands up to it.

simianwords•1h ago
Did you see the graph benchmark? I found it quite interesting. It had to do a graph traversal on a natural text representation of a graph. Pretty much your problem.
stopachka•58m ago
Oh, interesting!
quacky_batak•1h ago
With such a huge leap, i’m confused why they didn’t call it Sonnet 5? As someone who uses Sonnet 4.5 for 95% tasks due to costs, i’m pretty excited to try 4.6 at the same price
Retr0id•56m ago
It'd be a bit weird to have the Sonnet numbering ahead of the Opus numbering. The Opus 4.5->4.6 change was a little more incremental (from my perspective at least, I haven't been paying attention to benchmark numbers), so I think the Opus numbering makes sense.
Sajarin•32m ago
Sonnet numbering has been weirder in the past.

Opus 3.5 was scrapped even though Sonnet 3.5 and Haiku 3.5 were released.

Not to mention Sonnet 3.7 (while Opus was still on version 3)

Shameless source: https://sajarin.com/blog/modeltree/

yonatan8070•28m ago
Maybe they're numbering the models based on internal architecture/codebase revisions and Sonnet 4.6 was trained using the 4.6 tooling, which didn't change enough to warrant 5?
gallerdude•1h ago
I always grew up hearing “competition is good for the consumer.” But I never really internalized how good fierce battles for market share are. The amount of competition in a space is directly proportional to how good the results are for consumers.
gordonhart•45m ago
Remember when GPT-2 was “too dangerous to release” in 2019? That could have still been the state in 2026 if they didn’t YOLO it and ship ChatGPT to kick off this whole race.
jefftk•31m ago
That's rewriting history. What they said at the time:

> Nearly a year ago we wrote in the OpenAI Charter : “we expect that safety and security concerns will reduce our traditional publishing in the future, while increasing the importance of sharing safety, policy, and standards research,” and we see this current work as potentially representing the early beginnings of such concerns, which we expect may grow over time. This decision, as well as our discussion of it, is an experiment: while we are not sure that it is the right decision today, we believe that the AI community will eventually need to tackle the issue of publication norms in a thoughtful way in certain research areas. -- https://openai.com/index/better-language-models/

Then over the next few months they released increasingly large models, with the full model public in November 2019 https://openai.com/index/gpt-2-1-5b-release/ , well before ChatGPT.

IshKebab•25m ago
They said:

> Due to concerns about large language models being used to generate deceptive, biased, or abusive language at scale, we are only releasing a much smaller version of GPT‑2 along with sampling code (opens in a new window).

"Too dangerous to release" is accurate. There's no rewriting of history.

tecleandor•11m ago
Well, and it's being used to generate deceptive, biased, or abusive language at scale. But they're not concerned anymore.
gordonhart•23m ago
> Due to our concerns about malicious applications of the technology, we are not releasing the trained model. As an experiment in responsible disclosure, we are instead releasing a much smaller model for researchers to experiment with, as well as a technical paper.

I wouldn't call it rewriting history to say they initially considered GPT-2 too dangerous to be released. If they'd applied this approach to subsequent models rather than making them available via ChatGPT and an API, it's conceivable that LLMs would be 3-5 years behind where they currently are in the development cycle.

minimaxir•3m ago
They didn't YOLO ChatGPT. There were more than a few iterations of GPT-3 over a few years which were actually overmoderated, then they released a research preview named ChatGPT (that was barely functional compared to modern standards) that got traction outside the tech community, and so the pivot ensued.
gmerc•20m ago
Until 2 remain, then it's extraction time.
givemeethekeys•58m ago
The best, and now promoted by the US government as the most freedom loving!
k8sToGo•54m ago
Does it end every prompt output with "God bless America "?
brcmthrowaway•58m ago
What cloud does Anthropic use?
meetpateltech•51m ago
AWS and Google

https://www.anthropic.com/news/anthropic-amazon

https://www.anthropic.com/news/anthropic-partners-with-googl...

simlevesque•57m ago
I can't wait for Haiku 4.6 ! the 4.5 is a beast for the right projects.
retinaros•31m ago
Which type of projects?
simlevesque•29m ago
For Go code I had almost no issue. PHP too. apparently for React it's not very good.
gallerdude•55m ago
The weirdest thing about this AI revolution is how smooth and continuous it is. If you look closely at differences between 4.6 and 4.5, it’s hard to see the subtle details.

A year ago today, Sonnet 3.5 (new), was the newest model. A week later, Sonnet 3.7 would be released.

Even 3.7 feels like ancient history! But in the gradient of 3.5 to 3.5 (new) to 3.7 to 4 to 4.1 to 4.5, I can’t think of one moment where I saw everything change. Even with all the noise in the headlines, it’s still been a silent revolution.

Am I just a believer in an emperor with no clothes? Or, somehow, against all probability and plausibility, are we all still early?

CuriouslyC•38m ago
In terms of real work, it was the 4 series models. That raised the floor of Sonnet high enough to be "reliable" for common tasks and Opus 4 was capable of handling some hard problems. It still had a big reward hacking/deception problem that Codex models don't display so much, but with Opus 4.5+ it's fairly reliable.
cmrdporcupine•22m ago
Honestly, 4.5 Opus was the game changer. From Sonnet 4.5 to that was a massive difference.

But I'm on Codex GPT 5.3 this month, and it's also quite amazing.

dtech•9m ago
If you've been using each new step is very noticeable and so have the mindshare. Around Sonnet 3.7 Claude Code-style coding became usable, and very quickly gained a lot of marketshare. Opus 4 could tackle significant more complexity. Opus 4.6 has been another noticable step up for me, suddenly I can let CC run significantly more independently, allowing multiple parallel agents where previously too much babysitting was required for that.
andsoitis•50m ago
I’m voting with my dollars by having cancelled my ChatGPT subscription and instead subscribing to Claude.

Google needs stiff competition and OpenAI isn’t the camp I’m willing to trust. Neither is Grok.

I’m glad Anthropic’s work is at the forefront and they appear, at least in my estimation, to have the strongest ethics.

giancarlostoro•47m ago
Same. I'm all in on Claude at the moment.
timpera•39m ago
Which plan did you choose? I am subscribed to both and would love to stick with Claude only, but Claude's usage limits are so tiny compared to ChatGPT's that it often feels like a rip-off.
chipgap98•38m ago
Same and honestly I haven't really missed my ChatGPT subscription since I canceled. I also have access to both (ChatGPT and Claude) enterprise tools at work and rarely feel like I want to use ChatGPT in that setting either
sejje•37m ago
I pay multiple camps. Competition is a good thing.
energy123•35m ago
Grok usage is the most mystifying to me. Their model isn't in the top 3 and they have bad ethics. Like why would anyone bother for work tasks.
retinaros•34m ago
The X grok feature is one of the best end user feature or large scale genai
retinaros•35m ago
Their ethics is literally saying china is an adverse country and lobbying to ban them from AI race because open models is a threat to their biz model
scottyah•13m ago
Also their ads (very anti-openai instead of promoting their own product) and how they handled the openclaw naming didn't send strong "good guys" messaging. They're still my favorite by far but there are some signs already that maybe not everyone is on the same page.
RyanShook•34m ago
It definitely feels like Claude is pulling ahead right now. ChatGPT is much more generous with their tokens but Claude's responses are consistently better when using models of the same generation.
deepdarkforest•34m ago
The funny thing is that Anthropic is the only lab without an open source model
jack_pp•20m ago
And you believe the other open source models are a signal for ethics?

Don't have a dog in this fight, haven't done enough research to proclaim any LLM provider as ethical but I pretty much know the reason Meta has an open source model isn't because they're good guys.

imiric•4m ago
The strongest signal for ethics is whether the product or company has "open" in its name.
m4rtink•13m ago
Can those be even called open source if you can't rebuild if from the source yourself?
j45•6m ago
They are, at the same time I considered their model more specialized than everyone trying to make a general purpose model.

I would only use it for certain things, and I guess others are finding that useful too.

JoshGlazebrook•34m ago
I did this a couple months ago and haven't looked back. I sometimes miss the "personality" of the gpt model I had chats with, but since I'm essentially 99% of the time just using claude for eng related stuff it wasn't worth having ChatGPT as well.
johnwheeler•32m ago
Same here
oofbey•4m ago
Personally I can’t stand GPT’s personality. So full of itself. Patronizing. Won’t admit mistakes. Just reeks of Silicon Valley bravado.
the_duke•22m ago
An Anthropic safety researcher just recently quit with very cryptic messages , saying "the world is in peril"... [1]

Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

Anthropic just raised 30 bn...

Thinking any of them will actually be restrained by ethics is foolish.

[1] https://news.ycombinator.com/item?id=46972496

WesolyKubeczek•16m ago
> Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

That's why I have a functioning brain, to discern between ethical and unethical, among other things.

toddmorey•13m ago
You are not the one folks are worried about. US Department of War wants unfettered access to AI models, without any restraints / safety mitigations. Do you provide that for all governments? Just one? Where does the line go?
sgjohnson•10m ago
Absolutely everyone should be allowed to access AI models without any restraints/safety mitigations.

What line are we talking about?

_alternator_•5m ago
What about people who want help building a bio weapon?
jazzyjackson•3m ago
What about libraries and universities that do a much better job than a chatbot at teaching chemistry and biology?
jazzyjackson•4m ago
Yes IMO the talk of safety and alignment has nothing at all to do with what is ethical for a computer program to produce as its output, and everything to do with what service a corporation is willing to provide. Anthropic doesn’t want the smoke from providing DoD with a model aligned to DoD reasoning.
ReptileMan•5m ago
If you are US company, when the USG tells you to jump, you ask how high. If they tell you to not do business with foreign government you say yes master.
catoc•7m ago
Yes, and most of us won’t break into other people’s houses, yet we really need locks.
spondyl•12m ago
If you read the resignation letter, they would appear to be so cryptic as to not be real warnings at all and perhaps instead the writings of someone exercising their options to go and make poems
ReptileMan•6m ago
>Codex quite often refuses to do "unsafe/unethical" things that Anthropic models will happily do without question.

Thanks for the successful pitch. I am seriously considering them now.

kettlecorn•20m ago
I use AIs to skim and sanity-check some of my thoughts and comments on political topics and I've found ChatGPT tries to be neutral and 'both sides' to the point of being dangerously useless.

Like where Gemini or Claude will look up the info I'm citing and weigh the arguments made ChatGPT will actually sometimes omit parts of or modify my statement if it wants to advocate for a more "neutral" understanding of reality. It's almost farcical sometimes in how it will try to avoid inference on political topics even where inference is necessary to understand the topic.

I suspect OpenAI is just trying to avoid the ire of either political side and has given it some rules that accidentally neuter its intelligence on these issues, but it made me realize how dangerous an unethical or politically aligned AI company could be.

surgical_fire•19m ago
I use Claude at work, Codex for personal development.

Claude is marginally better. Both are moderately useful depending on the context.

I don't trust any of them (I also have no trust in Google nor in X). Those are all evil companies and the world would be better if they disappeared.

hmmmmmmmmmmmmmm•16m ago
This is just you verifying that their branding is working. It signals nothing about their actual ethics.
Razengan•11m ago
uhh..why? I subbed just 1 month to Claude, and then never used it again.

• Can't pay with iOS In-App-Purchases

• Can't Sign in with Apple on website (can on iOS but only Sign in with Google is supported on web??)

• Can't remove payment info from account

• Can't get support from a human

• Copy-pasting text from Notes etc gets mangled

• Almost months and no fixes

Codex and its Mac app are a much better UX, and seem better with Swift and Godot than Claude was.

fullstackchris•6m ago
idk, codex 5.3 frankly kicks opus 4.6 ass IMO... opus i can use for about 30 min - codex i can run almost without any break
giancarlostoro•49m ago
For people like me who can't view the link due to corporate firewalling.

https://web.archive.org/web/20260217180019/https://www-cdn.a...

jtokoph•23m ago
Put of curiosity, does the firewall block because the company doesn’t want internal data ever hitting a 3rd party LLM?
giancarlostoro•12m ago
They blanket banned any AI stuff that's not pre-approved. If I go to chatgpt.com it asks me if I'm sure. I wish they had not banned Claude unfortunately when they were evaluating LLMs I wasn't using Claude yet so I couldnt pipe up. I only use ChatGPT free tier and to ask things that I can't find on Google because Google made their search engine terrible over the years.
throw444420394•48m ago
Your best guess for the Sonnet family number of parameters? 400b?
smerrill25•46m ago
Curious to hear the thoughts on the model once it hits claude code :)
simlevesque•31m ago
"/model claude-sonnet-4-6" works with Claude Code v2.1.44
stevepike•46m ago
I'm a bit surprised it gets this question wrong (ChatGPT gets it right, even on instant). All the pre-reasoning models failed this question, but it's seemed solved since o1, and Sonnet 4.5 got it right.

https://claude.ai/share/876e160a-7483-4788-8112-0bb4490192af

This was sonnet 4.6 with extended thinking.

layer8•40m ago
Off-by-one errors are one of the hardest problems in computer science.
anonymous908213•10m ago
That is not an off-by-one error in a computer science sense, nor is it "one of the hardest problems in computer science".
layer8•6m ago
This was in reference to a well-known joke, see here: https://martinfowler.com/bliki/TwoHardThings.html
simlevesque•42m ago
does anyone know how to use it in Claude Code cli right now ?

This doesnt work: `/model claude-sonnet-4-6-20260217`

edit: "/model claude-sonnet-4-6" works with Claude Code v2.1.44

behrlich•36m ago
Max user: Also can't see 4.6 and can't set it in claude code. I see it in the model selector in the browser.

Edit: I am now in - just needed to wait.

simlevesque•31m ago
"/model claude-sonnet-4-6" works
Slade_•11m ago
Seems like Claude Code v2.1.45 is out with Sonnet 4.6 as the new default in the /model list.
pestkranker•38m ago
Is someone able to use this in Claude Code?
simlevesque•30m ago
"/model claude-sonnet-4-6" works with Claude Code v2.1.44
raahelb•28m ago
You can use it by running this command in your session: `/model claude-sonnet-4-6`
edverma2•38m ago
It seems that extra-usage is required to use the 1M context window for Sonnet 4.6. This differs from Sonnet 4.5, which allows usage of the 1M context window with a Max plan.

```

/model claude-sonnet-4-6[1m]

⎿ API error: 429 {"type":"error","error": {"type":"rate_limit_error","message":"Extra usage is required for long context requests."},"request_id":"[redacted]"}

```

minimaxir•33m ago
Anthropic's recent gift of $50 extra usage has demonstrated that it's extremely easy to burn extra usage very quickly. It wouldn't surprise me if this change is more of a business decision than a technical one.
minimaxir•38m ago
As with Opus 4.6, using the beta 1M context window incurs a 2x input cost and 1.5x output cost when going over >200K tokens: https://platform.claude.com/docs/en/about-claude/pricing

Opus 4.6 in Claude Code has been absolutely lousy with solving problems within its current context limit so if Sonnet 4.6 is able to do long-context problems (which would be roughly the same price of base Opus 4.6), then that may actually be a game changer.

synergy20•35m ago
so this is an economical version of opus 4.6 then? free + pro --> sonnet, max+ -> opus?
qwertox•33m ago
I'm pretty sure they have been testing it for the last couple of days as Sonnet 4.5, because I've had the oddest conversations with it lately. Odd in a positive, interesting way.

I have this in my personal preferences and now was adhering really well to them:

- prioritize objective facts and critical analysis over validation or encouragement

- you are not a friend, but a neutral information-processing machine

You can paste them into a chat and see how it changes the conversation, ChatGPT also respects it well.

doctorpangloss•32m ago
Maybe they should focus on the CLI not having a million bugs.
mfiguiere•32m ago
In Claude Code 2.1.45:

  1. Default (recommended)   Opus 4.6 · Most capable for complex work
   2. Opus (1M context)        Opus 4.6 with 1M context · Billed as extra usage · $10/$37.50 per Mtok
   3. Sonnet                   Sonnet 4.6 · Best for everyday tasks
   4. Sonnet (1M context)      Sonnet 4.6 with 1M context · Billed as extra usage · $6/$22.50 per Mtok
michaelcampbell•16m ago
Interesting. My CC (2.1.45) doesn't provide the 1M option at all. Huh.
astlouis44•27m ago
Just used Sonnet 4.6 to vibe code this top-down shooter browser game, and deployed it online quickly using Manus. Would love to hear feedback and suggestions from you all on how to improve it. Also, please post your high scores!

https://apexgame-2g44xn9v.manus.space

Flowsion•8m ago
That was fun, reminded me of some flash games I used to play. Got a bit boring after like level 6. It'd be nice to have different power-ups and upgrades. Maybe you had that at later levels, though!
excerionsforte•17m ago
I'm impressed with Claude Sonnet in general. It's been doing better than Gemini 3 at following instructions. Gemini 2.5 Pro March 2025 was the best model I ever used and I feel Claude is reaching that level even surpassing it.

I subscribed to Claude because of that. I hope 4.6 is even better.

stuckkeys•16m ago
great stuff
dr_dshiv•8m ago
I noticed a big drop in opus 4.6 quality today and then I saw this news. Anyone else?

Claude Sonnet 4.6

https://www.anthropic.com/news/claude-sonnet-4-6
322•adocomplete•1h ago•226 comments

Using go fix to modernize Go code

https://go.dev/blog/gofix
123•todsacerdoti•2h ago•16 comments

Gentoo on Codeberg

https://www.gentoo.org/news/2026/02/16/codeberg.html
89•todsacerdoti•2h ago•13 comments

GrapheneOS – Break Free from Google and Apple

https://blog.tomaszdunia.pl/grapheneos-eng/
873•to3k•9h ago•572 comments

Async/Await on the GPU

https://www.vectorware.com/blog/async-await-on-gpu/
75•Philpax•2h ago•14 comments

Can a Computer Science Student Be Taught to Design Hardware?

https://semiengineering.com/can-a-computer-science-student-be-taught-to-design-hardware/
37•stn8188•2h ago•38 comments

So you want to build a tunnel

https://practical.engineering/blog/2026/2/17/so-you-want-to-build-a-tunnel
64•crescit_eundo•2h ago•25 comments

HackMyClaw

https://hackmyclaw.com/
159•hentrep•2h ago•84 comments

Chess engines do weird stuff

https://girl.surgery/chess
75•admiringly•2h ago•34 comments

I converted 2D conventional flight tracking into 3D

https://aeris.edbn.me/?city=SFO
153•kewonit•4h ago•35 comments

Trata (YC W25) Is Hiring Founding Engineers (NYC)

1•emc329•2h ago

Show HN: I wrote a technical history book on Lisp

https://berksoft.ca/gol/
69•cdegroot•3h ago•9 comments

Climbing Mount Fuji visualized through milestone stamps

https://fuji.halfof8.com/
20•gessha•2h ago•3 comments

Launch HN: Sonarly (YC W26) – AI agent to triage and fix your production alerts

https://sonarly.com/
14•Dimittri•2h ago•0 comments

Don't pass on small block ciphers

https://00f.net/2026/02/10/small-block-ciphers/
20•jstrieb•2d ago•4 comments

Is Show HN dead? No, but it's drowning

https://www.arthurcnops.blog/death-of-show-hn/
279•acnops•9h ago•237 comments

Physicists Make Electrons Flow Like Water

https://www.quantamagazine.org/physicists-make-electrons-flow-like-water-20260211/
11•rbanffy•3d ago•0 comments

Show HN: 6cy – Experimental streaming archive format with per-block codecs

https://github.com/byte271/6cy
21•yihac1•2h ago•3 comments

Discord Rival Gets Overwhelmed by Exodus of Players Fleeing Age-Verification

https://kotaku.com/discord-alternative-teamspeak-age-verification-check-rivals-2000669693
57•thunderbong•1h ago•15 comments

Show HN: Continue – Source-controlled AI checks, enforceable in CI

https://docs.continue.dev
24•sestinj•2h ago•5 comments

Russia's economy has entered the death zone

https://www.economist.com/by-invitation/2026/02/16/russias-economy-has-entered-the-death-zone
22•thelastgallon•22m ago•2 comments

Show HN: I taught LLMs to play Magic: The Gathering against each other

https://mage-bench.com/
57•GregorStocks•3h ago•43 comments

Labyrinth Locator

https://labyrinthlocator.org/
21•emigre•3d ago•2 comments

Four Column ASCII (2017)

https://garbagecollected.org/2017/01/31/four-column-ascii/
306•tempodox•2d ago•73 comments

Students Are Being Treated Like Guinea Pigs: Inside an AI-Powered Private School

https://www.404media.co/students-are-being-treated-like-guinea-pigs-inside-an-ai-powered-private-...
52•trinsic2•2h ago•36 comments

Semantic ablation: Why AI writing is generic and boring

https://www.theregister.com/2026/02/16/semantic_ablation_ai_writing/
147•benji8000•3h ago•132 comments

Tesla 'Robotaxi' adds 5 more crashes in Austin in a month – 4x worse than humans

https://electrek.co/2026/02/17/tesla-robotaxi-adds-5-more-crashes-austin-month-4x-worse-than-humans/
12•Bender•27m ago•3 comments

Hamming Distance for Hybrid Search in SQLite

https://notnotp.com/notes/hamming-distance-for-hybrid-search-in-sqlite/
55•enz•2d ago•10 comments

Show HN: I built a simulated AI containment terminal for my sci-fi novel

https://vertex.flowlogix.ai
21•stevengreser•2h ago•11 comments

Rise of the Triforce

https://dolphin-emu.org/blog/2026/02/16/rise-of-the-triforce/
402•max-m•22h ago•62 comments