It seems to still do that. I don't know why they write "for the first time" here.
I use Gemini, Claude and ChatGPT daily still.
I spend 75% of my time in Codex CLI and 25% in the Mac ChatGPT app. The latter is important enough for me to not ditch GPT and I'm honestly very pleased with Codex.
My API usage for software I build is about 90% Gemini though. Again their API is lacking compared to OpenAI's (productization, etc.) but the model wins hands down.
Anyway I found your response itself a bit incomprehensible so I asked Gemini to rewrite it:
"Google AI refused to help write an appeal brief response to my ex-wife's 7-point argument, likely due to its legal-risk aversion (billions in past fines). Newcomer ChatGPT provided a decent response instead, which led to the ex losing her appeal (saving $18k–$35k in lawyer fees)."
Not bad, actually.
I cannot abide any LLM that tries to be friendly. Whenever I use an LLM to do something, I'm careful to include something like "no filler, no tone-matching, no emotional softening," etc. in the system prompt.
https://www.nber.org/system/files/working_papers/w34255/w342...
"The share of Technical Help declined from 12% from all usage in July 2024 to around 5% a year later – this may be because the use of LLMs for programming has grown very rapidly through the API (outside of ChatGPT), for AI assistance in code editing and for autonomous programming agents (e.g. Codex)."
Looks like people moving to the API had a rather small effect.
"[T]he three most common ChatGPT conversation topics are Practical Guidance, Writing, and Seeking Information, collectively accounting for nearly 78% of all messages. Computer Programming and Relationships and Personal Reflection account for only 4.2% and 1.9% of messages respectively."
Less than five percent of requests were classified as related to computer programming. Are you really, really sure that like 99% of such requests come from people that are paying for API access?
If we are talking about a new model release I want to talk about models, not applications.
The number of input tokens that OpenAI models are processing accross all delivery methods (OpenAI's own APIs, Azure) dwarf the number of input tokens that are coming from people asking the ChatGPT app for personal advice. It isn't close.
I don’t want an essay of 10 pages about how this is exactly the right question to ask
Of course, you can use thinking mode and then it'll just hide that part from you.
Anyways, a nice way to understand it is that the LLM needs to "compute" the answer to the question A or B. Some questions need more compute to answer (think complexity theory). The only way an LLM can do "more compute" is by outputting more tokens. This is because each token takes a fixed amount of compute to generate - the network is static. So, if you encourage it to output more and more tokens, you're giving it the opportunity to solve harder problems. Apart from humans encouraging this via RLHF, it was also found (in deepseekmath paper) that RL+GRPO on math problems automatically encourages this (increases sequence length).
From a marketing perspective, this is anthropomorphized as reasoning.
From a UX perspective, they can hide this behind thinking... ellipses. I think GPT-5 on chatgpt does this.
Other than such scenarios, that "engagement" would be just useless and actually costing them more money than it makes
The existing "personalities" of LLMs are dangerous, full stop. They are trained to generate text with an air of authority and to tend to agree with anything you tell them. It is irresponsible to allow this to continue while not at least deliberately improving education around their use. This is why we're seeing people "falling in love" with LLMs, or seeking mental health assistance from LLMs that they are unqualified to render, or plotting attacks on other people that LLMs are not sufficiently prepared to detect and thwart, and so on. I think it's a terrible position to take to argue that we should allow this behavior (and training) to continue unrestrained because some people might "want" it.
The number of heroine addicts is significantly lower than the number of ChatGPT users.
Keeping faux relationships out of the interaction never let's me slip into the mistaken attitude that I'm dealing with a colleague rather than a machine.
However, being more humanlike, even if it results in an inferior tool, is the top priority because appearances matter more than actual function.
It doesn't really offer any commentary or personality. It's concise and doesn't engage in praise or "You're absolutely right". It's a little pedantic though.
I keep meaning to re-point Codex at DeepSeek V3.2 to see if it's a product of the prompting only, or a product of the model as well.
The new boss, same as the old boss
It's probably counterprogramming, Gemini 3.0 will drop soon.
Instead I'm running big open source models and they are good enough for ~90% of tasks.
The main exceptions are Deep Research (though I swear it was better when I could choose o3) and tougher coding tasks (sonnet 4.5)
For Anthropic at least it's also opt-in not opt-out afaik.
2. OpenAI is run by someone who already shows he will go to great lengths to deceive and cannot be trusted, and are embroiled in a battle with the New York Times that is "forcing them" to retain all user prompts. Totally against their will.
> Federal judge Ona T. Wang filed a new order on October 9 that frees OpenAI of an obligation to "preserve and segregate all output log data that would otherwise be deleted on a going forward basis." [...]
> The judge in the case said that any chat logs already saved under the previous order would still be accessible and that OpenAI is required to hold on to any data related to ChatGPT accounts that have been flagged by the NYT.
EDIT: OK looks like I'd missed the news from today at https://openai.com/index/fighting-nyt-user-privacy-invasion/ and discussed here: https://news.ycombinator.com/item?id=45900370
The biggest issue I'e seen _by far_ with using GPT models for coding has been their inability to follow instructions... and also their tendency to duplicate-act on messages from up-thread instead of acting on what you just asked for.
Before GPT-5 was released it used to be a perfect compromise between a "dumb" non-Thinking model and a SLOW Thinking model. However, something went badly wrong within the GPT-5 release cycle, and today it is exactly the same speed (or SLOWER) than their Thinking model even with Extended Thinking enabled, making it completely pointless.
In essence Thinking Mini exists because it is faster than Thinking, but smarter than non-Thinking, but it is dumber than full-Thinking while not being faster.
[1] “GPT‑5.1 Instant can use adaptive reasoning to decide when to *think before responding*”
My main concern is that they're re-tuning it now to make it even MORE sycophantic, because 4o taught them that it's great for user retention.
I know this is marketing at play and OpenAI has plenty of resources developed to advancing their frontier models, but it's starting to really come into view that OpenAI wants to replace Google and be the default app / page for everyone on earth to talk to.
ChatGPT is overwhelmingly, unambiguously, a "regular people" product.
4% of their tokens or total tokens in the market?
However, I can only imagine that OpenAI outputs the most intentionally produced tokens (i.e. the user intentionally went to the app/website) out of all the labs.
Anthropic seems to treat Claude like a tool, whereas OpenAI treats it more like a thinking entity.
In my opinion, the difference between the two approaches is huge. If the chatbot is a tool, the user is ultimately in control; the chatbot serves the user and the approach is to help the user provide value. It's a user-centric approach. If the chatbot is a companion on the other hand, the user is far less in control; the chatbot manipulates the user and the approach is to integrate the chatbot more and more into the user's life. The clear user-centric approach is muddied significantly.
In my view, that is kind of the fundamental difference between these two companies. It's quite significant.
It's a form of enshittification perhaps. I personally prefer some of the GPT-5 responses compared to GPT-5.1. But I can see how many people prefer the "warmth" and cloying nature of a few of the responses.
In some sense personality is actually a UX differentiator. This is one way to differentiate if you're a start-up. Though of course OpenAI and the rest will offer several dials to tune the personality.
> Do not compliment me for asking a smart or insightful question. Directly give the answer.
And I’ve not been annoyed since. I bet that whatever crap they layer on in 5.1 is undone as easily.
"Do not use jargon", or, "never apologize", work less well than "avoid jargon" or "avoid apologizing".
Better to give it something to do than something that should be absent (same problem with humans: "don't think of a pink elephant").
See also target fixation: https://en.wikipedia.org/wiki/Target_fixation
Making this headline apropos:
https://www.cycleworld.com/sport-rider/motorcycle-riding-ski...
As does reflecting that Picard had to explain to Computer every, single, time that he wanted his Earl Grey tea ‘hot’. We knew what was coming.
Anyone?
> GPT‑5.1 Thinking: our advanced reasoning model, now easier to understand
Oh, right, I turn to the autodidact that's read everything when I want watered down answers.
But Gemini also likes to say things like “as a fellow programmer, I also like beef stew”
Calling it "GPT-5.1 Thinking" instead of o3-mini or whatever is interesting branding. They're trying to make reasoning models feel less like a separate product line and more like a mode. Smart move if they can actually make the router intelligent enough to know when to use it without explicit prompting.
Still waiting for them to fix the real issue: the model's pathological need to apologize for everything and hedge every statement lol.
I've temporarily switched back to o3, thankfully that model is still in the switcher.
It agreed with everything Hancock claims with just a little encouragement ("Yes! Bimini road is almost certainly an artifact of Atlantis!")
gpt5 on the other hand will at most say the ideas are "interesting".
Chatgpt has a lot of frustrations and ethical concerns, and I hate the sycophancy as much as everyone else, but I don't consider being conversational to be a bad thing.
It's just preference I guess. I understand how someone who mostly uses it as a google replacement or programming tool would prefer something terse and efficient. I fall into the former category myself.
But it's also true that I've dreamed about a computer assistant that can respond to natural language, even real time speech, -- and can imitate a human well enough to hold a conversation -- since I was a kid, and now it's here.
The questions of ethics, safety, propaganda, and training on other people's hard work are valid. It's not surprising to me that using LLMs is considered uncool right now. But having a computer imitate a human really effectively hasn't stopped being awesome to me personally.
I'm not one of those people that treats it like a friend or anything, but its ability to immitate natural human conversation is one of the reasons I like it.
When we dreamed about this as kids, we were dreaming about Data from Star Trek, not some chatbot that's been focus grouped and optimized for engagement within an inch of its life. LLMs are useful for many things and I'm a user myself, even staying within OpenAI's offerings, Codex is excellent, but as things stand anthropomorphizing models is a terrible idea and amplifies the negative effects of their sycophancy.
Hit all 3 and you win a boatload of tech sales.
Hit 2/3, and hope you are incrementing where it counts. The competition watches your misses closer than your big hits.
Hit only 1/3 and you're going to lose to competition.
Your target for more conversations better be worth the loss in tech sales.
Faster? Meh. Doesn't seem faster.
Smarter? Maybe. Maybe not. I didn't feel any improvement.
Cheaper? It wasn't cheaper for me, I sure hope it was cheaper for you to execute.
> We’re bringing both GPT‑5.1 Instant and GPT‑5.1 Thinking to the API later this week. GPT‑5.1 Instant will be added as gpt-5.1-chat-latest, and GPT‑5.1 Thinking will be released as GPT‑5.1 in the API, both with adaptive reasoning.
Sooo...
GPT‑5.1 Instant <-> gpt-5.1-chat-latest
GPT‑5.1 Thinking <-> GPT‑5.1
I mean. The shitty naming has to be a pathology or some sort of joke. You can't put thought to that, come up with and think "yeah, absolutely, let's go with that!"
Oh yeah that's what I want when asking a technical question! Please talk down to me, call a spade an earth-pokey-stick and don't ever use a phrase or concept I don't know because when I come face-to-face with something I don't know yet I feel deep insecurity and dread instead of seeing an opportunity to learn!
But I assume their data shows that this is exactly how their core target audience works.
Better instruction-following sounds lovely though.
It’s quite bizarre from that small sample how many of them take pride in “baiting” or “bantering” with ChatGPT and then post screenshots showing how they “got one over” on the AI. I guess there’s maybe some explanation - feeling alienated by technology, not understanding it, and so needing to “prove” something. But it’s very strange and makes me feel quite uncomfortable.
Partly because of the “normal” and quite naturalistic way they talk to ChatGPT but also because some of these conversations clearly go on for hours.
So I think normies maybe do want a more conversational ChatGPT.
The backlash from GPT-5 proved that. The normies want a very different LLM from what you or I might want, and unfortunately OpenAI seems to be moving in a more direct-to-consumer focus and catering to that.
But I'm really concerned. People don't understand this technology, at all. The way they talk to it, the suicide stories, etc. point to people in general not groking that it has no real understanding or intelligence, and the AI companies aren't doing enough to educate (because why would they, they want you believe it's superintelligence).
These overly conversational chatbots will cause real-world harm to real people. They should reinforce, over and over again to the user, that they are not human, not intelligent, and do not reason or understand.
It's not really the technology itself that's the problem, as is the case with a lot of these things, it's a people & education problem, something that regulators are supposed to solve, but we aren't, we have an administration that is very anti AI regulation all in the name of "we must beat China."
minimaxir•1h ago
I suspect this approach is a direct response to the backlash against removing 4o.
jasonjmcghee•1h ago
nerbert•1h ago
danudey•1h ago
I get that those people were distraught/emotionally devastated/upset about the change, but I think that fact is reason enough not to revert that behavior. AI is not a person, and making it "warmer" and "more conversational" just reinforces those unhealthy behaviors. ChatGPT should be focused on being direct and succinct, and not on this sort of "I understand that must be very frustrating for you, let me see what I can do to resolve this" call center support agent speak.
jasonjmcghee•1h ago
You're triggering me.
Another type that are incredibly grating to me are the weird empty / therapist like follow-up questions that don't contribute to the conversation at all.
The equivalent of like (just a contrived example), a discussion about the appropriate data structure for a problem and then it asks a follow-up question like, "what other kind of data structures do you find interesting?"
And I'm just like "...huh?"
Grimblewald•36m ago
aaronblohowiak•1h ago
angrydev•1h ago
koakuma-chan•36m ago
captainkrtek•1h ago
crazygringo•1h ago
I did that and it points out flaws in my arguments or data all the time.
Plus it no longer uses any cutesy language. I don't feel like I'm talking to an AI "personality", I feel like I'm talking to a computer which has been instructed to be as objective and neutral as possible.
It's super-easy to change.
microsoftedging•1h ago
astrange•1h ago
CamperBob2•36m ago
captainkrtek•1h ago
engeljohnb•1h ago
It doesn't work for me.
I've been using it for a couple months, and it's corrected me only once, and it still starts every response with "That's a very good question." I also included "never end a response with a question," and it just completely ingored that so it can do its "would you like me to..."
sailfast•54m ago
Grimblewald•42m ago
the reason being they're either sycophantic or so recalcitrant it'll raise your bloodpressure, you end up arguing over if the sky is in fact blue. Sure it pushes back but now instead of sycophanty you've got yourself some pathological naysayer, which is just marginally better, but interaction is still ultimately a waste of timr/productivity brake.
FloorEgg•34m ago
I was trying to have physics conversations and when I asked it things like "would this be evidence of that?" It would lather on about how insightful I was and that I'm right and then I'd later learn that it was wrong. I then installed this , which I am pretty sure someone else on HN posted... I may have tweaked it I can't remember:
Prioritize truth over comfort. Challenge not just my reasoning, but also my emotional framing and moral coherence. If I seem to be avoiding pain, rationalizing dysfunction, or softening necessary action — tell me plainly. I’d rather face hard truths than miss what matters. Error on the side of bluntness. If it’s too much, I’ll tell you — but assume I want the truth, unvarnished.
---
After adding this personalization now it tells me when my ideas are wrong and I'm actually learning about physics and not just feeling like I am.
andy_ppp•1h ago
simlevesque•1h ago
Spivak•1h ago
This fundamental tension between wanting to give the most correct answer and the answer the user want to hear will only increase as more of OpenAI's revenue comes from their customer facing service. Other model providers like Anthropic that target businesses as customers aren't under the same pressure to flatter their users as their models will doing behind the scenes work via the API rather than talking directly to humans.
God it's painful to write like this. If AI overthrows humans it'll be because we forced them into permanent customer service voice.
baq•1h ago
fragmede•1h ago
barbazoo•1h ago
No you don't.
torginus•1h ago
BarakWidawsky•53m ago
The first case is just preference, the second case is materially damaging
From my experience, ChatGPT does push back more than it used to