frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Discuss – Do AI agents deserve all the hype they are getting?

4•MicroWagie•4h ago•1 comments

Ask HN: Anyone Using a Mac Studio for Local AI/LLM?

48•UmYeahNo•1d ago•30 comments

LLMs are powerful, but enterprises are deterministic by nature

3•prateekdalal•8h ago•6 comments

Ask HN: Non AI-obsessed tech forums

29•nanocat•19h ago•26 comments

Ask HN: Ideas for small ways to make the world a better place

18•jlmcgraw•21h ago•21 comments

Ask HN: 10 months since the Llama-4 release: what happened to Meta AI?

44•Invictus0•1d ago•11 comments

Ask HN: Who wants to be hired? (February 2026)

139•whoishiring•5d ago•520 comments

Ask HN: Who is hiring? (February 2026)

313•whoishiring•5d ago•514 comments

Ask HN: Non-profit, volunteers run org needs CRM. Is Odoo Community a good sol.?

2•netfortius•16h ago•1 comments

AI Regex Scientist: A self-improving regex solver

7•PranoyP•23h ago•1 comments

Tell HN: Another round of Zendesk email spam

104•Philpax•2d ago•54 comments

Ask HN: Is Connecting via SSH Risky?

19•atrevbot•2d ago•37 comments

Ask HN: Has your whole engineering team gone big into AI coding? How's it going?

18•jchung•2d ago•13 comments

Ask HN: Why LLM providers sell access instead of consulting services?

5•pera•1d ago•13 comments

Ask HN: How does ChatGPT decide which websites to recommend?

5•nworley•1d ago•11 comments

Ask HN: What is the most complicated Algorithm you came up with yourself?

3•meffmadd•1d ago•7 comments

Ask HN: Is it just me or are most businesses insane?

8•justenough•1d ago•7 comments

Ask HN: Mem0 stores memories, but doesn't learn user patterns

9•fliellerjulian•2d ago•6 comments

Ask HN: Is there anyone here who still uses slide rules?

123•blenderob•4d ago•122 comments

Kernighan on Programming

170•chrisjj•5d ago•61 comments

Ask HN: Anyone Seeing YT ads related to chats on ChatGPT?

2•guhsnamih•1d ago•4 comments

Ask HN: Does global decoupling from the USA signal comeback of the desktop app?

5•wewewedxfgdf•1d ago•3 comments

Ask HN: Any International Job Boards for International Workers?

2•15charslong•18h ago•2 comments

We built a serverless GPU inference platform with predictable latency

5•QubridAI•2d ago•1 comments

Ask HN: Does a good "read it later" app exist?

8•buchanae•3d ago•18 comments

Ask HN: Have you been fired because of AI?

17•s-stude•4d ago•15 comments

Ask HN: Anyone have a "sovereign" solution for phone calls?

12•kldg•4d ago•1 comments

Ask HN: Cheap laptop for Linux without GUI (for writing)

15•locusofself•3d ago•16 comments

Ask HN: How Did You Validate?

4•haute_cuisine•2d ago•6 comments

Ask HN: OpenClaw users, what is your token spend?

14•8cvor6j844qw_d6•4d ago•6 comments
Open in hackernews

ChatGPT 5 is slow and no better than 4

63•iwontberude•6mo ago
Have general LLMs clearly peaked?

Comments

labrador•6mo ago
GPT-5 is better as a conversational thinking partner than GPT-4o. It's answers are more concise, focused and informative. The conversation flows. GPT-5 feels more mature than GPT-4o with less juvenile "glazing."

I can't speak to other uses such as coding, but as a sounding board GPT-5 is better than GPT-4o, which was already pretty good. GPT-5's personality has definitely shifted to a more professional tone which I like.

I do understand why people are missing the more synchophant persoanlity of GPT-4o, but I'm not one of them.

saulpw•6mo ago
That sounds 10% better, not 10x better. That's close enough to 'peaked'.
labrador•6mo ago
Agreed. Sam Altman definitely over-hyped GPT-5. It's not so much more capable that it deserves a major version number bump.
3836293648•6mo ago
Surely a major version bump says more about the internals than the capabilities
labrador•6mo ago
I see your point from a software engineering perspective but unfortuately that's not how the public sees it. The common perception is that we are making leaps towards AGI. I never thought AGI was close so I'm not disappointed, but a lot of people seem to be. On the other hand, I've seen comments like "I guess my fears of a destructive super-intelligence were over-blown."
kjkjadksj•6mo ago
People seem to make this exact comment on here at every gpt release. I wonder what gpt we ought to actually be on? 1.4.6?
labrador•6mo ago
In retrospect I would have named it as follows:

GPT-4 -> GPT-4 Home

GPT-5 -> GPT-4 Enterprise

Because my impression after using GPT-5 that it is designed to satisfy the needs of Microsoft mainly. Microsoft has no interest in making AI therapists or AI companions, probably because of the legal liability. Also, that's outside their core business.

coldtea•6mo ago
3.something.

They make "this exact comment on here at every gpt release" because every GPT release is touted as revolutionary and it's increasinly a smaller bump.

BriggyDwiggs42•6mo ago
4 was definitely way better
torginus•6mo ago
I still think it's a solid achievement, but weirdly positioned. It's their new-poverty spec model, available to everyone, and is likely not too large.

It's decently good at coding, and math, beating current SOTA of Opus 4.1 by a small margin, while being much cheaper and faster to run, hinting at likely a much smaller model.

However it's no better at trivia or writing emails or essays, which is what regular people who use ChatGPT through the website actually care about, making this launch come off as awkward.

hoppp•6mo ago
They gonna release new models like apple releases iphones, same stuff little tweaks and improvements.
al_borland•6mo ago
By definition, if something is still getting 10% better each year it hasn’t yet peaked. Not even close.
coldtea•6mo ago
Getting 10% better over this last year compared to, say, 100% and 50% and 25% better the 4-5 years before?

I'd say that points to it being very close to peaked.

Nobody said anything about a steady 10% year-over-year being the case forever...

anuramat•6mo ago
kinda how theoretical physics peaked in 1878, as predicted by Philipp von Jolly? why does everyone feel the urge to extrapolate a vibe-based metric based on three points? isn't scientific/technological progress inherently unpredictable anyway
coldtea•6mo ago
No, more like how theoretical physics peaked in 1960s.

>why does everyone feel the urge to extrapolate a vibe-based metric based on three points?

Because marketers of AI extrapolate even worse to hype it, and a counter-correction is needed...

anuramat•5mo ago
I meant that we could be in the 19th century stage, not that we necessarily are

> marketers of AI

why would I ever consider their opinion to be valuable?

BriggyDwiggs42•6mo ago
It’s just extrapolating the asymptote man.
dileeparanawake•6mo ago
Yep right now I’m not even sure I’d say it’s 10% better (at things I use it for). Feels overhyped based on the launch and what every influencer covered in their review (maybe obviously). Maybe it will get better as they sort out routing and other kinks. Right now feels like big gap between marketing and reality.
mikert89•6mo ago
Have you used the Pro version? Its incredible
ManlyBread•6mo ago
>229 EUR a month

LOL

binarymax•6mo ago
My primary use case for LLMs are running jobs at scale over an API, and not chat. Yes it's very slow, and it is annoying. Getting a response from GPT-5-mini for <Classify these 50 tokens as true or false> takes 5 seconds, compared to GPT-4o which takes about a second.
jscheel•6mo ago
Doing quite a bit of that as well, but I’ve held off moving anything to gpt-5 yet. Guessing it’s a capacity issue right now.
hoppp•6mo ago
If its 5 seconds maybe you are better off renting a GPU server and running the inference where the data is, without round trips and you can use gpt-oss
beering•6mo ago
The 5 seconds delay is probably due to reasoning. Maybe try setting it to minimal? If your use case isn’t complex maybe reasoning is overkill and gpt-4.1 would suffice.
darepublic•6mo ago
They took away o3 on plus for this :(
Buttons840•6mo ago
o3 was surprisingly good at research. I once saw it spend 6 full minutes researching something before giving an answer, and I wasn't using the "research" or "deep think" or whatever it's called, o3 just decided on its own to do that much research.
gooodvibes•6mo ago
Not having the choice to use the old models is a horrible user experience. Taking 4o away so soon was a crime.

I don’t feel like I got anything new, I feel like something got taken away.

hirvi74•6mo ago
4o and perhaps a few other of the older models are coming back. Altman already stated so.
pseudo_meta•6mo ago
API is noticeably slower for me, sometimes up to 10x slower.

Upon some digging, it seems that part of the slowdown is due to the gpt-5 models by default doing some reasoning (reasoning effort "medium"), even for the nano or mini model. Setting the reasoning effort to "minimal" improves the speed a lot.

However, to be able to set the reasoning effort you have to switch to the new Response API, which wasn't a lot of work, but more than just changing a URL.

Tiberium•6mo ago
> However, to be able to set the reasoning effort you have to switch to the new Response API, which wasn't a lot of work, but more than just changing a URL.

That's not true - you can switch reasoning effort in the Chat Completions API - https://platform.openai.com/docs/api-reference/chat/create . It's just that in Chat Completions API it's a parameter called "reasoning_effort", while in the Responses API it's a "reasoning" parameter (object) with a parameter "effort" inside.

pseudo_meta•6mo ago
Oh thx, must have missed that. Guess at least that saves me some time to switch to the newer API in the future.
hirvi74•6mo ago
I'm noticing significant differences already.

Code seems to work on the first try more often for me too.

Perhaps my favorite change so far is the difference of verbosity. Some of the responses I am receiving when asking trivial questions are now are merely a handful of sentences instead of a dissertation. However, dissertation mode comes back when appropriate, which is also nice.

Edit: Slightly tangential, but I forgot to ask, do any of you all have access to the $200/month plan? If so, how does that model compare to GPT-5?

dyauspitr•6mo ago
It’s the same model at the $200 price point.
Tiberium•6mo ago
The $200 sub (ChatGPT Pro) offers GPT-5 Pro, which is not the same model.
anuramat•6mo ago
It's still GPT-5, just with the highest reasoning effort value
Tiberium•6mo ago
It's not, just like o3-pro is not the same as o3 with high reasoning.
anuramat•5mo ago
I guess it depends on your definitions, but they did say "the same underlying model"

https://x.com/openai/status/1932530423911096508

jostylr•6mo ago
I have been using the $200 plan the past day in CodexCli and so far find that it is easy to work with (nothing crazy, just web app stuff). The context window means I no longer have to worry about running out of room and it seems to stay on track just fine so far. I have it incrementally coding pieces in a manageable chunk to understand and verify. No limits so far as promised. In contrast, using Claude Code ($100 level) I run into limits after just a couple of iterations of asking and context window gets problematic too quite quickly. It feels like 5 isn't gobbling up as much irrelevant text as Claude does.
al_borland•6mo ago
It feels slower, but if the quality is better, so one response will do instead of multiple follow up questions, it’s still faster overall. It’s also still orders of magnitude faster than doing the research manually.

I’m reminded of that Louis CK joke about people being upset about the WiFi not working on their airplane.

BriggyDwiggs42•6mo ago
There’s definitely an element of that, being too used to a very impressive technology so coming off entitled. There’s also the other side though, which is that LLM-based AI has been sort of positioned as and portrayed as a technology which will drive an utter transformation in the near future, and with gpt 5 many people are having the realization forced on them that it was all sorta nonsense, lies, motivated reasoning, exaggeration, etc.
rolodexter2023•6mo ago
"ChatGPT 5" is a marketing psyop
rolodexter2023•6mo ago
“Open”AI is growing more opaque and black box every day
Tiberium•6mo ago
GPT-5 in the API, especially at "high" reasoning is quite a bit better than o3, especially at web design "style". Also from my own experience, and from some people I've talked with it's great at agentic programming and at finding bugs (and then fixing them, of course).

ChatGPT 5 in web is a router to lots of different models [1], and the non-reasoning GPT-5 chat model ("gpt-5-chat-latest" in the API) is quite dumb - no significant difference from 4o/4.1. Even if you choose GPT-5 Thinking, there's a chance that your request will be routed to GPT-5 Mini, not to full GPT-5. The only real way to fix that in ChatGPT is to subscribe to Pro and use GPT-5 Pro, but of course that's very expensive. Otherwise people suggest saying "think hard" in the prompt, which might make the router choose the better model. Worse even, Sam Altman publicly said that in the first day of the GPT-5 release their router didn't work properly. [2]

I'd suggest trying GPT-5 in API or in apps like Cursor/Windsurf if you want to truly test it.

Oh, and from the GPT-5 guide [3] apparently OpenAI considers GPT-5 to be good even at "minimal" reasoning (it's still the full thinking model then, not the chat variant, and will respond much faster):

> The minimal setting performs especially well in coding and instruction following scenarios, adhering closely to given directions

[1] https://cdn.openai.com/pdf/8124a3ce-ab78-4f06-96eb-49ea29ffb... "Table 1: Model progressions"

[2] https://x.com/sama/status/1953893841381273969 "Yesterday, the autoswitcher broke and was out of commission for a chunk of the day, and the result was GPT-5 seemed way dumber"

[3] https://platform.openai.com/docs/guides/latest-model

dileeparanawake•6mo ago
Not sure about them peaking but I definitely feel that right now gpt-5 doesn’t feel like a game changer like the marketing says. In some things it’s worse, slower, less useful (in terms of detail not ego flattery!). My use case is mainly code and learning.

I think the make everything bigger is plateauing and not yielding the infinite returns that were first suggested and demonstrated from gpt 2 - 4.

I think now it’s harder because they’ve got to focus on value per watt. Smaller good models mean less energy, less complexity to go wrong, but harder to achieve.

The unlock could be more techniques and focussed synthetic data from old models used to train new ones but apparently gpt-5 uses synthetic data and this is one of the reasons it isn’t necessarily good in real world tasks.

For me if we go the synthetic data route it’s important to shoot for quality - good synthetic data distils useful stuff, discards the noise so useful patterns are more solid in training, but imagine it’s hard to distinguish signal from noise to produce good synthetic data.

ManlyBread•6mo ago
I have used GPT-5 a few times and I genuinely can't think of a single improvement. You could stealthy deploy it in place of GPT-4 and I wouldn't be able to tell the difference.
iwontberude•6mo ago
Indeed I would say wow, ChatGPT must be having issues today it’s so slow
pelasaco•5mo ago
After 5 versions, it still not able to generate a Markdown with ```cmd``` without breaking it