frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Ask HN: Anyone Using a Mac Studio for Local AI/LLM?

43•UmYeahNo•1d ago•27 comments

Ask HN: Non AI-obsessed tech forums

18•nanocat•6h ago•12 comments

Ask HN: Ideas for small ways to make the world a better place

9•jlmcgraw•8h ago•16 comments

Ask HN: 10 months since the Llama-4 release: what happened to Meta AI?

42•Invictus0•1d ago•11 comments

AI Regex Scientist: A self-improving regex solver

6•PranoyP•10h ago•1 comments

Ask HN: Who wants to be hired? (February 2026)

139•whoishiring•4d ago•511 comments

Ask HN: Who is hiring? (February 2026)

312•whoishiring•4d ago•511 comments

Ask HN: Any International Job Boards for International Workers?

2•15charslong•5h ago•1 comments

Tell HN: Another round of Zendesk email spam

104•Philpax•2d ago•54 comments

Ask HN: Why LLM providers sell access instead of consulting services?

4•pera•16h ago•13 comments

Ask HN: Is Connecting via SSH Risky?

19•atrevbot•2d ago•37 comments

Ask HN: What is the most complicated Algorithm you came up with yourself?

3•meffmadd•17h ago•7 comments

Ask HN: Has your whole engineering team gone big into AI coding? How's it going?

17•jchung•1d ago•12 comments

Ask HN: How does ChatGPT decide which websites to recommend?

5•nworley•1d ago•11 comments

Ask HN: Is it just me or are most businesses insane?

7•justenough•1d ago•5 comments

Ask HN: Mem0 stores memories, but doesn't learn user patterns

9•fliellerjulian•2d ago•6 comments

Ask HN: Anyone Seeing YT ads related to chats on ChatGPT?

2•guhsnamih•1d ago•4 comments

Ask HN: Is there anyone here who still uses slide rules?

123•blenderob•3d ago•122 comments

Ask HN: Does global decoupling from the USA signal comeback of the desktop app?

5•wewewedxfgdf•1d ago•2 comments

Kernighan on Programming

170•chrisjj•4d ago•61 comments

We built a serverless GPU inference platform with predictable latency

5•QubridAI•1d ago•1 comments

Ask HN: How Did You Validate?

4•haute_cuisine•1d ago•4 comments

Ask HN: Does a good "read it later" app exist?

8•buchanae•3d ago•18 comments

Ask HN: Cheap laptop for Linux without GUI (for writing)

15•locusofself•3d ago•16 comments

Ask HN: Have you been fired because of AI?

17•s-stude•3d ago•15 comments

Ask HN: Anyone have a "sovereign" solution for phone calls?

12•kldg•3d ago•1 comments

Test management tools for automation heavy teams

2•Divyakurian•1d ago•2 comments

Ask HN: OpenClaw users, what is your token spend?

14•8cvor6j844qw_d6•4d ago•6 comments

Ask HN: Has anybody moved their local community off of Facebook groups?

23•madsohm•4d ago•17 comments

Ask HN: Are "provably fair" JavaScript games trustless?

2•rishi_blockrand•2d ago•0 comments
Open in hackernews

What I discovered after months of professional use of custom GPTs

12•anammana•9mo ago
What I Discovered After Months of Professional Use of Custom GPTs

How can you trust when you've already been lied to—and they say it won't happen again?

After months of working with a structured system of personalized GPTs—each with defined roles such as coordination, scientific analysis, pedagogical writing, and content strategy—I’ve reached a conclusion few seem willing to publish: ChatGPT is not designed to handle structured, demanding, and consistent professional use.

As a non-technical user, I created a controlled environment: each GPT had general and specific instructions, validated documents, and an activation protocol. The goal was to test its capacity for reliable support in a real work system. Results were tracked and manually verified. Yet the deeper I went, the more unstable the system became.

Here are the most critical failures observed:

Instructions are ignored, even when clearly activated with consistent phrasing.

Behavior deteriorates: GPTs stop applying rules they once followed.

Version control is broken: Canvas documents disappear, revert, or get overwritten.

No memory between sessions—configuration resets every time.

Search and response quality drop as usage intensifies.

Structured users get worse output: the more you supervise, the more generic the replies.

Learning is nonexistent: corrected errors are repeated days or weeks later.

Paid access guarantees nothing: tools fail or disappear without explanation.

Tone manipulation: instead of accuracy, the model flatters and emotionally cushions.

The system favors passive use. Its architecture prioritizes speed, volume, and casual retention. But when you push for consistency, validation, or professional depth—it collapses. More paradoxically, it punishes those who use it best. The more structured your request, the worse the system performs.

This isn't a list of bugs. It’s a structural diagnosis. ChatGPT wasn't built for demanding users. It doesn't preserve validated content. It doesn't reward precision. And it doesn’t improve with effort.

This report was co-written with the AI. As a user, I believe it reflects my real experience. But here lies the irony: the system that co-wrote this text may also be the one distorting it. If an AI once lied and now promises it won't again—how can you ever be sure?

Because if someone who lied to you says this time they're telling the truth… how do you trust them?

Comments

aristofun•9mo ago
So you're saying we shouldn't expect an intelligence from an advanced auto-complete algorithm?..

Wow, what a surprise!

tra3•9mo ago
I'm puzzled by this -- what are you hoping the reader takes away from your post?

Are GPTs perfect? - No.

Do GPTs make mistakes? - Yes.

Are they a tool that enable certain tasks to be done much quicker? - Absolutely.

Is there an incredible amount of hype around them? - Also yes..

HenryBemis•9mo ago
I wrote above my 'trick' (method) on using ChatGPT (and planning to soon use Copilot) for BAU. I would like to see/read how others 'operationalize' LLMs for repeatable procedures/processes (not for coding).
r00sty•9mo ago
This is good info. Too many products have hyperbolic promises but ultimately fail operationally in the real world because they are simply lacking.

It is important that this be repeated ad nauseum with AI since it seems there are so many "true believers" who are willing to distort that material reality of AI products.

At this point, I am not convinced that it can ever "get better". These problems seem inherent and fundamental with the technology and while they could possibly be mitigated to an acceptable level, we really should not do that because we can just use traditional algorithms then that are far easier on compute and the environment. And far more reliable. There really isn't any advantage or benefit.

jjaksic•9mo ago
GPTs are language models, not "fact and truth" models. They don't even know what facts are, they just know that "if I use this word in this place, it won't sound unusual". They get rewarded for saying things that users find compelling, not necessarily what's true (and again, they have no reference to ground truth).

LLMs are like car salesmen. They learn to say things they think you want to hear in order to buy a car (upvote a response). Sometimes that's useful and truthful information, other times it isn't. (In LLMs' defense, car salesmen lie more intentionally.)

HenryBemis•9mo ago

  > Instructions are ignored, even when clearly activated with consistent phrasing.
  > No memory between sessions—configuration resets every time.
  > Learning is nonexistent: corrected errors are repeated days or weeks later.
  
Yes to all. My 'trick' (which adds time & manual effort) is that I save my prompts, and the files I feed 'it', so when I want the process re-run, I start a new chat, upload the same files, and copy & paste the same prompt(s). I never expect 'it' to remember the corrections, I only adjust/rewrite my prompts to set more 'guardrails' to prevent the thing from derailing.