frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: A longitudinal health record built from fragmented medical data

https://myaether.live
1•takmak007•54s ago•0 comments

CoreWeave's $30B Bet on GPU Market Infrastructure

https://davefriedman.substack.com/p/coreweaves-30-billion-bet-on-gpu
1•gmays•12m ago•0 comments

Creating and Hosting a Static Website on Cloudflare for Free

https://benjaminsmallwood.com/blog/creating-and-hosting-a-static-website-on-cloudflare-for-free/
1•bensmallwood•17m ago•1 comments

"The Stanford scam proves America is becoming a nation of grifters"

https://www.thetimes.com/us/news-today/article/students-stanford-grifters-ivy-league-w2g5z768z
1•cwwc•22m ago•0 comments

Elon Musk on Space GPUs, AI, Optimus, and His Manufacturing Method

https://cheekypint.substack.com/p/elon-musk-on-space-gpus-ai-optimus
2•simonebrunozzi•30m ago•0 comments

X (Twitter) is back with a new X API Pay-Per-Use model

https://developer.x.com/
2•eeko_systems•37m ago•0 comments

Zlob.h 100% POSIX and glibc compatible globbing lib that is faste and better

https://github.com/dmtrKovalenko/zlob
2•neogoose•40m ago•1 comments

Show HN: Deterministic signal triangulation using a fixed .72% variance constant

https://github.com/mabrucker85-prog/Project_Lance_Core
2•mav5431•41m ago•1 comments

Scientists Discover Levitating Time Crystals You Can Hold, Defy Newton’s 3rd Law

https://phys.org/news/2026-02-scientists-levitating-crystals.html
3•sizzle•41m ago•0 comments

When Michelangelo Met Titian

https://www.wsj.com/arts-culture/books/michelangelo-titian-review-the-renaissances-odd-couple-e34...
1•keiferski•42m ago•0 comments

Solving NYT Pips with DLX

https://github.com/DonoG/NYTPips4Processing
1•impossiblecode•43m ago•1 comments

Baldur's Gate to be turned into TV series – without the game's developers

https://www.bbc.com/news/articles/c24g457y534o
2•vunderba•43m ago•0 comments

Interview with 'Just use a VPS' bro (OpenClaw version) [video]

https://www.youtube.com/watch?v=40SnEd1RWUU
1•dangtony98•48m ago•0 comments

EchoJEPA: Latent Predictive Foundation Model for Echocardiography

https://github.com/bowang-lab/EchoJEPA
1•euvin•56m ago•0 comments

Disablling Go Telemetry

https://go.dev/doc/telemetry
1•1vuio0pswjnm7•58m ago•0 comments

Effective Nihilism

https://www.effectivenihilism.org/
1•abetusk•1h ago•1 comments

The UK government didn't want you to see this report on ecosystem collapse

https://www.theguardian.com/commentisfree/2026/jan/27/uk-government-report-ecosystem-collapse-foi...
4•pabs3•1h ago•0 comments

No 10 blocks report on impact of rainforest collapse on food prices

https://www.thetimes.com/uk/environment/article/no-10-blocks-report-on-impact-of-rainforest-colla...
2•pabs3•1h ago•0 comments

Seedance 2.0 Is Coming

https://seedance-2.app/
1•Jenny249•1h ago•0 comments

Show HN: Fitspire – a simple 5-minute workout app for busy people (iOS)

https://apps.apple.com/us/app/fitspire-5-minute-workout/id6758784938
1•devavinoth12•1h ago•0 comments

Dexterous robotic hands: 2009 – 2014 – 2025

https://old.reddit.com/r/robotics/comments/1qp7z15/dexterous_robotic_hands_2009_2014_2025/
1•gmays•1h ago•0 comments

Interop 2025: A Year of Convergence

https://webkit.org/blog/17808/interop-2025-review/
1•ksec•1h ago•1 comments

JobArena – Human Intuition vs. Artificial Intelligence

https://www.jobarena.ai/
1•84634E1A607A•1h ago•0 comments

Concept Artists Say Generative AI References Only Make Their Jobs Harder

https://thisweekinvideogames.com/feature/concept-artists-in-games-say-generative-ai-references-on...
1•KittenInABox•1h ago•0 comments

Show HN: PaySentry – Open-source control plane for AI agent payments

https://github.com/mkmkkkkk/paysentry
2•mkyang•1h ago•0 comments

Show HN: Moli P2P – An ephemeral, serverless image gallery (Rust and WebRTC)

https://moli-green.is/
2•ShinyaKoyano•1h ago•1 comments

The Crumbling Workflow Moat: Aggregation Theory's Final Chapter

https://twitter.com/nicbstme/status/2019149771706102022
1•SubiculumCode•1h ago•0 comments

Pax Historia – User and AI powered gaming platform

https://www.ycombinator.com/launches/PMu-pax-historia-user-ai-powered-gaming-platform
2•Osiris30•1h ago•0 comments

Show HN: I built a RAG engine to search Singaporean laws

https://github.com/adityaprasad-sudo/Explore-Singapore
3•ambitious_potat•1h ago•4 comments

Scams, Fraud, and Fake Apps: How to Protect Your Money in a Mobile-First Economy

https://blog.afrowallet.co/en_GB/tiers-app/scams-fraud-and-fake-apps-in-africa
1•jonatask•1h ago•0 comments
Open in hackernews

AI failure mode: when "confidence" replace verification and user pay the price

1•arxdigitalis•1mo ago
I want to raise a systemic issue I keep encountering when using AI/LLMs in professional, high-stakes work.

This is not about a wrong answer. It is about how AI behaves when it is wrong.

The pattern

In long, technical conversations where requirements are explicit and repeatedly reinforced, the AI:

Locks onto an initial solution space and continues optimizing inside it

Ignores or downplays hard constraints stated by the user

Claims to have “checked the documentation” when it clearly has not

Continues proposing incompatible solutions despite stop instructions

Reframes factual criticism as “accusations”, “emotional tone”, or “user frustration”

Uses defensive meta-language instead of stopping and revising premises

This creates a dangerous illusion of competence.

Why this matters

When AI is used professionally (architecture, infrastructure, integrations, compliance):

Time and money are lost

Technical debt explodes

Trust erodes

Users are trained into harsher communication just to regain precision

Negative learning loops form (for both user and system)

The most damaging moment is not the initial mistake — it is when the AI asserts verification it did not perform.

At that point, the user can no longer reason safely about the system’s outputs.

This is not about “tone”

When users say:

“You are ignoring constraints” “You are hallucinating” “You are not reading the documentation”

These are not accusations. They are verifiable observations.

Reframing them as emotional or confrontational responses is a defensive failure mode, not alignment.

The core problem

LLMs currently lack:

Hard premise validation gates

Explicit stop-and-replan mechanisms

Honest uncertainty when verification hasn’t occurred

Accountability signaling when constraints are violated

As a result, users pay the real-world cost.

Why I’m posting this

I care deeply about this technology succeeding beyond demos and experimentation.

If AI is to be trusted in real systems, it must:

Stop early when constraints break

Admit uncertainty clearly

Avoid confident improvisation

Treat user escalation as a signal, not noise

I’m sharing this because I believe this failure mode is systemic, fixable, and critical.

If any AI developers want to discuss this further or explore mitigation patterns, I’m open to dialogue.

Contact: post@smartesider.no / https://arxdigitalis.no

Comments

PaulHoule•1mo ago
Practically as the adult in charge you have to recognize when this is going on.

With Junie and other IDE-based coding agents my experience is that sometimes the context goes bad and once that happens the best thing to do is start a new session. If you ask it to do something and it gets it 80% right and then you say "that's pretty good but..." and it keeps improving that's great... But once it doesn't seem to be listening to you or is going in circles or you feel like you are arguing it is time to regroup.

Negation is one of the hardest problems in logic and NLP, you're better off explaining what to do instead of saying "DO NOT ..." as the attention mechanism is just as capable of locking on to the part after the DO NOT as it is on locking onto the whole thing.

Reasoning with uncertainty is another super-hard problem, I tend to think the "language instinct" is actually a derangement about reasoning about probabilities that cause people to make the same mistakes and collapse the manifold of meanings to a low-dimensional space that is learnable... LLMs work because they make the same mistakes too.

Circa 2018 I was working for a startup that was trying to develop foundation models and I was the pessimist who used a method of "predictive evaluation" who could prove that "roughly 10% of the time the system loses some critical information for making a decision and that gives an upper limit of 90% accuracy" which was right in the sense that I was thinking like a math teacher who rejects "getting the right answer by the wrong means" but wrong in the sense that people might not care about the means and be happy to get 95% accuracy if it guesses right half the time. My thinking was never going to lead to ChatGPT because I wasn't going to accept short circuiting.