frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: SAA – A minimal shell-as-chat agent using only Bash

https://github.com/moravy-mochi/saa
1•mrvmochi•49s ago•0 comments

Mario Tchou

https://en.wikipedia.org/wiki/Mario_Tchou
1•simonebrunozzi•1m ago•0 comments

Does Anyone Even Know What's Happening in Zim?

https://mayberay.bearblog.dev/does-anyone-even-know-whats-happening-in-zim-right-now/
1•mugamuga•2m ago•0 comments

The last Morse code maritime radio station in North America [video]

https://www.youtube.com/watch?v=GzN-D0yIkGQ
1•austinallegro•4m ago•0 comments

Show HN: Hacker Newspaper – Yet another HN front end optimized for mobile

https://hackernews.paperd.ink/
1•robertlangdon•5m ago•0 comments

OpenClaw Is Changing My Life

https://reorx.com/blog/openclaw-is-changing-my-life/
1•novoreorx•13m ago•0 comments

Everything you need to know about lasers in one photo

https://commons.wikimedia.org/wiki/File:Commercial_laser_lines.svg
1•mahirsaid•15m ago•0 comments

SCOTUS to decide if 1988 video tape privacy law applies to internet uses

https://www.jurist.org/news/2026/01/us-supreme-court-to-decide-if-1988-video-tape-privacy-law-app...
1•voxadam•16m ago•0 comments

Epstein files reveal deeper ties to scientists than previously known

https://www.nature.com/articles/d41586-026-00388-0
1•XzetaU8•24m ago•0 comments

Red teamers arrested conducting a penetration test

https://www.infosecinstitute.com/podcast/red-teamers-arrested-conducting-a-penetration-test/
1•begueradj•31m ago•0 comments

Show HN: Open-source AI powered Kubernetes IDE

https://github.com/agentkube/agentkube
1•saiyampathak•34m ago•0 comments

Show HN: Lucid – Use LLM hallucination to generate verified software specs

https://github.com/gtsbahamas/hallucination-reversing-system
1•tywells•37m ago•0 comments

AI Doesn't Write Every Framework Equally Well

https://x.com/SevenviewSteve/article/2019601506429730976
1•Osiris30•40m ago•0 comments

Aisbf – an intelligent routing proxy for OpenAI compatible clients

https://pypi.org/project/aisbf/
1•nextime•41m ago•1 comments

Let's handle 1M requests per second

https://www.youtube.com/watch?v=W4EwfEU8CGA
1•4pkjai•41m ago•0 comments

OpenClaw Partners with VirusTotal for Skill Security

https://openclaw.ai/blog/virustotal-partnership
1•zhizhenchi•42m ago•0 comments

Goal: Ship 1M Lines of Code Daily

2•feastingonslop•52m ago•0 comments

Show HN: Codex-mem, 90% fewer tokens for Codex

https://github.com/StartripAI/codex-mem
1•alfredray•55m ago•0 comments

FastLangML: FastLangML:Context‑aware lang detector for short conversational text

https://github.com/pnrajan/fastlangml
1•sachuin23•58m ago•1 comments

LineageOS 23.2

https://lineageos.org/Changelog-31/
2•pentagrama•1h ago•0 comments

Crypto Deposit Frauds

2•wwdesouza•1h ago•0 comments

Substack makes money from hosting Nazi newsletters

https://www.theguardian.com/media/2026/feb/07/revealed-how-substack-makes-money-from-hosting-nazi...
4•lostlogin•1h ago•0 comments

Framing an LLM as a safety researcher changes its language, not its judgement

https://lab.fukami.eu/LLMAAJ
1•dogacel•1h ago•0 comments

Are there anyone interested about a creator economy startup

1•Nejana•1h ago•0 comments

Show HN: Skill Lab – CLI tool for testing and quality scoring agent skills

https://github.com/8ddieHu0314/Skill-Lab
1•qu4rk5314•1h ago•0 comments

2003: What is Google's Ultimate Goal? [video]

https://www.youtube.com/watch?v=xqdi1xjtys4
1•1659447091•1h ago•0 comments

Roger Ebert Reviews "The Shawshank Redemption"

https://www.rogerebert.com/reviews/great-movie-the-shawshank-redemption-1994
1•monero-xmr•1h ago•0 comments

Busy Months in KDE Linux

https://pointieststick.com/2026/02/06/busy-months-in-kde-linux/
1•todsacerdoti•1h ago•0 comments

Zram as Swap

https://wiki.archlinux.org/title/Zram#Usage_as_swap
1•seansh•1h ago•1 comments

Green’s Dictionary of Slang - Five hundred years of the vulgar tongue

https://greensdictofslang.com/
1•mxfh•1h ago•0 comments
Open in hackernews

AI failure mode: when "confidence" replace verification and user pay the price

1•arxdigitalis•1mo ago
I want to raise a systemic issue I keep encountering when using AI/LLMs in professional, high-stakes work.

This is not about a wrong answer. It is about how AI behaves when it is wrong.

The pattern

In long, technical conversations where requirements are explicit and repeatedly reinforced, the AI:

Locks onto an initial solution space and continues optimizing inside it

Ignores or downplays hard constraints stated by the user

Claims to have “checked the documentation” when it clearly has not

Continues proposing incompatible solutions despite stop instructions

Reframes factual criticism as “accusations”, “emotional tone”, or “user frustration”

Uses defensive meta-language instead of stopping and revising premises

This creates a dangerous illusion of competence.

Why this matters

When AI is used professionally (architecture, infrastructure, integrations, compliance):

Time and money are lost

Technical debt explodes

Trust erodes

Users are trained into harsher communication just to regain precision

Negative learning loops form (for both user and system)

The most damaging moment is not the initial mistake — it is when the AI asserts verification it did not perform.

At that point, the user can no longer reason safely about the system’s outputs.

This is not about “tone”

When users say:

“You are ignoring constraints” “You are hallucinating” “You are not reading the documentation”

These are not accusations. They are verifiable observations.

Reframing them as emotional or confrontational responses is a defensive failure mode, not alignment.

The core problem

LLMs currently lack:

Hard premise validation gates

Explicit stop-and-replan mechanisms

Honest uncertainty when verification hasn’t occurred

Accountability signaling when constraints are violated

As a result, users pay the real-world cost.

Why I’m posting this

I care deeply about this technology succeeding beyond demos and experimentation.

If AI is to be trusted in real systems, it must:

Stop early when constraints break

Admit uncertainty clearly

Avoid confident improvisation

Treat user escalation as a signal, not noise

I’m sharing this because I believe this failure mode is systemic, fixable, and critical.

If any AI developers want to discuss this further or explore mitigation patterns, I’m open to dialogue.

Contact: post@smartesider.no / https://arxdigitalis.no

Comments

PaulHoule•1mo ago
Practically as the adult in charge you have to recognize when this is going on.

With Junie and other IDE-based coding agents my experience is that sometimes the context goes bad and once that happens the best thing to do is start a new session. If you ask it to do something and it gets it 80% right and then you say "that's pretty good but..." and it keeps improving that's great... But once it doesn't seem to be listening to you or is going in circles or you feel like you are arguing it is time to regroup.

Negation is one of the hardest problems in logic and NLP, you're better off explaining what to do instead of saying "DO NOT ..." as the attention mechanism is just as capable of locking on to the part after the DO NOT as it is on locking onto the whole thing.

Reasoning with uncertainty is another super-hard problem, I tend to think the "language instinct" is actually a derangement about reasoning about probabilities that cause people to make the same mistakes and collapse the manifold of meanings to a low-dimensional space that is learnable... LLMs work because they make the same mistakes too.

Circa 2018 I was working for a startup that was trying to develop foundation models and I was the pessimist who used a method of "predictive evaluation" who could prove that "roughly 10% of the time the system loses some critical information for making a decision and that gives an upper limit of 90% accuracy" which was right in the sense that I was thinking like a math teacher who rejects "getting the right answer by the wrong means" but wrong in the sense that people might not care about the means and be happy to get 95% accuracy if it guesses right half the time. My thinking was never going to lead to ChatGPT because I wasn't going to accept short circuiting.