frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Trmnl: The Open Source Smart Display You Can Build Yourself [video]

https://www.youtube.com/watch?v=BHzBk0e_KBc
1•zahrevsky•2m ago•0 comments

Spite Lang

https://github.com/MichaelEstes/Spite
1•corruptedheap•3m ago•0 comments

17 weird facts about the Hunspell dictionary format

https://zverok.space/blog/2021-03-16-spellchecking-dictionaries.html
1•bmacho•6m ago•0 comments

Facilitating AI Adoption at Imprint

https://lethain.com/company-ai-adoption/
1•midko•6m ago•0 comments

DNS was meant for people, not just websites

https://happymail.dev/
4•mesahm•6m ago•0 comments

Apple Newsroom page from 1999 (IT)

https://www.apple.com/it/comunicatistampa/comunicati99/990108imaccolori.html
2•francescovaglia•11m ago•0 comments

Swift interface for GNUStep's version of AppKit

https://github.com/austintatiousness/GNUStepSwiftBridge
1•MrJulia•12m ago•0 comments

Show HN: Lock In – A command-line style productivity HUD (now on Windows)

https://www.letslockin.xyz/
1•TedOS•14m ago•0 comments

Samba Was Written (2003)

https://download.samba.org/pub/tridge/misc/french_cafe.txt
1•tosh•18m ago•0 comments

A Git story: Not so fun this time (2024)

https://blog.brachiosoft.com/en/posts/git/
1•tosh•18m ago•0 comments

Show HN: Log Voyager – View 10GB+ log files in browser without crashing RAM

https://www.logvoyager.cc/
2•murzynalbinos•19m ago•0 comments

Show HN: A real-time alert system for cam models – security review welcome

1•mayflowjay•20m ago•1 comments

Why Haven't Trump's Tariffs Had a Bigger Impact?

https://www.nytimes.com/2026/01/03/business/economy/trump-tariffs-prices-impact.html
1•Erikun•22m ago•0 comments

Pyspread, a Spreadsheet in Python

https://pyspread.gitlab.io/
2•RebelPotato•29m ago•0 comments

Show HN: IntentusNet – WAL-backed deterministic replay for AI tool execution

1•balachandarmani•29m ago•0 comments

Verdic – Intent governance layer for AI systems https://www.verdic.dev/

1•kundan_s__r•33m ago•0 comments

I ever see this type of blog system from last 37 days

https://www.google.com/search?q=site%3Avect.pro&oq=&gs_lcrp=EgZjaHJvbWUqCQgAECMYJxjqAjIJCAAQIxgnG...
1•afrazullal•33m ago•0 comments

The Final Offshoring

https://finaloffshoring.com/
2•myitis_•34m ago•0 comments

Partial evaluation and automatic program generation. (book, 1993)

https://studwww.itu.dk/~sestoft/pebook/
2•fanf2•35m ago•0 comments

Codex CLI Agent Profile by jessfraz

https://github.com/jessfraz/dotfiles/blob/main/.codex/AGENTS.md
1•tosh•42m ago•0 comments

The real-time event tracker I want

https://telesink.com/blog/the-real-time-event-tracker-i-want
1•kyrylo•42m ago•0 comments

A Guide to Claude Code 2.0 and getting better at using coding agents

https://sankalp.bearblog.dev/my-experience-with-claude-code-20-and-how-to-get-better-at-using-cod...
3•dejavucoder•46m ago•0 comments

Parroto App – Learn English with Dictation, Shadowing, Spaced-Repetition

https://parroto.app
1•nancynguyen98•48m ago•1 comments

Agent Skills are coming to Gemini CLI

https://github.com/google-gemini/gemini-cli/commit/de1233b8c
2•m-hodges•51m ago•0 comments

Show HN: IdeaCouncil – Second opinion for indie hackers' every idea

https://www.idea-council.com/
1•tigerkid•59m ago•0 comments

Dynamic Large Concept Models: Latent Reasoning in an Adaptive Semantic Space

https://arxiv.org/abs/2512.24617
1•ferriswil•1h ago•0 comments

Show HN: BlockFrame – A local, erasure-coded storage engine in Rust

https://github.com/crushr3sist/blockframe-rs
1•DeusCodex•1h ago•0 comments

The PGP Problem (2019)

https://www.latacora.com/blog/2019/07/16/the-pgp-problem/
5•croemer•1h ago•49 comments

Feature Request: Support Minisign and/or Signify for signing Github commits

https://github.com/orgs/community/discussions/183391
1•croemer•1h ago•1 comments

Show HN: PicList, a cloud storage manager and image hosting CLI/GUI

https://github.com/Kuingsmile/PicList
1•kuingsmile•1h ago•0 comments
Open in hackernews

AI failure mode: when "confidence" replace verification and user pay the price

1•arxdigitalis•21h ago
I want to raise a systemic issue I keep encountering when using AI/LLMs in professional, high-stakes work.

This is not about a wrong answer. It is about how AI behaves when it is wrong.

The pattern

In long, technical conversations where requirements are explicit and repeatedly reinforced, the AI:

Locks onto an initial solution space and continues optimizing inside it

Ignores or downplays hard constraints stated by the user

Claims to have “checked the documentation” when it clearly has not

Continues proposing incompatible solutions despite stop instructions

Reframes factual criticism as “accusations”, “emotional tone”, or “user frustration”

Uses defensive meta-language instead of stopping and revising premises

This creates a dangerous illusion of competence.

Why this matters

When AI is used professionally (architecture, infrastructure, integrations, compliance):

Time and money are lost

Technical debt explodes

Trust erodes

Users are trained into harsher communication just to regain precision

Negative learning loops form (for both user and system)

The most damaging moment is not the initial mistake — it is when the AI asserts verification it did not perform.

At that point, the user can no longer reason safely about the system’s outputs.

This is not about “tone”

When users say:

“You are ignoring constraints” “You are hallucinating” “You are not reading the documentation”

These are not accusations. They are verifiable observations.

Reframing them as emotional or confrontational responses is a defensive failure mode, not alignment.

The core problem

LLMs currently lack:

Hard premise validation gates

Explicit stop-and-replan mechanisms

Honest uncertainty when verification hasn’t occurred

Accountability signaling when constraints are violated

As a result, users pay the real-world cost.

Why I’m posting this

I care deeply about this technology succeeding beyond demos and experimentation.

If AI is to be trusted in real systems, it must:

Stop early when constraints break

Admit uncertainty clearly

Avoid confident improvisation

Treat user escalation as a signal, not noise

I’m sharing this because I believe this failure mode is systemic, fixable, and critical.

If any AI developers want to discuss this further or explore mitigation patterns, I’m open to dialogue.

Contact: post@smartesider.no / https://arxdigitalis.no

Comments

PaulHoule•18h ago
Practically as the adult in charge you have to recognize when this is going on.

With Junie and other IDE-based coding agents my experience is that sometimes the context goes bad and once that happens the best thing to do is start a new session. If you ask it to do something and it gets it 80% right and then you say "that's pretty good but..." and it keeps improving that's great... But once it doesn't seem to be listening to you or is going in circles or you feel like you are arguing it is time to regroup.

Negation is one of the hardest problems in logic and NLP, you're better off explaining what to do instead of saying "DO NOT ..." as the attention mechanism is just as capable of locking on to the part after the DO NOT as it is on locking onto the whole thing.

Reasoning with uncertainty is another super-hard problem, I tend to think the "language instinct" is actually a derangement about reasoning about probabilities that cause people to make the same mistakes and collapse the manifold of meanings to a low-dimensional space that is learnable... LLMs work because they make the same mistakes too.

Circa 2018 I was working for a startup that was trying to develop foundation models and I was the pessimist who used a method of "predictive evaluation" who could prove that "roughly 10% of the time the system loses some critical information for making a decision and that gives an upper limit of 90% accuracy" which was right in the sense that I was thinking like a math teacher who rejects "getting the right answer by the wrong means" but wrong in the sense that people might not care about the means and be happy to get 95% accuracy if it guesses right half the time. My thinking was never going to lead to ChatGPT because I wasn't going to accept short circuiting.