frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The $LANG Programming Language

226•dang•13h ago•44 comments

Ask HN: Iran's 120h internet shutdown, phones back. How to stay resilient?

83•us321•19h ago•84 comments

Ask HN: ADHD – How do you manage the constant stream of thoughts and ideas?

91•chriswright1664•13h ago•109 comments

Ask HN: Quantum Computation, Computers and Programming

27•rramadass•1d ago•23 comments

Tell HN: I Downgraded from macOS Tahoe to Sequoia

5•inatreecrown2•2h ago•6 comments

Ask HN: Vxlan over WireGuard or WireGuard over Vxlan?

41•mlhpdx•17h ago•69 comments

Ask HN: Personal website featured on HN, list of restaurants in NYC

4•laffOr•1h ago•0 comments

Ask HN: Discrepancy between Lichess and Stockfish

20•HNLurker2•17h ago•11 comments

Ask HN: What are you working on? (January 2026)

255•david927•2d ago•843 comments

Ask HN: Looking for Windows contributors for meeting-detection engine

7•Ayobamiu•20h ago•0 comments

Tell HN: DigitalOcean's managed services broke each other after update

76•neilfrndes•1d ago•46 comments

Tell HN: Intel could blow up the Console Wars if it had the guts

6•noumenon1111•17h ago•7 comments

Ask HN: How to find gaps and oppurtunities in the AI era?

5•SRMohitkr•7h ago•3 comments

Ask HN: Who remembers AWS Spot's auction era before the 2017 pricing change?

3•aleroawani•18h ago•0 comments

Ask HN: 500 citation MSc CS, stuck in a low-trust region. How to move forward?

18•throwawaysafely•20h ago•12 comments

Tell HN: The Google Tenor GIF API has been shut down

19•dfajgljsldkjag•20h ago•12 comments

Ask HN: Are you underutilizing your insurance too?

4•nemath•17h ago•4 comments

Ask HN: Learning Discoverability

2•learnwithmattc•22h ago•0 comments

Ask HN: How are you preventing LLM hallucinations in production systems?

2•kundan_s__r•13h ago•10 comments

Is "AI vibe coding" making prototyping worse inside real companies?

15•arapkuliev•21h ago•5 comments

Ask HN: What made you move back to HTML-to-PDF in production?

6•gokulsiva•22h ago•5 comments

Ask HN: Does anyone else think that humanoid robots is a bubble?

5•NewUser76312•17h ago•9 comments

Eleva.js – A 2.3KB JavaScript framework with signals and no virtual DOM

3•TarekRaafat•20h ago•0 comments

Experiment: Using NotebookLM as a cynical code reviewer (via custom prompts)

2•practicalaifg•20h ago•0 comments

Gh Account Permabanned – Help?

10•nicomeemes•21h ago•9 comments

Ask HN: Job seekers, what's working / not working?

18•Jabbs•1d ago•19 comments

Ask HN: Story about a CEO going off on a user who left feedback?

3•VladVladikoff•22h ago•2 comments

Ask HN: How do you prevent AI agents from going rogue in production?

3•techbuilder4242•22h ago•1 comments

Ask HN: When has a "dumb" solution beaten a sophisticated one for you?

12•amadeuswoo•3d ago•15 comments

Ask HN: Salesforce, SAP, or ServiceNow: Which Is Most Ripe for Disruption?

7•Saurabh_Kumar_•22h ago•2 comments
Open in hackernews

Ask HN: How are you preventing LLM hallucinations in production systems?

2•kundan_s__r•13h ago
Hi HN,

For those running LLMs in real production environments (especially agentic or tool-using systems): what’s actually worked for you to prevent confident but incorrect outputs?

Prompt engineering and basic filters help, but we’ve still seen cases where responses look fluent, structured, and reasonable — yet violate business rules, domain boundaries, or downstream assumptions.

I’m curious:

Do you rely on strict schemas or typed outputs?

Secondary validation models or rule engines?

Human-in-the-loop for certain classes of actions?

Hard constraints before execution (e.g., allow/deny lists)?

What approaches failed for you, and what held up under scale and real user behavior?

Interested in practical lessons and post-mortems rather than theory.

Comments

al_borland•13h ago
I’ve just been ignoring my boss every time he says something about how we should leverage AI. What we’re building doesn’t need it and can’t tolerate hallucinations. They just want to be able to brag up the chain that AI is being used, which is the wrong reason to use it.

If I was forced to use it, I’d probably be writing pretty extensive guardrails (outside of the AI) to make sure it isn’t going off the rails and the results make sense. I’m doing that anyway with all user input, so I guess I’d be treating all LLM generated text as user input and assuming it’s unreliable.

kundan_s__r•13h ago
That’s a very sane stance. Treating LLM output as untrusted input is probably the correct default when correctness matters.

The worst failures I’ve seen happen when teams half-trust the model — enough to automate, but still needing heavy guardrails. Putting the checks outside the model keeps the system understandable and deterministic.

Ignoring AI unless it can be safely boxed isn’t anti-AI — it’s good engineering.

stephenr•12h ago
I've found that I can use a very similar approach to the one I've used when handling the risks associated with blockchain, cryptocurrencies, "web scale" infrastructure, and of course the chupacabra.
kundan_s__r•9h ago
Fair enough. A healthy dose of skepticism has served us well for every overhyped wave so far. The difference this time seems to be that AI systems don’t just fail noisily — they fail convincingly, which changes how risk leaks into production.

Treating them with the same paranoia we applied to web scale infra and crypto is probably the right instinct. The chupacabra deserved it too.

stephenr•8h ago
> they fail convincingly

To someone that has paid zero attention and/or deliberately ignores any coverage about the numerous (and often hilarious) ways that spicy autocomplete just completely shits the bed? Sure, maybe.

kundan_s__r•6h ago
That’s fair — if you’re already skeptical and paying attention, the failures are obvious and often funny. The risk tends to show up more with non-experts or downstream systems that assume the output is trustworthy because it looks structured and confident.

Autocomplete failing loudly is annoying; autocomplete failing quietly inside automation is where things get interesting.

stephenr•5h ago
> The risk tends to show up more with non-experts

This hits a key point that isn't emphasised enough. A few interactions with technology and people have shaped my view:

I fiddled with Apple's Image Playground thing sometime last year, and it was quite rewarding to see a result from a simple description. It wasn't exactly what I'd asked for, but it was close, kind of. As someone who has basically zero artistic ability it was fun to be able to create something like that. I didn't think much about it at the time, but recently I thought about this again, and I keep that in mind when seeing people who are waxing poetic about using spicy autocomplete to "write code" or "analyse business requests" or whatever the fuck it is they're using it for. Of course it seems fucking magical and fool proof if you don't know how to do the thing you're asking it for, yourself.

I had to fly back to my childhood home at very short notice in August to see my (as it turned out, dying) father in hospital. I spoke to more doctors, nurses and specialists in two weeks, than I think I have ever spoken to about my own health in 40+ years). I was relaying the information from doctors to my brother via text message. His initial response to said information was to send me back a Chat fucking GPT summary/analysis of what I'd passed along to him... because apparently my own eyeballs seeing the physical condition of our father, and a doctor explaining the cause, prognosis and chances of recovery were not reliable enough. Better ask Dr Spicy Autocomplete for a fucking second opinion I guess.

So now my default view about people who choose to use spicy autocomplete for anything besides shits and giggles like "Write a Star Trek fan fiction where Captain Jack Sparrow is in charge of DS9", or "Draw <my wife> as a cute little bunny" is essentially "yeah of course it looks like infallible magic, if you don't know how to do it yourself".

kundan_s__r•4h ago
The real risk with LLMs isn’t when they fail loudly — it’s when they fail quietly and confidently, especially for non-experts or downstream systems that assume structured output equals correctness.

When you don’t already understand the domain, AI feels infallible. That’s exactly when unvalidated outputs become dangerous inside automation, decision pipelines, and production workflows.

This is why governance can’t be an afterthought. AI systems need deterministic validation against intent and execution boundaries before outputs are trusted or acted on — not just better prompts or post-hoc monitoring.

That gap between “sounds right” and “is allowed to run” is where tools like Verdic Guard are meant to sit.

Agent_Builder•6h ago
We ran into this while building GTWY.ai. What reduced hallucinations for us wasn’t more prompting or verification layers, but narrowing what the agent was allowed to do at each step. When inputs, tools, and outputs were explicit, the model stopped confidently inventing things. Fewer degrees of freedom beat smarter models.
kundan_s__r•6h ago
This matches our experience too. The biggest reduction in hallucinations usually comes from shrinking the action space, not improving the prompt. When inputs, tools, and outputs are explicitly constrained, the model stops “being creative” in places where creativity is actually risk.

It’s less about smarter models and more about making the system boring and deterministic at each step.