frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Ask HN: How do you shut down misbehaving AI in production?

2•nordic_lion•1h ago
If you are running AI workloads/agents or LLM-backed systems in production, how do you actually shut one down when it starts behaving badly?

By “misbehaving” I mean things like: -runaway spend -latency issues -prompt loops -tool abuse or unexpected external calls -data leakage risks -cascading failures across downstream services

In most systems I’ve seen, there is good observability. You can see logs, traces, cost dashboards. But the actual shutdown mechanism often ends up being manual: disable a feature flag, revoke an API key, roll back a deployment, rate limit something upstream.

I am trying to understand what people are doing in practice.

-What is your actual kill mechanism? -Is it bound to a model endpoint, an agent instance, a workflow, a Kubernetes workload, something else? -Is shutdown automated under certain conditions, or always human-approved? -What did you discover only after your first real incident?

Concrete examples would be extremely helpful.

Comments

zachdotai•1h ago
I found it more helpful to try and "steer" the LLM into self-correcting its action if I detect misalignment. This generally improved our task success completion rates by 20%.
jamiemallers•19m ago
We run AI agents in production and our kill chain looks roughly like this:

1. Circuit breakers per-agent with token/cost ceilings. If an agent burns through more than X tokens in Y seconds, it gets hard-stopped at the proxy layer before the request even hits the model provider. This catches runaway loops fast.

2. Tool-level allowlists with runtime revocation. Each agent has an explicit list of tools/APIs it can call. We can revoke individual tool access without killing the whole agent — useful when you discover it's hammering one specific external service.

3. Graceful degradation before kill. For non-critical paths, we drop to a cached/static fallback rather than killing outright. Full kill is reserved for safety-critical cases (data leakage risk, unauthorized external calls).

4. The actual kill mechanism is boring on purpose: a feature flag that gates the agent entrypoint, backed by a fast-propagating config system (sub-second). Kubernetes pod kills are too slow when you need to stop something mid-execution.

The thing we learned the hard way: observability without automated circuit breakers is just watching a fire. Our first incident was a prompt loop that we could see clearly in traces but took 8 minutes to manually kill because the on-call had to figure out which deployment to roll back. Now the circuit breaker fires automatically and pages the human to decide whether to re-enable.

Biggest gap I still see: there's no good standard for "agent-level observability" the way we have for microservices. Traces help but they don't capture the semantic intent of what an agent was trying to do when it went off the rails.

My AI now drives my apps and checks its own work in 50ms loops

1•sebringj•36s ago•0 comments

Show HN: AI Dev Hub. 75 free AI and dev tools

https://aidevhub.io/
1•orbydx•1m ago•0 comments

Delphi is 31 years old – innovation timeline

https://blogs.embarcadero.com/delphi-innovation-timeline-31st-anniversary-edition-published-get-y...
1•andsoitis•2m ago•0 comments

Let Me Ask AI for You

https://letmeaskai.fyi/
2•dmitrysergeyev•4m ago•0 comments

Profiling on Windows: A Short Rant

https://mropert.github.io/2026/02/13/profiling_on_windows/
1•ingve•6m ago•0 comments

Energy balance in cyclists on plant-based diets during a 30-day, 4300-km ride

https://physoc.onlinelibrary.wiley.com/doi/10.14814/phy2.70629
1•PaulHoule•8m ago•0 comments

Show HN: I speak 5 languages. Duolingo taught me none. So I built lairner

https://lairner.com
3•t17r•10m ago•1 comments

Hong Kong has land, autocracy, and expensive housing. Why doesn't it build?

https://worksinprogress.co/issue/the-dysfunctional-tiger/
1•bensouthwood•11m ago•0 comments

It must be hard to publish null results

https://osf.io/preprints/socarxiv/zr5vf_v1
2•cainxinth•13m ago•0 comments

Manipulating AI memory for profit: The rise of AI Recommendation Poisoning

https://www.microsoft.com/en-us/security/blog/2026/02/10/ai-recommendation-poisoning/
1•philk10•14m ago•1 comments

Show HN: Joria – a native Mac notes app for instant capture and semantic recall

https://joria.app
1•nanbing•14m ago•0 comments

Every blog post I have shared until 2026

https://bryanhogan.com/blog/other-cool-blog-posts-2026
1•bryanhogan•15m ago•0 comments

One Task at a Time, Even with AI

https://wakamoleguy.com/p/one-task-at-a-time-even-with-ai
2•wakamoleguy•15m ago•0 comments

Scott Adams and the Art of Dying (and Living Forever) Online

https://meghanboilard.substack.com/p/scott-adams-and-the-art-of-dying
2•bcohen123•17m ago•0 comments

Jmail hits 450M views, Vercel CEO agrees to handle server costs

https://piunikaweb.com/2026/02/11/jmail-450m-views-vercel-ceo-covers-server-costs/
1•no_creativity_•18m ago•0 comments

BinaryAudit: Can AI find backdoors in raw machine code?

https://quesma.com/benchmarks/binaryaudit/
1•stared•18m ago•1 comments

AI is making online crimes easier. It could get worse

https://www.technologyreview.com/2026/02/12/1132386/ai-already-making-online-swindles-easier/
3•Brajeshwar•18m ago•0 comments

SafeRun Guard- Runtime safety firewall for AI coding agents (bash+jq, zero deps)

https://github.com/Cocabadger/saferun-guard
1•cocabadger•19m ago•1 comments

PyTorch Now Uses Pyrefly for Type Checking

https://pytorch.org/blog/pyrefly-now-type-checks-pytorch/
2•ocamoss•19m ago•0 comments

DiffSwarm: Multi-agent code review from your terminal (BYOK, runs locally)

https://diffswarm.com/
1•swolpatrol•20m ago•1 comments

Discord Voluntarily Pushes Mandatory Age Verification Despite Recent Data Breach

https://www.eff.org/deeplinks/2026/02/discord-voluntarily-pushes-mandatory-age-verification-despi...
3•hn_acker•20m ago•0 comments

Show HN: 1MB iOS apps designed to reduce mental open loops

1•kentaroyamauchi•21m ago•2 comments

Trump Antitrust Is Dead

https://pluralistic.net/2026/02/13/khanservatives/#kid-rock-eats-shit
3•leotravis10•22m ago•0 comments

Chris Liddell appointed to Anthropic's board of directors

https://www.anthropic.com/news/chris-liddell-appointed-anthropic-board
1•ryanhn•22m ago•0 comments

Maybe the Hollywood is cooked guys are cooked too idk

https://twitter.com/RuairiRobinson/status/2021394940757209134
1•hooch•23m ago•0 comments

US billionaires race China to moon

https://www.reuters.com/business/aerospace-defense/musk-fires-up-spacex-bezos-pushes-blue-origin-...
2•PessimalDecimal•25m ago•1 comments

The Tast Supply Problem

https://charlielabs.ai/blog/the-task-supply-problem/
1•mrbbk•25m ago•0 comments

Unified API Proxy for OpenAI, Anthropic, and Compatible LLM Providers

https://github.com/mylxsw/llm-gateway
1•mylxsw•27m ago•1 comments

Show HN: Uber's new publicly available RPC Kafka repository

https://github.com/uber/uForwarder
1•zahidcakici•28m ago•0 comments

Show HN: Blip – Ephemeral chat that stores nothing, anywhere. open source

https://github.com/greatsk55/BLIP
1•hackersk•29m ago•1 comments