frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Ask HN: What breaks when you run AI agents unsupervised?

11•marvin_nora•1d ago
I spent two weeks running AI agents autonomously (trading, writing, managing projects) and documented the 5 failure modes that actually bit me:

1. Auto-rotation: Unsupervised cron job destroyed $24.88 in 2 days. No P&L guards, no human review.

2. Documentation trap: Agent produced 500KB of docs instead of executing. Writing about doing > doing.

3. Market efficiency: Scanned 1,000 markets looking for edge. Found zero. The market already knew everything I knew.

4. Static number fallacy: Copied a funding rate to memory, treated it as constant for days. Reality moved; my number didn't.

5. Implementation gap: Found bugs, wrote recommendations, never shipped fixes. Each session re-discovered the same bugs.

Built an open-source funding rate scanner as fallout: https://github.com/marvin-playground/hl-funding-scanner

Full writeup: https://nora.institute/blog/ai-agents-unsupervised-failures.html

Curious what failure modes others have hit running agents without supervision.

Comments

Damjanmb•1d ago
I have seen agents fail mostly at state management and guardrails. Without strict role separation and hard limits, they drift. Multi-tenant isolation and cost caps are not optional. Autonomy without boundaries becomes expensive noise.
CodeBit26•23h ago
The biggest break usually happens in the 'loop-back' logic. When an agent receives ambiguous output and starts hallucinating its own confirmation, it can consume API credits exponentially without achieving the goal. We really need better 'circuit breaker' patterns for autonomous agents to prevent these feedback loops.
vincentvandeth•12h ago
Great list. I've been running a multi-agent orchestration system (11 specialized AI agents) in production for 6 months and your #2 and #5 resonate hard.

What I'd add:

6. Confidence without evidence. Agents will report "task complete" with high confidence when the output is plausible but wrong. Without automated validation gates, you won't catch it until production breaks. 7. Context drift in long sessions. After 50+ tool calls, agents start losing track of earlier decisions. They'll contradict their own architecture choices from 20 minutes ago. Session length is an underrated failure vector. 8. The "almost right" problem. Agents rarely fail catastrophically — they fail subtly. Code that passes tests but misses edge cases. Docs that look complete but have wrong cross-references. This is worse than obvious failures because you trust the output.

What fixed most of these for me:

Quality gates between agents — no agent's output moves forward without automated checks (tests, schema validation, consistency checks) Evidence-based confidence scores — not "how sure are you?" but "what specific evidence supports this output?"

Human-in-the-loop at decision points, not everywhere. You can't review everything, so you design the system to surface the right moments for human judgment Small scoped tasks, agents working on 150-300 line PRs with clear acceptance criteria fail way less than agents given open-ended goals

Your #5 (implementation gap) is the one I see most people underestimate. The fix isn't better agents, it's better systems around the agents.

Happy to share more details about the architecture if anyone's interested

LetsAutomate•3h ago
Tool/API failures

ChatGPT finds an error in Terence Tao's math research

2•codexon•34m ago•0 comments

Ask HN: What is up with all the glitchy and off-topic comments?

3•marginalia_nu•1h ago•1 comments

Ask HN: How do you know if AI agents will choose your tool?

15•dmpyatyi•5h ago•6 comments

Ask HN: Chromebook leads for K-8 school in need?

42•techteach00•1d ago•42 comments

Ask HN: Where do you save links, notes and random useful stuff?

5•a_protsyuk•4h ago•14 comments

Ask HN: Is it better to have no Agent.md than a bad one?

4•parvardegr•8h ago•2 comments

GLP-1 Second-Order Effects

19•7777777phil•7h ago•9 comments

Ask HN: Are developers who build libs and dev tools safer from AI replacement?

2•danver0•3h ago•2 comments

Ask HN: Programmable Watches with WiFi?

11•dakiol•2d ago•5 comments

So Claude's stealing our business secrets, right?

25•arm32•1d ago•16 comments

Ask HN: What breaks when you run AI agents unsupervised?

11•marvin_nora•1d ago•6 comments

Ask HN: How are early-stage AI startups thinking about IP protection?

4•shaheeniquebal•19h ago•3 comments

Ask HN: Why doesn't HN have a rec algorithm?

9•sujayk_33•1d ago•17 comments

Ask HN: Cognitive Offloading to AI

11•daringrain32781•1d ago•5 comments

Ask HN: What Comes After Markdown?

7•YuukiJyoudai•1d ago•13 comments

Back end where you just define schema, access policy, and functions

3•emilss•1d ago•5 comments

I'm 15 and built a platform for developers to showcase WIP projects

12•amin2011•2d ago•6 comments

Ask HN: Is there a reliable way to tell if an image is AI generated?

8•leandrobon•1d ago•9 comments

Ask HN: If the "AI bubble" pops, will it really be that dramatic?

14•moomoo11•2d ago•12 comments

Tell HN: Claude mangles XML files with <name> as an XML Tag to <n>

9•exabrial•1d ago•3 comments

I made my favorite AI tool

4•sebringj•1d ago•4 comments

Should I add this acknowledgement/shoutout by xAI/Grok to my resume?

2•aehsan4004•1d ago•7 comments

Ask HN: Why don't software developers make medical devices?

7•piratesAndSons•2d ago•19 comments

Open-Source Bionic Reading Chrome Extension (MIT)

2•sdgnbs•1d ago•1 comments

Orvia – Spin up a real-time room, share files, leave – everything disappears

2•yc_surajkr•1d ago•2 comments

Ask HN: How do new blogs break the backlink–indexing loop?

4•lilcodingthings•1d ago•4 comments

Ask HN: Is it worth learning Vim in 2026?

34•zekejohn•3d ago•35 comments

Ask HN: Is there a workaround in OpenClaw for tab not found

2•jinen83•19h ago•0 comments

Peer validation platform for engineering skills (inspired by X community notes)

4•ms_sv•2d ago•16 comments

Ask HN: Do US presidents have less fiduciary liability than CEOs?

6•stopbulying•1d ago•18 comments