frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Show HN: An "earned autonomy" architecture for AI agents using Subjective Logic

https://kenschachter.substack.com/p/earned-autonomy
1•ken_neth•1h ago
Most agent systems treat autonomy as binary: the agent either does the thing or asks permission first. In practice, this means you end up rubber-stamping a stream of approval requests until you stop paying attention. The system designed to keep you in control trains you to stop caring.

To manage operations for my independent video game studio, I built a trust system that works more like onboarding a new hire. Agents start in draft mode (every action needs approval), and earn autonomy over time based on their track record in specific task categories.

The core idea: each agent maintains a separate Beta distribution per task category (support triage, expense reports, publisher emails, etc.). A Beta distribution is basically a track record parameterized by successes and failures. But raw E[p] = α/(α+β) can't tell the difference between "9 successes, 0 failures" and "90 successes, 10 failures" since both give E[p] = 0.90. So I use Jøsang's Subjective Logic to map these to opinion tuples that explicitly separate belief from uncertainty. High uncertainty means "not enough data yet," which is different from "we know this agent is bad."

Every action passes through a gate:

  VoI = stakes × (1 - trust) × uncertainty
Low VoI = auto-execute. High VoI = draft for human review. Static trust thresholds set the maximum autonomy level an agent can reach (Auto-Execute, Soft-Execute, Draft, Restricted), and VoI acts as a secondary gate that can restrict it further based on context — an agent might qualify for auto-execute in general, but a high-stakes situation still gets flagged.

Three things that made the biggest difference:

1. Edit distance feedback. If you rewrite half an email before hitting "approve," the system notices. A 0% edit = full trust credit. A 71%+ rewrite = penalty. This single change prevented agents from reaching auto-execute on work users were quietly fixing.

2. Time-based decay. Trust scores decay daily for inactive categories (λ = 0.95). If an agent hasn't done a task in two months, it gets supervised again. This also handles model upgrades, since the track record was earned on a different model.

3. Weakest-link chains. Multi-step workflows (send welcome email → create project → schedule meeting → notify team) use a weakest-link model. If any step needs approval, the whole chain surfaces as one inbox item. Nothing runs until you approve the full picture.

The core mapping from track record to opinion looks like this:

  def beta_to_opinion(alpha, beta, base_rate=0.5):
      n = alpha + beta
      return Opinion(
          belief=(alpha - 1) / n,
          disbelief=(beta - 1) / n,
          uncertainty=2 / n,
          base_rate=base_rate,
      )
The math is all well-established (Beta distributions, Subjective Logic, Value of Information). The part that worked was combining them into something that mirrors how trust actually develops between people.

Article with full implementation details, code examples, and diagrams: https://kenschachter.substack.com/p/earned-autonomy

Touchscreen OLED MacBook Pro Coming in 2026: Dynamic Island, Redesigned Controls

https://www.macrumors.com/2026/02/24/touchscreen-macbook-pro-dynamic-island/
1•mikhael•2m ago•0 comments

Lattice-proxy – 93% token compression for LLM APIs (drop-in replacement)

https://latticeproxy.io
1•Negative_ron•4m ago•0 comments

Fed's Cook says AI triggering big changes, sees possible unemployment rise

https://www.reuters.com/business/feds-cook-says-ai-triggering-big-changes-sees-possible-short-ter...
3•geox•4m ago•0 comments

Show HN: Linex – A daily challenge: placing pieces on a board that fights back

https://www.playlinex.com/
1•Humanista75•5m ago•0 comments

Claude says its DeepSeek when asked in Chinese

https://twitter.com/stevibe/status/2026227392076018101
1•smusamashah•5m ago•0 comments

UK fines Reddit for not checking user ages aggressively enough

https://arstechnica.com/tech-policy/2026/02/uk-fines-reddit-for-not-checking-user-ages-aggressive...
1•iamnothere•6m ago•0 comments

Show HN: Tau Router – Using Number-Theory to Partition Long-Context Retrieval

https://github.com/davidames0523/tau-router
1•davidames0523•6m ago•0 comments

Show HN: Claud-ometer – See your Claude Code usage, costs, and sessions locally

https://github.com/deshraj/Claud-ometer
1•deshraj•7m ago•0 comments

Show HN: MCP app for viewing/generating 3D Gaussian Splatting in ChatGPT

https://github.com/kstonekuan/splatter-mcp-app
1•kstonekuan•7m ago•0 comments

Sandcastles Made of Bits

https://jsfour.substack.com/p/sandcastles-made-of-bits
2•TheOsiris•8m ago•0 comments

Show HN: VoooAI – natural language to multi-modal AI pipelines

https://voooai.com/
1•hkljjkl•9m ago•0 comments

Every privacy concession in history has been permanent

https://nbtv.substack.com/p/enough-is-enough
2•Cider9986•9m ago•1 comments

Show HN: Ctrl and Click on text to jump to its translation key in VSCode

https://chromewebstore.google.com/detail/i18nkeylocator/nkoandfnjiopdjmhbcnggpeomnmieadi
1•yusufA207•12m ago•0 comments

Xkcd Simulation for Real Packages

https://nesbitt.io/xkcd-2347/
1•michaelw•12m ago•0 comments

New Research Suggests Myopia Could Be Caused by How We Use Our Eyes Indoors

https://www.sunyopt.edu/new-research-suggests-myopia-could-be-caused-by-how-we-use-our-eyes-indoors/
1•gnabgib•13m ago•0 comments

AI and My Crisis of Meaning

https://brids.bearblog.dev/ai-and-my-crisis-of-meaning/
2•mindwok•16m ago•0 comments

Men develop heart disease 7 years before women

https://www.empirical.health/blog/men-vs-women-heart-disease/
1•brandonb•17m ago•0 comments

AI isn't killing SaaS – it's killing single-purpose SaaS

2•JB_5000•18m ago•0 comments

Show HN: WaveGuard – Anomaly detection using wave physics simulation (GPU, MCP)

https://github.com/gpartin/WaveGuardClient
1•waveguard•18m ago•0 comments

Show HN: Jqueue: A job queue that runs on a single JSON file

https://github.com/janbjorge/jqueue
1•jeeybee•24m ago•0 comments

AST-Guided Translation of Natural Language into First-Order Logic with LLMs

https://arxiv.org/abs/2602.13237
2•PaulHoule•24m ago•0 comments

How Long Is the Coast of Britain? (1967)

https://www.jstor.org/stable/1721427
1•Hooke•24m ago•0 comments

Hegseth threatens to blacklist Anthropic over 'woke AI' concerns

https://www.npr.org/2026/02/24/nx-s1-5725327/pentagon-anthropic-hegseth-safety
1•cdrnsf•26m ago•0 comments

Ask HN: Do you use trading AI agents?

1•alphathetaquant•26m ago•1 comments

People-Shaped Problems

https://herbertlui.net/people-shaped-problems/
1•herbertl•26m ago•0 comments

In Pursuit of High-Fidelity GPU Kernel Benchmarking

https://standardkernel.com/blog/in-pursuit-of-high-fidelity-gpu-kernel-benchmarking/
1•matt_d•27m ago•0 comments

Chile's Project Cybersyn (1971-1973)

https://en.wikipedia.org/wiki/Project_Cybersyn
1•alhazrod•27m ago•0 comments

Mac mini plant was planned under the Biden administration

https://appleinsider.com/articles/26/02/24/apples-us-mac-mini-assembly-plant-was-planned-during-t...
3•alwillis•28m ago•0 comments

Cortex 2.0 – The Future of Robotic Intelligence

https://cortex2.sereact.ai
1•pancakenetwork•28m ago•0 comments

GRC Engineering

https://mattgoodrich.com/posts/grc-engineering/
1•mooreds•28m ago•0 comments