KarnEvil9 is a TypeScript runtime that implements the DeepMind delegation paper from earlier this year (Tomasev et al.). The core idea is pretty simple: every action goes into a SHA-256 hash-chain journal, agents earn trust through a Bayesian scoring model, and there are actual economic stakes via escrow bonds. If an agent screws up, it loses its bond. If it keeps failing, the futility monitor kills the loop.
The fun part was testing it on Zork I. I set up three agents in a swarm: one plans moves, one executes them against a Z-machine, one independently verifies game state. The governance layer immediately blocked the agent from attacking the troll because it classified "attack" as high-risk. Took me a while to realize the fix wasn't to whitelist attack commands, it was to make the system trust-aware so an agent with a good track record can take riskier actions.
The other thing I didn't expect: when Eddie (the autonomous agent that runs 24/7 on this) hit the Anthropic API credit wall, the futility monitor halted everything, and Eddie's next plan included switching to cheaper models for routine code reviews. Nobody told it to optimize costs. That came out of the delegation framework's cost-awareness primitives.
Happy to answer questions. https://oldeucryptoboi.com