frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN

https://vibecolors.life/
1•tusharnaik•51s ago•0 comments

OpenAI is Broke and so is everyone else [video][10M]

https://www.youtube.com/watch?v=Y3N9qlPZBc0
1•Bender•1m ago•0 comments

We interfaced single-threaded C++ with multi-threaded Rust

https://antithesis.com/blog/2026/rust_cpp/
1•lukastyrychtr•2m ago•0 comments

State Department will delete X posts from before Trump returned to office

https://text.npr.org/nx-s1-5704785
2•derriz•2m ago•0 comments

AI Skills Marketplace

https://skly.ai
1•briannezhad•2m ago•1 comments

Show HN: A fast TUI for managing Azure Key Vault secrets written in Rust

https://github.com/jkoessle/akv-tui-rs
1•jkoessle•3m ago•0 comments

eInk UI Components in CSS

https://eink-components.dev/
1•edent•3m ago•0 comments

Discuss – Do AI agents deserve all the hype they are getting?

1•MicroWagie•6m ago•0 comments

ChatGPT is changing how we ask stupid questions

https://www.washingtonpost.com/technology/2026/02/06/stupid-questions-ai/
1•edward•7m ago•0 comments

Zig Package Manager Enhancements

https://ziglang.org/devlog/2026/#2026-02-06
2•jackhalford•9m ago•1 comments

Neutron Scans Reveal Hidden Water in Martian Meteorite

https://www.universetoday.com/articles/neutron-scans-reveal-hidden-water-in-famous-martian-meteorite
1•geox•9m ago•0 comments

Deepfaking Orson Welles's Mangled Masterpiece

https://www.newyorker.com/magazine/2026/02/09/deepfaking-orson-welless-mangled-masterpiece
1•fortran77•11m ago•1 comments

France's homegrown open source online office suite

https://github.com/suitenumerique
3•nar001•13m ago•1 comments

SpaceX Delays Mars Plans to Focus on Moon

https://www.wsj.com/science/space-astronomy/spacex-delays-mars-plans-to-focus-on-moon-66d5c542
1•BostonFern•13m ago•0 comments

Jeremy Wade's Mighty Rivers

https://www.youtube.com/playlist?list=PLyOro6vMGsP_xkW6FXxsaeHUkD5e-9AUa
1•saikatsg•14m ago•0 comments

Show HN: MCP App to play backgammon with your LLM

https://github.com/sam-mfb/backgammon-mcp
2•sam256•16m ago•0 comments

AI Command and Staff–Operational Evidence and Insights from Wargaming

https://www.militarystrategymagazine.com/article/ai-command-and-staff-operational-evidence-and-in...
1•tomwphillips•16m ago•0 comments

Show HN: CCBot – Control Claude Code from Telegram via tmux

https://github.com/six-ddc/ccbot
1•sixddc•17m ago•1 comments

Ask HN: Is the CoCo 3 the best 8 bit computer ever made?

2•amichail•19m ago•1 comments

Show HN: Convert your articles into videos in one click

https://vidinie.com/
3•kositheastro•22m ago•1 comments

Red Queen's Race

https://en.wikipedia.org/wiki/Red_Queen%27s_race
2•rzk•22m ago•0 comments

The Anthropic Hive Mind

https://steve-yegge.medium.com/the-anthropic-hive-mind-d01f768f3d7b
2•gozzoo•25m ago•0 comments

A Horrible Conclusion

https://addisoncrump.info/research/a-horrible-conclusion/
1•todsacerdoti•25m ago•0 comments

I spent $10k to automate my research at OpenAI with Codex

https://twitter.com/KarelDoostrlnck/status/2019477361557926281
2•tosh•26m ago•1 comments

From Zero to Hero: A Spring Boot Deep Dive

https://jcob-sikorski.github.io/me/
1•jjcob_sikorski•27m ago•0 comments

Show HN: Solving NP-Complete Structures via Information Noise Subtraction (P=NP)

https://zenodo.org/records/18395618
1•alemonti06•32m ago•1 comments

Cook New Emojis

https://emoji.supply/kitchen/
1•vasanthv•34m ago•0 comments

Show HN: LoKey Typer – A calm typing practice app with ambient soundscapes

https://mcp-tool-shop-org.github.io/LoKey-Typer/
1•mikeyfrilot•37m ago•0 comments

Long-Sought Proof Tames Some of Math's Unruliest Equations

https://www.quantamagazine.org/long-sought-proof-tames-some-of-maths-unruliest-equations-20260206/
1•asplake•38m ago•0 comments

Hacking the last Z80 computer – FOSDEM 2026 [video]

https://fosdem.org/2026/schedule/event/FEHLHY-hacking_the_last_z80_computer_ever_made/
2•michalpleban•39m ago•0 comments
Open in hackernews

LLM Economist – Mechanism Design for Simulated Agent Societies

https://github.com/sethkarten/LLM-Economist
2•milkkarten•6mo ago

Comments

milkkarten•6mo ago
We simulate large-scale agent societies where heterogeneous personas work, adapt, and vote—governed by an in-context planner optimizing social welfare.

The system models decentralized governance, dynamic tax policy, and institutional evolution—entirely via in-context reinforcement learning, no fine-tuning required.

Full paper (arXiv): https://arxiv.org/abs/2507.15815

slwvx•6mo ago
I like the idea of simulating a society! I don't pretend to understand everything that you're doing, so please correct me where I'm wrong below.

The right side of Fig 5a shows that your LLM tool has 80% tax for people making between 0 and $11.6k/year, then drops to about 30% for the next tax bracket, with other tax brackets moving around all over the place. This seems to be designed to induce people to NOT pay taxes.For all its faults, I think the US progressive system is fairly rational and does a pretty good job of inducing people to actually pay taxes [1]; specifically the (effectively) negative tax rate in the US for low-income people gets them in the habit of paying taxes. I.e. whatever underlying model of social welfare you are assuming to get the great social welfare on the right side of Fig 5a seems to not model real people. I wonder if some LLM hallucinations are going on under the hood to create the strange behavior in Fig 5a.

Some questions: You don't seem to model the US system of tax credits; is that right? Also, is there a Saez tax below $47.2k in Fig 5a? What about between $244k and $609k? I.e. is the Saez tax ever under the LLM tax?

[1] https://blogs.worldbank.org/en/governance/why-does-progressi...

milkkarten•6mo ago
These are the marginal tax rates not the effective tax rate (e.g. 80% of first $10k, 30% of $10k-20k). We do not model tax credits here. We try to keep the system as simple as possible so that we can effectively evaluate changes. As is, the Economic theory is intractable once we introduce bounded rationality from purely rational. We do think in future work we can potentially work out some smoothness in the overall tax rate but try to let the LLM planner try what it thinks is best in order to help test the in-context optimization capabilities.

Also, while there is a complicated tax code in the US, in our simulation there is no way for agents to avoid paying taxes :)

The Saez tax rates are perturbed from the LLM Economist's tax rates to find the theoretically optimal values according to the economic theory.

Thanks for the interest and I hope that this helps clarify some of the details.

slwvx•6mo ago
Thanks for the further details!

Ah, the fact that they are marginal rates makes marginally more sense, but it still seems to me that the SWF in fig 5a has very little relation to the real world.

> Also, while there is a complicated tax code in the US, in our simulation there is no way for agents to avoid paying taxes :)

Seems like an obvious thing to add. I.e. if you believe the World Bank when they say "People are more willing to pay tax when taxes are progressive" [1], then it seems worthwhile to update your model to include this.

[1] https://blogs.worldbank.org/en/governance/why-does-progressi...

MutedEstate45•6mo ago
Interesting approach, but I'm curious about the practical cost considerations. A 1,000-agent simulation could easily be hundreds of thousands of API calls. The repo recommends gpt-4o-mini over gpt-4 and supports local Llama models, but there's no guidance on the performance trade-offs.

Would love to see cost-per-experiment breakdowns and quality benchmarks across model tiers. Does a local Llama 3.1 8B produce meaningful economic simulations or do you need the reasoning power of frontier models? This could be the difference between $5 and $500 experiments.

milkkarten•6mo ago
Using smaller, cheaper agents is one of the goals of the work. There is a Pareto frontier though: by using smaller, faster, cheaper agents, the number of steps required to converge increases. We touch upon this briefly in the paper
MutedEstate45•6mo ago
Thanks. That Pareto trade-off is exactly what I'm trying to quantify not just qualify. For example, if I've got a $50 budget, what's the sweet spot?

Scenario A: 100 agents × GPT-4o-mini × 500 steps Scenario B: 500 agents × local Llama 3-8B × 1,000+ steps

A quick table like "X agents × Y model × Z steps → tokens, $, convergence score" in the README would let new users budget experiments without having to read the whole paper plus run expensive experiments just to discover basic resource planning.

milkkarten•6mo ago
We ran each method in under 24 hours on a singular H100. I understand your point and think we will include this in future iterations of our work since this is very interesting from the user perspective. Though, in the paper we focus more on algorithmic concerns.
MutedEstate45•6mo ago
I'll look out for future iterations. Thanks and good luck with the paper.