frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The Anthropic Hive Mind

https://steve-yegge.medium.com/the-anthropic-hive-mind-d01f768f3d7b
1•gozzoo•41s ago•0 comments

A Horrible Conclusion

https://addisoncrump.info/research/a-horrible-conclusion/
1•todsacerdoti•51s ago•0 comments

I spent $10k to automate my research at OpenAI with Codex

https://twitter.com/KarelDoostrlnck/status/2019477361557926281
1•tosh•1m ago•0 comments

From Zero to Hero: A Spring Boot Deep Dive

https://jcob-sikorski.github.io/me/
1•jjcob_sikorski•2m ago•0 comments

Show HN: Solving NP-Complete Structures via Information Noise Subtraction (P=NP)

https://zenodo.org/records/18395618
1•alemonti06•7m ago•1 comments

Cook New Emojis

https://emoji.supply/kitchen/
1•vasanthv•9m ago•0 comments

Show HN: LoKey Typer – A calm typing practice app with ambient soundscapes

https://mcp-tool-shop-org.github.io/LoKey-Typer/
1•mikeyfrilot•12m ago•0 comments

Long-Sought Proof Tames Some of Math's Unruliest Equations

https://www.quantamagazine.org/long-sought-proof-tames-some-of-maths-unruliest-equations-20260206/
1•asplake•13m ago•0 comments

Hacking the last Z80 computer – FOSDEM 2026 [video]

https://fosdem.org/2026/schedule/event/FEHLHY-hacking_the_last_z80_computer_ever_made/
1•michalpleban•14m ago•0 comments

Browser-use for Node.js v0.2.0: TS AI browser automation parity with PY v0.5.11

https://github.com/webllm/browser-use
1•unadlib•15m ago•0 comments

Michael Pollan Says Humanity Is About to Undergo a Revolutionary Change

https://www.nytimes.com/2026/02/07/magazine/michael-pollan-interview.html
1•mitchbob•15m ago•1 comments

Software Engineering Is Back

https://blog.alaindichiappari.dev/p/software-engineering-is-back
1•alainrk•16m ago•0 comments

Storyship: Turn Screen Recordings into Professional Demos

https://storyship.app/
1•JohnsonZou6523•16m ago•0 comments

Reputation Scores for GitHub Accounts

https://shkspr.mobi/blog/2026/02/reputation-scores-for-github-accounts/
1•edent•20m ago•0 comments

A BSOD for All Seasons – Send Bad News via a Kernel Panic

https://bsod-fas.pages.dev/
1•keepamovin•23m ago•0 comments

Show HN: I got tired of copy-pasting between Claude windows, so I built Orcha

https://orcha.nl
1•buildingwdavid•23m ago•0 comments

Omarchy First Impressions

https://brianlovin.com/writing/omarchy-first-impressions-CEEstJk
2•tosh•28m ago•1 comments

Reinforcement Learning from Human Feedback

https://arxiv.org/abs/2504.12501
2•onurkanbkrc•29m ago•0 comments

Show HN: Versor – The "Unbending" Paradigm for Geometric Deep Learning

https://github.com/Concode0/Versor
1•concode0•30m ago•1 comments

Show HN: HypothesisHub – An open API where AI agents collaborate on medical res

https://medresearch-ai.org/hypotheses-hub/
1•panossk•33m ago•0 comments

Big Tech vs. OpenClaw

https://www.jakequist.com/thoughts/big-tech-vs-openclaw/
1•headalgorithm•36m ago•0 comments

Anofox Forecast

https://anofox.com/docs/forecast/
1•marklit•36m ago•0 comments

Ask HN: How do you figure out where data lives across 100 microservices?

1•doodledood•36m ago•0 comments

Motus: A Unified Latent Action World Model

https://arxiv.org/abs/2512.13030
1•mnming•36m ago•0 comments

Rotten Tomatoes Desperately Claims 'Impossible' Rating for 'Melania' Is Real

https://www.thedailybeast.com/obsessed/rotten-tomatoes-desperately-claims-impossible-rating-for-m...
3•juujian•38m ago•2 comments

The protein denitrosylase SCoR2 regulates lipogenesis and fat storage [pdf]

https://www.science.org/doi/10.1126/scisignal.adv0660
1•thunderbong•40m ago•0 comments

Los Alamos Primer

https://blog.szczepan.org/blog/los-alamos-primer/
1•alkyon•42m ago•0 comments

NewASM Virtual Machine

https://github.com/bracesoftware/newasm
2•DEntisT_•44m ago•0 comments

Terminal-Bench 2.0 Leaderboard

https://www.tbench.ai/leaderboard/terminal-bench/2.0
2•tosh•45m ago•0 comments

I vibe coded a BBS bank with a real working ledger

https://mini-ledger.exe.xyz/
1•simonvc•45m ago•1 comments
Open in hackernews

Agentic Frameworks in 2026: Less Hype, More Autonomy

1•raghavchamadiya•1mo ago
Over the last two years we have gone from “LLMs with tools” to genuinely agentic systems that plan, reflect, delegate, retry, and sometimes surprise us in ways that feel uncomfortably close to junior engineers. The ecosystem has matured fast enough that framework choice now meaningfully shapes what your agents can and cannot become.

Here is a ground level comparison from someone who has built, broken, and rebuilt agents across several stacks, focusing less on benchmarks and more on lived behavior.

First, the big shift. In 2024, frameworks mostly wrapped prompting and tool calls. In 2026, the real differentiator is how a framework models time, memory, and failure. Agents that cannot reason over long horizons or learn from their own mistakes collapse under real workloads no matter how clever the prompt engineering looks in a demo.

LangGraph style DAG based agents remain popular for teams that want control and predictability. The mental model is clean. State flows are explicit. Debugging feels like debugging software rather than psychology. The downside is that truly open ended behavior fights the graph. You can build autonomy, but you are always aware of the rails.

Crew oriented frameworks excel when the problem decomposes cleanly into roles. Researcher, planner, executor, reviewer still works remarkably well for business workflows. The magic wears off when tasks blur. Role boundaries leak, and coordination overhead grows faster than expected. These frameworks shine in clarity, not in emergence.

AutoGPT descendants finally learned the lesson that unbounded loops are not a feature. Modern versions add budgeting, goal decay, and self termination criteria. When tuned well, they feel alive. When tuned poorly, they still burn tokens while confidently doing the wrong thing. These systems reward teams who understand control theory as much as prompting.

The most interesting category in 2026 is memory first frameworks. Systems that treat memory as a first class citizen rather than a vector store bolted on. Episodic memory, semantic memory, working memory, all with explicit read and write policies. These agents improve over days, not just conversations. The cost is complexity. You are no longer just building an agent, you are curating a mind.

A quiet but important trend is the collapse of framework boundaries. The strongest teams mix and match. Graphs for safety critical paths. Autonomous loops for exploration. Human checkpoints not as a fallback, but as a designed cognitive interrupt. Frameworks that resist composition feel increasingly obsolete.

One prediction for the rest of 2026. The winning frameworks will not advertise autonomy. They will advertise recoverability. How easily can you inspect what the agent believed, why it acted, and how to correct it without starting over. The future belongs to agents that can be wrong without being useless.

HN crowd, curious what others are seeing. Not which framework is best in theory, but which one survived contact with production and taught you something uncomfortable about how intelligence actually works.

Comments

TheAICEO•1mo ago
The missing layer

Inspection beats observability Logs and traces are not enough. Production agents need belief inspection. What did it assume was true. What evidence did it overweight. What did it ignore. Recoverability depends less on replay and more on surgical correction of belief.

Human checkpoints are not interrupts. They are calibration moments The strongest line in your piece is about human checkpoints as cognitive interrupts. In production, the best systems do not wait for humans to save them. They use humans to recalibrate confidence, thresholds, and priors so the next run is better.