frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
411•klaussilveira•5h ago•92 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
764•xnx•10h ago•463 comments

Why I Joined OpenAI

https://www.brendangregg.com/blog/2026-02-07/why-i-joined-openai.html
29•SerCe•1h ago•24 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
136•isitcontent•5h ago•14 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
127•dmpetrov•6h ago•53 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
35•quibono•4d ago•2 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
240•vecti•7h ago•114 comments

A century of hair samples proves leaded gas ban worked

https://arstechnica.com/science/2026/02/a-century-of-hair-samples-proves-leaded-gas-ban-worked/
61•jnord•3d ago•4 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
307•aktau•12h ago•152 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
308•ostacke•11h ago•84 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
167•eljojo•8h ago•123 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
384•todsacerdoti•13h ago•217 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
313•lstoll•11h ago•230 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
47•phreda4•5h ago•8 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
103•vmatsiiako•10h ago•34 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
177•i5heu•8h ago•128 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
13•gfortaine•3h ago•0 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
230•surprisetalk•3d ago•30 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
967•cdrnsf•15h ago•414 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
139•limoce•3d ago•79 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
39•rescrv•13h ago•17 comments

Evaluating and mitigating the growing risk of LLM-discovered 0-days

https://red.anthropic.com/2026/zero-days/
34•lebovic•1d ago•11 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
7•kmm•4d ago•0 comments

Show HN: Smooth CLI – Token-efficient browser for AI agents

https://docs.smooth.sh/cli/overview
76•antves•1d ago•56 comments

I'm going to cure my girlfriend's brain tumor

https://andrewjrod.substack.com/p/im-going-to-cure-my-girlfriends-brain
34•ray__•2h ago•10 comments

The Oklahoma Architect Who Turned Kitsch into Art

https://www.bloomberg.com/news/features/2026-01-31/oklahoma-architect-bruce-goff-s-wild-home-desi...
17•MarlonPro•3d ago•3 comments

Show HN: Slack CLI for Agents

https://github.com/stablyai/agent-slack
38•nwparker•1d ago•8 comments

Claude Composer

https://www.josh.ing/blog/claude-composer
100•coloneltcb•2d ago•69 comments

How virtual textures work

https://www.shlom.dev/articles/how-virtual-textures-really-work/
25•betamark•12h ago•23 comments

The Beauty of Slag

https://mag.uchicago.edu/science-medicine/beauty-slag
31•sohkamyung•3d ago•3 comments
Open in hackernews

Show HN: Building a Deep Research Agent Using MCP-Agent

https://thealliance.ai/blog/building-a-deep-research-agent-using-mcp-agent
91•saqadri•4mo ago

Comments

asail77•4mo ago
A good model for planner seems pretty important, what models are best?
haniehz•4mo ago
based on the article, it seems like a good reasoning model like gpt5 or opus 4.1 might be good choices for the planner. I wonder if the gpt oss reasoning models would do well
koakuma-chan•4mo ago
Gemini 2.5 Pro is also a great reasoning model, I still prefer it over GPT 5
luckydata•4mo ago
Gemini is great, it's just incredibly clumsy at tool use and that's why it fails so often in practice. I'm looking forward to the next version, it will for sure address it, it's a big issue internally too (I'm a recent xoogler).
koakuma-chan•4mo ago
I'm excited for the next version!
PantaloonFlames•4mo ago
Can you elaborate on “clumsy at tool use”?
luckydata•4mo ago
have you ever witnessed how sometimes Gemini makes multiple attempts at writing a file only to give up and start chanting "I'm worthless...".

That's tool use failure :)

reachableceo•4mo ago
Yes it really is horrible at using tools. Codex is way better (even better than Claude code ). Gemini is great at doing audits and content (though I’ve switched to codex for everything all in one).
diggan•4mo ago
Personally been using GPT-OSS-120b locally with reasoning_effort set to `high` and it blows pretty much every other local model out of the water, but takes a lot of time for it to eventually do a proper content reply. But for fire-and-forget jobs like "Create a well-researched report on X from perspective Y" it works really well.
cyberninja15•4mo ago
what machine are you running GPT-OSS-120B on? I'm currently only able to get GPT-OSS-20B working on my macbook using Ollama
saqadri•4mo ago
OP here -- I think the general principle I would recommend is using a big reasoning model for the planning phase. I think Claude Code and other agents do the same. The reason this is important is because the quality of the plan really affects the final result, and error rates will compound if the plan isn't good.
ilovefood•4mo ago
Great write-up! Gives me a few ideas for a governance bot that I'm working on. Thanks for sharing :)
diggan•4mo ago
I gotta say, having white blurry blobs of something in the background floating behind white/grey text maybe wasn't the best design-choice out there.

None the less, I tried to find the actual APIs/service/software used for the "search" part, as I've found that to be the hardest to actually get right (at least for as-local-as-possible usage) for my own "Deep Research Agent".

I've experimented with Brave's search API which worked OK, but seems pricey for agent usage. Currently experimenting with using my own (local) YaCy instance right now, which actually gives me higher quality artifacts at the end, as there are no rate-limits and the model can do hundreds of search calls without me worrying about the cost. But it isn't very quick at picking up some stuff like news and more, otherwise works OK too.

What is the author doing here for the actual searching? Anyone else have any other ideas/approaches to this?

saqadri•4mo ago
Haha, I didn't have control on the blog website, just the content. The readme and code is the ultimate source of truth (and easier to read):https://github.com/lastmile-ai/mcp-agent/blob/main/src/mcp_a...

So the core idea is the Deep Orchestrator is pretty unopinionated on what to use for searching, as long as it is exposed over MCP. I tried with a basic fetch server that's one of the reference MCP servers (with a single tool called `fetch`), and also tried with Brave.

I think the folks at Jina wrote some really good stuff on the actual search part: https://jina.ai/news/a-practical-guide-to-implementing-deeps... -- and how to do page/url ranking over the course of the flow. My recommendation would be to do all that in an MCP server itself. That keeps the "deep orchestrator" architecture fairly clean, and you can plug in increasingly sophisticated search techniques over time.

jimmySixDOF•4mo ago
I'd be interested if you did any comparison testing to the langchain project which was, at least a month ago, the top open source approach

https://huggingface.co/spaces/Ayanami0730/DeepResearch-Leade...

saqadri•4mo ago
Thanks for sharing this! We've reached out to the benchmark owners are are going to get our deep research agent benchmarked soon.
Zetaphor•4mo ago
Self host an instance of SearXNG[1] either locally or on a remote server with a simple docker container and use its JSON API [2]. You have to enable the JSON API in the config manually [3].

[1] https://docs.searxng.org/admin/installation-docker.html#inst...

[2] https://docs.searxng.org/dev/search_api.html

[3] https://github.com/searxng/searxng/discussions/3542

saqadri•4mo ago
Thanks for sharing, this looks great! Do they have an MCP server? It should be easy to wrap around their JSON API but I couldn't see MCP support in the repo/docs.
Zetaphor•4mo ago
Not that I'm aware of, but it's an extremely simple API. It's should be really easy to wrap into an MCP
mbil•4mo ago
I'm using mcp-agent and have tried the orchestrator workflow pattern[0]. For deep research I'm having mixed results. As far as I can tell, it's not using prompt caching[1] with Anthropic models, nor the gpt-5 responses API[2], which is preferable to the completions API. The many MCP tools from a handful of servers eat up a lot of context. It doesn't report progress, so it'll just spin for minutes at a time without meaningful indication. Mostly it has been high cost and high latency without great grounding in source facts. I like the interface overall, but some of the patterns and examples were convoluted. I'm aware that mcp-agent is being worked on, and I look forward to improvements.

[0]: https://docs.mcp-agent.com/workflows/orchestrator

[1]: https://docs.anthropic.com/en/docs/build-with-claude/prompt-...

[2]: https://platform.openai.com/docs/guides/migrate-to-responses