frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

minikeyvalue

https://github.com/commaai/minikeyvalue/tree/prod
1•tosh•33s ago•0 comments

Neomacs: GPU-accelerated Emacs with inline video, WebKit, and terminal via wgpu

https://github.com/eval-exec/neomacs
1•evalexec•5m ago•0 comments

Show HN: Moli P2P – An ephemeral, serverless image gallery (Rust and WebRTC)

https://moli-green.is/
1•ShinyaKoyano•9m ago•0 comments

How I grow my X presence?

https://www.reddit.com/r/GrowthHacking/s/UEc8pAl61b
1•m00dy•10m ago•0 comments

What's the cost of the most expensive Super Bowl ad slot?

https://ballparkguess.com/?id=5b98b1d3-5887-47b9-8a92-43be2ced674b
1•bkls•11m ago•0 comments

What if you just did a startup instead?

https://alexaraki.substack.com/p/what-if-you-just-did-a-startup
2•okaywriting•18m ago•0 comments

Hacking up your own shell completion (2020)

https://www.feltrac.co/environment/2020/01/18/build-your-own-shell-completion.html
1•todsacerdoti•21m ago•0 comments

Show HN: Gorse 0.5 – Open-source recommender system with visual workflow editor

https://github.com/gorse-io/gorse
1•zhenghaoz•21m ago•0 comments

GLM-OCR: Accurate × Fast × Comprehensive

https://github.com/zai-org/GLM-OCR
1•ms7892•22m ago•0 comments

Local Agent Bench: Test 11 small LLMs on tool-calling judgment, on CPU, no GPU

https://github.com/MikeVeerman/tool-calling-benchmark
1•MikeVeerman•23m ago•0 comments

Show HN: AboutMyProject – A public log for developer proof-of-work

https://aboutmyproject.com/
1•Raiplus•23m ago•0 comments

Expertise, AI and Work of Future [video]

https://www.youtube.com/watch?v=wsxWl9iT1XU
1•indiantinker•24m ago•0 comments

So Long to Cheap Books You Could Fit in Your Pocket

https://www.nytimes.com/2026/02/06/books/mass-market-paperback-books.html
3•pseudolus•24m ago•1 comments

PID Controller

https://en.wikipedia.org/wiki/Proportional%E2%80%93integral%E2%80%93derivative_controller
1•tosh•28m ago•0 comments

SpaceX Rocket Generates 100GW of Power, or 20% of US Electricity

https://twitter.com/AlecStapp/status/2019932764515234159
2•bkls•29m ago•0 comments

Kubernetes MCP Server

https://github.com/yindia/rootcause
1•yindia•30m ago•0 comments

I Built a Movie Recommendation Agent to Solve Movie Nights with My Wife

https://rokn.io/posts/building-movie-recommendation-agent
4•roknovosel•30m ago•0 comments

What were the first animals? The fierce sponge–jelly battle that just won't end

https://www.nature.com/articles/d41586-026-00238-z
2•beardyw•38m ago•0 comments

Sidestepping Evaluation Awareness and Anticipating Misalignment

https://alignment.openai.com/prod-evals/
1•taubek•38m ago•0 comments

OldMapsOnline

https://www.oldmapsonline.org/en
1•surprisetalk•41m ago•0 comments

What It's Like to Be a Worm

https://www.asimov.press/p/sentience
2•surprisetalk•41m ago•0 comments

Don't go to physics grad school and other cautionary tales

https://scottlocklin.wordpress.com/2025/12/19/dont-go-to-physics-grad-school-and-other-cautionary...
2•surprisetalk•41m ago•0 comments

Lawyer sets new standard for abuse of AI; judge tosses case

https://arstechnica.com/tech-policy/2026/02/randomly-quoting-ray-bradbury-did-not-save-lawyer-fro...
5•pseudolus•41m ago•0 comments

AI anxiety batters software execs, costing them combined $62B: report

https://nypost.com/2026/02/04/business/ai-anxiety-batters-software-execs-costing-them-62b-report/
1•1vuio0pswjnm7•41m ago•0 comments

Bogus Pipeline

https://en.wikipedia.org/wiki/Bogus_pipeline
1•doener•43m ago•0 comments

Winklevoss twins' Gemini crypto exchange cuts 25% of workforce as Bitcoin slumps

https://nypost.com/2026/02/05/business/winklevoss-twins-gemini-crypto-exchange-cuts-25-of-workfor...
2•1vuio0pswjnm7•43m ago•0 comments

How AI Is Reshaping Human Reasoning and the Rise of Cognitive Surrender

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646
3•obscurette•43m ago•0 comments

Cycling in France

https://www.sheldonbrown.com/org/france-sheldon.html
2•jackhalford•45m ago•0 comments

Ask HN: What breaks in cross-border healthcare coordination?

1•abhay1633•45m ago•0 comments

Show HN: Simple – a bytecode VM and language stack I built with AI

https://github.com/JJLDonley/Simple
2•tangjiehao•48m ago•0 comments
Open in hackernews

SWE-Grep and SWE-Grep-Mini: RL for Fast Multi-Turn Context Retrieval

https://cognition.ai/blog/swe-grep
97•meetpateltech•3mo ago

Comments

marstall•3mo ago
SWE-1 has been being booped up by WindSurf to me lately and I've been impressed - often (enough?) getting me the same answers as GPT5 etc., but almost instantly. Gotta say speed is nice.
swyx•3mo ago
nice, what does booped up mean? is this gen z lingo?
marstall•3mo ago
ha more like how i talk to my two year old. WindSurf's Cascade sidebar tool (which i use in RubyMine) has a stable of LLMs and it somewhat randomly switches the active one out from time to time. So I get a taste of what different ones are like, it's kind of cool.
tifa2up•3mo ago
Searched for 'hi' and it took 166s to return a response using this model: https://pasteboard.co/oB4VqVC5FGkl.png

Claude Code took 0.1s, Cursor CLI 19s

mgambati•3mo ago
If you ask a real question, then you might get real results.
silasalberti•3mo ago
hey I'm from the SWE-grep team - feel free to ask me any questions :)
daralthus•3mo ago
this would be useful outside of coding. could you release a benchmark so we can have more models tuned for this?
kwillets•3mo ago
Are you actually using grep here? How much data are you searching?
swyx•3mo ago
no - grep is just the closest analogy/use case that we have for it. if we end up releasing the CLI it should be as handy and nobrainer as using ripgrep

idk what you expect from a question about "how much data". its tool based search. its a lot.

kwillets•3mo ago
I'm just learning about agentic search so I'm a bit adrift.

One of my side projects is a full text index for pattern search, and I'm trying to understand how it might fit with that. You mention tool call overhead, but is that a significant part of the latency in the multi-turn scenario, or is it the coding agent being forced into a serial processing pattern?

swyx•3mo ago
hey sorryjust saw this. i do think its majority serial processing, BUT, parallel calling the same tools also gets issues that i honestly havent spent the time to dig into (something something locks and threading). all i know is ive been stuck in very very super slow/tool calls myself in Windsurf/other AI IDEs and that was a drag.

for another take on latency attribution see https://x.com/silasalberti/status/1979310181424206143

bluelightning2k•3mo ago
No question just wanted to say good job and thanks as a user. Same with deepwiki and codemaps.
llllm•3mo ago
Did you intend to answer them, or you just wanted the questions?
foodbaby•3mo ago
What base model did you use?
swyx•3mo ago
(coauthor) main charts/evals here https://x.com/cognition/status/1978867021669413252

you can try the https://playground.cognition.ai/ here

i wrote a longer explainer here https://x.com/swyx/status/1978874342743343254 but saving you the click

this was a perspective cut from the blogpost, but let me explain why subagents kill long context

Like you can spend $500m building 100 million context models, and they would be 1) slow, 2) expensive to use, 3) have huge context rot. O(n) is the lower bound.

Cog's approach is something you learn in day 1 of CS50 - divide and parallelize. Embeddings are too dumb, Agentic Search is too slow. So train limited-agency (max 4 turns), natively parallel tool calling (avg parallelism of 7-8, custom toolset) fast (2800tok/s) subagents to give the performance of Agentic Search under an acceptable "Flow Window" that feels immaterially slower than Embeddings.

The benefit of this is threefold:

- 8 ^ 4 toolcalls cover a very large code search space. can compound subagent calls if more needed.

- predictable cost & end to end latency

- subagent outputs "clean" contexts, free of context failure modes like context poisoning and context rot

we originally called this Rapid Agentic Search, to contrast with RAG. but Fast Context rolls off the tongue better.

-- Second perspective --

The Fundamental Equation of Coding Agents is:

Coding Agent Performance = Ability to Read the Right Files * Ability to Generate the Right Diffs

Fast Context is Cognition's first solution for the Read. As codebases get larger and and tasks get more complex, Reads get more important. the average production codebase first query in Cascade is >60% just searching and reading files.

But if this were just about speed, it might not be that exciting. I think there are unappreciated effects in performance as well when you have very good context. In other words:

Context Engineering is Actually Very Important. Too important for humans and hardcoded rules.

The swe-greps are the first dedicated context engineer agent models.

vessenes•3mo ago
Thanks for the summary. I noticed from the announcement you trained on parallel tool calling to save on serial round tripping. This is awesome.

Most LLM coding is so slow that you're permanently out of flow state, and in 'manager' state right now - I'm interested in a future where you've got enough fast low TTFT support that an engineer could maintain flow state and have sort of super power type productivity at the same time, and this tool makes me think of that.

That is, it looks fast enough to be used as a sort of sidebar info tool, as in "what you're coding might need / refer to these other parts of the codebase" -- effectively increasing an engineer's working memory. Super cool. And obviously useful for an AI engineer as well. Thanks for the writeup!

SafeDusk•3mo ago
Any plans to offer this as a tool/MCP server for other coding agents or is it going to be Windsurf exclusive?
swyx•3mo ago
we have other things in store that can be used by other coding agents, this one was tuned to use custom fast search tools that kinda wouldnt be useful in other agents
ntntnt•3mo ago
lol dead thread, cognition begging to grab some traction in this space.
kburman•3mo ago
I thought https://playground.cognition.ai/ was just returning some cached query results, but no, they’re actually spinning up real VMs and running live queries without any authentication or restrictions. That must be costing them a fortune.
groby_b•3mo ago
Currently, all queries are returning "We're under load and processing too many requests. Please try again later."

So that's how that is going ;)

awsanswers•3mo ago
LLM product managers: Show me what's in the context convenient to where I am prompting. Likely the user knowing and editing the precise context between requests will be a user task for a long time
breadislove•3mo ago
guys please release the benchmark or the benchmark code. like this is just "trust me bro"
swyx•3mo ago
well thats what the playground is for! playground.cognition.ai
breadislove•3mo ago
yeah but if people would like to double check the results it would be nice to have the actual benchmark. especially given that your playground is broken...

"We ran into an error processing your request. Please try again"

seanobannon•3mo ago
This link redirects to https://cognition.ai/blog/swe-grep now?
swyx•3mo ago
got a lot of traffic and was taken down temporarily for a couple reasons - team got it online again last night
bluelightning2k•3mo ago
This is really cool. Thank you for this. I'm a Windsurf user since launch and was VERY pleasantly surprised to see this pop up.

I also enjoyed the tech write-up. It's good to see REAL substantial engineering like this which is both highly impressive and highly productized.

bluelightning2k•3mo ago
Actually I do have a question! How come things as substantial as this were just released and not part of a "wave" ? I quite liked the waves way of doing things! Great work either way.
SafeDusk•3mo ago
Kickstarting an exploratory open version here https://github.com/aperoc/op-grep since it doesn't look like they will do it.
unturned3•3mo ago
This has very little resemblance of SWE-grep haha. At least fine-tune a small pre-trained LLM or something on a retrieval dataset. But no, this literally tries to train a small RNN from scratch to retrieve results given a natural language query...