Based on this post's current position on the front page it kind of seems to fall in line with a pattern we've all been seeing the past few months: HN is finally majority onboard with believing in the usefullness of coding agents and is celebrating this by rediscovering each and every personal "I improved CC by doing [blank] thing" from scratch project.
That's all whatever. Fine. But what I'm really curious about is does the HN community actually look at the random LLM-generated statistic-vomit text posted by creators like this and find themselves convinced?
I ask because if you're new to random stat vomit you're going to find yourself having to deal with it all the time soon, and I've yet to find good meta discussions about how we find signal in this noise. I used to use HN or selected reddit community upvotes as a first pass "possibly important" signal, but its been getting worse and worse, illustrated by posts like this getting upvoted to the top without any genuine discssion.
I do not understand why this project in particular have set you off.
Their README looks much better than many I've seen on HN:
- no annoying verbosity, that is so prevalent in AI-generated text - not too many buzzwords (they're not saying "agentic" every sentence) - it is very clear what exactly project is supposed to do and why it can be useful
Personally, I upvoted this because I wanted to do something similar for a long time but never got around to it.
It's much easier to give your agents the LSP for the language(s) it's working on.
concensure•1h ago
The Solution: Semantic uses a local AST (Abstract Syntax Tree) parser to build a Logical Node Graph of the codebase. Instead of guessing what is relevant, it deterministically retrieves the specific functional skeletons and call-site signatures required for a task. The Shift: From "Token Savings" to "Step Savings"
Earlier versions of this project focused on minimizing tokens per call. However, our latest benchmarks show that investing more tokens into high-precision context leads to significantly fewer developer intervention steps. Latest A/B Benchmark (2026-03-27)
Run Variant Token Delta (per call) Step Savings (vs Baseline) Task Success Baseline (2026-03-13) -18.62% — 11/11 Hardened A +8.07% — 11/11 Enhanced (2026-03-27) -6.73% +27.78% 11/11 Key Takeaways: Detailed breakdowns of the task suite and methodology are available in docs/AB_TEST_DEV_RESULTS.md.