frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Apple's Multibillion-Dollar Push to Make Chips in the U.S. [video]

https://www.youtube.com/watch?v=ktFlaBhpMu8
1•keepamovin•33s ago•0 comments

Amazon would rather blame its own engineers than its AI

https://www.theregister.com/2026/02/24/amazon_blame_human_not_ai/
1•beardyw•3m ago•0 comments

Slow Tuesday Night by R. A. Lafferty (1965)

https://www.baen.com/Chapters/9781618249203/9781618249203___2.htm
1•monort•5m ago•0 comments

Why the Intelligence Crisis Scenario Is Wrong

https://deadneurons.substack.com/p/why-the-intelligence-crisis-scenario
1•nr378•5m ago•0 comments

Data Crew's Route Optimiser Framework

https://tech.marksblogg.com/data-crew-route-optimiser-solver-framework.html
1•marklit•5m ago•0 comments

Managed Iceberg for Streaming with PostgreSQL Simplicity – RisingWave Open Lake

https://risingwave.com/lakehouse/
1•AnneWodell•6m ago•0 comments

FinCrew: Multi-Agent AI Financial Intelligence

1•adnan_builds•6m ago•1 comments

Pentagon sets Friday deadline for Anthropic to abandon ethics rules for AI

https://www.politico.com/news/2026/02/24/hegseth-sets-friday-deadline-for-anthropic-to-drop-its-a...
2•borski•10m ago•0 comments

Orbital datacenters are a pie-in-the-sky idea: Gartner

https://www.theregister.com/2026/02/25/gartner_orbiting_datacenter_peak_insanity/
1•beardyw•11m ago•0 comments

AMD and Meta strike $100B AI deal that includes 10% stock deal

https://www.tomshardware.com/tech-industry/artificial-intelligence/amd-meta-100-billion-deal
1•pjmlp•11m ago•0 comments

Core Banking Is a Terrible Idea. It Always Was

https://andrewbaker.ninja/2026/02/24/core-banking-is-a-terrible-idea-it-always-was/
2•jinonoel•12m ago•0 comments

Show HN: RAgent – Claude Code on a VPS So Remote Control Never Drops

https://github.com/Chris-bzst/ragent
1•chris-bzst•16m ago•0 comments

ShipGrowth – Discover, Compare and Submit Best AI Tools

https://shipgrowth.dev
1•duanhjlt•18m ago•0 comments

Stripe is reportedly eyeing deal to buy some or all of PayPal

https://techcrunch.com/2026/02/24/stripe-is-reportedly-eyeing-deal-to-buy-some-or-all-of-paypal/
1•taubek•20m ago•0 comments

Software engineers could go extinct this year, says Claude Code creator

https://fortune.com/2026/02/24/will-claude-destroy-software-engineer-coding-jobs-creator-says-pri...
3•bfmalky•21m ago•1 comments

Waymo Expands Autonomous Rides to Dallas, Houston, San Antonio, and Orlando

https://waymo.com/blog/2026/02/dallas-houston-san-antonio-orlando
4•integralpilot•22m ago•1 comments

Pg_doom

https://github.com/DreamNik/pg_doom
3•fla•26m ago•0 comments

Asahi Linux in the Cloud: Scaleway Launches Dedicated M2 Pro Mac Mini Servers

https://www.scaleway.com/en/mac-mini-asahi-linux/
3•Lwrless•26m ago•0 comments

Claw-Guard.org – Agentic Monetisation Middleware That Works

https://claw-guard.org
1•gmerc•26m ago•0 comments

WiseTech Global to cut 2k jobs as AI ends era of 'manually writing code'

https://www.abc.net.au/news/2026-02-25/wisetech-job-losses-losing-2000-over-next-two-years-coding...
1•BrissyCoder•27m ago•1 comments

Show HN: AGX v2 – From multi-agent chat to execution graph

https://github.com/ramarlina/agx
1•Mendrika•30m ago•0 comments

What's so hard about continuous learning?

https://www.seangoedecke.com/continuous-learning/
2•rbanffy•35m ago•0 comments

Show HN: Chorus – Open-source Agent and human collaboration platform on AI-DLC

https://github.com/Chorus-AIDLC/Chorus
2•fennu637•36m ago•0 comments

Show HN: I let Claude autonomously deploy OpenClaw and write an honest review

https://blog.rezvov.com/deploying-openclaw-sixteen-incidents-one-day
1•alexrezvov•36m ago•0 comments

Michael Faraday: Scientist and Nonconformist(1996)

http://silas.psfc.mit.edu/Faraday/
1•o4c•37m ago•0 comments

Show HN: ClawMoat – Open-source runtime security for AI agents (zero deps, <1ms)

https://github.com/darfaz/clawmoat
1•ildar•40m ago•0 comments

North American Computational Linguistics Open Competition

https://naclo.org/practice.php
2•Antibabelic•43m ago•0 comments

DSGym: A holistic framework for evaluating and training data science agents

https://www.together.ai/blog/dsgym
1•roody_wurlitzer•44m ago•0 comments

Show HN: crai – Get notified when your AI CLI finishes thinking

https://github.com/turtlekazu/crai
1•turtlekazu•45m ago•0 comments

Show HN: What data brokers sell your profile for vs. what ads earn from you

https://data.attentionworth.com/
1•withshakespeare•47m ago•1 comments
Open in hackernews

Show HN: Context Mode – 315 KB of MCP output becomes 5.4 KB in Claude Code

https://github.com/mksglu/claude-context-mode
37•mksglu•1h ago
Every MCP tool call dumps raw data into Claude Code's 200K context window. A Playwright snapshot costs 56 KB, 20 GitHub issues cost 59 KB. After 30 minutes, 40% of your context is gone.

I built an MCP server that sits between Claude Code and these outputs. It processes them in sandboxes and only returns summaries. 315 KB becomes 5.4 KB.

It supports 10 language runtimes, SQLite FTS5 with BM25 ranking for search, and batch execution. Session time before slowdown goes from ~30 min to ~3 hours.

MIT licensed, single command install:

/plugin marketplace add mksglu/claude-context-mode

/plugin install context-mode@claude-context-mode

Benchmarks and source: https://github.com/mksglu/claude-context-mode

Would love feedback from anyone hitting context limits in Claude Code.

Comments

handfuloflight•1h ago
One moment you're speaking about context but talking in kilobytes, can you confirm the token savings data?

And when you say only returns summaries, does this mean there is LLM model calls happening in the sandbox?

mksglu•1h ago
Hey! Thank you for your comment! There are test examples in the README. Could you please try them? Your feedback is valuable.
mksglu•1h ago
For your second question: No LLM calls. Context Mode uses algorithmic processing — FTS5 indexing with BM25 ranking and Porter stemming. Raw output gets chunked and indexed in a SQLite database inside the sandbox, and only the relevant snippets matching your intent are returned to context. It's purely deterministic text processing, no model inference involved.
handfuloflight•26m ago
Excellent, thank you for your responses. Will be putting it through a test drive.
mksglu•16m ago
Sure, thank you for your comment!
sim04ful•1h ago
Looks pretty interesting. How could i use this on other MCP clients e.g OpenCode ?
mksglu•1h ago
Hey! Thank you for your comment! You can actually use an MCP on this basis, but I haven't tested it yet. I'll look into it as soon as possible. Your feedback is valuable.
nightmunnas•1h ago
nice, I'd love to se it for codex and opencode
mksglu•55m ago
Thanks! Context Mode is a standard MCP server, so it works with any client that supports MCP — including Codex and opencode.

Codex CLI:

  codex mcp add context-mode -- npx -y context-mode
Or in ~/.codex/config.toml:

  [mcp_servers.context-mode]
  command = "npx"
  args = ["-y", "context-mode"]
opencode:

In opencode.json:

  {
    "mcp": {
      "context-mode": {
        "type": "local",
        "command": ["npx", "-y", "context-mode"],
        "enabled": true
      }
    }
  }
We haven't tested yet — would love to hear if anyone tries it!
vicchenai•51m ago
The BM25+FTS5 approach without LLM calls is the right call - deterministic, no added latency, no extra token spend on compression itself.

The tradeoff I want to understand better: how does it handle cases where the relevant signal is in the "low-ranked" 310 KB, but you just haven't formed the query that would surface it yet? The compression is necessarily lossy - is there a raw mode fallback for when the summarized context produces unexpected downstream results?

Also curious about the token count methodology - are you measuring Claude's tokenizer specifically, or a proxy?

mksglu•45m ago
Great questions.

--

On lossy compression and the "unsurfaced signal" problem:

Nothing is thrown away. The full output is indexed into a persistent SQLite FTS5 store — the 310 KB stays in the knowledge base, only the search results enter context. If the first query misses something, you (or the model) can call search(queries: ["different angle", "another term"]) as many times as needed against the same indexed data. The vocabulary of distinctive terms is returned with every intent-search result specifically to help form better follow-up queries.

The fallback chain: if intent-scoped search returns nothing, it splits the intent into individual words and ranks by match count. If that still misses, batch_execute has a three-tier fallback — source-scoped search → boosted search with section titles → global search across all indexed content.

There's no explicit "raw mode" toggle, but if you omit the intent parameter, execute returns the full stdout directly (smart-truncated at 60% head / 40% tail if it exceeds the buffer). So the escape hatch is: don't pass intent, get raw output.

On token counting:

It's a bytes/4 estimate using Buffer.byteLength() (UTF-8), not an actual tokenizer. Marked as "estimated (~)" in stats output. It's a rough proxy — Claude's tokenizer would give slightly different numbers — but directionally accurate for measuring relative savings. The percentage reduction (e.g., "98%") is measured in bytes, not tokens, comparing raw output size vs. what actually enters the conversation context.