frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Moonshine Open-Weights STT models – higher accuracy than WhisperLargev3

https://github.com/moonshine-ai/moonshine
243•petewarden•11h ago•51 comments

Show HN: Quantifying opportunity cost with a deliberately "simple" web app

https://shouldhavebought.com/
15•b0bbi•17h ago•11 comments

Show HN: Scheme-langserver – Digest incomplete code with static analysis

https://github.com/ufo5260987423/scheme-langserver
4•ufo5260987423•1d ago•0 comments

Show HN: Emdash – Open-source agentic development environment

https://github.com/generalaction/emdash
153•onecommit•15h ago•59 comments

Show HN: WinterMute Local-first OSINT workbench with native Tor and AI analysis

https://wintermute.stratir.com
2•VanceVP•46m ago•0 comments

Show HN: ArcticKey – Managed Redis (Valkey) Hosted in the EU

https://arctickey.com
2•simskiie•50m ago•1 comments

Show HN: Context Mode – 315 KB of MCP output becomes 5.4 KB in Claude Code

https://github.com/mksglu/claude-context-mode
56•mksglu•3h ago•17 comments

Show HN: Recursively apply patterns for pathfinding

https://pattern-pathfinder.vercel.app/?fixtureId=%7B%22path%22%3A%22site%2Fexamples%2F_intro.fixt...
22•seveibar•11h ago•5 comments

Show HN: Workz – Zoxide for Git worktrees (auto node_modules and .env, AI-ready)

https://github.com/rohansx/workz
2•rohansx•2h ago•0 comments

Show HN: Tag Promptless on any GitHub PR/Issue to get updated user-facing docs

34•prithvi2206•15h ago•6 comments

Show HN: enveil – hide your .env secrets from prAIng eyes

https://github.com/GreatScott/enveil
194•parkaboy•1d ago•122 comments

Show HN: Chaos Monkey but for Audio Video Testing (WebRTC and UDP)

https://github.com/MdSadiqMd/AV-Chaos-Monkey
38•MdSadiqMd•2d ago•2 comments

Show HN: PgDog – Scale Postgres without changing the app

https://github.com/pgdogdev/pgdog
318•levkk•1d ago•58 comments

Show HN: A free tool to turn your boring screenshots brutalist in seconds

https://neo.retroui.dev
3•devarifhossain•4h ago•0 comments

Show HN: Babyshark – Wireshark made easy (terminal UI for PCAPs)

https://github.com/vignesh07/babyshark
142•eigen-vector•1d ago•45 comments

Show HN: Sowbot – Open-hardware agricultural robot (ROS2, RTK GPS)

https://sowbot.co.uk/
178•Sabrees•1d ago•45 comments

Show HN: X86CSS – An x86 CPU emulator written in CSS

https://lyra.horse/x86css/
259•rebane2001•1d ago•90 comments

Show HN: Declarative open-source framework for MCPs with search and execute

https://hyperterse.com
9•samrith•12h ago•2 comments

Show HN: Steerling-8B, a language model that can explain any token it generates

https://www.guidelabs.ai/post/steerling-8b-base-model-release/
319•adebayoj•1d ago•87 comments

Show HN: A Visual Editor for Karabiner

https://github.com/zkwokleung/karabiner-config-editor
4•zkwokleung•6h ago•0 comments

Show HN: AI Timeline – 171 LLMs from Transformer (2017) to GPT-5.3 (2026)

https://llm-timeline.com/
169•ai_bot•2d ago•57 comments

Show HN: StreamHouse – S3-native Kafka alternative written in Rust

https://github.com/gbram1/streamhouse
5•gbram•6h ago•2 comments

Show HN: Cellarium: A Playground for Cellular Automata

https://github.com/andrewosh/cellarium
39•andrewosh•4d ago•1 comments

Show HN: ProdRescue AI – Turn Slack war-rooms and raw logs into incident reports

https://www.prodrescueai.com/
4•devrimozcay•13h ago•0 comments

Show HN: Ghist – Task management that lives in your repo

https://github.com/unnecessary-special-projects/ghist
16•nxnze•16h ago•2 comments

Show HN: CIA World Factbook Archive (1990–2025), searchable and exportable

https://cia-factbook-archive.fly.dev/
486•MilkMp•2d ago•99 comments

Show HN: Mnemosyne – Cognitive memory OS for AI agents (zero LLM calls)

https://github.com/28naem-del/mnemosyne
5•mnemosy•11h ago•1 comments

Show HN: Open-source KYC plugin for Claude – 95min→27min, £85K→£240/year

https://github.com/vyayasan/kyc-analyst
4•vyayasan•12h ago•3 comments

Show HN: Praxis, my personal take on Compound Engineering with AI

https://github.com/DFilipeS/praxis
4•DFilipeS•12h ago•0 comments

Show HN: OpenPawz – Open-source desktop AI agent platform (Rust/Tauri, 75 tools)

https://github.com/OpenPawz/openpawz
2•openpawz•12h ago•1 comments
Open in hackernews

Show HN: Context Mode – 315 KB of MCP output becomes 5.4 KB in Claude Code

https://github.com/mksglu/claude-context-mode
56•mksglu•3h ago
Every MCP tool call dumps raw data into Claude Code's 200K context window. A Playwright snapshot costs 56 KB, 20 GitHub issues cost 59 KB. After 30 minutes, 40% of your context is gone.

I built an MCP server that sits between Claude Code and these outputs. It processes them in sandboxes and only returns summaries. 315 KB becomes 5.4 KB.

It supports 10 language runtimes, SQLite FTS5 with BM25 ranking for search, and batch execution. Session time before slowdown goes from ~30 min to ~3 hours.

MIT licensed, single command install:

/plugin marketplace add mksglu/claude-context-mode

/plugin install context-mode@claude-context-mode

Benchmarks and source: https://github.com/mksglu/claude-context-mode

Would love feedback from anyone hitting context limits in Claude Code.

Comments

handfuloflight•2h ago
One moment you're speaking about context but talking in kilobytes, can you confirm the token savings data?

And when you say only returns summaries, does this mean there is LLM model calls happening in the sandbox?

mksglu•2h ago
Hey! Thank you for your comment! There are test examples in the README. Could you please try them? Your feedback is valuable.
mksglu•2h ago
For your second question: No LLM calls. Context Mode uses algorithmic processing — FTS5 indexing with BM25 ranking and Porter stemming. Raw output gets chunked and indexed in a SQLite database inside the sandbox, and only the relevant snippets matching your intent are returned to context. It's purely deterministic text processing, no model inference involved.
handfuloflight•1h ago
Excellent, thank you for your responses. Will be putting it through a test drive.
mksglu•1h ago
Sure, thank you for your comment!
sim04ful•2h ago
Looks pretty interesting. How could i use this on other MCP clients e.g OpenCode ?
mksglu•2h ago
Hey! Thank you for your comment! You can actually use an MCP on this basis, but I haven't tested it yet. I'll look into it as soon as possible. Your feedback is valuable.
nightmunnas•2h ago
nice, I'd love to se it for codex and opencode
mksglu•2h ago
Thanks! Context Mode is a standard MCP server, so it works with any client that supports MCP — including Codex and opencode.

Codex CLI:

  codex mcp add context-mode -- npx -y context-mode
Or in ~/.codex/config.toml:

  [mcp_servers.context-mode]
  command = "npx"
  args = ["-y", "context-mode"]
opencode:

In opencode.json:

  {
    "mcp": {
      "context-mode": {
        "type": "local",
        "command": ["npx", "-y", "context-mode"],
        "enabled": true
      }
    }
  }
We haven't tested yet — would love to hear if anyone tries it!
vicchenai•2h ago
The BM25+FTS5 approach without LLM calls is the right call - deterministic, no added latency, no extra token spend on compression itself.

The tradeoff I want to understand better: how does it handle cases where the relevant signal is in the "low-ranked" 310 KB, but you just haven't formed the query that would surface it yet? The compression is necessarily lossy - is there a raw mode fallback for when the summarized context produces unexpected downstream results?

Also curious about the token count methodology - are you measuring Claude's tokenizer specifically, or a proxy?

mksglu•2h ago
Great questions.

--

On lossy compression and the "unsurfaced signal" problem:

Nothing is thrown away. The full output is indexed into a persistent SQLite FTS5 store — the 310 KB stays in the knowledge base, only the search results enter context. If the first query misses something, you (or the model) can call search(queries: ["different angle", "another term"]) as many times as needed against the same indexed data. The vocabulary of distinctive terms is returned with every intent-search result specifically to help form better follow-up queries.

The fallback chain: if intent-scoped search returns nothing, it splits the intent into individual words and ranks by match count. If that still misses, batch_execute has a three-tier fallback — source-scoped search → boosted search with section titles → global search across all indexed content.

There's no explicit "raw mode" toggle, but if you omit the intent parameter, execute returns the full stdout directly (smart-truncated at 60% head / 40% tail if it exceeds the buffer). So the escape hatch is: don't pass intent, get raw output.

On token counting:

It's a bytes/4 estimate using Buffer.byteLength() (UTF-8), not an actual tokenizer. Marked as "estimated (~)" in stats output. It's a rough proxy — Claude's tokenizer would give slightly different numbers — but directionally accurate for measuring relative savings. The percentage reduction (e.g., "98%") is measured in bytes, not tokens, comparing raw output size vs. what actually enters the conversation context.

rcarmo•1h ago
Nice trick. I’m going to see how I can apply it to tool calls in pi.dev as well
mksglu•1h ago
That means a lot, thank you! Would love to hear your feedback once you try it — and an upvote would be much appreciated if you find it useful
robbomacrae•1h ago
Really cool. A tangential task that seems to be coming up more and more is masking sensitive data in these calls for security and privacy. Is that something you considered as a feature?
mksglu•47m ago
Good question.

The SQLite database is ephemeral — stored in the OS temp directory (/tmp/context-mode-{pid}.db) and scoped to the session process. Nothing persists after the session ends. For sensitive data masking specifically: right now the raw data never leaves the sandbox (it stays in the subprocess or the temp SQLite store), and only stdout summaries enter the conversation. But a dedicated redaction layer (regex-based PII stripping before indexing) is an interesting idea worth exploring. Would be a clean addition to the execute pipeline.

virgilp•25m ago
> Nothing persists after the session ends.

Does that mean that if I exit claude code and then later resume the session, the database is already lost? When exactly does the session end?

mksglu•24m ago
Yes — the database is tied to the MCP server process, so it's created fresh on each claude launch and lost when you exit; resuming a session starts a new process with a new empty database.