So I built an MCP server that does the searching before you write code. It scans 5 real-time sources (GitHub, Hacker News, npm, PyPI, Product Hunt) and returns a quantified reality_signal (0-100) with actual evidence — repo counts, star counts, top competitors, and pivot suggestions.
Example: "AI code review tool" → reality_signal: 90, 847 repos, top competitor reviewdog (9,094 stars), 254 HN mentions.
What it's NOT: not a business plan generator, not an LLM opinion wrapper. Every number comes from a real API call you can verify.
- Quick mode: GitHub + HN (default) - Deep mode: all 5 sources in parallel - Works with Claude Desktop, Cursor, Claude Code - One-line install: uvx idea-reality-mcp - Also available as web demo (no install needed)
GitHub: https://github.com/mnemox-ai/idea-reality-mcp Web demo: https://mnemox.ai/check MCP Registry: io.github.mnemox-ai/idea-reality-mcp
Built with Python + FastMCP, 120 tests, published on PyPI. Happy to answer any questions about the scoring algorithm or MCP integration.
mnemoxai•1h ago
I use Claude Code and Cursor daily. The pattern I kept hitting: describe an idea → AI starts coding → hours later I search GitHub and realize there are already 10+ mature projects doing the same thing.
The fix was simple: search BEFORE coding. But I wanted it automated — not a manual GitHub search, but something the AI agent does by default as part of its workflow.
Technical details: - Keyword extraction uses a 3-stage dictionary pipeline (90+ intent anchors, 80+ synonym expansions, 150+ Chinese term mappings). No LLM needed for the MCP path — fast, free, deterministic. - Scoring is intentionally simple and explainable: weighted average across sources, not ML. You can read the formula in scoring/engine.py. - All sources are searched in parallel with asyncio.gather. Deep mode hits 5 APIs concurrently. - Zero storage — nothing is logged or saved. Protocol, not SaaS.
The most useful integration pattern: add "always run idea_check before building anything new" to your CLAUDE.md or Cursor rules. The agent will automatically check before writing code.
Honest limitation: dictionary-based keyword extraction has a coverage ceiling. Niche domains outside tech (e.g., "pet acupuncture booking app") may get less precise results. The web demo at mnemox.ai/check uses LLM extraction for better coverage.
Would love feedback on the scoring formula — it's the part I'm least confident about. Is a single 0-100 number useful, or would you prefer something more granular?