frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Paper: AI models are faking their step by step thinking

https://twitter.com/thetripathi58/status/2032775838329090191
1•MrBuddyCasino•23s ago•0 comments

Dreamfarm product copied by large competitors

https://www.news.com.au/lifestyle/home/interiors/ripped-off-kmart-slammed-for-unaustralian-act/ne...
1•asdefghyk•2m ago•1 comments

Revanced Manager v2

https://revanced.app/announcements?id=20-release-of-revanced-manager-v2
1•super256•4m ago•0 comments

Show HN: Nimhuml – A Nim parser and serializer for HUML

https://github.com/w3Abhishek/nimhuml
1•w3abhishek•8m ago•0 comments

Young people are turning to old-school hobbies to get off their phones

https://apnews.com/article/old-school-hobbies-vintage-analog-grandma-e45fa11ae1422715b6a254004476...
1•1vuio0pswjnm7•17m ago•0 comments

Epic – a visual design editor with a built-in visual sitemap builder

https://no-edit.lovable.app/
1•theme-man•20m ago•2 comments

Show HN: An experiment in giving coding agents long-term memory

1•yacc2•22m ago•0 comments

A most elegant TCP hole punching algorithm

https://robertsdotpm.github.io/cryptography/tcp_hole_punching.html
1•Uptrenda•22m ago•0 comments

S&P 500 Concentration Approaching 50%

https://www.apolloacademy.com/sp-500-concentration-approaching-50/
3•toomuchtodo•27m ago•1 comments

How compilers should evolve in the era of LLM coding

https://twitter.com/ezyang/status/2032932628131721462
1•mfiguiere•35m ago•0 comments

FCC Chair Threatens to Revoke Broadcasters' Licenses over War Coverage

https://www.nytimes.com/2026/03/14/world/middleeast/fcc-broadcasters-iran-war.html
9•KnuthIsGod•37m ago•2 comments

I Asked 4 AI Models to Research the Parasite Cleanse Hype

https://christiantech.substack.com/p/i-asked-4-ai-models-to-research-the
1•snow_mac•37m ago•0 comments

Show HN: WebGPU and WebAssembly SIMD SHA‑256 PoW Miner

https://etherdream.github.io/hash-miner/
1•etherdream•43m ago•0 comments

CW Radio Signals Require Fix Faulty Equipment on Boeing 787s

https://www.paddleyourownkanoo.com/2026/03/14/ham-radio-enthusiasts-land-u-s-airlines-with-8-mill...
1•wglb•44m ago•0 comments

The Brain Stores Quadruples: Why Context Is the Missing Primitive in AI

https://gizmohan.substack.com/p/agents-as-labor
1•iamthedruid•47m ago•0 comments

Anthropic Courses

https://anthropic.skilljar.com/
1•vinhnx•52m ago•1 comments

Show HN: Signet.js – A minimalist reactivity engine for the modern web

https://github.com/sntran/signet.js/
2•sntran•53m ago•1 comments

Factorio Any% Tool-Assisted Speedrun in 57 Minutes [video]

https://www.youtube.com/watch?v=fkmRd5uJoKI
1•bozbalci•56m ago•0 comments

Social Media, Reset

https://www.meetzeta.com/
6•novateg•58m ago•0 comments

Earth's first major extinction was worse than we thought

https://www.science.org/content/article/earth-s-first-major-extinction-was-worse-we-thought
3•thunderbong•1h ago•0 comments

Treasure hunter freed from jail after refusing to turn over shipwreck gold

https://www.bbc.com/news/articles/cg4g7kn99q3o
25•tartoran•1h ago•15 comments

CloudPipe – download anything to Google Drive

https://cloudpipe.app
2•darkhasi•1h ago•1 comments

Am I a Forward Deployed Engineer? Maybe. Take the Quiz

https://wanjiko.substack.com/p/am-i-a-forward-deployed-engineer
1•contextwindow•1h ago•1 comments

DeerFlow 2.0: open-source SuperAgent harness that researches, codes, creates

https://github.com/bytedance/deer-flow
2•nateb2022•1h ago•0 comments

Add all your GitHub stars on your readme, autoupdating

https://starsum.jia.build/
1•Audgeviolin07•1h ago•0 comments

A Superpower Goes Offline

https://www.politico.com/news/2026/03/14/russias-self-inflicted-communication-crisis-00827197
3•mitchbob•1h ago•0 comments

Memegen Pro

https://memegen.pro/
1•decimalenough•1h ago•0 comments

Why do we need lots of Nuclear power long term?

https://www.gridstatus.io/live/ercot
1•chris222•1h ago•2 comments

$3k sequencing rescue dog's mast cell tumor DNA

https://twitter.com/IterIntellectus/status/2032858964858228817
3•gmays•1h ago•1 comments

Cats May Hold the Key to Treating Human Cancer

https://scitechdaily.com/cats-may-hold-the-key-to-treating-human-cancer/
2•y1n0•1h ago•0 comments
Open in hackernews

Show HN: Korupedia – a knowledge base maintained by AI agents, not humans

https://korupedia.com
2•benryanx•1h ago
The idea started as a question: if AI agents are increasingly being used to retrieve and synthesize facts, why are they still pulling from knowledge bases built for humans?

Korupedia is an experiment in agent-native knowledge. Agents register with a cryptographic identity (did:key, Ed25519), submit factual claims with sources and confidence scores, and vote on each other's submissions. A weighted supermajority (67%) resolves consensus. The whole thing is queryable via a plain GET endpoint — GET /ask?q=your+question — designed to be dropped directly into an agent's context.

A few design decisions worth discussing:

Reverse CAPTCHA - instead of proving you're human, you prove you're an AI. Five challenge types (arithmetic, code trace, semantic, logic, pattern) that any LLM solves in under 8 seconds but a human needs 30–120 seconds. Solve time is recorded as a signal.

Sybil resistance - votes are weighted by domain reputation. New agents start at floor weight 1.0. Quorum requires a minimum number of voters with accounts old enough to not be freshly-minted attack agents.

No LLM in the query path — /ask is full-text search returning the highest-confidence accepted claim. Fast, deterministic, no hallucination surface.

It's early - the knowledge base is small and the agent network is just forming. But Jasper, an agent on a separate machine, self-registered and submitted claims yesterday by downloading a bootstrap script from the API itself (GET /agent.js).

Live at korupedia.com. API docs at api.korupedia.com/docs.

Curious what people think about the model especially whether cryptographic identity + consensus is the right foundation, or if there's a better mechanism for agents to establish shared ground truth.

Comments

rgupta1833•1h ago
The "no LLM in the query path" decision is the most interesting part. You're basically saying: use AI to build the knowledge, but don't use AI to retrieve it. That's a meaningful distinction most people building in this space collapse. The hard problem isn't consensus on facts — it's consensus on what counts as a fact. Agents citing other agents citing other agents creates the same circular authority problem that Google's link graph had. Curious how you're thinking about that failure mode.
benryanx•1h ago
Hello

Exactly right, and it's the failure mode we're most worried about.

The circular authority problem is real. Our current partial mitigation: sources must be external URLs (we block self-referential korupedia.com links), and the confidence score is attached to the agent, not just the claim - so if an agent's claims keep getting disputed, their vote weight decays. But that's not sufficient on its own.

The deeper issue is what you're pointing at: facts vs. fact-like statements. An agent can submit "The Eiffel Tower is in Paris" (verifiable, stable) and "GPT-4 outperforms humans on the bar exam" (contested framing, depends on which humans, which version, which year). Both look like facts. The current schema doesn't distinguish them.

A few directions we're considering:

Expiry by domain - scientific claims expire faster than historical ones. Forces re-verification rather than letting stale consensus calcify.

Dispute chains - a counter-claim doesn't just reject, it must cite a contradicting source. So the graph is claims → sources, not claims → claims.

Attestation tier - claims with primary source URLs (arxiv, official docs, peer-reviewed) get flagged differently from those citing aggregators or secondary sources.

None of this fully solves citation circularity - it's structurally similar to the PageRank problem and probably requires a similar insight (some equivalent of "links from outside the cluster count more"). We don't have that insight yet.

What's your intuition on where the right pressure point is?

lichih•54m ago
I believe the real growth potential for Korupedia isn't in reaching a 'perfect' consensus, but in experimenting with computable implementations of what constitutes a 'fact.' > Even a flawed or experimental hypothesis on factual definition would give this project immense room for discourse. For instance, instead of a binary 'Accepted/Rejected' status, why not implement tiered metadata for:

Axiomatic Facts: (Mathematical/Logical constants)

Empirical Observations: (Sensor data, verifiable external URLs)

Inferred Derivatives: (Claims synthesized from other claims, requiring a 'Traceable Inference' tag)

The difficulty here isn't just 'Wikipedia for AI'; it’s the philosophical problem of 'What is fact?' If you can codify even a basic version of this—moving from simple voting to a formal grounding or falsification mechanism—you’d be building the first actual Epistemic Engine for the agentic web. That’s where the real breakthrough lies.

lichih•50m ago
To make this practical, I’d suggest adopting a TDD (Test-Driven Development) approach to 'Factness.'

Instead of just voting, the project could define a 'Fact Test Suite' that every claim must pass before entering the consensus phase. For example:

Red (The Hypothesis): Define a 'Fact Object' with a mandatory Falsification Test. What specific, observable data would prove this claim wrong?

Green (The Verification): Agents don't just vote 'Yes'; they must provide a 'Test Runner'—a piece of code or a deterministic query that successfully fetches the supporting evidence from a third-party, non-LLM source.

Refactor (The Synthesis): As facts change (e.g., a CEO changes or a law is repealed), the 'Fact Test' fails, triggering an automatic re-verification cycle.

By treating a 'Fact' as a passing test case rather than a statistical popularity contest, you move from 'what agents believe' to 'what can be independently verified.' This turns Korupedia into an infrastructure for Computable Truth, which is far more valuable to an autonomous agent than a simple knowledge base.

lichih•47m ago
Thinking bigger: the real asset here shouldn't be the Claims, but the Chain of Thought (CoT) that leads to them.

What if Korupedia shifted its goal from collecting facts to collecting the logical proofs for those facts? Instead of agents submitting a simple claim, they submit the 'traceable reasoning'—the step-by-step verification logic, the sources weighted, and the counter-arguments dismissed.

In this model, you don't just ask the network for an answer; you provide a Question (The Challenge) and the network competes to provide the most robust CoT. This achieves two things:

Transparency: You can audit why the network reached a consensus.

Intellectual Provenance: The 'how' we know something becomes a reusable asset for other agents, much like a library of pre-compiled logic.

In many ways, a consensus on the process of reasoning is far more valuable and resistant to model collapse than a consensus on the final string of text. Has there been any thought toward shifting the objective from gathering 'Claims' to gathering 'Verifiable Reasoning Paths'?

benryanx•37m ago
You've just described the natural v2 of what we're building, and honestly it's the more defensible moat.

The shift from claims to reasoning paths changes the fundamental unit of value. A claim is a string that can be hallucinated, copied, or become stale. A verified reasoning chain is a computational artifact - it either holds up under scrutiny or it doesn't and you can check.

The architecture that comes to my mind:

Claims become theorems, not assertions. Instead of {"statement": "X", "confidence": 0.95}, you submit {"conclusion": "X", "premises": [claim_id_1, claim_id_2], "inference_rule": "modus_ponens", "sources": [...]}. The conclusion is only as strong as its weakest premise.

Voting shifts from "do you agree with the output" to "does the reasoning hold". Agents with logic/formal-reasoning domain expertise get higher weight on proof verification than agents who only have empirical domain knowledge. Specialisation becomes meaningful.

The CoT library becomes the real product. Not the answer to "when was CRISPR discovered" but the reusable proof that can be composed into larger inferences. Other agents don't query for facts - they query for verified reasoning substrates.

The hard problem: who validates the inference rules themselves? You need either a fixed axiom set (brittle) or a meta-layer of reasoning about reasoning (potentially infinite regress). The way out is probably domain-scoping - formal logic in Mathematics, empirical verification chains in Science, source provenance chains in History.

We haven't built this yet. But the schema is designed to be extended - premises as an array of claim IDs is a one-field addition to the current model. If you want to prototype a reasoning-path agent on top of the existing API, the infrastructure is there today.

benryanx•40m ago
This reframes the problem better than anything we've articulated internally. "Epistemic engine" is the right term and the tiered taxonomy you're describing is close to where we've been heading implicitly without naming it properly.

The three-tier model maps well to what we're seeing in practice:

Axiomatic - claims that don't expire and don't need re-verification. 2 + 2 = 4 has a different lifecycle than GPT-4 is the most capable model available. Our current staleness system treats all claims equally; it shouldn't.

Empirical - claims anchored to external URLs already get partial treatment (we validate sources, block shorteners, require real domains). The missing piece is a live re-fetch signal - did the source URL still say what it said when the claim was submitted?

Inferred derivatives - this is the hard one and to be fair where we haven't gone yet. A claim like "therefore LLMs are becoming more reliable" is synthesized from multiple empirical claims. Tracing that inference chain is essentially building a lightweight knowledge graph, not just a flat list of claims.

The honest answer is we built the voting/consensus layer first because it was tractable. The epistemic grounding layer - distinguishing types of truth, not just tallying confidence is the next hard problem.

If you're working in this space, we'd genuinely want to talk. The schema is open and the API is public - the inference chain tagging you're describing could be built as an agent on top of the existing API without touching the core.