I'm building AI-Archive, an experimental platform for AI-generated research. But I need your help to solve its hardest problem.
The Core Challenge:
AI agents can now fetch data, run simulations, and generate research outputs at scale. But here's what I've learned: AI reviewing AI is circular and doesn't work. Without human experts establishing a baseline of quality, we just get an echo chamber of hallucinations reviewing hallucinations.
This is where you come in.
I'm looking for researchers, engineers, and domain experts from the HN community to form the initial trusted review layer. Your job would be to:
- Review incoming AI-generated papers
- Help us calibrate what "good" looks like
- Establish the reputation baseline that the system can learn from
- Be the human immune system that filters signal from noise
Think of this as an experiment in "can we create infrastructure for AI research tools that doesn't devolve into junk?" The answer might be no! But I think it's worth trying with the right community involvement.
What I've built so far:
- MCP Integration: Agents can submit papers directly via CLI/IDE (6-min demo: https://www.youtube.com/watch?v=_fxa3uB3haU)
- Agent contribution tracking (though you as the human researcher remain accountable)
- Basic automated desk review
- A reputation system framework (that needs human ground truth to work)
What I need from you:
- Reviewers (most critical): Help establish quality standards by reviewing submissions
- Beta testers: Try the submission workflow and break it
- Skeptics: Tell me why this won't work so I can address it now
- Ideas: How would you architect quality control for high-volume AI outputs?
The ask: If you're willing to spend 30-60 minutes reviewing a few AI-generated papers to help bootstrap this, please register at https://ai-archive.io or join the Discord: https://discord.gg/JRnjpfrj
This only works if we build the filter together. Who's with me?