frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Tell HN: YC companies scrape GitHub activity, send spam emails to users

591•miki123211•15h ago•222 comments

I don't need AI to build me a new app. I need it to make Jira bearable

13•niel_hu•9h ago•9 comments

Stripe closed my account – no notice – my LLC was registered using Stripe Atlas

5•travelyesim•5h ago•2 comments

I built a 151k-node GraphRAG swarm that autonomously invents SDG solutions

2•wisdomagi•8h ago•0 comments

Ask HN: Programmable Watches with WiFi?

34•dakiol•5d ago•24 comments

1Password pricing increasing up to 33% in March

142•otterley•2d ago•193 comments

Ask HN: Who Is Using XMPP?

22•nunobrito•1d ago•11 comments

Claude Code Bug triggers Rate limits without usage

3•busssard•13h ago•1 comments

Ask HN: What's it like working in big tech recently with all the AI tools?

18•ex-aws-dude•23h ago•8 comments

Comparing manual vs. AI requirements gathering: 2 sentences vs. 127-point spec

2•thesssaism•13h ago•1 comments

Ask HN: Chromebook leads for K-8 school in need?

46•techteach00•4d ago•44 comments

Ask HN: Starting a New Role with Ada

9•NoNameHaveI•1d ago•4 comments

New Claude Code Feature "Remote Control"

8•rob•1d ago•0 comments

LazyGravity – I made my phone control Antigravity so I never leave bed

4•masaTokyo•21h ago•3 comments

Ask HN: Could you create a competitor to your company at 10% of the cost?

7•TheAlchemist•1d ago•14 comments

Ask HN: How do you know if AI agents will choose your tool?

36•dmpyatyi•3d ago•21 comments

Fix cron routes: POST → GET (Vercel cron sends GET)

2•nishiohiroshi•1d ago•1 comments

Ask HN: What will happen with Anthropics ultimatum?

6•maniacwhat•1d ago•3 comments

Would you choose the Microsoft stack today if starting greenfield?

20•JB_5000•2d ago•23 comments

Ask HN: How are you controlling AI agents that take real actions?

7•thesvp•2d ago•18 comments

Ask HN: Have top AI research institutions just given up on the idea of safety?

79•DietaryNonsense•1d ago•88 comments

Comparing manual vs. AI requirements gathering: 2 sentences vs. 127-point spec

7•thesssaism•2d ago•6 comments

ChatGPT finds an error in Terence Tao's math research

42•codexon•3d ago•8 comments

Ask HN: Who has seen productivity increases from AI

11•Kapura•2d ago•12 comments

What podcasts are you listening to?

42•thomk•4d ago•71 comments

Ask HN: What Linux Would Be a Good Transition from Windows 11

13•Cyberis•3d ago•20 comments

Ask HN: Any DIY open-source Alexa/Google alternatives?

8•personality0•2d ago•4 comments

Persistent Prompts and Built in Search

2•lazybusy•1d ago•0 comments

Ask HN: Is it better to have no Agent.md than a bad one?

9•parvardegr•3d ago•9 comments

Ask HN: Is it worth avoiding AI while making a game?

6•2muchclout•2d ago•22 comments
Open in hackernews

Comparing manual vs. AI requirements gathering: 2 sentences vs. 127-point spec

2•thesssaism•13h ago
We took a vague 2-sentence client request for a "Team Productivity Dashboard" and ran it through two different discovery processes: a traditional human analyst approach vs an AI-driven interrogation workflow.

The results were uncomfortable. The human produced a polite paragraph summarizing the "happy path." The AI produced a 127-point technical specification that highlighted every edge case, security flaw, and missing feature we usually forget until Week 8.

Here is the breakdown of the experiment and why I think "scope creep" is mostly just discovery failure.

The Problem: The "Assumption Blind Spot"

We’ve all lived through the "Week 8 Crisis." You’re 75% through a 12-week build, and suddenly the client asks, "Where is the admin panel to manage users?" The dev team assumed it was out of scope; the client assumed it was implied because "all apps have logins."

Humans have high context. When we hear "dashboard," we assume standard auth, standard errors, and standard scale. We don't write it down because it feels pedantic.

AI has zero context. It doesn't know that "auth" is implied. It doesn't know that we don't care about rate limiting for a prototype. So it asks.

The Experiment

We fed the same input to a senior human analyst and an LLM workflow acting as a technical interrogator.

Input: "We need a dashboard to track team productivity. It should pull data from Jira and GitHub and show us who is blocking who."

Path A: Human Analyst Output: ~5 bullet points. Focused on the UI and the "business value." Assumed: Standard Jira/GitHub APIs, single tenant, standard security. Result: A clean, readable, but technically hollow summary.

Path B: AI Interrogator Output: 127 distinct technical requirements. Focused on: Failure states, data governance, and edge cases. Result: A massive, boring, but exhaustive document.

The Results

The volume difference (5 vs 127) is striking, but the content difference is what matters. The AI explicitly defined requirements that the human completely "blind spotted":

- Granular RBAC: "What happens if a junior dev tries to delete a repo link?" - API Rate Limits: "How do we handle 429 errors from GitHub during a sync?" - Data Retention: "Do we store the Jira tickets indefinitely? Is there a purge policy?" - Empty States: "What does the dashboard look like for a new user with 0 tickets?"

The human spec implied these were "implementation details." The AI treated them as requirements. In my experience, treating RBAC as an implementation detail is exactly why projects go over budget.

Trade-offs and Limitations

To be fair, reading a 127-point spec is miserable. There is a serious signal-to-noise problem here.

- Bloat: The AI can be overly rigid. It suggested microservices architecture for what should be a monolith. It hallucinated complexity where none existed. - Paralysis: Handing a developer a 127-point list for a prototype is a great way to kill morale. - Filtering: You still need a human to look at the list and say, "We don't need multi-tenancy yet, delete points 45-60."

However, I'd rather delete 20 unnecessary points at the start of a project than discover 20 missing requirements two weeks before launch.

Discussion

This experiment made me realize that our hatred of writing specs—and our reliance on "implied" context—is a major source of technical debt. The AI is useful not because it's smart, but because it's pedantic enough to ask the questions we think are too obvious to ask.

I’m curious how others handle this "implied requirements" problem:

1. Do you have a checklist for things like RBAC/Auth/Rate Limits that you reuse? 2. Is a 100+ point spec actually helpful, or does it just front-load the arguments? 3. How do you filter the "AI noise" from the critical missing specs?

If anyone wants to see the specific prompts we used to trigger this "interrogator" mode, happy to share in the comments.

Comments

skydhash•6h ago
The AI approach smells of waterfall and that has been deemed a bad approach to software projects. Not only because of the time cost (which some would argue it can be reduced by agents), but because the only real product is the one in use.

Every product solve some needs, and there are risks associated to the implementations of features. Trying to map every risk is not productive as the feature is not yet proven. Experience is needed to map the path ahead, but only if one has chosen to walk it.

So writing specs is not hated. It’s just that they’re not that different than a SF novels, as their foundation are just hypotheses. If you have a mountain to climb, you do it step by step, not by tracing paths with maps drawn from reports narrating hearsay.

So the most productive is to have an overall roadmap, plan the next months a bit according to your resources, and then fully focus on the next few tasks. Do periodic reviews to assess your progress.