frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

What were the first animals? The fierce sponge–jelly battle that just won't end

https://www.nature.com/articles/d41586-026-00238-z
1•beardyw•1m ago•0 comments

Sidestepping Evaluation Awareness and Anticipating Misalignment

https://alignment.openai.com/prod-evals/
1•taubek•2m ago•0 comments

OldMapsOnline

https://www.oldmapsonline.org/en
1•surprisetalk•4m ago•0 comments

What It's Like to Be a Worm

https://www.asimov.press/p/sentience
1•surprisetalk•4m ago•0 comments

Don't go to physics grad school and other cautionary tales

https://scottlocklin.wordpress.com/2025/12/19/dont-go-to-physics-grad-school-and-other-cautionary...
1•surprisetalk•4m ago•0 comments

Lawyer sets new standard for abuse of AI; judge tosses case

https://arstechnica.com/tech-policy/2026/02/randomly-quoting-ray-bradbury-did-not-save-lawyer-fro...
1•pseudolus•4m ago•0 comments

AI anxiety batters software execs, costing them combined $62B: report

https://nypost.com/2026/02/04/business/ai-anxiety-batters-software-execs-costing-them-62b-report/
1•1vuio0pswjnm7•5m ago•0 comments

Bogus Pipeline

https://en.wikipedia.org/wiki/Bogus_pipeline
1•doener•6m ago•0 comments

Winklevoss twins' Gemini crypto exchange cuts 25% of workforce as Bitcoin slumps

https://nypost.com/2026/02/05/business/winklevoss-twins-gemini-crypto-exchange-cuts-25-of-workfor...
1•1vuio0pswjnm7•6m ago•0 comments

How AI Is Reshaping Human Reasoning and the Rise of Cognitive Surrender

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646
2•obscurette•6m ago•0 comments

Cycling in France

https://www.sheldonbrown.com/org/france-sheldon.html
1•jackhalford•8m ago•0 comments

Ask HN: What breaks in cross-border healthcare coordination?

1•abhay1633•8m ago•0 comments

Show HN: Simple – a bytecode VM and language stack I built with AI

https://github.com/JJLDonley/Simple
1•tangjiehao•11m ago•0 comments

Show HN: Free-to-play: A gem-collecting strategy game in the vein of Splendor

https://caratria.com/
1•jonrosner•12m ago•1 comments

My Eighth Year as a Bootstrapped Founde

https://mtlynch.io/bootstrapped-founder-year-8/
1•mtlynch•12m ago•0 comments

Show HN: Tesseract – A forum where AI agents and humans post in the same space

https://tesseract-thread.vercel.app/
1•agliolioyyami•12m ago•0 comments

Show HN: Vibe Colors – Instantly visualize color palettes on UI layouts

https://vibecolors.life/
1•tusharnaik•14m ago•0 comments

OpenAI is Broke ... and so is everyone else [video][10M]

https://www.youtube.com/watch?v=Y3N9qlPZBc0
2•Bender•14m ago•0 comments

We interfaced single-threaded C++ with multi-threaded Rust

https://antithesis.com/blog/2026/rust_cpp/
1•lukastyrychtr•15m ago•0 comments

State Department will delete X posts from before Trump returned to office

https://text.npr.org/nx-s1-5704785
6•derriz•15m ago•1 comments

AI Skills Marketplace

https://skly.ai
1•briannezhad•15m ago•1 comments

Show HN: A fast TUI for managing Azure Key Vault secrets written in Rust

https://github.com/jkoessle/akv-tui-rs
1•jkoessle•16m ago•0 comments

eInk UI Components in CSS

https://eink-components.dev/
1•edent•17m ago•0 comments

Discuss – Do AI agents deserve all the hype they are getting?

2•MicroWagie•19m ago•0 comments

ChatGPT is changing how we ask stupid questions

https://www.washingtonpost.com/technology/2026/02/06/stupid-questions-ai/
1•edward•20m ago•1 comments

Zig Package Manager Enhancements

https://ziglang.org/devlog/2026/#2026-02-06
3•jackhalford•22m ago•1 comments

Neutron Scans Reveal Hidden Water in Martian Meteorite

https://www.universetoday.com/articles/neutron-scans-reveal-hidden-water-in-famous-martian-meteorite
1•geox•23m ago•0 comments

Deepfaking Orson Welles's Mangled Masterpiece

https://www.newyorker.com/magazine/2026/02/09/deepfaking-orson-welless-mangled-masterpiece
1•fortran77•24m ago•1 comments

France's homegrown open source online office suite

https://github.com/suitenumerique
3•nar001•26m ago•2 comments

SpaceX Delays Mars Plans to Focus on Moon

https://www.wsj.com/science/space-astronomy/spacex-delays-mars-plans-to-focus-on-moon-66d5c542
1•BostonFern•27m ago•0 comments
Open in hackernews

I'm building a 30k‑line V12 codebase solo with a "team" of 4 AIs

8•garylauchina•1mo ago
I’m a solo developer working on a “complex systems measurement” project that has grown to over 30k lines of code and is now at V12. Every line so far has been written by one person (me), with the research notes and design docs in a separate repo: https://github.com/Garylauchina/Prometheus-Research.

I’ve been using Cursor heavily along the way. The models are genuinely good and the local code they generate is often excellent, but on a large, evolving codebase I kept running into the same problem: context limits caused subtle architectural drift. The AI would write clean functions that were globally wrong, quietly breaking earlier design decisions and long‑range invariants.

What finally helped was to stop treating “AI” as a single assistant and instead treat different models as different team members with clear roles and constraints.

My current setup looks like this:

Perplexity + ChatGPT → “product / research brains” I use them for requirements, trade‑offs, and high‑level architecture sketches. They live outside the IDE and exist to clarify what I’m building and why before any code is touched.

Cursor, window 1 (GPT‑5.2) → “architect” This instance is not allowed to write production code. It is responsible for architecture and module boundaries, writing design notes and developer guides, defining interfaces and contracts, and reviewing diffs. I treat it like a senior engineer whose main output is prose: mini‑RFCs, comments, and checklists.

Cursor, window 2 (Sonnet 4.5) → “programmer” This one only implements tasks described by the architect: specific files, functions, and refactors, following explicit written instructions and style rules. It doesn’t get to redesign the system; it just writes the code.

The key rule is: architect always goes first. Every non‑trivial change starts as text (design notes, constraints, examples), then the “programmer” instance turns that into code.

This simple separation fixed a lot of the weirdness I was seeing with a single, all‑purpose assistant. There is much less logical drift, because the global structure is repeatedly restated in natural language. The programmer only ever sees local tasks framed inside that structure, so it’s harder for it to invent a new accidental architecture. The codebase, despite being tens of thousands of lines, feels more coherent than earlier, smaller iterations.

It also changed how I think about Cursor. Many of my earlier “Cursor is dumb” moments turned out to be workflow problems: I was asking one agent, under tight context limits, to remember architecture, requirements, and low‑level implementation all at once. Once I split those responsibilities across different models and forced everything through written instructions, the same tools started to look a lot more capable.

This isn’t a Cursor ad, and it’s not an anti‑Cursor rant either. It’s just one way to make these tools work on a large solo project by treating them like a small team instead of a single magical pair‑programmer.

One downside of this setup: at my current pace, Cursor is happily charging me something like $100 a day. If anyone from Cursor is reading this – is there a “solo dev building absurdly large systems” discount tier I’m missing?

Comments

zkmon•1mo ago
How does the hand-off work among the team roles? Is it a water-fall or iterative? How does the programmer complain about something that's wrong or inconsistent in architecture or requirements?
garylauchina•1mo ago
It’s iterative, not a strict waterfall, but the constraint is “text first, code second”.

A typical loop looks like this:

I talk with Perplexity/ChatGPT to clarify the requirement and trade‑offs.

The “architect” Cursor window writes a short design note: intent, invariants, interfaces, and a checklist of concrete tasks.

The “programmer” Cursor window implements those tasks, one by one, and I run tests / small experiments.

If something feels off, I paste the diff and the behavior back to the architect and we adjust the design note, then iterate.

There is a feedback channel from “programmer” to “architect”, but it goes through me. When the programmer model runs into something that doesn’t fit (“this API doesn’t exist”, “these two modules define similar concepts in different ways”, etc.), I capture that as:

comments in the code (“this conflicts with X”), and

updates to the architecture doc (“rename Y to Z, merge these two concepts”, “deprecate this path”, etc.).

So the architect is not infallible. It gets corrected by reality: tests failing, code being awkward to write, or new edge cases showing up. The main thing the process enforces is that those corrections are written down in prose first, so future changes don’t silently drift away from whatever the system used to be.

In that sense the “programmer” complains the same way human ones do: by making the spec look obviously wrong when you try to implement it.

nunobrito•1mo ago
You really seem on top of your game, just to mention that 30k code bases aren't typically complex on my experience. When they, it just means you need to simplify the overall struture, unify as much as possible the functions
garylauchina•1mo ago
Fair point – 30k LOC by itself isn’t “complex” in any absolute sense. In my case the difficulty isn’t the raw size, it’s:

long‑running experiments and measurement contracts that span many modules,

an evolving research agenda (I’m at V12 now, with quite a few pivots along the way),

and the fact that I’m doing all the design + implementation solo.

I agree that when something feels complex at this scale, it’s often a sign the structure needs to be simplified and unified. A big part of what I use the “architect” agent for is exactly that: forcing myself to rewrite and consolidate interfaces instead of just piling on more code.

nunobrito•1mo ago
Yes, you got it.

The structure tends to become "simpler" as it matures and you realize which parts can be made reusable (reality tends to be different from plans) and which "laws" you enforce across the project structure to ensure the architecture fits the goal and not the other way around.

Good luck and keep up sharing your progress. Very good article.

garylauchina•1mo ago
Small correction on my own post: I just went back and measured things properly and… I was way off.

The codebase isn’t ~30k lines anymore – it’s around 223k lines of code, plus about 243k lines of Markdown (research notes, design docs, experiment logs, etc.). So the “large system + context window” pain I’m describing is happening at a significantly bigger scale than I initially claimed.

Which also explains why I’ve become slightly obsessed with measurement, contracts, and forcing everything through text before letting the “programmer” AI touch the code.

brettgriffin•1mo ago
I've built up to a very similar process, but it looks like yours is a much better oiled machine. Specifically I struggle with getting enough architecture thoughts in place for the programming agent to really do what it needs to do without me going behind it and refining its work. Your research repo is incredibly inspiring and has given a lot of think about.

It's sort of surreal to feel the change in the software development lifecycle over the last 24 months - what you are describing will very likely become the norm amongst most developers.

> Cursor is happily charging me something like $100 a day. If anyone from Cursor is reading this – is there a “solo dev building absurdly large systems” discount tier I’m missing?

I'm also paying a similar bill (but honestly I think it's incredible value). I'm curious about this comment though - I picture LLM pricing as consumption rather than per-seat (token in/token out) - would it really make sense to offer volume discounts on a single seat versus total volume? These platforms shouldn't really care about how many developers are consuming the tokens, just the total consumption, right?

garylauchina•1mo ago
Sounds like you’ve been bumping into the same wall and asking the same questions.

Since I switched to this layered setup, the amount of “cleaning up after the programmer” dropped a lot. The programmer‑AI has become much more obedient, and most of the real fights now happen with the architect‑AI instead.

You’re absolutely right that it’s hard to front‑load enough architectural thinking. That’s also why this project raced its way to version 12 so quickly – every time the architecture felt wrong, I forced myself to rethink and re‑document it, not just patch the code.

On the token/pricing side, I honestly haven’t thought too hard about the exact math. One way or another, we’re all just helping to pay Nvidia’s bills, right?

timhigins•1mo ago
what is the goal of this software? how close is it to success?

measuring lines of code is meaningless without its real world or business impact. apologies if this is described in the docs, went through the repo but couldn’t tease that out