sometimes it even misses things in CLAUDE.md
(Full disclosure, they use and talk about the agent I'm building)
It's just the herd following the herd...
It's able to identify an ai-simulated frustrated user nine times out of 10.
If these are routine, in what kind of state is the repository? All of those easily can and should’ve been done properly at write/merge time in any half-decent code base. Sure, sometimes one case slips by, but if these are routine fixes, there is something deeply wrong with the process.
> I can't write a function anymore without thinking about whether I'm teaching the system or just solving today's problem.
This isn’t a positive thing.
> When you're done reading this, you'll have the same affliction.
No, not all. What you have described is a deeply broken system which will lead to worse developers and even worse software. I hope your method propagates as little as possible.
> But AI outputs aren't deterministic—a prompt that works once might fail the next time.
> So I have Claude run the test 10 times. When it only identifies frustration in four out of 10 passes, Claude analyzes why it failed the other six times.
All of this is wasteful and insane. You don’t learn anything or understand your own system, you just throw random crap at it a bunch of times and then pick some of it that sticks. No wonder it’s part of your routine to have to fix basic things. It’s bonkers that this is lauded as an improvement over doing things right.
dinvlad•5mo ago
Perhaps I'm overlooking something though, it would be great to see more concrete details of their implementation, rather than this high-level inspirational description..
Torn•5mo ago
CLAUDE.md and working memory only goes so far — it never religiously follows them and does not truly 'learn' from past collaboration. In fact the more you put into a CLAUDE.md the more it seems to degrade
politelemon•5mo ago