This is not a good way to do anything. The models are sychophantic, all you need to do in order to get them to agree with you is keep prompting: https://www.sfgate.com/tech/article/calif-teen-chatgpt-drug-...
At least you complied with the next sentence :)
EDIT: Whoa, I didn't check your link before I posted. That's terribly sad. While I agree that LLMs can be sycophantic, I don't think Yegge was debating with Claude about drug use in this situation. Other references might have worked better to support you claim like this first page result when I search for "papers on llm sycophancy": https://pmc.ncbi.nlm.nih.gov/articles/PMC12592531/
We already went over how Stack Overflow was in decline before LLMs.
SaaS is not about build vs. buy, it's about having someone else babysit it for you. Before LLMs, if you wanted shitty software for cheap, you could try hiring a cheap freelancer on Fiverr or something. Paying for LLM tokens instead of giving it to someone in a developing country doesn't really change anything. PagerDuty's value isn't that it has an API that will call someone if there's an error, you could write a proof of concept of that by hand in any web framework in a day. The point is that PagerDuty is up even if your service isn't. You're paying for maintenance and whatever SLA you negotiate.
Steve Yegge's detachment from reality is sad to watch.
This is why I think Ruby is such a great language for LLMs. Yeah, it's token-efficient, but that's not my point [0]. The DWIM/TIMTOWTDI [1] culture of Ruby libraries is incredible for LLMs. And LLMs help to compound exactly that.
For example, I recently published a library, RatatuiRuby [2], that feeds event objects to your application. It includes predicates like `event.a?` for the "a" key, and `event.enter?` for the Enter key. When I was implementing using the library, I saw the LLM try `event.tilde?`, which didn't exist. So... I added it! And dozens more [3]. It's great for humans and LLMs, because the friction of using it just disappears.
EDIT: I see that this was his later point exactly! FTA:
> What I did was make their hallucinations real, over and over, by implementing whatever I saw the agents trying to do [...]
[0]: Incidentally, Matz's static typing design, RBS, keeps it even more token-efficient as it adds type annotations. The types are in different files than the source code, which means they don't have to be loaded into context. Instead, only static analysis errors get added to context, which saves a lot of tokens compared to inline static types.
[1]: Do What I Mean / There Is More Than One Way To Do It
[2]: https://www.ratatui-ruby.dev
[3]: https://git.sr.ht/~kerrick/ratatui_ruby/commit/1eebe98063080...
We are entering the absurd phase where we are beginning to turn all of earth into paperclips.
All software is gonna be agents orchestrating agents?
Oh how I wish I would have learned a useful skill.
That sounds pretty hyperbolic. Everyone? Next “wave”?
He needs an editor, I’m sure he can afford one.
I look forward to him confronting his existence as he gets to be as old as his neighbor. It will be a fun spectacle. He can tell us all about how he was right all along as to the meaning of life. For decades, no less.
I've used these tools on-and-off an awful lot, and I decided last month to entirely stop using LLMs for programming (my one exception is if I'm stuck on a problem longer than 2-3 hours). I think there is little cost to not getting acquainted with these tools, but there is a heavy cognitive cost to offloading critical thinking work that I'm not willing to pay yet. Writing a design document is usually just a small part of the work. I tend to prototype and work within the code as a living document, and LLMs separate me from incurring the cost of incorrect decisions fully.
I will continue to use LLMs for my weird interests. I still use them to engage on spiritual questions since they just act as mirrors on my own thinking and there is no right answer (my side project this past year was looking through the Christian Gospels and some of the Nag Hammadi collection from a mystical / non-dual lens).
Too many people are running a LLM or Opus in a code cycle or new set of Markdown specs (sorry Agents) and getting some cool results and then writing thought-pieces on what is happening to tech.. its just silly and far to immediate news cycle driven (moltbot, gastown etc really?)
Reminds me of how current news cycle in politics has devolved into hour by hour introspection and no long view or clear headed analyis -we lose attention before we even digest that last story - oh the nurse had a gun, no he spit at ICE, masks on ICE, look at this new angle on the shooting etc.. just endless tweet level thoughts turned into youtube videos and 'in-depth' but shallow thought-pieces..
its impossible to separate the hype from baseline chatter let alone what the real innovation cycle is and where it is really heading.
Sadly this has more momentum then the actual tech trends and serves to guide them chaotically in terms of business decisions -then when confused C suite leaders who follow the hype make stupid decisons we blame them..all while pushing their own stock picks...
Don't get me started on the secondary Linkedin posts that come out of these cycles - I hate the low barrier to entry in connected media sometimes.. it feels like we need to go back to newspapers and print magazines. </end rant>
My other thought, that I can't articulate that well is....what about testing? Sure LLMs can generate tons of code but so what? If your two sentence prompt is for a tiny feature that's one thing. If you ask Claude to "build me a todo system" the results will likely rapidly diverge from what you're expecting. The specification for the system is the code, right? I just don't see how this can scale.
Even if 10x cheaper, internally built Saas tools don't come with service level agreements, a vendor to blame/cancel if it goes wrong or a built-in defense of "But we picked the Gartner top quadrant tool".
Also, is this even true? The author's only evidence was to link to a book about vibe coding. I'd be interested to hear anecdotes of companies who are even attempting this.
Edit: wow, and he's a co-author of that book. This guy really just said "source: me"
It's far more challenging to win the 'build' argument on a cost savings approach, because even the least-savvy CIO/CTO understands that the the price of the vendor software is a proof point grounded in the difficulty for other firms to build these capabilities themselves. If there's merit to these claims, the first evidence we'll see is certain domains of enterprise software (like everything Atlassian does) getting more and more crowded, and less and less expensive, as the difficulty of competing with a tier-1 software provider drops and small shops spring up to challenge the incumbents.
Now his bloviated blogposts only speak of a man extremely high on his own supply. Long, pointless, meandering, self-aggrandising. It really is easier to dump this dump into an LLM to try to summarize it than spend time trying to understand what he means.
And he means very little.
The gist: I am great and amazing and predicted the inevitable orchestration of agents. I also call the hundreds of thousands of lines of extremely low quality AI slop "I spent the last year programming". Also here are some impressive sounding terms that I pretend I didn't pull out of my ass to sound like I am a great philosopher with a lot of untapped knowledge. Read my book. Participate in my meme coin pump and dump schemes. The future is futuring now and in the future.
Convinced an LLM to agree with you? What a feat!
Yegge's latest posts are not exactly half AI slop - half marketing same (for Beads and co), but close enough.
As a technical person today, I wouldn't pay a $10/month SaaS subscription if I can login to my VPS and tell claude to install [alternate free software] self-hosted on it. The thing is, everyone is going to have access to this in a few years (if nothing else it will be through the next generation of ChatGPT/Claude artifacts), and the free options are going to get much better to fit any needs common enough to have a significant market size.
You probably need another moat like network effects or unique content to actually survive.
A quick look at gastown makes me think we all are.
LOLWUT?
Counter-factual much?
1. Paying money for the software or access to it.
2. Allowing a fraction of the attention to be siphoned off and sold to advertisers while they use the software.
I don't think advertisers want to pay much for the "mindshare" of mindless bots. And I'm not sure that agents have wallets they can use to pony up cash with. Hopefully someone will figure out a business model here, but Yegge's article certainly doesn't posit one.
yodon•1d ago
I'd recommend starting with Stratechery's articles on on Platforms and Aggregators[0], and a semester long course on Porter's Five Forces[1].
[0]https://stratechery.com/2019/shopify-and-the-power-of-platfo...
[1]https://en.wikipedia.org/wiki/Porter%27s_five_forces_analysi...