I've wanted to make video games forever. It's fun, and scratches an itch that no other kind of programming does. But making a game is a mountain of work that is almost completely unassailable for an individual in their free time. The sheer volume of assets to be created stops anything from ever being more than a silly little demo. Now, with Gemini 3.1, I can build an asset pipeline that generates an entire game's worth of graphics in minutes, and actually be able to build a game. And the assets are good. With the right prompting and pipeline, Gemini can now easily generate extremely high quality 2d assets with consistent art direction and perfect prompt adherence. It's not about asking AI to make a game for you, it's about enabling an individual to finally be able to realize their vision without having to resort to generic premade asset libraries.
Game development just isn’t something AI can do well. Good games are not just recreations of existing titles.
As with anything else, 95% of it will always be crap. Taste is now the great differentiator.
This is precisely what I'm running into as well. There's a few SaaS solutions that are ok, but I gave up after an attempt at building a pipeline for it. Sticking with building 4X/strategy card games that don't need character animations for now until the models catch up.
I have a news feed, work tab for managing issues/PRs, markdown editor with folders, calendar, AI powered buttons all over the place (I click a button, it does something interesting with Claude code I can't do programmatically).
Why don't I share it? Because it's highly personal, others would find it doesn't fit their own workflow.
Not OP, just guessing.
Before that, it single-shot an app for me where I can copy-paste a table (or a subsection of it) from Excel and print it out perfectly aligned on label sticker paper; it does instantly what used to take me an hour each time, when I had to fight Microsoft Word (mail merge) and my Canon printer's settings to get the text properly aligned on labels, and not cut off because something along the way decided to scale content or add margins or such.
Neither of these tools is immediately usable for others. They're not meant to, and that's fine.
For 20 minutes of time, I had a simple TTS/STT app that allows me to have a voice conversation with my AI assistant.
I vibe'd a basic ticketing system in just under an hour that does what we need. So not 20 mins, but more like 45-60.
I had to create a bunch of GitHub and Linear apps. Without me even asking Codex whipped up a web page and a local server to set them up, collecting the OAuth credentials, and forward them to the actual app.
Took two minutes, I used it to set up the apps in three clicks each, and then just deleted the thing.
Code as transient disposable artifacts.
You can get a throw away app in 5 mins, before I wouldn't even bother.
I'd rather not pay monthly for something (like water) that I'm turning on and off and may not even need for weeks. But paying per-liter is currently more expensive so that's what we currently do.
I think the future is going to be local models running on powerful GPUs that you have on-prem or in your homelab, so you don't need your wallet perpetually tethered to a company just to turn the hose on for a few minutes.
Do you never open a code editor?
Emacs with Hyperbole[0]?
But it requires A LOT of work to make sure it is actually safe for people and organizations. And no, an .md file saying “PLEASE DONT PWN ME, KTHX” isn’t it at all. “Alignment” is only part of the equation.
If you’re not afraid to dive into rabbitholes, here is how it works: http://community.safebots.ai/t/layer-4-browser-extensions-pe...
But based on the hype (100x productivity!), there should be a deluge of high quality mobile apps, Saas offerings, etc. There is a huge profit incentive to create quality software at a low price.
Yet, the majority of new apps and services that I see are all AI ecosystem stuff. Wrappers around LLMs, or tools to use LLMs to create software. But I’m not really seeing the output of this process (net new software).
Oh and sadly, llm’s are useless for the imaginative part too. Shucks eh.
I have a list of ideas a mile long that gets longer every day, and LLMs help me burn through that list significantly faster.
However, the older I get, the more distraught I get that most people I meet "IRL" are simply not sitting on a list of problems they simply lack time to solve. I have... a lot of emotions around this, but it seems to be the norm.
If someone doesn't see or experience problems and intuitively start working out how they would fix them if they only had time, the notion that they could pair program effectively ideas that they didn't previously have with an LLM is absurd.
Problem is that all these companies trying to push AI experiences know that giving users unfettered access to their data to build further customization is corporate suicide.
If someone did make a mobile app, how would it get up take? Coding has never been the hard part about a successful software product.
Because it's better to sell shovels than to pan for gold.
In the current state of LLMs, the average no-experience, non-techy person was never going to make production software with it, let alone actually launch something profitable. Coding was never the hard part in the first place, sales, marketing & growth is.
LLMs are basically just another devtool at this point. In the 90s, IDEs/Rapid App Development was a gold rush. LLMs are today's version of that. Both made developer's life's better, but neither resulted in a huge rush of new, cheap software from the masses.
I do think there would be value in sharing your setup at some point if you get around to it, I think a lot of builders are in the same boat and we're all trying to figure out what the right interface for this is (or at least right for us personally).
a) there are likely many more active, eager contributors all of a sudden, and
b) there's suddenly a huge amount of new papers published every week about algorithms and techniques that said contributors then eagerly implement (usually of dubious benefit).
More cynically, one might also hypothesize that
c) code quality has dropped, so more frequent releases are required to fix broken programs.
And you can even see the number of new games that disclosed using generative AI (~21% in 2025). [2]
And that's probably significantly undercounting because I doubt everyone voluntarily discloses when they use tools like Claude Code (and it's not clear how much Valve cares about code-assistance). [3]
Also no one is buying or playing a lot of these games.
[1] https://steamdb.info/stats/releases/
[2] https://steamdb.info/stats/releases/?tagid=1368160
[3] https://store.steampowered.com/news/group/4145017/view/38624...
Self plug, but basically that’s the TL;DR https://robertdelu.ca/2026/02/02/personal-software-era/
I am just tired boss. I am not going to look at your app.
I really dont know how to respond to these requests. I am going to hide out and not talk to anyone till this fad passes.
Reminds of the trend where everyone was dj wanting you to listen their mixtrack they made on abbleton live
~18 months ago a friend of mine had a very viable, good idea for a physical product, but very fuzzy on the details of where to begin. My skillset backfilled everything he was missing to go from idea to reality to in-market.
I began at arm's length with just advice and validation, then slowly got involved with CAD and prototyping to make sure it kept moving forward, then infrastructure/admin, graphic design, digital marketing and support, etc, while he worked on manufacturing, physical marketing, networking, fulfillment, sales, etc.
Long story short, because I both deeply believe in the vision and know that teamwork makes the dream work, I am fully, completely, inextricably involved LOL -- and I don't have a single complaint about it either, but man, watch out, because if you don't believe in the vision but do have skills/expertise they're lacking, and opt out, friends and family will be the quickest and most aggrieved people you'll ever meet that think you're gatekeeping them from success.
This is true, and I bet there are thousands of people who are in this stage right now - having gotten there far faster than they would have without Claude Code - which makes me predict that the point made in the article will not age well. I think it’s just a matter of a bit more time before the deluge starts, something on the order of six more months.
I launched a vibe coded product a few months ago. I spent the majority of my time
- making sure the copy / presentation was effective on product website
- getting signing certificates (this part SUCKS and is expensive)
- managing release version binaries without a CDN (stupid)
- setting up LLC, website, domain, email, google search indexing, etc, etc
But there's not one tool there that triggered a major boost in output or number of apps / libraries / products created - unless I missed something.
Sure, total output has increased, especially since the early 2010's thanks to both Github becoming the social network of software development, and (arguably) Node / JS becoming one of the most popular languages/runtimes out there attracting a lot of developers to publish a lot of tools. But that's not down to productivity or output boosting developments.
Yup. And for most purposes, that's enough. An app does not have to be productized and shipped to general audience to be useful. In fact, if your goal is to solve some specific problem for yourself, your friends/family, community or your team, then the "last step" you mention - the one that "takes majority of time and effort" - is entirely unnecessary, irrelevant, and a waste of time.
The productivity boost is there, but it's not measured because people are looking for the wrong thing. Products on the market are not solutions to problems, they're tools to make money. The two are correlated, because of bunch of obvious reasons (people need money, solving a problem costs money, people are happy to pay for solutions, etc.), but they're still distinct. AI is dropping the costs of "solving the problem" part, much more than that of "making a product", so it's not useful to use the lack of the latter as evidence of lack of the former.
I've started tons of scratch my own itch projects. There's adoption, UX, onboarding costs even if you're the only audience.
TLDR: i don't even use my own projects. I churn.
To a point, but I think this overstates it by quite a bit. At the moment I'm weighing some tradeoffs around this myself. I'm currently making an app for a niche interest of mine. I have a few acquaintances who would find it useful as well but I'm not sure if I want to take that on. If I keep the project for personal use I can make a lot of simplifying decisions like just running it on my own machine and using the CLI for certain steps.
To deploy this to for non-tech users I need to figure out a whole deployment approach, make the UI more polished, and worry more about bugs and uptime. It sucks to get invested in some software that then constantly starts breaking or crashing. GenAI will help with this somewhat, but certainly won't drop the extra coding time cost down to zero.
That's true, but even the "last step" is being accelerated. The 10% that takes 90% of the time has itself been cut in half.
An example is turning debug logs and bug reports into bugfixes, and performance stats into infrastructure migrations.
The time required to analyze, implement, and deploy those has been reduced by a large amount.
It still needs to be coupled with software engineering skills - to decide between multiple solutions generated by an LLM, but the acceleration is significant.
-0.75 years.
Software development output (features, bugs, products) - especially at smaller companies like startups - has already accelerated significantly, while software development hiring has stayed flat or declined. So there has been a dramatic increase in human-efficiency. To me, that seems like a result, although it's cold comfort as a software engineer.
You probably won't see this reflected as a multiplication of new apps because the app consumer's attention is already completely tapped. There's very little attention surface area left to capture.
Before AI for the last 8 or so years now first at a startup then working in consulting mostly with companies new to AWS or they wanted a new implementation, it’s been:
1. Gather requirements
2. Do the design
3. Present the design and get approval and make sure I didn’t miss anything
4. Do the infrastructure as code to create the architecture and the deployment pipeline
5. Design the schema and write the code
6. Take it through UAT and often go back to #4 or #5
7. Move it into production
8. Monitoring and maintenance.
#4 and #5 can be done easily with AI for most run of the mill enterprise SaaS implementations especially if you have the luxury of starting from the ground up “post AI”. This is something you could farm off to mid level ticket takers before AI.
Just doesn't have the same ring to it.
superpowers/get-shit-done type bloated workflows that try to do everything.
this seems a bit different but still in the same mental category for me
0. It runs way too fast and far ahead. You need to slow it down, force planning only and explicitly present a multi-step (i.e. numbered plan) and say "we'll do #1 first, then do the rest in future steps".
take-away: This is likely solved with experience and changing how I work - or maybe caring less? The problem is the model can produce much faster than you can consume, but it runs down dead ends that destroy YOUR context. I think if you were running a bunch of autonomous agents this would be less noticeable, but impact 1-3 negatively and get very expensive.
1. lots of "just plain wrong" details. You catch this developing or testing because it doesn't work, or you know from experience it's wrong just by looking at it. Or you've already corrected it and need to point out the previous context.
take-away: If you were vibe coding you'd solve all these eventually. Addressing #0 with "MORE AI" would probably help (i.e. AI to play/validate, etc).
2. Serious runtime issues that are not necessarily bugs. Examples: it made a lot of client-side API endpoints public that didn't even need to exist, or at least needed to be scoped to the current auth. It missed basic filtering and SQL clauses that constrained data. It hardcoded important data (but not necessarily secrets) like ports, etc. It made assumptions that worked fine in development but could be big issues in public.
take-away: AI starts to build traps here. Vibe coders are in big trouble because everything works but that's not really the end goal. Problems could range from 3am downtime call-outs to getting your infrastructure owned or data breaches. More serious: experienced devs who go all-in on autonomous coding might be three months from their last manual code review and be in the same position as a vibe coder. You'd need a week or more to onboard and figure out what was going on, and fix it, which is probably too late.
3. It made (at least) one huge architectural mistake (this is a pretty simple project so I'm not sure there's space for more). I saw it coming but kept going in the spirit of my experiment.
take-away: TBD. I'm going to try and use AI to refactor this, but it is non trivial. It could take as long as the initial app did to fix. If you followed the current pro-AI narrative you'd only notice it when your app started to intermittently fail - or you got you cloud provider's bill.
However, PyPi is not really the best way to measure this as the amount of people who take time to wrap their code into a proper package, register into PyPi, push a package, etc... is quite low. Very narrow sampling window.
I do think AI will directly fuel the creation of a lot of personal apps that will not be published anywhere. AI lower the barrier of entry, as we all know, so now regular folks with a bit of technical knowledge can just build the app they want tailored to their needs. I think we´ll see a lot of that.
I see the same with AI. Some companies will use AI quietly and productively without much fuzz. Others are just using it as a marketing tool or an ego trip by execs but no real understanding.
i know maybe this is not to your analysis as its about open source stuff, but this is the sentiment i see with some companies. rather than have 10x output which their clients dont need, they produce things cheaper and earn more money from what they produce. (and later lose that revenue to a breach :p)
Number of iOS apps has exploded since ChatGPT came out, according to Sensor Tower: https://i.imgur.com/TOlazzk.png
Furthermore, most productivity gains will be in private repos, either in a work setting or individuals' personal projects.
When you are developing library it's exact opposite - you really care about how it works and which interface it provides so you end up writing it mostly by hand.
First, I find that I'm using a lot fewer libraries in general because I am less constrained by the mental models imposed by library authors upon what I'm actually trying to do. Libraries are often heavy and by nature abstract low-level calls from API. These days, I'm far more likely to have 2-3 functions that make those low-level calls directly without any conceptual baggage.
Second, I am generalizing but a reasonable assertion can be made that publishing a package is implicitly launching an open source project, however small in scope or audience. Running OSS projects is a) extremely demanding b) a lot of pain for questionable reward. When you put something into the universe you're taking a non-zero amount of responsibility for it, even just reputationally. Maintainers burn out all of the time, and not everyone is signed up for that. I don't think there's going to be anything remotely like a 1:1 Venn for LLM use and package publishing.
I would counter-argue that in most cases, there might already be too many libraries for everything under the sun. Consolidation around the libraries that are genuinely amazing is not a terrible thing.
Third, one of the most recurring sentiments in these sorts of threads is that people are finally able to work through the long lists of ideas they had but would have never otherwise gotten around to. Some of those ideas might have legs as a product or OSS project, but a lot of them are going to be thought experiments or solve problems for the person writing them, and IMO that's a W not an L.
Fourth, once most devs are past the "vibe" party trick phase of LLM adoption, they are less likely to squat out entire projects and far, far more likely to return to doing all of the things that they were doing before; just doing them faster and with less typing up-front.
In other words, don't think project-level. Successful LLM use cases are commit-level.
But people are desperate for data right? Desperate to prove that AI hasn't done shit.
Maybe. But this much is true. If AI keeps improving and if the trendline keeps going, we're not going to need data to prove something equivalent to the ground existing.
It's like looking at tire sales to wonder about where the EV cars are.
Besides, it's working for me. If it isn't working for others I don't want to convince them of anything. I do want to hear from other people for whom it's working, though, so I'm happy to share when things work for me.
I have no plans of publishing them or making the open source, so it will not be a part of this metric. I believe others are doing this too.
I don't blame people for responding to the title instead of the article, because the article itself doesn't bother to answer its own question.
You do realize that "The author means software in general" is already a concession that they don't actually address the question in the title, right?
Yes, you do produce more code. But LoC produced is never a healthy metric. Reviewing the LLM generated code, polishing the result and getting it to production-level quality still very much requires a human-in-the-loop with dedicated time and effort.
On the other hand, people who vibe code and claims to be 10x productive, who produces numerous PRs with large diffs usually bog down the overall productivity of teams by requiring tenuous code reviews.
Some of us are forced to fast-track this review process so as to not slow down these "star developers" which leads to the slow erosion in overall code quality which in my opinion would more than offset the productivity gains from using the AI tools in the first place.
They're in the app stores. Apple's review times are skyrocketing at the moment due to the influx of new apps.
> So, let’s ask again, why? Why is this jump concentrated in software about AI?...Money and hype
The AI field right now is drowning in hype and jumping from one fad to another.Don't get me wrong: there are real productivity gains to be had, but the reality is that building small one-offs and personal tools is not the same thing as building, operationalizing, and maintaining a large system used by paying customers and performing critical business transactions.
A lot of devs are surrendering their critical thinking facilities to coding agents now. This is part of why the hype has to exist: to convince devs, teams, and leaders that they are "falling behind". Hand over more of your attention (and $$$) to the model providers, create the dependency, shut off your critical thinking, and the loop manifests itself.
The providers are no different from doctors pushing OxyContin in this sense; make teams dependent on the product. The more they use the product, the more they build a dependency. Junior and mid-career devs have their growth curves fully stunted and become entirely reliant on the LLM to even perform basic functions. Leaders believe the hype and lay off teams and replace them with agents, mistaking speed for velocity. The more slop a team codes with AI, the more they become reliant on AI to maintain the codebase because now no one understands it. What do you do now? Double down; more AI! Of course, the answer is an AI code reviewer!. Nothing that more tokens can't solve.
I work with a team that is heavily, heavily using AI and I'm building much of the supporting infrastructure to make this work. But what's clear is that while there are productivity gains to be had, a lot of it is also just hype to keep the $$$ flowing.
The tools for generating AI code aren't yet capable of producing code that is decent enough for general purpose use cases, with good robust tests, and clean and quality.
Measuring python packages to indicate AI-based production is like measuring saw production to measure the effectiveness of the steam engine. You need to look at houses and communities being built, not the tools.
But since late last year even when it’s not part of the requirements leading app dev + cloud consulting projects, I’ll throw in a feature complete internal web admin site to manage everything for a project with a UI that looks like something I would have done 25 years ago with a decent UX.
They are completely vibe coded, authenticated with Amazon Cognito and the only things I verify are that unauthenticated users can’t access endpoints, the permissions of the lambda hosting environment (IAM role) and the database user it’s using permissions.
Only at most 5 people will ever use the website at a time - but yeah I get scalability for free (not that it matters) because it’s hosted on Lambda. (yes with IAC)
The website would not exist at all if it weren’t for AI.
Now just to be clear, if a website is meant for real people and the customer’s customers. I’ll insist on a real web designer and a real web developer be assigned to the project with me.
I’ve done some experiments with reading gedcom files, and I think I’m quite close to a demoable version of a genealogy app.
Biggest thing is a tool for remotely working musicians. It’s about 10000 lines of well written rust, it is a demoable state and I wish I could work more on it but I just started a new job.
But yeah, this wouldn’t have been possible if I hadn’t been a very experienced dev who knows how to get things live. Also I’ve found a way to work with LLMs that works for me, I can quickly steer the process in the right way and I understand the code thats written, again it’s possible that a lot of real experience is needed for this.
- The 80/20 rule still applies. We’ve optimized the 20% of time part (a lot!) but all the hype is only including the 80% of work part. It looks amazing and is, but you can’t escape the reality of ~80% of the time is still needed on non-trivial projects.
- Breathless AI CEO hype because they need money. This stuff costs a lot. This has passed on to run of the mill CEOs that want to feel ahead of things and smart.
- You should be shipping faster in many cases. Lots of hype but there is real value especially in automating lots of communication and organization tasks.
Except none of them are open source so they don't show up in this article's metrics.
But it's fine. Keep your head in the sand. It doesn't change the once in a lifetime shift we are currently experiencing.
Im not saying that AI is bad, infact, its the opposite, its one of the most important tools that I have seen introduced in my lifetime. Its like a calculator. Its not going to turn everyone into a mathematician, but it will turn those who have an understanding of math into faster mathematician.
Personally, I see the paid or adware software market shrinking, not growing, as a testament to the success of LLMs in coding.
There are many small, different, and one-time tasks that don’t fit full blown apps. Which I would characterize an AI building a novel app as building a house out of random bits of lumber. It will work but will have no cohesive process and sounds like a nightmare.
Plus you all have LLMs at home. I have my version that takes care of exactly my needs and you can have yours.
Same with vibe-coded stuff.
As far as totally new products - I built one (Habit.am - wordless journaling for mental health) and new products require new habits, people trying new things, its not that easy to change people's behavior. It would be much easier for me to sell my little app if it was a literal plain old journal.
Apparently new iOS app submissions jumped by 24% last year:
> According to Appfigures Explorer, Apple's App Store saw 557K new app submissions in 2025, a whopping 24% increase from 2024, and the first meaningful increase since 2016's all-time high of 1M apps.
The chart shows stagnant new iOS app submissions until AI.
Here's a month by month bar chart from 2019 to Feb 2026: https://www.statista.com/statistics/1020964/apple-app-store-...
Also, if you hang out in places with borderline technical people, they might do things like vibe-code a waybar app and proudly post it to r/omarchy which was the first time they ever installed linux in their life.
Though I'd be super surprised if average activity didn't pick up big on Github in general. And if it hasn't, it's only because we overestimate how fast people develop new workflows. Just by going by my own increase in software output and the projects I've taken on over the last couple months.
Finally, December 2025 (Opus 4.5 and that new Codex one) was a big inflection point where AI was suddenly good enough to do all sorts of things for me without hand-holding.
curl 'https://cdn.statcdn.com/Statistic/1020000/1020964-blank-754.png' \
-H 'Origin: https://www.statista.com' \
-H 'Referer: https://www.statista.com/' \
--output chart.png
Assuming it's a real chart, that will give you the image with the uptick in the last year.Really the one thing that conclusively has changed is that the 'ask it on stackoverflow' has become 'ask it an LLM'. Around 95% of the stackoverflow questions can be answered by an LLM with access to the documentation, not sure what will happen to the other 5%. I don't think stackoverflow will survive a 20-fold reduction in size, if only because their stance on not allowing repeat questions means that exponential growth was the main thing preventing them from becoming stale.
Right.
I don't think you even need cynicism or whatever you felt you were having impolite thoughts about:
I'd expect the top mature libraries to be the most resistant to AI tool use for various reasons. They already have established processes, they don't accept drive-by PR spam, the developers working on them might be the least likely to be early adopters, and -- perhaps most importantly -- the todo list of those projects might need the most human comms, like directional planning rather than the sort of yolo feature impl you can do in a one-man greenfield.
All to further bury signals you might find elsewhere in broader ecosystems.
Also using PyPI as a benchmark is incredibly myopic. Github's 2025 Octoverse[0] is more informative. In that report, you can see a clear inflection point in total users[1] and total open source contributions[2].
The report also notes:
> In 2025, 81.5% of contributions happened in private repositories, while 63% of all repositories were public
[0]: https://github.blog/news-insights/octoverse/octoverse-a-new-...
[1]: https://github.blog/wp-content/uploads/2025/10/octoverse-202...
[2]: https://github.blog/wp-content/uploads/2025/10/octoverse-202...
Pre-ChatGPT, in ~2020, there were about 5,000 new packages per month. Starting in 2025 (the actual year agents took off), there is a clear uptick in packages that is consistently about 10,000 or 2X the pre-ChatGPT era.
In general, the rate of increase is on a clear exponential. So while we might not see a step change in productivity, there comes a point where the average developer is in fact 10X productive than before. It just doesn't feel so crazy because it can about in discrete 5% boosts.
I also disagree with the dataset being a good indicator of productivity. I wouldn't actually suspect the number of packages or the frequency of updates to track closely with productivity. My first order guess would that AI would actually be deflationary. Why spend the time to open source something that AI can gen up for anyone on a case by case basis specific to the project. it takes a certain level of dedication and passion for a person to open source a project and if the AI just made it for them, then they haven't actually made the investment of their time and effort to make them feel justified in publishing the package.
The metrics I would expect to go up are actually the size of codebases, the number of forks of projects that create hyper customized versions of tools and libraries, and other metrics like that.
Overall, I'd predict AI is deflationary on the number of products that exist. If AI removes the friction involved with just making a custom solution, then the amount of demand for middleman software should actually fall as products vertically integrate and reduce dependencies.
Like I've been making things, and making changes to things, but I haven't published any of that because, well they're pretty specific to my needs. There are also things which I won't consider publishing for now, even if generally useful because, well the moat has moved from execution effort to ideas, and we all want to maintain some kind of moat to boost our market value (while there's still one). Everyone has reasonable access to the same capabilities now, so everyone can reasonably make what they need according to their exact specs easily, quickly and cheaply.
So while there are many things being made with AI, there is ever-decreasing reasons to publish most of it. We're in an era of highly personalized software, which just isn't worth generalizing and sharing as the effort is now greater than creating from scratch or modifying something already close enough.
There's a very real problem of low effort AI slop, but throwing out the baby with the bathwater is not the solution.
That said, I do kind of wonder if the old model of open source just isn't very good in the AI era. Maybe when AI gets a lot better, but for now it does take real human effort to review and test. If contributors were reviewing and testing like they should be doing, it wouldn't be an issue, but far too many people just run AI and don't even look at it before sending the PR. It's not the maintainers job to do all the review and test of a low-effort push. That's not fair to them, and even discarding that it's a terrible model for software that you share with anyone else.
Having your code snatched and its copyright disregarded, to the benefit of some rando LLM vendor. People can just press "pause" and wait until they see whether they fuel something that brings joy to the world. (Which it might in the end. Or not.)
Yep, also a huge factor. Why publish something you built with an AI assistant if you know it's going to be immediately dunked on not because the quality may be questionable, but because someone sees an em-dash, or an AI coauthor, and immediately goes on a warpath? Heck I commented[0] on the attitude just a few hours ago. I find it really irritating.
[0] https://github.com/duriantaco/fyn/issues/4#issuecomment-4117...
Of course these are specific workplaces designed around moving tickets on a board, not high-agentic, fast-moving startups or independent projects—but they might represent a lot of the developer workforce.
I also know this is not everyone's experience and probably a rare favorable outcome of productivity gain captured by a worker that is not and won't stay the norm.
the real growth is in apps that use ai as a feature, not ai-first packages. like every saas just quietly added an llm call somewhere in their stack. thats hard to measure from dependency graphs.
While it’s interesting to see that in open source software the increase is not dramatic, this ignores however many people are now gen-coding software they will never publish just for them, or which winds up on hosting platforms like Replit.
In the last few months, Gemini (and I) have written for highly personal, very niche apps that are perfect for my needs, but I would never dream of releasing. Things like cataloguing and searching my departed mom‘s recipe cards, or a text message based budget tracker for my wife and I to share.
These things would never be released or available as of source or commercial applications in the way that I wanted them, and it took me less time to have them built with AI then it would have taken me to Research existing alternatives and adapt my workflow/use case to fit whatever I found.
So yeah, there are more apps but I would venture to say you’ll never see most of them…
But that's not really what we were promised.
"THE APPLE APP STORE IS DROWNING IN AI SLOP" https://x.com/shiri_shh/status/2036307020396241228
superkuh•1h ago
It's a great change for a human person. I'm not pretending I'm making something other people would buy nor do I want to. That's the point.