Of course, stack-overflow has every incentive to push this narrative, which the LLM makers will counter vigorously. The end-consumers (developers) will be the jury.
[1] https://www.gartner.com/en/articles/hype-cycle-for-artificia...
The Gartner hype cycle is astrology for MBAs: it cannot fail, it can only be failed by those lacking faith to see its wisdom.
This disillusionment is evidenced by lower proportion of people seeking to take such injections now.
In Realityland, the COVID vaccine was an overwhelming success.
The problem with LLMs is they treat all context as significant. Even telling it to ignore previous context is just confusing it more. So Cursor gets hung up on dumb shit like insignificant linter errors or insisting that extracting a block of code to a function will fix a bug (what?). You add more cursor rules to keep it from doing these things, which doesn't focus it, it just pollutes the context more. It doesn't know when to ignore files it's already read. It just gets shittier and shittier.
I appreciate that I can tell someone where the problem is and they'll know what is or isn't irrelevant.
Edit: Here is something Claude Sonnet 4 told me today. I would expect this if I was talking to it for a while, not the second request.
"Used excludeSwitches (plural) instead of excludeSwitches"
And the discourse around all this is also the same: the detractors point out these flaws, and the proponents chime in with "Yeah well humans suck too" and around and around we go...
If you use AI to write the code, did you ever make it to the 1x level to begin with? How can you be 2x smarter to debug it if you didn’t reach the 1x level initially?
Apparently the solution will be 10x prompters.
press X to doubt
"No, the developers do that."
"So you must take the answers to the person with the question?"
"Well, no... my LLM does that!"
"...what would you say you DO here?"
"I'm an answer website! What don't you people understand?!"
https://en.wikipedia.org/wiki/No_Silver_Bullet
Coding is 10% or 20% of the work, LLMs don’t radically improve throughput of software organizations
https://www.reddit.com/r/ExperiencedDevs/comments/1lwk503/st...
I see working with a LLM is like pair programming, if anything I end up wit better quality in the end because they see things in my blind spots, but there is no radical speed up.
I bet that is surprising, at least if you're astonishingly naive and believe everything a company tells you when they're trying to sell you something.
Survey of current and ex-Stack Overflow users shows that they use Stack Overflow.
belter•20h ago
We don't even need human subjects. GitHub contains a natural experiment: Millions of high-quality commits and PRs from before the GenAI explosion (pre-Jan 2021). Compare velocity, complexity, and quality metrics before and after, the data is sitting there, waiting to be analyzed.
Instead, we're drowning in anecdotes while petabytes of empirical evidence remain untouched.
steveBK123•19h ago
Everyone is selling AI or being sold AI.
CJefferson•18h ago
bigtex•18h ago
lelanthran•18h ago
It's much more remarkable, to me, that the AI vendors have been unable to publish a study on the effects of agents on developer velocity.
I mean, if it was really a 10x (or even 2x) boost to productivity, why aren't they seeing it? The returns on each generation have been diminishing compared to the cost. Shouldn't it be the other way around if AI was such a productivity boost?
AIPedant•18h ago
Software developers have had many flame wars over the decades because the lack of data forces the conversation to be anecdotal and ideological:
- should we use dynamic/gradual typing and prioritize developer productivity, or static typing to help enforce correctness?
- agile vs waterfall
- OO vs procedural vs functional
- is Rust's fussiness around memory management more trouble than it's worth for large projects?
- when should you use a 3rd-party library vs doing it yourself?
So this really is nothing new. You are badly underestimating the scientific challenges, instead just hoping big data will plow through. It won't.
belter•18h ago
AIPedant•18h ago