frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

US to paint southern border wall black to make it hotter

https://english.elpais.com/usa/2025-08-20/us-to-paint-southern-border-wall-black-to-make-it-hotter-and-harder-to-climb.html
1•geox•1m ago•0 comments

Las Vegas leaders are banking on trees for cooling: The science is complicated

https://azdailysun.com/news/state-and-regional/las-vegas-leaders-are-banking-on-trees-for-cooling-the-science-is-complicated/article_82bcdb83-af86-4f57-9c23-5db24eb9b6d9.html
1•PaulHoule•1m ago•0 comments

Type Foundry Directory: Language support, licensing models, metrics, logo usage

https://typefoundry.directory
1•gregwolanski•2m ago•0 comments

AI Analyzed My Health Data and Uncovered How I Threw My Back Out

https://akhurana.substack.com/p/how-ai-analyzed-6-months-of-my-health
1•LanceJones•4m ago•0 comments

Scaling domain expertise in complex, regulated domains

https://openai.com/index/blue-j/
1•gmays•4m ago•0 comments

AI in open source: fighting bias in AI analysis

https://rainer.gerhards.net/2025/08/when-humans-and-ais-overthink-a-complex-rsyslog-crash-that-wasnt.html
1•rgerhards•5m ago•0 comments

Physics Grifters: Eric Weinstein, Sabine Hossenfelder a Crisis of Credibility

https://timothynguyen.org/2025/08/21/physics-grifters-eric-weinstein-sabine-hossenfelder-and-a-crisis-of-credibility/
2•Emerson1•6m ago•0 comments

The Simple Vision Breakthrough That Might End Lasik

https://scitechdaily.com/the-shockingly-simple-vision-breakthrough-that-might-end-lasik/
2•bilsbie•7m ago•0 comments

Iterative DFS with stack-based graph traversal

https://dwf.dev/blog/2024/09/23/2024/dfs-iterative-stack-based
1•cpp_frog•10m ago•0 comments

The Great ChatGPT Traffic Miscounting Problem

https://www.rivalsee.com/blog/chatgpt-traffic-miscounting-problem
2•grahac•13m ago•0 comments

Perplexity AI's Motion to Dismiss Dow Jones Lawsuit Is Denied in Full [pdf]

https://storage.courtlistener.com/recap/gov.uscourts.nysd.630270/gov.uscourts.nysd.630270.65.0.pdf
2•pcaharrier•13m ago•1 comments

Claude Code is for more than just code

https://www.ai-supremacy.com/p/claude-code-is-growing-crazy-fast-vibe-coding
3•wordsaboutcode•14m ago•0 comments

Amazon Q Developer for VS Code Vulnerable to Invisible Prompt Injection

https://embracethered.com/blog/posts/2025/amazon-q-developer-interprets-hidden-instructions/
1•spillman•14m ago•0 comments

John Bragg: The Unknown Billionaire Who Controls Half The

https://fs.blog/knowledge-project-podcast/outliers-john-bragg/
1•feross•17m ago•0 comments

Building Realistic Benchmarks for Evaluating Retrieval on Technical Documents

https://arxiv.org/abs/2504.13128
1•fzliu•18m ago•0 comments

Marines managed to get past an AI powered camera "undetected" by hiding in boxes

https://rudevulture.com/marines-managed-to-get-past-an-ai-powered-camera-undetected-thanks-to-hiding-in-boxes/
2•voxadam•20m ago•1 comments

What Are Oklch Colors?

https://jakub.kr/components/oklch-colors
1•pbardea•21m ago•0 comments

"Surprises" in BLS Jobs Revisions Became More Frequent After 2020

https://medium.com/@baogorek/surprises-in-bls-jobs-revisions-became-more-frequent-after-2020-bba2f257fe1a
2•benogorek•23m ago•0 comments

Coding Wedge: Are Developers and Coding Automation Key to LLM Competition?

https://www.decodingdiscontinuity.com/p/the-coding-wedge-gpt-5-openai-orchestration
1•ChrisInFrance•23m ago•0 comments

Israeli military's own data indicates civilian death rate of 83% in Gaza war

https://www.theguardian.com/world/ng-interactive/2025/aug/21/revealed-israeli-militarys-own-data-indicates-civilian-death-rate-of-83-in-gaza-war
7•NomDePlum•23m ago•1 comments

Starlink Courts Luxury Airlines to Grow In-Flight Wi-Fi Business

https://www.bloomberg.com/news/articles/2025-08-21/musk-s-starlink-wi-fi-in-talks-with-emirates-saudia-flydubai
1•JumpCrisscross•26m ago•0 comments

Show HN: DockFlow – Dock presets for instant workflow switching on macOS

1•pugdogdev•26m ago•0 comments

Brace for a crash before the golden age of AI

https://www.ft.com/content/a76f238d-5543-4c01-9419-52aaf352dc23
1•FabHK•26m ago•0 comments

ICE's Official X Account Tags DOJ in Repost of State Rep's Clearly Protected Spe

https://www.techdirt.com/2025/08/20/ices-official-x-account-tags-doj-in-repost-of-state-reps-clearly-protected-speech/
3•taimurkazmi•27m ago•0 comments

Why Be Reactive?

https://crank.js.org/blog/why-be-reactive/
2•bikeshaving•28m ago•1 comments

Claude Opus refuses to answer biotech questions

2•vapemaster•28m ago•0 comments

Poland foiled cyberattack on big city's water supply, deputy PM says

https://www.reuters.com/en/poland-foiled-cyberattack-big-citys-water-supply-deputy-pm-says-2025-08-14
2•acossta•30m ago•1 comments

The Baby Paradox in Haskell

https://blog.jle.im/entry/the-baby-paradox-in-haskell.html
1•jle•31m ago•0 comments

The Extraordinary Rise in the Wealth of Older American Households

https://www.nber.org/papers/w34131
2•bikenaga•31m ago•0 comments

Zoom patches worrying security Windows flaw

https://www.techradar.com/pro/security/zoom-patches-worrying-security-windows-flaw-make-sure-youre-protected-update-now
2•acossta•32m ago•1 comments
Open in hackernews

95% of Companies See 'Zero Return' on $30B Generative AI Spend

https://thedailyadda.com/95-of-companies-see-zero-return-on-30-billion-generative-ai-spend-mit-report-finds/
255•speckx•1h ago

Comments

rogerkirkness•1h ago
AI is a deflationary technology, sort of like how most of the cost of TVs is the cost of shipping them from where they are made to where they are used. So wouldn't the returns show up in 'less work needs to be done' slowly over time?
lazide•1h ago
Except generative AI is mostly an arms race, with content being generated being processed by some other AI on the other side. And any humans unfortunate enough to be in the middle regretting being born.
seydor•1h ago
I'm confused, which point of the hype cycle is this?
jameskilton•1h ago
The Trough of Disillusion. Though we definitely aren't down there yet, it's coming.
m_fayer•57m ago
You just wait, when we achieve agi it’ll drive hype cycles at a speed humans have no chance of keeping up with. We will always effectively be in all points of the hype cycle at once. Money will wash randomly through the economy like soap through your laundry. You ain’t seen nothing yet.
spogbiper•1h ago
Sounds like 95% of companies are potential clients for my consulting services
uncircle•56m ago
Do you sell overpriced generative AI solutions, or do you consult them on how to pivot away from idiotic generative AI?
onlyrealcuzzo•54m ago
Can't tell if you're in on it, but he's implying this waste is no different than average for consulting
uncircle•51m ago
I take offence at comparing my consulting services writing real software by hand like we did in 2021 with generative AI spambots.

(I'm not really offended honestly. Startups will come crying to un-vibe the codebases soon enough.)

ehutch79•54m ago
YES!
theyinwhy•36m ago
So much about can't have the cake and eat it
ehutch79•32m ago
Sell them the cake and then get paid to eat it.
nextworddev•18m ago
Good thing consulting is one thing AI is decent at
AstroBen•1h ago
The link to their referenced study doesn't seem to work?

This is confusing.. it's directly saying AI is improving employee productivity, but that's not leading to more business profit... how does that happen?

macintux•55m ago
Possible conclusion: most of the work that employees do has no direct impact on earnings.
AznHisoka•53m ago
Maybe they’re measuring productivity by flawed metrics. One could write 10x as much code, but that doesnt mean that will equate to more profit
addaon•53m ago
> AI is improving employee productivity, but that's not leading to more business profit... how does that happen?

One trivial way is that the increase of productivity is less than the added cost of the tools. Which suggests that (either due to their own pricing, or just mis-judgement) the AI companies are mis-pricing their tools. If the tool adds $5000 in productivity, it should be priced at $4999, eventually -- the AI companies have every motivation to capture nearly all of the value, but they need to leave something, even if just a penny, for the purchasing company to motivate adoption. If they're pricing at $5001, there's no motivation to use the tool at all; but of course at $4998 they're leaving money on the table. There's no stable equilibrium here where the purchasing companies end up with a /significant/ increase in (productivity - cost of that productivity), of course.

bilbo0s•49m ago
the AI companies are mis-pricing their tools

Sounds like the AI companies are not so much mispricing, as the companies using the tools are simply paying wayyy too much for the privilege.

As long as the companies keep paying, the AI companies are gonna keep the usage charges as high as possible. (Or at least, at a level as profitable to themselves as possible.) It's unreasonable to expect AI companies to unilaterally lower their prices.

philipodonnell•1h ago
Anyone have a link to the actual report?
theonionspeaks•55m ago
https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Bus...
cstejerean•1h ago
Why do I have a feeling that most of that $30B was spent on paying for consultants, most of which were also essentially making things up as they went along.
bilbo0s•53m ago
I don't know?

For some reason, I'm thinking most of the money went to either inferencing costs or NVidia.

wood_spirit•52m ago
Are there really that many consultants floating around still? I remember the heyday of the 2000s and kind of thought the outside consultants were largely disappearing and nowadays companies have trained themselves to jump on the hype train without paying consultants to push them?
varispeed•59m ago
The funniest part isn’t that AI hasn’t delivered profits. It’s that the only “value” most people got from LLMs was accidentally rediscovering what Google used to be before it turned into an ad-riddled casino.

Executives mistook that novelty for a business revolution. After years of degraded search, SEO spam, and “zero-click” answers, suddenly ChatGPT spat out a coherent paragraph and everyone thought: my god, the future is here. No - you just got a glimpse of 2009 Google with autocomplete.

So billions were lit on fire chasing “the sliced bread moment” of finally finding information again - except this time it’s wrapped in stochastic parroting, hallucinations, and a SaaS subscription. The real irony is that most of these AI pilots aren’t “failing to deliver ROI” - they’re faithfully mirroring the mediocrity of the organisations deploying them. Brittle workflows meet brittle models, and everyone acts surprised.

The pitch was always upside-down. These things don’t think, don’t learn, don’t adapt. They remix. At best they’re productivity duct tape for bored middle managers. At worst they’re a trillion-dollar hallucination engine being sold as “strategy.”

The MIT study basically confirms what was obvious: if you expect parrots to run your company, you get birdshite for returns.

aqme28•59m ago
This is why it's so good to sell shovels, so-to-speak.

In this case, that's NVDA

tovej•46m ago
The shovel business is good as long as the gold rush lasts. Once the gold rush is over, you're going to have to deal with a significant decrease in volume, unless you can find other customers.

Crypto's over, gaming isn't a large enough market to fill the hole, the only customers that could fill the demand would be military projects. Considering the arms race with China, and the many military applications of AI, that seems the most likely to me. That's not a pleasant thought, of course.

The alternative is a massive crash of the stock price, and considering the fact that NVIDIA makes up 8% of everyone's favorite index, that's not a very pleasant alternative either.

It seems to me that an ultra-financialized economy has trouble with controlled deceleration, once the hypetrain is on it's full-throttle until you hit a wall.

taormina•39m ago
There aren’t enough GPUs for average gamers to buy anything vaguely recent and they would love to be able to. Making the best GPUs on the planet is still huge and the market is quite large. Scalping might finally die at this rate, but NVDA wasn’t making any of the scalping money anyway so who cares? Data centers and gamers still need every GPU NVDA can make.
tovej•10m ago
Oh there's definitely a market but it's not as big or worth as much the AI market. Gamers don't need GB200s or H100s, and AMD beats Nvidia on price in most segments. Nvidia isn't going to die, but gamers won't fill the demand.

Data centers might, but then they'll need something else to compute, and if AI fails to deliver on the big disruptive promises it seems unlikely that other technologies will fill those shoes.

I'm just saying that something big will have to change, either Nvidias story or share price. And the story is most likely to pivot to military applications.

lgats•41m ago
or steel and wood for making shovels, TSMC
abrinz•57m ago
Skill issue
mushufasa•57m ago
1 in 20 doesn't sound great. But you have to mediate that with

- everyone and their mother are doing a "generative ai program" right now, a lot of times just using the label to try to get their project funded, ai being an afterthought

- if the 1 out of 20 projects is game-changing, then you could argue right now people should actually be willing to spend even more on the opportunity, maybe the number should actually be 1 in 100. (The VC model is about having big success 1 in 10 times.)

- studies of ongoing business activities are inherently methodologically limited by the data available; I don't have a ton of confidence that these researchers' numbers are authoritative -- it's inherently impossible to truly report on internal R&D spend especially a private companies without inside information, and if you have the inside information you likely don't have the full picture.

JCM9•57m ago
We are entering the “Trough of disillusionment.” These hype cycles are very predictable. GPT-5 being panned as a disappointment after endless hype may go down as GenAI’s “jump the shark” moment.

It’s all fun and games until the bean counters start asking for evidence of return on investment. GenAI folks better buckle up. Bumps ahead. The smart folks are already quietly preparing for a shift to ride the next hype wave up while others ride this train to the trough’s bottom.

Cue a bunch of increasingly desperate puff PR trying to show this stuff returns value.

highwaylights•53m ago
I wouldn’t be surprised if 95% of companies knew this was a money pit but felt obligated to burn a pile of money on it so as not to hurt the stock price.
JCM9•49m ago
FOMO on the way up to the peak is a powerful force. Now that we’re sliding down the other side FOMO turns into “WTF did we just spend all that money on again?”
lenerdenator•43m ago
I also wouldn't be surprised if bean counters were expecting a return in an unreasonable amount of time.

"Hey, guys, listen, I know that this just completely torched decades of best practices in your field, but if you can't show me progress in a fiscal year, I have to turn it down." - some MBA somewhere, probably, trying and failing yet again to rub his two brain cells together for the first time since high school.

Just agentic coding is a huge change. Like a years-to-grasp change, and the very nature of the changes that need to be made keep changing.

dingnuts•35m ago
Sam Altman and company have been promising full on AGI. THAT'S the price shock.

Agents may be good (I haven't seen it yet, maybe it's a skill issue but I'm not spending hundreds of dollars to find out and my company seems reluctant to spend thousands to find out) but they are definitely, definitely not general superintelligence like SamA has been promising

at all

really is sinking in

these might be useful tools, yes, but the market was sold science fiction. We have a useful supercharged autocomplete sold as goddamn positronic brains. The commentariat here perhaps understood that (definitely not everyone) but it's no surprise that there's a correction now that GPT-5 isn't literally smarter than 95% of the population when that's how it was being marketed

wredcoll•18m ago
It's real good for stock prices though. Reminds me of tesla.
beepbooptheory•33m ago
"Actually its good we aren't making money, this actually proves how revolutionary the technology is. You really need to think about adapting to our new timeline."

You really set yourself up with a nice glass house trying to make fun of the money guys when you are essentially just moving your own goal posts. It was annoying two (or three?) years ago when we were all talking about replacing doctors and lawyers, now it just cant help but feel like a parody of itself in some small way.

Spivak•17m ago
How dare the business ask for receipts of value being produced in actual dollars! Those idiots don't know anything.
lubesGordi•24m ago
Agreed agentic coding is a huge change. Smart startups will be flying but aren't representative. Big companies won't change because the staff will just spend more time shopping online instead of doing more than what is asked of them. Maybe increased retail spend is a better measure of AI efficacy.
generic92034•34m ago
In Germany there is the additional issue of companies only really starting to invest into the hype when the hype cycle is already at its end, in other parts of the world. And do not imagine that would lower the investments or shorten the amount of time spent on the hype. The C level can never admit errors, the middle management only sees a way to promotion by following the hype.
pgwhalen•31m ago
It's hard to define what it means for a company to know something, but as a person inside a company spending on gen AI efforts, I'm pretty confident that we're not investing in it just to maintain an elevated valuation (we're a mature, privately owned company).
empath75•19m ago
There were similar headlines in the late 80s and early 90s as IT in general was widely seen to have been a money wasting bust. Most people who try to use new technologies, especially early adopters waste a shitload of money and don't accomplish very much.
frozenport•50m ago
Yo what’s the next hype cycle that smart folks like us should be working on?
belter•42m ago
The US mix of government and enterprise. You want Capital you need to be MAGA

"Donald Trump and Silicon Valley's Billionaire Elegy" - https://www.wired.com/story/donald-trump-and-silicon-valleys...

gmd63•37m ago
Smartest people will be working against that. You're thinking of opportunistic people with myopia.
belter•30m ago
The law firms failed to do so.
quotemstr•32m ago
Defense will increasingly become a national priority over the next few decades. Pax Americana is teetering.
deepdarkforest•24m ago
That and consumer robotics. The latter will explode if (big if) RL and llm reasoning get combined into something solid. Lots and lots of smart people are working on it already of course, we are seeing great improvements but nothing really usable. i think we will finally get to a real hype stage in maybe 3-4 years
Terr_•29m ago
The next hype-cycle might not relate to software. I'm thinking of the (smaller, shorter) phase where "nanotechnology" was getting slapped onto everything including laundry detergent.
no_wizard•41m ago
Gemini keeps being rather impressive though, even their iterative updates have improvements, though I'm seeing a significant slowdown in the improvements (both quantity and in how much they improve) suggestion a wall may be approaching.

That said, technologies like this can also go through a rollercoaster pattern itself. Lots of innovation and improvement, followed by very little improvement but lots of research, which then explodes more improvements.

I think LLMs have a better chance at following that pattern than computer vision did when that hype cycle was all the rage

belter•39m ago
Prepare for the crash:

"Spending on AI data centers is so massive that it’s taken a bigger chunk of GDP growth than shopping" - https://fortune.com/2025/08/06/data-center-artificial-intell...

seatac76•24m ago
Good point. I wonder if the Windsurf folks saw the writing on the wall and cashed out when they could.
deadbabe•8m ago
If GPT-5 had been released last year instead, they could have probably kept the hype manageable. But they waited too long and too greedily, and the launch fell flat. And in some cases even negative as I’m seeing some bad PR about people who got too attached to their GPT-4o lovers and hate the new GPT-5.
mattlondon•6m ago
Perhaps just a trough of disillusionment for OpenAI. Anthropic and Google keep delivering.
CharlesW•57m ago
Duplicate discussion from this week (162 comments): https://news.ycombinator.com/item?id=44941118

Here's the source report, not linked to by this content farm's AI-written article: https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Bus...

pawelduda•55m ago
I blame bad prompts /s
eldenring•53m ago
This is how America ends up being ahead of the rest of world with every new technology breakthrough. They spend a lot of money, lose a lot of money, take risks, and then end up being too far for others to catch up.

Trying to claim victory against AI/US Companies this early is a dangerous move.

ta20240528•45m ago
> This is how America ends up being ahead of the rest of world with every new technology breakthrough.

Too young to remember GSM?

eldenring•39m ago
I don't think its controversial to say that in net, the US has had a good streak so far with new technologies. Yeah there have been some pretty big misses, but obviously the wins more than make up for them.
vdupras•42m ago
That must be exactly what Napoleon felt like when he entered Moscow. Unstoppable army, sure riding right into what is obviously hubris, sure having have had a bit of a setback against Kutusov, but man were they on a winning streak!
onlypassingthru•34m ago
Wasn't half of Napoleon's army gone by the time he reached Moscow?[0] It's the sunk cost fallacy writ large with dead bodies, no?

[0]https://www.researchgate.net/figure/Napoleon-march-graphic-C...

vdupras•31m ago
Certainly, and they lost the other half on the way home. On the way to Moscow, there was still something to look forward to. In Moscow, with no supply, dying like flies, all they had to look forward to was a long march through the Almighty Intense Winter, during which they'd be dying like flies too.
stephankoelle•52m ago
In some products, certain AI features have become expected. If a product doesn’t include them, it risks losing customers, making it a net negative for the market. At this point, companies either invest in AI or risk falling behind.
Scarblac•44m ago
Can you name some of those products?
phplovesong•52m ago
No shit. AI is not "AI", its just a meme word for getting VC cash, because the original term (ML) did not raise any capital. Play stupid games, win stupid prizes.
IgorPartola•51m ago
What are the actual use cases that can generate revenue or at least save costs today? I can think of:

1. Generate content to create online influence. This is at this point probably way oversaturated and I think more sophisticated models will not make it better.

2. Replace junior developers with Claude Code or similar. Only sort of works. After all, you can only babysit one of these at a time no matter how senior you are so realistically it will make you, what, 50% more productive?

3. Replace your customer service staff. This may work in the long run but it saves money instead of making money so its impact has a hard ceiling (of spending just the cost of electricity).

4. Assistive tools. Someone to do basic analysis, double check your writing to make it better, generate secondary graphic assets. Can save a bit of money but can’t really make you a ton because you are still the limiting factor.

Aside: I have tried it for editing writing and it works pretty well but only if I have it do minimal actual writing. The more words it adds, the worse the essay. Having it point out awkward phrasing and finding missing parts of a theme is genuinely helpful.

5. AI for characters in video games, robot dogs, etc. Could be a brave new frontier for video games that don’t have such a rigid cause/effect quest based system.

6. AI girlfriends and boyfriends and other NSFW content. Probably a good money maker for a decade or so before authentic human connections swing back as a priority over anxiety over speaking to humans.

What use cases am I missing?

wedn3sday•45m ago
One use case I'd love to see an easy plug-and-play solution for is a RAG build around companies vast internal documentation/wikis/codebase to help developers onboard and find information faster. I would love to see less of people trying to replace humans with language models and more of people trying to use language models to make humans jobs less frustrating.
OutOfHere•15m ago
In all the companies I have worked at and have looked at such docs, unfortunately this doesn't really work because those internal documentation sites are statistically never up to date or even close. They are hilariously unclearly written or out of date.

As for relying on the code base, that is good for code, not for onboarding/deployment/operations/monitoring/troubleshooting that have manual steps.

arahman4710•15m ago
Check out https://asksolo.ai/

We connect with slack/notion/code/etc so that you can do the following:

1. Ask questions about how your code/product works 2. Generate release notes instantly 3. Auto update your documentation when your code changes

We primarily rely on the codebase since it is never out of date

jaimebuelta•44m ago
Stock images. I’ve already seen trining courses (for compliance reasons) using AI videos. A bit cringey, but I imagine cheaper than shooting real people.
wredcoll•14m ago
> but I imagine cheaper than shooting real people

How much does that cost these days? Do you still have to fly to remote islands?

plantain•43m ago
+50% productivity for 200$/mo is outstanding value! Most countries have 0-2% productivity growth per year!
spogbiper•39m ago
I am working on a project that uses LLM to pull certain pieces of information from semi-structured documents and then categorize/file them under the correct account. it's about 95% accurate and we haven't even begun to fine tune it. i expect it will require human in the loop checks for the foreseeable future, but even with a human approval of each item, its going to save the clerical staff hundreds of hours per year. There are a lot of opportunities in automating/semi-automating processes like this, basically just information extraction and categorization tasks.
beepbooptheory•30m ago
How will you know in practice which 5% is wrong?
spogbiper•27m ago
the system presents a summary that a human has to approve, with everything laid out to make that as easy as possible, links to all the sources etc
kjkjadksj•26m ago
Isn’t that something you can do with non ai tooling to 100% accuracy?
spogbiper•18m ago
in some similar cases yes, and this client has tried to accomplish that for literally decades without success. i don't want to be too detailed for reasons, but basically they cannot standardize the input to the point where anything non AI has been able to parse it very well.
systemerror•24m ago
The big issue with LLMs is that they’re usually right — like 90% of the time — but that last 10% is tough to fix. A 10% failure rate might sound small, but at scale, it's significant — especially when it includes false positives. You end up either having to live with some bad results, build something to automatically catch mistakes, or have a person double-check everything if you want to bring that error rate down.
spogbiper•20m ago
yes, the entire design relies on a human to check everything. basically it presents what it thinks should be done, and why. the human then agrees or does not. much work is put into streamlining this but ultimately its still human controlled
wredcoll•14m ago
At the risk of being obvious, this seems set up for failure in the same way expecting a human to catch an automated car's mistakes is. Although I assume mistakes here probably don't matter very much.
whatever1•16m ago
All of the AI projects promise that they just need some fine tuning to go from poc to actual workable product. Nobody was able to fine tune them.

Sorry this is some bull. Either it works or it doesn’t.

LPisGood•8m ago
> its going to save the clerical staff hundreds of hours per year

How many hundreds of hours is your team spending to get there? What is the ROI on this vs investing that money elsewhere?

kingkawn•35m ago
#5 is an enormous use case that when well implemented will permanently replace prescribed character arcs

It is uniquely susceptible because the gaming market is well acclimated to mediocre writing and one dimensional character development that’s tacked on to a software product, so the improvements of making “thinking” improvisational characters can be immense.

Another revenue potential you’ve missed is visual effects, where AI tools allow what were previously labor intensive and expensive projects to be completed in much less time and with less, but not no, human input per frame

shantara•17m ago
>#5 is an enormous use case that when well implemented will permanently replace prescribed character arcs

I mostly disagree. Every gaming AI character demo I've seen so far is just adds more irrelevant filler dialogue between the player and the game they want to play. It's the same problem that some of the older RPG games had, thinking that 4 paragraphs of text is always better than 1.

kingkawn•9m ago
I agree the implementation isn’t great, but mostly it’s because the devs aren’t well versed yet in setting the parameters for the AI’s personality and how rapidly it gets to the point. That’s true of all chatbot AIs out of the box at the moment it seems, but is fixable with an eye to the artistry of the output
b8•23m ago
AI Customer Service is very frustrating to work with as a end user.
siliconc0w•12m ago
The problem is that a lot of friction is intentional. Companies want there to be friction to return an item or cancel a subscription. Insurance companies want there to be friction to evaluate policies or appeal denied claims. Companies create legal friction to make competition harder. The friction is the point so AI isn't a solution. If you were to add AI they'd just find a way to create new friction.
Spivak•8m ago
I mean look, if the customer service department is just trying to frustrate you into not contacting them then AI is just a new tool in the belt for that. Sad, but improvement for them is explicitly worse for you.

But if you're actually trying to provide good customer service because people are paying you for it any paying per case then you wouldn't dare put a phone menu or AI chat bot in-between them and the human. The person handles all the interaction with the client and then uses AI where it's useful to speed up the actual work.

SalmoShalazar•21m ago
Even “only sort of works” is too generous for point #2. A dozen Claude Code agents spamming out code is… something. But it still does not replace a human, at all, even a junior one. It’s something else entirely.
empath75•15m ago
> Replace junior developers with Claude Code or similar.

I don't know why everyone goes to "replacing". Were a bunch of computer programmers replaced when compilers came out that made writing machine code a lot easier? Of course not, they were more productive and accomplished a lot more, which made them more valuable, not less.

jsmith99•11m ago
MS Copilot is quite useful for meeting minutes and summaries etc. Still not nearly as useful as good handwritten notes but saves loads of time.
m3kw9•51m ago
Is hard to measure return when you prompt a question and it comes out with a partial solution.
cm2012•51m ago
The report itself can only be viewed by filling out a form - this article is so details light as to be useless.
lacy_tinpot•50m ago
Why do people so desperately want to see AI fail?
lkramer•40m ago
I think people want to see the mindless application of crappy LLM chatbots and AI summaries everywhere fail. At least that's my position. I would also like the notion that "development can be sped up 6-700% by applying AI" would go away.
taormina•38m ago
Why do people so desperately want to see AI succeed? The financial investment explains it for some.
lacy_tinpot•29m ago
Why do want a technology that's able to automate human drudgery? Feels like an absurd question.
jurking_hoff•25m ago
Because it doesn’t. The jobs that will be left will be exclusively drudgery.
lacy_tinpot•12m ago
It doesn't because we already know it doesn't because just like think about it and it clearly doesn't. So everyone trying to make the technology is a big dumb dumb for even trying.

Like is the conclusion we shouldn't even try? This kind of thinking ridiculous.

snozolli•33m ago
Because it's going to destroy knowledge work and the entire middle class.
lacy_tinpot•29m ago
I thought we hated doing the menial office jobs though.
snozolli•16m ago
Maybe spoiled brats did. I like "menial" office jobs a lot more than starving or doing humiliating acts on OnlyFans to pay my bills, which is the future we're barreling toward.

What menial about knowledge work, anyway?

recallingmemory•31m ago
Humans like knowing their next paycheck is safe. AI is a disruption to that feeling of security.
jurking_hoff•26m ago
It’s not just that. It’s just that the mask is off and now the intent is stated clear: we are going to strip you of all your security and leave you high and dry.
rusted1•26m ago
We dont. People just tired of listening to the pipe dreams of these weirdos CEOs who sell AI.
exasperaited•10m ago
I don’t want to see it fail.

I do want people to understand what they are discarding when they use it to replace human creativity and human contact. All the evidence is that AI bros still think artists are evil gatekeepers.

And I absolutely want to see the bubble burst. I see absolutely no reason to be excited on the behalf of Silicon Valley vampires and their latest, soul-crushing innovation. If it’s making Andreessen Horowitz a stack of money, its value to humanity is already more questionable.

jampa•49m ago
The biggest mistake people are making is treating AI as a product instead of a feature.

While people are doing their work, they don't think, "Oh man, I am really excited to talk with AI today, and I can't wait to talk with a chatbot."

People want to do their jobs without being too bored and overwhelmed, and that's where AI comes in. But of course, we cannot hype features; we sell products after all, so that's the state we are in.

If you go to Notion, Slack, or Airtable, the headline emphasizes AI first instead of "Text Editor, Corporate Chat etc".

The problem is that AI is not "the thing", it is the "tool that gets you to the thing".

jameshart•46m ago
Agreed. We’ve got the potential to build real bicycles for the mind here and marketing departments are jumping right in to trying to sell people spandex cycling shorts.
the_snooze•40m ago
I wouldn't mind it if it were presented as yet another tool in the box. Maybe have a one-time popup saying "Hey, there's this thing, here's a cool use case, go check it out on your own terms."

In reality, AI sparkles and logos and autocompletes are everywhere. It's distracting. It makes itself the star of the show instead of being a backup dancer to my work. It could very well have some useful applications, but that's for users to decide and adapt to their particular needs. The ham-fisted approach of shoving it into every UI front-and-center signals a gross sense of desperation, neediness, and entitlement. These companies need to learn how to STFU sometimes.

ryandrake•33m ago
I wouldn't even call it a feature. It's enabling technology. I've never once said "I would like AI in [some product]." I say: "I would like to be able to [do this task]." If the company adds that feature to a product, I'll buy it. I don't care if the company used AI, traditional algorithms, or sorcery to make the feature work--I just care that it does what I want it to do.

Too many companies are just trying to spoon AI into their product somehow, as if AI itself is a desired feature, and are forgetting to find an actual user problem for it to actually solve.

TimCTRL•27m ago
I like this take, in fact i feel a little uneasy when i see startups mention "MCP" on their landing pages! Its a protocol and its like saying we use HTTP here.

I could be wrong but, all in all, buy a .com for your "ai" product, such that you survive the Dot-ai bubble [1]

[1] - https://en.wikipedia.org/wiki/Dot-com_bubble

biophysboy•25m ago
1000% agree, so many AI "applications" right now are solutions looking for a problem.
rank0•11m ago
Seriously! The product itself is supposed to be the valuable thing…regardless of the underlying technology.
TrackerFF•48m ago
I'm guessing every VC firm out there is hoping that they'll be the ones that picked the winners - the AI companies that will rule the world in N years.
os2warpman•47m ago
TIL 5% of companies are lying about their financials. (or 5% of companies are selling the pickaxes to the unfortunate miners)
ApeWithCompiler•47m ago
For the future I will relabel "AI" as "Ain't Interested". But despite the missing return of investment, it only needs a manager or a few, invested enough. They will push layoffs and restructuring regardless of better advice.
md3911027514•44m ago
It’s interesting how self-reports of productivity can be wrong.

For example a study from METR found that developers felt that AI sped them up by 20%, but it empirically it slowed them down by 19%. https://metr.org/blog/2025-07-10-early-2025-ai-experienced-o...

throitallaway•41m ago
Sometimes it's like an autocomplete on steroids and reads my mind. Other times it suggests things that make no sense and completely gets in the way.
sdeframond•37m ago
An interesting study indeed. Not enough data points but still way better than anecdata and self reporting!
rapind•23m ago
I suspect there's a lot of nuance you can't really capture in a study like this.

How you use AI will depend on the model, the tools (claude-code vs cursor vs w/e), your familiarity and process (planning phases, vibe coding, etc.), and the team size (solo dev versus large team), your seniority and attention to detail, and hard to measure effects like an increased willingness to tackle harder problems you may have procrastinated on otherwise.

I suspect we're heading to a plateau. I think there's a ton of polish that can be done with existing models to improve the coding experience and interface. I think that we're being massively subsidized by investors racing to own this market, but by the time they can't afford to subsidize it anymore, it'll be such a commodity that the prices won't go up and might even go down regardless of their individual losses.

As someone who knows they are benefitting from AI (study shmuddy), I'm perfectly fine with things slowing down since it's already quite good and stands to be much better with a focus on polish and incremental improvements. I wouldn't invest in these AI companies though!

biophysboy•22m ago
I sort of wonder if AI pushes people to stay "in the weeds". I've noticed that the speed of my software development hinges a lot on decisions that require me to take a step back (e.g. Am I using the right tool? Is this feature necessary? etc)
xyst•44m ago
I am not surprised at all. The capabilities of AI have been vastly oversold. The only winners here are the billionaires that have had early equity/investment in genAI.

They got a majority of the country hooked into AI without truly understanding its current limitations. This is just like digital currency bubble/fad that popped a couple of years ago.

What most companies got out of it is a glorified chatbot (ie, something that was possible in 2014…) at 1000X the cost.

What a sad joke. Innovation in this country is based on a lie, fueled by FOMO.

jawns•43m ago
Full disclosure: I'm currently in a leadership role on an AI engineering team, so it's in my best interest for AI to be perceived as driving value.

Here's a relatively straightforward application of AI that is set to save my company millions of dollars annually.

We operate large call centers, and agents were previously spending 3-5 minutes after each call writing manual summaries of the calls.

We recently switched to using AI to transcribe and write these summaries. Not only are the summaries better than those produced by our human agents, they also free up the human agents to do higher-value work.

It's not sexy. It's not going to replace anyone's job. But it's a huge, measurable efficiency gain.

vasco•40m ago
I wonder if the human agents agree the AI summaries are better than their summaries. I was nodding as I read and then told myself "yeah but it wouldn't be able to summarize the meetings I have", so I wonder if this only works in 3rd person.
dymk•37m ago
Have you tried having it summarize the meetings you have?
mbStavola•37m ago
Part of me also wonders if people may agree that its better simply because they don't actually have to do the summarization anymore. Even if it is worse by some %, that is an annoying task you are no longer responsible for; if anything goes wrong down the line, "ah the AI must've screwed up" is your way out.
roflc0ptic•35m ago
I’m inclined to believe that call center employees don’t have a lot of incentive to do a good job/care, so a lossy AI could quite plausibly be higher quality than a human
evereverever•12m ago
That re-synthesis of information is incredibly valuable to storing it in your own memory.

Of course, we can just rely on knowing nothing just to look things up, but I want more for thinking peoples.

jcims•23m ago
I built a little solution to record and transcribe all of my own meetings. I have many meetings (30hr week+) and I can't keep pace with adequate note-taking while participating in them all.

I'm finding that the summarization of individual meetings very useful, I'm also finding that the ability to send in transcripts across meetings, departments, initiatives whatever to be very effective at surfacing subtexts and common pain points much more effectively than I can.

I'm also using it to look at my own participation in meetings to help me see how I interact with others a (little) bit more objectively and it has helped me find ways to improve. (I don't take its advice directly lol, just think about observations and determine myself if it's something that's important and worth thinking about)

garciasn•14m ago
We have automated data -> PowerPoint reporting generation tools in production. We added LLM 'observations' to read in the data for a particular slide that's either in tabular or chart form, and provide insights for human analysts to use as a starting point for actual client-facing insights later.

We employ a loop where we generate the PPTX, it goes to the team for polish, they submit the final PPTX back, and the system assesses the differences between provided 'observations' from the LLMs and what the humans eventually delivered.

In 90% of cases, during our regular human review sessions, the LLMs (after much tweaking of context and prompts to make them sound less...flowery) are providing subjectively better insights than many of the humans. There are a wide variety of reasons for this, but it's primarily available time to deliver (too much to do; too wide of a range of slides to cover; etc, etc, etc). If the slide insights are just 'reporting the weather', the LLM can do that just fine. However, where the LLMs struggle, but humans outpace them, is in the real meaty insights that cross into areas where context about the business or the overall efforts are unavailable to the LLM.

In our case, we have the subjectively 'best' SMEs and insight writers evaluating the insights cross-client deliverables on a regular cadence and helping us tweak not only the LLMs but better training the humans on how to write better insights. Assuming there is a human eval in the loop, you're probably going to get a better system in place over time.

What I see in businesses trying to leverage LLMs is that they just assume they can replace everything w/the LLM and move along w/o supervision. This is nonsense and that sort of assumption, while likely driven by the AI-marketing-hype-cycle, is just batshit crazy to me. Due diligence is required at every step and it will pay off, just like any technology, but we want to pretend it isn't.

doubled112•39m ago
At work we've tried AI summaries for meetings, but we spent so much time fixing those summaries that we started writing our own again.

Is there some training you applied or something specific to your use case that makes it work for you?

nsxwolf•36m ago
We stopped after it kept transcribing a particular phrase of domain jargon as “child p*rn”, again and again.
cube00•35m ago
Unless a case goes down the legal road, nobody is ever bothering to read old call summaries in a giant call center.

When was the last time you called a large company and the person answering was already across all the past history without you giving them a specific case number?

doubled112•34m ago
Does an AI summary hold up in court? Or would you still need to review a transcript or recording anyway?
cube00•31m ago
You can store low quality audio cheaply on cold storage so I suspect that's the real legal record if it got that far.
shawabawa3•29m ago
My guess is that the summaries are never actually read, so accuracy doesn't actually matter and the AI could equally be replaced with /dev/null
butlike•39m ago
How can you double-check the work? Also, what happens when the AI transcription is wrong in a way that would have terminated the employee. You can't fire a model.

Finally, who cares about millions saved (while considering the above introduced risk), when trillions are on the line?

throitallaway•38m ago
I presume they're not using these notes for anything mission or life critical, so anything less than 100% accuracy is OK.
butlike•36m ago
I disagree with the concept of affluvic notes. All notes are intrinsically actionable; it's why they're a note in the first place. Any note has unbounded consequence depending on the action taken from it.
wredcoll•18m ago
You're being downvoted, I suspect for being a tad hyperbolic, but I think you are raising a really important point, which is just the ever more gradual of removing a human's ability to disobey the computer system running everything. And the lack of responsibility for following computer instructions.

It's a tad far-fetched in this specific scenario, but an AI summary that says something like "cancel the subscription for user xyz" and then someone else takes action on that, and XYZ is the wrong ID, what happens?

PaulRobinson•35m ago
Having a human read a summary is way faster than getting them to write it. If they want to edit it, they can.

AI today is terrible at replacing humans, but OK at enhancing them.

Everyone who gets that is going to find gains - real gains, and fast - and everyone who doesn't, is going to end up spending a lot of money getting into an almost irreversible mistake.

butlike•31m ago
"Reading a summary is faster, so enhancing humans with AI is going to receive boons or busts to the implementer."

Now, summary, or original? (Provided the summary is intentionally vague to a fault, for arguments sake on my end).

pluc•38m ago
You could do that offline 20 years ago with Dragon Naturally Speaking and a one-time licence.
dymk•36m ago
You could get a transcript, not a summary
loloquwowndueo•35m ago
Only if your audio was crystal clear, you spoke like a robot very slowly, and each customer has to do a 30-minute “please read this text slowly to train the speech recognition software” preamble before talking to the actual human.
glimshe•38m ago
This makes sense. AI is obviously useful for many things. But people wouldn't invest tens of billions to summarize data center calls and similar tasks. Replacing data center workers isnt where the money is - it's replacing 100K-200K/year workers.
generic92034•38m ago
> It's not going to replace anyone's job.

Is it not, in the scenario you are describing? You are saying the agents are free now to do higher-value work. Why were there not enough agents before, especially if higher-value work was not done?

hobs•37m ago
Because call centers are cost centers - nobody pays a dime more than they have to in these situations and its all commodity work.
generic92034•33m ago
But that means the so-called "higher-value" work does not need to be done, so agents can be fired.
cube00•34m ago
It's such a useless platitude. The "higher value work" is answer more calls so we can have less staff on queue.
positron26•38m ago
Given that people skimp on work that is viewed as trash anyway, how were you getting value out of the summaries in the first place?
thrown-0825•31m ago
they weren't

its likely a checkbox for compliance or some policy a middle manager put in place that is now tied to a kpi

positron26•16m ago
Could be CRM, leaving summaries for the next person. I suppose it would sound like I'm implying a prior.
Terr_•38m ago
I think the biggest issue is accurately estimating the LLM failure risk, and what impacts the company is willing to tolerate in the long term. (As distinct from what the company is liable to permit through haste and ignorance.)

With LLMs the risk is particularly hard to characterize, especially when it comes to adversarial inputs.

hobs•38m ago
That is a good thing, but that's also just a training gap - I worked with tech support agents for years in gigantic settings and taking notes while you take an action is difficult to train but yields tangible results, clarity from the agent on what they are doing step by step, and builds a shared method of documenting things that importantly focuses on the important details and doesn't miss a piece which may be considered trivial by an outsider but (for instance) defines SOP for call escalation and the like.
dsr_•37m ago
Pro-tip: don't write the summary at all until you need it for evidence. Store the call audio at 24Kb/s Opus - that's 180KB per minute. After a year or whatever, delete the oldest audio.

There, I've saved you more millions.

lotsofpulp•35m ago
The summaries can help automate performance evaluation. If the employee disputes it, I imagine they pull up the audio to confirm.
anoojb•29m ago
Also entrenches plausible deniability and makes legal contests way more cumbersome for plantiffs to resolve.
paulddraper•29m ago
This works unless you want to automate something with the transcripts, stats, feedback.
Spivak•22m ago
Why wouldn't it, once you actually have that project you have the raw audio to generate the transcripts. Only spend the money at the last second when you know you need it.

Edit: Tell me more how preemptively spending five figures to transcribe and summarize calls in case you might want to do some "data engineering" on it later is a sound business decision. What if the model is cheaper down the road?

alooPotato•27m ago
you want to be able to search over summaries so you need to generate them right away
krainboltgreene•23m ago
Pro-tip: You won't ever do that.
alooPotato•22m ago
we do
ninininino•17m ago
Advanced organizations (think not startups, but companies that have had years of decades of profit in the public market) might have solved all the low-hanging fruit problems and have staff doing things like automated quality audits (search summaries for swearing, abusive language, etc).
ch4s3•15m ago
I would imagine OP is probably mining service call summaries to find common service issues, or at least that's what I would do.
deadbabe•11m ago
Do you want to search summaries, or do you want to save millions of dollars per year?
morkalork•5m ago
I can assure you that people care very much about searching and mining calls, especially for compliance and QA reasons.
andix•16m ago
Imagine a follow-up call of a customer. They are referring to earlier calls and the call center agents needs to check what it was about. So they can skim/read the transcripts while talking to the customer. I guess it's really hard to listen to transcripts while you're on the phone.
dsr_•13m ago
That would be awesome!

But in fact, customer call centers tend not to be able to even know that you called in yesterday, three days ago and last week.

This is why email-ticketing call centers are vastly superior.

ssharp•9m ago
I've always guessed that they are able to tell when you called/what you called about, but they simply don't give that level of information to their frontline folks.
ethagknight•5m ago
Im imagining my actual experience of being transferred for the 3rd or 4th time, repeating my name and address for the 3rd or 4th time, restating my problem for the 3rd or 4th time... feels like theres an implementation problem, not a technological problem.
doorhammer•16m ago
Sentiment analysis, nuanced categorization by issue, detecting new issues, tracking trends, etc, are the bread and butter of any data team at a f500 call center.

I'm not going to say every project born out of that data makes good business sense (big enough companies have fluff everywhere), but ime anyway, projects grounded to that kind of data are typically some of the most straight-forward to concretely tie to a dollar value outcome.

FirmwareBurner•10m ago
>Store the call audio at 24Kb/s Opus - that's 180KB per minute

Why OPUS though? There's dedicated audio codecs in the VoiP/telecom industry that are specifically designed for the best size/quality for voice call encoding.

tux3•36m ago
> it's a huge, measurable efficiency gain.

> It's not going to replace anyone's job

Mechanically, more efficiency means less people required for the same output.

I understand there is no evidence that any other sentence can be written about jobs. Still, you should put more text in between those two sentences. Reading them so close together creates audible dissonance.

flkiwi•22m ago
You're not accounting for teams already being understaffed and overtasked in many situations, with some AI tools allowing people to get back to doing their jobs. We aren't expecting significant headcount changes, but we are expecting significant performance and quality improvements for the resources we have.
tux3•19m ago
The reason that caused the team to be understaffed and overtasked has not gone away because of AI. I am expecting the team to stay understaffed and overtasked, for the same reason it was before: it's less expensive. With or without an LLM summarizing phone calls.
belter•36m ago
Are the summaries reviewed by the agents? And if not how do you handle hallucinations, or transcribe of wrong insurance policy id for example? Like, customer wants to cancel insurance policy AB-2345D and transcribe says wants to cancel insurance policy AD-2345B >
pedrocr•36m ago
> agents were previously spending 3-5 minutes after each call writing manual summaries of the calls

Why were they doing this at all? It may not be what is happening in this specific case but a lot of the AI business cases I've seen are good automations of useless things. Which makes sense because if you're automating a report that no one reads the quality of the output is not a problem and it doesn't matter if the AI gets things wrong.

In operations optimization there's a saying to not go about automating waste, cut it out instead. A lot of AI I suspect is being used to paper over wasteful organization of labor. Which is fine if it turns out we just aren't able to do those optimizations anyway.

nulbyte•21m ago
As a customer of many companies who has also worked in call centers, I can't tell you how frustrating it is when I, as a customer, have to call back and the person I speak with has no record or an insufficient record of my last call. This has required me to repeat myself, resend emails, and wait all over again.

It was equally frustrating when I, as a call center worker, had to ask the custmer to tell me what should already have been noted. This has required me to apologize and to do someone else's work in addition to my own.

Summarizing calls is not a waste, it's just good business.

Shank•36m ago
Who reads the summaries? Are they even useful to begin with? Or did this just save everyone 3-5 minutes of meaningless work?
vosper•32m ago
AI reads them and identifies trends and patterns, or answers questions from PMs or others?
doorhammer•23m ago
Not the op, but I did work supporting three massive call centers for an f500 ecom.

It's 100% plausible it's busy work but it could also be for: - Categorizing calls into broad buckets to see which issues are trending - Sentiment analysis - Identifying surges of some novel/unique issue - Categorizing calls across vendors and doing sentiment analysis that way (looking for upticks in problem calls related to specific TSPs or whatever) - etc

False positives and negatives aren't really a problem once you hit a certain scale because you're just looking for trends. If you find one, you go spot-check it and do a deeper dive to get better accuracy.

Which is also how you end up with some schlepp like me listening to a few hundreds calls in a day at 8x speed (back when I was a QA data analyst) to verify the bucketing. And when I was doing it everything was based on phonetic indexing, which I can't imagine touching llms in terms of accuracy, and it still provided a ton of business value at scale.

recallingmemory•34m ago
So long as precision isn't important, I suppose. Hallucination within summaries is the issue I keep running into which prevents me from incorporating it into any of our systems.
thrown-0825•31m ago
I have seen ai summaries of calls get people into trouble because the ai hallucinated prices and features that didn't exist
trevor-e•34m ago
This is a great use-case of AI.

However I strongly doubt your point about "It's not going to replace anyone's job" and that "they also free up the human agents to do higher-value work". The reality in most places is that fewer agents are now needed to do the same work as before, so some downsizing will likely occur. Even if they are able to switch to higher-value work, some amount of work is being displaced somewhere in the chain.

And to be clear I'm not saying this is bad at all, I'm just surprised to see so many deluded by the "it won't replace jobs" take.

varispeed•34m ago
What’s the actual business value of a “summary” though? A transcript is the record. A tag or structured note (“warranty claim,” “billing dispute,” “out of scope”) is actionable. But a free-form blob of prose? That’s just narrative garnish - which, if wrong or biased, is worse than useless.

Imagine a human agent or AI summarises: “Customer accepted proposed solution.” Did they? Or did they say “I’ll think about it”? Those aren’t the same thing, but in the dashboard they look identical. Summaries can erase nuance, hedge words, emotional tone, or the fact the customer hung up furious.

If you’re running a call centre, the question is: are you using this text to drive decisions, or is it just paperwork to make management feel like something is documented? Because “we saved millions on producing inaccurate metadata nobody really needs” isn’t quite the slam dunk it sounds like.

amluto•31m ago
I’d like to see a competent AI replace the time that doctors and nurses spend tediously transcribing notes into a medical record system. More time spent doing the actual job is good for pretty much everyone.
simmerup•27m ago
Until it hallucinates and the AI has written something wrong about you in your official medical record
nottorp•25m ago
Competent, yes. But the current ones are likely to transcribe "recommend amputating the left foot" as "recommend amputating the right foot". Still want it?
beart•22m ago
But... that is the actual job. A clear medical history is very important, and I'm not ready yet to cut out my doctor from that process.

This reminds me of the way juniors tend to think about things. That is, writing code is "the actual job" and commit messages, documentation, project tracking, code review, etc. are tedious chores that get in the way. Of course, there is no end to the complaints of legacy code bases not having any of those things and being difficult to work with.

jordanb•29m ago
We use Google meet and it has Gemini transcriptions of our meetings.

They are hilariously inaccurate. They confuse who said what. They often invert the meaning "Joe said we should go with approach x" where Joe actually said we should not do X. It also lacks context causing it to "mishear" all of our internal jargon to "shit my iPhone said" levels.

ricardonunez•15m ago
I don’t know how it can confuse because input on mic is relatively straight forward to get. I use fathom and others and they are accurate, better than manual taken. Interesting take, that I don’t memorize 100% on the calls anymore since I rely on note takers, I only remember the major points but when I read the notes, everything comes clear.
rowanseymour•12m ago
Same here. It's frustrating that it doesn't seem to have contextual awareness of who we are and the things we work on so things like names of our products, names of big clients, that we use repeatedly in meetings, are often butchered.
orphea•7m ago
Oh, that's what happening. I thought my English is just terrible :(
actsasbuffoon•29m ago
That’s the thing. There’s value in AI, it’s just not worth half a trillion dollars to train a new model that’s 0.4% better on benchmarks. Meta is never going to get a worthwhile return on spending $100M on individual engineers.

But that doesn’t mean AI is without its uses. We’re just in that painful phase where the hype needs to die down and we treat LLMs as what they really are; an interesting new tool in the toolkit that provides some new ways to solve problems. It’s almost certainly not going to turn into AGI any time soon. It’s not worth trillions. It’s certainly worth something, though.

I think the financials on developing new frontier models are terrible. But I’ve already built multiple AI projects for my company that are making money and we’ve got extremely happy customers.

Investors thought one company was going to win the AI Wars and make a quadrillion dollars. Instead it’s probably going to be 10,000 startups that will build interesting products based on AI, and training new models won’t actually be a good financial move.

Imanari•12m ago
Could you broadly describe the AI projects you have built?
nuker•28m ago
> recently switched to using AI to transcribe and write these summaries

Did users knew that conversation was recorded?

prophesi•19m ago
You would be hard-pressed to find a call center that _doesn't_ start every phone call with a warning that the conversation may be recorded.
creaturemachine•14m ago
Yeah the standard "this call may be recorded for quality or training purposes" preamble shouldn't cover for slurping your voiceprint to further the butchering of client service that this call centre is here for.
watwut•14m ago
Typical call center call is recorded and you are told so by the start of the conversation. I had quite a few of those.
Capricorn2481•25m ago
> Not only are the summaries better than those produced by our human agents

We have someone using Firefly for note taking, and it's pretty bad. Frequently gets details wrong or extrapolates way too much from a one-off sentence someone said.

How do you verify these are actually better?

ghalvatzakis•23m ago
I lead an AI engineering team that automated key parts of an interviewing process, saving thousands of hours each month by handling thousands of interviews. This reduced repetitive, time-consuming tasks and freed human resources to focus on higher-value work
the_snooze•19m ago
I'm under the impression that one of the most critical responsibilities a lead has is to establish and maintain a good working culture. Properly vetting new additions feeds directly into that. Why offload it to AI?
ghalvatzakis•12m ago
Just to clarify, these aren’t interviews for job positions
pjmorris•22m ago
I wonder how this change affects what the agents remember about the calls, and how that affects their performance on future calls. And I wonder whether agent performance, as measured by customer satisfaction, will decline over time, and whether that will affect the bottom line.
stronglikedan•16m ago
I wouldn't allow myself to be held accountable for anything in a summary I didn't write.
MangoToupe•14m ago
> We operate large call centers, and agents were previously spending 3-5 minutes after each call writing manual summaries of the calls.

This is a tiny fraction of all work done. This is work people were claiming to have solved 15 years ago. Who cares?

didibus•13m ago
Where is the money being saved? Are you reducing the number of agents? Otherwise, it should actually cost more, before you simply had customers wait longer to speak to the next agent no? Or do you sell "support calls" so you're able to sell more of them given the same number of agents?
ozgune•13m ago
Previously discussed here: https://news.ycombinator.com/item?id=44941118

It's also disappointing that MIT requires you to fill out a form (and wait for) access to the report. I read four separate stories based on the report, and they all provide a different perspective.

Here's the original pdf before MIT started gating it: https://web.archive.org/web/20250818145714/https://nanda.med...

doorhammer•12m ago
I'm curious, have you noticed an impact on agent morale with this?

Specifically: Do they spend more time actually taking calls now? I guess as long as you're not at the burnout point with utilization it's probably fine, but when I was still supporting call centers I can't count the number of projects I saw trying to push utilization up not realizing how real burnout is at call centers.

I assume that's not news to you, of course. At a certain utilization threshold we'd always start to see AHTs creep up as agents got burned out and consciously or not started trying to stay on good calls.

Guess it also partly depends on if you're in more of a cust serv call center or sales.

I hated working as an actual agent on the phones, but call center ops and strategy at scale has always been fascinating.

dsr_•41m ago
"I'm sure my company is among the 5% of superusers," said CEO Chad "Bro" Brochad, who later admitted he had not checked.
jurking_hoff•14m ago
> who later admitted he had not checked.

Incorrect. He did check, and decided to lie.

Terr_•41m ago
Can you really put a price on a good excuse for layoffs that deflects rancor from employees while sounding like positive news to investors? :p

... Well, probably yes, but I don't have the data to do it.

tqi•40m ago
The link to the report pdf redirects to a landing page[1] that makes me think this "study" is run-of-the-mill content marketing. The CTA at the bottom of the page says "Join the Revolution / Be at the forefront of creating the agentic web. Get exposure and faster adoption for your AI products and tools," which certainly doesn't give the impression that this is an objective report. Either way I can't speak to the quality of the study directly because it is walled off behind a contact information form, which is another bad sign.

From the article though:

> But researchers found most use cases were limited to boosting individual productivity rather than improving a company’s overall profits.

What does that even mean?

[1] Website: https://nanda.media.mit.edu/, FAQ: https://projnanda.github.io/projnanda/#/faq_nanda

kingstnap•40m ago
I mean of course they aren't. So many people are offering so much stuff for free or at a pittance.

I honestly don't think it matters though. Feel free to disagree with me but I think the money is irrelevant.

The only thing that actually matters is the long run is the attention, time, and brain space of other people. After all that's where fiat currency actually derives it's value. These Gen AI companies have captured a lot of that extremely quickly.

OpenAI might have "burned" billions but they way they have wrung themselves into seemingly every university student's computer, every CEOs mind, the policy decisions of world leaders, ever other hackernews post, is nothing short of miraculous...

justinator•40m ago
It HAS lead to widespread enshitification of products and services though!
mannyv•35m ago
My daughter, as an intern, created a whole set of prompts for a metal fab that extracted all the metal parts out of a CAD file (or PDF) and dimensions so it's easier for them to bid.

Saved them hours of work.

Of course, they didn't spend on "AI" per se.

Most people don't know how to meta their job functions, so AI won't really be worth it. And the productivity gains may not be measurable ie: "I did this in 5 minutes instead of 500, so I was able to goof off more."

ryeguy_24•34m ago
Does anyone have this mystical report?
nerevarthelame•27m ago
https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Bus...
rconti•33m ago
I'm no AI apologist, but for one, this is how investments, particularly speculative investments, work. They're investing money now in the _hopes_ of a future return. It's pretty early days still. Secondly, of _course_ the huge AI players are doing everything they can to overpromise and convince corporations to throw cash at them to keep the party going.

I think the real problem is, it's just a bit too early, but every CEO out there dreams of being lauded for their visionary take on AI, and nobody wants to miss the bus. It's high-leverage tech, so if it (some day) does what it's supposed to do, and you miss making the investment at the right time, you're done.

maerF0x0•33m ago
If everyone does it, then no one gets an advantage, because all see a productivity boost.

If you do not do it, you get left behind and cannot compete in the marketplace.

I took a business systems administration course like 20 years ago, and they knew this was the case. As far as we can tell it's always been the case.

IT doesn't create massive moats/margins because price competition erodes the gap. And yet if you do not keep up you lose.

It's definitely a boon for humanity though, in the industries where technology applies things have been very obviously getting much cheaper over time.

(Most notably American housing has been very very resistant to technological change and productivity gains, a part of the story why housing has gone way up) - https://youtu.be/VfYp9qkUnt4?si=D-Jpmojtn7zV5E8T

manishsharan•32m ago
I am not interested in the results of the 95%.

I want to know more about the 5% who got it right. What are their use cases ?

nextworddev•32m ago
Lol, I guess “they” want a drawdown in QQQs
revskill•29m ago
Because they are doom to fail.
morelandjs•27m ago
I’d have to guess that new startup founders are building leaner teams leading to slower burn rates and longer run ways. That has value. I think share skepticism is warranted is whether or not old behemoths can retrofit AI efficiency gains into their existing organizational structures.
crowcroft•26m ago
IMO 'Zero return' is really just saying that these companies never had a plan for their AI implementation in the first place, and so never had any idea what they were trying to achieve and measure in the first place.

The article does call out clear issues companies have with AI workflows etc. and those are likely real problems, but if you're saying *zero* return those aren't the root cause problems.

seatac76•24m ago
Ed Zitron was right all along eh.
ChrisArchitect•22m ago
Earlier discussion on the report:

AI is predominantly replacing outsourced, offshore workers

https://news.ycombinator.com/item?id=44940944

PDF report that was taken down/walled: https://web.archive.org/web/20250818145714/https://nanda.med...

pjmlp•22m ago
I can hardly wait for the bubble to burst, now everyone is forced to meet OKRs on using AI at work, while at the same time, being forbidden to use the data that actually makes the tools usable.
turnsout•22m ago
The AI wave is so reminiscent of the early days of the internet. Right now we're in about 1999, a time of tremendous hype. Business that weren't even in tech were wondering how much they need to know about the HTTP spec. People from the marketing team were being pulled off their print ad campaigns, given a copy of Dreamweaver, and told to make the company website. We hadn't even figured out the roles.

It 100% turned out to be a bubble and yet, if anything, the internet was under-hyped. The problem in 1999 was that no one really knew how it was going to play out. Which investments would be shrewd in retrospect, and which ones would be a money pit?

When an innovation hits, it takes time to figure out whether you're selling buggy whips, or employing drivers who can drive any vehicle.

Plenty of companies sunk way too much money into custom websites back in 99, but would we say they were wrong to do it? They may have overspent at a time when a website couldn't justify the ROI within 12 months, but how could they know? A few short years later, a website was virtually required for every business.

So are companies really seeing "zero return" on their AI spend, or are they paying for valuable lessons about how AI applies to their businesses? There may be zero ROI today, but all you need to do is look at the behavior of normal people to see that AI is not going anywhere. Smart companies are experimenting.

foxfired•21m ago
Atlassian added this new AI feature to create audio summaries of Confluence documents. It's really impressive. No one reads documentation. I know it because I'll send someone a documentation, then they'll ask me a question that is answered in the second paragraph, meaning they didn't read it.

So their feature is not just text to speech, but a reading of a summarized version of the articles. But here is the problem. The documentation has no fluff. You don't want a summary, you want the actual details. When you are reading the document that describes how the recovery fee is calculated, you want to know exactly how it is calculated.

I've ran it on multiple documents and it misses key information. An unsuspecting user might take it at face value. So this feature looks impressive, but it misses the entire point of documentation. Which is *preserving the details*.

GloriousMEEPT•20m ago
One thing I ponder after remembering previous tech booms, is that they often left behind something extremely valuable (fiber). With all the GPU datacenters rolling out, what next bubble can take advantage of this boon, if it is one?
eulgro•16m ago
It will probably all be used to mine bitcoin.
OutOfHere•10m ago
Bitcoin is mined only with custom ASICs, no longer with GPUs.
wheelerwj•15m ago
My guess is that this title could also be written as, “The value of AI projects are being captured by just 5% of companies.”

It’s pretty clear to anyone who’s using this technology that it’s significant. Theres still tons to work out and the exact impact is still unknown. But this cat isn’t going back in the bag.

resiros•13m ago
Here is the report: https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Bus...

The story there is very different than what's in the article.

Some infos:

- 50% of the budgets (the one that fails) went to marketing and sales

- the authors still see that AI would offer automation equaling $2.3 trillion in labor value affecting 39 million positions

- top barriers for failure is Unwillingness to adopt new tools, Lack of executive sponsorship

Lots of people here are jumping to conclusions. AI does not work. I don't think that's what the report says.

baal80spam•5m ago
> Lots of people here are jumping to conclusions. AI does not work. I don't think that's what the report says.

Well...

"It is difficult to get a man to understand something when his salary depends upon his not understanding it"

bishnu•11m ago
The illustrative example to me is the construction of railroads at the beginning of the Industrial revolution. Unquestionably useful, unquestionably important there were nevertheless at least 3 speculative financial bubbles centered on railway construction.