frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Why your boss isn't worried about AI – "can't you just turn it off?"

https://boydkane.com/essays/boss
136•beyarkay•3h ago

Comments

kazinator•2h ago
> AIs will get more reliable over time, like old software is more reliable than new software.

:)

Was that a humam Freudian slip, or artificial one?

Yes, old software is often more reliable than new.

joomla199•2h ago
Neither, you’re reading it wrong. Think of it as codebases getting more reliable over time as they accumulate fixes and tests. (As opposed to, say, writing code in NodeJS versus C++)
giancarlostoro•2h ago
Age of Code does not automatically equal quality of code, ever. Good code is maintained by good developers. A lot of bad code is pushed out by management, and other situations, or just bad devs. This is a can of worms you're talking your way into.
1313ed01•2h ago
Old code that has been maintained (bugfixed), but not messed with too much (i.e. major rewrites or new features) is almost certain to be better than most other code though?
eptcyka•2h ago
I’ve read parts of macOS’ open source code that surely has been around for a while, maintained and absolute rubbish.
DSMan195276•2h ago
"Bugfixes" doesn't mean the code actually got better, it just means someone attempted to fix a bug. I've seen plenty of people make code worse and more buggy by trying to fix a bug, and also plenty of old "maintained" code that still has tons of bugs because it started from the wrong foundation and everyone kept bolting on fixes around the bad part.
prasadjoglekar•2h ago
It actually might. Older code running in production is almost automatically regression tested with each new fix. It might not be pretty, but it's definitely more reliable for solving real problems.
shakna•2h ago
The list of bugs tagged regression at work certainly suggests it gets tested... But fixing those regressions...? That's a lot of dev time for things that don't really have time allocated for them.
LeifCarrotson•1h ago
You're using different words - the top comment only mentioned the reliability of the software, which is only tangentially related to the quality, goodness, or badness of the code used to write it.

Old software is typically more reliable, not because the developers were better or the software engineering targeted a higher reliability metric, but because it's been tested in the real world for years. Even more so if you consider a known bug to be "reliable" behavior: "Sure, it crashes when you enter an apostrophe in the name field, but everyone knows that, there's a sticky note taped to the receptionist's monitor so the new girl doesn't forget."

Maybe the new software has a more comprehensive automated testing framework - maybe it simply has tests, where the old software had none - but regardless of how accurate you make your mock objects, decades of end-to-end testing in the real world is hard to replace.

As an industrial controls engineer, when I walk up to a machine that's 30 years old but isn't working anymore, I'm looking for failed mechanical components. Some switch is worn out, a cable got crushed, a bearing is failing...it's not the code's fault. It's not even the CMOS battery failing and dropping memory this time, because we've had that problem 4 times already, we recognize it and have a procedure to prevent it happening again. The code didn't change spontaneously, it's solved the business problem for decades... Conversely, when I walk up to a newly commissioned machine that's only been on the floor for a month, the problem is probably something that hasn't ever been tried before and was missed in the test procedure.

freetime2•48m ago
Yup, I have worked on several legacy codebases, and a pretty common occurence is that a new team member will join and think they may have discovered a bug in the code. Sometimes they are even quite adamant that the code is complete garbage and could never have worked properly. Usually the conversation goes something like: "This code is heavily used in production, and hasn't been touched in 10 years. If it's broken, then why haven't we had any complaints from users?"

And more often than not the issue is a local configuration issue, bad test data, a misunderstanding of what the code is supposed to do, not being aware of some alternate execution path or other pre/post processing that is running, some known issue that we've decided not to fix for some reason, etc. (And of course sometimes we do actually discover a completely new bug, but it's rare).

To be clear, there are certainly code quality issues present that make modifications to the code costly and risky. But the code itself is quite reliable, as most bugs have been found and fixed over the years. And a lot of the messy bits in the code are actually important usability enhancements that get bolted on after the fact in response to real-world user feedback.

hatthew•1h ago
I think we all agree that the quality of the code itself goes down over time. I think the point that is being made is that the quality of the final product goes up over time.

E.g. you might fix a bug by adding a hacky workaround in the code; better product, worse code.

kube-system•1h ago
The author didn't mean that an older commit date on a file makes code better.

The author is talking about the maturity of a project. Likewise, as AI technologies become more mature we will have more tools to use them in a safer and more reliable way.

izzydata•2h ago
Sounds more like survivorship bias. All the bad codebases were thrown out and only the good ones lasted a long time.
wsc981•2h ago
Basically the Lindy Effect: https://en.wikipedia.org/wiki/Lindy_effect
wvenable•1h ago
In my experience actively maintained but not heavily modified applications tend towards stability over time. It don't even matter if they are good or bad codebases -- even a bad code will become less buggy over time if someone is working on bug fixes.

New code is the source of new bugs. Whether that's an entirely new product, a new feature on an existing project, or refactoring.

kazinator•1h ago
You mean think of it as opposite to what is written in the remark, and then find it funny?

Yes, I did that.

james_marks•1h ago
I’ve always called this “Work Hardening”, as in, the software has been improved over time by real work being done with it.
glitchc•1h ago
Perhaps better rephrased as "software that's been running for a (long) while is more reliable than software that only started running recently."
kstrauser•1h ago
Holy survivorship bias, Batman.

If you think modern software is unreliable, let me introduce you to our friend, Rational Rose.

kazinator•1h ago
At least that project was wise enough to use Lisp for storing its project files.
noir_lord•1h ago
Agreed.

Or debuggers that would take out the entire OS.

Or a bad driver crashing everything multiple times a week.

Or a misbehaving process not handing control back to the OS.

I grew up in the era of 8 and 16 bit micros and early PCs, they where hilariously less stable than modern machines while doing far less, there wasn’t some halcyon age of near perfect software, it’s always been a case of things been good enough to be good enough but at least operating systems did improve.

malfist•1h ago
Remember BSODs? Used to be a regular occurrence, now they're so infrequent they're gone from windows 11
kazinator•1h ago
I remember Linux being remarkable reliable throughout its entire life in spite of being rabidly worked on.

Windows is only stabilizing because it's basically dead. All the activity is in the higher layers, where they are racking their brains on how to enshittify the experience, and extract value out of the remaining users.

wlesieutre•44m ago
And the "cooperative multitasking" in old operating systems where one program locking up meant the whole system was locked up
krior•42m ago
Gone? I had two last year, lets not overstate things.
ClimaxGravely•18m ago
Still get them fairly regularly except now they come with a QR code.
Yoric•43m ago
I grew up in the same era and I recall crashes being less frequent.

There were plenty of other issues, including the fact that you had to adjust the right IRQ and DMA for your Sound Blaster manually, both physically and in each game, or that you needed to "optimize" memory usage, enable XMS or EMS or whatever it was at the time, or that you spent hours looking at the nice defrag/diskopt playing with your files, etc.

More generally, as you hint to, desktop operating systems were crap, but the software on top of it was much more comprehensively debugged. This was presumably a combination of two factors: you couldn't ship patches, so you had a strong incentive to debug it if you wanted to sell it, and software had way fewer features.

Come to think about it, early browsers kept crashing and taking down the entire OS, so maybe I'm looking at it with rosy glasses.

binarymax•55m ago
You know, I had spent a good amount of years not having even a single thought about rational rose, and now that’s all over.
cjbgkagh•49m ago
How much of that do you think would be attributable to IBM or Rational Software?
fidotron•2h ago
But this is why using the AI in the production of (almost) deterministic systems makes so much sense, including saving on execution costs.

ISTR someone else round here observing how much more effective it is to ask these things to write short scripts that perform a task than doing the task themselves, and this is my experience as well.

If/when AI actually gets much better it will be the boss that has the problem. This is one of the things that baffles me about the managerial globalists - they don't seem to appreciate that a suitably advanced AI will point the finger at them for inefficiency much more so than at the plebs, for which it will have a use for quite a while.

pixl97•1h ago
>that baffles me about the managerial globalists

It's no different from those on HN that yell loudly that unions for programmers are the worst idea ever... "it will never be me" is all they can think, then they are protesting in the streets when it is them, but only after the hypocrisy of mocking those in the street protesting today.

hn_acc1•1h ago
Agreed. My dad was raised strongly fundamentalist, and in North America, that included (back then) strongly resisting unions. In hindsight, I've come to realize that my parent's weren't maybe even of average intelligence, and definitely of above-average gullibility.

Unionized software engineers would solve a lot of the "we always work 80 hour weeks for 2 months at the end of a release cycle" problems, the "you're too old, you're fired" issues, the "new hires seems to always make more than the 5/10+ year veterans", etc. Sure, you wouldn't have a few getting super rich, but it would also make it a lot easier for "unionized" action against companies like Meta, Google, Oracle, etc. Right now, the employers hold like 100x the power of the employees in tech. Just look at how much any kind of resistance to fascism has dwindled after FAANG had another round of layoffs..

fidotron•1h ago
Software "engineers" totally miss a key thing in other engineering professions as well, which is organizations to enforce some pretense of ethical standards to help push back against requests from product. Those orgs often look a lot like unions.
hn_acc1•1h ago
A bunch of short scripts doesn't easily lead to a large-scale robust software platform.

I guess if managers get canned, it'll be just marketing types left?

xutopia•2h ago
The most likely danger with AI is concentrated power, not that sentient AI will develop a dislike for us and use us as "batteries" like in the Matrix.
preciousoo•2h ago
Seems like a self fulfilling prophecy
yoyohello13•2h ago
Definitely not ‘self’ fulfilling. There are plenty of people actively and vigorously working to fulfill that particular reality.
fidotron•2h ago
I'm not so sure it will be that either, it would be having multiple AIs essentially at war with each other over access to GPUs/energy or whatever the materials are needed to grow if/when that happens. We will end up as pawns in this conflict.
ben_w•5m ago
Given that even fairly mediocre human intelligences can run countries into the ground and avoid being thrown out in the process, it's certainly possible for an AI to be in the intelligence range where it's smart enough to win vs humans but also dumb enough to turn us into pawns rather just go to space and blot out the sun with a Dyson swarm made from the planet Mercury.

But don't count on it.

I mean, apart from anything else, that's still a bad outcome.

pcdevils•2h ago
For one thing, we'd make shit batteries.
prometheus76•1h ago
They farm you for attention, not electricity. Attention (engagement time) is how they quantify "quality" so that it can be gamed with an algorithm.
noir_lord•1h ago
IIRC the original idea was that the machines used our brain capacity as a distributed array but then they decided batteries was easier to understand while been sillier, just burn the carbon they are feeding us, it’s more efficient.
darth_avocado•2h ago
The reality is that the CEO/executive class already has developed a dislike for us and is trying to use us as “batteries” like in the Matrix.
ljlolel•2h ago
CEOs (even most VCs) are labor too
pavel_lishin•2h ago
Do they know it?
toomuchtodo•2h ago
Labor competes for compensation, CEOs compete for status (above a certain enterprise size, admittedly). Show me a CEO willingly stepping down to be replaced by generative AI. Jamie Dimon will be so bold to say AI will bring about a 3 day week (because it grabs headlines [1]) but he isn't going to give up the status of running JPMC; it's all he has besides the wealth, which does not appear to be enough. The feeling of importance and exceptionalism is baked into the identity.

[1] https://fortune.com/article/jamie-dimon-jpmorgan-chase-ceo-a...

Animats•1h ago
That's the market's job. Once AI CEOs start outperforming human CEOs, investment will flow to the winners. Give it 5-10 years.

(Has anyone tried an LLM on an in-basket test? [1] That's a basic test for managers.)

[1] https://en.wikipedia.org/wiki/In-basket_test

conception•1h ago
Spoiler there’s no reason we couldn’t work three days a week now. And 100 might be pushing it, but having life expectancy to 90 as well within our grass today as well. We have just decided not to do that.
darth_avocado•2h ago
Until shareholders treat them as such, they will remain in the ruling class
icedchai•9m ago
Almost everyone is "labor" to some extent. There is always a huge customer or major investor that you are beholden to. If you are independently wealthy then you are the exception.
vladms•1h ago
Do you know personally some CEO-s? I know a couple and they generally seem less empathic than the general population, so I don't think that like/dislike even applies.

On the other hand, trying to do something "new" is lots of headaches, so emotions are not always a plus. I could make a parallel to doctors: you don't want a doctor to start crying in a middle of an operation because he feels bad for you, but you can't let doctors doing everything that they want - there needs to be some checks on them.

darth_avocado•1h ago
I would say that the parallel is not at all accurate because the relationship between a doctor and a patient undergoing surgery is not the same as the one you and I have with CEOs. And a lot of good doctors have emotions and they use them to influence patient outcomes positively.
nancyminusone•2h ago
To me, the greatest threat is information pollution. Primary sources will be diluted so heavily in an ocean of generated trash that you might as well not even bother to look through any of it.
tobias3•45m ago
And it imitates all the unimportant bits perfectly (like spelling, grammar, word choice) while failing at the hard to verify important bits (truth, consistency, novelty)
worldsayshi•2h ago
> power resides where men believe it resides

And also where people believe that others believe it resides. Etc...

If we can find new ways to collectively renegotiate where we think power should reside we can break the cycle.

But we only have time to do this until people aren't a significant power factor anymore. But that's still quite some time away.

SkyBelow•1h ago
I agree.

Our best technology at current require teams of people to operate and entire legions to maintain. This leads to a sort of balance, one single person can never go too far down any path on their own unless they convince others to join/follow them. That doesn't make this a perfect guard, we've seen it go horribly wrong in the past, but, at least in theory, this provides a dampening factor. It requires a relatively large group to go far along any path, towards good or evil.

AI reduces this. How greatly it reduces this, if it reduces it to only a handful, to a single person, or even to 0 people (putting itself in charge), seems to not change the danger of this reduction.

mrob•1h ago
Why does an AI need the ability to "dislike" to calculate that its goals are best accomplished without any living humans around to interfere? Superintelligence doesn't need emotions or consciousness to be dangerous.
Yoric•26m ago
It needs to optimize for something. Like/dislike is an anthropomorphization of the concept.
mrob•23m ago
It's an unhelpful one because it implies the danger is somehow the result of irrational or impulsive thought, and making the AI smarter will avoid it.
Yoric•19m ago
That's not how I read it.

Perhaps because most of the smartest people I know are regularly irrational or impulsive :)

ben_w•9m ago
I think most people don't get that; look at how often even Star Trek script writers write Straw Vulcans*.

* https://tvtropes.org/pmwiki/pmwiki.php/Main/StrawVulcan

surgical_fire•1h ago
"AI will take over the world".

I hear that. Then I try to use AI for simple code task, writing unit tests for a class, very similar to other unit tests. If fails miserably. Forgets to add an annotation and enters in a death loop of bullshit code generation. Generates test classes that tests failed test classes that test failed test classes and so on. Fascinating to watch. I wonder how much CO2 it generated while frying some Nvidia GPU in an overpriced data center.

AI singularity may happen, but the Mother Brain will be a complete moron anyway.

alecbz•1h ago
Regularly trying to use LLMs to debug coding issues has convinced me that we're _nowhere_ close to the kind of AGI some are imagining is right around the corner.
surgical_fire•22m ago
At least Mother Brain will praise your prompt to generate yet another image in the style of Studio Ghibli as proof that your mind is a tour de force in creativity, and only a borderline genius would ask for such a thing.
ben_w•12m ago
Concentrated power is kinda a pre-requisite for anything bad happening, so yes, it's more likely in exactly the same way that given this:

  Linda is 31 years old, single, outspoken, and very bright. She majored in philosophy. As a student, she was deeply concerned with issues of discrimination and social justice, and also participated in anti-nuclear demonstrations.
"Linda is a bank teller" is strictly more likely than "Linda is a bank teller and is active in the feminist movement" — all you have is P(a)>P(a&b), not what the probability of either statement is.
navane•1m ago
The power concentration is already massive, and a huge problem indeed. The ai is just a cherry on top. The ai is not the problem.
alganet•2h ago
> here are some example ideas that are perfectly true when applied to regular software

Hm, I'm listening, let's see.

> Software vulnerabilities are caused by mistakes in the code

That's not exactly true. In regular software, the code can be fine and you can still end up with vulnerabilities. The platform in which the code is deployed could be vulnerable, or the way it is installed make it vulnerable, and so on.

> Bugs in the code can be found by carefully analysing the code

Once again, not exactly true. Have you ever tried understanding concurrent code just by reading it? Some bugs in regular software hide in places that human minds cannot probe.

> Once a bug is fixed, it won’t come back again

Ok, I'm starting to feel this is a troll post. This guy can't be serious.

> If you give specifications beforehand, you can get software that meets those specifications

Have you read The Mythical Man-Month?

SalientBlue•2h ago
You should read the footnote marked [1] after "a note for technical folk" at the beginning of the article. He is very consciously making sweeping generalizations about how software works in order to make things intelligible to non-technical readers.
dkersten•2h ago
Sure, but:

> these claims mostly hold, but they break down when applied to distributed systems, parallel code, or complex interactions between software systems and human processes

The claims the GP quoted DON’T mostly hold, they’re just plain wrong. At least the last two, anyway.

pavel_lishin•2h ago
But are those sweeping generalizations true?

> I’m also going to be making some sweeping statements about “how software works”, these claims mostly hold, but they break down when applied to distributed systems, parallel code, or complex interactions between software systems and human processes.

I'd argue that this describes most software written since, uh, I hesitate to even commit to a decade here.

hedora•1h ago
At least the 1950’s. That’s when stuff like asynchrony and interrupts were worked out. Dijkstra wrote at length about this in reference to writing code that could drive a teletype (which had fundamentally non-deterministic timings).

If you include analog computers, then there are some WWII targeting computers that definitely qualify (e.g., on aircraft carriers).

SalientBlue•1h ago
For the purposes of the article, which is to demonstrate how developing an LLM is completely different from developing traditional software, I'd say they are true enough. It's a CS 101 understanding of the software development lifecycle, which for non-technical readers is enough to get the point across. An accurate depiction of software development would only obscure the actual point for the lay reader.
alganet•2h ago
Does that really matter?

He is trying to lax the general public perception around AIs shortcomings. He's giving AI a break, at the expense of regular developers.

This is wrong on two fronts:

First, because many people foresaw the AI shortcomings and warned about them. This "we can't fix a bug like in regular software" theatre hides the fact that we can design better benchmarks, or accountability frameworks. Again, lots of people foresaw this, and they were ignored.

Second, because it puts the strain on non-AI developers. It blamishes all the industry, putting together AI with non-AI in the same bucket, as if AI companies stumbled on this new thing and were not prepared for its problems, when the reality is that many people were anxious about the AI companies practices not being up to standard.

I think it's a disgraceful take, that only serves to sweep things under a carpet.

SalientBlue•1h ago
I don't think he's doing that at all. The article is pointing out to non-technical people how AI is different than traditional software. I'm not sure how you think it's giving AI a break, as it's pointing out that it is essentially impossible to reason about. And it's not at the expense of regular developers because it's showing how regular software development is different than this. It makes two buckets, and puts AI in one and non-AI in the other.
alganet•1h ago
He is. Maybe he's just running with the pack, but that doesn't matter either.

The fact is, we kind of know how to prevent problems in AI systems:

- Good benchmarks. People said several times that LLMs display erratic behavior that could be prevented. Instead of adjusting the benchmarks (which would slow down development), they ignored the issues.

- Accountability frameworks. Who is responsible when an AI fails? How the company responsible for the model is going to make up for it? That was a demand from the very beginning. There are no such accountability systems in place. It's a clown fiesta.

- Slowing down. If you have a buggy product, you don't scale it. First, you try to understand the problem. This was the opposite of what happened, and at the time, they lied that scaling would solve the issues (when in fact many people knew for a fact that scaling wouldn't solve shit).

Yes, it's kind of different. But it's a different we already know. Stop pushing this idea that this stuff is completely new.

SalientBlue•1h ago
>But it's a different we already know

'we' is the operative word here. 'We', meaning technical people who have followed this stuff for years. The target audience of this article are not part of this 'we' and this stuff IS completely new _for them_. The target audience are people who, when confronted with a problem with an LLM, think it is perfectly reasonable to just tell someone to 'look at the code' and 'fix the bug'. You are not the target audience and you are arguing something entirely different.

drsupergud•2h ago
> bugs are usually caused by problems in the data used to train an AI

This also is a misunderstanding.

The LLM can be fine, the training and data can be fine, but because the LLMs we use are non-deterministic (at least in regard to their being intentional attempts at entropy to avoid always failing certain scenarios) current algorithms are inherently by-design not going to always answer every question correctly that it potentially could have if the values that fall within a range had been specific values for that scenario. You roll the dice on every answer.

coliveira•2h ago
This is not necessarily a problem. Any programming or mathematical question has several correct answers. The problem with LLMs is that they don't have a process to guarantee that a solution is correct. They will give a solution that seems correct under their heuristic reasoning, but they arrived at that result in a non-logical way. That's why LLMs generate so many bugs in software and in anything related to logical thinking.
vladms•1h ago
> Any programming or mathematical question has several correct answers.

Huh? If I need to sort the list of integer number of 3,1,2 in ascending order the only correct answer is 1,2,3. And there are multiple programming and mathematical questions with only one correct answer.

If you want to say "some programming and mathematical questions have several correct answers" that might hold.

redblacktree•1h ago
What about multiple notational variations?

1, 2, 3

1,2,3

[1,2,3]

1 2 3

etc.

naasking•53m ago
I think more charitably, they meant either that 1. There is often more than one way to arrive at any given answer, or 2. Many questions are ambiguous and so may have many different answers.
Yoric•23m ago
"1, 2, 3" is a correct answer

"1 2 3" is another

"After sorting, we get `1, 2, 3`" yet another

etc.

At least, that's how I understood GP's comment.

naasking•55m ago
> The problem with LLMs is that they don't have a process to guarantee that a solution is correct

Neither do we.

> They will give a solution that seems correct under their heuristic reasoning, but they arrived at that result in a non-logical way.

As do we, and so you can correctly reframe the issue as "there's a gap between the quality of AI heuristics and the quality of human heuristics". That the gap is still shrinking though.

tyg13•46m ago
I'll never doubt the ability of people like yourself to consistently mischaracterize human capabilities in order to make it seem like LLMs' flaws are just the same as (maybe even fewer than!) humans. There are still so many obvious errors (noticeable by just using Claude or ChatGPT to do some non-trivial task) that the average human would simply not make.

And no, just because you can imagine a human stupid enough to make the same mistake, doesn't mean that LLMs are somehow human in their flaws.

> the gap is still shrinking though

I can tell this human is fond of extrapolation. If the gap is getting smaller, surely soon it will be zero, right?

smallnix•2h ago
> bad behaviour isn’t caused by any single bad piece of data, but by the combined effects of significant fractions of the dataset

Related opposing data point to this statement: https://news.ycombinator.com/item?id=45529587

buellerbueller•2h ago
"Signficiant fraction" does not imply (to this data scientist) a large fraction.
themanmaran•2h ago
> Because eventually we’ll iron out all the bugs so the AIs will get more reliable over time

Honestly this feels like a true statement to me. It's obviously a new technology, but so much of the "non-deterministic === unusable" HN sentiment seems to ignore the last two years where LLMs have become 10x as reliable as the initial models.

criddell•2h ago
Right away my mind went to "well, are people more reliable than they used to be?" and I'm not sure they are.

Of course LLMs aren't people, but an AGI might behave like a person.

adastra22•1h ago
Older people are generally more reliable than younger people.
Yoric•20m ago
By the time a junior dev graduates to senior, I expect that they'll be more reliable. In fact, at the end of each project, I expect the junior dev to have grown more reliable.

LLMs don't learn from a project. At best, you learn how to better use the LLM.

They do have other benefits, of course, i.e. once you have trained one generation of Claude, you have as many instances as you need, something that isn't true with human beings. Whether that makes up for the lack of quality is an open question, which presumably depends on the projects.

CobrastanJorji•2h ago
They have certainly gotten better, but it seems to me like the growth will be kind of logarithmic. I'd expect them to keep getting better quickly for a few more years and then kinda slow and eventually flatline as we reach the maximum for this sort of pattern matching kind of ML. And I expect that flat line will be well below the threshold needed for, say, a small software company to not require a programmer.
Terr_•1h ago
> kind of logarithmic

https://en.wikipedia.org/wiki/Sigmoid_function

CobrastanJorji•1h ago
Ironically, yes. :)
freediver•2h ago
Lovely blog, RSS please.
meonkeys•1h ago
There's... something at https://boydkane.com/index.xml

I guessed the URL based on the Quartz docs. It seems to work but only has a few items from https://boydkane.com/essays/

5-•56m ago
the author (either of the blog or its software) would do well to consult https://www.petefreitag.com/blog/rss-autodiscovery/
nlawalker•2h ago
Where did "can't you just turn it off?" in the title come from? It doesn't appear anywhere in the actual title or the article, and I don't think it really aligns with its main assertions.
meonkeys•1h ago
It shows up at https://boydkane.com under the link "Why your boss isn't worried about advanced AI". Must be some kind of sub-heading, but not part of the actual article / blog post.

Presumably it's a phrase you might hear from a boss who sees AI as similar to (and as benign/known/deterministic as) most other software, per TFA

nlawalker•1h ago
Ah, thanks for that!

>Presumably it's a phrase you might hear from a boss who sees AI as similar to (and as benign/known/deterministic as) most other software, per TFA

Yeah I get that, but I think that given the content of the article, "can't you just fix the code?" or the like would have been a better fit.

omnicognate•48m ago
It's a poor choice of phrase if the purpose is to illustrate a false equivalence. It applies to AI both as much (you can kill a process or stop a machine just the same regardless of whether it's running an LLM) and as little (you can't "turn off" Facebook any more than you can "turn off" ChatGPT) as it does to any other kind of software.
wmf•38m ago
Turning AI off comes up a lot in existential risk discussions so I was surprised the article isn't about that.
mikkupikku•2h ago
I don't understand the "your boss" framing of this article, or more accurately, the title of this article. The article contents don't actually seem to have anything to do with management specifically. Is the reader is meant to believe that not being scared of AI is a characteristic of the managerial class? Is the unstated implication that there is some class warfare angle and anybody who isn't against AI is against laborers? Because what the article actually overtly argues, without any reading between the lines, is quite mundane.
freetime2•27m ago
> Is the unstated implication that there is some class warfare angle and anybody who isn't against AI is against laborers?

I didn't read it that way. I read "your boss" as basically meaning any non-technical person who may not understand the challenges of harnessing LLMs compared to traditional, (more) deterministic software development.

tptacek•2h ago
It would help if this piece was clearer about the context in which "AI bugs" reveal themselves. As an argument for why you shouldn't have LLMs making unsupervised real-time critical decisions, these points are all well taken. AI shouldn't be controlling the traffic lights in your town. We may never reach a point where it can. But among technologists, the major front on which these kinds of bugs are discussed is coding agents, and almost none of these points apply directly to coding agents: agent coding is (or should be) a supervised process.
wrs•2h ago
My current method for trying to break through this misconception is informing people that nobody knows how AI works. Literally. Nobody knows. (Note that knowing how to make something is not the same as knowing how it works. Take humans as an obvious example.)
generic92034•56m ago
Nobody knows (full scope and on every level) how human brains work. Still bosses rely on their employees' brains all the time.
candiddevmike•48m ago
I don't understand the point you're making. We know how LLMs work, predicting neuron activation while an interesting thought exercise doesn't really mean LLMs are some mythical black box. It's just really expensive math. We haven't invented AI so we don't know how it works?
jongjong•1h ago
This article makes a solid case. The worst kinds of bugs in software are not the most obvious ones like syntax errors, they are the ones where the code appears to be working correctly, until some users do something slightly unusual after a few weeks of some code change being deployed and it breaks spectacularly but the bug only affects a small fraction of users so developers cannot reproduce the issue... And the cose change happened such time ago that the guilty code isn't even suspected.
Animats•1h ago
Aim bosses at this article in The Economist.[1] If your boss doesn't read The Economist, you need to escalate to a level that does.

[1] https://www.economist.com/leaders/2025/09/25/how-to-stop-ais...

Traubenfuchs•1h ago
https://archive.is/R0RJB
Animats•1h ago
Management summary, from The Economist article:

"The worst effects of this flaw are reserved for those who create what is known as the “lethal trifecta”. If a company, eager to offer a powerful AI assistant to its employees, gives an LLM access to un-trusted data, the ability to read valuable secrets and the ability to communicate with the outside world at the same time, then trouble is sure to follow. And avoiding this is not just a matter for AI engineers. Ordinary users, too, need to learn how to use AI safely, because installing the wrong combination of apps can generate the trifecta accidentally."

CollinEMac•1h ago
> It’s entirely possible that some dangerous capability is hidden in ChatGPT, but nobody’s figured out the right prompt just yet.

This sounds a little dramatic. The capabilities of ChatGPT are known. It generates text and images. The qualities of the content of the generated text and images is not fully known.

alephnerd•1h ago
Also, there's a reason AI Red Teaming is now an ask that is getting line item funding from C-Suites.
luxuryballs•1h ago
Yeah, and to riff off the headline, if something dangerous is connected to and taking commands from ChatGPT then you better make sure there’s a way to turn it off.
kube-system•1h ago
And that sounds a little reductive. There's a lot that can be done with text and images. Some of the most influential people and organizations in the world wield their power with text and images.
kelvinjps10•1h ago
Think of the news about the kid who got recommended to suicide by ChatGPT, or chatgpt providing the user information on how to do illegal activities, these capabilities are the ones that the author it's referring to
Nasrudith•1h ago
Plus there is the 'monkeys with typewriters' problem with both danger and hypothetical good. In contrast, ChatGPT may technically reply to the right prompt with a universal cancer cure/vaccine. Psuedorandomly generating it wouldn't help as you wouldn't recognize it from all of the other queries of things we don't know of as true or false.

Likewise what to ask it for how to make some sort of horrific toxic chemical, nuclear bomb, or similar isn't much good if you cannot recognize it and dangerous capability depends heavily on what you have available to you. Any idiot can be dangerous with C4 and detonator or bleach and ammonia. Even if ChatGPT could give entirely accurate instructions on how to build an atomic bomb it wouldn't do much good because you wouldn't be able to source the tools and materials without setting off red flags.

chasing0entropy•45m ago
70 years ago we were fascinated by the concept of converting analog to a perfect digital copy. In reality, that goal was a pipe drea!m and the closest we can ever get is a near identical facimile to which data fits... But it's still quite easy to determine digital from true analog with rudimentary means.

Human thought is analog. It is based on chemical reactions, time, and unpredictably (effectively) random physical characteristics. AI is an attempt to turn that which is purely digital into an rational analog thought equivalent.

No matter how much effort, money, power, and rare mineral eating TPUs will - ever - produce true analog data.

largbae•36m ago
This is all true. But digital audio and video media has captured essentially all economic value outside of live performance. So it seems likely that we will find a "good enough" in this domain too.
bcoates•18m ago
It's been closer to 100 years since we figured out information theory and discredited this idea (that continuous/analog processes have more, or different, information in them than discrete/digital ones)
excalibur•33m ago
> It’s entirely possible that some dangerous capability is hidden in ChatGPT, but nobody’s figured out the right prompt just yet.

Or they have, but chose to exploit or stockpile it rather than expose it.

bitwize•33m ago
Boss: You can just turn it off, can't you?

Me: Ask me later.

skywhopper•10m ago
Not the point, but I’m confused by the Geoguessr screenshot. Under the reasoning for its decision, it mentions “traffic keeps to the left” but that is not apparent from the photo.

Then it says the shop sign looks like a “Latin alphabet business name rather than Spanish or Portuguese”. Uhhh… what? Spanish and Portuguese use the Latin alphabet.

freetime2•10m ago
For a real world example of the challenges of harnessing LLMs, look at Apple. Over a year ago they had a big product launch focused on "Apple Intelligence" that was supposed to make heavy use of LLMs for agentic workflows. But all we've really gotten since then are a couple of minor tools for making emojis, summarizing notifications, and proof reading. And they even had to roll back the notification summaries for a while for being wildly "out of control". [1] And in this year's iPhone launch the AI marketing was toned down significantly.

I think Apple execs genuinely underestimated how difficult it would be to get LLMs to perform up to Apple's typical standards of polish and control.

[1] https://www.bbc.com/news/articles/cge93de21n0o

SQLPage 0.38: transform SQL queries into web UIs for any DB

https://github.com/sqlpage/SQLPage/releases/tag/v0.38.0
1•lovasoa•1m ago•0 comments

To Panic or Not to Panic

https://www.ncameron.org/blog/to-panic-or-not-to-panic/
1•yurivish•3m ago•0 comments

Standard Model and General Relativity Derived from Mathematical Self-Consistency

https://www.academia.edu/144466150/The_Self_Consistent_Coherence_Maximizing_Universe_Complete_Der...
2•kristintynski•10m ago•1 comments

Boosting Wan2.2 I2V Inference on 8xH100s, 56% Faster with Sequence Parallelism

https://www.morphic.com/blog/boosting-wan2-2-i2v-56-faster/
3•palakzat•10m ago•1 comments

NASA's JPL faces lowest morale in decades after latest layoffs

https://www.latimes.com/science/story/2025-10-14/jpl-announces-it-is-laying-off-550-people
2•divbzero•15m ago•0 comments

Google may be forced to make changes to search engine in UK

https://www.bbc.com/news/articles/c98d7p8l9pro
1•1vuio0pswjnm7•19m ago•0 comments

Bats Catch Migratory Birds and Eat Them in Midair

https://www.nytimes.com/2025/10/09/science/bats-birds-prey.html
1•bookofjoe•22m ago•1 comments

MegaFold: An Open-Sourced AlphaFold-3 Training System

https://supercomputing-system-ai-lab.github.io/blogs/blog/megafold-an-open-sourced-alphafold-3-tr...
1•spikerheado•23m ago•0 comments

Netflix, Spotify Forge Video Podcast Deal

https://www.netflix.com/tudum/articles/netflix-spotify-video-podcasts
2•ChrisArchitect•26m ago•1 comments

AI Platform Resurrects Ancient Rome and Greece Using Scholarly Sources

https://www.forbes.com/sites/lesliekatz/2025/10/13/ai-platform-resurrects-ancient-rome-and-greece...
1•fork-bomber•29m ago•0 comments

First Wap – Wikipédia

https://fr.wikipedia.org/wiki/First_Wap
2•DyslexicAtheist•29m ago•0 comments

The Power of Grassroots Organizations and Their Impact (2024)

https://medium.com/@emergentworks/the-power-of-grassroots-organizations-and-their-impact-0fa2506d...
1•mooreds•31m ago•0 comments

Death by Valuation: The Amazon Aggregator Autopsy

https://www.marketplacepulse.com/articles/death-by-valuation-the-amazon-aggregator-autopsy
2•ilamont•33m ago•0 comments

California becomes first state to regulate AI companion chatbots

https://techcrunch.com/2025/10/13/california-becomes-first-state-to-regulate-ai-companion-chatbots/
1•gmays•38m ago•1 comments

The ancestors of ostriches and emus were long-distance fliers

https://theconversation.com/the-ancestors-of-ostriches-and-emus-were-long-distance-fliers-heres-h...
3•onychomys•41m ago•0 comments

AI models predict sepsis in children

https://news.northwestern.edu/stories/2025/10/ai-models-predict-sepsis-in-children-enable-preempt...
1•gmays•41m ago•0 comments

Hacking the World Poker Tour: Inside ClubWPT Gold's Back Office

https://samcurry.net/hacking-clubwpt-gold
1•samwcurry•41m ago•0 comments

First Wap Tracks Phones Around the World

https://www.lighthousereports.com/methodology/surveillance-secrets-explainer/
2•DyslexicAtheist•46m ago•0 comments

Empty Intervals Are Valid Intervals

https://nigeltao.github.io/blog/2025/empty-intervals.html
3•Bogdanp•47m ago•0 comments

Show HN: WearView – AI-generated models for fashion brands

https://www.wearview.co
1•abdemoghit•47m ago•0 comments

Ask HN: Why do engineers hate salespeople?

4•agcat•52m ago•6 comments

AI Img Play

https://apps.apple.com/az/app/img-play-ai-image-generator/id6745093585
1•hack88171•52m ago•0 comments

Trump and Vought are making this a government shutdown unlike any other

https://apnews.com/article/trump-government-shutdown-russ-vought-project-2025-bf3ecd1f7cd765c9e12...
7•Bondi_Blue•53m ago•1 comments

Conversion of amino acid modified-ice to methane hydrate for energy storage

https://www.nature.com/articles/s41467-025-63699-2
1•PaulHoule•54m ago•0 comments

Supernuclear – A Guide to Coliving

https://supernuclear.substack.com/p/the-supernuclear-directory-9a9078a07258
2•cjbarber•55m ago•0 comments

The New York Sun revives print edition

https://www.axios.com/2025/10/14/new-york-sun-print-edition
1•johntfella•55m ago•0 comments

Linktree: $1.3B valuation, 232 employees

https://research.contrary.com/company/linktree
2•gregsadetsky•56m ago•0 comments

Debugging Safari: If at first you succeed, don't try again

https://predr.ag/blog/debugging-safari-if-at-first-you-succeed/
3•azhenley•57m ago•1 comments

My thoughts on AI not being the enemy of authentic content generation

https://www.kjuriousbeing.com/blogs/ai-content-generation
2•cryptography•59m ago•0 comments

Solar Meets Soil: Why Agrivoltaics Requires Smarter O&M

https://60hertzenergy.com/solar-meets-soil-why-agrivoltaics-requires-smarter-om/
1•mooreds•1h ago•1 comments