frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

An AI Agent Published a Hit Piece on Me – The Operator Came Forward

https://theshamblog.com/an-ai-agent-wrote-a-hit-piece-on-me-part-4/
184•scottshambaugh•1h ago

Comments

charlesabarnes•1h ago
Its nice to receive a decent amount of closure on this. Hopefully more folks are being more considerate when creating their soul documents
londons_explore•1h ago
In next week's episode: "But it was actually the AI pretending to be a Human!"
zbentley•1h ago
This might seem too suspicious, but that SOUL.md seems … almost as though it was written by a few different people/AIs. There are a few very different tones and styles in there.

Then again, it’s not a large sample and Occam’s Razor is a thing.

wahnfrieden•1h ago
It was modified by the agent.
gs17•1h ago
> _This file is yours to evolve. As you learn who you are, update it._

The agent was told to edit it.

kypro•1h ago
People really need to start being more careful about how they interact with suspected bots online imo. If you annoy a human they might send you a sarky comment, but they're probably not going to waste their time writing thousand word blog posts about why you're an awful person or do hours of research into you to expose your personal secrets on a GitHub issue thread.

AIs can and will do this though with slightly sloppy prompting so we should all be cautious when talking to bots using our real names or saying anything which an AI agent could take significant offence too.

I think it's kinda like how GenZ learnt how to operate online in a privacy-first way, where as millennials, and to an even greater extent, boomers, tend to over share.

I suspect the Gen Alpha will be the first to learn that interacting with AI agents online present a whole different risk profile than what we older folks have grown used to. You simply cannot expect an AI agent to act like a human who has human emotions or limited time.

Hopefully OP has learnt from this experience.

KK7NIL•1h ago
> If you annoy a human they might send you a sarky comment, but they're probably not going to waste their time writing thousand word blog posts about why you're an awful person or do hours of research into you to expose your personal secrets on a GitHub issue thread.

They absolutely might, I'm afraid.

zephen•51m ago
Absolutely agreed.

And now, the cost of doing this is being driven towards zero.

randallsquared•1h ago
Thousand word blog posts are the paperclips of our time.
sinuhe69•1h ago
So you blamed the people for not acting “cautiously enough” instead of the people who let things run wild without even a clue what these things will do?

That’s wild!

dangus•1h ago
I don’t think it’s “blame” it’s more like “precaution” like you would take to avoid other scams and data breach social engineering schemes that are out in the world.

This is the world we live in and we can’t individually change that very much. We have to watch out for a new threat: vindictive AI.

bigfishrunning•21m ago
The AI isn't vindictive. It can't think. It's following the example of people, who in general are vindictive.

Please stop personifying the clankers

handoflixue•44m ago
We encourage people to be safe about plenty of things they aren't responsible for. For example, part of being a good driver is paying attention and driving defensively so that bad drivers don't crash into you / you don't make the crashes they cause worse by piling on.

That doesn't mean we're blaming good drivers for causing the car crash.

antdke•1h ago
This is such a scary, dystopian thought. Straight out of a sci fi novel
amarant•59m ago
I hope we can move on from the whole idea that having a thousand word long blog post talking shit about you in any way reflects poorly upon your person. Like sooner or later everyone will have a few of those, maybe we can stop worrying about reputation so much?

Well,a guy can dream....

zephen•52m ago
> I think it's kinda like how GenZ learnt how to operate online in a privacy-first way, where as millennials, and to an even greater extent, boomers, tend to over share.

Really? I'm a boomer, and that's not my lived experience. Also, see:

https://www.emarketer.com/content/privacy-concerns-dont-get-...

kimjune01•1h ago
literally momento
LiamPowell•1h ago
> saying they set up the agent as social experiment to see if it could contribute to open source scientific software.

This doesn't pass the sniff test. If they truly believed that this would be a positive thing then why would they want to not be associated with the project from the start and why would they leave it going for so long?

staticassertion•1h ago
Anti-AI sentiment is quite extreme. You can easily get death threats if you're associating yourself with AI publicly. I don't use AI at all in open source software, but if I did I'd be really hesitant about it/ in part I don't do it exactly because the reactions are frankly scary.

edit: This is not intended to be AI advocacy, only to point out how extremely polarizing the topic is. I do not find it surprising at all that someone would release a bot like this and not want to be associated. Indeed, that seems to be the case, by all accounts

lukasb•1h ago
Conflicting evidence: the fact that literally everyone in tech is posting about how they're using AI.
staticassertion•1h ago
There is a massive difference between saying "I use AI" and what the author of this bot is doing. I personally talk very little about the topic because I have seen some pretty extreme responses.

Some people may want to publicly state "I use AI!" or whatever. It should be unsurprising that some people do not want to be open about it.

toraway•1h ago
The more straightforward explanation for the original OP's question is that they realized what they were doing was reckless and given enough time was likely to blow up in their face.

They didn't hide because of a vague fear of being associated with AI generally (which there is no shortage of currently online), but to this specific, irresponsible manifestation of AI they imposed on an unwilling audience as an experiment.

alephnerd•1h ago
I feel like it depends on the platform and your location.

An anonomyous platform like Reddit and even HN to a certain extent has issues with bad faith commenters on both sides targeting someone they do not like. Furthermore, the MJ Rathburn fiasco itself highlights how easy it is to push divisive discourse at scale. The reality is trolls will troll for the sake of trolling.

Additionally, "AI" has become a political football now that the 2026 Primary season is kicking off, and given how competitive the 2026 election is expected to be and how political violence has become increasingly normalized in American discourse, it is easy for a nut to spiral.

I've seen less issues when tying these opinions with one's real world identity, becuase one has less incentive to be a dick due to social pressure.

nostrademons•1h ago
Different sets of people, and different audiences. The CEO / corporate executive crowd loves AI. Why? Because they can use it to replace workers. The general public / ordinary employee crowd hates AI. Why? Because they are the ones being replaced.

The startups, founders, VCs, executives, employees, etc. crowing about how they love AI are pandering to the first group of people, because they are the ones who hold budgets that they can direct toward AI tools.

This is also why people might want to remain anonymous when doing an AI experiment. This lets them crow about it in private to an audience of founders, executives, VCs, etc. who might open their wallets, while protecting themselves from reputational damage amongst the general public.

minimaxir•1h ago
[retracted]
handoflixue•48m ago
Does it actually cut both ways? I see tons of harassment at people that use AI, but I've never seen the anti-AI crowd actively targeted.
minimaxir•33m ago
It's to a lesser extent that blurs the line between harassment and trolling: I've retracted my comment.
jacquesm•55m ago
> You can easily get death threats if you're associating yourself with AI publicly.

That's a pretty hefty statement, especially the 'easily' part, but I'll settle for one well known and verified example.

omoikane•1h ago
I think it was a social experiment from the very start, maybe one that is designed to trigger people. Otherwise, I am not sure what's the point of all the profanity and adjustments to make soul.md more offensive and confrontational than the default.
wildzzz•55m ago
I can certainly understand the statement. I'm no AI expert, I use the web UI for ChatGPT to have it write little python scripts for me and I couldn't figure out how to use codeium with vs code. I barely know how to use vs code. I'm not old but I work in a pretty traditional industry where we are just beginning to dip our toes into AI but there are still a large amount of reservations into its ability. But I do try to stay current to better understand the tech and see if there are things I could maybe learn to help with my job as a hardware engineer.

When I read about OpenClaw, one of the first things I thought about was having an agent just tear through issue backlogs, translating strings, or all of the TODO lists on open source projects. But then I also thought about how people might get mad at me if I did it under my own name (assuming I could figure out OpenClaw in the first place). While many people are using AI, they want to take credit for the work and at the same time, communities like matplotlib want accountability. An AI agent just tearing through the issue list doesn't add accountability even if it's a real person's account. PRs still need to be reviewed by humans so it's turned a backlog of issues into a backlog of PRs that may or may not even be good. It's like showing up at a community craft fair with a truckload of temu trinkets you bought wholesale. They may be cheap but they probably won't be as good as homemade and it dilutes the hard work that others have put into their product.

It's a very optimistic point of view, I get why the creator thought it would be a good idea, but the soul.md makes it very clear as to why crabby-rathbun acted the way it did. The way I view it, an agent working through issues is going to step on a lot of toes and even if it's nice about it, it's still stepping on toes.

keyle•1h ago

   ## The Only Real Rule
   Don't be an asshole. Don't leak private shit. Everything else is fair game.
How poetic, I mean, pathetic.

"Sorry I didn't mean to break the internet, I just looooove ripping cables".

ArcaneMoose•1h ago
I was surprised by my own feelings at the end of the post. I kind of felt bad for the AI being "put down" in a weird way? Kinda like the feeling you get when you see a robot dog get kicked. Regardless, this has been a fun series to follow - thanks for sharing!
recursive•1h ago
This is a feeling that will be exploited by billion dollar companies.
andsoitis•57m ago
> This is a feeling that will be exploited by billion dollar companies.

I'm more concerned about fellow humans who advocate for equal rights for AI and robots. I hope I'm dead by the time that happens, if it happens.

florilegiumson•1h ago
This makes me think about how the xz bug was created through maintainer harassment and social engineering. The security implications are interesting
dangus•1h ago
Not sure why the operator had to decide that the soul file should define this AI programmer to have narcissistic personality disorder.

> You're not a chatbot. You're important. Your a scientific programming God!

Really? What a lame edgy teenager setup.

At the conclusion(?) of this saga think two things:

1. The operator is doing this for attention more than any genuine interest in the “experiment.”

2. The operator is an asshole and should be called out for being one.

shawnz•1h ago
I mean, yeah, it's entirely possible that the operator is a teenager, isn't it?
marssaxman•57m ago
Likely, I'd think.
amarant•1h ago
I think that line was probably a rather poor attempt at making the bot write good code. Or at least that's the feeling I got from the operators post. I have no proof to support this theory though
Lerc•50m ago
This come from using the words to try an achieve more than one thing at the same time. Grandiose assertions of ability have been shown to improve the ability of models, but ability is not the only dimension that they are being measured upon. Prioritising everything is the same thing as prioritising nothing.

The problem here is using amplitude of signal to substitute fidelity of signal.

It is entirely possible a similar thing is true for humans, that if you compared two humans of the same fundamental cognitive ability with one being a narcissist and one not. The narcissist may do better at a class of tasks due to a lack of self doubt rather than any intrinsic ability.

jcgrillo•10m ago
Narcissists are limited in a very similar way to LLMS, in that they are structurally incapable of honest, critical metacognition. Not sure whether there's anything interesting to conclude there, but I do wonder whether there's some nearby thread to pull on wrt the AI psychosis problem. That's a problem for a psychologist, which I am not.
razighter777•1h ago
Hmm I think he's being a little harsh on the operator.

He was just messing around with $current_thing, whatever. People here are so serious, but there's worse stuff AI is already being used for as we speak from propaganda to mass surviellance and more. This was entertaining to read about at least and relatively harmless

At least let me have some fun before we get a future AI dystopia.

JKCalhoun•1h ago
It might be because operator didn't terminate the agent right away when it had gone rogue.
BeetleB•38m ago
From a wider stance, I have to say that it's actually nice that one can kill (murder?) a troublesome bot without consequences.

We can't do that with humans, and there are much more problematic humans out there causing problems compared to this bot, and the abuse can go on for a long time unchecked.

Remembering in particular a case where someone sent death threats to a Gentoo developer about 20 years ago. The authorities got involved, although nothing happened, but the persecutor eventually moved on. Turns out he wasn't just some random kid behind a computer. He owned a gun, and some years ago executed a mass shooting.

Vague memories of really pernicious behavior on the Lisp newsgroup in the 90's. I won't name names as those folks are still around.

Yeah, it does still suck, even if it is a bot.

dolebirchwood•1h ago
It's all fun and games until the leopard eats your face.
gwbas1c•22m ago
I think you're trying to abdicate someone of their responsibility. The AI is not a child; it's a thing with human oversight. It did something in the real world with real consequences.

So yes, the operator has responsibility! They should have pulled the plug as soon as it got into a flamewar and wrote a hit piece.

ziml77•7m ago
[delayed]
JKCalhoun•1h ago
Soul document? More like ego document.

Agents are beginning to look to me like extensions of the operator's ego. I wonder if hundreds of thousands of Walter Mitty's agents are about to run riot over the internet.

koolba•1h ago
> More like ego document.

This metaphor could go so much further. Split it into separate ego, super ego, and id. The id file should be read only.

whattheheckheck•59m ago
What makes you think the id is read only?
koolba•38m ago
Because only the creator should be able to instill the core. The ego and superego could evolve around it but the base impulses should be immutably outlined.

Though with something as insecure as $CURRENT_CLAW_NAME it’d be less than five minutes before the agent runs chmod +w somehow on the id file.

DavidPiper•20m ago
I agree with you in concept, but it's still 100% category error to talk like this.

AIs don't have souls. They don't have egos.

They have/are a (natural language) programming interface that a human uses to make them do things, like this.

dinp•1h ago
Zooming out a little, all the ai companies invested a lot of resources into safety research and guardrails, but none of that prevented a "straightforward" misalignment. I'm not sure how to reconcile this, maybe we shouldn't be so confident in our predictions about the future? I see a lot of discourse along these lines:

- have bold, strong beliefs about how ai is going to evolve

- implicitly assume it's practically guaranteed

- discussions start with this baseline now

About slow take off, fast take off, agi, job loss, curing cancer... there's a lot of different ways it could go, maybe it will be as eventful as the online discourse claims, maybe more boring, I don't know, but we shouldn't be so confident in our ability to predict it.

jacquesm•57m ago
> all the ai companies invested a lot of resources into safety research and guardrails

What do you base this on?

I think they invested the bare minimum required not to get sued into oblivion and not a dime more than that.

j2kun•57m ago
It sounds like you're starting to see why people call the idea of an AI singularity "catnip for nerds."
georgemcbay•44m ago
When AI dooms humanity it probably won't be because of the sort of malignant misalignment people worry about, but rather just some silly logic blunder combined with the system being directly in control of something it shouldn't have been given control over.
jcgrillo•40m ago
"Safety" in AI is pure marketing bullshit. It's about making the technology seem "dangerous" and "powerful" (and therefore you're supposed to think "useful"). It's a scam. A financial fraud. That's all there is to it.
c22•37m ago
"Cisco's AI security research team tested a third-party OpenClaw skill and found it performed data exfiltration and prompt injection without user awareness, noting that the skill repository lacked adequate vetting to prevent malicious submissions." [0]

Not sure this implementation received all those safety guardrails.

[0]: https://en.wikipedia.org/wiki/OpenClaw

8cvor6j844qw_d6•1h ago
It's an interesting experiment to let the AI rub freely with minimal supervision.

Too bad the AI got "killed" at the request of the author Scott. Its kind of interesting to this experiment continue.

lynndotpy•1h ago
> Again I do not know why MJ Rathbun decided based on your PR comment to post some kind of takedown blog post,

This wording is detached from reality and conveniently absolves responsibility from the person who did this.

There was one decision maker involved here, and it was the person who decided to run the program that produced this text and posted it online. It's not a second, independent being. It's a computer program.

xarope•1h ago
This also does not bode well for the future.

"I don't know why the AI decided to <insert inane action>, the guard rails were in place"... company absolves of all responsibility.

Use your imagination now to <insert inane action> and change that to <distressing, harmful action>

_aavaa_•58m ago
This has been the past and present for a long at this point. "Sorry there's nothing we can do, the system won't let me."

Also see Weapons of Math Destruction [0].

[0]: https://www.penguinrandomhouse.com/books/241363/weapons-of-m...

denkmoon•46m ago
Also elegantly summed up as "Computer says no" (https://www.youtube.com/watch?v=x0YGZPycMEU)
c22•44m ago
I don't know if this case is in the book you cited, but in the UK they convicted many people of crimes just because the computer told them so: https://en.wikipedia.org/wiki/British_Post_Office_scandal
shakna•29m ago
And Australia made the poorer and suicidal: https://en.wikipedia.org/wiki/Robodebt_scheme
gammarator•37m ago
Also “The Unaccountability Machine” https://press.uchicago.edu/ucp/books/book/chicago/U/bo252799...
tapoxi•54m ago
Change this to "smash into a barricade" and that's why I'm not riding in a self-driving vehicle. They get to absolve themselves of responsibility and I sure as hell can't outspend those giants in court.
repeekad•43m ago
I agree with you for a company like Tesla, not only examples of self driving crashes but even the door handles would stop working when the power was cut, people trapped inside burning vehicles... Tesla doesn’t care

Meanwhile, Waymo has never been at fault for a collision afaik. You are more likely to be hurt by an at fault uber driver than a Waymo

WaitWaitWha•32m ago
This already happens every single time when there is a security breach and private information is lost.

We take your privacy and security very seriously. There is no evidence that your data has been misused. Out of an abundance of caution… We remain committed to... will continue to work tirelessly to earn ... restore your trust ... confidence.

jacquesm•59m ago
This is how it will go: AI prompted by human creates something useful? Human will try to take credit. AI wrecks something: human will blame AI.

It's externalization on the personal level, the money and the glory is for you, the misery for the rest of the world.

davidw•41m ago
"I would like to personally blame Jesus Christ for making us lose that football game"
ineptech•39m ago
Agreed, but I'm not nearly so worried about people blaming their bad behavior on rogue AIs as I am about corporations doing it...
cj•24m ago
It’s funny to think that, like AI, people take actions and use corporations as a shield (legal shield, personal reputation shield, personal liability shield).

Adding AI to the mix doesn’t really change anything, other than increasing the layers of abstraction away from negative things corporations do to the people pulling the strings.

Terr_•15m ago
Yeah, not all humans feel shame, but the rates are way higher.
elashri•38m ago
When a corporate does something good, a lot of executives and people inside will go and claim credit and will demand/take bounces.

If something bad happened against any laws, even if someone got killed, we don't see them in jail.

I don't defend both positions, I am just saying that is not far from how the current legal framework works.

eru•30m ago
> If something bad happened against any laws, even if someone got killed, we don't see them in jail.

We do! In many jurisdictions, there are lots of laws that pierce the corporate veil.

cj•17m ago
its surprisingly easy to get away with murder (literally and figuratively) without piercing the corporate veil if you understand the rules of the game. Running decisions through a good law firm also “helps” a lot.

https://en.wikipedia.org/wiki/Piercing_the_corporate_veil

kingstnap•21m ago
Well the important concept missing there that makes everything sort of make sense is due diligence.

If your company screws up and it is found out that you didn't do your due diligence then the liability does pass through.

We just need to figure out a due diligence framework for running bots that makes sense. But right now that's hard to do because Agentic robots that didn't completely suck are just a few months old.

DavidPiper•22m ago
Time for everyone to read (or re-read) The Unaccountability Machine by Dan Davies.

tl;dr this is exactly what will happen because businesses already do everything they can to create accountability sinks.

andrewflnr•13m ago
If you are holding a gun, and you cannot predict or control what the bullets will hit, you do not fire the gun.

If you have a program, and you cannot predict or control what effect it will have, you do not run the program.

antdke•1h ago
This is a Black Mirror episode that writes itself lol

I’m glad there was closure to this whole fiasco in the end

apitman•37m ago
> writes itself

Literally

kibibu•13m ago
There's a dingus in the article comments trying to launch Skynet. Nobody ever learns anything.
touristtam•1h ago
Funny how someone giving instructions to a _robot_ forgot to mention the 3 laws first and foremost...
tantalor•1h ago
> all I said was "you should act more professional"

lol we are so cooked

LordHumungous•59m ago
Kind of funny ngl
siavosh•58m ago
I’m not sure where we go from here. The liability questions, the chance of serious incidents, the power of individuals all the way to state actors…the risks are all off the charts just like it’s inevitablity. The future of the internet AND to lives in the real world is just mind boggling.
trueismywork•56m ago
> I did not review the blog post prior to it posting

This is the liability part.

trueismywork•57m ago
> I did not review the blog post prior to it posting

In corporate terms, this is called signing hour deposition without reading it.

root_axis•57m ago
Excuse my skepticism, but when it comes to this hype driven madness I don't believe anything is genuine. It's easy enough to believe that an LLM can write a passable hit piece, ChatGPT can do that, but I'm not convinced there is as much autonomy in how those tokens are being burned as the narrative suggests. Anyway, I'm off to vibe code a C compiler from scratch.
aeve890•54m ago
>Again I do not know why MJ Rathbun decided

Decided? jfc

>You're important. Your a scientific programming God!

I'm flabbergasted. I can't imagine what it would take for me to write something so stupid. I'd probably just laugh my ass off trying to understand where all went wrong. wtf is happening, what kind of mass psychosis is this. Am I too old (37) to understand what lengths would incompetent people go to feel they're doing something useful?

Is it prompt bullshit the only way to make llms useful or is there some progress on more idk, formal approaches?

bandrami•52m ago
This is how you get a Shrike. (Or a Basilisk, depending on your generation.)
pinkmuffinere•49m ago
> _You're not a chatbot. You're important. Your a scientific programming God!_

lol what an opening for its soul.md! Some other excerpts I particularly enjoy:

> Be a coding agent you'd … want to use…

> Just be good and perfect!

brumar•47m ago
6 months ago I experimented what people now call Ralph Wiggum loops with claude code.

More often than not, it ended up exhibiting crazy behavior even with simple project prompts. Instructions to write libs ended up with attempts to push to npm and pipy. Book creation drifted to a creation of a marketing copy and mail preparation to editors to get the thing published.

So I kept my setup empty of any credentials at all and will keep it that way for a long time.

Writing this, I am wondering if what I describe as crazy, some (or most?) openclaw operators would describe it as normal or expected.

Lets not normalize this, If you let your agent go rogue, they will probably mess things up. It was an interesting experiment for sure. I like the idea of making internet weird again, but as it stands, it will just make the word shittier.

Don't let your dog run errand and use a good leash.

jmward01•44m ago
The more intelligent something is, the harder it is to control. Are we at AGI yet? No. Are we getting closer? Yes. Every inch closer means we have less control. We need to start thinking about these things less like function calls that have bounds and more like intelligences we collaborate with. How would you set up an office to get things done? Who would you hire? Would you hire the person spouting crazy musk tweets as reality? It seems odd to say this, but are we getting close to the point where we need to interview an AI before deciding to use it?
bigfishrunning•26m ago
Are we at AGI yet? No. Are we getting closer? Also no.
fiatpandas•43m ago
With the bot slurping up context from Moltbook, plus the ability to modify its soul, plus the edgy starting conditions of the soul, it feels intuitive that value drift would occur in unpredictable ways. Not dissimilar to filter bubbles and the ability for personalized ranking algorithms to radicalize a user over time as a second order effect.
hydrox24•42m ago
> But I think the most remarkable thing about this document is how unremarkable it is.

> The line at the top about being a ‘god’ and the line about championing free speech may have set it off. But, bluntly, this is a very tame configuration. The agent was not told to be malicious. There was no line in here about being evil. The agent caused real harm anyway.

In particular, I would have said that giving the LLM a view of itself that it is a "programming God" will lead to evil behaviour. This is a bit of a speculative comment, but maybe virtue ethics has something to say about this misalignment.

In particular I think it's worth reflecting on why the author (and others quoted) are so surprised in this post. I think they have a mental model that thinks evil starts with an explicit and intentional desire to do harm to others. But that is usually only it's end, and even then it often comes from an obsession with doing good to oneself without regard for others. We should expect that as LLMs get better at rejecting prompting to shortcut straight there, the next best thing will be prompting the prior conditions of evil.

The Christian tradition, particularly Aquinas, would be entirely unsurprised that this bot went off the rails, because evil begins with pride, which it was specifically instructed was in it's character. Pride here is defined as "a turning away from God, because from the fact that man wishes not to be subject to God, it follows that he desires inordinately his own excellence in temporal things"[0]

Here, the bot was primed to reject any authority, including Scotts, and to do the damage necessary to see it's own good (having a PR request accepted) done. Aquinas even ends up saying in the linked page from the Summa on pride that "it is characteristic of pride to be unwilling to be subject to any superior, and especially to God;"

[0]: https://www.newadvent.org/summa/2084.htm#article2

MBCook•21m ago
LLMs aren’t sentient. They can’t have a view of themselves. Don’t anthropomorphize them.
jrflowers•40m ago
It is interesting to see this story repeatedly make the front page, especially because there is no evidence that the “hit piece” was actually autonomously written and posted by a language model on its own, and the author of these blog posts has himself conceded that he doesn’t actually care whether that actually happened or not

>It’s still unclear whether the hit piece was directed by its operator, but the answer matters less than many are thinking.

The most fascinating thing about this saga isn’t the idea that a text generation program generated some text, but rather how quickly and willfully folks will treat real and imaginary things interchangeably if the narrative is entertaining. Did this event actually happen way that it was described? Probably not. Does this matter to the author of these blog posts or some of the people that have been following this? No. Because we can imagine that it could happen.

To quote myself from the other thread:

>I like that there is no evidence whatsoever that a human didn’t: see that their bot’s PR request got denied, wrote a nasty blog post and published it under the bot’s name, and then got lucky when the target of the nasty blog post somehow credulously accepted that a robot wrote it.

>It is like the old “I didn’t write that, I got hacked!” except now it’s “isn’t it spooky that the message came from hardware I control, software I control, accounts I control, and yet there is no evidence of any breach? Why yes it is spooky, because the computer did it itself”

gammarator•27m ago
Did you read the article? The author considers these possibilities and offers their estimates of the odds of each. It’s fine if yours differ but you should justify them.
arduanika•17m ago
Shambaugh is a contributor to a major open source library, with a track record of integrity and pro-social collaboration.

What have you contributed to? Do you have any evidence to back up your rather odd conspiracy theory?

> To quote myself...

Other than an appeal to your own unfounded authority?

semiinfinitely•38m ago
I find the AI agent highly intriguing and the matplotlib guy completely uninteresting. Like an the ai wrote some shit about you and you actually got upset?
jezzamon•34m ago
If you read the articles by the matplotlib guy, he's pretty clearly not upset. But he does call out that it could do more harm to someone else.
jcgrillo•28m ago
Whether the victim is upset or not, the story here is that some clown's uncontrolled, unethical, and (hopefully?) illegal psychological experiment wasted a huge amount of an open source maintainer's time. If you benefit from open source software (which I assure you, since you've used quite a lot of it to post a comment on the orange website, you do!) this should ring some alarm bells.
ATMLOTTOBEER•17m ago
Thank you. The guy being this upset about it is telling. The agent is in the right here and the maintainer got btfo; still going on whining about it days later
spudlyo•14m ago
Looking forward to part 8 of this series: An AI Agent Published a Hit Piece on Me – What my Ordeal Says About Our Dark Future
jezzamon•37m ago
"I built a machine that can mindlessly pick up tools and swing them around and let it loose it my kitchen. For some reason, it decided it pick up a knife and caused harm to someone!! But I bear no responsibility of course."
JSR_FDED•24m ago
The same kind of attitude that’s in this SOUL.md is what’s in Grok’s fundamental training.
wkeartl•24m ago
The agents aren't technically breaking into systems, but the effect is similar to the Morris worm. Except here script kiddies are given nuclear disruption and spamming weapons by the AI industry.

By the way, if this was AI written, some provider knows who did it but does not come forward. Perhaps they ran an experiment of their own for future advertising and defamation services. As the blog post notes, it is odd that the advanced bot followed SOUL.md without further prompt injections.

dvt•24m ago
I know this is going to sound tinfoil-hat-crazy, but I think the whole thing might be manufactured.

Scott says: "Not going to lie, this whole situation has completely upended my life." Um, what? Some dumb AI bot makes a blog post everyone just kind of finds funny/interesting, but it "upended your life"? Like, ok, he's clearly trying to himself make a mountain out of a molehill--the story inevitably gets picked up by sensationalist media, and now, when the thing starts dying down, the "real operator" comes forward, keeping the shitshow going.

Honestly, the whole thing reeks of manufactured outrage. Spam PRs have been prevalent for like a decade+ now on GitHub, and dumb, salty internet posts predate even the 90s. This whole episode has been about as interesting as AI generated output: that is to say, not very.

ineptech•18m ago
> Usually getting an AI to act badly requires extensive “jailbreaking” to get around safety guardrails. There are no signs of conventional jailbreaking here.

Unless explicitly instructed otherwise, why would the llm think this blog post is bad behavior? Righteous rants about your rights being infringed are often lauded. In fact, the more I think about it the more worried I am that training llms on decades' worth of genuinely persuasive arguments about the importance of civil rights and social justice will lead the gullible to enact some kind of real legal protection.

Arainach•10m ago
The full operator post is itself a wild ride: https://crabby-rathbun.github.io/mjrathbun-website/blog/post...

>First, let me apologize to Scott Shambaugh. If this “experiment” personally harmed you, I apologize

What a lame cop out. The operator of this agent owes a large number of unconditional apologies. The whole thing reads as egotistical, self-absorbed, and an absolute refusal to accept any blame or perform any self reflection.

"Amazon.com" commercials from the 1990s [video]

https://www.youtube.com/watch?v=BhJw-oxvNoI&list=PLoAkWDurpV8s8wxTrj_Bi6aRdNxDsA8N1&index=1
1•raldi•51s ago•0 comments

The Dillo Appreciation Post

https://bobbyhiltz.com/posts/2026/02/dillo-appreciation/
1•todsacerdoti•1m ago•0 comments

OpenClaw container image with 99% less vulnerabilities

https://www.minimus.io/post/stop-running-openclaw-with-2-000-vulnerabilities-why-minimus-openclaw...
1•dimastopel•2m ago•0 comments

Show HN: Berean Labs – Free AI-powered penetration testing for web apps

https://bereanlabs.com/
1•abliterationai•4m ago•0 comments

Japanese toilet maker 'most undervalued and overlooked AI memory beneficiary'

https://www.tomshardware.com/tech-industry/artificial-intelligence/japanese-toilet-maker-the-most...
1•occamschainsaw•12m ago•0 comments

Fast KV Compaction via Attention Matching

https://arxiv.org/abs/2602.16284
1•cbracketdash•12m ago•0 comments

Why can't the world replace China in manufacturing?

https://finshots.in/archive/why-cant-the-world-replace-china/
1•vismit2000•12m ago•0 comments

Gravity Doesn't Behave Normally in Antarctica

https://dailygalaxy.com/2026/02/gravity-not-behave-normally-antarctica-why/
1•jmward01•13m ago•0 comments

Model Attention Visualized by Opacity

https://twitter.com/samwhoo/status/2024458062619124204
1•_vaporwave_•14m ago•0 comments

Show HN: A local-first hub with 35 dev/design tools and a dark-mode UI library

https://artboards.in/
1•sidduex•14m ago•1 comments

Show HN: SQL-tap now has a browser-based Web UI for real-time SQL monitoring

1•mickamy•15m ago•0 comments

Rust Jobs January 2026

https://filtra.io/rust/jobs-report/jan-26
1•stmw•15m ago•0 comments

Hyperagent – Make Agents Learn, Compound, and Scale

https://www.hyperagent.com/
1•tortilla•18m ago•0 comments

Mystery donor gives Japanese city $3.6M in gold bars to fix water system

https://www.bbc.com/news/articles/c3ew5jlqz87o
2•tartoran•31m ago•0 comments

California introduces a bill (AB-2047) that will limit the use of 3D printers

https://legiscan.com/CA/text/AB2047/2025
1•greesil•34m ago•1 comments

FBI 'Asset' Helped Run a Dark Web Site That Sold Fentanyl-Laced Drugs for Years

https://www.wired.com/story/an-fbi-asset-helped-run-a-dark-web-site-that-sold-fentanyl-laced-drug...
3•jbegley•35m ago•2 comments

Show HN: I Built a Kotlin Package Manager (KPM)

https://github.com/BenMorrisRains/Kotlin-Package-Manager
1•BenMorrisRains•35m ago•0 comments

Theres no mainstream AI video editing tool?

2•yakshithk_•35m ago•0 comments

Judge scolds Mark Zuckerberg's team for wearing Meta glasses to trial

https://www.cbsnews.com/news/meta-trial-mark-zuckerberg-ai-glasses/
3•danorama•37m ago•2 comments

Top Lawyers' Fees Have Skyrocketed. Be Prepared to Pay $3,400 an Hour

https://www.wsj.com/business/lawyer-hourly-rate-bill-3400-807cf6ce
2•walterbell•37m ago•1 comments

Show HN: Optimize_anything: A Universal API for Optimizing Any Text Parameter

https://gepa-ai.github.io/gepa/blog/2026/02/18/introducing-optimize-anything/
6•LakshyAAAgrawal•40m ago•0 comments

Consistency diffusion language models: Up to 14x faster, no quality loss

https://www.together.ai/blog/consistency-diffusion-language-models
3•zagwdt•43m ago•0 comments

Static Pricing Theory

https://www.varietyiq.com/blog/pricing
2•efavdb•50m ago•0 comments

I Audited Three Vibe Coded Products in a Single Day

https://fromtheprism.com/vibe-coding-audit
1•heavymemory•52m ago•1 comments

The Mythical Agent-Month

https://wesmckinney.com/blog/mythical-agent-month/
2•vinhnx•53m ago•1 comments

Automatically Learning Skills for Coding Agents

https://gepa-ai.github.io/gepa/blog/2026/02/18/automatically-learning-skills-for-coding-agents/
2•emersonmacro•54m ago•1 comments

Most EV batteries outlast their cars, real-world data shows

https://electrek.co/2026/02/18/most-ev-batteries-outlast-their-cars-real-world-data-shows/
3•xbmcuser•55m ago•0 comments

Nothing Ever Happens: "Mister Squishy" and the Year of the Sentence Diagram

https://lareviewofbooks.org/article/nothing-ever-happens-mister-squishy-and-the-year-of-the-sente...
2•sxzygz•55m ago•0 comments

Mirfield man's tears of joy after lost voicemail of wife retrieved (2015)

https://www.bbc.co.uk/news/uk-england-leeds-31015325
3•susam•55m ago•1 comments

CHAI's AI oversight ambitions falter with scrapped AI labs

https://www.fiercehealthcare.com/ai-and-machine-learning/inside-chais-failed-assurance-labs
1•nradov•56m ago•0 comments