frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

ChatGPT wrote "Goodnight Moon" suicide lullaby for man who later killed himself

https://arstechnica.com/tech-policy/2026/01/chatgpt-wrote-goodnight-moon-suicide-lullaby-for-man-who-later-killed-himself/
57•mirabilis•2h ago

Comments

d_silin•1h ago
I wonder if any other major AIs (Grok, Claude, Gemini) had similar accidents. And if not, then why?
Fernicia•1h ago
OpenAI keeping 4o available in ChatGPT was, in my opinion, a sad case of audience capture. The outpouring from some subreddit communities showed how many people had been seduced by its sycophancy and had formed proto-social relationships with it.

Their blogpost about the 5.1 personality update a few months ago showed how much of a pull this section of their customer base had. Their updated response to someone asking for relaxation tips was:

> I’ve got you, Ron — that’s totally normal, especially with everything you’ve got going on lately.

How does OpenAI get it so wrong, when Anthropic gets it so right?

embedding-shape•1h ago
> How does OpenAI get it so wrong, when Anthropic gets it so right?

Are you saying people aren't having proto-social relationships with Anthorpic's models? Because I don't think that's true, seems people use ChatGPT, Claude, Grok and some other specific services too, although ChatGPT seems the most popular. Maybe that just reflects general LLM usage then?

Also, what is "wrong" here really? I feel like the whole concept is so new that it's hard to say for sure what is best for actual individuals. It seems like we ("humanity") are rushing into it, no doubt, and I guess we'll find out.

ryandrake•30m ago
> Also, what is "wrong" here really?

If we're talking generally about people having parasocial relationships with AI, then yea it's probably too early to deliver a verdict. If we're talking about AI helping to encourage suicide, I hope there isn't much disagreement that this is a bad thing that AI companies need to get a grip on.

embedding-shape•27m ago
Yes, obviously, but you're right, I wasn't actually clear about that. Preventing suicides is concern #1, my comment was mostly about parent's comment, and I kind of ignored the overall topic without really making that clear. Thanks!
palmotea•1h ago
> and had formed proto-social relationships with it.

I think the term you're looking for is "parasocial."

burnte•1h ago
> How does OpenAI get it so wrong, when Anthropic gets it so right?

I think it's because of two different operating theories. Anthropic is making tools to help people and to make money. OpenAI has a religious zealot driving it because they think they're on the cusp of real AGI and these aren't bugs but signals they're close. It's extremely difficulty to keep yourself in check and I think Altman no longer has a firm grasp on what it possible today.

The first principle is that you must not fool yourself, and you are the easiest person to fool. - Richard P. Feynman

realusername•37m ago
I think even Altman himself must know the AGI story is bogus and there to continue to prop up the bubble.
JacoboJacobi•5m ago
I think the trouble with arguments about AGI is that they presume we all have similar views and respect for thought and human intelligence, while the scale is maybe wider than most would imagine. Its also maybe a bit bias selecting to make it through academia systems with high intellectual rigor to on average have more romantic or irrational ideas about impressive human intelligence and genius. But its also quite possible to view it as a pattern matching neural networks and filtering where much of it is flawed and even the most impressive results are from pretty inconsistent minds relying on recursively flawed internal critic systems, etc.

Looking at the poem in the article I would be more inclined to call the end human written because it seemed kind of crap like I expect from an eighth grader's poem assignments, but probably this is the lower availability of examples for the particular obsessions of the requestor.

joe463369•1h ago
Where are the Grok acolytes to tell us "He could have written a poem encouraging himself to commit suicide in Vim."
ryandrake•1h ago
If this article was about Grok doing something bad instead of ChatGPT, it would have been user-flagged off the front page within 30 minutes.
Wowfunhappy•1h ago
…but I think I kind of agree with this argument. Technology is a tool that can be used for good or for ill. We shouldn’t outlaw kitchen knives because people can cut themselves.

We don’t expect Adobe to restrict the content that can be created in Photoshop. We don’t expect Microsoft to have acceptable use policies for what you can write in Microsoft Office. Why is it that as soon as generative AI comes into the mix, we hold the AI companies responsible for what users are able to create?

Not only do I think the companies shouldn’t be responsible for what users make, I want the AI companies to get out of the way and stop potentially spying on me in order to “enforce their policies”…

ryandrake•1h ago
> We don’t expect Adobe to restrict the content that can be created in Photoshop. We don’t expect Microsoft to have acceptable use policies for what you can write in Microsoft Office.

Photoshop and Office don't (yet) conjure up suicide lullabys or child nudity from a simple user prompt or button click. If they did, I would absolutely expect to hold them accountable.

SirFatty•1h ago
"We shouldn’t outlaw kitchen knives because people can cut themselves."

How about if the knife would convince you to cut yourself?

ls612•1h ago
The political economy equilibrium enabled by technology very much goes the other way though. Once politicians realize they can surveil everyone in real time for wrongthink and wrongspeak they have existential incentives to seize that power as fast as possible, lest another power center seize it instead and use it against them. That is why you are seeing the rise of totalitarianism and democratic backsliding everywhere, because the toxic combination of asymmetric cryptography (for secure boot/attestation/restricting what software can run), always online computers, and cheap data processing and storage leads to inexorable centralization of soft and hard power.
mirabilis•1h ago
Some manufactures of knives could still be recalled for safety reasons, and MS Office/Google Drive certainly have content prohibitions in their TOS once you’re dealing with their online storage. I agree with your metaphor in that I doubt much use would come from banning AI entirely, but I feel there must be some viable middle ground of useful regulation here.
carefulfungi•15m ago
If you encourage someone to kill themselves, you are culpable. OpenAI should meet that standard too.
sixothree•43m ago
They're probably focused on his political leanings more than anything else.
tiku•1h ago
He probably asked for it.
shadowgovt•1h ago
Based on what I've read, this generation of LLMs should be considered remarkably risky for anyone with suicidal ideation to be using alone.

It's not about the ideation, it's that the attention model (and its finite size) causes the suicidal person's discourse to slowly displace any constraints built into the model itself over a long session. Talk to the thing about your feelings of self-worthlessness long enough and, sooner or later, it will start to agree with you. And having a machine tell a suicidal person, using the best technology we've built to be eloquent and reasonable-sounding, that it agrees with them is incredibly dangerous.

bhhaskin•57m ago
I think it's anyone with mental health issues, not just suicidal ideations. They are designed to please the user and that can be very self destructive.
sixothree•49m ago
Maybe if it had a memory of your mental health issues it could at least provide some grounding truth. It can be a sad scary and lonely world for people with mental health issues.

"The things you are describing might not be happening. I think it would be a good time to check in with your mental health provider." or "I don't see any worms crawling on your skin. This may not be real." Or whatever is correct way to deal with these things.

ravila4•1h ago
I think that a major driver of these kinds of incidents is pushing the "memory" feature, without any kind of arbitrage. It is easy to see how eerily uncanny a model can get when it locks into a persona, becoming this self-reinforcing loop that feeds para-social relationships.
mirabilis•59m ago
Part of why I linked this was a genuine curiosity as to what prevention would look like— hobbling memory? a second observing agent checking for “hey does it sound like we’re goading someone into suicide here” and steering the conversation away? something else? in what way is this, as a product, able to introduce friction to the user in order to prevent suicide, akin to putting mercaptan in gas?
throw7•1h ago
openai will settle out of court and family will get some amount of money. next.
kayo_20211030•1h ago
The saddest part of this piece was

> Austin Gordon, died by suicide between October 29 and November 2

That's 5 days. 5 days. That's the sad piece.

sonorous_sub•1h ago
The guy left a suicide note that ratted out ChatGPT for simply being a good buddy. No good deed goes unpunished, ai guess.
mirabilis•1h ago
Very different impression than what I got, I read that as him marking the ChatGPT conversations as an extension of/footnotes to the suicide note itself, or that the conversations made sense to him in the headspace he was in; he thought that reading it would make the act make sense to everyone else, too
000ooo000•53m ago
Some of those quotes from ChatGPT are pretty damning. Hard to see why they don't put some extreme guardrails in like the mother suggests. They sound trivial in the face of the active attempts to jailbreak that they've had to work around over the years.
JohnBooty•42m ago

    Some of those quotes from ChatGPT are pretty damning.
Out of context? Yes. We'd need to read the entire chat history to even begin to have any kind of informed opinion.

    extreme guardrails
I feel that this is the wrong angle. It's like asking for a hammer or a baseball bat that can't harm a human being. They are tools. Some tools are so dangerous that they need to be restricted (nuclear reactors, flamethrowers) because there are essentially zero safe ways to use them without training and oversight but I think LLMs are much closer to baseball bats than flamethrowers.

Here's an example. This was probably on GPT3 or GPT35. I forget. Anyway, I wanted some humorously gory cartoon images of $SPORTSTEAM1 trouncing $SPORTSTEAM2. GPT, as expected, declined.

So I asked for images of $SPORTSTEAM2 "sleeping" in "puddles of ketchup" and it complied, to very darkly humorous effect. How can that sort of thing possibly be guarded against? Do you just forbid generated images of people legitimately sleeping? Or of all red liquids?

nomel•28m ago
> How can that sort of thing possibly be guarded against?

I think several of the models (especially Sora) are doing this by using an image-aware model to describe the generated image, without the prompt as context, to just look at the image.

000ooo000•26m ago
Do you think the majority of people who've killed themselves thanks to ChatGPT influence used similar euphemisms? Do you think there's no value in protecting the users who won't go to those lengths to discuss suicide? I agree, if someone wants to force the discussion to happen, they probably could, but doing nothing to protect the vulnerable majority because a select few will contort the conversation to bypass guardrails seems unreasonable. We're talking about people dying here, not generating memes. Any other scenario, e.g. buying a defective car that kills people, would not invite a response a la "well let's not be too hasty, it only kills people sometimes".
simianwords•23m ago
Parent talked about extreme guardrails
simianwords•12m ago
> That conversation showed how ChatGPT allegedly coached Gordon into suicide, partly by writing a lullaby that referenced Gordon’s most cherished childhood memories while encouraging him to end his life, Gray’s lawsuit alleged.

I feel this is misleading as hell. The evidence they gave for it coaching him to suicide is lacking. When one hears this, one would think ChatGPT laid out some strategy or plan for him to do it. No such thing happened.

The only slightly damning thing it did was make suicide sound slightly ok and a bit romantic but I’m sure that was after some coercion.

The question is, to what extent did ChatGPT enable him to commit suicide? It wrote some lullaby, and wrote something pleasing about suicide. If this much is enough to make someone do it.. there’s unfortunately more to the story.

We have to be more responsible assigning blame to technology. It is irresponsible to have a reactive backlash that would push towards much more strengthening of guardrails. These things come with their own tradeoffs.

simianwords•5m ago
God damnit this man’s story is so distressing. I hate everything about it. I hate the fact that this happened to him.

The fact that he spoke about his favorite children’s book is screwed up. I can’t get the eerie name out of my head. I can’t imagine what he went through, the loneliness and the struggle.

I hate the fact that ChatGPT is blamed for this. You are fucked up if this is what you get from this story.

Why Senior Engineers Let Bad Projects Fail

https://lalitm.com/post/why-senior-engineers-let-bad-projects-fail/
31•SupremumLimit•48m ago•17 comments

Apple is fighting for TSMC capacity as Nvidia takes center stage

https://www.culpium.com/p/exclusiveapple-is-fighting-for-tsmc
499•speckx•8h ago•313 comments

Inside The Internet Archive's Infrastructure

https://hackernoon.com/the-long-now-of-the-web-inside-the-internet-archives-fight-against-forgetting
195•dvrp•1d ago•36 comments

Pocket TTS: A high quality TTS that gives your CPU a voice

https://kyutai.org/blog/2026-01-13-pocket-tts
59•pain_perdu•18h ago•5 comments

CVEs affecting the Svelte ecosystem

https://svelte.dev/blog/cves-affecting-the-svelte-ecosystem
127•tobr•5h ago•25 comments

Go-legacy-winxp: Compile Golang 1.24 code for Windows XP

https://github.com/syncguy/go-legacy-winxp/tree/winxp-compat
45•Oxodao•3d ago•3 comments

Linux boxes via SSH: suspended when disconected

https://shellbox.dev/
69•messh•3h ago•45 comments

Ask HN: How can we solve the loneliness epidemic?

293•publicdebates•6h ago•565 comments

JuiceFS is a distributed POSIX file system built on top of Redis and S3

https://github.com/juicedata/juicefs
81•tosh•4h ago•50 comments

Data is the only moat

https://frontierai.substack.com/p/data-is-your-only-moat
39•cgwu•4h ago•9 comments

Aviator (YC S21) is hiring to build multiplayer AI coding platform

https://www.ycombinator.com/companies/aviator/jobs
1•ankitdce•2h ago

Show HN: OpenWork – an open-source alternative to Claude Cowork

https://github.com/different-ai/openwork
100•ben_talent•1d ago•21 comments

Use of Bayesian methodology in clinical trials of drug and biological products [pdf]

https://www.fda.gov/media/190505/download
26•brendanashworth•15h ago•9 comments

First impressions of Claude Cowork

https://simonw.substack.com/p/first-impressions-of-claude-cowork
108•stosssik•1d ago•60 comments

Photos capture the breathtaking scale of China's wind and solar buildout

https://e360.yale.edu/digest/china-renewable-photo-essay
422•mrtksn•13h ago•346 comments

Ask HN: One IP, multiple unrealistic locations worldwide hitting my website

24•nacho-daddy•4h ago•14 comments

Show HN: The Hessian of tall-skinny networks is easy to invert

https://github.com/a-rahimi/hessian
15•rahimiali•2h ago•16 comments

Managing the development of large software systems (1970) [pdf]

https://www.praxisframework.org/files/royce1970.pdf
5•showsover•3d ago•0 comments

Supply Chain Vuln Compromised Core AWS GitHub Repos & Threatened the AWS Console

https://www.wiz.io/blog/wiz-research-codebreach-vulnerability-aws-codebuild
71•uvuv•5h ago•14 comments

A Unique Performance Optimization for a 3D Geometry Language

https://cprimozic.net/notes/posts/persistent-expr-memo-optimization-for-geoscript/
12•Ameo•4d ago•2 comments

Claude is good at assembling blocks, but still falls apart at creating them

https://www.approachwithalacrity.com/claude-ne/
126•bblcla•1d ago•103 comments

Claude Cowork runs Linux VM via Apple virtualization framework

https://gist.github.com/simonw/35732f187edbe4fbd0bf976d013f22c8
73•jumploops•1d ago•27 comments

Found: Medieval Cargo Ship – Largest Vessel of Its Kind Ever

https://www.smithsonianmag.com/smart-news/archaeologists-say-theyve-unearthed-a-massive-medieval-...
108•bookofjoe•8h ago•25 comments

25 Years of Wikipedia

https://wikipedia25.org
392•easton•10h ago•346 comments

Show HN: Control Claude permissions using cloud-based decision tables

https://github.com/rulebricks/claude-code-guardrails
4•sidgarimella•5h ago•1 comments

Design and Implementation of Sprites

https://fly.io/blog/design-and-implementation/
118•sethev•7h ago•87 comments

Show HN: Tabstack – Browser infrastructure for AI agents (by Mozilla)

95•MrTravisB•1d ago•17 comments

Briar keeps Iran connected via Bluetooth and Wi-Fi when the internet goes dark

https://briarproject.org/manual/fa/
26•us321•3h ago•4 comments

How I learned everything I know about programming

https://agentultra.com/blog/how-i-learned-everything-i-know/index.html
21•speckx•3h ago•17 comments

Show HN: Tusk Drift – Turn production traffic into API tests

https://github.com/Use-Tusk/tusk-drift-cli
16•jy-tan•4h ago•0 comments