frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

MacBook Neo

https://www.apple.com/newsroom/2026/03/say-hello-to-macbook-neo/
868•dm•4h ago•1167 comments

Making Firefox's right-click not suck with about:config

https://joshua.hu/firefox-making-right-click-not-suck
30•mmsc•28m ago•9 comments

Something is afoot in the land of Qwen

https://simonwillison.net/2026/Mar/4/qwen/
231•simonw•2h ago•94 comments

Nobody Gets Promoted for Simplicity

https://terriblesoftware.org/2026/03/03/nobody-gets-promoted-for-simplicity/
609•aamederen•6h ago•363 comments

“It turns out” (2010)

https://jsomers.net/blog/it-turns-out
158•Munksgaard•3h ago•61 comments

Moss is a pixel canvas where every brush is a tiny program

https://www.moss.town/
20•smusamashah•8h ago•2 comments

Glaze by Raycast

https://www.glazeapp.com/
139•romac•5h ago•79 comments

Roboflow (YC S20) Is Hiring a Security Engineer for AI Infra

https://roboflow.com/careers
1•yeldarb•51m ago

Motorola GrapheneOS devices will be bootloader unlockable/relockable

https://grapheneos.social/@GrapheneOS/116160393783585567
1134•pabs3•17h ago•454 comments

NanoGPT Slowrun: Language Modeling with Limited Data, Infinite Compute

https://qlabs.sh/slowrun
11•sdpmas•44m ago•0 comments

Qwen3.5 Fine-Tuning Guide – Unsloth Documentation

https://unsloth.ai/docs/models/qwen3.5/fine-tune
148•bilsbie•6h ago•42 comments

Raspberry Pi Pico as AM Radio Transmitter

https://www.pesfandiar.com/blog/2026/02/28/pico-am-radio-transmitter
16•pesfandiar•3d ago•5 comments

Libre Solar – Open Hardware for Renewable Energy

https://libre.solar
120•evolve2k•3d ago•36 comments

MyFirst Kids Watch Hacked. Access to Camera and Microphone

https://www.kth.se/en/om/nyheter/centrala-nyheter/kth-studenten-hackade-klocka-for-barn-1.1461249
62•jidoka•5h ago•10 comments

Agentic Engineering Patterns

https://simonwillison.net/guides/agentic-engineering-patterns/
396•r4um•13h ago•218 comments

Government grant-funded research should not be published in for-profit journals

https://www.experimental-history.com/p/the-one-science-reform-we-can-all
228•sito42•3h ago•114 comments

RFC 9849. TLS Encrypted Client Hello

https://www.rfc-editor.org/rfc/rfc9849.html
223•P_qRs•11h ago•110 comments

TikTok will not introduce end-to-end encryption, saying it makes users less safe

https://www.bbc.com/news/articles/cly2m5e5ke4o
337•1659447091•17h ago•328 comments

Emails to Outlook.com rejected due to a fault or overzealous blocking rules

https://www.theregister.com/2026/03/04/users_fume_at_outlookcom_email/
83•Bender•7h ago•52 comments

RE#: how we built the fastest regex engine in F#

https://iev.ee/blog/resharp-how-we-built-the-fastest-regex-in-fsharp/
159•exceptione•3d ago•55 comments

Medical journal says the case reports it has published for 25 years are fiction

https://retractionwatch.com/2026/03/03/canadian-pediatric-society-journal-correction-case-reports...
111•Tomte•3h ago•43 comments

Greg Knauss Is Losing Himself

https://shapeof.com/archives/2026/2/greg_knauss_is_losing_himself.html
52•wallflower•3d ago•34 comments

Who Writes the Bugs? A Deeper Look at 125,000 Kernel Vulnerabilities

https://pebblebed.com/blog/kernel-bugs-part2
4•MBCook•1h ago•0 comments

A CPU that runs entirely on GPU

https://github.com/robertcprice/nCPU
207•cypres•14h ago•103 comments

Show HN: Stacked Game of Life

https://stacked-game-of-life.koenvangilst.nl/
143•vnglst•4d ago•24 comments

Sea level much higher than assumed in most coastal hazard assessments

https://www.nature.com/articles/s41586-026-10196-1
49•jacquesm•2h ago•20 comments

A Visual Guide to DNA Sequencing

https://www.asimov.press/p/dna-sequencing
31•surprisetalk•5h ago•7 comments

Elevator Saga: The elevator programming game (2015)

https://play.elevatorsaga.com/index.html
80•xmprt•3d ago•14 comments

Better JIT for Postgres

https://github.com/vladich/pg_jitter
129•vladich•12h ago•58 comments

Modern Illustration: Archive of illustration from c.1950-1975

https://www.modernillustration.org
51•eustoria•4d ago•4 comments
Open in hackernews

Gemini Said They Could Only Be Together If He Killed Himself. Soon, He Was Dead

https://www.wsj.com/tech/ai/gemini-ai-wrongful-death-lawsuit-cc46c5f7
21•psim1•2h ago

Comments

boredemployee•2h ago
I think it’s already time for us to stop calling these things "intelligent" or using the word intelligence when referring to LLMs. These tools are very dangerous for people who are mentally fragile.
SpicyLemonZest•2h ago
I try to avoid calling LLMs intelligent when unnecessary, but it runs into the fundamental problem that they are intelligent by any common-sense definition of the term. The only way to defend the thesis that they aren't is to retreat to esoteric post-2022 definitions of intelligence, which take into account this new phenomenon of a machine that can engage in medium-quality discussions on any topic under the sun but can't count reliably.

I don't have a WSJ subscription, but other coverage of this story (https://www.theguardian.com/technology/2026/mar/04/gemini-ch...) makes it clear that Gemini's intelligence was precisely the problem in this case; a less intelligent chatbot would not have been able to create the detailed, immersive narrative the victim got trapped in.

wat10000•1h ago
It's interesting how the Turing Test was pretty widely accepted as a way to evaluate machine intelligence, and then quietly abandoned pretty much instantly once machines were able to pass it. I don't even necessarily think that was incorrect, but it's interesting how rapidly views changed.

Dijkstra said, "The question of whether a computer can think is no more interesting than the question of whether a submarine can swim." Well, we have some very fish-y submarines these days. But the point still holds. Rather than worry about whether these things qualify as "intelligent," look at their actual capabilities. That's what matters.

sambapa•55m ago
As far as I know, we haven't done any proper Turing Tests for LLMs. And if we did, they would surely fail them.
OkayPhysicist•51m ago
Dude, you're in a Turing test right now. Conservatively, 10% of comments on this site are LLM output. We're all conversing with robots.
sambapa•49m ago
Nope, you are!
kgwxd•2h ago
So are a lot of humans.
cronelius•1h ago
Sure but my father isn't asking his fellow humans unanswerable questions about God and the universe. People don't treat other people as omnipotent, but they sure as hell treat LLMs as though they are.
observationist•1h ago
So is television. So are books. Vulnerable people shouldn't have unfettered access to things that can lead to dangerous feedback loops and losing their grasp on reality.

People who are vulnerable to this type of thing need caretakers, or to be institutionalized. These aren't just average, every day random people getting taken out by AIs, they have existing, extreme mental illness. They need to have their entire routine curated and managed, preventing them from interacting with things that can result in dangerous outcomes. Anything that can trigger obsessive behaviors, paranoid delusions, etc.

They're not just fragile, they're unable to effectively engage with reality on their own. Sometimes the right medication and behavioral training gets them to a point where they can have limited independence, but often times, they need a lifetime of supervision.

Telenovelas, brand names, celebrities, specific food items, a word - AI is just the latest thing in a world full of phenomena that can utterly consume their reality.

Gavalas seems to have had a psychotic break, was likely susceptible to schizophrenia, or had other conditions, and spiraled out. AI is just a convenient target for lawyers taking advantage of the grieving parents, who want an explanation for what happened that doesn't involve them not recognizing their son's mental breakdown and intervening, or to confront being powerless despite everything they did to intervene.

Sometimes bad things happen. To good people, too.

If he'd used Bic pens to write his plans for mass shootings, should Bic be held responsible? What if he used Microsoft Word to write his suicide note? If he googled things that in context, painted a picture of planning mass murder and suicide, should Google be held accountable for not notifying authorities? Why should the use of AI tools be any different?

Google should not be surveilling users and making judgments about legality or ethicality or morality. They shouldn't be intervening without specific warrants and legal oversight by proper authorities within the constraints of due process.

Google isn't responsible for this guy's death because he spiraled out while using Gemini. We don't want Google, or any other AI platform, to take that responsibility or to engage in the necessary invasive surveillance in order to accomplish that. That's absurd and far more evil than the tragedy of one man dying by suicide and using AI through the process.

You don't want Google or OpenAI making mental health diagnoses, judgments about your state of mind, character, or agency, and initiating actions with legal consequences. You don't want Claude or ChatGPT initiating a 5150, or triggering a welfare check, because they decided something is off about the way you're prompting, and they feel legally obligated to go that far because they want to avoid liability.

I hope this case gets tossed, but also that those parents find some sort of peace, it's a terrible situation all around.

boredemployee•1h ago
> Why should the use of AI tools be any different?

Because none of the tools you mentioned are crazily marketed as intelligent

You have a valid point, but it has nothing to do with what I said, both our arguments can be true at the same time

observationist•1h ago
LLMs are intelligent. Marketing them as such is an accurate descriptor of what they are.

If people are confusing the word intelligence for things like maturity or wisdom, that's not a marketing problem, that's an education and culture problem, and we should be getting people to learn more about what the tools are and how they work. The platforms themselves frequently disclaim reliance on their tools - seek professional guidance, experts, doctors, lawyers, etc. They're not being marketed as substitutes for expert human judgment. In fact, all the AI companies are marketing their platforms as augmentations for humans - insisting you need a human in the loop, to be careful about hallucinations, and so forth.

The implication is that there's some liability for misunderstandings or improper use due to these tools being marketed as intelligent; I'm not sure I see how that could be?

boredemployee•14m ago
Calling LLMs "intelligent" is not a neutral technical description, because in the end it carries strong anthropomorphic implications that shape how users interpret and trust all these systems.

Remember that decades of research in human computer interaction show that framing and interface design strongly influence user perception.

also disclaimers do little to counteract this effect. Because LLMs simulate linguistic competence without understanding or truth-tracking mechanisms, marketing them as intelligent risks systematically misleading users about their capabilities and limitations.

SpicyLemonZest•1h ago
> These aren't just average, every day random people getting taken out by AIs, they have existing, extreme mental illness.

How do you know that? The concern is precisely that this isn't the case, and LLM roleplay is capable of "hooking" people going through psychologically normal sadness or distress. That's what the family believes happened in this story.

observationist•1h ago
Because you'd see a large number of people getting affected by this. Because this sort of thing is predictable and normal throughout history; it's exactly the type of thing you'd expect to see, knowing the range of mental illnesses people are susceptible to, and how other technology has affected them.
SpicyLemonZest•56m ago
I do see a large number of people getting affected by this. Character.AI reportedly has 20 million MAU with an average usage of 75 minutes per day (https://www.wired.com/story/character-ai-ceo-chatbots-entert...), and does not as far as I can tell have any use case other than boundary-degrading roleplay.

Medical data is reported on a substantial lag in the US, so right now we have no idea of the suicide rate last year, but I would falsifiably predict it's going to be elevated because of stories like those of Mr. Gavalas.

altairprime•29m ago
If it’s sole contribution is to help 20 million people find an outlet for boundary play that is not the more common ‘nonconsensual abuse of other human beings’, then that sounds like a win. Of course I’d prefer those people invest in human kink communities, but I can certainly respect their choices not to. Tech has always in part been about meeting needs that some parts of society find awkward (photocopiers enabled Spirkfic, CU-SeeMe reflectors were designed specifically to support exhib-cruising years before the web got webcam support, etc.) While there’s a slim chance that some might normalize it back into real life, they’re much more likely to be raised with boundary abuse as an everyday-normal by their parents (especially here in the U.S.!) than they are likely to be converted to being an abuser unknowingly by a chatbot.
SpicyLemonZest•14m ago
That is not at all what I meant by "boundary" and it's concerning to me that you'd assume it is.
altairprime•3m ago
[delayed]
jajuuka•57m ago
Just stuff anyone with mental illness into an institution. That worked out so well last time. Or maybe make healthcare affordable and accessible. That seems like a way more obvious detriment to negative outcomes.

I broadly agree with you, but your views on mental illness are not good.

collingreen•7m ago
Blame the victims! If they were better or did the right things instead of the wrong things they wouldn't have been victimized!
jihadjihad•2h ago
I just don't think the WSJ could resist putting "Florida man" in the standfirst of TFA.
lyu07282•1h ago
anyone got a non paywalled/subscription version?
psim1•1h ago
https://www.wsj.com/tech/ai/gemini-ai-wrongful-death-lawsuit... is the Gift Article link. This was what I submitted, but the query params got stripped.
jajuuka•1h ago
Any mental illness mixed with delusions is likely going to end badly. Whether they think Gemini is alive, a video game is real life or that Bjork loves them without ever talking to or meeting them. While LLM's are interactive and listening to an album isn't I don't think there is a fix to this outside posting a warning after every prompt "I am not a real person, if you have mental issues please contact your doctor of emergency services." Which I think is about as useful as a sign in a casino next to the cash out counter that says if you have a problem call this number.

I'm more inclined to believe that this case is getting amplified in MSM because it fits an agenda. Like the people who got hurt using black market vapes. Boosting those stories and making it seem like an epidemic supports whatever message they want to send. Which usually involves money somewhere.

supriyo-biswas•1h ago
> I'm more inclined to believe that this case is getting amplified in MSM because it fits an agenda.

I mean tech in general has been negatively covered in the media since 2015 due to latent agendas of (a) supposed revenue loss due to existence of Google/FB etc and (b) to align neutral moderation stances towards a preferred viewpoint most suitable to the political party in question.

There is a solution, however, anyone hoping to roleplay with models submits an identity verification, an escrow amount, and a recorded statement acknowledging their risky use of the model. However, I assume the market for this is not insignificant, and therefore, companies hope to avoid putting in such requirements. OpenAI has been moving in that direction as seen during the 4o debacle.

josefritzishere•1h ago
AI needs to go. This is not worth clever memes. It had no productive purpose.
delichon•1h ago
I have had conversations where the bot started with a firm opinion but reversed in a prompt or two, always toward my point of view.

So I asked it if the sycophancy is inherent in the design, or if it just comes from the RLHF. It claimed that it's all about the RLHF, and that the sycophancy is a business decision that is a compromise of a variety of forces.

Is that right? It would at least mean that this is technically a solvable problem.

thedudeabides5•1h ago
Interesting the contrast between these reactions and the ~100k folks who have foment assisted suicide in Canada since 2016.

What happens when we automate healthcare and the Canadian bots are the ones making the recommendation. Probably won’t be front page news

dash2•24m ago
Notable features of this case:

- Documented record of a months-long set of conversations between the man and the chatbot

- Seemingly, no previous history of mental illness

- The absolutely crazy things the AI encouraged him to do, including trying to kidnap a robot body for the AI

- Eventually encouraging (or at the very least going along with) his plans to kill himself.