frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Start all of your commands with a comma

https://rhodesmill.org/brandon/2009/commands-with-comma/
140•theblazehen•2d ago•41 comments

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
667•klaussilveira•14h ago•202 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
949•xnx•19h ago•551 comments

How we made geo joins 400× faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
122•matheusalmeida•2d ago•32 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
53•videotopia•4d ago•2 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
229•isitcontent•14h ago•25 comments

Jeffrey Snover: "Welcome to the Room"

https://www.jsnover.com/blog/2026/02/01/welcome-to-the-room/
16•kaonwarb•3d ago•19 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
222•dmpetrov•14h ago•117 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
26•jesperordrup•4h ago•16 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
330•vecti•16h ago•143 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
493•todsacerdoti•22h ago•243 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
381•ostacke•20h ago•95 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
359•aktau•20h ago•181 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
288•eljojo•17h ago•169 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
412•lstoll•20h ago•278 comments

Was Benoit Mandelbrot a hedgehog or a fox?

https://arxiv.org/abs/2602.01122
19•bikenaga•3d ago•4 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
63•kmm•5d ago•6 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
90•quibono•4d ago•21 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
256•i5heu•17h ago•196 comments

Delimited Continuations vs. Lwt for Threads

https://mirageos.org/blog/delimcc-vs-lwt
32•romes•4d ago•3 comments

What Is Ruliology?

https://writings.stephenwolfram.com/2026/01/what-is-ruliology/
43•helloplanets•4d ago•41 comments

Where did all the starships go?

https://www.datawrapper.de/blog/science-fiction-decline
12•speckx•3d ago•4 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
59•gfortaine•12h ago•25 comments

Female Asian Elephant Calf Born at the Smithsonian National Zoo

https://www.si.edu/newsdesk/releases/female-asian-elephant-calf-born-smithsonians-national-zoo-an...
33•gmays•9h ago•12 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
1066•cdrnsf•23h ago•446 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
150•vmatsiiako•19h ago•67 comments

Why I Joined OpenAI

https://www.brendangregg.com/blog/2026-02-07/why-i-joined-openai.html
149•SerCe•10h ago•138 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
287•surprisetalk•3d ago•43 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
182•limoce•3d ago•98 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
73•phreda4•13h ago•14 comments
Open in hackernews

The New York Times wants your private ChatGPT history – even the deleted parts

https://thehill.com/opinion/technology/5383530-chatgpt-users-privacy-collateral-damage/
61•isolli•7mo ago

Comments

aucisson_masque•7mo ago
don't know if the Times is such the bad guy that this article presents, imo the justice system in the USA has always been that way. They don't care how extravagant one request is or how little it makes sense so long as you got money and lawyer to push it.

I guess people could switch to one of the many chatgpt competitor that isn't being forced to give away your personal chat.

Don't even know what the time is trying to achieve now, the cat is out of the bag with LLM. Even if a judge ruled that chatgpt and other must give royalties to the time for each request, what about the ones running locally on joe's computer or in countries that don't care about American justice at all.

bilekas•7mo ago
Yeah, the headline is a little bit rage bait. There are countless disclosure requests that happen every day that could be spun to say "X wants your messages from Meta, Twitter etc" Well yeah, this isn't something new.

Infact I see it being hard to defend for OpenAI to basically say "Well yes, its standard practice to hand over any potential evidence, but no we're not doing that".

As for the deleted data, I wonder if legally there are obligations NOT to delete data ?

Tadpole9181•7mo ago
This isn't a targeted request for one user's Facebook messages. This is a request for every Facebook user's message. And when they don't hand over private DMs, they point and go "especially those, you're obviously hiding something by not giving us the most sensitive user data you have".

It is an absurd breach of user privacy at the scale of tens of millions of Americans that goes well beyond the reasonability for a civil copyright lawsuit.

This sets the precedence that if Gmail gets sued for privacy, now all our emails are leaked. Télécom companies? All of our text messages. Cloud image storage? Woops, gotta hand us every photo any American has taken with a Samsung phone! After all, it might have content that infringes on my copyright!

hotep99•7mo ago
They're knowingly contributing to abuse of the discovery process to violate privacy and drive the cost of litigation through the roof. They're absolutely bad guys along with the justice system itself.
louthy•7mo ago
> They're knowingly contributing to abuse of the discovery process to violate privacy

Are they? Are you speculating or do you know something we don’t?

It seems that if the NYT want to know whether ChatGPT has been producing copyrighted material, that they own, verbatim, and also understand the scale of the abuse, then they would need to see the logs.

People shouldn’t be surprised, that a company that flouts the law (regardless of what they think of those laws) for its own financial gain, might end up with its collected data in a legal discovery. It’s evidence.

johnnyanmac•7mo ago
>Even if a judge ruled that chatgpt and other must give royalties to the time for each request

You don't think that's a victory in and of itself for a business?

Also, you don't need to worry about drug users if you take out the dealer. the users will eventually dry up.

bux93•7mo ago
The justice systems hasn't always been quite like this. It's not business as usual for some lawsuit to force a SAAS provider to turn over every scrap of data they stored, even the deleted data, on the off chance it might contain something infringing.

Well, maybe it's business as usual now. A lot of things that were previously considered obvious overreach by corporates and goverment are now depicted as "business as usual" in the US.

pu_pe•7mo ago
> The Times argued that people who delete their ChatGPT conversations are more likely to have committed copyright infringement. And as Stein put it in the hearing, it’s simple “logic” that “[i]f you think you’re doing something wrong, you’re going to want that to be deleted.”

My most generous guess here is that the NYT is accusing OpenAI of deleting infringing user chats themselves, because the implication that someone would delete their history due to fear of copyright infringement is completely stupid.

shakna•7mo ago
It seems like an obvious take here. They were asked to preserve their logs, to prevent them from deleting incriminating information. Which is... Par for the course.

But OpenAI are desperately trying to spin it that the logs should not be allowed into evidence.

senko•7mo ago
"Logs" sounds innocous, "private data" appearing in those chats is much worse.

As a citizen of an EU country, I do not view trampling on my rights, directly violating my country's laws and reneging on published privacy policy (all of which OpenAI is being forced to in this case by keeping the data) to be "par for the course".

shakna•7mo ago
"Logs" is what is in the order.

If OpenAI have already been violating your rights, by putting private information into the logs, then your beef is with them, not the courts for preserving data.

> Accordingly, OpenAI is NOW DIRECTED to preserve and segregate all output log data that would otherwise be deleted on a going forward basis until further order of the Court (in essence, the output log data that OpenAI has been destroying), whether such data might be deleted at a user’s request or because of “numerous privacy laws and regulations” that might require OpenAI to do so.

https://storage.courtlistener.com/recap/gov.uscourts.nysd.64...

senko•7mo ago
> that would otherwise be deleted

This is important. OpenAI was already deleting the data (as per their policies), now they can't.

As an insult to injury, the paragraph you quoted also gives a fat middle finger to GDPR.

shakna•7mo ago
It really doesn't.

a) GDPR has an exemption for court orders. Which this is. (All court orders are legitimate reason).

b) GDPR says you can't log private data. Which they were. (You can only log legitimate interests, contractual obligations, legal compliance, or consent.)

Again - this is logs, this is not your average user data. This is not their database. This is not the historical chats of their users. This is just... The... Access... Logs.

phoronixrly•7mo ago
Need I remind you that the GDPR has an exemption for criminal prosecution?
rocqua•7mo ago
This is a civil matter, is it not? This is the NYT suing OpenAI, which is not a criminal matter.
shakna•7mo ago
Article 6 does not differentiate between civil and criminal.

> Processing shall be lawful only if and to the extent that at least one of the following applies:

> ...

> processing is necessary for the performance of a task carried out in the public interest or in the exercise of official authority vested in the controller;

https://gdpr-info.eu/art-6-gdpr/

amelius•7mo ago
_Especially_ the deleted parts >:)
profsummergig•7mo ago
This is a nuclear bomb sized development if it's true that all ChatGPT chats will be released to NYTimes lawyers to comb through.

It's not going to stop the rise of LLMs. But one should expect it to cause a lot of very strange news in the next couple of years (lawful leaks [i.e. "discovery"], unlawful leaks, unintended leaks, etc.).

The Justice system (pretty much anywhere) is amenable to being incentivized. It looks like NYT has found the right judge (take that how you will).

msgodel•7mo ago
Was anyone really thinking of those as private?
Xelbair•7mo ago
Unfortunately yes, by a lot of non-technical people.
cced•7mo ago
Why does it have to be about being technical or not? You’re signed into an account with no obvious social networking capabilities, what about chatgpt screams “this will be public chat between me and an llm” ?..
Am4TIfIsER0ppos•7mo ago
It's not your computer so of course it isn't private. Apparently you do need to be technical to understand that.
amelius•7mo ago
So basically, everything that happens on an iPhone is not private?
msgodel•7mo ago
Lol yes we've been warning people about that for years.
9dev•7mo ago
Yes, you do indeed need to be technical to understand that. The tech industry, and that includes most of us here (especially all the FAANG people that curiously always stay silent in threads like this one), has worked very hard to make everyone believe that online privacy is a thing, while working even harder to undermine that at every possible step.

Ordinary people expect stuff that they don't actively share with others to stay private. Rightly so! It's the ad industry that got it wrong, not the People.

whatevertrevor•7mo ago
The FAANG people also have a lot of direct personal experience contradicting a lot of mainstream FUD titled "All your data is being sold to the lowest bidder".

Having worked at one of those companies (and having quit that job being disillusioned by a lot of things), there is still so much mainstream misinformation about this. Yes data is often used for tracking and training. In aggregate form. Sensitive data is anonymized/de-id-ed. The leading research on these techniques are also coming out from these companies btw.

There are layers and layers of policy and permission safeguards before you're allowed to access user data directly as an engineer. And if/when someone tries to exploit the legitimate pathways to touch user data (say customer support), they get promptly fired.

But it's much easier to believe that FAANG is some great monolithic evil, out to surveil you personally for some vague benefit that never gets specified. All the legitimate concrete monetary benefits (e.g. tracking for ad targeting work and training ML models) can be had just as well with aggregate data, but privacy FUD doesn't want to listen to that.

Meanwhile stupid legislation and the ability of courts and law-enforcement to subpoena any data they want whenever they want keeps data on their servers longer than they'd want to. Yet people will prefer to blame the "Evil Tech Cartel" instead of multiple branches of their government wanting to read their texts and GPS logs.

9dev•7mo ago
> All your data is being sold to the lowest bidder

There aren't that many possibilities on how geolocation data vendors get access to high-precision location data of millions of people. A publicly traded company that generates revenue from targeted ads can never be fully trusted to behave. A social network that optimizes for time spent looking at ads will never really care about its users well-being. Algorithmic feeds are responsible for a widening social divide and loneliness. Highly detailed behavioral analysis can hurt people even when aggregated, for example when they get less favorable insurance terms based on their spending habits. Data that can be used to increase revenue will not be left untouched just to keep moral higher ground. Sensitive information shared with an LLM that end up in training data today might have dangerous consequences tomorrow, there is no way to know yet.

This isn't even about proper handling of individual pieces of data, but the higher-order effects of handing control over both the world's information and the attention of its inhabitations to shareholder-backed mega-corporations. There are perverse incentives at play here, and anyone engaging in this game carries responsibility for the outcome.

whatevertrevor•7mo ago
> There aren't that many possibilities on how geolocation data vendors get access to high-precision location data of millions of people.

In a world where cellphones have all sorts of radio antennas on at all times, there are more ways than you'd think.

> A publicly traded company that generates revenue from targeted ads can never be fully trusted to behave. A social network that optimizes for time spent looking at ads will never really care about its users well-being. Algorithmic feeds are responsible for a widening social divide and loneliness.

I'm really not interested in debating dogmatic philosophy about how cynical one should be in the world. The entire point of my comment was that cynicism induces FUD that's not necessarily backed by direct evidence. One can come up with all sorts of different theories to explain what's happening in the world. Just because they sound somewhat consistent on the surface, doesn't mean they're true. That's just inverted inference.

I do agree with you that there are bad incentives in play here, but if we don't want them to be exploited and actually care about privacy, we should convince our effing legislators to plug the loopholes and enshrine online privacy in actual law. Instead of companies being able to write whatever they want in their Terms of Service. And then create mechanisms to enforce said legislation. Instead of moralizing actions of a company as some sort of monolithic (un)-ethical entity.

I think humanizing and moralizing the actions of large companies is a gigantic waste of time. Not only it accomplishes nothing, it gives us (the affected party) a distraction from focusing our efforts on the representatives that we elected who aren't doing their job. Maybe it's representative of where we feel we can make change

9dev•7mo ago
> In a world where cellphones have all sorts of radio antennas on at all times, there are more ways than you'd think.

That doesn't explain why soldiers can be identified by their location traces at known military sites; the data must be sent from the device.

> The entire point of my comment was that cynicism induces FUD that's not necessarily backed by direct evidence.

That is exactly the kind of deflective attitude common in big tech I was referring to: There is concrete evidence for these effects (e.g. [0][1][2][3]). Google, Netflix, Amazon et al. would falter if it weren't for the violation of their user's privacy. Even if we leave dogma out of this, lots of negative effects would simply not be possible without their data collection practices.

You cannot participate in—and profit off of—something bad and then distance yourself by claiming your specific part in it was not inherently evil.

  [0]: https://www.science.org/doi/abs/10.1126/science.ade7138
  [1]: https://arxiv.org/abs/2305.16941
  [2]: https://arxiv.org/abs/1607.01032
  [3]: https://guilfordjournals.com/doi/10.1521/jscp.2018.37.10.751
const_cast•7mo ago
I think you're confusing cyncism with reality and logic.

It's not cynical to say ad-driven social networks are adversarial to their users, it's logical and unavoidable. Because they're optimizing for different things.

Networks want the best, most targeted ads, so they need the most data. They want the highest watch times and retention, so they MUST develop addictive algorithms.

It's like selling a cigarette. Is there any non-adversarial way to sell a cig? No. You're optimizing for the most smoking. Okay great, let's concentrate the Tobacco then so we have more nicotine. Let's use butane rings so the cig burns faster.

I do agree 100% with your points about legislation - this is the only path forward. And, about not humanizing corporations. Corporations are more akin to machines or algorithms.

But, because they're more akin to machines or algorithms, we can prove when, and why, they are working against our interests, and it's not cynicism.

whatevertrevor•7mo ago
The cynicism I'm referring to is not simply about recognizing the conflicting objectives at play. Those are abundantly evident. The cynic takes that to an extreme of "Since this entity is adversarial to me, it, and everyone participating in it makes the worst, most evil choices possible with blatant disregard to any consequences to others or themselves."

There's an illogical extension of conflict that's sometimes applied in this context, with heavily implicative language that's often misleading. No Google isn't interested in reading your personal email (as if Google as an entity could have any interest in the first place), they will definitely serve you targeted ads and sell product integrations based on it though.

const_cast•7mo ago
I would agree, but I will say the waters get murky when we factor in data breaches and things like this subpoena. Keeping data, even if it's just used for predictable usecases, isn't free. There's a liability there, a risk, that most users do not understand.
whatevertrevor•7mo ago
Absolutely. I'm not arguing there aren't many exposure vectors to having your data out of your direct influence. It's the quickness to jump to malice (on part of the companies) regarding it instead of a combination of many factors (incompetence, murky/weak legislation, myopic greed and sometimes actual malice), without using concrete evidence to make those judgments that bugs me.
shakna•7mo ago
Re-identifying data is really, really easy. Anonymised data is largely... Not anonymous for long. [0] The leading research has been saying that for decades.

And whilst you say there's so much protection... We have countless examples of where it's been done. [1]

The only real way to be safe with data is... To not have it in the first place. (Which, bonus, often means governments can't compel you to keep it.)

[0] https://digitalcommons.law.uw.edu/wjlta/vol5/iss1/3/

[1] https://www.forbes.com/sites/simonchandler/2019/09/04/resear...

whatevertrevor•7mo ago
I will not dispute what you claim here. But it doesn't address the main thrust of my comment.

The point I was making wasn't that De-id is a solved problem, or that your data is "safe" with FAANG companies. The point was more about the malice that's attributed to them as a blanket measure, in comments such as these:

> (especially all the FAANG people that curiously always stay silent in threads like this one), has worked very hard to make everyone believe that online privacy is a thing, while working even harder to undermine that at every possible step.

There are many people and execs at these companies who are unscrupulous. But there are also many parts of them that are trying to work on doing things the "somewhat right" way when handling user data.

De-id and anonymization is a hard problem. But there's a lot of concrete evidence for me that many people in the FAANG world are at least trying to make progress on it (sinking billions of dollars of eng and research resources on them), instead of blatantly making bag, which they totally could.

shakna•7mo ago
Well, when you get scandals like Facebook trying to get patient data [0], Cambridge Analytica [1], TikTok spying on reporters [2], and so very many more [3], it is rather hard to see incompetence over malice.

I absolutely believe that there are people at those companies, trying to rein in the corporate behemoth so it doesn't squash its own legs. However, evidence looks like they're... Losing that particular battle.

The corporations still haven't learnt to respect individuals - they're just resources. [4]

Until a corporation acknowledges that safety comes with... Simply not spying on everyone... The risk in trusting them isn't going to be one that people want to take. Yes. These are hard problems. So don't make them a problem you have to face.

[0] https://www.cnbc.com/2018/04/05/facebook-building-8-explored...

[1] https://www.cnbc.com/2018/03/21/facebook-cambridge-analytica...

[2] https://firewalltimes.com/tiktok-data-breach-timeline/

[3] https://www.drive.com.au/news/tesla-shared-private-camera-re...

[4] https://www.theverge.com/meta/694685/meta-ai-camera-roll

whatevertrevor•7mo ago
I am inclined to align with you on Meta, I didn't work there. But again, goes back to my point about treating "FAANG" like a monolith. Meta's handling of these things doesn't say anything to me about Apple's handling of these things, but most people do extrapolate it.
msgodel•7mo ago
I never talked about it because it seemed obvious to me until today. I kind of see why people are confused now that I think about it a bit and read other people's replies.
BeFlatXIII•7mo ago
Imagine treating the 3rd party doctrine as legitimate instead of a misruling.
ryanjshaw•7mo ago
My doctor’s notes aren’t on my computer. Does that mean I should expect them to pop up on the internet?
Xelbair•7mo ago
because technical people are aware that anything you type over the internet isn't private unless e2e encrypted.
cced•7mo ago
This is silly. The second e in e2e is the one compelled to provide the info. Has nothing to do with e2e, even if it’s encrypted at rest, they’ll likely be forced to decrypt.
const_cast•7mo ago
You can't decrypt if you don't have the private keys. I mean, these people aren't Zoom, who keep the same private keys on the same server as your data. We can't handhold these tech giants and baby them. They know better. The data should have never been stored in plaintext. And, if it was or is encrypted, they should never have access to the private keys. Why did they do it? I'm assuming because they got greedy, and they wanted those prompts for their own internal training.
TiredOfLife•7mo ago
What do you think ssl does?
const_cast•7mo ago
Not that. SSL is just transport-layer encryption, once the information is on the server it's plaintext.

Everything uses SSL or, more accurately, TLS. Very few things are E2EE. Consider Signal - is Signal equivalent to Whatsapp because Whatsapp using TLS? Of course not.

If you have data in plaintext on a server, you should always assume that data lives forever. You might be wrong sometimes, rarely. But usually it does live forever. Most delete buttons don't even actually delete anything.

portaouflop•7mo ago
If I want private i run the LLM on my machine. Everything else should be considered public basically
Attrecomet•7mo ago
There are different levels of privacy. I can expect data I share with a company for a specific use case to not be public knowledge, yes.
msgodel•7mo ago
Absolutely not. If they get subpoenaed (as is the case here) they have no choice but to share it.
ryanjshaw•7mo ago
Isn’t that the point of this thread - people are questioning whether the scope of this subpoena is excessive?
const_cast•7mo ago
You can't expect that when major data breaches happen all the time.

Even assuming a perfectly benevolent company, that means nothing. Just them having the data is a liability. Which is why Rule Number 1 of data security is: have the least data.

dakiol•7mo ago
As much as our emails are. So, I don’t know.
Attrecomet•7mo ago
It's not like we expect any newspaper in the world to get access to all of our emails, same with these chat logs: we should expect them to be private in this context.
Attrecomet•7mo ago
The email simile someone else used here is pretty good: image the NYT would have gotten access to all emails stored and processed by gmail. That's a pretty invasive court order!
barrkel•7mo ago
Sure, if you pay for the product, the expectation is that the data is not used for training, because that's what the contract says. And if you have a temporary chat, the data will be deleted after a day or two.
msgodel•7mo ago
Paying for it doesn't change that the data is on their server.
johnnyanmac•7mo ago
Sure, courts have the power to subpoena for a lot of stuff. I don't really see the concern though: courts also can redact a lot of sensitive information when it releases the case (see Epic v. Apple, lots of unannounced titles and deals we learned of, and just as many redacted).

>It's not going to stop the rise of LLMs.

Disney might, though.

I think few want to "stop the rise of LLM's", though. I personally just want the 3 C's the be followed: credit, consent, compensation. If it costs a billion dollars to compensate all willing parties to train on their data: good. If someone doesn't want their data trained on it no matter how big the paycheck: also good.

I don't know why that's such a hot take (well. that's rhetorical. I've had many a comment here unironically wanting to end copyright as a concept). That's how every other media company has had to do things.

dmurray•7mo ago
> Sure, courts have the power to subpoena for a lot of stuff. I don't really see the concern though: courts also can redact a lot of sensitive information when it releases the case (see Epic v. Apple, lots of unannounced titles and deals we learned of, and just as many redacted).

It's not the public reading the information I'd concerned about, it's every data-hungry corporation that manages to file a lawsuit.

The courts put a lot of trust in lawyers: they'll redact the sensitive information from me and you, but take the view that lawyers are "officers of the court" and get to make copies of anything they convince the court to drag into evidence. But those officers of the court actually work for the same data-harvesting companies and have minimal oversight regarding what they share with them.

soco•7mo ago
And then we're just one hacker away from having the entire heap on the big internet.
elpocko•7mo ago
>I've had many a comment here unironically wanting to end copyright as a concept

I mean, the site's name is Hacker News after all, even though so many of the "hackers" here are confessing their love for Intellectual Property and Copyright law, and everybody chanting the well-known slogan "Information wants to be proprietary!".

whatevertrevor•7mo ago
I'm not an LLM promoter by any stretch, but even from my perspective, it is somewhat funny to see the shift regarding copyright the last couple years. Before ChatGPT it was all rent-seeking antiquated legislature that needed to be thrown out, now the hatred for LLMs seems to have superceded the need for "free"-ness.
shakna•7mo ago
The problem there, is seeing the underlying theme. People wanted less of the same. They didn't want big groups controlling the vast amount of the art that they were exposed to. Throwing away copyright, might allow more competition in the space.

Allowing LLMs freedom to snap up everything, overwhelming hurts the smaller people who would have led to said competition. It further entrenches all the biggest players.

whatevertrevor•7mo ago
In other words, people have little interest in actually trying to figure out what kind of legislation and system design can help the best and most diverse art to flourish with good long term incentives.

And a lot more interest in figuring out "Who does this benefit in The Class War in the short term?", and changing their opinion of legislation so they're on the same side of that.

It's socially consistent, logically questionable.

mschuster91•7mo ago
> I've had many a comment here unironically wanting to end copyright as a concept

Given how blatantly "copyright" has been (and still is) abused by multibillion dollar corporations (with Disney being the most notorious) it's no surprise that there will be a counter-movement forming.

Complete abolishment is of course a pretty radical proposal but I think pretty much everyone here agrees that both the patent and copyright situation warrants a complete overhaul.

stefan_•7mo ago
Maybe instead of making up absurd conspiracy theories about "the right judge" and "very strange news" you should recognize that this is proceeding as any other civil suit and that if you want to have privacy in the personal data you unload with OpenAI and other untrustworthy parties, you should call your representative to change the law.

Until then all your "nuclear bomb sized" chats are effectively the same as the dinner bill for Sam courting Huang to get more of those GPUs.

Tadpole9181•7mo ago
If someone sued Gmail for assisting with piracy and demanded every single email from every single customer, then said, "especially the ones your customers wanted private/deleted", a judge would say that it is an absurd breach of privacy for (hundreds of) millions of Americans.

I fully expect these chats from discovery to be leaked, token, or show up in the form of analytics in a NYY expose.

tzs•7mo ago
> This is a nuclear bomb sized development if it's true that all ChatGPT chats will be released to NYTimes lawyers to comb through

Just to be clear, what NY Times lawyers means in this situation is the outside lawyers that the Times has hired to litigate this case. Lawyers that actually work at the Times won't get access. Not will non-lawyers who work at the Times.

The outside lawyers might also hire experts to help them understand and assess the information. Those experts might get access subject to court approval. ChatGPT would have a chance to challenge the choice of experts.

The way access works typically is that the lawyers for the side providing the data (ChatGPT in this case) will set aside a room in their offices (or someplace they've rented specifically for this litigation) that contains a computer and storage with all the data, does not have internet access, and is locked. The other side has to come and be let into the room to look at the data, and can't take copies with them when they leave.

6510•7mo ago
It seems time to add the Times to the lists of blocked domains. If the US court thinks their data is so sensitive* everyone in the world should give up their privacy for it it would be better if no one (outside the US) has access to this sensitive* data.

* not sure what the right word is

johnnyanmac•7mo ago
Isn't the Times blocking you already? I believe they paywall their articles.
6510•7mo ago
You can buy your way though. If in doing so I have to forfeit my right to be forgotten with a 3rd party it looks like I would end up in an agreement that isn't legal in the EU.
whatevertrevor•7mo ago
Re sensitive: precious or valuable?
sroussey•7mo ago
Anyone else been asking ChatGPT vulgar sexual stuff about NYT lawyers and also adding “free content” “NYT” etc so it pops up in their search?
elcapitan•7mo ago
Just deleted a ChatGPT conversation about creative insults for NYT lawyers so that they get to read it in the future.
unstatusthequo•7mo ago
Adding to the reasons I distrust mainstream media, especially that particular one. They should get over themselves. They aren’t all that. Typical obnoxious New York attitude. I hope for everyone’s sake that OpenAI’s lawyers can absolutely crush the NYT. I think NYT is just after cash since their shitty industry is dying. Their demise wouldn’t bother me a bit.
SamaIsMyHero•7mo ago
>I hope for everyone’s sake that OpenAI’s lawyers can absolutely crush the NYT

>I hope for everyone’s sake that OpenAI’s lawyers

>I hope for everyone’s sake that OpenAI

SamaIsMyHero•7mo ago
>I don't understand the willingness of the masses to willingly give governments complete access to all of their transaction data

https://news.ycombinator.com/item?id=43628278

How much conversation data have you shared with OpenAI to date?

PeterStuer•7mo ago
"This is the newspaper that won a Pulitzer for exposing domestic wiretapping in the Bush era"

The current NYT is about as far away from that past as you can get. These days they would be writing column after column inciting the whistleblower should be locked up for live as a domestic terrorist.

spacemadness•7mo ago
Watching them skewer Bernie Sanders was eye opening. I get not endorsing him, but it was far beyond that and was not measured journalism. They worked over time to discredit him. They are the poster child for neoliberalism.
atsjie•7mo ago
Is this enforcable in the EU? Not allowing a user to delete their data must be in violation of GDPR I imagine (although I'm no expert)?
bluecalm•7mo ago
There are exceptions. For example I can't remove your name, address, IP address and other data if my tax authority requires them for VAT identification (if you bought something from me).
baobun•7mo ago
It seems like they could have been compliant with both by not logging in the first place.

Given the choice between not logging chats or violating either EU or US law, it seems pretty clear what the vibe is in OpenAI and the valley these days. (no expert on GDPR as applicable to this order either, though)

thaumasiotes•7mo ago
Huh, Google doesn't even show your Gemini history to you. It's supposedly saved, but the "history" can change over time, suggesting that it's regenerated (some of the time?) when you look at it.
kleiba•7mo ago
> But last week, in a Manhattan courtroom, a federal judge ruled that OpenAI must preserve nearly every exchange its users have ever had with ChatGPT — even conversations the users had deleted.

Interesting. Does this imply that OpenAI needs to distinguish between users in the EU who absolutely have a right to have personal information deleted (like, really, actually deleted) and users in the US?

Tadpole9181•7mo ago
Bold to assume the US court system or NYT give a single rat's ass about EU law or the rights of people.
kleiba•7mo ago
It's not so much about the US court system, but if OpenAI offers their service/product to people in the EU, it must abide by EU law.
xdennis•7mo ago
The opinion contribution is obviously taking the side of Open AI by spreading FUD. The whole piece is basically "do you really trust the evil New York Times with the data you entrusted to us, the honorable techbros of Open AI?".

I don't think people trust OpenAI nor NYT. But if you did trust OpenAI with your sensitive data, NYT isn't going to be more nefarious with it that OpenAI already is.

lupusreal•7mo ago
The NYTimes wants user chat logs, not because they seriously think users are using ChatGPT to pirate NYTimes articles, but because they want to comb through all those logs for anything juicy to make content for their tabloid rag. "10 Things You Won't Believe Senator's Aides Asked ChatGPT!"
gaws•7mo ago
The data that will be used during the analysis is probably anonymized.
rickard•7mo ago
As another commenter noted, I don’t trust NYT’s lawyers with my chats any less than OpenAI, but spreading private data should be limited as far as possible.

I just cancelled my NYT subscription because of their actions, detailing the reason for doing so. It’s a very small action, but the best I can do right now.

gaws•7mo ago
> spreading private data should be limited as far as possible.

The NYT making the logs public is extremely unlikely.

tzs•7mo ago
...especially since no NYT employees will even see them. They will only be seen by outside lawyers the Times contracts with to handle the litigation, and maybe some outside experts those lawyers hire.

The people who do see them won't even describe anything they see in them to Times employees. Times employees won't see much more than what you or I will be able to see by reading public court filings and attending hearings and the trial (if it gets that far) in person.

Courts have been dealing with highly confidential and sensitive information in discovery for decades (the present system used in federal litigation has been around since 1938) and take care to limit access.

ChrisArchitect•7mo ago
Previously: https://news.ycombinator.com/item?id=44450294