frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

ArXiv Declares Independence from Cornell

https://www.science.org/content/article/arxiv-pioneering-preprint-server-declares-independence-co...
216•bookstore-romeo•4h ago•56 comments

FSF Threatens Anthropic over Infringed Copyright: Share Your LLMs Freel

https://www.fsf.org/blogs/licensing/2026-anthropic-settlement
74•m463•3d ago•39 comments

Google details new 24-hour process to sideload unverified Android apps

https://arstechnica.com/gadgets/2026/03/google-details-new-24-hour-process-to-sideload-unverified...
769•0xedb•15h ago•858 comments

Push events into a running session with channels

https://code.claude.com/docs/en/channels
321•jasonjmcghee•8h ago•181 comments

Full Disclosure: A Third (and Fourth) Azure Sign-In Log Bypass Found

https://trustedsec.com/blog/full-disclosure-a-third-and-fourth-azure-sign-in-log-bypass-found
129•nyxgeek•7h ago•26 comments

Building a Reader for the Smallest Hard Drive

https://www.willwhang.dev/Reading-MK4001MTD/
25•voctor•3d ago•1 comments

Drugwars for the TI-82/83/83 Calculators (2011)

https://gist.github.com/mattmanning/1002653/b7a1e88479a10eaae3bd5298b8b2c86e16fb4404
132•robotnikman•8h ago•50 comments

Cockpit is a web-based graphical interface for servers

https://github.com/cockpit-project/cockpit
249•modinfo•12h ago•141 comments

Return of the Obra Dinn: spherical mapped dithering for a 1bpp first-person game

https://forums.tigsource.com/index.php?topic=40832.msg1363742#msg1363742
347•PaulHoule•3d ago•44 comments

How the Turner twins are mythbusting modern technical apparel

https://www.carryology.com/insights/how-the-turner-twins-are-mythbusting-modern-gear/
209•greedo•2d ago•103 comments

Show HN: Three new Kitten TTS models – smallest less than 25MB

https://github.com/KittenML/KittenTTS
415•rohan_joshi•16h ago•158 comments

Drawvg Filter for FFmpeg

https://ayosec.github.io/ffmpeg-drawvg/
7•nolta•2d ago•1 comments

4Chan mocks £520k fine for UK online safety breaches

https://www.bbc.com/news/articles/c624330lg1ko
357•mosura•17h ago•599 comments

A Journey Through Infertility

https://pudding.cool/2026/03/ivf/
28•tchanukvadze•2d ago•14 comments

Noq: n0's new QUIC implementation in Rust

https://www.iroh.computer/blog/noq-announcement
198•od0•14h ago•25 comments

Astral to Join OpenAI

https://astral.sh/blog/openai
1345•ibraheemdev•19h ago•829 comments

FSFE supporters affected: Payment provider Nexi cancelled us

https://fsfe.org/news/2026/news-20260316-01.en.html
32•rasjani•1h ago•6 comments

Clockwise acquired by Salesforce

https://www.getclockwise.com
116•nigelgutzmann•12h ago•60 comments

Be intentional about how AI changes your codebase

https://aicode.swerdlow.dev
112•benswerd•11h ago•47 comments

Launch HN: Voltair (YC W26) – Drone and charging network for power utilities

71•wweissbluth•15h ago•26 comments

Scaling Karpathy's Autoresearch: What Happens When the Agent Gets a GPU Cluster

https://blog.skypilot.co/scaling-autoresearch/
166•hopechong•15h ago•71 comments

How many branches can your CPU predict?

https://lemire.me/blog/2026/03/18/how-many-branches-can-your-cpu-predict/
72•chmaynard•1d ago•46 comments

Linux Page Faults, MMAP, and userfaultfd for faster VM boots

https://www.shayon.dev/post/2026/65/linux-page-faults-mmap-and-userfaultfd/
28•shayonj•1d ago•2 comments

NanoGPT Slowrun: 10x Data Efficiency with Infinite Compute

https://qlabs.sh/10x
142•sdpmas•13h ago•29 comments

OpenBSD: PF queues break the 4 Gbps barrier

https://undeadly.org/cgi?action=article;sid=20260319125859
198•defrost•19h ago•60 comments

From Oscilloscope to Wireshark: A UDP Story (2022)

https://www.mattkeeter.com/blog/2022-08-11-udp/
112•ofrzeta•13h ago•23 comments

Waymo Safety Impact

https://waymo.com/safety/impact/
301•xnx•12h ago•310 comments

Launch HN: Canary (YC W26) – AI QA that understands your code

53•Visweshyc•16h ago•20 comments

Last love: a romance in a care home (2023)

https://www.theguardian.com/lifeandstyle/2023/nov/23/last-love-a-romance-in-a-care-home
23•NaOH•3d ago•8 comments

CAIveat Emptor: What You Tell AI Can and Will Be Used Against You

https://natlawreview.com/article/caiveat-emptor-what-you-tell-ai-can-and-will-be-used-against-you
3•petethomas•5m ago•0 comments
Open in hackernews

FSF Threatens Anthropic over Infringed Copyright: Share Your LLMs Freel

https://www.fsf.org/blogs/licensing/2026-anthropic-settlement
74•m463•3d ago

Comments

mjg59•2h ago
Where's the threat? The FSF was notified that as part of the settlement in Bartz v. Anthropic they were potentially entitled to money, but in this case the works in question were released under a license that allowed free duplication and distribution so no harm was caused. There's then a note that if the FSF had been involved in such a suit they'd insist on any settlement requiring that the trained model be released under a free license. But they weren't, and they're not.

(Edit: In the event of it being changed to match the actual article title, the current subject line for this thread is " FSF Threatens Anthropic over Infringed Copyright: Share Your LLMs Freel")

lelanthran•1h ago
It's just an indication to model trainers that they should take care to omit FSF software from training.

Not a nothing burger, but not totally insignificant either.

mjg59•1h ago
Is it? The FSF's description of the judgement is that the training was fair use, but that the actual downloading of the material may have been a copyright infringement. What software does the FSF hold copyright to that can't be downloaded freely? Under what circumstances would the FSF be in a position to influence the nature of a settlement if they weren't harmed?
jfoster•1h ago
Is harm necessary to show in a copyright infringement case?
mjg59•28m ago
Copyright infringement causes harm, so if there's no harm there's no infringement. You can freely duplicate GFDLed material, so downloading it isn't an infringement. If training a model on that downloaded material is fair use then there's no infringement.
eschaton•1h ago
It’s pretty fucking simple: If GPL code is integrated into Claude, the Claude needs to be distributed under the terms of the GPL.
mjg59•1h ago
If it's pretty fucking simple, can you point to the statement in the linked post that supports this assertion? What it says is "According to the notice, the district court ruled that using the books to train LLMs was fair use", and while I accept that this doesn't mean the same would be true for software, I don't see anything in the FSF's post that contradicts the idea that training on GPLed software would also be fair use. I'm not passing a value judgement here, I'm a former board member of the FSF and I strongly believe in the value and effectiveness of copyleft licenses, I'm just asking how you get from what's in the post to such an absolute assertion.
boramalper•1h ago
Yet another instance of people jumping to comments based on the title of the submission alone. They don't mention GPL even once in that post...
sunnyps•1h ago
It's pretty fucking simple: a judge needs to decide that, not armchair lawyers on HN.
Bombthecat•1h ago
We know AI will be pushed through. No matter the laws
agile-gift0262•1h ago
what I keep wondering is what kind of laws will be rendered useless with the precedent they'll cause. Can this be beginning of the end of copyright and intellectual property?
Bombthecat•39m ago
In a way, I think so. Just let the code recreate existing code, say it's AI code and doesn't break any copyright laws
tsimionescu•27m ago
Copyright, possibly. Intellectual property more broadly, no. AI has 0 impact on trademark law, quite clearly (which is anchored in consumer protection, in principle). Patent law is perhaps more related, but it's still pretty far.
teiferer•1h ago
> but in this case the works in question were released under a license that allowed free duplication and distribution so no harm was caused.

FSF licenses contain attribution and copyleft clauses. It's "do whatever you want with it provided that you X, Y and Z". Just taking the first part without the second part is a breach of the license.

It's like renting a car without paying and then claiming "well you said I can drive around with it for the rest of the day, so where is the harm?" while conveniently ignoring the payment clause.

You maybe confusing this with a "public domain" license.

jcul•1h ago
This article is talking about a book though, not software.

"Sam Williams and Richard Stallman's Free as in freedom: Richard Stallman's crusade for free software"

"GNU Free Documentation License (GNU FDL). This is a free license allowing use of the work for any purpose without payment."

I'm not familiar with this license or how it compares to their software licenses, but it sounds closer to a public domain license.

kennywinker•53m ago
It sounds that way a bit from the one sentence. But that’s not the case at all.

> 4. MODIFICATIONS

> You may copy and distribute a Modified Version of the Document under the conditions of sections 2 and 3 above, provided that you release the Modified Version under precisely this License, with the Modified Version filling the role of the Document, thus licensing distribution and modification of the Modified Version to whoever possesses a copy of it. In addition, you must do these things in the Modified Version:

Etc etc.

In short, it is a copyleft license. You must also license derivative works under this license.

Just fyi, the gnu fdl is (unsurprisingly) available for free online - so if you want to know what it says, you can read it!

mjg59•29m ago
And the judgement said that the training was fair use, but that the duplication might be an infringement. The GFDL doesn't restrict duplication, only distribution, so if training on GFDLed material is fair use and not the creation of a derivative work then there's no damage.
onion2k•22m ago
For this to stand up in court you'd need to show that an LLM is distributing "a modified version of the document".

If I took a book and cut it up into individual words (or partial words even), and then used some of the words with words from every other book to write a new book, it'd be hard to argue that I'm really "distributing the first book", even if the subject of my book is the same as the first one.

This really just highlights how the law is a long way behind what's achievable with modern computing power.

karel-3d•42m ago
FDL is famously annoying.

wikipedia used to be under FDL and they lobbied FSF to allow an escape hatch to Commons for a few months, because FDL was so annoying.

Dylan16807•1h ago
They don't need the "do whatever" permission if everything they do is fair use. They only need the downloading permission, and it's free to download.
mjg59•20m ago
If what you do with a copyrighted work is covered by fair use it doesn't matter what the license says - you can do it anyway. The GFDL imposes restrictions on distribution, not copying, so merely downloading a copy imposes no obligation on you and so isn't a copyright infringement either.

I used to be on the FSF board of directors. I have provided legal testimony regarding copyleft licenses. I am excruciatingly aware of the difference between a copyleft license and the public domain.

politelemon•1h ago
The title is:

The FSF doesn't usually sue for copyright infringement, but when we do, we settle for freedom

Kwpolska•1h ago
Classic FSF, completely detached from reality. Did they propose any way for Anthropic to continue earning any revenue, paying its employees, and developing new models, if they give away their current models for free?
mcv•1h ago
It's not the FSF's job to provide Anthropic with a business model. If it turns out that their business model depends entirely on copyright violation, they might not have a business model. That's true regardless of whether you think the case has any merit.
kouteiheika•1h ago
Although it might not satisfy FSF there is a very simple way to do it - commit to release your models for free X months after they're first made available.
slopinthebag•1h ago
What the fuck? How is that their problem?

"Yeah we can't prosecute this person for stealing your car, because you haven't considered how they're going to get to work"

solid_fuel•55m ago
> Classic Hollywood, completely detached from reality. Did they propose any way for The Pirate Bay to continue earning any revenue, paying for its hosting, and developing new features, if they aren't allowed to redistribute movies for free?
Shitty-kitty•51m ago
Should the FCC be proposing how Robocaller's continue earning any revenue in its decisions?
tomhow•36m ago
Please don't fulminate on HN. The guidelines make it clear we're trying for something better here. https://news.ycombinator.com/newsguidelines.html
pwdisswordfishy•31m ago
OH NO THE POOR CAPITALISTS
slopinthebag•1h ago
Good. I want to see more lawsuits going after these hyper scalers for blatantly disregarding copyright law while simultaneously benefiting from it. In a just world they would all go down and we would be left with just the OSS models. But we don't live in a fair world :(
rvz•49m ago
> Among the works we hold copyrights over is Sam Williams and Richard Stallman's Free as in freedom: Richard Stallman's crusade for free software, which was found in datasets used by Anthropic as training inputs for their LLMs.

This is the reason why AI companies won't let anyone inspect which content was in the training set. It turns out the suspicions from many copyright holders (including the FSF) was true (of course).

Anthropic and others will never admit it, hence why they wanted to settle and not risk going to trial. AI boosters obviously will continue to gaslight copyright holders to believe nonsense like: "It only scraped the links, so AI didn't directly train on your content!", or "AI can't see like humans, it only see numbers, binary or digits" or "AI didn't reproduce exactly 100% of the content just like humans do when tracing from memory!".

They will not share the data-set used to train Claude, even if it was trained on AGPLv3 code.

bobokaytop•34m ago
The framing of 'share your weights freely' as a remedy is interesting but underspecified. The FSF's argument is essentially that training on copyrighted code without permission is infringement, and the remedy should be open weights. But open weights don't undo the infringement -- they just make a potentially infringing artifact publicly available. That's not how copyright remedies work. What they're actually asking for is more like a compulsory license, which Congress would have to create. The demand for open weights as a copyright remedy is a policy argument dressed up as a legal one.
kavalg•29m ago
It looks like the stance of FSF is for proliferation of the copyleft to trained LLMs

> "Therefore, we urge Anthropic and other LLM developers that train models using huge datasets downloaded from the Internet to provide these LLMs to their users in freedom"

mjg59•24m ago
No, it looks like the stance of the FSF is that models should be free as a matter of principle, the same as their stance when it comes to software. Nothing in the linked post contradicts the description that the judgement was that the training was fair use.
Topfi•22m ago
A related topic that I have in the past thought about is, whether LLM derived code would necessitate the release under a copyleft license because of the training data. Never saw a cogent analysis that explained either why or why not this is the case beyond practicality due to models having been utilized in closed source codebases already…
mjg59•16m ago
The short answer is that we don't know. The longer answer based purely on this case is that there's an argument that training is fair use and so copyleft doesn't have any impact on the model, but this is one case in California and doesn't inherently set precedent in the US in general and has no impact at all on legal interpretations in other countries.
grodriguez100•22m ago
Is the FSF threatening Anthropic? The way I read it looks like they are not:

> We are a small organization with limited resources and we have to pick our battles, but if the FSF were to participate in a lawsuit such as Bartz v. Anthropic and find our copyright and license violated, we would certainly request user freedom as compensation.

Sounds more like “we can’t and won’t sue, but this is the kind of compensation that we think would be appropriate”

charcircuit•20m ago
>share complete training inputs with every user of the LLM

They don't have the rights to distribute the training data.