frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

We built another object storage

https://fractalbits.com/blog/why-we-built-another-object-storage/
60•fractalbits•2h ago•9 comments

Java FFM zero-copy transport using io_uring

https://www.mvp.express/
25•mands•5d ago•6 comments

How exchanges turn order books into distributed logs

https://quant.engineering/exchange-order-book-distributed-logs.html
49•rundef•5d ago•17 comments

macOS 26.2 enables fast AI clusters with RDMA over Thunderbolt

https://developer.apple.com/documentation/macos-release-notes/macos-26_2-release-notes#RDMA-over-...
467•guiand•18h ago•237 comments

AI is bringing old nuclear plants out of retirement

https://www.wbur.org/hereandnow/2025/12/09/nuclear-power-ai
33•geox•1h ago•25 comments

Sick of smart TVs? Here are your best options

https://arstechnica.com/gadgets/2025/12/the-ars-technica-guide-to-dumb-tvs/
433•fleahunter•1d ago•362 comments

Photographer built a medium-format rangefinder, and so can you

https://petapixel.com/2025/12/06/this-photographer-built-an-awesome-medium-format-rangefinder-and...
78•shinryuu•6d ago•9 comments

Apple has locked my Apple ID, and I have no recourse. A plea for help

https://hey.paris/posts/appleid/
865•parisidau•10h ago•445 comments

GNU Unifont

https://unifoundry.com/unifont/index.html
287•remywang•18h ago•68 comments

A 'toaster with a lens': The story behind the first handheld digital camera

https://www.bbc.com/future/article/20251205-how-the-handheld-digital-camera-was-born
42•selvan•5d ago•18 comments

Beautiful Abelian Sandpiles

https://eavan.blog/posts/beautiful-sandpiles.html
83•eavan0•3d ago•16 comments

Rats Play DOOM

https://ratsplaydoom.com/
332•ano-ther•18h ago•123 comments

Show HN: Tiny VM sandbox in C with apps in Rust, C and Zig

https://github.com/ringtailsoftware/uvm32
167•trj•17h ago•11 comments

OpenAI are quietly adopting skills, now available in ChatGPT and Codex CLI

https://simonwillison.net/2025/Dec/12/openai-skills/
481•simonw•15h ago•271 comments

Computer Animator and Amiga fanatic Dick Van Dyke turns 100

109•ggm•6h ago•23 comments

Will West Coast Jazz Get Some Respect?

https://www.honest-broker.com/p/will-west-coast-jazz-finally-get
10•paulpauper•6d ago•2 comments

Formula One Handovers and Handovers From Surgery to Intensive Care (2008) [pdf]

https://gwern.net/doc/technology/2008-sower.pdf
82•bookofjoe•6d ago•33 comments

Show HN: I made a spreadsheet where formulas also update backwards

https://victorpoughon.github.io/bidicalc/
179•fouronnes3•1d ago•85 comments

Freeing a Xiaomi humidifier from the cloud

https://0l.de/blog/2025/11/xiaomi-humidifier/
126•stv0g•1d ago•51 comments

Obscuring P2P Nodes with Dandelion

https://www.johndcook.com/blog/2025/12/08/dandelion/
57•ColinWright•4d ago•1 comments

Go is portable, until it isn't

https://simpleobservability.com/blog/go-portable-until-isnt
119•khazit•6d ago•101 comments

Ensuring a National Policy Framework for Artificial Intelligence

https://www.whitehouse.gov/presidential-actions/2025/12/eliminating-state-law-obstruction-of-nati...
169•andsoitis•1d ago•217 comments

Poor Johnny still won't encrypt

https://bfswa.substack.com/p/poor-johnny-still-wont-encrypt
52•zdw•10h ago•64 comments

YouTube's CEO limits his kids' social media use – other tech bosses do the same

https://www.cnbc.com/2025/12/13/youtubes-ceo-is-latest-tech-boss-limiting-his-kids-social-media-u...
84•pseudolus•3h ago•67 comments

Slax: Live Pocket Linux

https://www.slax.org/
41•Ulf950•5d ago•5 comments

50 years of proof assistants

https://lawrencecpaulson.github.io//2025/12/05/History_of_Proof_Assistants.html
107•baruchel•15h ago•17 comments

Gild Just One Lily

https://www.smashingmagazine.com/2025/04/gild-just-one-lily/
29•serialx•5d ago•5 comments

Capsudo: Rethinking sudo with object capabilities

https://ariadne.space/2025/12/12/rethinking-sudo-with-object-capabilities.html
75•fanf2•17h ago•44 comments

Google removes Sci-Hub domains from U.S. search results due to dated court order

https://torrentfreak.com/google-removes-sci-hub-domains-from-u-s-search-results-due-to-dated-cour...
193•t-3•11h ago•34 comments

String theory inspires a brilliant, baffling new math proof

https://www.quantamagazine.org/string-theory-inspires-a-brilliant-baffling-new-math-proof-20251212/
167•ArmageddonIt•22h ago•154 comments
Open in hackernews

Company Reminder for Everyone to Talk Nicely About the Giant Plagiarism Machine

https://www.mcsweeneys.net/articles/a-company-reminder-for-everyone-to-talk-nicely-about-the-giant-plagiarism-machine
82•zdw•6mo ago

Comments

minimaxir•6mo ago
[edit: retracted kneejerk take]
JonChesterfield•6mo ago
Linkedin has stuff like this on it and I'm pretty sure it's sincere.

Also Google wouldn't give me an antonym for "satire", only the output of a LLM which thinks synonym is the same thing as antonym.

leephillips•6mo ago
Nouns do not have antonyms.

Edit: I should have said not all nouns have antonyms.

monster_truck•6mo ago
What? Yes they do. ridge/groove, heaven/hell, war/peace, north/south, predator/prey etc etc
Centigonal•6mo ago
Not true. Some concrete nouns don't have antonyms, but "good," "black," "heat," "invisibility," and many more abstract nouns have clear antonyms.
gotoeleven•6mo ago
sincerence ? earnestence? Of course using those words will make whatever you're writing sound like satire.
Pulcinella•6mo ago
Even "plagiarism" is putting way to positive a spin on it. "Rampant copyright infringement" is more accurate.

I'm sure we all have our own feelings about IP law, but remember what happens to regular people who try stuff like this. I don't think the RIAA, Disney, or Nintendo (or the government) are going to be pleased to hear "it's not piracy! It's a transformative experience protected by fair use!"

steveBK123•6mo ago
Millenials might be the generation that both got threatened with jail for music copyright infringement violations as youth AND gets to have their job threatened by automated mass corporate copyright infringement in adulthood!
hnthrow90348765•6mo ago
"Haha, yeah, those scrappy Millennials - who knows where their breaking point is but I'm sure there's a fintech app for making that bet"
svaha1728•6mo ago
We wanted Aaron Swartz and we got Sam Altman.
azemetre•6mo ago
Truly the darkest timeline.
rrauenza•6mo ago
I had this argument presented to me and I wasn't sure what to do with it.

> Humans are allowed to "absorb" art around them into their brains and generate derivative art. People may copy Miyazaki's style... why shouldn't an AI farm be allowed to?

Let's put aside for a moment that AI may have "consumed" some art without a license (e.g., "google books" - did google purchase every book?).

NooneAtAll3•6mo ago
except lawyers keep saying "fanart is actually technically illegal" and resinging/changing lyrics in songs isn't enough to be protected by "fair use" stuff

if anything, I'd campaign for "we should limit copyright because it already doesn't work for Ai"

rrauenza•6mo ago
I wasn't intending to include fan art.

Copying Miyazaki's style ... or copying Monet's style... those aren't fan art.

thiht•6mo ago
Why not? I’m a fan of X and make art in the style of X using the tools at my disposal (AI). Why is that not fanart?
rrauenza•6mo ago
They are. I meant the subset of fan art of the infringing type where you make fan art of, say, Mickey Mouse and sell stickers.
mcphage•6mo ago
> People may copy Miyazaki's style... why shouldn't an AI farm be allowed to?

People may take a penny from the tray at the 7-11, so why can't an AI farm take pennies from all the trays? Or take them from a much bigger tray and do it a couple of million times?

davidclark•6mo ago
The same legal rule applies to both for determining whether something is a derivative work.

No one is stopping you from using similar proportions or colors as Miyazaki to draw a character. You are also allowed to draw your own interpretation of an electric mouse-like monster.

Copyright infringement occurs if that character looks exactly like say Totoro or Pikachu. That is not “in the style of”, that is copying.

A problem with LLMs is that since their corpus is so large, it is difficult to identify when any given output is crossing that line because a single observer’s knowledge of the works influencing the output is limited. You might feed it a picture of your grandfather and it returns an almost exact copy of a grandfather character from a Miyazaki film you haven’t seen. If you don’t share the output with others, it might never be noticed that the infringement occurred.

The given argument conflates the slightest influence with direct copying. It is a reductive take that, personally, I’ve found emblematic of pro-LLM arguments.

rrauenza•6mo ago
Thanks for helping pick apart the argument presented to me.

I don't like the idea that photos I've published on, say, flickr have been pulled into these. Especially stuff I've published with creative commons non-commercial use.

SilasX•6mo ago
Learning from copyrighted works to create new ones has never been protected by copyright[1], and has never needed separate licensing rights. Until 2022, no one even suggested it, to a rounding error. If anything, people would have been horrified at the idea of being dinged because your novel clearly drew inspiration from another work.

That narrative only got picked up because people needed a reason to demonize evil corps that they already hated for unrelated reasons.

[1] Yes, if you create "new" works from your learning that are basically copies, that has always been infringement. I'm talking about the general case.

satyanash•6mo ago
> narrative only got picked up because people needed a reason to demonize evil corps

Either they aren't evil in which case they're being demonized, or they're already evil in which case demonization is redundant.

Keeping aside the motives of people, what is clear is that scale effects of AI cannot be ignored. An AI "learning" millions of pieces of content in a short span is not the same as humans spending time, effort and energy to replicate someone's style. You can argue that its 'neural nets' in both cases, but the massive scale is what separates the two.

A village is not a large family, a city is not a large village, ... and all that.

SilasX•6mo ago
>Either they aren't evil in which case they're being demonized, or they're already evil in which case demonization is redundant.

If you were trying to be charitable rather than clever, you would have read "evil corps" as "corps that the critic regards as evil".

>Keeping aside the motives of people, what is clear is that scale effects of AI cannot be ignored.

Okay, so just give some kind of standard -- any clear, articulable standard -- for how and why the scale matters. It's a cop-out to just rest your case on a hand-wavy "it changes at scale".

Pulcinella•6mo ago
"It's not piracy! I'm learning from all this media I didn't pay for!"
caseyy•6mo ago
> Learning from copyrighted works to create new ones has never been protected by copyright

The term "learning" (I presume from "machine learning") shoulders a lot of weight. If we describe the situation more precisely, it involves commercially exploiting literature and other text media to produce a statistical corpus of texts, which is then commercially exploited. It's okay if that is licensed, but none of the AI companies bothered to license said original texts. Some (allegedly) just downloaded torrents of books, which is clear as day piracy. It has little to do with "learning" as used in common English — a person naturally retaining some knowledge of what they've consumed. Plain English "learning" doesn't describe the whole of what's happening with LLMs at all! It's a borrowed term, so let's not pretend it isn't.

What's happening is closer to buying some music cassettes, ripping parts of songs off them into various mixtapes, and selling them. The fact that the new cassettes "learned" the contents of the old ones, or that the songs are now jumbled up, doesn't change that the mixtape maker never had a license to copy the bits of music for commercial exploitation in the first place. After the infringement is done, the rest is smoke and mirrors...

SilasX•6mo ago
>The term "learning" (I presume from "machine learning") shoulders a lot of weight. If we describe the situation more precisely, it involves commercially exploiting literature and other text media to produce a statistical corpus of texts, which is then commercially exploited.

It's "commercially exploiting literature" in the same sense that an author would if they read a bunch of novels and then wrote their own based on what the learned from the pre-existing text. The whole point in dispute is whether that turns into infringement when an AI does it.

By labeling only one of them as "commercially exploiting literature" but not the other, you're failing to distinguish them in any meaningful way, and basically arguing from name-calling.

>It has little to do with "learning" as used in common English — a person naturally retaining some knowledge of what they've consumed. Plain English "learning" doesn't describe the whole of what's happening with LLMs at all! It's a borrowed term, so let's not pretend it isn't.

That's fair, that you can't just call them both "learning" and call it a day. But then the burden's on you to show how machine learning breaks from the time-honored tradition of license-free learning/"updating what you write based on having viewed other works". What's different? What is it about machine learning that makes it infringement in a way that it isn't when humans update their weights from having seen copyrighted works?

>What's happening is closer to buying some music cassettes, ripping parts of songs off them into various mixtapes, and selling them. The fact that the new cassettes "learned" the contents of the old ones, or that the songs are now jumbled up, doesn't change that the mixtape maker never had a license to copy the bits of music for commercial exploitation in the first place.

Okay, but (as above) to make that case, you'd need to identify where "acceptable" learning/"updating what you write based on having viewed other works" crosses over into the infringing mixtape example, and I have yet to see anyone try beyond "they're evil corps, it must be bad somehow".

dragonwriter•6mo ago
> Learning

Human learning doesn't involve making a copy (or any other use of an exclusive rights) as defined in copyright law (the human brain not being a fixed medium), AI training does, because digital storage is.

AI training may fall into the Fair Use exception in the US, but it absolutely does not fall through the same gap that makes human learning not even eequire fair use analysis since it doesn't meet the definitions ser out for a violation in the first place.

SilasX•6mo ago
>Human learning doesn't involve making a copy (or any other use of an exclusive rights) as defined in copyright law (the human brain not being a fixed medium), AI training does, because digital storage is.

That's just false -- AI models themselves only store parameter weights, which represent a high-level, aggregated understanding across all data that was learned on, i.e. what human brains do. This is clear from all the examples where you have to painstakingly trick them into producing exact text.

And even if they did store something that's "effectively a copy", that's no more copyright infringement than when Google caches a site in order to handle search queries. It's not copyright infringement until they start redistributing [non-fair-use] content from the sites.

yusefnapora•6mo ago
Good thing my MP3 files only store a psycho-acoustic model of that Metallica album!

I mean sure, if you go to painstaking lengths, you can trick your computer into making some noise that seems vaguely similar to the copyrighted work it was trained on, but I trust the consumer to make their own fair use evaluation.

dragonwriter•6mo ago
Sorry, that was expressed less precisely than I intended.

Human learning doesn't involve fixing anything, whether or not it would be a copy, into a fixed medium, and therefore even the question of whether or not it would be a copy if it were in a fixed medium, much less the question of whether or not such a copy would be exempted due to fair use, does not arise.

AI training does involve fixing something into a fixed medium, which makes it disanalogous to human learning. This raises the question of whether the thing fixes (parameter weights) is or is not legally a lossy mechanical copy of the training corpus, which further potentially raises the question of whether the incorporation of individual copyright protected works into the training corpus combined with its use in training would (before considering exceptions like Fair Use) violate the copyright on the works involved and whether, if it would, the use nevertheless satisfies the requirements of one or more of those exceptions (Fair Use being the one usually argued for.)

> And even if they did store something that's "effectively a copy", that's no more copyright infringement than when Google caches a site in order to handle search queries.

Among the reasons that Google removed outside access to its cached copies is that the evolution of the web meant that providing them increasingly potentially had negative revenue on the original content providers, which harms the Fair Use case for caching, since effect on the market of the original work is a fair use factor. (Google's cache was ruled as fair use based on the factual situation -- including implied consent, which AI model trainers have a much weaker arguent for -- in 2006, but changes in the facts relevant to fair use can change the outcome.)

But AI training is not so analogous to Google's cache (much less the situation of Google's cache in 2006) that one can simply leap with no analysis from one being fair use to the other in the first place. That's applying wishful thinking, not Fair Use analysis.

SilasX•6mo ago
>Human learning doesn't involve fixing anything,... AI training does involve fixing something into a fixed medium, which makes it disanalogous to human learning.

No, it breaks one part of the analogy, which has never been considered relevant: how long the learning persists. Yes, computers can store weights of a AI model much longer than any human could. But storing this "impact of having viewed a copyrighted work" has never been a factor in considering something infringing, regardless of how long it's stored. Courts don't consider it infringement if you simply use what you have learned from reading previous novels (the updates to your brain's neural weights) in producing new content.

Your argument is saying that the "fixedness of storing model weights" (aggregated high level understanding) can cause it to be infringement. That's without precedent. It would imply that if you're really good at "fixing" (remembering) the style of an author you read 50 years ago, it somehow crosses over into infringement because you "fixed that understanding into a medium" (your brain). That's not how it works at all.

>Among the reasons that Google removed outside access to its cached copies

I wasn't referring to the cached part of a site that Google serves to users, but the undistributed cache that they hold merely to know which sites to point you to, so you're not addressing the analogy. Here Google does store an exact copy (of at least some portions) and even then it's not considered copyright infringement until they start redistributing that content (or at least, too much of it).

My point was that acceptance of this practice even further bolsters the case that AI models aren't infringing, because, even if they did store exact copies, that generally not considered infringement until they start serving close-enough copies of the original copyrighted content.

username135•6mo ago
Its hard not to demonize large corps that often enjoy the governments legal largess when there are many examples of individuals with ruined lives for the same behavior.
gojomo•6mo ago
There is no de jure legal requirement that the RIAA, Disney, Nintendo, or the government be "pleased to hear" about new technology.

And, while copyright prohibits some sorts of reproduction of copyrighted materials, it doesn't give rightsholders veto power over all downstream uses of legal copies.

rrauenza•6mo ago
How could the author not have called it the Giant Plagiarism Tool or Giant Plagiarism Technology ...
nh23423fefe•6mo ago
People who live in the past use bad metaphors like this.
nashashmi•6mo ago
Might be a trademark problem?
readthenotes1•6mo ago
Disappointed. I thought this was going to be about Harvard.
fblp•6mo ago
I wish this wasn't flagged, I find some of these satirical pieces on hn most thought provoking.
codr7•6mo ago
I can see it's not being well received by the AI apologizing squad.

Let's just flag anything that gets in the way of profit and peace of mind.

Truth will find you.