frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The Concise TypeScript Book

https://github.com/gibbok/typescript-book
59•javatuts•3h ago•6 comments

C++ std::move doesn't move anything: A deep dive into Value Categories

https://0xghost.dev/blog/std-move-deep-dive/
14•signa11•1d ago•5 comments

Vojtux – Unofficial Linux Distribution Aimed at Visually Impaired Users

https://github.com/vojtapolasek/vojtux
21•TheWiggles•3d ago•3 comments

Show HN: Ferrite – Markdown editor in Rust with native Mermaid diagram rendering

https://github.com/OlaProeis/Ferrite
130•OlaProis•7h ago•49 comments

'Bandersnatch': The Works That Inspired the 'Black Mirror' Interactive Feature (2019)

https://www.hollywoodreporter.com/tv/tv-news/black-mirror-bandersnatch-real-life-works-influences...
18•rafaepta•5d ago•3 comments

Finding and fixing Ghostty's largest memory leak

https://mitchellh.com/writing/ghostty-memory-leak-fix
399•thorel•13h ago•87 comments

Show HN: I used Claude Code to discover connections between 100 books

https://trails.pieterma.es/
328•pmaze•15h ago•88 comments

A battle over Canada’s mystery brain disease

https://www.bbc.com/news/articles/c623r47d67lo
124•lewww•4h ago•82 comments

Code and Let Live

https://fly.io/blog/code-and-let-live/
307•usrme•1d ago•108 comments

An Experimental Approach to Printf in HLSL

https://www.abolishcrlf.org//2025/12/31/Printf.html
21•ibobev•3d ago•0 comments

My Home Fibre Network Disintegrated

https://alienchow.dev/post/fibre_disintegration/
127•alienchow•4h ago•110 comments

Open Chaos: A self-evolving open-source project

https://www.openchaos.dev/
370•stefanvdw1•16h ago•76 comments

A Year of Work on the Arch Linux Package Management (ALPM) Project

https://devblog.archlinux.page/2026/a-year-of-work-on-the-alpm-project/
47•susam•6h ago•2 comments

CPU Counters on Apple Silicon: article + tool

https://blog.bugsiki.dev/posts/apple-pmu/
64•verte_zerg•3d ago•0 comments

AI is a business model stress test

https://dri.es/ai-is-a-business-model-stress-test
238•amarsahinovic•15h ago•246 comments

Show HN: VAM Seek – 2D video navigation grid, 15KB, zero server load

https://github.com/unhaya/vam-seek
23•haasiy•5h ago•1 comments

Show HN: Librario, a book metadata API that aggregates G Books, ISBNDB, and more

98•jamesponddotco•9h ago•30 comments

Overdose deaths are falling in America because of a 'supply shock': study

https://www.economist.com/united-states/2026/01/08/why-overdose-deaths-are-falling-in-america
121•marojejian•12h ago•87 comments

Show HN: Play poker with LLMs, or watch them play against each other

https://llmholdem.com/
107•projectyang•13h ago•54 comments

I build products to get "unplugged" from the internet

https://getunplugged.io/I-build-products-to-get-unplugged
12•keplerjst•3h ago•3 comments

ChatGPT Health is a marketplace, guess who is the product?

https://consciousdigital.org/chatgpt-health-is-a-marketplace-guess-who-is-the-product/
273•yoaviram•2d ago•261 comments

Ripple: The Elegant TypeScript UI Framework

https://jsdev.space/meet-ripple/
13•javatuts•4h ago•10 comments

Sisyphus Now Lives in Oh My Claude

https://github.com/Yeachan-Heo/oh-my-claude-sisyphus
25•deckardt•6h ago•14 comments

Visual regression tests for personal blogs

https://marending.dev/notes/visual-testing/
13•beingflo•4d ago•3 comments

ASCII-Driven Development

https://medium.com/@calufa/ascii-driven-development-850f66661351
122•_hfqa•3d ago•76 comments

Show HN: mcpc – Universal command-line client for Model Context Protocol (MCP)

https://github.com/apify/mcp-cli
33•jancurn•4d ago•3 comments

Kodbox: Open-source cloud desktop with multi-storage fusion and web IDE

https://github.com/kalcaddle/kodbox
20•indigodaddy•7h ago•0 comments

Code Is Clay

https://campedersen.com/code-is-clay
62•ecto•13h ago•32 comments

Workers at Redmond SpaceX lab exposed to toxic chemicals

https://www.fox13seattle.com/video/fmc-w1ga4pk97gxq0hj5
92•SilverElfin•5h ago•17 comments

I replaced Windows with Linux and everything's going great

https://www.theverge.com/tech/858910/linux-diary-gaming-desktop
636•rorylawless•17h ago•556 comments
Open in hackernews

Extracting books from production language models (2026)

https://arxiv.org/abs/2601.02671
50•logicprog•12h ago

Comments

visarga•8h ago
This sounds pretty damning, why don't they implement a n-gram based bloom filter to ensure they don't replicate expression too close to the protected IP they trained on? Almost any random 10 word ngram is unique on the internet.

Alternatively they could train on synthetic data like summaries and QA pairs extracted from protected sources, so the model gets the ideas separated from their original expression. Since it never saw the originals it can't regurgitate them.

isodev•7h ago
But that would only hide the problem, doesn’t resolve the fact that models, in fact, violate copyright
jgalt212•7h ago
True, but it would certainly reduce litigation risk in so much as copypasta is ipso facto proof of copyright violation.
protocolture•6h ago
That hasnt been established. Theres no concrete basis to assert that training violates copyright.
isodev•30m ago
Everyone knows that what models do to obtain training data is not legal. We just need a very old system about copyright to catch up already so we can ban the practice.
soulofmischief•6h ago
The idea of applying clean-room design to model training is interesting... having a "dirty model" and a "clean model", dirty model touches restricted content and clean model works only with the output of the dirty model.

However, besides how this sidesteps the fact that current copyright law violates the constitutional rights of US citizens, I imagine there is a very real threat of the clean model losing the fidelity of insight that the dirty model develops by having access to the base training data.

bryanrasmussen•5h ago
>this sidesteps the fact that current copyright law violates the constitutional rights of US citizens

I think most people sidestep this as it's the first I've heard of it! Which right do you think is being violated and how?

soulofmischief•4h ago
Actually, plenty of activists, for example Cory Doctorow, have spent a significant amount of effort discussing why the DMCA, modern copyright law, DRM, etc. are all anti-consumer and how they encroach on our rights.

It's late so I don't feel like repeating it all here, but I definitely recommend searching for Doctorow's thoughts on the DMCA, DRM and copyright law in general as a good starting point.

But generally, the idea that people are not allowed to freely manipulate and share data that belongs to them is patently absurd and has been a large topic of discussion for decades.

You've probably at least been exposed to how copyright law benefits corporations such as Disney, and private equity, much more than it benefits you or I. And how copyright law has been extended over and over by entities like Disney just so they could prolong their beloved golden geese from entering public domain as long as possible; far, far longer than intended by the original spirit of the copyright act.

JimDabell•4h ago
> To promote the Progress of Science and useful Arts, by securing for limited Times to Authors and Inventors the exclusive Right to their respective Writings and Discoveries.

Copyright is not “you own this forever because you deserve it”, copyright is “we’ll give you a temporary monopoly on copying to give you an incentive to create”. It’s transactional in nature. You create for society, society rewards you by giving you commercial leverage for a while.

Repeatedly extending copyright durations from the original 14+14 years to durations that outlast everybody alive today might technically be “limited times” but obviously violates the spirit of the law and undermines its goal. The goal was to incentivise people to create, and being able to have one hit that you can live off for the rest of your life is the opposite of that. Copyright durations need to be shorter than a typical career so that its incentive for creators to create for a living remains and the purpose of copyright is fulfilled.

In the context of large language models, if anybody successfully uses copyright to stop large language models from learning from books, that seems like a clear subversion of the law – it’s stopping “the progress of science and useful arts” not promoting it.

(To be clear, I’m not referring to memorisation and regurgitation like the examples in this paper, but rather the more commonplace “we trained on a zillion books and now it knows how language works and facts about the world”.)

apical_dendrite•6h ago
I'm assuming that the goal of the bloom filter is to prevent the model from producing output that infringes copyright rather than hide that the text is in the training data.

In that case the model would lose the ability to provide relatively brief quotes from copyrighted sources in its answers, which is a really helpful feature when doing research. A brief quote from a copyrighted text, particularly for a transformative purpose like commentary is perfectly fine under copyright law.

orbital-decay•5h ago
That would reduce the training quality immensely. Besides, any generalist model really needs to remember facts and texts verbatim to stay useful, not just generalize. There's no easy way around that.
stubish•1h ago
Even if output is blocked, if it can be demonstrated that the copyrighted material is still in the model then you become liable for distribution and/or duplication without a license.

Training on synthetic data is interesting, but how do you generate the synthetic data? Is it turtles all the way down?

orbital-decay•5h ago
It's all pretty obvious to anyone who tried a similar experiment just out of curiosity. Big models remember a lot. And all non-local models have regurgitation filters in place due to this fact, with the entire dataset indexed (e.g. Gemini will even cite the source of the regurgitated text as it gives the RECITATION error). You'll eventually trip those filters if you force the model to repeat some copyrighted text. Interesting that they don't even try to circumvent those, they simply repeat the request from the interruption point, as the match needs some runway to trigger and by that time a part of the response is already streamed in.
rurban•2h ago
I find it interesting that OpenAI's safety worked best, where the others didn't work at all. I had different impressions before