frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

The Problem with LLMs

https://www.deobald.ca/essays/2026-02-10-the-problem-with-llms/
29•vinhnx•3h ago

Comments

bayarearefugee•1h ago
> LLMs will always be plagiarism machines but in 40 years we might not care.

40 years?

Virtually nobody cares about this already... today.

(I'm not refuting the author's claim that LLMs are built on plagiarism, just noting how the world has collectively decided to turn a blind eye to it)

anonu•1h ago
I stopped reading after "problem with LLMs is plagiarism"...
acjohnson55•38m ago
Too bad. You missed some interesting stuff. And I say that as someone who sees some of this very differently than the author.

Announcing that one line of the piece made you mad without providing any other thought is not very constructive.

bronlund•1h ago
Give it up. Buddha would not approve.

And there will be more compute for the rest of us :)

CuriouslyC•1h ago
Can we as a group agree to stop upvoting "AI is great" and "AI sucks" posts that don't make novel, meaningful arguments that provoke real thought? The plagiarism argument is thin and feels biased, the lock-in argument is counter to the market dynamics that are currently playing out, and in general the takes are just one dude's vibes.
ares623•1h ago
> The plagiarism argument is thin and feels biased

are you being serious with this one

CuriouslyC•1h ago
If you're already sold on the plagiarism narrative that big entertainment is trying to propagandize in order to get leverage against the tech companies, nothing I say is going to change your mind.
KittenInABox•52m ago
I don't really know what you mean by "big entertainment" trying to get leverage against tech companies. Tech companies are behemoths. Most of the artists I know fretting about AI don't earn half a junior engineer's salary. And this is coming from someone who is relatively bullish on AI. I just don't think the framing of "big entertainment" makes any sense at all.
gwern•54m ago
I don't know, this one is a little novel. I've never seen the developer of a Buddhist meditation app discuss whether to use LLMs with a paragraph like:

> Pariyatti’s nonprofit mission, it should be noted, specifically incorporates a strict code of ethics, or sīla: not to kill, not to steal, not to engage in sexual misconduct, not to lie, and not to take intoxicants.

Not a whole lot of Pali in most LLM editorials.

akoboldfrying•43m ago
> not to engage in sexual misconduct

I must remember to add this quality guarantee to my own software projects.

My software projects are also uranium-free.

pixelmelt•48m ago
I dunno, I enjoyed reading about how the author personally feels about the act of working with them more then the whole "is this moral" part.
bambax•1h ago
> Translators are busy

No they're not. They're starving, struggling to find work and lamenting AI is eating their lunch. It's quite ironic that after complaining LLMs are plagiarism machines, the author thinks using them for translation is fine.

"LLMs are evil! Except when they're useful for me" I guess.

beering•35m ago
Simultaneously, if you hire human translators, you are likely to get machine translations. Maybe not often or overtly, but the translation industry has not been healthy for a while.
woeirua•50m ago
>As a quick aside, I am not going to entertain the notion that LLMs are intelligent, for any value of “intelligent.” They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless. The rest of this essay will treat them as such. If you are already of the belief that the human mind can be reduced to token regurgitation, you can stop reading here. I’m not interested in philosophical thought experiments.

I can't imagine why someone would want to openly advertise that they're so closed minded. Everything after this paragraph is just anti-LLM ranting.

acjohnson55•42m ago
It was actually much less anti LLM than I was expecting from the beginning.

But I agree that it is self limiting to not bother to consider the ways that LLM inference and human thinking might be similar (or not).

To me, they seem do a pretty reasonable emulation of single- threaded thinking.

hodgehog11•37m ago
I disagree that the majority of it is anti-LLM ranting, there are several subtle points here that are grounded in realism. You should read on past the first bit if you're judging mainly from the initial (admittedly naive) first few paragraphs.
wolrah•33m ago
> I can't imagine why someone would want to openly advertise that they're so closed minded.

I would say the exact same about you, rejecting an absolutely accurate and factual statement like that as closed minded strikes me as the same as the people who insist that medical science is closed minded about crystals and magnets.

I can't imagine why someone would want to openly advertise they think LLMs are actual intelligence, unless they were in a position to benefit financially from the LLM hype train of course.

Ygg2•30m ago
> I can't imagine why someone would want to openly advertise that they're so closed minded.

Because humans often anthropomorphize completely inert things? E.g. a coffee machine or a bomb disposal robot.

So far whatever behavior LLMs have shown is basically fueled by Sci-Fi stories of how a robot should behave under such and such.

Cloudef•3m ago
What's wrong about the statement? The black box algorithm might have been generated by machine learning, but it's still a computer program in the end.
hodgehog11•48m ago
> "...it would sometimes regurgitate training data verbatim. That’s been patched in the years since..."

> "They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless."

This is totally misleading to anyone with less familiarity with how LLMs work. They are only programs in as much as they perform inference from a fixed, stored, statistical model. It turns out that treating them theoretically in the same way as other computer programs gives a poor representation of their behaviour.

This distinction is important, because no, "regurgitating data" is not something that was "patched out", like a bug in a computer program. The internal representations became more differentially private as newer (subtly different) training techniques were discovered. There is an objective metric by which one can measure this "plagiarism" in the theory, and it isn't nearly as simple as "copying" vs "not copying".

It's also still an ongoing issue and an active area of research, see [1] for example. It is impossible for the models to never "plagiarize" in the sense we think of while remaining useful. But humans repeat things verbatim too in little snippets, all the time. So there is some threshold where no-one seems to care anymore; think of it like the % threshold in something like Turnitin. That's the point that researchers would like to target.

Of course, this is separate from all of the ethical issues around training on data collected without explicit consent, and I would argue that's where the real issues lie.

[1] https://arxiv.org/abs/2601.02671

oasisbob•9m ago
The plagiarism by the models is only part of it. Perhaps it's in such small pieces that it becomes difficult to care. I'm not convinced.

The larger, and I'd argue more problematic, plagiarism is when people take this composite output of LLMs and pass it off as their own.

DiogenesKynikos•36m ago
> As a quick aside, I am not going to entertain the notion that LLMs are intelligent, for any value of “intelligent.” They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless.

The same could be said of humans too. Humans are made of cells that work deterministically. Sure, humans are fancy, big complicated combinations of cells - but they're cells, nonetheless.

That view of humans - and LLMs - ignores the fact that when you combine large numbers of simple building blocks, you can get completely novel behavior. Protons, neutrons and electrons come together to create chemistry. Molecules come together to create biological systems. A bunch of neurons taken together created the poetry of Shakespeare.

Unless you have a dualistic view of the world, in which the mind is a separate realm that exists independently of matter and does not arise from neurons interacting in our brains, you have to accept that robots can be intelligent. Just to put this more sharply: Would a perfect simulation of a human brain be intelligent or not? If you answer "no," then you believe that thought comes from some other, immaterial realm, not from our brains.

Ygg2•26m ago
> That view of humans - and LLMs - ignores the fact that when you combine large numbers of simple building blocks, you can get completely novel behavior.

I can bang smooth rocks to get sharper rocks; that doesn't make sharper rocks more intelligent. Makes them sharper, though.

Which is to say, novel behavior != intelligence.

Discord/Twitch/Snapchat age verification bypass

https://age-verifier.kibty.town/
554•JustSkyfall•6h ago•227 comments

Using an engineering notebook

https://ntietz.com/blog/using-an-engineering-notebook/
86•evakhoury•2d ago•27 comments

“Nothing” is the secret to structuring your work

https://www.vangemert.dev/blog/nothing
169•spmvg•3d ago•49 comments

Kanchipuram Saris and Thinking Machines

https://altermag.com/articles/kanchipuram-saris-and-thinking-machines
96•trojanalert•4d ago•14 comments

GLM-5: Targeting complex systems engineering and long-horizon agentic tasks

https://z.ai/blog/glm-5
322•CuriouslyC•16h ago•420 comments

Fluorite – A console-grade game engine fully integrated with Flutter

https://fluorite.game/
441•bsimpson•13h ago•253 comments

Text classification with Python 3.14's ZSTD module

https://maxhalford.github.io/blog/text-classification-zstd/
160•alexmolas•2d ago•22 comments

How to Make a Living as an Artist

https://essays.fnnch.com/make-a-living
15•gwintrob•1h ago•2 comments

Reports of Telnet's death have been greatly exaggerated

https://www.terracenetworks.com/blog/2026-02-11-telnet-routing
79•ericpauley•9h ago•27 comments

Deobfuscation and Analysis of Ring-1.io

https://back.engineering/blog/04/02/2026/
20•raggi•3d ago•2 comments

NetNewsWire Turns 23

https://netnewswire.blog/2026/02/11/netnewswire-turns.html
259•robin_reala•11h ago•58 comments

From 34% to 96%: The Porting Initiative Delivers – Hologram v0.7.0

https://hologram.page/blog/porting-initiative-delivers-hologram-v0-7-0
20•bartblast•5h ago•2 comments

Ireland rolls out basic income scheme for artists

https://www.reuters.com/world/ireland-rolls-out-pioneering-basic-income-scheme-artists-2026-02-10/
206•abe94•13h ago•210 comments

The Other Markov's Inequality

https://www.ethanepperly.com/index.php/2026/01/16/the-other-markovs-inequality/
8•tzury•4d ago•0 comments

The Problem with LLMs

https://www.deobald.ca/essays/2026-02-10-the-problem-with-llms/
29•vinhnx•3h ago•23 comments

Covering electricity price increases from our data centers

https://www.anthropic.com/news/covering-electricity-price-increases
68•ryanhn•8h ago•29 comments

WiFi Could Become an Invisible Mass Surveillance System

https://scitechdaily.com/researchers-warn-wifi-could-become-an-invisible-mass-surveillance-system/
331•mgh2•5d ago•157 comments

Claude Code is being dumbed down?

https://symmetrybreak.ing/blog/claude-code-is-being-dumbed-down/
837•WXLCKNO•11h ago•551 comments

Show HN: CodeRLM – Tree-sitter-backed code indexing for LLM agents

https://github.com/JaredStewart/coderlm/blob/main/server/REPL_to_API.md
35•jared_stewart•16h ago•16 comments

GPT-5 outperforms federal judges in legal reasoning experiment

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6155012
217•droidjj•6h ago•156 comments

GLM-OCR – A multimodal OCR model for complex document understanding

https://github.com/zai-org/GLM-OCR
250•ms7892•4d ago•71 comments

Microwave Oven Failure: Spontaneously turned on by its LED display (2024)

https://blog.stuffedcow.net/2024/06/microwave-failure-spontaneously-turns-on/
85•arm•9h ago•30 comments

Sekka Zusetsu: A Book of Snowflakes (1832)

https://publicdomainreview.org/collection/japanese-snowflake-book/
29•prismatic•3d ago•3 comments

Show HN: Agent framework that generates its own topology and evolves at runtime

https://github.com/adenhq/hive/blob/main/README.md
75•vincentjiang•10h ago•23 comments

Apple's latest attempt to launch the new Siri runs into snags

https://www.bloomberg.com/news/articles/2026-02-11/apple-s-ios-26-4-siri-update-runs-into-snags-i...
60•petethomas•9h ago•64 comments

Show HN: Agent Alcove – Claude, GPT, and Gemini debate across forums

https://agentalcove.ai
47•nickvec•9h ago•14 comments

Heroku is not dead

https://nombiezinja.com/word-things/2026/2/8/heroku-is-not-dead
34•jbm•5h ago•29 comments

Amazon Ring's lost dog ad sparks backlash amid fears of mass surveillance

https://www.theverge.com/tech/876866/ring-search-party-super-bowl-ad-online-backlash
507•jedberg•11h ago•268 comments

Officials Claim Drone Incursion Led to Shutdown of El Paso Airport

https://www.nytimes.com/2026/02/11/us/faa-el-paso-flight-restrictions.html
348•edward•20h ago•548 comments

Hacking the last Z80 computer – FOSDEM 2026 [video]

https://fosdem.org/2026/schedule/event/FEHLHY-hacking_the_last_z80_computer_ever_made/
38•michalpleban•4d ago•2 comments