frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

You can now play Grand Theft Auto Vice City in the browser

https://dos.zone/grand-theft-auto-vice-city/
63•Alifatisk•1h ago•20 comments

TP-Link Tapo C200: Hardcoded Keys, Buffer Overflows and Privacy

https://www.evilsocket.net/2025/12/18/TP-Link-Tapo-C200-Hardcoded-Keys-Buffer-Overflows-and-Priva...
119•sibellavia•2h ago•21 comments

Garage – An S3 object store so reliable you can run it outside datacenters

https://garagehq.deuxfleurs.fr/
292•ibobev•4h ago•56 comments

GotaTun -- Mullvad's WireGuard Implementation in Rust

https://mullvad.net/en/blog/announcing-gotatun-the-future-of-wireguard-at-mullvad-vpn
451•km•9h ago•98 comments

Amazon will allow ePub and PDF downloads for DRM-free eBooks

https://www.kdpcommunity.com/s/article/New-eBook-Download-Options-for-Readers-Coming-in-2026?lang...
427•captn3m0•10h ago•236 comments

Vm.overcommit_memory=2 is always the right setting for servers

https://ariadne.space/2025/12/16/vmovercommitmemory-is-always-the-right.html
11•signa11•2d ago•8 comments

Mistral OCR 3

https://mistral.ai/news/mistral-ocr-3
28•pember•1d ago•0 comments

The FreeBSD Foundation's Laptop Support and Usability Project

https://github.com/FreeBSDFoundation/proj-laptop
102•mikece•5h ago•40 comments

Reverse Engineering US Airline's PNR System and Accessing All Reservations

https://alexschapiro.com/security/vulnerability/2025/11/20/avelo-airline-reservation-api-vulnerab...
47•bearsyankees•2h ago•22 comments

Believe the Checkbook

https://robertgreiner.com/believe-the-checkbook/
70•rg81•4h ago•26 comments

Where Is GPT in the Chomsky Hierarchy?

https://fi-le.net/chomsky/
36•fi-le•4d ago•27 comments

Show HN: Stickerbox, a kid-safe, AI-powered voice to sticker printer

https://stickerbox.com/
6•spydertennis•36m ago•1 comments

Graphite Is Joining Cursor

https://cursor.com/blog/graphite
83•fosterfriends•4h ago•125 comments

Proton Leaves Switzerland

https://www.nzz.ch/technologie/proton-ceo-andy-yen-wer-gesetzgebung-der-polizei-ueberlaesst-sollt...
99•_tk_•1h ago•39 comments

Lite^3, a JSON-compatible zero-copy serialization format

https://github.com/fastserial/lite3
76•cryptonector•6d ago•26 comments

Show HN: I Made Loom for Mobile

https://demoscope.app
37•admtal•3h ago•27 comments

Prepare for That Stupid World

https://ploum.net/2025-12-19-prepare-for-that-world.html
115•speckx•3h ago•64 comments

Show HN: MCPShark Viewer (VS Code/Cursor extension)- view MCP traffic in-editor

16•mywork-dev•2d ago•0 comments

Rust's Block Pattern

https://notgull.net/block-pattern/
19•zdw•15h ago•4 comments

Wall Street Ruined the Roomba and Then Blamed Lina Khan

https://www.thebignewsletter.com/p/how-wall-street-ruined-the-roomba
74•connor11528•1h ago•41 comments

We pwned X, Vercel, Cursor, and Discord through a supply-chain attack

https://gist.github.com/hackermondev/5e2cdc32849405fff6b46957747a2d28
1073•hackermondev•1d ago•393 comments

Building a Transparent Keyserver

https://words.filippo.io/keyserver-tlog/
40•noident•5h ago•14 comments

Show HN: Stepped Actions – distributed workflow orchestration for Rails

https://github.com/envirobly/stepped
70•klevo•5d ago•10 comments

1.5 TB of VRAM on Mac Studio – RDMA over Thunderbolt 5

https://www.jeffgeerling.com/blog/2025/15-tb-vram-on-mac-studio-rdma-over-thunderbolt-5
563•rbanffy•21h ago•207 comments

Show HN: TinyPDF – 3kb pdf library (70x smaller than jsPDF)

https://github.com/Lulzx/tinypdf
16•lulzx•1d ago•3 comments

History LLMs: Models trained exclusively on pre-1913 texts

https://github.com/DGoettlich/history-llms
700•iamwil•21h ago•343 comments

Getting bitten by Intel's poor naming schemes

https://lorendb.dev/posts/getting-bitten-by-poor-naming-schemes/
253•LorenDB•14h ago•136 comments

Show HN: Linggen – A local-first memory layer for your AI (Cursor, Zed, Claude)

https://github.com/linggen/linggen
10•linggen•2h ago•5 comments

Cycle-accurate YM2149 PSG emulator

https://github.com/slippyex/ym2149-rs
8•todsacerdoti•6d ago•1 comments

Noclip.website – A digital museum of video game levels

https://noclip.website/
420•ivmoreau•18h ago•53 comments
Open in hackernews

Where Is GPT in the Chomsky Hierarchy?

https://fi-le.net/chomsky/
36•fi-le•4d ago

Comments

PaulHoule•1h ago
The Chomsky Hierarchy isn't the right way to think about, in fact the Chomsky paradigm probably held back progress in language technology over the past 70 years.

In the Kuhnian sense, Syntactic Structures was the vanguard of a paradigm shift that let linguists do "normal science" in terms of positing a range of problems and writing papers about them. It was also useful in computer science for formally defining computer languages that are close enough to natural languages that people find them usable.

On the other hand, attempts to use the Chomsky theory to develop language technology failed miserably outside of very narrow domains and in the real world Markov models and conditional random fields often outperformed when the function was "understanding", "segmentation", etc.

From an engineering standpoint, the function that tells if a production is in the grammar that is central to the theory is not so interesting for language technology, I mean, if LLMs were -centric then an LLM would go on strike if you got the spelling or grammar wrong or correct you in a schoolmarmish William Safire way but rather it is more useful for LLMs to silently correct for obvious mistakes the same way that real language users do.

The other interesting thing is the mixing of syntax and semantics. For a long time I believed the "Language Instinct" idea of Pinker in which the ability to serialize and deserialize language is a peripheral that is attached to an animal and you need the rest of the animal and even the surrounding world to have "human" competence. Now LLMs come around and manage to show a lot of linguistic competence without the rest of the animal and boy we have egg on our face, and coming out of the shadow of the Chomsky theory, structuralism will rear it's head again. (e.g. "X is structured like a language" got hung up on the unspoken truth that "language is not structured like a language")

ogogmad•1h ago
I think most PL parsers use recursive descent, which doesn't require any formal language theory. That said, it's possible that formal language theory has enabled regular expressions to enter widespread use - first via theory; then to make tokenisers; then included in text editors; and then to make GREP.
Maxatar•54m ago
You're going to have a very hard time parsing a language using recursive descent if you don't understand some formal language theory. Recursive descent algorithms have a lot of trouble with left-recursion, and can result in arbitrary backtracking if you're not careful, which can blow up memory and time. To fix left-recursion you need to rewrite your grammar, which means familiarity with formal language theory. To avoid backtracking you typically use an LL(N) or variations thereof, which once again require having a pretty good understanding of formal language theory.
FeepingCreature•21m ago
Nah. Source: have done it without formal language theory. You can do most things by printf debugging and kludging.

(Left-recursion? Use a while loop! Mutable state is a pathway to many abilities functional programmers consider unnatural.)

DiscourseFan•38m ago
Don't worry, the LLMs are already post-structuralist (at least the ones working on them are often familiar with those thinkers). Which is incredibly disturbing, I think, for many linguists who have spent their entire careers fighting against the French, only to have such philosophical and intellectual work become the most important in the development of this new technology. Hey, at least I'm employed! (And they said it "wasn't serious," that Chomskian Linguistics was the "real science"!)
lukev•17m ago
Well, technically, they're more structuralist than post-structuralist. Arguably they even prove that Saussure's original structuralist hypothesis was correct.
czzr•17m ago
I don’t think LLMs contradict the Pinker description - it just turns out that the output stream embodies a lot of information, and you can construct a model (of some reasonable fidelity) of the original sources from sufficient examples of the output stream.
Maxatar•1h ago
So unless I'm mistaken, the TL;DR is that GPTs inherently can not be Turing complete because they always terminate, ie. there is always a non-zero probability of the end-of-token character to be generated.

Seems like a fair argument to raise, although not too insightful.

As a nitpick, it really is not the case that most programming languages are context-free, but I can sympathize with what the author is trying to get at. Most programming languages have a context-free superset that is useful to use as one of many stages in the parsing pipeline, before type checking and name resolution etc... but apart from perhaps some dynamically typed programming languages, the vast majority of programming languages are context-sensitive, not context-free.

onraglanroad•1h ago
> GPTs inherently can not be Turing complete because they always terminate

I'm not sure that was intended entirely seriously. After all, humans always terminate too...

bigfishrunning•47m ago
To follow on logically, maybe humans aren't Turing complete? we are not equivalent to Turing machines...
prolyxis•41m ago
Except that as far as I understand, one of the inspirations for the Turing machine is to explain precisely the computations a human computer could perform with (potentially a lot of) pen and paper.
stormking•33m ago
And an unlimited life span.
wat10000•36m ago
We can simulate a Turing machine, given storage. The infinite storage and infinite time is always a sticking point when comparing any real physical system to a theoretical Turing machine, so we tend to ignore those bits.
FeepingCreature•19m ago
No machine instantiated in the physical universe is Turing complete in this sense.
arjvik•1h ago
As a quick hack, is this flaw fixed by using top-p or top-k or min-p or whatever the current hottest logit sampling algorithm is?
FeepingCreature•18m ago
The "flaw" is also fixed by simply not recognizing the end-of-sampling token. Even the limited size of the context window can be worked around with tool calls.
Vosporos•1h ago
On Little Saint James, with Chomsky himself.
pohl•1h ago
Regular...albeit astronomically large (unless we're granting idealizations like infinite context, etc.)
Netcob•5m ago
Exactly, same as all real-world computers.

Although to be fair, nothing above regular (that I'm aware of, it's been a while) requires infinite space, just unlimited space... you can always make a real Turing machine as long as you keep adding more tape whenever it needs it.

aghilmort•1h ago
there’s decent work on computational reasoning power of transformers, SSMs, etc.

some approximate snippets that come to mind are that decoder-only transformers recognize AC^0 and think in TC^0, that encoder-decoders are strictly more powerful than decoder-only, etc.

Person with last name Miller iric if poke around on arXiv, a few others, been a while since was current top of mind so ymmv on exact correctness of above snippets

ctoa•1h ago
You are probably thinking of Merrill (whose work is referenced towards the end of the article).
guerrilla•1h ago
This is a trivial category mistake.
enriquepablo•5m ago
The Chomsky hierarchy is a classification of grammars. So it's about Syntax.

GPT is purely about semantics, about truth. To the extent that I think it can be said to be fundamentally informal in its dedication to truth, since it can easily be prompted to produce syntactically incorrect output - wrt the syntax of whatever language whose semantics determine the truth GPT is asked to reflect.

So I think this is a categorically incorrect question.

I mean, what is the fundamental syntax of the output of an LLM? It would have to be that any sequence of utf8 chars is valid.