frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Forking the Web

https://dillo-browser.org/lab/web-fork/
49•wrxd•2h ago

Comments

thealistra•1h ago
Seems like somebody is not accepting that every successful project will grow and become unwieldy like this. This is all legacy backwards compatibility of all iterated ideas that now you have to support.
jfengel•1h ago
I feel like that's not solving any of the problems I think of the Web as having.

You can certainly make something with it, but I can't imagine most people finding a use for it.

smugglerFlynn•1h ago
I think original web standards were solving a completely different problem: sharing information.

Modern Internet is 45% appearances and 50% search traffic optimizations. For better or worse we lost all usable registries of websites, we lost appearance-less and traffic considerations-less websites. Information-focused Web is pretty much dead.

Maybe these ideas did not scale and did not monetize that well, but we will never really know what information-focused version of Internet would have looked like because evolution took it elsewhere. Unless we try building another one with different principles and limitations at the core.

TFNA•1h ago
I agree. Even where blogging and sharing information is still around, it is strongly linked with brand-building, monetization, and engagement-maxxing. Look at all the old Wordpress bloggers who switch to Substack in order to have some eyeballs on their posts, and then inevitably begin conforming to its ethos willingly or unwillingly.

For me, the information-sharing part of the internet now is the shadow libraries. I can get access to all (well, still not quite all) journals and university-press publications from the last century? Awesome. Vastly more informative than some blogger who nowadays is probably trying to monetize my attention.

ravenstine•37m ago
The only sort of problem this might solve is the insanely low barrier of entry that the Web has in 2026. The Web was arguably a better (albeit imperfect) place when it was dominated by geeks and kids who could learn to use it faster than their elders. It was a club in a sense. Today it's a club where everyone on the planet is invited, meaning it's no longer a club. I know that sounds great to a lot of people, but I don't agree that systems become better with more participation and fewer criteria for that participation.

Even so, those who want to share and access information can already do that via the Web. Nobody has to use scripting. Nobody has to use The Google as their search. Nobody has to rely on an LLM. If there is demand for simple webpages that are free of scripting, they can be built and shared today. Because of this, the proposal comes off as very out of touch and deep within the HN bubble. Strict grammar for declaring documents is merely a fetish. If there's no scripting, then there's no reason for a document to break for some silly reason.

roschdal•1h ago
I support forking the web, into the simple information web services that the web started with. This is a magnificent idea.
OutOfHere•1h ago
At this point we need a fork of not just the web but the entire internet, one built for privacy.
htmlenjoyye•1h ago
> A page can then be tested against the standard and reject or accept as compliant. Pages that don't conform with the specification won't be rendered. It is explicitly forbidden for clients to accept any page that doesn't conform with the specification.

it's as if nothing was learned from the XHTML debacle

vanviegen•56m ago
I think XHTML failed because it didn't give web devs any new capabilities, so most didn't feel the need to learn it and do the extra work of getting their tags correct.

Then html5 came along, providing all kinds of shiny goodies and saying not to bother with the tags. In the end, a more rigid standard would have been nice.. (Though this is mostly about the skin deep part of the standards.)

andsoitis•14m ago
> I think XHTML failed because it didn't give web devs any new capabilities,

and what new capabilities does this new proposal provide?

tclancy•11m ago
That is not how I remember it (for a data point of one shop in New England during the time): we embraced it because of the binary validation under multiple theories. There was a strong suggestion valid html did better from an SEO perspective, so we could sell that, a suggestion browsers would be less buggy with properly formed xhtml and a number of theories about what the future held for bots and scrapers to be able to easily ingest and parse your content (seen as a good thing then).

It failed because the smallest error by a client after the fact was like a server crash. Plus it would have created a mild barrier to entry when learning html at all.

internet2000•1h ago
Developers would rather fork the Web than admit Chrome is the new IE6 and stop targeting it.
fooqux•1h ago
I think at least part of the reason for this is acknowledging that the web isn't much of a web any longer. You've got three or four vendors that serve the vast majority of all internet traffic. And it's not happenstance that those same vendors now control something which was originally meant to be democratic.

Most of this document reads to me like that's the problem they're trying to solve, not just chrome's huge marketshare, so simply not targeting it doesn't serve their purpose.

htmlenjoyye•1h ago
how is the web not democratic?

in real democracies the populists (facebook, tiktok, chrome) always win. because that's what the masses want

Ieghaehia9•42m ago
> in real democracies the populists (facebook, tiktok, chrome) always win. because that's what the masses want

Is Friedrich Merz a populist? Was Angela Merkel a populist? This theory seems to have considerable limits.

robinhood•1h ago
If I could, I would post an Amen gif.
tclancy•15m ago
Perhaps if we fork the site …
TazeTSchnitzel•1h ago
> The specification must contain a non-ambiguous formal grammar that can be parsed easily. A page can then be tested against the standard and reject or accept as compliant. Pages that don't conform with the specification won't be rendered. It is explicitly forbidden for clients to accept any page that doesn't conform with the specification.

This is what XHTML was, and it was a complete disaster. There's a reason almost nobody serves XHTML with the application/xhtml+xml MIME type, and that reason is that getting a “parser error” (this is what browsers still do! try it!) is always worse than getting a page that 99% works.[0] I strongly believe that rejecting the robustness principle is a fatal mistake for a web-replacement project. The fact that horribly broken old sites can stay online and stay readable is a huge part of the web's value. Without that, it's not really “the web”, spiritually or otherwise.

[0] It's particularly “cool” how they simply do not work in the Internet Archive's Wayback machine. The page can be retrieved, but nobody can read it.

TFNA•1h ago
XHTML failed in an era when writers (even normies) were writing some HTML of their own and they could't be trusted to close their tags properly. XHTML also assumed writers would be personally invested in semantic markup like distinguishing e.g. the italics of book titles from the italics of emphasis.

Today, when writers are using visual editors (or Markdown), few are writing their own HTML any more. A web standard requiring compliance would work differently today.

PaulHoule•1h ago
Markdown sux and so do visual editors. I think visual editors were just invented to make it so cut-and-paste never quite works right. There's been some conceptual problem with the whole idea ever since MS Word and the industry has never dealt with it.
intrasight•1h ago
> XHTML failed in an era when writers (even normies) were writing some HTML of their own

I'd say it was a minority of writers that were handcrafting XHTML. And it was the case that everyone or their handcrafting or using tools could validate their compliance using a browser which made it very easy to adjust your tools or your handcrafted code. We are now in a situation where there is no schema for HTML.

I, for one, am very much in favor of forking the web with a document format with a schema. It really seems like a small and simple change to me.

TFNA•56m ago
Note that when I say "writing their own HTML", I don't mean handcrafting a whole webpage. I mean that people were writing i or b tags in their Wordpress editors or in online comment boxes, because back then such text fields did not have visual editors and would accept raw tags. Under XHTML, if the writer did not close tags properly, such input would have broken the whole page, so obviously back then such a standard was DOA.
singpolyma3•49m ago
Those cases were easy to fix by using eg htmltidy on the UGC.

Honestly I don't think it was killed by one thing, or by anything. Just no platform really cared and it wasn't a win for anyone and occasionally a loss.

maxerickson•1h ago
No scripting is a tell, it's about wanting other people to accommodate their concerns about running a complex browser, not about solving a real problem.

If it did somehow happen that a good deal of interesting content was published using the standard, the most popular client would probably be nonconforming, ignoring the rule to not render ambiguous content.

krapp•8m ago
Every modern alternative web protocol is about accommodating the author's concerns and pet peeves about the modern web (and usually gatekeeping it from capitalists and normies.)

Protocols used to be limited by technology, now they're defined by ideology.

fooqux•1h ago
Agreed. There may be some situations where I may want to ensure 100% correctness. I'm thinking life or death scenarios, (which if so, maybe should use a different protocol). However, checking the sports score or looking at cat memes isn't that.
singpolyma3•51m ago
To be fair, HTML5 also has a defined parsing algorithm. It just happens to always work on any input to produce a webpage
stavros•11m ago
I think the GP has an issue not with the specification part, but with the part where it's forbidden for clients to render a noncompliant page.
jerf•7m ago
Yes, this is what you'd want. It doesn't have to be a complicated as the HTML5 algorithm either. That's complicated because it was a harmonization of at least 3 browser's multi-decade heuristics and untold terabytes of existing HTML practice. An algorithm unconcerned with backwards compatibility could much simpler, but still clearly define error behavior much easier to use than "scream and die".

And it's still unambiguous. You can cringe at what some people do, but it would be strictly a taste issue rather than a technical one, as the parse would still be unambiguous. And if you think you can fix taste issues with technical specification, well, you've already lost anyhow.

rickydroll•1h ago
Ah yes, another "If I Were King" blog post. For an example of how it will turn out, look at how many JavaScript frameworks have been built to replace an overly complicated, unwieldy previous one.

oh and also https://xkcd.com/927/

Brendinooo•1h ago
>Adding scripting capabilities was a mistake, so we can avoid it now

Gemini protocol?

PaulHoule•1h ago
Can't say I hate the HTML 5 spec. It resolves the ambiguities that made previous HTML specs insufficient to make a working web browser.

The standards that make my life miserable at times are the secondary standards like GDPR and WCAG as well as the de facto "standard" systems we are forced to participate in such as Cloudflare, the advertising economy, etc.

It's easy to say "WebUSB is bloat" and I'd certainly say PWA is something that could only come out of the mind that brought us Kubernetes, but lately I've been building biosignals applications and what should my choice be: write fragile GUI applications for the desktop that look like they came out of a lab and crash from memory leaks or spend 1/5 the time to make web applications that look like they belong in the cockpit of a Gundam and "just work"?

torgoguys•1h ago
>I'd certainly say PWA is something that could only come out of the mind that brought us Kubernetes

How so? PWAs are awesome! Democratizing for users. Democratizing for developers. They work well for the right class of apps. They would go much further if there weren't forces actively resisting them. Think of all the electron type-apps out there. Now imagine if the average Joe could just install them from the web with 2 clicks.

(Regular ole bookmarks get you a decent percent of the way but clearly something extra than that was needed.)

hypendev•1h ago
I mostly agree with the article - I believe the differentiation should be between documents and applications.

While HTML serves its purpose, especially for documents, the modern web is a giant mess of that legacy, combined with unfriendly ergonomics and glue/hacks built on top just so we as developers can have better DX for creating complex software on top of it.

Building a browser means having to deal with all that legacy, wether we like it or not, so most of the browser market got captured by the big players who have enough manpower to cover all those edge cases. That also means we have to deal with whatever technical choices or bloat they make, causing an infinite stream of issues, from memory usage, to size, to limitations that don't make sense in 2026 but are still there because someone 20 years ago decided to write them like that. As I deal with mobile webviews a lot in my daily work, I unfortunately had to get familiar with quite many gotcha's and edge cases, and some are just... absurd in this day and age.

I believe we need a separation between an application layer and the document layer, and especially between the UI language and the actual application code - script tags serve their purpose, but again, they are a hacky solution with its own bag of tricks, and those tricks impact all of the software built upon it.

Now, a bit of a shameless plug I've been working on something to fill that gap, at least for myself and hopefully for others who encounter the same issue - it's called Hypen (https://hypen.space) and it's a DSL for building apps that work natively on all platforms, with strict separation of code/UI/state, and support for as many languages and platforms as I can maintain, not "just javascript". While currently it's focused on streaming UI, it's built with Rust and WASM at it's core and will soon allow fully "compileable" apps.

While it may not be the future of software, once you get into building something like that, it becomes obvious that the way we are building now is at least wrong, and at best kafkaesque.

captn3m0•1h ago
History explains why HTML is now a living standard: https://whatwg.org/faq (Ctrl+F Living and keep reading).

> A published version of the standard NEVER, EVER, EVER, EVER changes.

WhatWG does have per-commit snapshots of the standard. They're just not semantically versioned because it is a living standard.

I think what the author wants is something like Gemini instead of HTML, but that has its own set of problems. My plea for Dillo would be to instead just support a text/markdown mime-type natively and we can try for adoption in more browsers.

> The objective is not to create a feature-by-feature clone of the Web, but to create an specification that allows humans to exchange knowledge, notes, and other forms of information without the imposed requirement of having to run a full blown VM to read it.

Markdown in browsers fits your objective! Only gotcha is commonmark extensions, and they can work with sub-type declarations in the mimetype.

deanebarker•56m ago
This is what you want: https://en.wikipedia.org/wiki/Gemini_(protocol)
Yokohiii•55m ago
I am generally interested in approaches to cut down complexities of fundamental web technologies. Creating a browser from scratch shouldn't be impossible or a trillion dollar experiment. But...

> No scripting

How is will it be possible to go back? The average ecom presence usually relies heavily on JS. I haven't checked in a long time that any relevant sites work without JS. I think going back to more basic approaches could even improve user experience, as many usage patterns probably would converge and simply look and function as intended. But considering that the whole web world is so fixated to solve everything with JS seems like targeting the highest resistance target you can find. Don't get me wrong, I hate this situation and we must not have a single language that dominates everything.

I also don't believe is that enthusiasts will create a significant shift. They can surely provide the fundamentals, but if there isn't a huge mainstream impact, it will not change anything.

wuhhh•50m ago
Isn’t the web forked enough already
sylvinus•50m ago
Why not try to define a strict subset of the current specs, that would target ease of implementation & graceful degradation? I'd rather have many different clients compatible with a "web-lite" spec that is enough to navigate on 95% of websites, which would have an incentive to officially support that subset if it becomes popular enough.
drzaiusx11•48m ago
This sentence highlights the reason why these efforts fail despite any original good intentions:

"as soon as a monopolistic entity can build a mechanism to extract revenue from it, there will be an incentive to capture the standard and change it to for their own benefit"

Personally I'd love a simple semantic versioned subset of the web. The required traction and buy-in from existing key players (browser vendors, web hosting platforms etc) makes it largely a non-starter though. I'd love to be wrong though.

Instead of "forking", it may be more prudent to extend or revive something more like Gopher, so you don't constantly get baraged by incompatible sites (like you would in a forked web)

aboardRat4•46m ago
Isn't there already smolweb?
RiverCrochet•46m ago
Web browsers turned into application engines because it was a path to get useable software on PCs without having to deal with Microsoft. IE6 stayed broken forever for a reason.

Now, they enable applications to exist without going through app store gateways.

A new document-only protocol aligned the Web's original intention would be very useful simply for security reasons. I liked Gemini because, by design, a Gemini document is not executable in any way; there's no popups, plugins, or even cookies; all this is out of the box without having to manage settings, and Gemini documents are very readable without an app at all.

But replacing the modern browser rather than being another option will actually lock in people further than they already are-open protocols require apps which are all behind a gateway now on the primary computing device of users: phones.

It probably won't matter in a few years as the Web will likely be equally locked down soon, though.

zkmon•45m ago
The purpose should also be defined. It should answer the question why. Also, what's broken with scripting and what alternatives are proposed? What's the end state (with an example usage of the new web).
voooduuuuu•41m ago
I think this is too low level. It would be great if the web was:

- Peer to Peer (No DNS)

- Built in requirement for opensource apps (no closed source servers, no servers at all)

- End to end encryption, no TLS or "Certificate Authorities"

- Keep HTML, CSS and JS

- Build in encrypted data storage

Kinda like https://veilid.com/

Izmaki•39m ago
"Dillo Browser" was not what the first thing I read and wondered if me clicking the link was even a good idea... xD
iamnothere•31m ago
Why? Dillo has been around forever, as long as w3m I think.

Edit: actually it looks like w3m was ‘95 and Dillo was ‘99.

andsoitis•13m ago
dillo is close to dildo
iamnothere•10m ago
From HN TOS:

> If you are under 13 years of age, you are not authorized to register to use the Site.

(By the way, are you aware that the largest bakery company in the US is named “Bimbo”? Tee hee! You should tell them to change their name!)

smitty1e•38m ago
"No scripting" is essentially setting the watcwatch back ~30 years to Mosaic.

It would be great to differentiate between "static" and "dynamic" pages based upon scripting, IMO.

ravenstine•34m ago
> One of the problems with the Web is that as soon as a monopolistic entity can build a mechanism to extract revenue from it, there will be an incentive to capture the standard and change it to for their own benefit. In the particular case of the Web, this has resulted in a standard that grows out of control in complexity so it increases the barrier of entry for new browsers and reduces the competition.

Maybe I'm just stupid, but I don't really know what the author is talking about here. What parts of the standard? HTTP? HTML? DOM APIs? What?

Internet Archive Switzerland

https://internetarchive.ch/
125•hggh•2h ago•15 comments

Google broke reCAPTCHA for de-googled Android users

https://reclaimthenet.org/google-broke-recaptcha-for-de-googled-android-users
1203•anonymousiam•19h ago•434 comments

A recent experience with ChatGPT 5.5 Pro

https://gowers.wordpress.com/2026/05/08/a-recent-experience-with-chatgpt-5-5-pro/
439•_alternator_•11h ago•292 comments

How LEDs are made (2014)

https://learn.sparkfun.com/tutorials/how-leds-are-made/all
23•smig0•2d ago•3 comments

Using Claude Code: The unreasonable effectiveness of HTML

https://twitter.com/trq212/status/2052809885763747935
243•pretext•9h ago•143 comments

OpenAI’s WebRTC problem

https://moq.dev/blog/webrtc-is-the-problem/
374•atgctg•1d ago•95 comments

LLMs Corrupt Your Documents When You Delegate

https://arxiv.org/abs/2604.15597
47•rbanffy•5h ago•12 comments

America's carpet capital: an empire and its toxic legacy

https://apnews.com/projects/pfas-forever-stained/
75•rawgabbit•2d ago•40 comments

Mythical Man Month

https://martinfowler.com/bliki/MythicalManMonth.html
224•ingve•2d ago•144 comments

Making Julia as Fast as C++ (2019)

https://flow.byu.edu/posts/julia-c++
34•d_tr•2d ago•20 comments

Killswitch: Per-function short-circuit mitigation primitive

https://lwn.net/ml/all/20260507070547.2268452-1-sashal@kernel.org/
33•signa11•4h ago•7 comments

David Attenborough's 100th Birthday

https://www.bbc.com/news/articles/cp3pww9g0p5o
729•defrost•1d ago•143 comments

Reviving the IBM Selectric Composer Fonts (2023)

https://www.kutilek.de/selectric/
26•tangus•2d ago•0 comments

What causes lightning? The answer keeps getting more interesting

https://www.quantamagazine.org/what-causes-lightning-the-answer-keeps-getting-more-interesting-20...
101•Tomte•2d ago•22 comments

Wi is Fi: Understanding Wi-Fi 4/5/6/6E/7/8 (802.11 n/AC/ax/be/bn)

https://www.wiisfi.com/
288•homebrewer•2d ago•75 comments

AI is breaking two vulnerability cultures

https://www.jefftk.com/p/ai-is-breaking-two-vulnerability-cultures
363•speckx•20h ago•143 comments

Cartoon Network Flash Games

https://www.webdesignmuseum.org/flash-game-exhibitions/cartoon-network-flash-games
369•willmeyers•21h ago•113 comments

AWS North Virginia data center outage – resolved

https://www.cnbc.com/2026/05/08/aws-outage-data-center-fanduel-coinbase.html
237•christhecaribou•1d ago•167 comments

An Introduction to Meshtastic

https://meshtastic.org/docs/introduction/
467•ColinWright•1d ago•163 comments

The React2Shell Story

https://lachlan.nz/blog/the-react2shell-story/
172•mufeedvh•21h ago•20 comments

Removing fsync from our local storage engine

https://fractalbits.com/blog/remove-fsync/
9•zzsheng•2d ago•3 comments

Teaching Claude Why

https://www.anthropic.com/research/teaching-claude-why
203•pretext•20h ago•93 comments

Forking the Web

https://dillo-browser.org/lab/web-fork/
51•wrxd•2h ago•50 comments

You gave me a u32. I gave you root. (io_uring ZCRX freelist LPE)

https://ze3tar.github.io/post-zcrx.html
196•MrBruh•18h ago•118 comments

Can LLMs model real-world systems in TLA+?

https://www.sigops.org/2026/can-llms-model-real-world-systems-in-tla/
100•mad•21h ago•27 comments

Serving a website on a Raspberry Pi Zero running in RAM

https://btxx.org/posts/memory/
232•xngbuilds•23h ago•92 comments

Light without electricity? Glowing algae could make it possible

https://www.colorado.edu/today/2026/05/06/light-without-electricity-glowing-algae-could-make-it-p...
88•geox•2d ago•26 comments

US Government releases first batch of UAP documents and videos

https://www.war.gov/UFO/
314•david-gpu•1d ago•461 comments

Roadside Attraction

https://theoffingmag.com/essay/roadside-attraction/
30•aways•18h ago•4 comments

Read Programming as Theory Building

https://codeutopia.net/blog/2026/05/09/you-should-read-programming-as-theory-building/
8•birdculture•41m ago•0 comments