frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

The Art of Multiprocessor Programming 2nd Edition Book Club

https://eatonphil.com/2025-art-of-multiprocessor-programming.html
161•eatonphil•4h ago•23 comments

Telo MT1

https://www.telotrucks.com/
59•turtleyacht•1h ago•46 comments

We may not like what we become if A.I. solves loneliness

https://www.newyorker.com/magazine/2025/07/21/ai-is-about-to-solve-loneliness-thats-a-problem
238•defo10•7h ago•465 comments

6 Weeks of Claude Code

https://blog.puzzmo.com/posts/2025/07/30/six-weeks-of-claude-code/
18•mike1o1•2d ago•71 comments

Show HN: WebGPU enables local LLM in the browser – demo site with AI chat

https://andreinwald.github.io/browser-llm/
83•andreinwald•4h ago•26 comments

Hiding secret codes in light protects against fake videos

https://news.cornell.edu/stories/2025/07/hiding-secret-codes-light-protects-against-fake-videos
43•CharlesW•2h ago•37 comments

Compressing Icelandic name declension patterns into a 3.27 kB trie

https://alexharri.com/blog/icelandic-name-declension-trie
164•alexharri•7h ago•65 comments

Online Collection of Keygen Music

https://keygenmusic.tk
61•mifydev•3d ago•5 comments

Great Question (YC W21) Is Hiring a VP of Engineering (Remote)

https://www.ycombinator.com/companies/great-question/jobs/ONBQUqe-vp-of-engineering
1•nedwin•1h ago

The /o in Ruby regex stands for "oh the humanity "

https://jpcamara.com/2025/08/02/the-o-in-ruby-regex.html
64•todsacerdoti•3h ago•19 comments

ThinkPad designer David Hill on unreleased models

https://www.theregister.com/2025/08/02/thinkpad_david_hill_interview/
100•LorenDB•5h ago•33 comments

Browser extension and local backend that automatically archives YouTube videos

https://github.com/andrewarrow/starchive
21•fcpguru•2h ago•5 comments

Financial lessons from my family's experience with long-term care insurance

https://www.whitecoatinvestor.com/financial-lessons-father-long-term-care-insurance/
50•wallflower•4h ago•54 comments

The Rubik's Cube Perfect Scramble (2024)

https://www.solutionslookingforproblems.com/post/the-rubik-s-cube-perfect-scramble
59•notagoodidea•4h ago•17 comments

Unikernel Guide: Build and Deploy Lightweight, Secure Apps

https://tallysolutions.com/technology/introduction-to-unikernel-2/
33•Bogdanp•3h ago•7 comments

Caches: LRU vs. Random

https://danluu.com/2choices-eviction/
75•gslin•2d ago•13 comments

How to reverse engineer an analog chip: the TDA7000 FM radio receiver

https://www.righto.com/2025/08/reverse-engineering-analog-TDA7000.html
14•nynyny7•2h ago•4 comments

Microsoft is open sourcing Windows 11's UI framework

https://www.neowin.net/news/microsoft-is-taking-steps-to-open-sourcing-windows-11-user-interface-framework/
137•bundie•10h ago•147 comments

Show HN: Wordle-style game for Fermi questions

https://www.fermiquestions.org/
6•danielfetz•1h ago•10 comments

Why Exercise Is a Miracle Drug

https://www.derekthompson.org/p/the-sunday-morning-post-why-exercise
189•zdw•2h ago•155 comments

Helion begins work on Washington nuclear fusion plant

https://www.nucnet.org/news/microsoft-backed-fusion-company-begins-work-on-washington-nuclear-fusion-plant-7-4-2025
4•mpweiher•2d ago•0 comments

Cerebras Code

https://www.cerebras.ai/blog/introducing-cerebras-code
421•d3vr•20h ago•165 comments

VSCode extension for syntax highlighting multi-line YAML strings

https://github.com/harrydowning/vscode-yaml-embedded-languages
16•moondev•3h ago•1 comments

The case for having roommates even when you can afford to live alone

https://supernuclear.substack.com/p/the-case-for-having-roommates-even
36•surprisetalk•4h ago•57 comments

Iceberg, the Right Idea – The Wrong Spec – Part 2 of 2: The Spec

https://www.database-doctor.com/posts/iceberg-is-wrong-2.html
9•lsuresh•4h ago•1 comments

'Communities' of extreme life seen for first time in deep ocean

https://www.bbc.com/news/articles/c3wnqe5j99do
11•moontoast•2d ago•2 comments

Robert Wilson has died

https://www.theartnewspaper.com/2025/08/01/robert-wilson-playwright-director-artist-obituary
69•paulpauper•15h ago•16 comments

Coffeematic PC – A coffee maker computer that pumps hot coffee to the CPU

https://www.dougmacdowell.com/coffeematic-pc.html
266•dougdude3339•20h ago•81 comments

Character Bitmap Graphics on the Pet 2001

https://www.masswerk.at/nowgobang/2025/character-bitmaps-on-the-pet2001
14•masswerk•4h ago•5 comments

Why leather is best motorcycle protection [video]

https://www.youtube.com/watch?v=xwuRUcAGIEU
171•lifeisstillgood•2d ago•143 comments
Open in hackernews

Show HN: WebGPU enables local LLM in the browser – demo site with AI chat

https://andreinwald.github.io/browser-llm/
83•andreinwald•4h ago
Browser LLM demo working on JavaScript and WebGPU. WebGPU is already supported in Chrome, Safari, Firefox, iOS (v26) and Android.

Demo, similar to ChatGPT https://andreinwald.github.io/browser-llm/

Code https://github.com/andreinwald/browser-llm

- No need to use your OPENAI_API_KEY - its local model that runs on your device

- No network requests to any API

- No need to install any program

- No need to download files on your device (model is cached in browser)

- Site will ask before downloading large files (llm model) to browser cache

- Hosted on Github Pages from this repo - secure, because you see what you are running

Comments

andsoitis•4h ago
very cool. improvement would be if the input text box is always on screen, rather than having to manually scroll down as the screen fills.
maxmcd•3h ago
Looks like this is a wrapper around: https://github.com/mlc-ai/web-llm

Which has a full web demo: https://chat.webllm.ai/

refulgentis•2h ago
I am to see it regardless - projects been very low activity for months. Just last night I was thinking about ripping it out before launch. No observable future.

EDIT: Ah, wrapper ~= using it as a dependency, no more

swores•2h ago
Is this correct?

It doesn't seem so to me, either from the way it works or from what little of the code I've looked at...

But I don't have time to do more than the quick glance I just did at a few of the files of each and need to run, so hopefully someone cleverer than me who won't need as much time as me to answer the question could confirm while I'm afk

refulgentis•1h ago
Entirely correct, c.f. LLM.ts. The bit to Show HN here, is about ~40 lines of code. Simple Typescript MVP of calling the library.

(source: maintains an LLM client that works across MLC/llama.cpp/3P providers; author of sibling comment that misunderstood initially)

pjmlp•3h ago
Beware of opening this on mobile Internet.
andreinwald•3h ago
Demo site is asking before download
lukan•2h ago
Well, I am on a mobile right now, can someone maybe share anything about the performance?
pjmlp•2h ago
Not everyone enjoys unlimited data plans, and 500 MB is schon a lot.
andreinwald•1h ago
On my Android device works pretty fast.

But keep in mind that it's small Llama-3.2-1B model, specially for less powerfull GPU.

asim•2h ago
What's the performance of a model like vs an OpenAI API? What's the comparable here? Edit: I see it's same models locally that you'd run using Ollama or something else. So basically just constrained by the size of the model, GPU and perf of the machine.
andreinwald•1h ago
Yes, its very similar to Ollama app, and Llama-3.2-1B model used
petermcneeley•2h ago
This demo only works if you have the webgpu feature "f16". You can find out if you have this by checking for the feature list in https://webgpureport.org/ . The page itself can of course check for this but since f16 support is common they probably just didnt bother.
andreinwald•2h ago
Also here https://caniuse.com/webgpu
cgdl•2h ago
Which model does the demo use?
andreinwald•1h ago
Llama 3.2 1B
scottfr•2h ago
There is a Prompt API in development that's available in both Chrome and Edge to give access to a local LLM. Chrome extensions have access to it and I believe websites can request access as part of an origin trial.

The model is fully managed by the browser. It's currently the Gemini Nano model on Chrome, and they are testing a version of the Gemma 3n model in beta channels. Edge uses phi-4-mini.

More information is available here: https://github.com/webmachinelearning/prompt-api

om8•1h ago
To have a gpu inference, you need a gpu. I have a demo that runs 8B llama on any computer with 4 gigs of ram

https://galqiwi.github.io/aqlm-rs/about.html

adastra22•1h ago
Any computer with a display has a GPU.
om8•1h ago
Sure, but integrated graphics usually lacks vram for LLM inference.
adastra22•10m ago
Which means that inference would be approximately the same speed (but compute offloaded) as the suggested CPU inference engine.
echoangle•1h ago
Fun demo but the model that's used seems to be pretty stupid:

> What's the best way to get to space?

>> Unfortunately, it's not currently possible for humans to travel to space in the same way that astronauts do. While there have been several manned missions to space, such as those to the International Space Station, the technology and resources required to make interstellar travel feasible are still in the early stages of development.

>> Currently, the fastest spacecraft is the Apollo missions, which took around 77 hours to reach the Moon. From there, they traveled to other planets, but the distances between them are so great that even at high speeds, it would take many years to reach them.

>> As for traveling to space itself, there are no commercial or private companies that offer space travel services, and the cost of such a journey is estimated to be in the tens of millions of dollars per person.

>> However, there are some concepts and technologies being explored that might one day make space travel more accessible and affordable. For example, private companies like SpaceX and Blue Origin are working on developing reusable rockets that could potentially take people to space more quickly and reliably.

>> But for now, let's just say that space travel is still in the future, and we'll have to wait a bit longer for humans to venture into the cosmos!

andreinwald•1h ago
It works on small Llama-3.2-1B model, specially for less powerfull GPU devices
tomrod•31m ago
It shows an interface of what could be.
andreinwald•1h ago
Model used: Llama 3.2 1B (small). Quality hould be similar with running Ollama app with same small model.
apitman•31m ago
Does anyone know why all of these WebGPU LLM demos have you download the models to browser storage rather than letting you open a gguf already on your local drive? I have several models downloaded already that I would be interested in trying.