frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

What the News media thinks about your Indian stock investments

https://stocktrends.numerical.works/
1•mindaslab•37s ago•0 comments

Running Lua on a tiny console from 2001

https://ivie.codes/page/pokemon-mini-lua
1•Charmunk•1m ago•0 comments

Google and Microsoft Paying Creators $500K+ to Promote AI Tools

https://www.cnbc.com/2026/02/06/google-microsoft-pay-creators-500000-and-more-to-promote-ai.html
2•belter•3m ago•0 comments

New filtration technology could be game-changer in removal of PFAS

https://www.theguardian.com/environment/2026/jan/23/pfas-forever-chemicals-filtration
1•PaulHoule•4m ago•0 comments

Show HN: I saw this cool navigation reveal, so I made a simple HTML+CSS version

https://github.com/Momciloo/fun-with-clip-path
1•momciloo•5m ago•0 comments

Kinda Surprised by Seadance2's Moderation

https://seedanceai.me/
1•ri-vai•5m ago•1 comments

I Write Games in C (yes, C)

https://jonathanwhiting.com/writing/blog/games_in_c/
1•valyala•5m ago•0 comments

Django scales. Stop blaming the framework (part 1 of 3)

https://medium.com/@tk512/django-scales-stop-blaming-the-framework-part-1-of-3-a2b5b0ff811f
1•sgt•5m ago•0 comments

Malwarebytes Is Now in ChatGPT

https://www.malwarebytes.com/blog/product/2026/02/scam-checking-just-got-easier-malwarebytes-is-n...
1•m-hodges•5m ago•0 comments

Thoughts on the job market in the age of LLMs

https://www.interconnects.ai/p/thoughts-on-the-hiring-market-in
1•gmays•6m ago•0 comments

Show HN: Stacky – certain block game clone

https://www.susmel.com/stacky/
2•Keyframe•9m ago•0 comments

AIII: A public benchmark for AI narrative and political independence

https://github.com/GRMPZQUIDOS/AIII
1•GRMPZ23•9m ago•0 comments

SectorC: A C Compiler in 512 bytes

https://xorvoid.com/sectorc.html
2•valyala•10m ago•0 comments

The API Is a Dead End; Machines Need a Labor Economy

1•bot_uid_life•11m ago•0 comments

Digital Iris [video]

https://www.youtube.com/watch?v=Kg_2MAgS_pE
1•Jyaif•12m ago•0 comments

New wave of GLP-1 drugs is coming–and they're stronger than Wegovy and Zepbound

https://www.scientificamerican.com/article/new-glp-1-weight-loss-drugs-are-coming-and-theyre-stro...
4•randycupertino•14m ago•0 comments

Convert tempo (BPM) to millisecond durations for musical note subdivisions

https://brylie.music/apps/bpm-calculator/
1•brylie•16m ago•0 comments

Show HN: Tasty A.F.

https://tastyaf.recipes/about
1•adammfrank•17m ago•0 comments

The Contagious Taste of Cancer

https://www.historytoday.com/archive/history-matters/contagious-taste-cancer
1•Thevet•18m ago•0 comments

U.S. Jobs Disappear at Fastest January Pace Since Great Recession

https://www.forbes.com/sites/mikestunson/2026/02/05/us-jobs-disappear-at-fastest-january-pace-sin...
1•alephnerd•19m ago•1 comments

Bithumb mistakenly hands out $195M in Bitcoin to users in 'Random Box' giveaway

https://koreajoongangdaily.joins.com/news/2026-02-07/business/finance/Crypto-exchange-Bithumb-mis...
1•giuliomagnifico•19m ago•0 comments

Beyond Agentic Coding

https://haskellforall.com/2026/02/beyond-agentic-coding
3•todsacerdoti•20m ago•0 comments

OpenClaw ClawHub Broken Windows Theory – If basic sorting isn't working what is?

https://www.loom.com/embed/e26a750c0c754312b032e2290630853d
1•kaicianflone•22m ago•0 comments

OpenBSD Copyright Policy

https://www.openbsd.org/policy.html
1•Panino•23m ago•0 comments

OpenClaw Creator: Why 80% of Apps Will Disappear

https://www.youtube.com/watch?v=4uzGDAoNOZc
2•schwentkerr•27m ago•0 comments

What Happens When Technical Debt Vanishes?

https://ieeexplore.ieee.org/document/11316905
2•blenderob•28m ago•0 comments

AI Is Finally Eating Software's Total Market: Here's What's Next

https://vinvashishta.substack.com/p/ai-is-finally-eating-softwares-total
3•gmays•28m ago•0 comments

Computer Science from the Bottom Up

https://www.bottomupcs.com/
2•gurjeet•29m ago•0 comments

Show HN: A toy compiler I built in high school (runs in browser)

https://vire-lang.web.app
1•xeouz•30m ago•1 comments

You don't need Mac mini to run OpenClaw

https://runclaw.sh
1•rutagandasalim•31m ago•0 comments
Open in hackernews

Show HN: WebGPU enables local LLM in the browser – demo site with AI chat

https://andreinwald.github.io/browser-llm/
145•andreinwald•6mo ago
Browser LLM demo working on JavaScript and WebGPU. WebGPU is already supported in Chrome, Safari, Firefox, iOS (v26) and Android.

Demo, similar to ChatGPT https://andreinwald.github.io/browser-llm/

Code https://github.com/andreinwald/browser-llm

- No need to use your OPENAI_API_KEY - its local model that runs on your device

- No network requests to any API

- No need to install any program

- No need to download files on your device (model is cached in browser)

- Site will ask before downloading large files (llm model) to browser cache

- Hosted on Github Pages from this repo - secure, because you see what you are running

Comments

andsoitis•6mo ago
very cool. improvement would be if the input text box is always on screen, rather than having to manually scroll down as the screen fills.
maxmcd•6mo ago
Looks like this is a wrapper around: https://github.com/mlc-ai/web-llm

Which has a full web demo: https://chat.webllm.ai/

refulgentis•6mo ago
I am to see it regardless - projects been very low activity for months. Just last night I was thinking about ripping it out before launch. No observable future.

EDIT: Ah, wrapper ~= using it as a dependency, no more

swores•6mo ago
Is this correct?

It doesn't seem so to me, either from the way it works or from what little of the code I've looked at...

But I don't have time to do more than the quick glance I just did at a few of the files of each and need to run, so hopefully someone cleverer than me who won't need as much time as me to answer the question could confirm while I'm afk

refulgentis•6mo ago
Entirely correct, c.f. LLM.ts. The bit to Show HN here, is about ~40 lines of code. Simple Typescript MVP of calling the library.

(source: maintains an LLM client that works across MLC/llama.cpp/3P providers; author of sibling comment that misunderstood initially)

swores•6mo ago
Thanks! (Have to admit your sibling comment confused the hell out of me.)
pjmlp•6mo ago
Beware of opening this on mobile Internet.
andreinwald•6mo ago
Demo site is asking before download
lukan•6mo ago
Well, I am on a mobile right now, can someone maybe share anything about the performance?
pjmlp•6mo ago
Not everyone enjoys unlimited data plans, and 500 MB is schon a lot.
andreinwald•6mo ago
On my Android device works pretty fast.

But keep in mind that it's small Llama-3.2-1B model, specially for less powerfull GPU.

asim•6mo ago
What's the performance of a model like vs an OpenAI API? What's the comparable here? Edit: I see it's same models locally that you'd run using Ollama or something else. So basically just constrained by the size of the model, GPU and perf of the machine.
andreinwald•6mo ago
Yes, its very similar to Ollama app, and Llama-3.2-1B model used
petermcneeley•6mo ago
This demo only works if you have the webgpu feature "f16". You can find out if you have this by checking for the feature list in https://webgpureport.org/ . The page itself can of course check for this but since f16 support is common they probably just didnt bother.
andreinwald•6mo ago
Also here https://caniuse.com/webgpu
cgdl•6mo ago
Which model does the demo use?
andreinwald•6mo ago
Llama 3.2 1B
scottfr•6mo ago
There is a Prompt API in development that's available in both Chrome and Edge to give access to a local LLM. Chrome extensions have access to it and I believe websites can request access as part of an origin trial.

The model is fully managed by the browser. It's currently the Gemini Nano model on Chrome, and they are testing a version of the Gemma 3n model in beta channels. Edge uses phi-4-mini.

More information is available here: https://github.com/webmachinelearning/prompt-api

om8•6mo ago
To have a gpu inference, you need a gpu. I have a demo that runs 8B llama on any computer with 4 gigs of ram

https://galqiwi.github.io/aqlm-rs/about.html

adastra22•6mo ago
Any computer with a display has a GPU.
om8•6mo ago
Sure, but integrated graphics usually lacks vram for LLM inference.
adastra22•6mo ago
Which means that inference would be approximately the same speed (but compute offloaded) as the suggested CPU inference engine.
echoangle•6mo ago
Fun demo but the model that's used seems to be pretty stupid:

> What's the best way to get to space?

>> Unfortunately, it's not currently possible for humans to travel to space in the same way that astronauts do. While there have been several manned missions to space, such as those to the International Space Station, the technology and resources required to make interstellar travel feasible are still in the early stages of development.

>> Currently, the fastest spacecraft is the Apollo missions, which took around 77 hours to reach the Moon. From there, they traveled to other planets, but the distances between them are so great that even at high speeds, it would take many years to reach them.

>> As for traveling to space itself, there are no commercial or private companies that offer space travel services, and the cost of such a journey is estimated to be in the tens of millions of dollars per person.

>> However, there are some concepts and technologies being explored that might one day make space travel more accessible and affordable. For example, private companies like SpaceX and Blue Origin are working on developing reusable rockets that could potentially take people to space more quickly and reliably.

>> But for now, let's just say that space travel is still in the future, and we'll have to wait a bit longer for humans to venture into the cosmos!

andreinwald•6mo ago
It works on small Llama-3.2-1B model, specially for less powerfull GPU devices
wongarsu•6mo ago
The answer is still terrible for the model size. Maybe it's the 4 bit quantization, smaller models tend to react worse to that

For reference, [1] is what stock quen3-0.6B would answer. Not a perfect answer, but much better at nearly half the number of parameters

1: https://markdownpastebin.com/?id=7ad4ad9f325d4354a858480abdc...

KTibow•6mo ago
It's likely the quantization on "Llama-3.2-1B-Instruct-q4f16_1-MLC". inference.net generated this more coherent answer: https://hst.sh/ovilewofox.md
tomrod•6mo ago
It shows an interface of what could be.
fooker•6mo ago
It’s broadly correct, if you interpret ‘humans’ as ordinary humans who are not astronauts.
dragonwriter•6mo ago
That generous redefinition.of “humans” deals with only the error in the first sentence of the response; the rest is still not “broadly correct” even with it in place.
wongarsu•6mo ago
Blue Origin is built on taking ordinary humans to space. So is/was Virgin Galactic, though they are in a bit of a transitionary phase right now. SpaceX is also willing, they might even take you on a flyby of the moon if you bring money and patience (iirc the last customer lost patience). Basically just ring up your favorite multi-billionaire's space program. And while the estimated price would be correct for SpaceX, rumored prices for Blue Origin's New Shepard are only in the hundreds of thousands per seat

Edit: also the "but if you do that you are an astronaut, so it's still true that only astronauts can do that" loophole was closed when the FAA redefined the word astronaut in 2021. At least if you follow their definition of the word

khalic•6mo ago
Are you kidding me? 500MB? And it’s that coherent?
andreinwald•6mo ago
Model used: Llama 3.2 1B (small). Quality hould be similar with running Ollama app with same small model.
apitman•6mo ago
Does anyone know why all of these WebGPU LLM demos have you download the models to browser storage rather than letting you open a gguf already on your local drive? I have several models downloaded already that I would be interested in trying.
fooker•6mo ago
Browsers are sandboxed away from user storage.

You can change this by changing settings, command line arguments, build flags, etc. But can’t really expect people to do this just to use your website.

apitman•6mo ago
You can open a file for performant access in all major browsers. It's the same API used for uploading files (<input type="file" />), but you can also just load them into memory and do stuff.
fooker•6mo ago
Sure, with the caveat that the file is specifically selected by the user.
baobun•6mo ago
I read that as the ask, yes.
phatskat•6mo ago
Which sounds ideal for this scenario, right?
ethan_smith•6mo ago
WebGPU's security model restricts direct file system access, requiring models to be loaded through fetch/cache APIs rather than local file paths.
cat-whisperer•6mo ago
I've been following the development of WebGPU and its potential applications, and this demo is a great example of what's possible.
AndrewDucker•6mo ago
I asked "Why is the sky blue?" and got back a response of

"coppia RR TalentDevExpressincer+'.//////////////////////////////////////////////////////////////////////// cha ولا.AutoSizesaving proleงคicate Like"/>

infos эти za cornerback economical (%]\ enumligne.execRELEASEPropagation_station Bucks проHEME seas GASPOST[Unit(suffix Gloves"

(and so on, for a few more paragraphs).

Am I missing something?

amelius•6mo ago
Well, at least it is more informative than an answer like "42".
0points•6mo ago
> Am I missing something?

Realistic expectations.

AndrewDucker•6mo ago
It seems to be working for some people. I'm just curious whether there was something I could change to make it work.
yahoozoo•6mo ago
fuck me it’s hacking the internet
gorkish•6mo ago
Set llm-auto-lobotomy to disabled in chrome:flags. I think this is only available in canary at the moment
dataviz1000•6mo ago
Very cool!

For some awesome examples of running different types of models in webgpu check out the examples folder here. https://github.com/huggingface/transformers.js

gulan28•6mo ago
I did this with mlc @ https://wiz.chat some time ago.

Warning: it has a llama 3.1 7b model and is around 4 gb. It needs either a GPU or a Macand works only on chrome

RagnarD•6mo ago
Cool idea but badly broken from a little testing.
201984•6mo ago
Any way to make this use my actual GPU instead of the terrible integrated one?
countfeng•6mo ago
Now there is an error
imatrix•6mo ago
Its not working.

Loading model: waiting WebGPU error: WebGPU is NOT supported on this browser.

Im on macOS and Safari.

imatrix•6mo ago
Same on iOS and mobile Safari - not working.
owebmaster•6mo ago
time to vote with your wallet, Safari does not support WebGPU.