frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

NanoChat – The best ChatGPT that $100 can buy

https://github.com/karpathy/nanochat
271•huseyinkeles•3h ago
https://x.com/karpathy/status/1977755427569111362

Comments

daft_pink•3h ago
Wow, how do we sign up for the Eurekalabs course and how much does it cost?
huseyinkeles•3h ago
Karpathy says nanochat will become the capstone project of the course LLM101n being developed by Eureka Labs.

I guess it’s still a work in progress? Couldn’t find any other information elsewhere.

Schiphol•2h ago
A bit more info [here](https://github.com/karpathy/LLM101n)
karpathy•1h ago
Still under development, remaining work includes tuning nanochat (current state being solid v0.1) and finalizing the in-between projects so that students can "unlock" all complexity that hides underneath: `torch.Tensor`, `torch.dist`, `.backward()`, '.compile()`, etc. And then the more ops heavy aspects.
BrokenCogs•24m ago
What's the pricing for the course/EurekaLabs? P.s. thanks for all you're doing
karimf•2h ago
I've always thought about the best way to contribute to humanity: number of people you help x how much you help them. I think what Karpathy is doing is one of the highest leverage ways to achieve that.

Our current world is build on top of open source projects. This is possible because there are a lot of free resources to learn to code so anyone from anywhere in the world can learn and make a great piece of software.

I just hope the same will happen with the AI/LLM wave.

viccis•18m ago
I recommend his ANN/LLM from scratch videos to people a lot because not only is he a clear instructor, but his code tends to be very Pythonic and just the right balance of terse but readable (not counting the Pytorch vectorization stuff, but that's not his fault, it's just complex). So I think people benefit just from watching and imitating his code style.
croes•17m ago
I‘m afraid the technology will do more damage because many people will abuse it for fake news and misinformation.
shafyy•13m ago
If it only were so easy
bkettle•5m ago
This free tradition in software is I think one of the things that I love so much, but I don't see how it can continue with LLMs due to the extremely high training costs and the powerful hardware required for inference. It just seems like software will necessarily require paying rent to the LLM hosts to keep up. I guess it's possible that we'll figure out a way to do local inference in a way that is accessible to everyone in the way that most other modern software tools are, but the high training costs make that seem unlikely to me.
flakiness•1h ago
Eureka Labs: https://github.com/EurekaLabsAI

What a prolific person Andrej is. It's been more than amazing to follow along!

TheAceOfHearts•1h ago
Here's the announcement post [0] from Karpathy, which provides a bit of additional context.

[0] https://x.com/karpathy/status/1977755427569111362

dang•40m ago
Thanks - we'll put that in the toptext as well
swyx•1h ago
> Thank you to chief LLM whisperer Alec Radford for advice/guidance.

oh man an Alec x Andrej podcast would BREAK THE INTERNET... just saying... going from glory days of GPT1 to now building GPT3? in 4 hours

codybontecou•1h ago
Please oh please. This would be perfect.
mhitza•1h ago
Should be "that you can train for $100"

Curios to try it someday on a set of specialized documents. Though as I understand the cost of running this is whatever GPU you can rent with 80GB of VRAM. Which kind of leaves hobbyists and students out. Unless some cloud is donating gpu compute capacity.

portaouflop•1h ago
If I have let’s say 40gb RAM does it not work at all or just take twice as long to train?
typpilol•54m ago
Won't work at all. Or if it does it'll be so slow since it'll have to go to the disk for every single calculation so it won't ever finish.
Onavo•1h ago
A GPU with 80GB VRAM costs around $1-3 USD an hour on commodity clouds (i.e. the non-Big 3 bare metal providers e.g. https://getdeploying.com/reference/cloud-gpu/nvidia-h100). I think it's accessible to most middle class users in first world countries.
antinomicus•7m ago
Isn’t the whole point to run your model locally?
Havoc•47m ago
>If your GPU(s) have less than 80GB, you'll have to tune some of the hyperparameters or you will OOM / run out of VRAM. Look for --device_batch_size in the scripts and reduce it until things fit. E.g. from 32 (default) to 16, 8, 4, 2, or even 1.

That sounds like it could run on a 24gb GPU. Batch size of 8 would imply 20gb mem, no?

...presumably just takes forever

faxmeyourcode•38m ago
This weekend I just cracked into nanoGPT (https://github.com/karpathy/nanoGPT), an older but fabulous learning exercise where you build and train a crappy shakespeare GPT with ~0.8M parameters on a cpu. Results are about what you'd expect from that, they suck, but you can start to feel the magic, especially if you're not a deep learning professional and you just want to poke around and hack on it.

I started writing up a blog post on my weekend with nanoGPT but it's not done yet... Would have been great to link to here lol oh well

andrewljohnson•32m ago
the shakespeare code tuned a little with different training data does a good job of generating Magic The Gathering commander decks
dmarcos•24m ago
I like the idea of specific-purpose toy models. How did you tune the code and what dataset you used?
ACCount37•24m ago
It's a useful exercise. A lot of the good ML work is first validated at small scale.

And this new example goes even further - adds instruction following and tool use SFT, as well as RLVR. Makes for a more useful baseline.

CountGeek•15m ago
So could I in practice train it on all my psychology books, materials, reports, case study and research papers and then run it on demand on a 1xH100 node - https://getdeploying.com/reference/cloud-gpu/nvidia-h100 whenever I have a specialised question?
zipy124•8m ago
You could but it would be significantly worse than fine-tuning or RAG with a pre-trained model, or using a smaller model since your dataset would be so small.
cyanydeez•5m ago
if the AI bubble is anything to be compared to, how is 100$ worth anything in GPT terms.

NanoChat – The best ChatGPT that $100 can buy

https://github.com/karpathy/nanochat
284•huseyinkeles•3h ago•39 comments

Show HN: SQLite Online – 11 years of solo development, 11K daily users

https://sqliteonline.com/
235•sqliteonline•6h ago•95 comments

Environment variables are a legacy mess: Let's dive deep into them

https://allvpv.org/haotic-journey-through-envvars/
144•signa11•2h ago•85 comments

From Millions to Billions

https://www.geocod.io/code-and-coordinates/2025-10-02-from-millions-to-billions/
27•mjwhansen•5d ago•1 comments

Spotlight on pdfly, the Swiss Army knife for PDF files

https://chezsoi.org/lucas/blog/spotlight-on-pdfly.html
264•Lucas-C•10h ago•81 comments

More random home lab things I've recently learned

https://chollinger.com/blog/2025/10/more-homelab-things-ive-recently-learned/
149•otter-in-a-suit•1w ago•69 comments

American solar farms

https://tech.marksblogg.com/american-solar-farms.html
151•marklit•8h ago•190 comments

JSON River – Parse JSON incrementally as it streams in

https://github.com/rictic/jsonriver
62•rickcarlino•5d ago•42 comments

The Sveriges Riksbank Prize in Economic Sciences in Memory of Alfred Nobel 2025

https://www.nobelprize.org/prizes/economic-sciences/2025/summary/
101•k2enemy•7h ago•120 comments

Optery (YC W22) – Hiring Tech Lead with Node.js Experience (U.S. & Latin America)

https://www.optery.com/careers/
1•beyondd•1h ago

CRDT and SQLite: Local-First Value Synchronization

https://marcobambini.substack.com/p/the-secret-life-of-a-local-first
9•marcobambini•4d ago•1 comments

Smartphones and being present

https://herman.bearblog.dev/being-present/
119•articsputnik•4h ago•81 comments

MPTCP for Linux

https://www.mptcp.dev/
84•SweetSoftPillow•9h ago•12 comments

AI and the Future of American Politics

https://www.schneier.com/blog/archives/2025/10/ai-and-the-future-of-american-politics.html
73•zdw•4h ago•27 comments

Control your Canon Camera wirelessly

https://github.com/JulianSchroden/cine_remote
70•nklswbr•6d ago•13 comments

Ofcom fines 4chan £20K and counting for violating UK's Online Safety Act

https://www.theregister.com/2025/10/13/4chan_ofcom_fine/
117•klez•4h ago•111 comments

Matrices can be your Friends

https://www.sjbaker.org/steve/omniv/matrices_can_be_your_friends.html
103•todsacerdoti•8h ago•74 comments

A16Z-backed data firms Fivetran, dbt Labs to merge in all-stock deal

https://www.reuters.com/business/a16z-backed-data-firms-fivetran-dbt-labs-merge-all-stock-deal-20...
87•mjirv•4h ago•29 comments

Two Paths to Memory Safety: CHERI and OMA

https://ednutting.com/2025/10/05/cheri-vs-oma.html
42•yvdriess•8h ago•29 comments

Clockss: Digital preservation services run by academic publishers and libraries

https://clockss.org/
45•robtherobber•5d ago•8 comments

LaTeXpOsEd: A Systematic Analysis of Information Leakage in Preprint Archives

https://arxiv.org/abs/2510.03761
62•oldfuture•10h ago•15 comments

Putting a dumb weather station on the internet

https://colincogle.name/blog/byo-weather-station/
130•todsacerdoti•5d ago•36 comments

Android's sideloading limits are its most anti-consumer move yet

https://www.makeuseof.com/androids-sideloading-limits-are-anti-consumer-move-yet/
231•josephcsible•3h ago•123 comments

Roger Dean – His legendary artwork in gaming history (Psygnosis)

https://spillhistorie.no/2025/10/03/legends-of-the-games-industry-roger-dean/
17•thelok•4h ago•2 comments

Jeep software update bricks vehicles, leaves owners stranded

https://www.thestack.technology/jeep-software-update-bricks-vehicles-leaves-owners-stranded/
49•croes•2h ago•2 comments

Ask HN: What are you working on? (October 2025)

309•david927•22h ago•864 comments

Tauri binding for Python through Pyo3

https://github.com/pytauri/pytauri
153•0x1997•5d ago•47 comments

Some graphene firms have reaped its potential but others are struggling

https://www.theguardian.com/business/2025/oct/13/lab-to-fab-are-promises-of-a-graphene-revolution...
61•robaato•10h ago•31 comments

Making regular GPS ultra-precise

https://norwegianscitechnews.com/2025/10/making-regular-gps-ultra-precise/
52•giuliomagnifico•6d ago•54 comments

MicroPythonOS – An Android-like OS for microcontrollers

https://micropythonos.com
159•alefnula•4d ago•57 comments