frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: I saw this cool navigation reveal, so I made a simple HTML+CSS version

https://github.com/Momciloo/fun-with-clip-path
2•momciloo•19m ago•0 comments

Show HN: Stacky – certain block game clone

https://www.susmel.com/stacky/
2•Keyframe•23m ago•0 comments

Show HN: A toy compiler I built in high school (runs in browser)

https://vire-lang.web.app
2•xeouz•45m ago•1 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
266•isitcontent•20h ago•33 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
365•vecti•22h ago•166 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
338•eljojo•23h ago•209 comments

Show HN: Kappal – CLI to Run Docker Compose YML on Kubernetes for Local Dev

https://github.com/sandys/kappal
17•sandGorgon•2d ago•5 comments

Show HN: Nginx-defender – realtime abuse blocking for Nginx

https://github.com/Anipaleja/nginx-defender
3•anipaleja•2h ago•0 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
80•phreda4•19h ago•15 comments

Show HN: Smooth CLI – Token-efficient browser for AI agents

https://docs.smooth.sh/cli/overview
94•antves•2d ago•70 comments

Show HN: MCP App to play backgammon with your LLM

https://github.com/sam-mfb/backgammon-mcp
3•sam256•4h ago•1 comments

Show HN: Slack CLI for Agents

https://github.com/stablyai/agent-slack
52•nwparker•1d ago•11 comments

Show HN: BioTradingArena – Benchmark for LLMs to predict biotech stock movements

https://www.biotradingarena.com/hn
27•dchu17•1d ago•12 comments

Show HN: Artifact Keeper – Open-Source Artifactory/Nexus Alternative in Rust

https://github.com/artifact-keeper
154•bsgeraci•1d ago•64 comments

Show HN: ARM64 Android Dev Kit

https://github.com/denuoweb/ARM64-ADK
18•denuoweb•2d ago•2 comments

Show HN: I'm 75, building an OSS Virtual Protest Protocol for digital activism

https://github.com/voice-of-japan/Virtual-Protest-Protocol/blob/main/README.md
7•sakanakana00•5h ago•1 comments

Show HN: I built Divvy to split restaurant bills from a photo

https://divvyai.app/
3•pieterdy•5h ago•1 comments

Show HN: Gigacode – Use OpenCode's UI with Claude Code/Codex/Amp

https://github.com/rivet-dev/sandbox-agent/tree/main/gigacode
19•NathanFlurry•1d ago•9 comments

Show HN: XAPIs.dev – Twitter API Alternative at 90% Lower Cost

https://xapis.dev
3•nmfccodes•2h ago•1 comments

Show HN: I Hacked My Family's Meal Planning with an App

https://mealjar.app
2•melvinzammit•7h ago•0 comments

Show HN: I built a free UCP checker – see if AI agents can find your store

https://ucphub.ai/ucp-store-check/
2•vladeta•8h ago•2 comments

Show HN: Micropolis/SimCity Clone in Emacs Lisp

https://github.com/vkazanov/elcity
173•vkazanov•2d ago•49 comments

Show HN: Daily-updated database of malicious browser extensions

https://github.com/toborrm9/malicious_extension_sentry
14•toborrm9•1d ago•8 comments

Show HN: Compile-Time Vibe Coding

https://github.com/Michael-JB/vibecode
10•michaelchicory•9h ago•3 comments

Show HN: Falcon's Eye (isometric NetHack) running in the browser via WebAssembly

https://rahuljaguste.github.io/Nethack_Falcons_Eye/
6•rahuljaguste•19h ago•1 comments

Show HN: Horizons – OSS agent execution engine

https://github.com/synth-laboratories/Horizons
24•JoshPurtell•1d ago•5 comments

Show HN: Slop News – HN front page now, but it's all slop

https://dosaygo-studio.github.io/hn-front-page-2035/slop-news
17•keepamovin•10h ago•6 comments

Show HN: Local task classifier and dispatcher on RTX 3080

https://github.com/resilientworkflowsentinel/resilient-workflow-sentinel
25•Shubham_Amb•1d ago•2 comments

Show HN: Fitspire – a simple 5-minute workout app for busy people (iOS)

https://apps.apple.com/us/app/fitspire-5-minute-workout/id6758784938
2•devavinoth12•13h ago•0 comments

Show HN: I built a RAG engine to search Singaporean laws

https://github.com/adityaprasad-sudo/Explore-Singapore
4•ambitious_potat•14h ago•4 comments
Open in hackernews

Show HN: Create-LLM – Train your own LLM in 60 seconds

https://github.com/theaniketgiri/create-llm
54•theaniketgiri•3mo ago
https://medium.com/@theaniketgiri/three-months-ago-i-wanted-...

Comments

kk58•3mo ago
Does this work on mac
theaniketgiri•3mo ago
Yep, works fine on Mac. Try the nano or tiny templates if you want quicker training runs
efilife•3mo ago
2 questions: how much of this project is AI generated and how much of only the readme is AI generated?
theaniketgiri•3mo ago
Mostly the repetitive stuff like README generation and pushing code with meaningful commit messages was handled by AI. The actual work and logic were done by me.
joshribakoff•3mo ago
What about the commit that added tens of thousands of lines of markdown claiming to be an AI summary?

Or the meaningful commit message of “.”

And the commit editing 1,000s of lines of python code mislabeled as a docs change?

theaniketgiri•3mo ago
Totally fair question!

Docs / Markdown: AI handled repetitive stuff like READMEs and summaries.

Core logic / Python: fully written by me.

Commit messages: some minimal ones just for quick iterations — the real work is in the code.

AI helped with boilerplate so I could ship faster; all functionality is hand-crafted.

joshribakoff•3mo ago
If the AI did the boilerplate that implies it was not fully written by you.

The “meaningful commit messages” — again are a single period as the message for a single commit for the entire python portion of the codebase.

My question was rhetorical. Whether the AI did it or a human did, it burns credibility to refer to things that don’t exist (like “meaningful commit messages”)

teruakohatu•3mo ago
Hacker News is a better place when we don’t attack people sharing their work. Your point was made.

Well done to the author for shipping code. I look forward to trying it out.

Grimblewald•3mo ago
> for sharing their work

If it was their work your point would hold.

theaniketgiri•3mo ago
To clarify the AI question once and for all:

What AI did: - Generated README templates (boilerplate markdown) - Suggested commit messages (I didn't always edit them) - Helped with documentation structure

What I wrote: - All Python training logic (train.py, trainer.py, callbacks) - All model architectures (gpt.py, tiny.py, small.py, etc.) - Tokenizer integration - Data pipeline - CLI scaffolding (

biinjo•3mo ago
Don’t feed the trolls. This was your idea and you made something that works. Who cares if its (partially) done by AI. Whomever is taking offense by people using AI for coding, is just having a hard time adapting to the current state of affairs.

It’s here, it’s happening. Try the project, if you like it thats great, if you don’t then move on.

And if you don’t intent to try it for whatever reason that’s fine as well but don’t be salty to the OP for sharing their passion project.

theaniketgiri•3mo ago
Thanks for the support! Appreciate you trying it out. Let me know if you hit any issues or have ideas for improvements.
Grimblewald•3mo ago
It isn't trolling if there is a genuine concern. It isn't about the fact it is all AI generated, I don't care about that personally. I do however care if someone lies about provenance.
Grimblewald•3mo ago
I find that hard to believe. You chose to put emojis into comments returned during script executions? You chose to store python scripts as strings in a TypeScript file rather than as python script files? You're even aware of the fact there are no python files in the project, rather it is strings in typescript files that get interpreted as python files, and you still refer to python files by filename in your comment like you expect it to exist in "your" codebase? You're competent enough to put together a project like this but then choose to use if-else for something solved better with match case? (LLM's do that since it's a recent addition to python and so LLM's avoid using match case, but humans rarely fail to use it)

Your medium article, your *.md and most of your code ALL looks LLM generated, which isn't as much a problem in my books, but lying about it is a huge problem.

theaniketgiri•3mo ago
Fair points and I get where you’re coming from. I’ve been very open that AI helped with repetitive parts (docs, boilerplate, commit messages). The functional code training logic, model architecture, CLI was written and tested by me. Some design choices (like storing scripts as strings or using if-else) were just pragmatic decisions made while iterating fast, not signs of AI authorship. Either way, the project is open source — you can inspect, critique, or even improve any part of it. I’m happy to take constructive feedback. My goal’s just to make LLM training more accessible.
Grimblewald•3mo ago
If you're iterating fast, then working on python files where you get the benefits of linting, syntax highlighting etc. and then reading the files as strings or bundled as required is MUCH faster than blindly mucking around in strings. As a side note, how fast do you type? Some of your multi-hundred line addition commits happen rather rapidly back to back.

The only time it would be faster to iterate with your scripts hard-coded into your TS files would be if an LLM is doing your iterating for you.

Why would anyone invest time and effort in a project where the author lies through their teeth about provenance? Why use your project that contributes, as it appears, nothing a LLM can't just give me? Why use this when I could just use an LLM to get the same directly in python without dicking around with npm?

x______________•3mo ago
You were called a troll in another reply but the comment you are replying to feels very AI. —'s and explaining to you what open you can do with open source?

Are you being trolled now?

Grimblewald•3mo ago
I miss when trolling just meant rage baiting, but yes almost certainly we're all being deceived by OP, if that carries an ulterior motive or stems from a simple lack of moral values is up for debtate but at this point it isnt worth the sustained energy.

This project being upvoted as much as it is reflects more on the community than it does OP at the end of the day.

010101010101•3mo ago
You’re absolutely right!
theaniketgiri•3mo ago
Thanks for the support!

And yeah, the commit history is messy - I was learning and shipping fast. Not perfect, but the tool works and people are using it.

Let me know if you have any questions when you try it!

darepublic•3mo ago
I don't quite understand how you get from this:

> I wanted to understand how these things work by building one myself.

Directly to this:

What if training an LLM was as easy as npx create-next-app?

I mean that the second thought seems to be the opposite of the first (what if the entirety of training llm was abstracted behind a simple command)

theaniketgiri•3mo ago
Great question - I should've been clearer.

When I started, I wanted to understand LLMs deeply. But I hit a wall: tutorials were either "hello world" toys or "here's 500 lines of setup before you start."

What I needed was: "give me working code quickly, THEN let me modify and learn."

That's what create-llm does. It scaffolds the boilerplate (like create-next-app), so you can spend time learning the interesting parts: - Why does vocab size matter? (adjust config, see results) - What causes overfitting? (train on small data, see it happen) - How do different architectures perform? (swap templates, compare)

It's "easy to start, deep to master." The abstraction gets you running in 60 seconds, then you dig into the code

seg_lol•3mo ago
The blogpost is some of the best LLM greentext I have seen for targeting the hn hivemind. Everything about this is :chefs kiss:
theaniketgiri•3mo ago
Thanks! The blog post is just my honest journey - spent way too much time trying to understand LLMs, figured others had the same frustration.

If you try create-llm, would love your feedback. Always looking to make it better.

3abiton•3mo ago
How does this differ from nanochat?
theaniketgiri•3mo ago
Good question! I think you mean nanoGPT (Karpathy's minimal GPT implementation)?

Key differences:

nanoGPT: - Minimal reference implementation (~300 lines) - Educational code for understanding transformers - Requires manual setup and configuration - Great for learning the internals

create-llm: - Production-ready scaffolding tool (like create-next-app) - One command: npx create-llm → complete project ready - Multiple templates (nano/tiny/small/base) - Built-in validation (warns about overfitting, vocab mismatches) - Includes tokenizer training, evaluation, deployment tools - Auto-detects issues before you waste GPU time

Think of it as: nanoGPT is the reference, create-llm is the framework.

nanoGPT teaches you HOW it works. create-llm lets you BUILD with what you learned.

You can actually use nanoGPT's architecture in create-llm templates - they're complementary tools!

Grimblewald•3mo ago
Unlike nanochat this is purely vibe-coded, improving vibes by 110%, with 112x more emoji. A key innovation that gets to the heart of the problem is that this project stores python files as strings in typscript files to help improve workflows. I imagine the author solved this engineering challenge to overcome existing limitations\emdash more efficient, interpretable, and maintainable code\emdash in existing projects.
theaniketgiri•3mo ago
The Python-in-TS bit made me smile But to clarify, it’s a standard TypeScript CLI — no such hacks involved, just template-based generation.
Grimblewald•3mo ago
Ok, but there is no reason to bake it into the TS scripts. You could write the python scripts and package them using standard tools. In my experience only an LLM would do that, since it makes sense to generate the code and templates to insert in one go. However, if a human were to do it, the python scripts would be their own files and they would be bundled / read in as strings when/as required. A gigantic lump of text in a string makes no sense in human paradigms, even if it makes perfect sense for an LLM to do it. For humans it is incredibly hostile to update and maintain.

As a side note, without looking it up, on your device, what is the process for typing an emdash?

theaniketgiri•3mo ago
Fair point I agree embedding code as strings isn’t ideal. I did it mainly to make npx create-llm portable without needing a Python setup during scaffolding. Definitely open to improving that happy to refactor if you have suggestions.
Grimblewald•3mo ago
That makes zero sense. Am I even speaking with a natural person right now?!? Your comments sound like llm bullshit and everything about this project reeks of it as well, from code to readme.
freakynit•3mo ago
Bro is angry about using llm to write code instead of being happy about a working code that makes it extremely easy for anyone to build their own nano gpt's.

LLM's are just the next evolution that assist you with coding tasks... similar to w3schools => blogs => stackoverflow, and now => llm's.

There's absolutely nothing wrong with using them. The problem is people who use them without reviewing their outputs.

theaniketgiri•3mo ago
Yeah, that’s how I see it too. LLMs are part of the toolkit now what matters is how you use them to actually ship something useful.
Grimblewald•3mo ago
True, however lying and lazy work are not part of the normal toolkit, even if they are in yours.
Grimblewald•3mo ago
Right, you hit the nail on the head.

> problem is people who use without reviewing outputs

the problem here is there is no reason to beleive OP reviewed or even fully understands the code. There is plenty of evidence of OP lying.

All in all best case scenario is we have a codebase made by someone dishonest and incompetant, and at worst it is plausible there is an intentional effort to obfuscated the codebase in an effort to slip people malware.

the entire thing stinks, and cannot be trusted. It deserves (oweing the dishonesty) a rather strong negative response and nothing more.

mkrishnan•3mo ago
It makes sense to me. its great to see tools like create-llm which help to kickstart llm development without spending lots of time putting low level stuffs
Grimblewald•3mo ago
Tools like that already exist. Several of them, and all of them made with a far greater level of care and quality. This project is not maintainable in it's current state and takes more work to get to a maintainable point than already actively maintained projects.
endofreach•3mo ago
While you are not wrong about this not being the wisest choice, i have seen it done before countless times already long before llms arrived. So i don‘t think it‘s such a clear sign of intense llm usage as you make it out to be.
3abiton•3mo ago
> Unlike nanochat this is purely vibe-coded, improving vibes by 110%

Karpathy clearly said that it wasn't vibe coded. Apparently it was more time consuming to fix gpt bugs than to do it by himself.

theaniketgiri•3mo ago
Thanks everyone for the feedback and discussion. For those asking technical questions - happy to help! The tool works on Mac/Linux/Windows, check the README for setup. For those concerned about the architecture - it follows standard scaffolding patterns (create-next-app, etc). TypeScript CLI generates Python projects. 82+ stars in 24 hours - grateful for everyone trying it out. Keep the feedback coming!
potamic•3mo ago
How did you test this? Did you train something?
theaniketgiri•3mo ago
Yeah, I did! I trained a few small ones — mostly the “nano” and “tiny” templates (a few million params) on datasets like Shakespeare and Alpaca. The goal was to make sure the training loop, tokenizer, and evaluation all worked smoothly.

Didn’t go for massive models — more about making the whole setup process quick and reliable. You can actually train the nano one on CPU in a few minutes just to see it working.

mkrishnan•3mo ago
This is great initiative. don't let anyone here discourage from doing something great like this.
theaniketgiri•3mo ago
Thanks a lot, really appreciate that Building something new always gets mixed reactions, but messages like yours keep me going.