frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Supabase MCP can leak your entire SQL database

https://www.generalanalysis.com/blog/supabase-mcp-blog
447•rexpository•6h ago•211 comments

Breaking Git with a carriage return and cloning RCE

https://dgl.cx/2025/07/git-clone-submodule-cve-2025-48384
246•dgl•6h ago•86 comments

Bootstrapping a side project into a profitable seven-figure business

https://projectionlab.com/blog/we-reached-1m-arr-with-zero-funding
167•jonkuipers•1d ago•35 comments

Smollm3: Smol, multilingual, long-context reasoner LLM

https://huggingface.co/blog/smollm3
201•kashifr•7h ago•36 comments

Radium Music Editor

http://users.notam02.no/~kjetism/radium/
133•ofalkaed•6h ago•26 comments

Dynamical origin of Theia, the last giant impactor on Earth

https://arxiv.org/abs/2507.01826
60•bikenaga•5h ago•17 comments

Brut: A New Web Framework for Ruby

https://naildrivin5.com/blog/2025/07/08/brut-a-new-web-framework-for-ruby.html
107•onnnon•5h ago•42 comments

Plants monitor the integrity of their barrier by sensing gas diffusion

https://www.nature.com/articles/s41586-025-09223-4
46•Bluestein•3d ago•9 comments

Taking over 60k spyware user accounts with SQL injection

https://ericdaigle.ca/posts/taking-over-60k-spyware-user-accounts/
143•mtlynch•5d ago•42 comments

Show HN: OffChess – Offline chess puzzles app

https://offchess.com
288•avadhesh18•14h ago•111 comments

Xenharmlib: A music theory library that supports non-western harmonic systems

https://xenharmlib.readthedocs.io/en/latest/
4•retooth•1h ago•0 comments

Can an email go 500 miles in 2025?

https://flak.tedunangst.com/post/can-an-email-go-500-miles-in-2025
253•zdw•4d ago•98 comments

Ask HN: What are some cool or underrated tech companies based in Canada?

61•pedrodelfino•2h ago•30 comments

GlobalFoundries to Acquire MIPS

https://mips.com/press-releases/gf-mips/
146•mshockwave•6h ago•97 comments

The Tradeoffs of SSMs and Transformers

https://goombalab.github.io/blog/2025/tradeoffs/
35•jxmorris12•4h ago•5 comments

Show HN: A rain Pomodoro with brown noise, ASMR, and Middle Eastern music

https://forgetoolz.com/rain-pomodoro
39•ShadowUnknown•6h ago•25 comments

Ceramic: A cross-platform and open-source 2D framework in Haxe

https://ceramic-engine.com/
43•-yukari•3d ago•3 comments

Show HN: Jukebox – Free, Open Source Group Playlist with Fair Queueing

https://www.jukeboxhq.com/
88•skeptrune•8h ago•33 comments

Inertial forces (indirect terms) in problems with a central body

https://astro.theoj.org/article/141682-on-inertial-forces-indirect-terms-in-problems-with-a-central-body
9•raattgift•3d ago•0 comments

On The Meaning of Ritual

https://alicemaz.substack.com/p/on-the-meaning-of-ritual
56•jger15•3d ago•50 comments

Particle Lenia Deluxe Edition

https://www.craftlinks.art/Notebook/particle-lenia/
25•CraftingLinks•3d ago•4 comments

Show HN: OpenAPI mocks that don't suck – realistic test data, quick setup

https://beeceptor.com/openapi-mock-server/
9•ankit84•4h ago•0 comments

New sphere-packing record stems from an unexpected source

https://www.quantamagazine.org/new-sphere-packing-record-stems-from-an-unexpected-source-20250707/
405•pseudolus•1d ago•204 comments

SVGs that feel like GIFs

https://koaning.io/posts/svg-gifs/
363•cantdutchthis•15h ago•97 comments

Blind to Disruption – The CEOs Who Missed the Future

https://steveblank.com/2025/07/08/blind-to-disruption-the-ceos-who-missed-the-future/
61•ArmageddonIt•10h ago•69 comments

Mercury: Ultra-fast language models based on diffusion

https://arxiv.org/abs/2506.17298
553•PaulHoule•1d ago•229 comments

Attimet (YC F24) – Quant Trading Research Lab – Is Hiring Founding Researcher

https://www.ycombinator.com/companies/attimet/jobs/6LaQIc5-founding-researcher-quant
1•kbanothu•11h ago

The Day You Became a Better Writer (2007)

https://dilbertblog.typepad.com/the_dilbert_blog/2007/06/the_day_you_bec.html
12•santiviquez•1d ago•0 comments

I used o3 to profile myself from my saved Pocket links

https://noperator.dev/posts/o3-pocket-profile/
496•noperator•1d ago•190 comments

Berry Script: lightweight embedded scripting language for microcontrollers

https://berry-lang.github.io/
91•hasheddan•3d ago•42 comments
Open in hackernews

Smollm3: Smol, multilingual, long-context reasoner LLM

https://huggingface.co/blog/smollm3
201•kashifr•7h ago

Comments

gardnr•6h ago
It's small (3B) and does great on benchmarks. This is a model for edge / mobile deployments so the gains over gemma3-4b are meaningful. It has dual mode reasoning / non_reasoning AND they released the full training method:

> We're releasing SmolLM3 with our engineering blueprint. It includes architecture details, exact data mixtures showing how we progressively boost performance across domains in a three-stage pretraining approach, and the methodology for building a hybrid reasoning model. Usually, achieving these results would require months of reverse engineering. Instead, we're providing the full methodology.

tiahura•6h ago
Can anyone estimate how much of the 3B is necessitated by multi-language support?
rockinghigh•6h ago
The vocabulary size is fairly small (128,256) for a multilingual model. I would guess it doesn't require many additional parameters to support these 5 languages as many tokens can be shared.
nateb2022•6h ago
https://web.archive.org/web/20250708164705/https://huggingfa...
_1•6h ago
Which small model is good for fine tuning to various enterprise data sets? Our business units are wanting to run small models in browser and on mobile devices, without dealing with RAG and cloud resources.
mhitza•6h ago
You really need to try them all out yourself and make sure you have proper benchmarks.

While machine learning is not my field, I've tried to finetune Mistral 7B (following their official guide and toolset) and the results did not satisfy. Had a few very specific questions from the dataset that no matter how much I've finetuned and tweaked the process it was not able to respond with correct information.

A mix of vector search + keyword search is still better at building the right question context than expecting it to learn all the information.

I've used the pretrained dataset approach. Maybe building syntethic questions and answers around the dataset yields better results but I didn't have time to experiment with that approach.

ivape•3h ago
How much data did you use to fine tune?
mhitza•3h ago
Kilobytes to megabytes of data. I was trying to fine-tune it for some specific legislation I was expecting to be able afterwards to ask about.
gardnr•6h ago
Small models are bad at knowing things. Trying to train knowledge in to small models is probably not the way you want to go. You could try building an offline embedded RAG system that is deployable as wasm. Some folks have been experiencing success with this.
_1•5h ago
We do use WebLLM and a hosted Weaviate database, but there are complaints about speed (both retrieval and time to first token as the context will get big). The Gemma 3n "nesting doll" approach sounds like it could be useful .. but haven't found anyone specifically doing it to add domain specific knowledge.
janalsncm•4h ago
Typically retrieval is the fast part in my experience. Have you considered cheaper retrieval methods? Bm25 does pretty well on its own. And you can augment your dataset by precomputing relevant queries for each doc.
simonw•5h ago
What are you hoping to achieve by fine-tuning a model in this way?
netdur•4h ago
I have fine-tuned Gemma 3N 2B and it's pretty good, but loads slow on my S23U, once it's loaded though, it works fine

Also tried SmolVLM 256M and 500M, they load faster and you can embed them in assets, they work if you know what you're doing

Just keep in mind that smaller models don't perform as well due to their limited parameters

Also on Android, since you can't ship files larger than 2GB due to Java compression issues, you need to download models separately, then you can't load the model from the download folder, you have to copy it into the app's own folder, this means a Gemma 3N 2B model that's 3.14 GB would need at least 7 GB of free space on the user's phone

WhitneyLand•6h ago
Mostly SOTA performance at the 3B level. A notable addition to the small but truly open club of models that provide full disclosure, code, recipes to reproduce their work.

Looks like ballpark a million dollars of GPU time if you want to train up one for yourself (4000 gpus/24 days).

Very nice write up that’s generous in sharing their learnings.

This is a solid and positive contribution.

YetAnotherNick•5h ago
It's 384 H100s for 24 days, costing less than half a million dollars.
Imustaskforhelp•4h ago
Pardon me, but is the dataset public.

Like if I really really just wanted to build it from scratch, could I do so? (not that I have that money but just curious)

hynky•4h ago
yes, both core web datasets are publicly available as well as the rest
Imustaskforhelp•4h ago
Thanks!

To be honest, if I might argue then that this is one of the best truly open source models that we have got.

There is AllenAI and (Elmo?) and there is also this one which does distributed training but I think this looks a lot like SOTA for 3B parameters to me.

Thanks for telling me, I am not going to lie, I am going to try to test it now! (Ima try some GGUF since ollama convenience)

peatmoss•1h ago
OLMo: https://allenai.org/olmo

AFAIK, they were the first open everything model.

refulgentis•2h ago
I spent about 10 minutes this AM cross-checking with Phi-4-mini benchmarks, as it was very odd to not include the leader in benchmarks and it seemed universally behind.

For context, I dev an LLM client, a core tenant is keeping local as close to cloud parity as much as is possible. (via llama.cpp)

Companies aren't taking local AI seriously on a sustained basis outside Microsoft.

Overall, I usually would bite my tongue. HF is a great citizen, and I doubt this'll be a one off. However, when I see superlatives affirmed, while leaving out the local SoTA for many many moons that is a godsend in this sector, I think it is good to, rather than shy away, stand up and say this.

adrianlzt•2h ago
From the blog post: "SmolLM3 supports tool calling, and its chat template incorporates two distinct sections for tool descriptions: XML Tools and Python Tools"
bitwize•6h ago
There's a British comedy skit lurking in here.

"So it's a small large language model?"

"Oh yes, very small."

"How can it be small and large at the same time?"

"Well, it's small by the standards of a large language model."

"So it's large."

"Oh yes, very large."

"Large compared to what?"

"Small language models."

"And so something like ChatGPT, what would that be exactly? A large large language model?"

"Yes, precisely. An LLLM."

netdur•4h ago
it's big little planet or small big planet?
janalsncm•4h ago
Standards have shifted as well. Gpt2 used to be considered “large” but it is half the size of this. Oh and also Sam Altman said it was too dangerous to release. At this point I consider anything too big to run on consumer grade hardware to be large, but an exact definition is a little silly to argue about.
a_wild_dandan•2h ago
Altman released GPT-2 despite expressing that doing so was a bad idea? That's wild.
Alifatisk•1h ago
I think Altman meant it's too dangerous to open-source GPT-2, therefore locked it in behind a service.
papichulo2023•3h ago
Do not mess with the Miniature giant space hamsters
_kb•30m ago
Australian. This is straight up Clarke and Dawe / Utopia.
msgodel•5h ago
Wow. Close to a Qwen3 distill with 75% the size. That's great!

I've been using the smollm base models for my own finetunes just because they're so high quality, it looks like I might be using them to drive local agents/code completion in the near future too.

Their RL algorithm looks interesting. I'm still using OpenAI's algorithm for my stuff, I've been meaning to check on the SoTA since I know my code is pretty outdated (It's crazy how fast that happens with this stuff.)

gdiamos•5h ago
Nice work anton et al.

I hope you continue the 50-100M parameter models.

I think there is a case for models that finish fast on CPUs in solve by llm test cases.

eachro•5h ago
From what I've heard, the llama3 models are fairly easy to fine-tune (please correct me if I'm wrong or if there are more amenable models here). How easy is it to finetune smollm3? I know a lot of the MoE LLMs have been quite fickle in this regard.
BarakWidawsky•5h ago
It’s interesting that it looks like they didn’t apply their own RL to the model, and instead fine tuned on reasoning traces from large datasets and generating reasoning traces from larger models
lewtun•5h ago
Indeed we opted for offline methods like Anchored Preference Optimization as we found in the Open R1 project that doing multi-task RL on small models is quite a hassle to get right. With offline methods, you focus much more on dataset curation / generation, but that still provides faster iteration cycles for the model scale we’re dealing with!
ivape•3h ago
Looks like it's the 3B models that are being shipped out to on device by default. Apple's on-device LLM is 3B, and I believe Canary is shipping Google nano:

https://developer.chrome.com/docs/ai/rewriter-api

ivape•3h ago
I wonder if this will be cheaper than llama 3.1 8b on OpenRouter.
danielhanchen•1h ago
I fixed some chat template issues for llama.cpp and other inference engines! To run it, do:

./llama.cpp/llama-cli -hf unsloth/SmolLM3-3B-GGUF:Q4_K_XL --jinja -ngl 99