frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Mistral 3 family of models released

https://mistral.ai/news/mistral-3
114•pember•55m ago

Comments

timpera•46m ago
Extremely cool! I just wish they would also include comparisons to SOTA models from OpenAI, Google, and Anthropic in the press release, so it's easier to know how it fares in the grand scheme of things.
constantcrying•28m ago
The lack of the comparison (which absolutely was done), tells you exactly what you need to know.
crimsoneer•20m ago
If someone is using these models, they probably can't or won't use the existing SOTA models, so not sure how useful those comparisons actually are. "Here is a benchmark that makes us look bad from a model you can't use on a task you won't be undertaking" isn't actually helpful (and definitely not in a press release).
constantcrying•7m ago
Completely agree, that there are legitimate reasons to prefer comparison to e.g. deepeek models. But that doesn't change my point, we both agree that the comparisons would be extremely unfavorable.
codybontecou•36m ago
Do all of these models, regardless of parameters, support tool use and structured output?
Y_Y•21m ago
In principle any model can do these. Tool use is just detecting something like "I should run a db query for pattern X" and structured output is even easier, just reject output tokens that don't match the grammar. The only question is how well they're trained, and how well your inference environment takes advantage.
simgt•30m ago
I still don't understand what the incentive is for releasing genuinely good model weights. What makes sense however is OpenAI releasing a somewhat generic model like gpt-oss that games the benchmarks just for PR. Or some Chinese companies doing the same to cut the ground from under the feet of American big tech. Are we really hopeful we'll still get decent open weights models in the future?
prodigycorp•20m ago
gpt-oss are really solid models. by far the best at tool calling, and performant.
talliman•14m ago
Until there is a sustainable, profitable and moat-building business model for generative AI, the competition is not to have the best proprietary model, but rather to raise the most VC money to be well positioned when that business model does arise.

Releasing a near stat-of-the-art open model instanly catapults companies to a valuation of several billion dollars, making it possible raise money to acquire GPUs and train more SOTA models.

Now, what happens if such a business model does not emerge? I hope we won't find out!

memming•3m ago
It’s funny how future money drive the world. Fortunately it’s fueling progress this time around.
NitpickLawyer•7m ago
> gpt-oss that games the benchmarks just for PR.

gpt-oss is killing the ongoing AIME3 competition on kaggle. They're using a hidden, new set of problems, IMO level, handcrafted to be "AI hardened". And gpt-oss submissions are at ~33/50 right now, two weeks into the competition. The benchmarks (at least for math) were not gamed at all. They are really good at math.

mirekrusin•3m ago
Because there is no money in making them closed.

Open weight means secondary sales channels like their fine tuning service for enterprises [0].

They can't compete with large proprietary providers but they can erode and potentially collapse them.

Open weights and research builds on itself advancing its participants creating environment that has a shot at proprietary services.

Transparency, control, privacy, cost etc. do matter to people and corporations.

[0] https://mistral.ai/solutions/custom-model-training

yvoschaap•29m ago
Upvoting for Europe's best efforts.
sebzim4500•13m ago
That's unfair to Europe. A bunch of AI work is done in London (Deepmind is based here for a start)
GaggiX•11m ago
London is not part of Europe anymore since Brexit /s
ot•8m ago
Is it so hard for people to understand that Europe is a continent, EU is a federation of European countries, and the two are not the same?
GaggiX•6m ago
I think you missed the joke
usrnm•3m ago
Europe isn't even a continent and has no real definition, so the whole thing is confusing by design
hnuser123456•25m ago
Looks like their own HF link is broken or the collection hasn't been made public yet. The 14B instruct model is here:

https://huggingface.co/mistralai/Ministral-3-14B-Instruct-25...

The unsloth quants are here:

https://huggingface.co/unsloth/Ministral-3-14B-Instruct-2512...

janpio•14m ago
Seems fixed now:

https://huggingface.co/collections/mistralai/mistral-large-3

https://huggingface.co/collections/mistralai/ministral-3

andhuman•25m ago
This is big. The first really big open weights model that understands images.
yoavm•10m ago
How is this different from Llama 3.2 "vision capabilities"?

https://www.llama.com/docs/how-to-guides/vision-capabilities...

Tiberium•17m ago
A bit interesting that they used Deepseek 3's architecture for their Large model :)
GaggiX•12m ago
The small dense model seems particularly good for their small sizes, I can't wait to test them out.
tucnak•11m ago
If the claims on multilingual and pretraining performance are accurate, this is huge! This may be the best-in-class multilingual stuff since the more recent Gemma's, where they used to be unmatched. I know Americans don't care much about the rest of the world, but we're still using our native tongues thank you very much; there is a huge issue with i.e. Ukrainian (as opposed to Russian) being underrepresented in many open-weight and weight-available models. Gemma used to be a notable exception, I wonder if it's still the case. On a different note: I wonder why scores on TriviaQA vis-a-vis 14b model lags behind Gemma 12b so much; that one is not a formatting-heavy benchmark.

Addressing the adding situation

https://xania.org/202512/02-adding-integers
170•messe•4h ago•48 comments

Learning Music with Strudel

https://terryds.notion.site/Learning-Music-with-Strudel-2ac98431b24180deb890cc7de667ea92
60•terryds•6d ago•13 comments

Nixtml: Static website and blog generator written in Nix

https://github.com/arnarg/nixtml
19•todsacerdoti•1h ago•1 comments

Advent of Compiler Optimisations 2025

https://xania.org/202511/advent-of-compiler-optimisation
220•vismit2000•6h ago•27 comments

Mistral 3 family of models released

https://mistral.ai/news/mistral-3
116•pember•55m ago•25 comments

Show HN: Marmot – Single-binary data catalog (no Kafka, no Elasticsearch)

https://github.com/marmotdata/marmot
20•charlie-haley•58m ago•3 comments

Python Data Science Handbook

https://jakevdp.github.io/PythonDataScienceHandbook/
56•cl3misch•3h ago•11 comments

YesNotice

https://infinitedigits.co/docs/software/yesnotice/
16•surprisetalk•6d ago•4 comments

A series of vignettes from my childhood and early career

https://www.jasonscheirer.com/weblog/vignettes/
76•absqueued•3h ago•45 comments

Apple Releases Open Weights Video Model

https://starflow-v.github.io
305•vessenes•10h ago•98 comments

What will enter the public domain in 2026?

https://publicdomainreview.org/features/entering-the-public-domain/2026/
376•herbertl•12h ago•231 comments

YouTube increases FreeBASIC performance (2019)

https://freebasic.net/forum/viewtopic.php?t=27927
107•giancarlostoro•2d ago•17 comments

Comparing AWS Lambda ARM64 vs. x86_64 Performance Across Runtimes in Late 2025

https://chrisebert.net/comparing-aws-lambda-arm64-vs-x86_64-performance-across-multiple-runtimes-...
83•hasanhaja•6h ago•37 comments

Zig's new plan for asynchronous programs

https://lwn.net/SubscriberLink/1046084/4c048ee008e1c70e/
47•messe•1h ago•31 comments

DeepSeek-v3.2: Pushing the frontier of open large language models [pdf]

https://huggingface.co/deepseek-ai/DeepSeek-V3.2/resolve/main/assets/paper.pdf
891•pretext•1d ago•420 comments

India orders smartphone makers to preload state-owned cyber safety app

https://www.reuters.com/sustainability/boards-policy-regulation/india-orders-mobile-phones-preloa...
820•jmsflknr•1d ago•590 comments

Proximity to coworkers increases long-run development, lowers short-term output

https://pallais.scholars.harvard.edu/publications/power-proximity-coworkers-training-tomorrow-or-...
80•delichon•1h ago•48 comments

Beej's Guide to Learning Computer Science

https://beej.us/guide/bglcs/
247•amruthreddi•2d ago•88 comments

Is 2026 Next Year?

https://www.google.com/search?q=is+2026+next+year&oq=is+2026+next+year
46•kjhughes•37m ago•12 comments

How Brian Eno Created Ambient 1: Music for Airports (2019)

https://reverbmachine.com/blog/deconstructing-brian-eno-music-for-airports/
129•dijksterhuis•8h ago•63 comments

Lazier Binary Decision Diagrams for set-theoretic types

https://elixir-lang.org/blog/2025/12/02/lazier-bdds-for-set-theoretic-types/
15•tvda•3h ago•1 comments

Rootless Pings in Rust

https://bou.ke/blog/rust-ping/
92•bouk•8h ago•61 comments

An LED panel that shows the aviation around you

https://github.com/AxisNimble/TheFlightWall_OSS
40•yzydserd•5d ago•7 comments

Tom Stoppard has died

https://www.bbc.com/news/articles/c74xe49q7vlo
142•mstep•2d ago•40 comments

Reverse math shows why hard problems are hard

https://www.quantamagazine.org/reverse-mathematics-illuminates-why-hard-problems-are-hard-20251201/
141•gsf_emergency_6•13h ago•28 comments

After Windows Update, Password icon invisible, click where it used to be

https://support.microsoft.com/en-us/topic/august-29-2025-kb5064081-os-build-26100-5074-preview-3f...
131•zdw•13h ago•127 comments

Man unexpectedly cured of HIV after stem cell transplant

https://www.newscientist.com/article/2506595-man-unexpectedly-cured-of-hiv-after-stem-cell-transp...
119•doener•5h ago•24 comments

URL in C (2011)

https://susam.net/url-in-c.html
72•birdculture•5d ago•19 comments

Codex, Opus, Gemini try to build Counter Strike

https://www.instantdb.com/essays/agents_building_counterstrike
264•stopachka•3d ago•102 comments

Ghostty compiled to WASM with xterm.js API compatibility

https://github.com/coder/ghostty-web
371•kylecarbs•21h ago•106 comments