frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
377•klaussilveira•4h ago•81 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
742•xnx•10h ago•456 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
112•dmpetrov•5h ago•49 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
132•isitcontent•5h ago•13 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
234•vecti•7h ago•112 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
21•quibono•4d ago•0 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
302•aktau•11h ago•150 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
302•ostacke•10h ago•80 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
156•eljojo•7h ago•117 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
375•todsacerdoti•12h ago•214 comments

A century of hair samples proves leaded gas ban worked

https://arstechnica.com/science/2026/02/a-century-of-hair-samples-proves-leaded-gas-ban-worked/
52•jnord•3d ago•3 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
301•lstoll•11h ago•227 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
42•phreda4•4h ago•7 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
100•vmatsiiako•9h ago•33 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
165•i5heu•7h ago•122 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
136•limoce•3d ago•75 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
35•rescrv•12h ago•17 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
223•surprisetalk•3d ago•29 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
951•cdrnsf•14h ago•411 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
7•kmm•4d ago•0 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
7•gfortaine•2h ago•0 comments

I'm going to cure my girlfriend's brain tumor

https://andrewjrod.substack.com/p/im-going-to-cure-my-girlfriends-brain
28•ray__•1h ago•4 comments

The Oklahoma Architect Who Turned Kitsch into Art

https://www.bloomberg.com/news/features/2026-01-31/oklahoma-architect-bruce-goff-s-wild-home-desi...
17•MarlonPro•3d ago•2 comments

Show HN: Smooth CLI – Token-efficient browser for AI agents

https://docs.smooth.sh/cli/overview
76•antves•1d ago•56 comments

Claude Composer

https://www.josh.ing/blog/claude-composer
94•coloneltcb•2d ago•67 comments

Evaluating and mitigating the growing risk of LLM-discovered 0-days

https://red.anthropic.com/2026/zero-days/
31•lebovic•1d ago•11 comments

Show HN: Slack CLI for Agents

https://github.com/stablyai/agent-slack
36•nwparker•1d ago•7 comments

How virtual textures work

https://www.shlom.dev/articles/how-virtual-textures-really-work/
22•betamark•12h ago•22 comments

Masked namespace vulnerability in Temporal

https://depthfirst.com/post/the-masked-namespace-vulnerability-in-temporal-cve-2025-14986
31•bmit•6h ago•3 comments

Evolution of car door handles over the decades

https://newatlas.com/automotive/evolution-car-door-handle/
38•andsoitis•3d ago•61 comments
Open in hackernews

Mistral Integration Improved in Llama.cpp

https://github.com/ggml-org/llama.cpp/pull/14737
95•decide1000•5mo ago

Comments

baggiponte•5mo ago
Wow I never realized how much mistral was “disconnected” from the ecosystem
hodgehog11•5mo ago
I appreciate Mistral (and others) releasing their weights for free. But given how llama.cpp underpins a lot of the programs which allow users to run open weight models, it is a little frustrating to have companies which brag about releasing models to the community, leave the community to their own devices to slowly try and actually implement their models.

I hear the reason for this is that llama.cpp keeps breaking basic things, so they have become an unreliable partner. Seems this is what Ollama is trying to address by diluting their connections to llama.cpp and directly contacting companies training these models to have simultaneous releases (e.g. GPT-OSS).

mattnewton•5mo ago
There are many different inference libraries and it's not clear which ones a small company like mistral should back yet IMO.

They do release high quality inference code, ie https://github.com/mistralai/mistral-inference

bastawhiz•5mo ago
There's more to it, though. The inference code you linked to is Python. Unless my software is Python, I have to ship a CPython binary to run the inference code, then wire it up (or port it, if you're feeling spicy).

Ollama brings value by exposing an API (literally over sockets) with many client SDKs. You don't even need the SDKs to use it effectively. If you're writing Node or PHP or Elixir or Clojurescript or whatever else you enjoy, you're probably covered.

It also means that you can swap models trivially, since you're essentially using the same API for each one. You never need to worry about dependency hell or the issues involved in hosting more than one model at a time.

As far as I know, Ollama is really the only solution that does this. Or at the very least, it's the most mature.

refulgentis•5mo ago
The relationship between Ollama and llama.cpp is massively closer than it must seem.

Ollama is llama.cpp with a nice little installer GUI and nice little server binary.

llama.cpp has a server binary as well, however, no nice installer GUI.

The only time recently Ollama had a feature llama.cpp didn't was they patched SWA in with Google, llama.cpp had it a couple weeks later.

Ollama is significantly behind llama.cpp in important areas, ex. the Gemma blog post, they note they'll get on tool calls and multimodal real soon now.

bastawhiz•5mo ago
I don't care about llama.cpp, just like I don't care about V8 when I reach for Node. And I suspect many other people don't, either. Lots of folks don't want to integrate a library. They don't want to download a model or weights. They want to `ollama run foo` and move on with their lives. I don't need to worry about whether my binary was compiled with the right flags on my MacBook versus a Linux server with an Nvidia GPU or setting gpu-layers or num_ctx.

> Ollama is significantly behind llama.cpp in important areas, ex. the Gemma blog post, they note they'll get on tool calls and multimodal real soon now.

If you don't use those things, you don't need to care. I'll just use another model that works.

And that's the thing really. Most folks don't give a shit about getting the maximum performance. They're probably not even keeping their GPU busy all the time. They just need it to work consistently without having to worry about nonsense. Llama.cpp simply isn't that tool.

refulgentis•5mo ago
Nah, llama.cpp is stable.

llama.cpp also got GPT-OSS early, like Ollama.

There's a lot of extremely subtle politics going on in the link.

Suffice it to say, as a commercial entity, there's a very clever way to put your thumb on the scale of what works and what doesn't without it being obvious to anyone involved, even the thumb.

hodgehog11•5mo ago
Stable for a power user, or stable for everyone? I don't have links on hand, but I could swear there have been instances where certain models rolled back support during llama.cpp development, and this was recent. Also llama.cpp adds features and support on a near-daily basis, how can this be LTS?

Don't get me wrong, llama.cpp is an amazing tool. But it's development is nowhere near as cautious as something like the Linux kernel, so there is room there for a more stable alternative. Not saying Ollama will do this, but llama.cpp won't be everything to everyone.

refulgentis•5mo ago
I'd start by noting all software adds features and code on a near-daily basis. (* modulo weekends and holidays and lack of interest in further development)

I'm not sure comparing to Linux kernel sheds light: what is different? Just Ubuntu/Red Hat LTS type stuff? What does LTS mean in the context of not-support-contracts and not-operating systems?

Steelmaning, I could say we mean....named branches? I guess a branch isn't a necessary condition...named versions?...that get fixes backported, but no new features.

Software where that's a commonly used approach are at least ~3 OOMs larger (i.e. are much more separable in terms of bug fixes vs. features and components) and hard to upgrade, i.e. it's hard for IT to force all N changes on end users since the last time they upgraded Linux machines, just to get a 0 day fix.

Here, it's a FOSS software library that needs to be part of an app to be useful, the consumers of the library are the ones would want to offer LTS.

I'm all ears if you dig up more info on a rollback or similar nasty scandal, but as it stands, I've been involved with it near-daily for 2 years now, including CI tests on every platform you can think of, and I've never, ever, heard of such a thing.

A guiding light here may be that Ollama inference is 99% llama.cpp or its consituents. From there, we notice a contradiction: if thats the case, how can we claim Ollama fulfills these ideas but llama.cpp doesn't? We could wave it away as they have a miraculous nose for what parts of llama.cpp won't fall victim to the issues we're worried about, but...well, here's one of my favorite quotes: "When faced with a contradiction, first, check your premises"

mhitza•5mo ago
llama.cpp still doesn't support gpt-oss tool calling. https://github.com/ggml-org/llama.cpp/pull/15158 (among other similar PRs)

But I also couldn't get vllm, or transformers serve, or ollama (400 response on /v1/chat/completions) working today with gpt-oss. OpenAI's cookbooks aren't really copy paste instructions. They probably tested on a single platform with preinstalled python packages which they forgot to mention :))

refulgentis•5mo ago
Re: gpt-oss tool calls support, I don't think that's true, I've been using it for days. Then again, I did write my own harmony parser...(Noting for audience as you imply, neither does Ollama. Thing here is you either gotta hope all your users have nicely formed templates in their ggufs (they do not) or sometimes step in to ex. here, note the OpenAI chat completions-alike API llama.cpp provides will output a text response that you'll need to parse into a tool call yourself, until they implement a harmony parser)
electroglyph•5mo ago
gpt-oss are still being actively fixed right this moment, and there have already been quite a few fixes.
flakiness•5mo ago
> We are using mistral-common internally for tokenization and want the community to use it to unlock full capacities of our models. As mistral-common is a Python library, we have opened a PR to add a REST API via FastAPI to make it easier for users who are not in the Python ecosystem.

A cpp binary depending on a python server is a bit sad.

I hope this is a stopgap measure and someone port it to C++ eventually:https://github.com/mistralai/mistral-common/blob/main/src/mi...

the_mitsuhiko•5mo ago
Isn’t llama.cpp already depending on Python anyways for the templating?
Maxious•5mo ago
It uses a cpp implementation of jinja https://github.com/google/minja