frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Wikipedia loses challenge against Online Safety Act

https://www.bbc.com/news/articles/cjr11qqvvwlo
357•phlummox•5h ago•357 comments

Why tail-recursive functions are loops

https://kmicinski.com/functional-programming/2025/08/01/loops/
37•speckx•3d ago•21 comments

I tried every todo app and ended up with a .txt file

https://www.al3rez.com/todo-txt-journey
664•al3rez•8h ago•435 comments

Neki – sharded Postgres by the team behind Vitess

https://planetscale.com/blog/announcing-neki
100•thdxr•4h ago•12 comments

OpenSSH Post-Quantum Cryptography

https://www.openssh.com/pq.html
312•throw0101d•10h ago•86 comments

Show HN: Play Pokémon to unlock your Wayland session

https://github.com/AdoPi/wlgblock
10•anajimi•1d ago•2 comments

The Value of Institutional Memory

https://timharford.com/2025/05/the-value-of-institutional-memory/
82•leoc•5h ago•40 comments

GitHub is no longer independent at Microsoft after CEO resignation

https://www.theverge.com/news/757461/microsoft-github-thomas-dohmke-resignation-coreai-team-transition
816•Handy-Man•6h ago•584 comments

Byte Buddy is a code generation and manipulation library for Java

https://bytebuddy.net/
54•mooreds•3d ago•17 comments

The Joy of Mixing Custom Elements, Web Components, and Markdown

https://deanebarker.net/tech/blog/custom-elements-markdown/
59•deanebarker•6h ago•23 comments

Claude Code is all you need

https://dwyer.co.za/static/claude-code-is-all-you-need.html
388•sixhobbits•8h ago•229 comments

Launch HN: Halluminate (YC S25) – Simulating the internet to train computer use

34•wujerry2000•6h ago•26 comments

Ollama and gguf

https://github.com/ollama/ollama/issues/11714
40•indigodaddy•4h ago•15 comments

UI vs. API. vs. UAI

https://www.joshbeckman.org/blog/practicing/ui-vs-api-vs-uai
60•bckmn•6h ago•23 comments

How Boom uses software to accelerate hardware development

https://bscholl.substack.com/p/move-fast-and-dont-break-safety-critical
47•flabber•1d ago•37 comments

Learn, Reflect, Apply, Prepare: The Four Daily Practices That Changed How I Live

https://opuslabs.substack.com/p/learn-reflect-apply-prepare
58•opuslabs•6h ago•10 comments

Trellis (YC W24) Is Hiring: Automate Prior Auth in Healthcare

https://www.ycombinator.com/companies/trellis/jobs/Cv3ZwXh-forward-deployed-engineers-all-levels-august-2025
1•jackylin•5h ago

The demographic future of humanity: facts and consequences [pdf]

https://www.sas.upenn.edu/~jesusfv/Slides_London.pdf
44•akyuu•5h ago•88 comments

I've seen 12 people hospitalized after losing touch with reality because of AI

https://twitter.com/KeithSakata/status/1954884361695719474
91•fortran77•1h ago•54 comments

Pricing Pages – A Curated Gallery of Pricing Page Designs

https://pricingpages.design/
159•finniansturdy•9h ago•47 comments

White Mountain Direttissima

https://whitemountainski.co/pages/white-mountain-direttissima
24•oftenwrong•3d ago•10 comments

36B solar mass black hole at centre of the Cosmic Horseshoe gravitational lens

https://academic.oup.com/mnras/article/541/4/2853/8213862?login=false
93•bookofjoe•7h ago•66 comments

A Guide Dog for the Face-Blind

https://asimov.blog/a-guide-dog-for-the-face-blind/
16•arto•3d ago•8 comments

Designing Software in the Large

https://dafoster.net/articles/2025/07/22/designing-software-in-the-large/
64•davidfstr•7h ago•23 comments

AP to end its weekly book reviews

https://dankennedy.net/2025/08/08/the-associated-press-tells-its-book-critics-that-its-ending-weekly-reviews/
65•thm•4h ago•22 comments

Porting to OS/2 (1987)

https://gitpi.us/article-archive/porting-to-os2/
36•rbanffy•4d ago•4 comments

Faster substring search with SIMD in Zig

https://aarol.dev/posts/zig-simd-substr/
168•todsacerdoti•12h ago•49 comments

Mistral Integration Improved in Llama.cpp

https://github.com/ggml-org/llama.cpp/pull/14737
78•decide1000•12h ago•7 comments

Optimizing my sleep around Claude usage limits

https://mattwie.se/no-sleep-till-agi
118•mattwiese•20h ago•95 comments

A simple pixel physics simulator in Rust using Macroquad

https://github.com/gale93/sbixel
42•sbirulo•4d ago•2 comments
Open in hackernews

Ollama and gguf

https://github.com/ollama/ollama/issues/11714
40•indigodaddy•4h ago

Comments

indigodaddy•4h ago
ggerganov explains the issue: https://github.com/ollama/ollama/issues/11714#issuecomment-3...
magicalhippo•4h ago
I noticed it the other way, llama.cpp failed to download the Ollama-downloaded gpt-oss 20b model. Thought it was odd given all the others I tried worked fine.

Figured it had to be Ollama doing Ollama things, seems that was indeed the case.

polotics•1h ago
ggerganov is my hero, and... it's a good thing this got posted so I saw in the comments that --flash-attn --cache-reuse 256 could help with my setup (M3 36GB + RPC to M1 16GB) figuring out what params to set and at what value is a lot of trial and error, Gemini does help a bit clarify what params like top-k are going to do in practice. Still the whole load-balancing with RPC is something I think I'm going to have to read the source of llama.cpp to really understand (oops I almost wrote grok, damn you Elon) Anyways ollama is still not doing distributed load, and yeah I guess using it is a stepping stone...
LeoPanthera•59m ago
The named anchor in this URL doesn't work in Safari. Safari correctly scrolls down to the comment in question, but then some Javascript on the page throws you back up to the top again.
dcreater•3h ago
I think the title buries the lede? Its specific to GPT-OSS and exposes the shady stuff Ollama is doing to acquiesce/curry favor/partner with/get paid by corporate interests
freedomben•1h ago
I think "shady" is a little too harsh - sounds like they forked an important upstream project, made incompatible changes that they didn't push upstream or even communicate with upstream about, and now have to deal with the consequences of that. If that's "shady" (despite being all out in the open) then nearly every company I've worked for has been "shady."
wsgeorge•1h ago
There's a reddit thread from a few months ago that sort of explains what people don't like about ollama, that "shadiness" parent references:

https://www.reddit.com/r/LocalLLaMA/comments/1jzocoo/finally...

llmthrowaway•1h ago
Confusing title - thought this was about Ollama finally supporting sharded GGUF (ie. the Huggingface default for large gguf over 48gb).

https://github.com/ollama/ollama/issues/5245

Sadly it is not and the issue still remains open after over a year meaning ollama cannot run the latest SOTA open source models unless they covert them to their proprietary format which they do not consistently do.

No surprise I guess given they've taken VC money, refuse to properly attribute the use things like llama.cpp and ggml, have their own model format for.. reasons? and have over 1800 open issues...

Llama-server, ramallama or whatever model switcher ggerganov is working on (he showed previews recently) feel like the way forward.

tarruda•1h ago
I recently discovered that ollama no longer uses llama.cpp as a library, and instead they link to the low level library (ggml) which requires them to reinvent a lot of wheel for absolutely no benefit (if there's some benefit I'm missing, please let me know).

Even using llama.cpp as a library seems like an overkill for most use cases. Ollama could make its life much easier by spawning llama-server as a subprocess listening on a unix socket, and forward requests to it.

One thing I'm curious about: Does ollama support strict structured output or strict tool calls adhering to a json schema? Because it would be insane to rely on a server for agentic use unless your server can guarantee the model will only produce valid json. AFAIK this feature is implemented by llama.cpp, which they no longer use.

arcanemachiner•30m ago
> I recently discovered that ollama no longer uses llama.cpp as a library, and instead they link to the low level library (ggml) which requires them to reinvent a lot of wheel for absolutely no benefit (if there's some benefit I'm missing, please let me know).

Here is some relevant drama on the subject:

https://github.com/ollama/ollama/issues/11714#issuecomment-3...

hodgehog11•28m ago
I got to speak with some of the leads at Ollama and asked more or less this same question. The reason they abandoned llama.cpp is because it does not align with their goals.

llama.cpp is designed to rapidly adopt research-level optimisations and features, but the downside is that reported speeds change all the time (sometimes faster, sometimes slower) and things break really often. You can't hope to establish contracts with simultaneous releases if there is no guarantee the model will even function.

By reimplementing this layer, Ollama gets to enjoy a kind of LTS status that their partners rely on. It won't be as feature-rich, and definitely won't be as fast, but that's not their goal.

halyconWays•13m ago
>(if there's some benefit I'm missing, please let me know).

Makes their VCs think they're doing more, and have more ownership, rather than being a do-nothing wrapper with some analytics and S3 buckets that rehost models from HF.

wubrr•1m ago
> Does ollama support strict structured output or strict tool calls adhering to a json schema?

As far as I understand this is generally not possible at the model level. Best you can do is wrap the call in a (non-llm) json schema validator, and emit an error json in case the llm output does not match the schema, which is what some APIs do for you, but not very complicated to do yourself.

Someone correct me if I'm wrong

12345hn6789•38m ago
Just days ago ollama devs claimed[0] that ollama no longer relies on ggml / llama.cpp. here is their pull request(+165,966 −47,980) to reimplement (copy) llama.cpp code in their repository.

https://github.com/ollama/ollama/pull/11823

[0] https://news.ycombinator.com/item?id=44802414#44805396

flakiness•25m ago
not against overall sentiment here, but quote the counterpoint from the linked HN comment to be fair:

> Ollama does not use llama.cpp anymore; we do still keep it and occasionally update it to remain compatible for older models for when we used it.

The linked PR is doing "occasionally update it" I guess? Note that "vendored" in the PR title often means to take a snapshot to pin a specific version.