frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

What the Eurostack Is Missing

https://pluralistic.net/2025/10/15/freedom-of-movement/#data-dieselgate
1•FromTheArchives•7m ago•0 comments

Show HN: maptail – Tail GeoIP data on a world map in realtime

https://maptail.live/
1•stagas•7m ago•0 comments

China's Rare Earth Restrictions Aim to Beat U.S. at Its Own Game

https://www.nytimes.com/2025/10/16/business/economy/china-rare-earths-supply-chain.html
1•curtistyr•12m ago•0 comments

Inverse Collatz's Tape

https://gbragafibra.github.io/2025/10/15/collatz_ant10.html
1•Fibra•14m ago•0 comments

Technology is my leverage, not design

https://herland.me/blog/technology-is-my-leverage-not-design/
1•gogoro•15m ago•0 comments

The Economic Cost of Antisemitism

https://critiqueanddigest.substack.com/p/the-economic-cost-of-antisemitism
2•omnibrain•16m ago•0 comments

The present and potential future of progressive image rendering

https://jakearchibald.com/2025/present-and-future-of-progressive-image-rendering/
1•FromTheArchives•19m ago•0 comments

Sora2 AI Video Generator

https://www.sora2-ai.top
1•detectmeai•19m ago•0 comments

Don't Stop Believin' in OpenAI

https://buildcognitiveresonance.substack.com/p/dont-stop-believin-in-openai
1•FromTheArchives•22m ago•0 comments

The Slack I Loved Is Slipping Away

https://sjg.io/writing/the-slack-i-loved-is-slipping-away/
1•simonjgreen•22m ago•1 comments

Where's the AI Design Renaissance?

https://www.learnui.design/blog/wheres-the-ai-design-renaissance.html
1•tobr•23m ago•0 comments

Haskell Weekly – Issue 494

https://haskellweekly.news/issue/494.html
2•amalinovic•29m ago•0 comments

MuPDF Explored (2022)

https://casper.mupdf.com/docs/mupdf_explored.pdf
1•nyir•30m ago•0 comments

Why does collapsing a bubble with a sound wave produce light?

https://akshatjiwannotes.blogspot.com/2025/10/why-does-collapsing-bubble-with-sound.html
3•akshatjiwan•31m ago•0 comments

AI agents are on the verge of being recognized as full-fledged workers

https://www.lemonde.fr/en/opinion/article/2025/10/16/ai-agents-are-on-the-verge-of-being-recogniz...
2•geox•38m ago•0 comments

The Beads Revolution: The Todo System That AI Agents Want to Use

https://steve-yegge.medium.com/the-beads-revolution-how-i-built-the-todo-system-that-ai-agents-ac...
1•SafeDusk•39m ago•0 comments

Waymo is bringing autonomous, driverless ride-hailing to London in 2026

https://9to5google.com/2025/10/15/waymo-london-2026/
5•pykello•40m ago•0 comments

WordPress Sub Menu or Mega Menu hidden behind other elements using Elementor [video]

https://www.youtube.com/watch?v=qsgfx1hlJwA
1•techwrath11•44m ago•0 comments

The Architect's Dilemma

https://www.oreilly.com/radar/the-architects-dilemma/
3•BerislavLopac•45m ago•0 comments

Understanding Spec-Driven-Development

https://martinfowler.com/articles/exploring-gen-ai/sdd-3-tools.html
4•BerislavLopac•45m ago•0 comments

3D Low Poly Character of Master Chef or Iron Chef in Blender

https://www.patreon.com/posts/3d-low-poly-of-103411260
1•techwrath11•45m ago•0 comments

Driverless taxis are coming to London's streets in the spring

https://www.londoncentric.media/p/driverless-taxis-waymo-wayve-are-coming-to-london
1•BerislavLopac•45m ago•0 comments

Introducing the Massive Legal Embedding Benchmark (MLEB)

https://isaacus.com/blog/introducing-mleb
4•ubutler•47m ago•4 comments

Running Kubernetes on a Million Nodes

https://bchess.github.io/k8s-1m/
2•yankcrime•49m ago•0 comments

I switched from Ruby to elixir and built a product to learn new tool

https://alexsinelnikov.blog/how-i-switched-from-ruby-to-elixir-and-to-learn-it-better-built-a-pro...
2•avdept•50m ago•0 comments

Agents, APIs, and Advertising: Lessons from Engineering Our MCP Server

https://medium.com/criteo-engineering/agents-apis-and-advertising-lessons-from-engineering-our-mc...
7•ouvreboite•52m ago•0 comments

The Rest of the World Is Following America's Retreat on EVs

https://www.wsj.com/business/autos/the-rest-of-the-world-is-following-americas-retreat-on-evs-e46...
1•measurablefunc•53m ago•2 comments

Comparison of Terminal Emulators

https://blog.randomstring.org/2025/09/26/a-comparison-of-terminal-emulators/
1•gasull•56m ago•0 comments

NASA let me test my weird chain theory in space [video]

https://www.youtube.com/watch?v=NtZaP8VMv0c
2•alexmolas•1h ago•0 comments

Homeless Man AI Prank Prompt: Risks, Ethics, and How to Use AI Responsibly

https://ray3.run/posts/homeless-man-ai-prank-prompt
2•combineimages•1h ago•0 comments
Open in hackernews

New coding models and integrations

https://ollama.com/blog/coding-models
83•meetpateltech•3h ago

Comments

qwe----3•2h ago
Just a paste of llama.cpp without attribution.
mchiang•2h ago
https://github.com/ollama/ollama?tab=readme-ov-file#supporte...
swyx•2h ago
i mean they have attributed but also it's open source software, i guess the more meaningful question is why didn't ggerganov build Ollama if it was that easy? or what is his company working on now?
speedgoose•2h ago
Ollama is more than a paste. But the support for GLM 4.6 is indeed coming from llama.cpp: https://github.com/ollama/ollama/issues/12505#issuecomment-3...

I don’t know how much Ollama contributes to llama.cpp

am17an•1h ago
The answer is 0
jhancock•2h ago
I've been using GLM-4.6 since its release this month. It's my new fav. Using it via Claude Code and the more simple Octofriend https://github.com/synthetic-lab/octofriend

Hosting through z.ai and synthetic.new. Both good experiences. z.ai even answers their support emails!! 5-stars ;)

mchiang•2h ago
Z.ai team is awesome and very supportive. I have yet to try synthetic.new. What's the reason for using multiple? Is it mainly to try different models or are you hitting some kind of rate limit / usage limit?
jhancock•2h ago
I tried synthetic.new prior to GLM-4.6...Starting in August...So I already had a subscription.

When z.ia launched GLM-4.6, I subscribed to their Coding Pro plan. Although I haven't been coding as heavy this month as the prior two months, I used to hit Claude limits almost daily, often twice a day. That was with both the $20 and $100 plans. I have yet to hit a limit with z.ai and the server response is at least as good as Claude.

I mention synthetic.new as it's good to have options and I do appreciate them sponsoring the dev of Octofriend. z.ai is a China company and I think hosts in Singapore. That could be a blocker for some.

mchiang•2h ago
Do you find yourself sticking with GLM 4.6 over Claude for some tasks? Or do you find yourself still wanting to reach for Claude?
jhancock•2h ago
I have been subscribing to both Claude and ChatGPT for over two years. Spent several months on Claude's $100 plan and couple months on ChatGPT's $200 plan but otherwise using their $20/month plans.

I cancelled Claude two weeks ago. Pure GLM-4.6 now and a tad of codex with my ChatGPT Pro subscription. I sometimes use ChatGPT for extended research stuff and non-tech.

hodgehog11•1h ago
My experience using GLM-4.6 with Charm Crush has been absolutely incredible, especially with high thinking. This is on pretty hard tasks too, e.g. proving small lemmas with Lean.

I've had much less luck with other agentic software, including Claude Code. For these kinds of tasks, only Codex seems to come close.

mike_d•2h ago
> For users with more than 300GB of VRAM, qwen3-coder:480b is also available locally.

I haven't really stayed up on all the AI specific GPUs, but are there really cards with 300GB of VRAM?

bakugo•2h ago
No, you need multiple GPUs. These models are not intended to be run by the average user.
Hamuko•2h ago
You can buy an M3 Ultra Mac Studio and configure it with 512 GB of memory shared between the CPU and the GPU. Will set you back about $9500.
bigyabai•2h ago
Been disappointed to see Ollama list models that are supported by the cloud product but not the Ollama app. It's becoming increasingly hard to deny that they're only interested in model inference just to turn a quick buck.
mchiang•2h ago
Qwen3-coder:30b is in the blog post. This is one that most users will be able to run locally.

We are in this together! Hoping for more models to come from the labs in varying sizes that will fit on devices.

bigyabai•1h ago
I'm looking forward to future ollama releases that might attempt parity with the cloud offerings. I've since moved onto the Ollama compatibility API on KoboldCPP since they don't have any such limits with their inference server.
mchiang•1h ago
I am super hopeful! Hardware is improving, inference costs will continue to decrease, models will only improve...
Balinares•28m ago
How does Qwen3-Coder:30B compare to Instruct-2507 as a coding agent backend? I was under the impression that Instruct was intended to supersede Coder?
colesantiago•1h ago
I know this is disappointing, but what business model would be best here for ollama?

1. Donationware - Let's be real, tokens are expensive and if they ask for everyone to chip in voluntarily people wouldn't do that and Ollama would go bust quickly.

2. Subscriptions (bootstrapped and no VCs) again like 1. people would have to pay for the cloud service as a subscription to be sustainable (would you?) or go bust.

3. Ads - Ollama could put ads in the free version but to remove them the users can pay for a higher tier, a somewhat good compromise, except developers don't like ads and don't like pay for their tools unless their company does it for them. No users = Ollama goes bust.

4. VCs - This is the current model which is why they have a cloud product and it keeps the main product free (for now). Again, if they cannot make money or sell to another company Ollama goes bust.

5. Fully Open Source (and 100% free) with Linux Foundation funding - Ollama could also go this route, but this means they wouldn't be a business anymore for investors and rely on the Linux Foundation's sponsors (Google, IBM, etc) for funding the LF to stay sustainable. The cloud product may stay for enterprises.

Ollama has already taken money from investors so they need to produce a return for them so 5. isn't an option in the long term.

6. Acquisition by another company - Ollama could get acquired and the product wouldn't change* (until the acquirer jacks up prices or messes with the product) which ultimately kills it anyway as the community moves on.

I don't see any other way that Ollama can not be enshittified without making a quick buck.

You just need to avoid VC backed tools and pay for bootstrapped ones without any ties to investors.

zozbot234•1h ago
Aren't these models consistently quite large and hard to run locally? It's possible that future Ollama releases will allow you to dynamically manage VRAM memory in a way that enables these models to run with acceleration on even modest GPU hardware (such as by dynamically loading layers for a single 'expert' into VRAM, and opportunistically batching computations that happen to rely on the same 'expert' parameters - essentially doing manually what mmap does for you in CPU-only inference) but these 'tricks' will nonetheless come at non-trivial cost in performance.
zozbot234•1h ago
For those interested in building Ollama locally, note that as of a few hours ago, experimental Vulkan Compute support (will not be in official binary releases as of yet) has been merged on the github main branch and you can test it on your hardware!
mchiang•1h ago
this one is exciting. It'll enable and accelerate a lot of devices on Ollama - especially around AMD GPUs not fully supported by ROCm, Intel GPUs, and iGPUs across different hardware vendors.
qqxufo•59m ago
Interesting to see more people mentioning GLM-4.6 lately — I’ve tried it briefly and it’s surprisingly strong for reasoning tasks. Curious how it compares to Claude 3.5 in coding throughput though?