frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Valve releases Steam Controller CAD files under Creative Commons license

https://www.digitalfoundry.net/news/2026/05/valve-releases-steam-controller-cad-files-under-creat...
1478•haunter•19h ago•479 comments

Appearing productive in the workplace

https://nooneshappy.com/article/appearing-productive-in-the-workplace/
1234•diebillionaires•19h ago•481 comments

SQLite Is a Library of Congress Recommended Storage Format

https://sqlite.org/locrsf.html
292•whatisabcdefgh•13h ago•78 comments

Diskless Linux boot using ZFS, iSCSI and PXE

https://aniket.foo/posts/20260505-netboot/
107•stereo-highway•8h ago•54 comments

Permacomputing Principles

https://permacomputing.net/principles/
157•andsoitis•9h ago•75 comments

Boris Cherny: TI-83 Plus Basic Programming Tutorial (2004)

https://www.ticalc.org/programming/columns/83plus-bas/cherny/
23•suoken•2d ago•11 comments

Photoshop's challenges with focus, pt. 2

https://unsung.aresluna.org/photoshops-challenges-with-focus-pt-2/
51•frizlab•2d ago•16 comments

Vibe coding and agentic engineering are getting closer than I'd like

https://simonwillison.net/2026/May/6/vibe-coding-and-agentic-engineering/
625•e12e•20h ago•679 comments

Chevrolet Performance eCrate package (400v/200hp)

https://www.chevrolet.com/performance-parts/crate-engines/ecrate
57•mindcrime•2d ago•29 comments

ProgramBench: Can Language Models Rebuild Programs from Scratch?

https://arxiv.org/abs/2605.03546
61•jonbaer•7h ago•35 comments

ZAYA1-8B: An 8B Moe Model with 760M Active Params Matching DeepSeek-R1 on Math

https://firethering.com/zaya1-8b-open-source-math-coding-model/
18•steveharing1•2h ago•18 comments

RSS feeds send me more traffic than Google

https://shkspr.mobi/blog/2026/05/rss-feeds-send-me-more-traffic-than-google/
131•SpyCoder77•10h ago•28 comments

Show HN: Trust – Coding Rust like it's 1989

https://github.com/wojtczyk/trust
34•wojtczyk•5h ago•11 comments

Making LLM Training Faster with Unsloth and NVIDIA

https://unsloth.ai/blog/nvidia-collab
57•segmenta•4h ago•9 comments

SingleRide: Longest route on NYC Subway without visiting the same station twice

https://singleride.nyc/
19•TMWNN•1d ago•9 comments

Show HN: Agent-skills-eval – Test whether Agent Skills improve outputs

https://github.com/darkrishabh/agent-skills-eval
28•darkrishabh•5h ago•5 comments

Google Cloud fraud defense, the next evolution of reCAPTCHA

https://cloud.google.com/blog/products/identity-security/introducing-google-cloud-fraud-defense-t...
316•unforgivenpasta•17h ago•324 comments

From Supabase to Clerk to Better Auth

https://blog.val.town/better-auth
263•stevekrouse•18h ago•193 comments

What I learned making an app for my family

https://mendelgreenberg.com/posts/ourcar/
61•chabad360•21h ago•29 comments

Show HN: Hallucinopedia

http://halupedia.com/
238•bstrama•18h ago•213 comments

The brave souls who bought a used, 340k-mile rental camper van

https://www.thedrive.com/news/meet-the-brave-souls-who-bought-a-used-340000-mile-rental-camper-van
7•PaulHoule•1d ago•0 comments

Pen pal programs endure in a digital age

https://apnews.com/article/pen-pals-letters-comeback-bc87e1b9c229665bafd368e19751d6ca
52•petethomas•1d ago•11 comments

Community firmware for the Xteink X4 e-paper reader

https://github.com/crosspoint-reader/crosspoint-reader
110•dmos62•1d ago•33 comments

The Mathematical Dance Inside Plant Cells

https://www.quantamagazine.org/the-hidden-mathematical-dance-inside-plant-cells-20260504/
50•isaacfrond•2d ago•2 comments

Show HN: Tilde.run – Agent sandbox with a transactional, versioned filesystem

https://tilde.run/
167•ozkatz•19h ago•118 comments

The Vatican's Website in Latin

https://www.vatican.va/latin/latin_index.html
157•ks2048•9h ago•96 comments

Learning the Integral of a Diffusion Model

https://sander.ai/2026/05/06/flow-maps.html
141•benanne•16h ago•21 comments

Building my own Vi text editor in BASIC

https://leetusman.com/nosebook/yvi
62•zeech•1d ago•32 comments

A Theory of Deep Learning

https://elonlit.com/scrivings/a-theory-of-deep-learning/
203•elonlit•1d ago•48 comments

Ted Turner has died

https://www.cnn.com/2026/05/06/us/ted-turner-death
278•pseudolus•20h ago•221 comments
Open in hackernews

ZAYA1-8B: An 8B Moe Model with 760M Active Params Matching DeepSeek-R1 on Math

https://firethering.com/zaya1-8b-open-source-math-coding-model/
18•steveharing1•2h ago

Comments

2ndorderthought•58m ago
I've been saying it for a long time now. I think small models are the future for LLMs. It's been fun seeing experiments to see just how much better models get by making them insanely large but it's not sustainable.

No I am not saying this model is a drop in Claude replacement. But I think in 2 years we might be really surprised what can be done in a desktop with commodity hardware, no connection to the internet, and a few models that span a subset of tasks.

Really happy to see amd put their hat in the ring. It's a good day for amd investors. I know a lot of AI bros will scoff at this, but having your first training run is a big deal for a new lab. AMD is on their way despite Nvidia having years of runway

zimi-24-imiz•49m ago
using C was 100 times as productive as assembly. what happened was not that we finished software 100 times faster, but that we did projects 100 times bigger in the same time

same thing with smol local LLMs versus the big ones in the sky. your smol local LLM will only be able to tackle projects which are not comercially valuable anymore, because people expect 100x scope and features. which is fine as a hobby/art project

yes, we'll do amazing things with local LLMs in 2 years, but the big LLMs will do things beyond imagination (assembly vs C)

2ndorderthought•35m ago
I disagree. I think people can make very good software by balancing their use of AI and their market knowledge. I still believe for the foreseeable future people can make wildly loved or mission critical software with 0 ai and have it be met with market interest.

I think we are going to see a surge in software claiming to do everything and becoming bloated and unsustainable.

I already see 1gpu local models 1 shotting games via vibe coding. I see people doing agentic programming, granted more slowly and cheaply than 12 Claude sessions.

The difference isn't as big as it was 2 months ago. In the past 45 days so many model releases have happened. Meanwhile frontier performance has stagnated and degraded. If it's a taste of what is to come I welcome it.

hparadiz•8m ago
I'm like two months into a vibe coded C project. My issues are the same as ever. How to pack memory. What syscalls to run and when. Is the program stable after running for 24 hours? When I want to make a change it's usually a trade off with something else. There's no accounting for taste among humans. Let alone among an LM. It's great at implementing my ideas but terrible at coming up with those ideas. Architecture is always going to be king.
steveharing1•46m ago
You couldn't be any more right!
zimi-24-imiz•45m ago
but he could be absolutely right
steveharing1•37m ago
He could be right but time will tell if we can really achieve that level in open source space because as you know Even in open source space companies go closed when they achieve something really efficient and frontier. I'm not talking about all but that's usually a pattern
2ndorderthought•31m ago
There are a lot of hats in the ring. I don't see Alibaba shutting down anytime soon. They make qwen.

Deepseek is doing valuations right now.

Moonshot is just getting started. Same with AMD. mistral is still working hard at it and has a customer base.

An Egyptian company dropped their first small model this month, Horus.

There are enough geopolitics at play that I expect this to be a very different outcome from typical startup market dynamics. If anything j worry about the big us labs longevity. The world is fed up with US tech it seems, and even for us citizens it's questionable the frontier labs have their interests in mind as they risk the entire economy.

adrian_b•6m ago
That is a danger, but for now it seems rather distant.

OpenAI has provided in the past a couple of open-weights models, but it does not seem to plan the release of any others.

But except for OpenAI and Anthropic, with this announcement Zyphra is the 12th company which has announced new improved open-weights models during the last couple of months.

A half of these 12 companies have launched not only small models with less than 128B parameters, but also big models with a number of parameters ranging from over 200B to over 1T.

So for now there is a healthy competition and the offerings in open-weights models are very diverse and numerous.

throwaw12•55m ago
> The math and coding part is impressive but the agentic one is not.

I think this is very important to eventually become a viable replacement for coding models. Because most of the time coding harnesses are leveraging tool calls to gather the context and then write a solution.

I am hopeful, that one day we can replace Claude and OpenAI models with local SOTA LLMs

2ndorderthought•51m ago
It's pretty close already. Check qwen3.6 27b if you haven't already. People are vibe and agentic coding with it on a single GPU.

It is more finicky than Claude but if you hand hold it a bit it's crazy.

gchamonlive•33m ago
I see that going around, and either the test cases are too simplistic or I'm doing something wrong. I have a server with a 3090 in it, enough to run qwen3.6, but I haven't had much luck using it with either codex or oh-my-pi. They work, but the model gets really slow with ~64k context and the attention degrades quickly. You'll sometimes execute a prompt, the model will load a test file and say something like "I was presented with a test file but no command. What should I do with it?".

So yeah, while it's true that qwen3.6 is good for agentic coding, it's not very good for exploring the codebase and coming up with plans. You need to pair it today with a model capable of ingesting the whole context and providing a detailed plan, and even then the implementation might take 10x the amount of time it'd take for sonnet or Gemini 3 to crunch through the plan.

2ndorderthought•25m ago
I agree for planning it's not there yet. But I wouldn't be surprised if something came out that was in a similar weight class.
steveharing1•50m ago
That's absolutely possible, its just as we move towards more advancement, We'll soon see Small models being smart enough to not be judged by parameter count but their reasoning and intelligence. You can see examples like Qwen 3.6 27B.
regexorcist•13m ago
Yeah this is key, a lot of people are still just looking at the number of params and thinking these models are toys. What Qwen 3.6 has shown is that reasoning and tool calling are just as important if not more.
yorwba•38m ago
Announcement blogpost: https://www.zyphra.com/post/zaya1-8b
Havoc•5m ago
0.76 active and it's vaguely competitive at coding sounds promising.

LM studio doesn't let me actually run this yet though: "Unsupported safetensors format: null"