frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
531•klaussilveira•9h ago•148 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
861•xnx•15h ago•520 comments

How we made geo joins 400× faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
72•matheusalmeida•1d ago•13 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
180•isitcontent•9h ago•21 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
182•dmpetrov•10h ago•81 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
294•vecti•11h ago•130 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
71•quibono•4d ago•13 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
343•aktau•16h ago•168 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
339•ostacke•15h ago•90 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
434•todsacerdoti•17h ago•226 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
238•eljojo•12h ago•147 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
374•lstoll•16h ago•252 comments

Delimited Continuations vs. Lwt for Threads

https://mirageos.org/blog/delimcc-vs-lwt
13•romes•4d ago•2 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
41•kmm•4d ago•3 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
6•videotopia•3d ago•0 comments

Show HN: ARM64 Android Dev Kit

https://github.com/denuoweb/ARM64-ADK
14•denuoweb•1d ago•2 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
220•i5heu•12h ago•162 comments

Why I Joined OpenAI

https://www.brendangregg.com/blog/2026-02-07/why-i-joined-openai.html
91•SerCe•5h ago•75 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
62•phreda4•9h ago•11 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
162•limoce•3d ago•82 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
38•gfortaine•7h ago•11 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
127•vmatsiiako•14h ago•54 comments

Female Asian Elephant Calf Born at the Smithsonian National Zoo

https://www.si.edu/newsdesk/releases/female-asian-elephant-calf-born-smithsonians-national-zoo-an...
18•gmays•4h ago•2 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
261•surprisetalk•3d ago•35 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
1029•cdrnsf•19h ago•428 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
55•rescrv•17h ago•18 comments

Show HN: Smooth CLI – Token-efficient browser for AI agents

https://docs.smooth.sh/cli/overview
84•antves•1d ago•60 comments

WebView performance significantly slower than PWA

https://issues.chromium.org/issues/40817676
18•denysonique•6h ago•2 comments

Zlob.h 100% POSIX and glibc compatible globbing lib that is faste and better

https://github.com/dmtrKovalenko/zlob
5•neogoose•2h ago•2 comments

I'm going to cure my girlfriend's brain tumor

https://andrewjrod.substack.com/p/im-going-to-cure-my-girlfriends-brain
109•ray__•6h ago•55 comments
Open in hackernews

Show HN: TetrisBench – Gemini Flash reaches 66% win rate on Tetris against Opus

https://tetrisbench.com/tetrisbench/
111•ykhli•1w ago

Comments

akomtu•1w ago
It would be more interesting to make it build a chess engine and compare it against Stockfish. The chess engine should be a standalone no-dependencies C/C++ program that fits in NNN lines of code.
gpm•1w ago
Comparing against stockfish isn't fair. That's comparing against enormous amounts of compute spent experimenting with strategies, training neutral nets, etc.

It will lose so badly there will be no point in the comparison.

Besides you could compare models (and harnesses) directly against eachother.

akomtu•1w ago
Stockfish is a good reference point, an objective measure of how far the LLM's advanced.
mikkupikku•1w ago
It's not. Maybe if you used old versions of stockfish that predate the neural net methods used by current versions, because otherwise you'd be comparing the hand-rolled (by an LLM) position evaluation functions against an NNUE and the results of that are a forgone conclusion; stockfish will stomp it every time.

Maybe that's the result you want for some sort of rhetorical reason, but it would nonetheless not be an informative test.

vunderba•1w ago
My back-of-the-envelope guess would be that 99% of LLMs given the task to build a chess engine would probably just end up implementing a flavor of negamax and calling it a day.

https://en.wikipedia.org/wiki/Negamax

ykhli•1w ago
oh that is super interesting. ty for the idea!
OGEnthusiast•1w ago
Gemini 3 Flash is at a very nice point along the price-performance curve. A good workhorse model, while supplementing it with Opus 4.5 / Gemini 3 Pro for more complex tasks.
bubblesorting•1w ago
Very cool! I am a good Tetris player (in the top 10% of players) and wanted to give brick yeeting against an LLM a spin.

Some feedback: - Knowing the scoring system is helpful when going 1v1 high score

- Use a different randomization system, I kept getting starved for pieces like I. True random is fine, throwing a copy of every piece into a bag and then drawing them one by one is better (7 bag), nearly random with some lookbehind to prevent getting a string of ZSZS is solid, too (TGM randomizer)

- Piece rotation feels left-biased, and keeps making me mis-drop, like the T pieces shift to the left if you spin 4 times. Check out https://tetris.wiki/images/thumb/3/3d/SRS-pieces.png/300px-S... or https://tetris.wiki/images/b/b5/Tgm_basic_ars_description.pn... for examples of how other games are doing it.

- Clockwise and counter-clockwise rotation is important for human players, we can only hit so many keys per second

- re-mappable keys are also appreciated

Nice work, I'm going to keep watching.

qsort•1w ago
The worst thing is that the delayed auto shift is slightly off and it messes my finesse. (I used to play competitive tetris as well, but between getting older -> worse reflexes and vision problems I can't really play anymore. Weirdly, finesse muscle memory is still working.)

I don't think the goal is to make a PvP simulator, it would be too easy to cheese or do weird strategies. It's mostly for LLMs to play.

bubblesorting•1w ago
Hello fellow Tetris nerd with a -sort username :)

On the topic of reflexes decaying (I'm getting there, in my late 30s): Have you played Stackflow? It's a number go up roguelite disguised as an arcade brick stacking game, but the gravity is low enough that it is effectively turn based. More about 'deck' building, less about chaining PCs and C-Spins.

kinduff•1w ago
Stackflow looks nice! I'm a Balatro fan and I didn't know about this variant.

By the way, kudos on your feedback. If I was OP, I would've been honored to get that type of fine-tuning comments.

vunderba•1w ago
I actually grew up playing the Spectrum HoloByte version of Tetris for PC, which only lets you rotate in one direction. As a result, I ended up playing NES Tetris for years as a kid before realizing it lets you rotate clockwise / counterclockwise!

https://en.wikipedia.org/wiki/Tetris_(Spectrum_HoloByte)

Keyframe•1w ago
cool to hear about more randomizers. I am not a great tetris player but I absolutely love the game. I put NES, 7-bag and my own take on randomizer (crap) in my online tetris-like experiment https://www.susmel.com/stacky/ (you can press c for more controls or h to see shortcuts)

One of my dream goals was to make a licensed low lag competitive game kind of like TGM, but I heard licensing is extremely cost-prohibitive so I kind of gave up on that goal. I remember I said to someone I was ready to pony up few tens of thousands for a license + cut, but reportedly it starts at an order of the magnitude higher.

ykhli•1w ago
Thanks so much for the amazing feedback!!! Will update the app to incorporate these
arendtio•1w ago
There are some concepts clashing here.

I mean, if you let the LLM build a testris bot, it would be 1000x better than what the LLMs are doing. So yes, it is fun to win against an AI, but to be fair against such processing power, you should not be able to win. It is only possible because LLMs are not built for such tasks.

i_cannot_hack•1w ago
Fun fact: Humans were not build for playing Tetris either!
westurner•1w ago
Task: play tetris

Task: write and optimize a tetris bot

Task: write and safely online optimize a tetris bot with consideration for cost to converge

openai/baselines (7 years ago) was leading on RL and then AlphaZero and Self-Attention Transformer networks.

LLMs are trained with RL, but aren't general purpose game theoretic RL agents?

westurner•1w ago
"Optimizing Tetris Gameplay Using Reinforcement Learning Framework with Adaptive Genetic Algorithms" (2025) https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5906702 .. https://scholar.google.com/scholar?cites=1615762352187216859...

"Outsmarting algorithms: A comparative battle between Reinforcement Learning and heuristics in Atari Tetris" (2025) https://dl.acm.org/doi/10.1016/j.eswa.2025.127251

burkaman•1w ago
It's actually 80% against Opus, 66% average against the 5 models it's tested with.
esafak•1w ago
I imagine this is because Tetris is visual and the Gemini models are strong visually.
bogtog•1w ago
I figure OP would try and give the models pure text forms of the game?

.....

l....

l....

l.ttt

l..t.

vunderba•1w ago
Interesting but frustratingly vague on details. How exactly are the models playing? Is it using some kind of PGN equivalent in Tetris that represents a on-going game, passing an ASCII representation, encoding as a JSON structure, or just directly sending screenshots of the game to the various LLMs?
storystarling•1w ago
It has to be turn-based. Even with Flash's speed, the inference latency would kill you in a real-time loop. They're likely pausing the game state after every tick to wait for the API response before resuming.
ykhli•1w ago
answered this in a comment above! It's not turn or visual layout based since LLMs are not trained that way. The representation is a JSON structure, but LLMs plug in algorithms and keeps optimizing it as the game state evolves
vunderba•1w ago
Thanks for the clarification! Kind of reminds me of the Brian Moore's AI clocks which uses several LLMs to continuously generate HTML/CSS to create an analog clock for comparisons.

https://clocks.brianmoore.com

ykhli•1w ago
Wow this is incredible!!
mhh__•1w ago
I suppose you could argue about whether it's an LLM at that point but vision is a huge part of frontier models now, no?
storystarling•1w ago
Curious how the token economics compare here to a standard agent loop. It seems like if you're using the LLM as a JIT to optimize the algorithm as the game evolves, the context accumulation would get expensive fast even with Flash pricing.
tiahura•1w ago
I'd like to see a nethackbench.
ykhli•1w ago
Thanks for all the questions! More details on how this works:

- Each model starts with an initial optimization function for evaluating Tetris moves.

- As the game progresses, the model sees the current board state and updates its algorithm—adapting its strategy based on how the game is evolving.

- The model continuously refines its optimizer. It decides when it needs to re-evaluate and when it should implement the next optimization function

- The model generates updated code, executes it to score all placements, and picks the best move.

- The reason I reframed this problem to a coding problem is Tetris is an optimization game in nature. At first I did try asking LLMs where to place each piece at every turn but models are just terrible at visual reasoning. What LLMs great at though is coding.

dakom•1w ago
How does it deal with latency? Afaict remote LLMs need seconds to process, but Tetris can move much faster..
p0w3n3d•1w ago
Guys, I don't know how to tell you but... Tetris can web solved without LLM...
brookman64k•1w ago
Why run something in a few CPU cyles on a 40 year old home computer when you can do the same (but worse) on a billion-dollar GPU cluster?
p0w3n3d•1w ago
Maybe this is about parsing the video? But still this should be done with OpenCV and then algorithmically...

A youtuber sentdex has a whole series on parsing the game's image and playing GTA https://www.youtube.com/playlist?list=PLQVvvaa0QuDeETZEOy4Vd...

gpm•1w ago
The world record for hatetris (tetris with deterministic somewhat adversarial piece selection instead of random piece selection) is still finite. The game is not solved. https://qntm.org/records
purplecats•1w ago
watch link?
segmondy•1w ago
... and what does this prove? what can you decide to use one LLM to solve over another based on this tetrisbench besides play tetris?
indigodaddy•1w ago
Is there a tl;dr on why this is? Does it just make faster decisions?
ykhli•1w ago
my unvalidated theory is that this comes down to the coding model’s training objective: Tetris is fundamentally an optimization problem with delayed rewards. Some models seem to aggressively over-optimize toward near term wins (clearing lines quickly), which looks good early but leads to brittle states and catastrophic failures later. Others appear to learn more stable heuristics like board smoothness, height control, long-term survivability even if that sacrifices short-term score

That difference in objective bias shows up very clearly in Tetris, but is much harder to notice in typical coding benchmarks. Just a theory though based on reviewing game results and logs

bityard•1w ago
Looks fun, but I'm not willing to give out my email address just to play a game.

Also, if the creator is reading this, you should know that Tetris Holdings is extremely aggressive with their trademark enforcement.

augusteo•1w ago
LLMs playing Tetris feels like testing a calculator's ability to write poetry. Interesting as a curiosity, but the results don't transfer to the tasks where these models actually excel.

Curious what the latency looks like per move. That seems like the actual bottleneck here.