frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
475•klaussilveira•7h ago•116 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
813•xnx•12h ago•487 comments

How we made geo joins 400× faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
33•matheusalmeida•1d ago•1 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
157•isitcontent•7h ago•17 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
156•dmpetrov•7h ago•67 comments

A century of hair samples proves leaded gas ban worked

https://arstechnica.com/science/2026/02/a-century-of-hair-samples-proves-leaded-gas-ban-worked/
92•jnord•3d ago•12 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
50•quibono•4d ago•6 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
260•vecti•9h ago•123 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
207•eljojo•10h ago•134 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
328•aktau•13h ago•158 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
327•ostacke•13h ago•86 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
411•todsacerdoti•15h ago•219 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
23•kmm•4d ago•1 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
337•lstoll•13h ago•242 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
52•phreda4•6h ago•9 comments

Delimited Continuations vs. Lwt for Threads

https://mirageos.org/blog/delimcc-vs-lwt
4•romes•4d ago•0 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
195•i5heu•10h ago•145 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
115•vmatsiiako•12h ago•38 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
152•limoce•3d ago•79 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
245•surprisetalk•3d ago•32 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
996•cdrnsf•16h ago•420 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
26•gfortaine•5h ago•3 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
46•rescrv•15h ago•17 comments

I'm going to cure my girlfriend's brain tumor

https://andrewjrod.substack.com/p/im-going-to-cure-my-girlfriends-brain
67•ray__•3h ago•30 comments

Evaluating and mitigating the growing risk of LLM-discovered 0-days

https://red.anthropic.com/2026/zero-days/
38•lebovic•1d ago•11 comments

Show HN: Smooth CLI – Token-efficient browser for AI agents

https://docs.smooth.sh/cli/overview
78•antves•1d ago•59 comments

How virtual textures work

https://www.shlom.dev/articles/how-virtual-textures-really-work/
30•betamark•14h ago•28 comments

Show HN: Slack CLI for Agents

https://github.com/stablyai/agent-slack
41•nwparker•1d ago•11 comments

Female Asian Elephant Calf Born at the Smithsonian National Zoo

https://www.si.edu/newsdesk/releases/female-asian-elephant-calf-born-smithsonians-national-zoo-an...
7•gmays•2h ago•2 comments

Evolution of car door handles over the decades

https://newatlas.com/automotive/evolution-car-door-handle/
41•andsoitis•3d ago•62 comments
Open in hackernews

There's no single best way to store information

https://www.quantamagazine.org/why-theres-no-single-best-way-to-store-information-20260116/
100•7777777phil•2w ago

Comments

pbreit•2w ago
Postgres is close.
imhoguy•2w ago
I would say Sqlite is closer, you find it on every phone, browser, server. I bet Sqlite files will be still readable in 2100. And I love Postgres.
rmwaite•2w ago
Relevant: https://sqlite.org/mostdeployed.html
mjevans•2w ago
Or (real) SQLite for reasonably scaled work.

I also like (old) .ini / TOML for small (bootstrap) config files / data exchange blobs a human might touch.

+

Re: PostgreSQL 'unfit' conversations.

I'd like some clearer examples of the desired transactions which don't fit well. After thinking about them in the background a bit I've started to suspect it might be an algorithmic / approach issue obscured by storage patterns that happen to be enabled by some other platforms which work 'at scale' supported by hardware (to a given point).

As an example of a pattern that might not perform well under PostgreSQL, something like lock-heavy multiple updates for flushing a transaction atomically. E.G. Bank Transaction Clearance like tasks. If every single double-entry booking requires it's own atomic transaction that clearly won't scale well in an ACID system. Rather the smaller grains of sand should be combined into a sandstone block / window of transactions which are processed at the same time and applied during the same overall update. The most obvious approach to this would be to switch from a no-intermediate values 'apply deduction and increment atomically' action to a versioned view of the global data state PLUS a 'pending transactions to apply' log / table (either/both can be sharded). At a given moment the transactions can be reconciled, for performance a cache for 'dirty' accounts can store the non-contested value of available balance.

__MatrixMan__•2w ago
There are, however, several objectively bad ways. In "Service Model" (a novel that I recommend) a certain collection of fools decides to sort bits by whether it's a 1 or a 0, ending up with a long list of 0's followed by a long list of 1's.
Rygian•2w ago
In a similar vein, someone decided that everyone should have subdirectories under home named "Pictures", "Videos", "Music", "Documents", …
lo_zamoyski•2w ago
That depends on the aim. The purpose of something determines how fitting the means are.

Also, let us not confuse "relative" with "not objective". My father is objectively my father, but he is objectively not your father.

__MatrixMan__•2w ago
I'm frequently bothered by misuse of "objective" but I stand by it here. In the case of storage, one criterion supersedes all others: can you get the information back out? If you can then there are merits to discuss relative to use case. If you can't then your storage mechanism is broken.
dsvf•2w ago
It _does_ open up amazing opportunities for compression though.
HPsquared•2w ago
That's fine so long as there's an index!
__MatrixMan__•2w ago
Presumably there was at some time, but they put it in long term bit storage also.
teraflop•2w ago
There's a similar anecdote in Iain M. Banks' The Player of Games.

https://scifi.stackexchange.com/questions/270578/negotiator-...

__MatrixMan__•2w ago
Thanks for pointing it out.

I'm putting together a database along the lines of https://www.whosampled.com/ except instead of samples in music it's fiction authors either:

- revisiting one another's ideas

- stumbling across the same idea independently of one other

Now I get to add an edge between Tchaikovsky and Banks. It's somewhat unsurprising. Tchaikovsky does this a lot.

hcs•2w ago
Related seriously: You can make it more likely to have long runs of the same letters, reversibly, with the Burrows-Wheeler Transform: https://en.wikipedia.org/wiki/Burrows%E2%80%93Wheeler_transf...

(No magic, though it still feels like it, BWT is most useful when there's repeated substrings.)

kittikitti•2w ago
Or it's the opposite, where the slowest possible retrieval time is the intended effect, as is the basis of many cryptographic algorithms.
adrianmonk•2w ago
Or it's neither, and the intended effect is zero variation in the retrieval time, as when trying to avoid leaking secrets via timing attacks.

(Or I guess, more generally, the intended effect is zero correlation between the information and the time it takes to retrieve it. If retrieval time were completely random, it would achieve the goal, but it wouldn't have zero variation.)

bob1029•2w ago
The best way to store information depends on how you intend to use (query) it.

The query itself represents information. If you can anticipate 100% of the ways in which you intend to query the information (no surprises), I'd argue there might be an ideal way to store it.

alphazard•2w ago
This is exactly right, and the article is clickbait junk.

Given the domain name, I was expecting something about the physics of information storage, and some interesting law of nature. Instead, the article is a bad introduction to data structures.

megaBiteToEat•2w ago
You both are affirming the title of the article.

"No single best way", meaning "it depends."

But don't let something like literacy get in the way of a opportunity to engage in meaningless outrage.

DixieDev•2w ago
This line of thought works for storage in isolation, but does not hold up if write speed is a concern.
convolvatron•2w ago
as a line of thought, it totally does. you just extend the workload description to include writes. where this get problematic is that the ideal structure for transactional writes is nearly pessimal from a read standpoint. which is why we seem to end up doubling the write overhead - once to remember and once to optimize. or highly write-centric approach like LSM

I'd love to be clued in on more interesting architectures that either attempt to optimize both or provide a more continuous tuning knob between them

cannonpalms•2w ago
So long as (fast/optimal) real-time access to new data is not a concern, you can introduce compaction to solve both problems.
bob1029•2w ago
> (fast/optimal) real-time access to new data

https://en.wikipedia.org/wiki/Optimal_binary_search_tree#Dyn...

sandworm101•2w ago
Speed can always be improved. If a method is too slow, run multiple machines in parralel. Longevity is different as it cannot scale. A million cd burners are together very fast, but the CDs wont last any longer. So the storage method is is the more profound tech problem.
roenxi•2w ago
Yes with the important caveat that a lot of the time people don't have a crystal ball, can't see the far future, don't know if their intents will materialise in practice 12 months down the line and should therefore store information in Postures until that isn't a feasible option any more.

A consequence of there being no generally superior storage mechanism is that technologists as a community should have an agreed default standard for storage - which happens to be relational.

jandrewrogers•2w ago
This is connected to the equivalence relationship between optimal indexing and optimal AGI. The "best" way is optimal for the entire universe of possible queries but has the downside of being profoundly computationally intractable.

Requiring perfect knowledge of how information will be used is brittle. It has the major benefit of making the algorithm design problem tractable, which is why we do it.

An alternative approach is to exclude large subsets of queries from the universe of answerable queries without enumerating the queries that the system can answer. The goal is to qualitatively reduce the computational intractability of the universal case by pruning it without over-specifying the queries it can answer such as in the traditional indexing case. This is approximately what "learned indexing" attempts to do.

ssivark•2w ago
What if the various potential queries demand different / conflicting compression schemes?

I'd say this is spiritually what the no-free-lunch theorems are about... Because whatever "AI model" / query system you build -- it is implicitly biased towards queries coming from one slice of futures.

andix•2w ago
It's always Markdown. Markdown is the best way to store information. ;)
_ea1k•2w ago
Claude Code vehemently agrees.
andix•2w ago
You're absolutely right!
eimrine•2w ago
Which implementation of Markdown is a correct Markdown? Why not org-mode syntax?
danans•2w ago
Pedantic, but the article is talking about the way we structure/organize information, not store it. When I think of the word store, I think of the physical medium. The way we organize the information is only partially related
pixl97•2w ago
I mean if we're talking about the physical storage of medium, the single most dense way would be to write it on the surface of a black hole. I still haven't figured out how to read it back though.
1970-01-01•2w ago
It's not pedantic, you are correctly using words as we understand them, and they are not. The headline needs a sharp correction. Editing jobs are in very short supply these days.
megaBiteToEat•2w ago
Oh come on. Programmers discuss how to "store" data in memory as a data model all the time.

You're reducing definitions and meaning too far to make an ultimately empty point just to contribute the thread.

If social medias only contribution is language policing, then it really should die off. What a waste of resources so functional illiterate nobodies can project ego.

1970-01-01•2w ago
No, I'll think I'll double down, because I do think I'm right here.

https://en.wikipedia.org/wiki/Data_storage is a different website from https://en.wikipedia.org/wiki/Data_store because they are different, slightly overlapping concepts.

ronsor•2w ago
There are plenty of good enough ways:

* For lossless compression of generic data, gzip or zstd.

* For text, documentation, and information without fancy formatting, markdown, which is effectively a plain-text superset.

* For small datasets, blobs, objects, and what not, JSON.

* For larger datasets and durable storage, SQLite3.

Whenever there's text involved, use UTF-8. Whenever there's dates, use ISO8601 format (UTC timezone) or Unix timestamps.

Following these rules will keep you happy 80% of the time.

nicbou•2w ago
One format I'm missing: storage for conversations and social media posts. Both are complex media (text + images/videos + metadata), and one is actually a collection of such posts.

How would you go about storing those in a somewhat human-readable format? My goal is to archive my chats and social media activity.

soulofmischief•2w ago
Why not just use WARC and a program that can read them? Do archives need to be human-readable?
ronsor•2w ago
The thing about archives is you either parse them now or parse them later. With how much JS and other crap is served in modern social media frontends, I'm not sure WARC is the best format for archiving from them.
ElectricalUnion•2w ago
But that is the point of WARC: otherwise, your archival method need some sort of general inteligence (ai or human behind the scenes) to store exacly what you need.

With WARC (and good WARC tooling like Browsetrix-crawler) you store everything HTTP the site sent.

ronsor•2w ago
Use a SQLite3 database. Have a table for the posts (or any other appropriate schema, depending on what metadata you have). Using SQLite3 has the advantage of future flexibility (new/different tables and schema as needed, full-text search, etc.).

You can have another table for attachments (images, videos, etc.). If they're small, store them directly in a BLOB. If they're not, store them alongside the database, and only store the relative path in the attachments table.

You may opt to convert images and videos to a single format (e.g. PNG and H.264 MP4), but you can lose information depending on the target format. It may be preferable to leave them in the original (or highest quality) format.

ifh-hn•2w ago
Depends on what you mean by humans readable. Sqlite, as the other sub comment mentions is good, but you could also just use a CSV file, unnormalised table, and sit the original media in the same or a sub folder. Hell, convert that CSV to a html table and you can display the data as a human readable local webpage. Through in some JS and you can navigate/filter it too.
notepad0x90•2w ago
would it be more accurate to say "to store using information, using information"? Since everything ultimately boils down to information, humans trying to store information is a bit recursive?
akhil08agrawal•2w ago
This clicked for me in a way I didn't expect.

I've been thinking about trade-offs as "pick two of three" in the abstract, but the bookshelf example made it concrete. The insight that matters is: if you know your query patterns, you can optimize differently.

As a PM, I keep trying to build systems that work for "every case." But this article reminded me that's the wrong goal. The hash table works because it accepts the space-time trade-off. The heap works because it embraces disorder for non-priority items.

Sometimes the best system isn't the most elegant one—it's the one that matches how you'll actually use it.

Good reminder to stop over-optimizing for flexibility I'll never need.

Thanks for sharing.

altmanaltman•2w ago
You're a PM and this basic-level watered down article barely discussing anything "clicked for you in a way" you didn't expect? Of course the best system is desinged based on requirements, how can a PM not know this before being a PM?
eliasdejong•2w ago
See also, RUM Conjecture: https://www.codementor.io/@arpitbhayani/the-rum-conjecture-1...

Conceptually similar to CAP, but with storage trade-offs. The idea is you can only pick 2 out of 3.

1970-01-01•2w ago
Oh I know this one. False. Compress it first, then encrypt. :)
vee-kay•2w ago
Millions of years of evolution has resulted in the human brain being the best way to store information.

I doubt we humans will be able to do better (faster, more capacity, more analytical, more intuitive, more logical) storage (at an individual level, not at mass scale, since that's kinda achieved already by the behemoths like Google, etc.) in a few thousand years of civilization.

Quantum computing may be the game changer though.

I read somewhere that the entirety of humanity's information, including all knowledge and data of past (of every human ever) and current, if stored via quantum computing - that quanta of quantum information will just be the size of a football.