frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
612•klaussilveira•12h ago•180 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
915•xnx•17h ago•545 comments

What Is Ruliology?

https://writings.stephenwolfram.com/2026/01/what-is-ruliology/
29•helloplanets•4d ago•22 comments

How we made geo joins 400× faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
102•matheusalmeida•1d ago•24 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
36•videotopia•4d ago•1 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
212•isitcontent•12h ago•25 comments

Jeffrey Snover: "Welcome to the Room"

https://www.jsnover.com/blog/2026/02/01/welcome-to-the-room/
5•kaonwarb•3d ago•1 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
206•dmpetrov•12h ago•101 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
316•vecti•14h ago•140 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
355•aktau•18h ago•181 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
361•ostacke•18h ago•94 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
471•todsacerdoti•20h ago•232 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
267•eljojo•15h ago•157 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
400•lstoll•18h ago•271 comments

Delimited Continuations vs. Lwt for Threads

https://mirageos.org/blog/delimcc-vs-lwt
25•romes•4d ago•3 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
82•quibono•4d ago•20 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
54•kmm•4d ago•3 comments

Was Benoit Mandelbrot a hedgehog or a fox?

https://arxiv.org/abs/2602.01122
9•bikenaga•3d ago•2 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
242•i5heu•15h ago•183 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
51•gfortaine•10h ago•16 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
138•vmatsiiako•17h ago•60 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
275•surprisetalk•3d ago•37 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
68•phreda4•11h ago•13 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
1052•cdrnsf•21h ago•433 comments

Why I Joined OpenAI

https://www.brendangregg.com/blog/2026-02-07/why-i-joined-openai.html
127•SerCe•8h ago•111 comments

Female Asian Elephant Calf Born at the Smithsonian National Zoo

https://www.si.edu/newsdesk/releases/female-asian-elephant-calf-born-smithsonians-national-zoo-an...
28•gmays•7h ago•10 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
173•limoce•3d ago•93 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
7•jesperordrup•2h ago•4 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
61•rescrv•20h ago•22 comments

Zlob.h 100% POSIX and glibc compatible globbing lib that is faste and better

https://github.com/dmtrKovalenko/zlob
17•neogoose•4h ago•9 comments
Open in hackernews

Spiral

https://spiraldb.com/post/announcing-spiral
261•jorangreef•4mo ago

Comments

all2•4mo ago
Spelling error "sttill"

> P.S. If you're sttill managing data in spreadsheets, this post isn't for you. Yet.

---

Since I discovered the ECS pattern, I've been curious about backing it with a database. One of the big issues seems to be IO on the database side. I wonder if Spiral might solve this issue.

lordnacho•4mo ago
If the ECS data is grid-like, perhaps you could use a columnar database for time series?

Then you could save every single state change and scroll back and forth. But I'm not sure if you were looking for that.

all2•4mo ago
This would make netcode state rollback easy. But I'm not sure the DB is the right place for that.
harwoodr•4mo ago
Have a look at something like spacetimeDB - caveat, I've only read about it and not directly used it:

https://github.com/ClockworkLabs/SpacetimeDB

all2•4mo ago
This was part of my inspiration. They do some interesting things with their systems, like hot reloads, that makes the dev loop a lot tighter.
SomeHacker44•4mo ago
"100KiB images"... This is odd. Most of my images are 2.5-4 MB. My raw images are 3-10x larger.
turnsout•4mo ago
I bet this refers to some common training use case that leverages 512px or 1024px images. Or it’s just Palantir scanning security camera frames.
pauldix•4mo ago
I've been following this team's work for a while and what they're doing is super interesting. The file format they created and put into the LF, Vortex, is very welcome innovation in the space: https://github.com/vortex-data/vortex

I'm excited to start doing some experimentation with Vortex to see how it can improve our products.

Great stuff, congrats to Will and team!

dist-epoch•4mo ago
https://vortex.dev doesn't work in my Firefox:

Application error: a client-side exception has occurred while loading vortex.dev (see the browser console for more information).

Console: unable to create webgl context

arusahni•4mo ago
Works for me. Mozilla/5.0 (X11; Linux x86_64; rv:142.0) Gecko/20100101 Firefox/142.0
miloignis•4mo ago
Presumably you don't have WebGL enabled or supported - the main page is just a cute 3D landing page.

You may be interested in https://github.com/vortex-data/vortex which of course has an overview and links to their docs and benchmark pages.

brunohaid•4mo ago
If anyone ever writes a post of why that error keeps happening with browsers that should support it, I'd be incredibly grateful. Keep seeing it in our (unrelated to OP company) Sentry logs and zero chance to reproduce them.
bflesch•4mo ago
I assume it's just people who do not have a graphics card
shakna•4mo ago
Handful of causes:

+ No hardware acceleration enabled.

+ Multiple graphics cards, and browser can't decide which to use.

+ Race conditions that can rarely cause a mount of 3d onto a 2d context (often happens to Unity).

dist-epoch•4mo ago
Privacy plugins which disable WebGL (fingerprinting)
reactordev•4mo ago
Anyone that can improve upon the parquet hell that is my life is gladly welcomed...
riku_iki•4mo ago
why you don't like parquet?
indoordin0saur•4mo ago
Parquet seems easy and straight-forward. The only issue I see people having with it is if they aren't used to non-human-readable formats and have to use special tools to look at it (as opposed to something like CSV). In that case this new file format will absolutely be worse.
reactordev•4mo ago
Not my issue at all. My issue is someone dumping 4gb of data into a parquet file thinking it’s fine…
riku_iki•4mo ago
I operate xxxGB files. What do you think is wrong with this?..
indoordin0saur•4mo ago
Is there a better file format for that?
paxys•4mo ago
Wasn't "3.0" supposed to be crypto? Is it AI now? It's had to keep track.
ionwake•4mo ago
I think AI is 4.0

EDIT> Maybe its how some poeple call the 4th dimension time when there is infact a 4th spatial dimension. So I guess if this is the 3rd Data dimension like what is the 4th one?

jppope•4mo ago
I think some of the crypto companies tried to get cute and leapfrog 3.0 going straight to 4.0, so that would put us at either 5.0, 4.0, 3.1, 2.2, or 2.1 depending on how you feel about the crypto space, and which groups you were validating
bee_rider•4mo ago
No, Web 3.0 was the Semantic Web. Thankfully, the silly idea of having major-number versions for the entire internet died when that it happen. Now we can safely ignore anybody who tries to do it.
adfm•4mo ago
You’re conflating concepts. FWIW, Web3 is snake oil or wishful thinking at best. As much as people like to bang on the old Web 2.0, it still holds up conceptually. And if you only know it as a buzz word, I suggest you go back and familiarize yourself with it if you’re looking for incremental change.

Who knows, maybe a Web 3.1 will deliver us from Enshitification.

stronglikedan•4mo ago
I think we're in a new era, so I consider this version of the web to be "AAI 1", and next year it will be "AAI 2", and so on. This era will be hereafter referred to as "in the year of the AI overlord", or "Anno Domini Artificialis Intellegentiae Artificialis" (according to google translate).
holoduke•4mo ago
So basically this is a file system that runs on your gpu?
djfobbz•4mo ago
So this Vortex engine is a combination of OLTP and OLAP on steroids?
maxmcd•4mo ago
Do they mention transactions anywhere? Maybe it will be OLAP?
didibus•4mo ago
It sounded only OLAP from the article.
cryptonector•4mo ago
I can't tell what this is about.
dkdbejwi383•4mo ago
Do you remember the days of “mongodb is web-scale”? It’s that but “spiral is ai-scale”
nwhnwh•4mo ago
So it will be irrelevant after a few years?
zzzeek•4mo ago
maybe just a few months, AI scale is much faster than web scale of course
steve_adams_86•4mo ago
Mongo is still very relevant

For better or worse

znort_•4mo ago
"I've been building data systems for long enough to be skeptical of “revolutionary” claims, and I’m uncomfortable with grandiose statements like “Built for the AI Era”. Nevertheless, ...

... i'm gonna make revolutionary claims and grandiose statements like "built for the ai era".

bee_rider•4mo ago
Probably either overcoming giant robots with the power of friendship and a giant drill, or a cursed village with an obsession-inducing whirlpool.
riku_iki•4mo ago
my reading that it will be some hyper-performant db thanks to some very low level optimization utilizing recent hw advancements and formats/pipelines unification and simplification.
didibus•4mo ago
I think I understood it as the database will basically store data in a binary format that can be fed into the GPU directly, and will also be optimized for streaming/batching large chunks of data at ounce.

So it's "optimized for machines to consume" meaning the GPU.

Their use case was training ML models where you need to feed the GPU massive datasets as part of training.

They seem to claim that training is now bottlenecked by how quickly you can feed the GPU, that otherwise the GPU is basically "waiting on IO" most of the time and not actual computing because the time goes in just grabbing the next piece of data, transforming it for GPU consumption, and then feeding it into the GPU.

But I'm not an expert, this is just my take from the article.

4ndrewl•4mo ago
The three eras of database systems starts with a client-server Postgres, but missed the daddy of the generation before that - xBase (ie dBase, FoxPro etc).
khaledh•4mo ago
It goes way before that. It starts with IDS (Integrated Data Store) from GE (1964), which was a network database system. Next was IBM's hierarchical database system IMS (Information Management System, 1966), still in use today. Then the CODASYL model (late 1960s), which was an effort to standardize the network model. And then Codd came up with the relational model in the early 70s, upon which an explosion of database systems were built (first is IBM System R, SQL, Oracle, DB2, Ingres). Then came the PC-based database systems you mentioned.
4ndrewl•4mo ago
Oh for sure. To suggest we're only on generation 3 of "databases" is way off the mark.
spankalee•4mo ago
I'm curious... I'm not a database or AI engineer. The last time I did GPU work was over a decade ago. What is the point of the "saturate an H100" metric?

I would think that a GPU isn't just sitting there waiting on a process that's in turn waiting for one query to finish to start the next query, but that a bunch of parallel queries and scans would be running, fed from many DB and object store servers, keeping the GPUs as utilized as possible. Given how expensive GPUs are, it would seem like a good trade to buy more servers to keep them fed, even if you do want to make the servers and DB/object store reads faster.

vouwfietsman•4mo ago
My guess is that just the raw data size, combined with the physical limitations of your RU, makes it hard for the GPU to be fully utilized. Instead you will always be stuck on CPU (decompressing/interpreting/uploading parquet) or bandwidth (transfer from s3) being the bottleneck.

Seems that they are targeting a low-to-no overhead path from s3 bucket to GPU, by targeting: same compression/faster random access, streamed encoding from S3 while in flight, zero copy to GPU.

Not 100% clear on the details, but I doubt that they can actually saturate the cpu/gpu bus, but rather just saturate the GPU utilization, which is itself dependent on multiple possible bottlenecks but generally not on bus bandwidth.

That's not criticism: it literally means you can't do better unless you improve the GPU utilization of your AI model.

otterley•4mo ago
The idea is that in a pipeline of work, throughput is limited by the slowest component. H100 GPUs have a lot of memory bandwidth. The question then becomes how to eliminate any bottlenecks between the data store and the GPU's memory.

First is the storage bottleneck. Network-attached storage is usually a bottleneck for uncached data. Then there is CPU work decoding data. Spiral claims that their table format is ready to load by the GPU so they can bypass various CPU-bound decoding stages. Once you eliminate storage and CPU bottlenecks, the remaining bottleneck is usually the PCI bus that sits between the host memory and the GPU, and they can't solve that themselves. (And no amount of parallelization can help when the bus is saturated.) What they can do is use the network, the host bus, and the GPU more efficiently by compressing and packing data with greater mechanical sympathy.

They've left unanswered how they're going to commercialize it, but my guess is that they're going to use a proprietary fork of Vortex that provides extra performance or features, or perhaps they'll offer commercial services or integrations that make it easier to use. The open-source release gives its customers a Reason to Believe, in marketing parlance.

zzzeek•4mo ago
This links to a super long winded blog post that sounds more like a toast at a wedding, so I went to the main page to try to see what their product is, and you just get a blitz of fancy animations of table diagrams and things and lots of very cheap sounding slogans pushed out like "Works with any data! Fully XYZ 2.0 compliant! Ties your shoes!"

basically im not sure where the product is hiding under all of this bluster but this doesnt feel very "hacker"-Y

vouwfietsman•4mo ago
Although I welcome a parquet successor, I am not particularly interested in a more complicated format. Random access time improvements are nice, but really what I would like just storing multiple tables in a single parquet file.

When I read "possible extension through embedded wasm encoders" I can already imagine the c++ linker hell required to get this thing included in my project.

I also don't think a lot of people need "ai scale".

alfalfasprout•4mo ago
also what does "ai scale" even mean?
aakkaakk•4mo ago
It’s obvious a jab at mongo’s ”web scale”. https://youtube.com/watch?v=b2F-DItXtZs
vouwfietsman•4mo ago
I think its a bit markety, but they explain it rather well: because of AI your data needs to be consumed by machines on an unprecedented scale, which requires new solutions to problems. Historically we mostly did large input -> small output, now we're doing large input -> large output. The existing tools are (supposedly) not ready.
alfalfasprout•4mo ago
no, I read that. It doesn't really add any more practical detail.
drdaeman•4mo ago
Storing multiple tables in a single file would be trivially solvable by storing multiple Parquet files in a most basic plain uncompressed tarball (to retain ability to access any part of any file without downloading the whole thing). Or maybe ar or cpio - tar has too many features (such as support for links) that are unnecessary here. Basically, anything well-standardized that implements a very basic directory structure, with a simple index located at a predictable offset.

If any tools would've supported that.

vouwfietsman•4mo ago
Couldn't agree more. If tooling would just settle on an arbitrary archive format our lives would be better.
nylonstrung•4mo ago
Lance already exists to solve Parquet problems but with drastically faster random access time
vouwfietsman•4mo ago
Lance is pretty far from a lingua franca. For instance the SDKs are only Rust/Python/Java, none of which I use.
nylonstrung•4mo ago
Sounds like we need more SDKs, not a new format
gcr•4mo ago
If you want "several tables and database-like semantics in one file," then what you want is DuckDB.

If you want modern parquet, then you want the Lance format (or LanceDB for DB-like CRUD semantics).

mellosouls•4mo ago
This is a pretty website but doesn't actually give us anything to actually look at, its just blurb.

For anybody confused, the "Vortex" stuff is the underlying data format used but isn't the database/whatever this website (by the creators of Vortex) is pushing.

kmoser•4mo ago
> Spiral is our database built on Vortex [...]

No surprise there's nothing to look at, since it's basically a press release posted on their blog.

mlhpdx•4mo ago
I stopped reading at “new era”. At this point in time with the deluge of content, start with a problem and solution in a concise statement if you want my attention. I’m not reading your opinion piece.
bflesch•4mo ago
Big ick from my side. Manifest-style marketing blog post talking about revolutionary things but it seems their main metric is in the image above the post: "hey, we've raised $22M in funding".

Landing pages of both spiral and vortex are GPU-hugging animations and void of any technical information. Empty nothing-statements like "machine scale". They claim 100x improvements but don't link any metrics.

Maybe this is a "don't hate the player, hate the game" situation but somehow the collective of likeminded AI engineers decided to upvote this post to #1 on HN.

indoordin0saur•4mo ago
> Vortex is designed to support decoding data directly from S3 to GPU, skipping the CPU bottleneck entirely.

If this is true I'm inclined to believe their claims.

bflesch•4mo ago
MY PERSONAL BOTTLENECK between S3 and GPU is my credit card and not some new cargo module by some already-rich AI engineer and a fancy marketing website that must've cost a couple hundred grand.

And if this module provides a benefit I'm sure it will find its way into our stack, just like PostgreSQL did. And PostgreSQL never had $22M to begin with - no shiny marketing, just technological skills.

The whole "donated by spiral" on the vortex.dev website also gives big tax write-off vibes.

IMO best case is that this will be a mongodb scenario, but with the current track record of tech grifters enshittifying everything they might find a creative new way.

xyzzy_plugh•4mo ago
> The whole "donated by spiral" on the vortex.dev website also gives big tax write-off vibes.

I've never heard of this sort of OSS work being used as a tax write-off. Could someone please either clarify, or enlighten me?

bflesch•4mo ago
Look at the website linked in the vortex website footer: https://lfprojects.org/ It has all the bells and whistles of using an expensive law firm.

I have no idea who exactly is behind this, but to me it does definitely not seem like a no-name open source genius, I assume it is some lucky AI grifter. They have two nicely designed, expensive marketing websites. They have all the legal documents for the parent LLC in Delaware.

The delaware corp "donates" the multi-million-worth tech to linux foundation, and uses it as tax write-off to offset gains from some other lucky AI grifter play the person did.

Just the chuzpe to self-compare yourself to something like PostgreSQL is what gets me. Why can't they just be rich and leave people doing actual work for the benefit of our common good be. No, they must make big blog posts claiming they are the next big thing after PostgreSQL.

So many red flags..

aduffy•4mo ago
Believe it or not, this is how the Linux Foundation organizes itself. It's more legwork than something simpler like Apache Foundation.

Basically in the US you need a legally recognized entity to hold intellectual property. "Donating" the project involves setting up a "Series LLC" that is nested underneath the top-level Linux Foundation corporation, and donating the IP into it.

Checkout https://docs.linuxfoundation.org/lfx/project-control-center/... and ctrl-f "LF Projects, LLC"

bflesch•4mo ago
Oh, thanks for pointing that out. I got it all mixed up.

But I think my argument still stands. Linux foundation is a 501(c)(6) nonprofit, see https://www.linuxfoundation.org/legal/bylaws

So you might still be able to do an "intellectual property transfer" to them and use it as a tax write-off. The "LF Projects LLC" is then the new owner, only the operating company who has the ongoing hosting contracts for the websites.

Edit: Not sure if a donation to 501(c)(6) can be used as write-off without using some other legal loopholes. Quick AI search told me that only 501(c)(3) can do the donation tax write-off thing.

I'm sure there are some good tax lawyers behind this, who am I to understand it as a mere mortal I am just jealous.

aduffy•4mo ago
We don't get a tax write-off.

The motivation is to move the IP and trademark into a separate organization so it's no longer owned by Spiral. This means we can't re-license it later, we'd have to fork it, because the Vortex trademark and all that is controlled by LF.

bflesch•4mo ago
Thanks for explaining it. All the best.
aduffy•4mo ago
> The whole "donated by spiral" on the vortex.dev website also gives big tax write-off vibes.

Donated is the Linux Foundation terminology.

Sadly the last time I filed a tax return there was no way to itemize a Github repo. Alas.

msteffen•4mo ago
There's this: https://bench.vortex.dev/, which links to https://github.com/vortex-data/vortex/tree/develop/bench-vor.... I haven't tried pulling the repo or anything but it seems like they might be runnable?

Of course I don't know what benchmarks or performance metrics they might have for the db layer, but it is something.

bflesch•4mo ago
Thanks, it was behind a "see benchmarks" link at the very bottom of https://spiraldb.com/. I noticed these claims on both vortex.dev and https://spiraldb.com/vortex website, but both had no hyperlink to any actual metrics.
raziel2p•4mo ago
> Vortex is designed to support decoding data directly from S3 to GPU, skipping the CPU bottleneck entirely.

how is this significant? surely either the network or the GPU calculations is the bottleneck here?

dwb•4mo ago
If you don't clearly detail what your new tech product or system is bad at, as well as what it's good at, I'm not interested. So much of engineering is about navigating the inevitable tradeoffs. Marketing should have no place in engineering.
derekhecksher•4mo ago
The AnyBlox paper from the folks at TUM, and linked to in the post, is a bit more interesting, imo, since it looks to solve the data systems x storage format problem in composable data architectures - https://gienieczko.com/anyblox-paper
aduffy•4mo ago
The AnyBlox paper has some very cool ideas, and the authors are friends.

In the paper you'll notice a large portion of it analyzes Vortex, both standalone and embedded. Definitely worth a read.

whalesalad•4mo ago
The hot new aesthetic these days is either "receipt printer" or "liquid glass". I dig it, tbh.
skywhopper•4mo ago
Man, they are really proud of that initial seed round funding aren’t they? Forgive me, but $22 million does not sound like enough to truly revolutionize data processing technology.

The gist seems to be that they can overcome network latency issues when dealing with huge numbers of smallish objects in S3-like storage systems that need to be fed into GPUs? Yeah, those formats and systems were not designed to feed that type of processor. You’re doing it wrong if this is your problem.

After a lot of nonsense, it sounds like they just reformat the data into something more efficient instead. But they forget about the network latency and blame CPUs for slowing things down? And what was that sidetrack about S3 permissions?

I wouldn’t jump right onto this… well, it’s not clear what this even is exactly. But you can probably wait it out.

bflesch•4mo ago
Even though Linux Foundation is only a 501(c)(6) nonprofit and not a 501(c)(3) nonprofit I wonder if all this legalese about "donation" might have very positive tax consequences for the donor.
datadrivenangel•4mo ago
$22 million gets you a potentially commercially viable narrow featured database and some time to get enough sales to validate the VCs throwing another $50m at you to flush it out and build out the sales arm seriously.
redwood•4mo ago
So it's for low change rate data that needs to be bulk processed during ML model training. Cool. But hardly the same thing as what you need for powering live AI applications... which is what I assumed this was upon reading the intro and the mention of Postgres..

Postgres (and MongoDB) are the king and prince of data due to their transactional capabilities.

rubenvanwyk•4mo ago
Interesting that Joran from Tigerbeetle posted this? So must be legit.
jorangreef•4mo ago
I’ve known Will for some time while they’ve been in stealth.
rubenvanwyk•4mo ago
How does Vortex compare to Lance? I imagine Lance is already a good solution for AI on CPUs.
donperignon•4mo ago
“ We work in person at our offices in London and New York. Face to face is better: if uncertain, the answer is “yes, get on the plane”. On Wednesdays, we wear pink.”

No comments.

sys13•4mo ago
I wonder how much we need this vs implementing it as part of Delta Lake or Iceberg
mritchie712•4mo ago
yeah, I think you could get much of the benefits if Iceberg (or ducklake) added support for `.vortex` (e.g. ducklake only supports `.parquet` right now)
dwroberts•4mo ago
> Remember that uncanny valley between 1KB and 25MB? The problem isn't the sizes—it's that Second Age systems force you to choose between two bad options: inline the data (killing performance) or store pointers (breaking governance). Spiral eliminates this false choice. We store 10KB embeddings directly in Vortex for microsecond access, intelligently batch 10MB blocks of images for optimal S3 throughput, and externalize 4GB videos without copying a single byte. One system, no compromises.

No compromises but isn’t ‘externalising’ a large video the equivalent of storing a pointer in the first example? Can’t really see any other way to understand what that means (it goes to an external system and you store where it is)