frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: LocalGPT – A local-first AI assistant in Rust with persistent memory

https://github.com/localgpt-app/localgpt
167•yi_wang•6h ago•58 comments

Haskell for all: Beyond agentic coding

https://haskellforall.com/2026/02/beyond-agentic-coding
82•RebelPotato•5h ago•21 comments

SectorC: A C Compiler in 512 bytes (2023)

https://xorvoid.com/sectorc.html
273•valyala•14h ago•52 comments

OpenClaw Is Changing My Life

https://reorx.com/blog/openclaw-is-changing-my-life/
16•novoreorx•1h ago•27 comments

Software factories and the agentic moment

https://factory.strongdm.ai/
213•mellosouls•16h ago•360 comments

LLMs as the new high level language

https://federicopereiro.com/llm-high/
82•swah•4d ago•149 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
172•surprisetalk•13h ago•172 comments

LineageOS 23.2

https://lineageos.org/Changelog-31/
20•pentagrama•2h ago•0 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
185•AlexeyBrin•19h ago•35 comments

Brookhaven Lab's RHIC concludes 25-year run with final collisions

https://www.hpcwire.com/off-the-wire/brookhaven-labs-rhic-concludes-25-year-run-with-final-collis...
76•gnufx•12h ago•60 comments

Stories from 25 Years of Software Development

https://susam.net/twenty-five-years-of-computing.html
178•vinhnx•16h ago•18 comments

The Architecture of Open Source Applications (Volume 1) Berkeley DB

https://aosabook.org/en/v1/bdb.html
12•grep_it•5d ago•0 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
337•jesperordrup•1d ago•102 comments

First Proof

https://arxiv.org/abs/2602.05192
139•samasblack•16h ago•81 comments

Show HN: I saw this cool navigation reveal, so I made a simple HTML+CSS version

https://github.com/Momciloo/fun-with-clip-path
89•momciloo•13h ago•18 comments

Substack confirms data breach affects users’ email addresses and phone numbers

https://techcrunch.com/2026/02/05/substack-confirms-data-breach-affecting-email-addresses-and-pho...
34•witnessme•3h ago•10 comments

The world heard JD Vance being booed at the Olympics. Except for viewers in USA

https://www.theguardian.com/sport/2026/feb/07/jd-vance-boos-winter-olympics
70•treetalker•39m ago•15 comments

Wood Gas Vehicles: Firewood in the Fuel Tank (2010)

https://solar.lowtechmagazine.com/2010/01/wood-gas-vehicles-firewood-in-the-fuel-tank/
39•Rygian•2d ago•13 comments

uLauncher

https://github.com/jrpie/launcher
11•dtj1123•4d ago•0 comments

Vouch

https://twitter.com/mitchellh/status/2020252149117313349
86•chwtutha•4h ago•23 comments

Al Lowe on model trains, funny deaths and working with Disney

https://spillhistorie.no/2026/02/06/interview-with-sierra-veteran-al-lowe/
109•thelok•15h ago•24 comments

Start all of your commands with a comma (2009)

https://rhodesmill.org/brandon/2009/commands-with-comma/
593•theblazehen•3d ago•216 comments

Show HN: A luma dependent chroma compression algorithm (image compression)

https://www.bitsnbites.eu/a-spatial-domain-variable-block-size-luma-dependent-chroma-compression-...
42•mbitsnbites•3d ago•6 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
318•1vuio0pswjnm7•20h ago•523 comments

FDA intends to take action against non-FDA-approved GLP-1 drugs

https://www.fda.gov/news-events/press-announcements/fda-intends-take-action-against-non-fda-appro...
117•randycupertino•9h ago•245 comments

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
909•klaussilveira•1d ago•277 comments

Where did all the starships go?

https://www.datawrapper.de/blog/science-fiction-decline
165•speckx•4d ago•247 comments

Selection rather than prediction

https://voratiq.com/blog/selection-rather-than-prediction/
37•languid-photic•4d ago•18 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
305•isitcontent•1d ago•39 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
149•videotopia•4d ago•49 comments
Open in hackernews

Apache Iceberg V3 Spec new features for more efficient and flexible data lakes

https://opensource.googleblog.com/2025/08/whats-new-in-iceberg-v3.html
87•talatuyarer•6mo ago

Comments

talatuyarer•6mo ago
This new version has some great new features, including deletion vectors for more efficient transactions and default column values to make schema evolution a breeze. The full article has all the details.
hodgesrm•6mo ago
This Google article was nice as a high level overview of Iceberg V3. I wish that the V3 spec (and Iceberg specs in general) were more readable. For now the best approach seems to be read the Javadoc for the Iceberg Java API. [0]

[0] https://javadoc.io/doc/org.apache.iceberg/iceberg-api/latest...

twoodfin•6mo ago
The Iceberg spec is a model of clarity and simplicity compared to the (constantly in flux via Databricks commits…) Delta protocol spec:

https://github.com/delta-io/delta/blob/master/PROTOCOL.md

eatonphil•6mo ago
To the contrary, the Delta Lake paper is extremely easy to read and implement the basics of (I did) and Iceberg has nothing so concise and clear.
twoodfin•6mo ago
If I implement what’s described in the Delta Lake paper, will I be able to query and update arbitrary Delta Lake tables as populated by Databricks in 2025?

(Would be genuinely excited if the answer is yes.)

eatonphil•6mo ago
Not sure (probably not). But it's definitely much easier to immediately understand IMO.
twoodfin•6mo ago
OK, but at least from my perspective, the point of OTF’s is to allow ongoing interoperability between query and update engines.

A “standard” getting semi-monthly updates via random Databricks-affiliated GitHub accounts doesn’t really fit that bill.

Look at something like this:

https://github.com/delta-io/delta/blob/master/PROTOCOL.md#wr...

Ouch.

ahmetburhan•6mo ago
Cool to see Iceberg getting these kinds of upgrades. Deletion vectors and default column values sound like real quality-of-life improvements, especially for big, messy datasets. Curious to hear if anyone’s tried V3 in production yet and what the performance looks like.
jamesblonde•6mo ago
Is it out yet?
amluto•6mo ago
> ALTER TABLE events ADD COLUMN version INT DEFAULT 1;

I’ve always disliked this approach. It conflates two things: the value to put in preexisting rows and the default going forward. I often want to add a column, backfill it, and not have a default.

Fortunately, the Iceberg spec at least got this right under the hood. There’s “initial-default”, which is the value implicitly inserted in rows that predate the addition of the column, and there’s “write-default”, which is the default for new rows.

drivenextfunc•6mo ago
Many companies seem to be using Apache Iceberg, but the ecosystem feels immature outside of Java. For instance, iceberg-rust doesn't even support HDFS. (Though admittedly, Iceberg's tendency to create many small files makes it a poor fit for HDFS anyway.)
hodgesrm•6mo ago
Seems like this is going to be a permanent issue, no? Library level storage APIs are complex and often quite leaky. That's based on looking at the innards of MySQL and ClickHouse for a while.

It seems quite possible that there will be maybe three libraries that can write to Iceberg (Java, Python, Rust, maybe Golang), while the rest at best will offer read access only. And those language choices will condition and be conditioned by the languages that developers use to write applications that manage Iceberg data.

ozgrakkurt•6mo ago
This was the same with arrow/parquet libraries as well. It takes a long time for all implementations to catch up
jamesblonde•6mo ago
When will open source v3 come out? It's supposed to be in Apache Iceberg 1.10, right?
talatuyarer•6mo ago
Yes 1.10 version will be first version for V3 spec. But not all features are implemented on runners such as Spark or Flink.
fabatka•6mo ago
I thought 1.9.0 already had at least some of the v3 features, like the variant type and column lineages? https://iceberg.apache.org/releases/#190-release

Of course I haven't seen any implementations supporting these yet.

talatuyarer•6mo ago
Yes, the specification will be finalized with version 1.10. Previous versions also include specification changes. Iceberg's implementation of V3 occurs in three stages: Specification Change, Core Implementation, and Spark/Flink Implementation.

So far only Variant is supported in Spark and with 1.10 Spark will support nano timestamp and unknowntype I believe.

jamesblonde•6mo ago
Any idea when 1.10 will be released?
talatuyarer•5mo ago
I believe we are very close to release candidate. We are waiting unknown type support for Apache Spark per latest email

https://lists.apache.org/thread/gd5smyln3v6k4b790t5d1vy4483m...

robertlagrant•6mo ago
> default column values

The way they implemented this seems really useful for any database.

nojito•6mo ago
It's a mismatch that this is on the official blog, but their implementation of Iceberg is still behind and doesn't have feature parity with the spec.

https://cloud.google.com/bigquery/docs/iceberg-tables#limita...

ahmetaltay•5mo ago
(Disclaimer: I work on the BigQuery team at Google, but my opinions are my own.)

You're right — our current implementation in BigLake doesn't have full feature parity with the V3 spec yet. We're actively working on it.

The key context is that the V3 spec is brand new, having been finalized only about two months ago. The official Apache Iceberg release that incorporates all these V3 features isn't even out yet. So, you'll find that the entire ecosystem, including major vendors, is in a similar position of implementing the new spec.

The purpose of our blog post was to celebrate this huge milestone for the open-source community and to share a technical deep-dive on why these new capabilities are so important.

sgarland•6mo ago
I read this [0] (I also recommend reading part 1 for background) a few weeks ago, and found it quite interesting.

The entire concept of data lakes seems odd to me, as a DBRE. If you want performant OLAP, then get an OLAP DB. If you want temporality, have a created_at column and filter. If the problem is that you need to ingest petabytes of data, fix your source: your OLTP schema probably sucks and is causing massive storage amplification.

[0]: https://database-doctor.com/posts/iceberg-is-wrong-2.html