frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

SectorC: A C Compiler in 512 bytes

https://xorvoid.com/sectorc.html
98•valyala•4h ago•16 comments

The F Word

http://muratbuffalo.blogspot.com/2026/02/friction.html
43•zdw•3d ago•11 comments

Brookhaven Lab's RHIC concludes 25-year run with final collisions

https://www.hpcwire.com/off-the-wire/brookhaven-labs-rhic-concludes-25-year-run-with-final-collis...
23•gnufx•2h ago•19 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
56•surprisetalk•3h ago•54 comments

Software factories and the agentic moment

https://factory.strongdm.ai/
98•mellosouls•6h ago•176 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
144•AlexeyBrin•9h ago•26 comments

Stories from 25 Years of Software Development

https://susam.net/twenty-five-years-of-computing.html
101•vinhnx•7h ago•13 comments

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
851•klaussilveira•1d ago•258 comments

I write games in C (yes, C)

https://jonathanwhiting.com/writing/blog/games_in_c/
139•valyala•4h ago•109 comments

First Proof

https://arxiv.org/abs/2602.05192
68•samasblack•6h ago•52 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
1093•xnx•1d ago•618 comments

Show HN: A luma dependent chroma compression algorithm (image compression)

https://www.bitsnbites.eu/a-spatial-domain-variable-block-size-luma-dependent-chroma-compression-...
7•mbitsnbites•3d ago•0 comments

Al Lowe on model trains, funny deaths and working with Disney

https://spillhistorie.no/2026/02/06/interview-with-sierra-veteran-al-lowe/
64•thelok•6h ago•10 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
235•jesperordrup•14h ago•80 comments

Start all of your commands with a comma (2009)

https://rhodesmill.org/brandon/2009/commands-with-comma/
519•theblazehen•3d ago•191 comments

Reinforcement Learning from Human Feedback

https://rlhfbook.com/
94•onurkanbkrc•9h ago•5 comments

Show HN: I saw this cool navigation reveal, so I made a simple HTML+CSS version

https://github.com/Momciloo/fun-with-clip-path
31•momciloo•4h ago•5 comments

Selection Rather Than Prediction

https://voratiq.com/blog/selection-rather-than-prediction/
13•languid-photic•3d ago•4 comments

Coding agents have replaced every framework I used

https://blog.alaindichiappari.dev/p/software-engineering-is-back
259•alainrk•8h ago•425 comments

A Fresh Look at IBM 3270 Information Display System

https://www.rs-online.com/designspark/a-fresh-look-at-ibm-3270-information-display-system
49•rbanffy•4d ago•9 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
187•1vuio0pswjnm7•10h ago•267 comments

France's homegrown open source online office suite

https://github.com/suitenumerique
615•nar001•8h ago•272 comments

72M Points of Interest

https://tech.marksblogg.com/overture-places-pois.html
36•marklit•5d ago•6 comments

We mourn our craft

https://nolanlawson.com/2026/02/07/we-mourn-our-craft/
348•ColinWright•3h ago•414 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
125•videotopia•4d ago•39 comments

Where did all the starships go?

https://www.datawrapper.de/blog/science-fiction-decline
99•speckx•4d ago•117 comments

Show HN: Kappal – CLI to Run Docker Compose YML on Kubernetes for Local Dev

https://github.com/sandys/kappal
33•sandGorgon•2d ago•15 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
211•limoce•4d ago•119 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
288•isitcontent•1d ago•38 comments

History and Timeline of the Proco Rat Pedal (2021)

https://web.archive.org/web/20211030011207/https://thejhsshow.com/articles/history-and-timeline-o...
20•brudgers•5d ago•5 comments
Open in hackernews

Show HN: CSV GB+ by Data.olllo – Open and Process CSVs Locally

https://apps.microsoft.com/detail/9pfcrwp46v22?hl=en-US&gl=US
51•olllo•8mo ago
I built CSV GB+ by Data.olllo, a local data tool that lets you open, clean, and export gigabyte-sized CSVs (even billions of rows) without writing code.

Most spreadsheet apps choke on big files. Coding in pandas or Polars works—but not everyone wants to write scripts just to filter or merge CSVs. CSV GB+ gives you a fast, point-and-click interface built on dual backends (memory-optimized or disk-backed) so you can process huge datasets offline.

Key Features: Handles massive CSVs with ease — merge, split, dedup, filter, batch export

Smart engine switch: disk-based "V Core" or RAM-based "P Core"

All processing is offline – no data upload or telemetry

Supports CSV, XLSX, JSON, DBF, Parquet and more

Designed for data pros, students, and privacy-conscious users

Register for 7-days free to pro try, pro versions remove row limits and unlock full features. I’m a solo dev building Data.olllo as a serious alternative to heavy coding or bloated enterprise tools.

Download for Windows: https://apps.microsoft.com/detail/9PFR86LCQPGS

User Guide: https://olllo.top/articles/article-0-Data.olllo-UserGuide

Would love feedback! I’m actively improving it based on real use cases.

Comments

xnx•8mo ago
Is this better than the free Tad (https://www.tadviewer.com/) which seems to do similar things for free?
rad_gruchalski•8mo ago
And on operating systems other than Windows...
olllo•8mo ago
Tad is a great tool—very clean and useful for quick exploration.

Data.olllo is focused more on local data processing, not just viewing—things like filtering, transforming, merging, and even running Python code (with AI assistance coming). It’s built for both small and large files with performance in mind, using many cores including Polars under the hood.

Also, good news: the macOS version is in the works and will be submitted to the Mac App Store soon!

dangerlibrary•8mo ago
It is 2025 and CSVs still dominate data interchange between organizations.

https://graydon2.dreamwidth.org/193447.html

esafak•8mo ago
parquet is also popular.
olllo•8mo ago
Absolutely—CSVs are still everywhere, especially for simple interchange between teams and tools. I designed Data.olllo with that in mind.

That said, I also plan to add support for Parquet and other formats soon—definitely agree it's gaining traction for larger, structured datasets.

paddy_m•8mo ago
Do you have a demo video?

What are you using for processing (polars)?

Marketing note: I'm sure you're proud of P Core/V Core, but that doesn't matter to your users, it's an implementation detail. At a maximum I'd write "intelligent execution that scales from small files to large files".

As an implementation note, I would make it simple to operate on just the first 1000 (10k or 100k) rows so responses are super quick, then once the users are happy about the transform, make it a single click to operate on the entire file with a time estimate.

Another feature I'd like in this vein is execute on a small subset, then if you find an error with a larger subset, try to reduce the larger subset to a small quick to reproduce version. Especially for deduping.

marcellus23•8mo ago
> Marketing note: I'm sure you're proud of P Core/V Core, but that doesn't matter to your users, it's an implementation detail. At a maximum I'd write "intelligent execution that scales from small files to large files".

Speaking personally, "intelligent execution that scales from small files to large files" sounds like marketing buzz that could mean absolutely nothing. I like that it mentions specifically switching between RAM and disk-powered engines, because that suggests it's not just marketing speak, but was actually engineered. Maybe P vs V Core is not the best way to market it, but I think it's worth mentioning that design.

olllo•8mo ago
Thanks for the thoughtful take—really appreciate both perspectives.

You're right that terms like "intelligent execution" can feel vague without concrete backing. My goal with mentioning P Core/V Core was to hint at the underlying design—switching between in-memory and disk-based engines like Polars and Vaex—without overwhelming with technical detail.

I’ll look for a better way to explain the idea clearly and briefly. Thanks again!

gopher_space•8mo ago
I wish every product had an engineer-only landing page I could set as a default in my browser. The number of companies that assume I'm familiar with their offering is astounding, and I'm usually looking for implementation docs just to figure out what it actually does.

I'm not saying we need a morlock/eloi toggle.

olllo•8mo ago
Thanks for the thoughtful feedback!

Yes, Data.olllo uses including Polars under the hood for fast and efficient processing. A demo video is in the works and should be up soon.

Good point about the "P Core/V Core" naming—I'll simplify that to focus more on the user benefit, like scaling from small to large files smoothly.

I also like your idea of running transformations on a sample first with a one-click full run—very aligned with the vision. And subset reproduction for errors is a great suggestion, especially for things like deduping. Appreciate it!

paddy_m•8mo ago
Feel free to get in touch. We are building similar tools
TheTaytay•8mo ago
Thank you for this. I find myself increasingly using CSVs (TSVs actually) as the data format of choice. I confess I wish this was written for Mac too, but I like the trend of (once again) moving data processing down to our super computers on our desk...
hilti•8mo ago
… I‘m trying to use our super computers in our pockets, like an iPhone ;-) But still struggling with the way how to present CSV data effectively on a small screen, although it‘s huge in terms of pixels compared to computer screens from the 90s

It‘s interesting to research how capable applications like Lotus123 have been even on low resolutions like 800x600 pixel compared to today’s standard

RyanHamilton•8mo ago
QStudio allows querying CSV on mac via DuckDB: https://www.timestored.com/qstudio/csv-file-viewer I've been improving the Mac version a lot lately, key bindings, icon, an App package to download. So if you find any problems please raise a github issue.
hermitcrab•8mo ago
If you are wrangling CSV/TSV files on Mac, it might be worth taking a look at Easy Data Transform.
paddy_m•8mo ago
Ok, if we are all tagging and promoting our own projects, check out mine.

I created Buckaroo to provide a better table viewing experience inside of notebooks. I also built a low code UI and auto cleaning to expedite the wrote data cleaning tasks that take up a large portion of data analysis. Autocleaning is heuristically powered - no LLMs, so it's fast and your data stays local. You can apply different autocleaning strategies and visually inspect the results. When you are happy with the cleaning, you can copy and paste the python code as a reusable function.

All of this is open source, and its extendable/customizable.

Here's a video walking through autocleaning and how to extend it https://youtu.be/A-GKVsqTLMI

Here's the repo: https://github.com/paddymul/buckaroo

olllo•8mo ago
Thank you! I completely agree—TSVs/CSVs are such a simple yet powerful format, and it's great to hear you're making good use of them. I'm also a big fan of doing as much as possible locally—our machines are incredibly capable these days. Good news: I'm currently working on the macOS version of Data.olllo and plan to submit it to the Mac App Store soon. Stay tuned!
crashabr•8mo ago
How does it compare to OpenRefine https://github.com/OpenRefine
bitbasher•8mo ago
What are "massive" CSVs? I have CSVs in the terabytes that need to be deduped by a specific column. Can it handle that? What if I want to run a function on the column to normalize it before the deduping?