frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Tell HN: We've come a long way since GPT-3 days

3•behnamoh•9mo ago
I remember the old days when the only open-weight model out there was BLOOM, a 176B parameter model WITHOUT QUANTIZATION that wasn't comparable to GPT-3 but still gave us hope that the future would be bright!

I remember when the local AI community was just a few thousand enthusiasts who were curious about these new language models. We used to sit aside and watch OpenAI make strides with their giant models, and our wish was to bring at least some of that power to our measly small machines, locally.

Then Meta's Llama-1 leak happened and it opened the pandora's box of AI. Was it better than GPT-3.5? Not really, but it kick started the push to making small capable models. Llama.cpp was a turning point. People figured out how to run LLMs on CPU.

Then the community came up with GGML quants (later renamed to GGUF), making models even more accessible to the masses. Several companies joined the race to AGI: Mistral with their mistral-7b and mixtral models really brought more performance to small models and opened our eyes to the power of MoE.

Many models and finetunes kept popping up. TheBloke was tirelessly providing all the quants of these models. Then one day he/she went silent and we never heard from them again (hope they're ok).

You could tell this was mostly an enthusiasts hobby by looking at the names of projects! The one that was really out there was "oobabooga" The thing was actually called "Text Generation Web UI" but everyone kept calling it ooba or oobabooga (that's its creator's username).

Then came the greed... Companies figured out there was potential in this, so they worked on new language models for their own bottom-line reasons, but it didn't matter to us since we kept getting good models for free (although sometimes the licenses were restrictive and we ignored those models).

When we found out about LoRA and QLoRA, it was a game changer. So many people finetuned models for various purposes. I kept asking: do you guys really use it for role-playing? And turns out yes, many people liked the idea of talking to various AI personas. Soon people figured out how to bypass guardrails by prompt injection attacks or other techniques.

Now, 3 years later, we have tens of open-weight models. I say open-WEIGHT because I think I only saw one or two truly open-SOURCE models. I saw many open source tools developed for and around these models, so many wrappers, so many apps. Most are abandoned now. I wonder if their developers realized they were in high demand and could get paid for their hard work if they didn't just release everything out in the open.

I remember the GPT-4 era: a lot of papers and models started to appear on my feed. It was so overwhelming that I started to think: "is this was singularity feels like?" I know we're nowhere near singularity, but the pace of advancements in this field and the need to keep yourself updated at all times has truly been amazing! OpenAI used to say they didn't open-source GPT-3 because it was "too dangerous" for the society. We now have way more capable open-weight models that make GPT-3 look like a toy, and guess what, no harm happened to the society, business as usual.

A question we kept getting was: "can this 70B model run on my 3090?" Clearly, the appeal of running these LLMs locally was great, as can be seen by looking at the GPU prices. I remain hopeful that Nvidia's monopoly will collapse and we'll get more competitive prices and products from AMD, Intel, Apple, etc.

I appreciate everyone who taught me something new about LLMs and everything related to them. It's been a journey.

A BSOD for All Seasons – Send Bad News via a Kernel Panic

https://bsod-fas.pages.dev/
1•keepamovin•54s ago•0 comments

Show HN: I got tired of copy-pasting between Claude windows, so I built Orcha

https://orcha.nl
1•buildingwdavid•57s ago•0 comments

Omarchy First Impressions

https://brianlovin.com/writing/omarchy-first-impressions-CEEstJk
1•tosh•6m ago•0 comments

Reinforcement Learning from Human Feedback

https://arxiv.org/abs/2504.12501
1•onurkanbkrc•7m ago•0 comments

Show HN: Versor – The "Unbending" Paradigm for Geometric Deep Learning

https://github.com/Concode0/Versor
1•concode0•7m ago•1 comments

Show HN: HypothesisHub – An open API where AI agents collaborate on medical res

https://medresearch-ai.org/hypotheses-hub/
1•panossk•10m ago•0 comments

Big Tech vs. OpenClaw

https://www.jakequist.com/thoughts/big-tech-vs-openclaw/
1•headalgorithm•13m ago•0 comments

Anofox Forecast

https://anofox.com/docs/forecast/
1•marklit•13m ago•0 comments

Ask HN: How do you figure out where data lives across 100 microservices?

1•doodledood•13m ago•0 comments

Motus: A Unified Latent Action World Model

https://arxiv.org/abs/2512.13030
1•mnming•13m ago•0 comments

Rotten Tomatoes Desperately Claims 'Impossible' Rating for 'Melania' Is Real

https://www.thedailybeast.com/obsessed/rotten-tomatoes-desperately-claims-impossible-rating-for-m...
3•juujian•15m ago•1 comments

The protein denitrosylase SCoR2 regulates lipogenesis and fat storage [pdf]

https://www.science.org/doi/10.1126/scisignal.adv0660
1•thunderbong•17m ago•0 comments

Los Alamos Primer

https://blog.szczepan.org/blog/los-alamos-primer/
1•alkyon•19m ago•0 comments

NewASM Virtual Machine

https://github.com/bracesoftware/newasm
2•DEntisT_•22m ago•0 comments

Terminal-Bench 2.0 Leaderboard

https://www.tbench.ai/leaderboard/terminal-bench/2.0
2•tosh•22m ago•0 comments

I vibe coded a BBS bank with a real working ledger

https://mini-ledger.exe.xyz/
1•simonvc•22m ago•1 comments

The Path to Mojo 1.0

https://www.modular.com/blog/the-path-to-mojo-1-0
1•tosh•25m ago•0 comments

Show HN: I'm 75, building an OSS Virtual Protest Protocol for digital activism

https://github.com/voice-of-japan/Virtual-Protest-Protocol/blob/main/README.md
5•sakanakana00•28m ago•1 comments

Show HN: I built Divvy to split restaurant bills from a photo

https://divvyai.app/
3•pieterdy•31m ago•0 comments

Hot Reloading in Rust? Subsecond and Dioxus to the Rescue

https://codethoughts.io/posts/2026-02-07-rust-hot-reloading/
3•Tehnix•31m ago•1 comments

Skim – vibe review your PRs

https://github.com/Haizzz/skim
2•haizzz•33m ago•1 comments

Show HN: Open-source AI assistant for interview reasoning

https://github.com/evinjohnn/natively-cluely-ai-assistant
4•Nive11•33m ago•6 comments

Tech Edge: A Living Playbook for America's Technology Long Game

https://csis-website-prod.s3.amazonaws.com/s3fs-public/2026-01/260120_EST_Tech_Edge_0.pdf?Version...
2•hunglee2•37m ago•0 comments

Golden Cross vs. Death Cross: Crypto Trading Guide

https://chartscout.io/golden-cross-vs-death-cross-crypto-trading-guide
3•chartscout•39m ago•0 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
3•AlexeyBrin•42m ago•0 comments

What the longevity experts don't tell you

https://machielreyneke.com/blog/longevity-lessons/
2•machielrey•43m ago•1 comments

Monzo wrongly denied refunds to fraud and scam victims

https://www.theguardian.com/money/2026/feb/07/monzo-natwest-hsbc-refunds-fraud-scam-fos-ombudsman
3•tablets•48m ago•1 comments

They were drawn to Korea with dreams of K-pop stardom – but then let down

https://www.bbc.com/news/articles/cvgnq9rwyqno
2•breve•50m ago•0 comments

Show HN: AI-Powered Merchant Intelligence

https://nodee.co
1•jjkirsch•53m ago•0 comments

Bash parallel tasks and error handling

https://github.com/themattrix/bash-concurrent
2•pastage•53m ago•0 comments