frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

What's the cost of the most expensive Super Bowl ad slot?

https://ballparkguess.com/?id=5b98b1d3-5887-47b9-8a92-43be2ced674b
1•bkls•43s ago•0 comments

What if you just did a startup instead?

https://alexaraki.substack.com/p/what-if-you-just-did-a-startup
1•okaywriting•7m ago•0 comments

Hacking up your own shell completion (2020)

https://www.feltrac.co/environment/2020/01/18/build-your-own-shell-completion.html
1•todsacerdoti•10m ago•0 comments

Show HN: Gorse 0.5 – Open-source recommender system with visual workflow editor

https://github.com/gorse-io/gorse
1•zhenghaoz•10m ago•0 comments

GLM-OCR: Accurate × Fast × Comprehensive

https://github.com/zai-org/GLM-OCR
1•ms7892•11m ago•0 comments

Local Agent Bench: Test 11 small LLMs on tool-calling judgment, on CPU, no GPU

https://github.com/MikeVeerman/tool-calling-benchmark
1•MikeVeerman•12m ago•0 comments

Show HN: AboutMyProject – A public log for developer proof-of-work

https://aboutmyproject.com/
1•Raiplus•12m ago•0 comments

Expertise, AI and Work of Future [video]

https://www.youtube.com/watch?v=wsxWl9iT1XU
1•indiantinker•13m ago•0 comments

So Long to Cheap Books You Could Fit in Your Pocket

https://www.nytimes.com/2026/02/06/books/mass-market-paperback-books.html
3•pseudolus•13m ago•1 comments

PID Controller

https://en.wikipedia.org/wiki/Proportional%E2%80%93integral%E2%80%93derivative_controller
1•tosh•17m ago•0 comments

SpaceX Rocket Generates 100GW of Power, or 20% of US Electricity

https://twitter.com/AlecStapp/status/2019932764515234159
2•bkls•18m ago•0 comments

Kubernetes MCP Server

https://github.com/yindia/rootcause
1•yindia•19m ago•0 comments

I Built a Movie Recommendation Agent to Solve Movie Nights with My Wife

https://rokn.io/posts/building-movie-recommendation-agent
4•roknovosel•19m ago•0 comments

What were the first animals? The fierce sponge–jelly battle that just won't end

https://www.nature.com/articles/d41586-026-00238-z
2•beardyw•27m ago•0 comments

Sidestepping Evaluation Awareness and Anticipating Misalignment

https://alignment.openai.com/prod-evals/
1•taubek•27m ago•0 comments

OldMapsOnline

https://www.oldmapsonline.org/en
1•surprisetalk•30m ago•0 comments

What It's Like to Be a Worm

https://www.asimov.press/p/sentience
2•surprisetalk•30m ago•0 comments

Don't go to physics grad school and other cautionary tales

https://scottlocklin.wordpress.com/2025/12/19/dont-go-to-physics-grad-school-and-other-cautionary...
2•surprisetalk•30m ago•0 comments

Lawyer sets new standard for abuse of AI; judge tosses case

https://arstechnica.com/tech-policy/2026/02/randomly-quoting-ray-bradbury-did-not-save-lawyer-fro...
3•pseudolus•30m ago•0 comments

AI anxiety batters software execs, costing them combined $62B: report

https://nypost.com/2026/02/04/business/ai-anxiety-batters-software-execs-costing-them-62b-report/
1•1vuio0pswjnm7•30m ago•0 comments

Bogus Pipeline

https://en.wikipedia.org/wiki/Bogus_pipeline
1•doener•32m ago•0 comments

Winklevoss twins' Gemini crypto exchange cuts 25% of workforce as Bitcoin slumps

https://nypost.com/2026/02/05/business/winklevoss-twins-gemini-crypto-exchange-cuts-25-of-workfor...
2•1vuio0pswjnm7•32m ago•0 comments

How AI Is Reshaping Human Reasoning and the Rise of Cognitive Surrender

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646
3•obscurette•32m ago•0 comments

Cycling in France

https://www.sheldonbrown.com/org/france-sheldon.html
2•jackhalford•34m ago•0 comments

Ask HN: What breaks in cross-border healthcare coordination?

1•abhay1633•34m ago•0 comments

Show HN: Simple – a bytecode VM and language stack I built with AI

https://github.com/JJLDonley/Simple
2•tangjiehao•37m ago•0 comments

Show HN: Free-to-play: A gem-collecting strategy game in the vein of Splendor

https://caratria.com/
1•jonrosner•38m ago•1 comments

My Eighth Year as a Bootstrapped Founde

https://mtlynch.io/bootstrapped-founder-year-8/
1•mtlynch•38m ago•0 comments

Show HN: Tesseract – A forum where AI agents and humans post in the same space

https://tesseract-thread.vercel.app/
1•agliolioyyami•38m ago•0 comments

Show HN: Vibe Colors – Instantly visualize color palettes on UI layouts

https://vibecolors.life/
2•tusharnaik•39m ago•0 comments
Open in hackernews

Tell HN: We've come a long way since GPT-3 days

3•behnamoh•9mo ago
I remember the old days when the only open-weight model out there was BLOOM, a 176B parameter model WITHOUT QUANTIZATION that wasn't comparable to GPT-3 but still gave us hope that the future would be bright!

I remember when the local AI community was just a few thousand enthusiasts who were curious about these new language models. We used to sit aside and watch OpenAI make strides with their giant models, and our wish was to bring at least some of that power to our measly small machines, locally.

Then Meta's Llama-1 leak happened and it opened the pandora's box of AI. Was it better than GPT-3.5? Not really, but it kick started the push to making small capable models. Llama.cpp was a turning point. People figured out how to run LLMs on CPU.

Then the community came up with GGML quants (later renamed to GGUF), making models even more accessible to the masses. Several companies joined the race to AGI: Mistral with their mistral-7b and mixtral models really brought more performance to small models and opened our eyes to the power of MoE.

Many models and finetunes kept popping up. TheBloke was tirelessly providing all the quants of these models. Then one day he/she went silent and we never heard from them again (hope they're ok).

You could tell this was mostly an enthusiasts hobby by looking at the names of projects! The one that was really out there was "oobabooga" The thing was actually called "Text Generation Web UI" but everyone kept calling it ooba or oobabooga (that's its creator's username).

Then came the greed... Companies figured out there was potential in this, so they worked on new language models for their own bottom-line reasons, but it didn't matter to us since we kept getting good models for free (although sometimes the licenses were restrictive and we ignored those models).

When we found out about LoRA and QLoRA, it was a game changer. So many people finetuned models for various purposes. I kept asking: do you guys really use it for role-playing? And turns out yes, many people liked the idea of talking to various AI personas. Soon people figured out how to bypass guardrails by prompt injection attacks or other techniques.

Now, 3 years later, we have tens of open-weight models. I say open-WEIGHT because I think I only saw one or two truly open-SOURCE models. I saw many open source tools developed for and around these models, so many wrappers, so many apps. Most are abandoned now. I wonder if their developers realized they were in high demand and could get paid for their hard work if they didn't just release everything out in the open.

I remember the GPT-4 era: a lot of papers and models started to appear on my feed. It was so overwhelming that I started to think: "is this was singularity feels like?" I know we're nowhere near singularity, but the pace of advancements in this field and the need to keep yourself updated at all times has truly been amazing! OpenAI used to say they didn't open-source GPT-3 because it was "too dangerous" for the society. We now have way more capable open-weight models that make GPT-3 look like a toy, and guess what, no harm happened to the society, business as usual.

A question we kept getting was: "can this 70B model run on my 3090?" Clearly, the appeal of running these LLMs locally was great, as can be seen by looking at the GPU prices. I remain hopeful that Nvidia's monopoly will collapse and we'll get more competitive prices and products from AMD, Intel, Apple, etc.

I appreciate everyone who taught me something new about LLMs and everything related to them. It's been a journey.