frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Ask HN: What in your opinion is the best model for vibecoding? My thoughts below

1•adinhitlore•4mo ago
So I've been vibecoding for like years but over the past 1-2 weeks it became an obsession to a point my eyes are literally red and inflamed right now since I can't stop it (slightly humours...i was feeling worse off yesterday, the redness is now gone).

anyways my takes:

1. The #1 place is VERY debatable for me, it's a toss between gpt 5 high, "claude thinking" both sonnet 4 and 4.1 opus and surprise,surprise: qwen 235b 'thinking' (the "hidden gem").

Their pros and cons:

gpt 5 high: Usually gives VERY long code so it'generous, no compute is saved, it's a bona fide model but it seems sometimes too aligned for my taste. For example: whenever i force it to design a novel text generation model, unless i am very speficic in my requirements it tries to dumb it down by making a pure n-gram model, which almost feels like an insult, basically saying "look we at openai are the best, here's a stupid markov chain for you to play with, but leave the big game to us". If however you phrase it more in detail and even if you show some pessimism it will not "echo back" the pessimism but rather try to convince you it can be done with some tweaks. The con: Usually it's just...not smart, this is easily seen when you go through the code and you see it had written code very specific to the example you gave, which is the number one symptom of bad programming, a variable/method should be as universal as possible, you don't need a template which only uploads ftp when you plan an upload via http and ftp, as a one example.

2. Claude: Initially i thought it's the best one and for pure coding it's "getting there" but for designing algorithms, gtpt 5 high and qwen 'thinking' outperform it with ideas. I'd say sonnet 4 32k is better for designing and opus for the actual coding, maybe depending on the task and programming language used it will perform differently. The good news is the actual code usually compiles with very few warnings and almost never errors, so it knows what its doing. Even gpt 5 high is worse and qwen will sometimes though rarely give you bad code that will produce an error be it in Python 3 or C/gcc.

Since i covered the 'good' here are the bad and the ugly:

Gemini, grok, amazon nova, whatever microsoft has: don't, just don't. Their shortcomings are so obvious to a point I'm convinced all the people who hype them online are either elon musk (for grok), bill gates (phi4 etc) or zuckerberg (llama). Their codes are very short so it's obvious they will not cover the features requested, compilation feels like 'quantum mechanics' 50/50 chance, the code is written in the worst way possible, and sometimes they even misinterpret entirely what your goal is. You may have some luck debugging with gemini 2.5 pro if you're patient, frankly even the gpt 4 on chatgpt.com version (not the "arena!") is bad for fixing errors but ok with the basic ones.

Another hidden gem: https://console.upstage.ai/playground/chat I'm not "shilling" for it, hard to believe i know, but i don't ignore it entirely because as an indie model i hope it's not too aligned so it may actually give you code that Yudkovsky and Yampolski consider "immediate risk to humanity, civilization and the galaxy".

My experience are 90% with C mostly a lot of Python too, little-to-no C# though back in the days vibecoding c# on gpt 4 sucked a lot.

My ultimate issue as of now is that while LLMs/transformers are great they still lack the innovation, human thought power to come up with original ideas, however they code way faster than human obviously and the code usually works with few warnings or errors - i think the focus towards 2030 should be the innovation power and complex designs of algorithms. Altman dreaming about "discovering new physics" seems a little bit ambitious given the current status quo. Again they're great and they help me a lot, looking forward to see their impact on larger scale on society!

Comments

reify•4mo ago
The 1925 Ford Model T Touring Car is the best bet.

It has amazing brakes for a 1920's car.

The best thing in my experience is, it does not rely on fantasy ai to drive it. you can just turn the key and Vrooom, away you go.

My local mechanic is be particularly pleased with my purchase and recommendation.

He says, he can repair my car without resorting to repairing the damage the ai mechanic did a few days earlier. which, in the long run saves me an awful lot of money on car maintenance.

I dont have to pay two people to fix one job.

isnt it amazing what humans can do.

incomingpain•4mo ago
https://aider.chat/docs/leaderboards/

I think this is my most favourite benchmark that fits me best.

Stuff I can run at home is down the list alot, but that's fine with me.

RFCs vs. READMEs: The Evolution of Protocols

https://h3manth.com/scribe/rfcs-vs-readmes/
1•init0•5m ago•1 comments

Kanchipuram Saris and Thinking Machines

https://altermag.com/articles/kanchipuram-saris-and-thinking-machines
1•trojanalert•5m ago•0 comments

Chinese chemical supplier causes global baby formula recall

https://www.reuters.com/business/healthcare-pharmaceuticals/nestle-widens-french-infant-formula-r...
1•fkdk•8m ago•0 comments

I've used AI to write 100% of my code for a year as an engineer

https://old.reddit.com/r/ClaudeCode/comments/1qxvobt/ive_used_ai_to_write_100_of_my_code_for_1_ye...
1•ukuina•10m ago•1 comments

Looking for 4 Autistic Co-Founders for AI Startup (Equity-Based)

1•au-ai-aisl•20m ago•1 comments

AI-native capabilities, a new API Catalog, and updated plans and pricing

https://blog.postman.com/new-capabilities-march-2026/
1•thunderbong•21m ago•0 comments

What changed in tech from 2010 to 2020?

https://www.tedsanders.com/what-changed-in-tech-from-2010-to-2020/
2•endorphine•26m ago•0 comments

From Human Ergonomics to Agent Ergonomics

https://wesmckinney.com/blog/agent-ergonomics/
1•Anon84•30m ago•0 comments

Advanced Inertial Reference Sphere

https://en.wikipedia.org/wiki/Advanced_Inertial_Reference_Sphere
1•cyanf•31m ago•0 comments

Toyota Developing a Console-Grade, Open-Source Game Engine with Flutter and Dart

https://www.phoronix.com/news/Fluorite-Toyota-Game-Engine
1•computer23•33m ago•0 comments

Typing for Love or Money: The Hidden Labor Behind Modern Literary Masterpieces

https://publicdomainreview.org/essay/typing-for-love-or-money/
1•prismatic•34m ago•0 comments

Show HN: A longitudinal health record built from fragmented medical data

https://myaether.live
1•takmak007•37m ago•0 comments

CoreWeave's $30B Bet on GPU Market Infrastructure

https://davefriedman.substack.com/p/coreweaves-30-billion-bet-on-gpu
1•gmays•48m ago•0 comments

Creating and Hosting a Static Website on Cloudflare for Free

https://benjaminsmallwood.com/blog/creating-and-hosting-a-static-website-on-cloudflare-for-free/
1•bensmallwood•54m ago•1 comments

"The Stanford scam proves America is becoming a nation of grifters"

https://www.thetimes.com/us/news-today/article/students-stanford-grifters-ivy-league-w2g5z768z
2•cwwc•58m ago•0 comments

Elon Musk on Space GPUs, AI, Optimus, and His Manufacturing Method

https://cheekypint.substack.com/p/elon-musk-on-space-gpus-ai-optimus
2•simonebrunozzi•1h ago•0 comments

X (Twitter) is back with a new X API Pay-Per-Use model

https://developer.x.com/
3•eeko_systems•1h ago•0 comments

Zlob.h 100% POSIX and glibc compatible globbing lib that is faste and better

https://github.com/dmtrKovalenko/zlob
3•neogoose•1h ago•1 comments

Show HN: Deterministic signal triangulation using a fixed .72% variance constant

https://github.com/mabrucker85-prog/Project_Lance_Core
2•mav5431•1h ago•1 comments

Scientists Discover Levitating Time Crystals You Can Hold, Defy Newton’s 3rd Law

https://phys.org/news/2026-02-scientists-levitating-crystals.html
3•sizzle•1h ago•0 comments

When Michelangelo Met Titian

https://www.wsj.com/arts-culture/books/michelangelo-titian-review-the-renaissances-odd-couple-e34...
1•keiferski•1h ago•0 comments

Solving NYT Pips with DLX

https://github.com/DonoG/NYTPips4Processing
1•impossiblecode•1h ago•1 comments

Baldur's Gate to be turned into TV series – without the game's developers

https://www.bbc.com/news/articles/c24g457y534o
3•vunderba•1h ago•0 comments

Interview with 'Just use a VPS' bro (OpenClaw version) [video]

https://www.youtube.com/watch?v=40SnEd1RWUU
2•dangtony98•1h ago•0 comments

EchoJEPA: Latent Predictive Foundation Model for Echocardiography

https://github.com/bowang-lab/EchoJEPA
1•euvin•1h ago•0 comments

Disablling Go Telemetry

https://go.dev/doc/telemetry
1•1vuio0pswjnm7•1h ago•0 comments

Effective Nihilism

https://www.effectivenihilism.org/
1•abetusk•1h ago•1 comments

The UK government didn't want you to see this report on ecosystem collapse

https://www.theguardian.com/commentisfree/2026/jan/27/uk-government-report-ecosystem-collapse-foi...
5•pabs3•1h ago•0 comments

No 10 blocks report on impact of rainforest collapse on food prices

https://www.thetimes.com/uk/environment/article/no-10-blocks-report-on-impact-of-rainforest-colla...
3•pabs3•1h ago•0 comments

Seedance 2.0 Is Coming

https://seedance-2.app/
1•Jenny249•1h ago•0 comments