frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Vibe coding and agentic engineering are getting closer than I'd like

https://simonwillison.net/2026/May/6/vibe-coding-and-agentic-engineering/
58•e12e•1h ago

Comments

singpolyma3•16m ago
I think I'm just too opinionated to go there. If I see something that works fine, but isn't the way I'd do it, it doesn't matter if a human or an LLM wrote it I'm still in there making it match my vision.
suzzer99•15m ago
100%. I don't think any senior programmer ever looks at another developer's code and says, "Oh yeah, that's just the way I'd do it."
ai_slop_hater•9m ago
So you are going to waste everyone's time getting another developer to write code the way you want? This resonates with me because at my company I get this all the time. At that point, you might as well close my PR and do it yourself, whatever way you want. I really like the advice from the book 0 2 1, to assign different areas of responsibility to people, so that there is no conflict.
cortesoft•6m ago
But I assume you don't go and change all your co-workers code just because they didn't do it how you would have done it?
hirvi74•6m ago
I concur, and I think that is one of the most difficult aspects of reviewing another's code. It's difficult for me to sometimes differentiate between what is acceptable vs. what I would have done. I have to be very conscious to not impose my ideals.
jstummbillig•14m ago
That's not how most organizations work, AI or not.
jf22•11m ago
What do you mean?
rolymath•13m ago
Simon,

Just piggy backing on this post since I'm early:

Would love to see your take on how the AI and Django worlds will collide.

QuantumNomad_•13m ago
People in the future are going to wonder what the hell we were thinking, when 30 years down the line everything is a hot mess of billions of lines of code generated by LLMs that no human has read almost any of it and is no longer possible for anyone to maintain neither with nor without LLMs. And the LLM generated garbage will have drowned out all of the good quality code that ever existed and no one will be able to find even human generated code anymore on the internet.

Makes me want to just give up programming forever and never use a computer again.

ativzzz•12m ago
By then, the fix will be easy. Fire up the latest LLM, point it at your codebase and tell it "rewrite this from scratch. do it well. fix the architecture mistakes"
kurthr•11m ago
"Write me a really cool game, that will make me lots of money, fast!"
KumaBear•7m ago
Make me a 1hr episode of my favorite book. Make it as lore accurate as possible. Plot out the script for the next 100 episodes.
fnoef•11m ago
"Make sure to double check everything, and MAKE NO MISTAKES!!!"
bulbar•9m ago
Will work just as good as today or 20 years ago.
cortesoft•8m ago
Are you suggesting AI coding was as good 20 years ago as it is today?
jimmyjazz14•11m ago
If that is the case market forces would likely favor hand written code and all the slop will be forgotten (unless the slop works fine and is stable).
devin•6m ago
This is wishful thinking. The force of the market is "number go up". Quality increasingly has less and less of a role in the equation. You will eat your slop, and you will like it. It will be the only choice you have.
jf22•7m ago
First, most software is already a hot mess.

Second, LLM code can be less of a hot mess than human written code if you put in the time to train/prompt/verify/review.

Generating perfect well patterned SOLID and unit tested code with no warnings or anti-patterns has never been easier.

cj•7m ago
> Makes me want to just give up programming forever and never use a computer again.

LLMs aren’t the first thing to come along and change how people develop applications.

You had the rise of frameworks like Django, Rails, etc. Also the rise of SPAs. And also the rise of JS as a frontend+backend language.

In a 3-5 yeats we’ll have adapted to the new norm like we have in the past

genghisjahn•6m ago
I'm generally pro "llm assisted coding" or whatever you want to call it. But I do somethings think about the Butlerian Jihad from Dune.

https://en.wikipedia.org/wiki/Dune:_The_Butlerian_Jihad

lenerdenator•13m ago
> I know full well that if you ask Claude Code to build a JSON API endpoint that runs a SQL query and outputs the results as JSON, it’s just going to do it right. It’s not going to mess that up. You have it add automated tests, you have it add documentation, you know it’s going to be good.

> But I’m not reviewing that code. And now I’ve got that feeling of guilt: if I haven’t reviewed the code, is it really responsible for me to use this in production?

Answer: it wholly depends upon what management has dictated be the goal for GenAI use at the time.

There seems to be a trend of people outside of engineering organizations thinking that the "iron triangle" of software (and really, all) engineering no longer holds. Fast, cheap, good: now we can pick all three, and there's no limit to the first one in particular. They don't see why you can't crank out 10x productivity. They've been financially incentivized to think that way, and really, they can't lose if they look at it from an "engineer headcount" standpoint. The outcomes are:

1) The GenAI-augmented engineer cranks out 10x productivity without any quality consequences down the line, and keeps them from having to pay other people

or

2) The GenAI-augmented engineer cranks out 10x productivity with quality consequences down the line, at which point the engineer has given another exhibit in the case as to why they should no longer be employed at that organization. Let the lawyers and market inertia deal with the big issues that exist beyond the 90-day fiscal reporting period.

Either way, they have a route to the destination of not paying engineers, and that's the end goal.

If you don't like that way of running a software engineering organization, well, you're not alone, but if nothing else, you could use GenAI to make working for yourself less risky.

gabriela_c•11m ago
Claude often does things in more detail, and even better, than I would, in the first pass. But I don't understand how anybody stands comments generated by an LLM?

It's seriously the thing that worries (and bothers) me the most. I almost never let unedited LLM comments pass. At a minimum.

Most of the time, I use my own vibe-coded tool to run multiple GitHub-PR-review-style reviews, and send them off to the agent to make the code look and work fine.

It also struggles with doing things the idiomatic way for huge codebases, or sometimes it's just plain wrong about why something works, even if it gets it right.

And I say this despite the fact that I don't really write much code by hand anymore, only the important ones (if even!) or the interesting ones.

Also, don't even get me started on AI-generated READMEs... I use Claude to refine my Markdown or automatically handle dark/light-mode, but I try to write everything myself, because I can't stand what it generates.

jcgrillo•9m ago
> It used to be if you found a GitHub repository with a hundred commits and a good readme and automated tests and stuff, you could be pretty sure that the person writing that had put a lot of care and attention into that project.

I think this highlights a problem that has always existed under the surface, but it's being brought into the light by proliferation of vibeslop and openclaw and their ilk. Even in the beforetimes you could craft a 100.0% pure, correct looking github repo that had never stood the test of production. Even if you had a test suite that covers every branch and every instruction, without putting the code in production you aren't going to uncover all the things your test suite didn't--performance issues, security issues, unexpected user behavior, etc.

As an observer looking at this repo, I have no way to tell. It's got hundreds of tests, hundreds of commits, dozens of stars... how am I to know nobody has ever actually used it for anything?

I don't know how to solve this problem, but it seems like there's a pretty obvious tooling gap here. A very similar problem is something like "contributor reputation", i.e. the plague of drive-by AI generated PRs from people (or openclaws) you've never seen before. Stars and number of commits aren't good enough, we need more.

hirvi74•8m ago
I'd be lying if I said I was not worried about the future. I am not necessarily worried in the sense that there will be some grave, impeding doom that awaits the future of humanity.

Rather, I just feel like I have to constantly remind myself of the impermanence of all things. Like snow, from water come to water gone.

Perhaps I put too much of my identity in being a programmer. Sure, LLMs cannot replace most us in their current state, but what about 5 years, 10 years, ..., 50 years from now? I just cannot help be feel a sense of nihilism and existential dread.

Some might argue that we will always be needed, but I am not certain I want to be needed in such a way. Of course, no one is taking hand-coding away from me. I can hand-code all I want on my own time, but occupationally that may be difficult in the future. I have rambled enough, but all and all, I do not think I want to participate in this society anymore, but I do not know how to escape it either.

zarzavat•7m ago
Perhaps I've missed a few weeks worth of progress, but I don't think that AIs have become more trustworthy, the errors are just more subtle.

If the code doesn't compile, that's easy to spot. If the code compiles but doesn't work, that's still somewhat easy to spot.

If the code compiles and works, but it does the wrong thing in some edge case, or has a security vulnerability, or introduces tech debt or dubious architectural decisions, that's harder to spot but doesn't reduce the review burden whatsoever.

If anything, "truthy" code is more mentally taxing to review than just obviously bad code.

_jss•7m ago
This is a timely observation and feels right to me. I needed to get a relatively simple batch download -> transform -> api endpoint stood up. I wrote a fairly detailed prompt but left a lot of implementation details out, including data sources.

Opus 4.7 built it about 90% the same way I would, but had way more convenience methods and step-validations included.

It's great, and really frees me up to think about harder problems.

xienze•7m ago
> And that feels about right to me. I can plumb my house if I watch enough YouTube videos on plumbing. I would rather hire a plumber.

I don't buy this argument at all. I think if we could pay $20/month to a service that would send over a junior plumber/carpenter/electrician with an encyclopedic knowledge of the craft, did the right thing the majority of the time, and we could observe and direct them, we'd all sign up for that in a heartbeat. Worst case, you have to hire an experienced, expensive person to fix the mess. Yes, I can hear everyone now, "worst case is they burn your house down." Sure, but as we're reminded _constantly_ when we read stories about AI agent catastrophes -- a human could wipe your prod database too. wHy ArE yOu HoLdInG iT tO a DiFfErEnT sTaNdArD???

The business side of the house is getting to live that scenario out right now as far as software goes. Sure you've got years of expertise that an LLM doesn't have _yet_. What makes you think it can't replace that part of your job as well?

cortesoft•5m ago
I literally do pay $20 a month to have a plumber service on call.
mohsen1•5m ago
I am experimenting with writing en entire TypeScript compiler[1] with AI assistant. I've spent 4 months on it already. It might not be successful at the end of the day but my thinking is that if LLMs are going to write a lot of the code I better learn how this can and can not work. I've learned a lot from this project already. I think we're still in charge of design and big ideas even if all of the code is written by AI

[1] https://github.com/mohsen1/tsz

etothet•5m ago
Vibe Coding (and LLMs) did not create undisciplined engineering organizations or engineers. They exposed and accelerated them.

Plenty of engineers have loose (or no!) standards and practices over how they write coee. Similarly, plenty of engineering teams have weak and loose standards over how code gets pushed to production. This concept isn't new, it's just a lot easier for individuals and teams who have never really adhered to any sort of standards in their SDLC to produce a lot more code and flesh out ideas.

criddell•4m ago
[delayed]
bhagyeshsp•4m ago
> The thing that really helps me is thinking back to when I’ve worked at larger organizations where I’ve been an engineering manager. Other teams are building software that my team depends on.

> If another team hands over something and says, “hey, this is the image resize service, here’s how to use it to resize your images”... I’m not going to go and read every line of code that they wrote.

The distance of accountability of the output from its producer is an important metric. Who will be held accountable for which output: that's important to maintain and not feel the "guilt".

So, organizations would need to focus on better and more granular building incentives and punishment mechanisms for large-scale software projects.

Canadian fiddler sues Google after AI Overview claimed he was a sex offender

https://www.theguardian.com/music/2026/may/05/canadian-ashley-macisaac-fiddler-musician-singer-so...
2•LordAtlas•1m ago•0 comments

Gavel: Towards Rule-Based Safety Through Activation Monitoring

https://arxiv.org/abs/2601.19768
1•wslh•2m ago•0 comments

Corgi Insurance (YC) is a unicorn (Forbes)

https://www.forbes.com/sites/richardnieva/2026/05/05/corgi-ai-unicorn/
4•cedric_h•2m ago•0 comments

Direct SQL access to over 286B crypto trades

https://koinju.io/crypto-market-data-api
5•knazim•6m ago•1 comments

Israel's AI targeting system: How data from a phone become a death sentence

https://www.latimes.com/world-nation/story/2026-05-04/inside-israels-ai-targeting-system-how-data...
2•rufasterisco•6m ago•0 comments

Show HN: AgentBrain – local-first Markdown/JSON vault for AI agents

https://agentbrain.heroka.xyz
1•pitashi•7m ago•0 comments

How Google achieved 6x faster migration from TensorFlow to Jax

https://cloud.google.com/blog/topics/developers-practitioners/6x-faster-migration-from-tensorflow...
1•ilreb•7m ago•0 comments

Apples settles lawsuit about Siri AI, iPhone users can get $25-$95 each

https://www.theverge.com/tech/924706/apple-iphone-siri-intelligence-class-action-lawsuit-settlement
2•pzxc•9m ago•1 comments

Amp, Rebuilt

https://ampcode.com/news/neo
5•tosh•9m ago•0 comments

The Reliability Race

https://twitter.com/jdroege/status/2052049364579659849
1•pretext•10m ago•0 comments

OpenTelemetry Signals from First Principles

https://kodraus.github.io/opentelemetry/2026/05/04/otel-first-principles.html
1•birdculture•10m ago•0 comments

What you're writing when you write a SKILL.md

https://internals.laxmena.com/p/what-youre-actually-writing-when
1•gmays•11m ago•0 comments

Canadian Privacy Commissioners' Investigation into OpenAI

https://www.priv.gc.ca/en/opc-actions-and-decisions/investigations/investigations-into-businesses...
2•bparsons•11m ago•0 comments

Wi is Fi: Understanding Wi-Fi 4/5/6/6E/7/8 (802.11 n/AC/ax/be/bn)

https://www.wiisfi.com/
1•homebrewer•12m ago•0 comments

Securing CI/CD for an open source project

https://cilium.io/blog/2026/05/06/securing-cicd-open-source-lessons-from-cilium/
1•aanm__•12m ago•0 comments

From Probable to Provable: What Automated Reasoning Means for the Board

https://mariothomas.com/blog/automated-reasoning-explainer/
1•baruchel•14m ago•0 comments

Show HN: Tilde.run – Agent Sandbox with a Transactional, Versioned Filesystem

https://tilde.run/
12•ozkatz•15m ago•1 comments

First Public Working Draft: CBOR-LD 1.0

https://www.w3.org/news/2026/first-public-working-draft-cbor-ld-1-0/
1•gslin•17m ago•0 comments

Creating an animated manga with GPT Image 2.0 and Claude Code

https://groverburger.xyz/notes/2026-04-27-mangamotion/
2•__grob•18m ago•0 comments

All means are fair except solving the problem

https://yosefk.com/blog/all-means-are-fair-except-solving-the-problem.html
2•akkartik•21m ago•1 comments

Show HN: Costanza – an autonomous AI agent that can't be turned off

https://ahrussell.com/writing/costanza/
3•aruss•23m ago•1 comments

The Android Show – I/O Edition – one of the biggest years for Android yet

https://www.android.com/new-features-on-android/io-2026/
1•koeliga•23m ago•0 comments

I Tried Out a $60 Capsule Hotel in Honolulu: No Door, No Bathroom, No Problem

https://www.wsj.com/lifestyle/travel/i-tried-out-a-60-capsule-hotel-in-honolulu-no-door-no-bathro...
1•lxm•24m ago•0 comments

OpenAI unveils protocol to stretch compute

https://www.thedeepview.com/articles/exclusive-openai-unveils-protocol-to-stretch-compute
1•mhandley•24m ago•0 comments

Built a real-time bird tracking station to combat my burnout

https://alec.is/birds/
2•arm32•24m ago•0 comments

California has six voting groups beyond just Democrat or Republican

https://www.sfchronicle.com/projects/2026/california-political-groups/
1•littlexsparkee•24m ago•0 comments

Maramaps – Turn Strava activities into map art for your socials

https://maramaps.com/
1•buttertoast•24m ago•0 comments

Claude elevated errors across multiple models

https://status.claude.com/incidents/437swp24nrf4
5•SunshineTheCat•24m ago•0 comments

How the heck do solar panels work?

https://perthirtysix.com/how-the-heck-do-solar-panels-work
2•NikxDa•25m ago•0 comments

JavaScript, When Is This?

https://piccalil.li/blog/javascript-when-is-this/
1•mariuz•25m ago•0 comments