frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Ask HN: How are you automating your coding work?

39•manthangupta109•2h ago
With the increase of vibe coding I am interested in knowing some creative ways people are automating their coding work.

Comments

jackfranklyn•2h ago
Claude Code has genuinely changed my workflow. Not in a "write the whole thing for me" way - more like having a really fast pair who knows the codebase.

The pattern that works best for me: I describe what I want at a high level, let it scaffold, then I read through and course-correct. The reading step is crucial. Blindly accepting generates technical debt faster than you can imagine.

Where it really shines is the tedious stuff - writing tests for edge cases, refactoring patterns across multiple files, generating boilerplate that follows existing conventions. Things that would take me 45 minutes of context-switching I can knock out in 5.

The automation piece I've landed on: I let it handle file operations and running commands, but I stay in the loop on architecture decisions. The moment you start rubber-stamping those, you end up with a codebase you don't understand.

mandeepj•1h ago
> The moment you start rubber-stamping those, you end up with a codebase you don't understand.

Yeah, treat it like an intern or junior engineer who needs constant feedback and reviews.

lucabraccani•1h ago
Which Claude Code model do you usually use? Any noticeable differences?
langarus•1h ago
I've began using Opus and I felt it was a class above all the rest. Used cursor and teste different models, but opus somehow was always much much better. Bought the max for 100$, totally worth it.
catigula•1h ago
Having AI generate tests is technical debt unless what you're doing is extremely trivial and well-trodden in which case you can basically gen all of the code and not care at all.

Tests are where the moat still exists because prior to creating tests the outcomes are unverifiable.

Yodel0914•32m ago
As somewhat of an AI-agnostic, I disagree. Writing tests is one of the things I find most useful about copilot. Of course you need to review them first correctness, but (especially for unit tests) it’s pretty good and getting it right first-time.
lordnacho•1h ago
I have similar observations. The time saved is things like going to some library I wrote to find the exact order of parameters, or looking up some API on the internet and adjusting my code to it. Inevitably if I did that the old way, I would screw up something trivial and get annoyed.

I rarely let it run for over 10 minutes unattended, but the benefits are not just pure time.

Being able to change the code without getting bogged down allows you to try more things. If I have to wait half an hour between iterations, I'm going to run into bedtime quite fast.

On top of this, I'm finding that the thing that takes the deepest attention is often, amazingly, trivial things. Fiddling with a regex takes attention, but it doesn't often decide the success of the project.

By contrast, the actual work, which is making technical decisions, is something I can do without concentrating in the same way. It's strange that the higher value thing feels less stressful.

Put these together and I'm more productive. I can string together a bunch of little things, and not have to be at my sharpest. Work can be scheduled for whenever, which means days are more flexible. More gets done, with less attention.

sjm-lbm•1h ago
Pretty similar here. Another thing I keep thinking is a phrase pilots use when flying airplanes using FMSes and autopilot: "never fly your airplane to a place you haven't already been to in your mind" - that is, don't ever just sit back and let the automation work, stay a step ahead of the automation and drop down to less automation when you aren't certain that it is doing the right thing.

When you send Claude Code something and already have an idea for what an acceptable solution looks like, you're a massive step ahead when it's time to read the diff and decide what you think of it. This does mean that every so often my productivity drops to basically zero as I try to understand what is actually happening before I just put the AI on the job, but so far it seems to be a good rule to keep in mind that allows me to use AI effectively while generating a code base that I still understand.

hmokiguess•1h ago
Each project gets its own share of supervision depending on how critical human intervention is needed.

I have some complex large and strict compliance projects that the AI is a pair programmer but I make most of the decisions, and I have smaller projects that, despite great impact on the bottom line, can be entirely done unsupervised due to the low risk factor of "mistakes" and the easiness of correcting them after the fact they are caught by the AI as well.

denysvitali•1h ago
With Happy (https://blog.denv.it/posts/im-happy-engineer-now/)
pbohun•1h ago
I'm not. I'm learning a little bit each day, making my brain better and myself more productive as I go.
bravura•1h ago
We use beads for everything. We label them as "human-spec" needed if they are not ready to implement. We label them as "qa-needed" if they cannot be verified through automatic tests.

I wrote beads-skills for Claude that I'll release soon to enforce this process.

2026 will be the year of agent orchestration for those of us who are frustrated having 10 different agents to check on constantly.

gastown is cool but too opinionated.

I'm excited about this promising new project: https://github.com/jzila/canopy

We're writing an internal tool to help with planning, which most people don't think is a problem but I think is a serious problem. Most plans are either too long and/or you end up repeating yourself.

Yodel0914•38m ago
Out of interest, what sort of products/systems are you building?
simonw•1h ago
One of my biggest unlocks has been embracing Claude Code for web - the cloud version - and making sure my projects are setup to work with it.

I mainly work in Python, and I've been ensuring that all of my projects have a test suite which runs cleanly with "uv run pytest" - using a dev dependency group to ensure the right dependencies are installed.

This means I can run Claude Code against any of my repos and tell it "run 'uv run pytest', then implement ..." - which is a shortcut for having it use TDD and write tests for the code it's building, which is essential for having coding agents produce working code that they've tested before they commit.

Once this is working well I can drop ideas directly into the Claude app on my iPhone and get 80% of the implementation of the idea done by the time I get back to a laptop to finish it off.

I wrote a bit about "uv run pytest" and dependency groups here: https://til.simonwillison.net/uv/dependency-groups

throwup238•1h ago
> and making sure my projects are setup to work with it.

MESA drivers are a godsend in Claude Code for web if working on non-web GUIs. It can take screenshots and otherwise interact with them.

corysama•1h ago
What does "MESA drivers" refer to here? I'm guessing it's not GPU drivers from https://mesa3d.org/
throwup238•14m ago
No those don’t work in most cloud VMs but MESA provides llvmpipe/softpipe implementations for Vulkan, OpenGL, etc. They’re software renders so relatively slow but work in headless VMs like Claude Code for web environments.
athrowaway3z•23m ago
Half my visits to HN are to check out a comment that explains the right uv inline dep syntax

   #!/usr/bin/env -S uv run --script
   # /// script
   # dependencies = [
   #   "requests<3",
   #   "rich",
   # ]
   # ///
   import requests, rich
   # ... script goes here`
   
so i can tell claude to write a self contained script it can later use.
sanderjd•5m ago
I've been trying to use workflows like this, but I quickly run into token limits.

I'm curious, for those of you who work like this, what level of subscription do you have?

onlyrealcuzzo•1h ago
For my real work? It has not been helpful so far.

For side projects? It's been a 10x+ multiplier.

jmathai•1h ago
The most useful automation for me has been a few simple commands. Here are some examples I use for repos in GitHub to resolve issues and PRs.

/gh-issue [issue number]

/gh-pr [pr number]

Edit: replaced links to private github repo to pastebin.

https://pastebin.com/5Spg4jGu

https://pastebin.com/hubWMqGa

johnQdeveloper•1h ago
Those both 404, private repo?
jmathai•1h ago
Meh, fixed. Replaced with pastebin links in and kept the repo private.
johnQdeveloper•44m ago
Thanks
growthloops•1h ago
ML Engineer here. For coding, I mostly use Cursor/Claude Code as a fast pair. I'll detail what I want at a high level, let it draft, then I incrementally make changes.

Where I've automated more aggressively is everywhere around the code. My main challenge was running experiments repeatedly across different systems and keeping track of the various models I ran and their metrics, etc. I started using Skyportal.ai as an ops-side agent. For me, it's mostly: take the training code I just iterated on, automatically install and configure the system with the right ML stack, run experiments via prompt, and see my model metrics from there.

epolanski•1h ago
I haven't automated anything to be honest, but LLMs are invaluable in connecting dots in repositories or exploring dependencies source codes.

The first saves me days of work/month by sparing me endless paper pages of notes trying to figure out why things work in a certain way in legacy work codebases. The second spares me from having to dig too much in partially outdated or lacking documentation or having to melt my brain understanding the architecture of every different dependency.

So I just put major deps in my projects in a `_vendor` directory that contains the source code of the dependencies and if I have doubts LLMs dig into it and their test to shed light.

What I haven't seen anybody yet accomplish is produce quality software by having AI write them. I'm not saying they can't help here, but the bottleneck is still reviewing and as soon as you get sloppy, codebase quality goes south, and the product quality follows soon after.

al_borland•1h ago
It’s usually just a slightly faster web search. When I try to have it do more, I end up spinning my wheels and then doing a web search.

I’ll sometimes have it help read really long error messages as well.

I got it to help me fix a reported security vulnerability, but it was a long road and I had to constantly work to keep it from going off the rails and adding insane amounts of complexity and extra code. It likely would have been faster for me to read up on the specific vulnerability, take a walk, and come back to my desk to write something up.

mjr00•1h ago
If I know what I want to code and it's a purely mechanical exercise to code it, I'll just tell Claude what to do and it does it. Pretty neat.

When I don't know what I want to do, I read existing code, think about it, and figure it out. Sometimes I'll sketch out ideas by writing code, then when I have something I like I'll get Claude to take my sketch as an example and having it go forward.

The big mistake I see people make is not knowing when to quit. Even with Opus 4.5 it still does weird things, and I've seen people end up arguing with Claude or trying to prompt engineer their way out of things when it would have been maybe 30 seconds of work to fix things manually. It's like people at shopping malls who spend 15 minutes driving in the parking lot to find a spot close to the door when they could have parked in the first spot they saw and walked to the door in less than a minute.

And as always, every line of code was written by me even if it wasn't written by me. I'm responsible for it, so I review all of it. If I wouldn't have written it on my own without AI assistance I don't commit it.

mountain_peak•29m ago
> The big mistake I see people make is not knowing when to quit.

This is sage advice. I spent the better part of a day trying to steer Gemini into correcting an inconsistency when I likely could have solved it in under an hour. I think persevering with Gemini was due to a number of factors, including novelty, stubbornness, and (unfortunately) not knowing in detail what Gemini had written up to that point.

I eventually studied the resulting code, which ended up having a number of nested 'hacks' and required refactoring - more time wasted, but still much faster overall.

cadamsdotcom•46m ago
The biggest principle is codification. Codify everything.

For instance, this skill of web development: https://raw.githubusercontent.com/vercel-labs/web-interface-...

That’s too much for a model to carry in its context while it’s trying to do actual work.

Far better is to give that skill.md to a model and have it produce several hundred lines of code with a shebang at the top. Now you haven’t got a skill, you’ve got a script. And it’s a script the model can run any time to check its work, without knowing what the script does, how, or why - it just sees the errors. Now all your principles of web dev can be checked across your codebase in a few hundred milliseconds while burning zero tokens.

TDD is codification too: codifying in executable form the precise way you want your logic to work. Enforce a 10ms timeout on every unit test and as a side effect your model won’t be able to introduce I/O or anything else that prevents parallel, randomized execution of your test suite. It’s awesome to be able to run ALL the tests hundreds of times per day.

Constantly checking your UI matches your design system? Have the model write a script that looks at your frontend codebase and refuses to let the model commit anything that doesn’t match the design system.

Codification is an insanely powerful thing to build into your mindset.

wrs•4m ago
[delayed]
anditherobot•32m ago
We're overlooking a critical metric in AI-assisted development: Token and Context Window to Utility Ratio.

AI coding tools are burning massive token budgets on boilerplate thousands of tokens just to render simple interfaces.

Consider the token cost of "Hello World":

- Tkinter: `import tkinter as tk; tk.Button(text="Hello").pack()`

- React: 500MB of node_modules, and dependencies

Right now context windows token limits are finite and costly. What do you think?

My prediction is that tooling that manage token and context efficiency will become essential.

tomduncalf•14m ago
But the model doesn't need to read the node_modules to write a React app, it just needs to write the React code (which it is heavily post-trained to be able to use). So the fair counter example is like:

function Hello() { return <button>Hello</buttton> }

scuff3d•28m ago
I find it most useful for getting up to speed on new libraries quickly, and for bouncing design ideas. I'll lay out what my goals are and the approaches I'm considering, and ask it to poke holes in them or to point out issues or things to keep in mind. Found it shockingly helpful in covering my blind spots
margorczynski•8m ago
At work unfortunately (?) we don't use any AI but there is movement to introduce it in some form (it is a heavily regulated area so it won't be YOLO coding using an agent for sure).

But my side projects which I kinda abandoned a long time ago are getting a second life and it is really fun just to direct the agent instead of slowly re-aquire all of the knowledge and waste time typing in all the stuff into the computer.

FreeBSD

https://docs.freebsd.org/en/books/handbook/
57•vermaden•37m ago•4 comments

Show HN: ChartGPU – WebGPU-powered charting library (1M points at 60fps)

https://github.com/ChartGPU/ChartGPU
393•huntergemmer•6h ago•130 comments

TeraWave Satellite Communications Network

https://www.blueorigin.com/terawave
79•T-A•2h ago•46 comments

Linux from Scratch

https://www.linuxfromscratch.org/lfs/view/stable/
261•Alupis•2h ago•71 comments

Skip Is Now Free and Open Source

https://skip.dev/blog/skip-is-free/
180•dayanruben•6h ago•54 comments

The WebRacket language is a subset of Racket that compiles to WebAssembly

https://github.com/soegaard/webracket
44•mfru•4d ago•4 comments

PicoPCMCIA – a PCMCIA development board for retro-computing enthusiasts

https://www.yyzkevin.com/picopcmcia/
94•rbanffy•4h ago•24 comments

Waiting for dawn in search: Search index, Google rulings and impact on Kagi

https://blog.kagi.com/waiting-dawn-search
152•josephwegner•3h ago•101 comments

Letting Claude Play Text Adventures

https://borretti.me/article/letting-claude-play-text-adventures
25•varjag•5d ago•2 comments

Slouching Towards Bethlehem – Joan Didion (1967)

https://www.saturdayeveningpost.com/2017/06/didion/
34•jxmorris12•3h ago•1 comments

Stanford scientists found a way to regrow cartilage and stop arthritis

https://www.sciencedaily.com/releases/2026/01/260120000333.htm
182•saikatsg•3h ago•47 comments

JPEG XL Test Page

https://tildeweb.nl/~michiel/jxl/
133•roywashere•4h ago•98 comments

Tell HN: Bending Spoons laid off almost everybody at Vimeo yesterday

283•Daemon404•5h ago•221 comments

Autonomous (YC F25) is hiring – AI-native financial advisor at 0% advisory fees

https://atg.science/
1•dkobran•4h ago

Claude's New Constitution

https://www.anthropic.com/news/claude-new-constitution
159•meetpateltech•5h ago•106 comments

Can you slim macOS down?

https://eclecticlight.co/2026/01/21/can-you-slim-macos-down/
123•ingve•13h ago•175 comments

Show HN: Rails UI

https://railsui.com/
64•justalever•2h ago•49 comments

Show HN: Company hiring trends and insights from job postings

https://jobswithgpt.com/company-profiles/
22•sp1982•3h ago•4 comments

SmartOS

https://docs.smartos.org/
151•ofrzeta•6h ago•61 comments

Nested Code Fences in Markdown

https://susam.net/nested-code-fences.html
162•todsacerdoti•8h ago•45 comments

Show HN: See the carbon impact of your cloud as you code

https://dashboard.infracost.io/
45•hkh•6h ago•12 comments

Without benchmarking LLMs, you're likely overpaying

https://karllorey.com/posts/without-benchmarking-llms-youre-overpaying
115•lorey•1d ago•66 comments

I Made Zig Compute 33M Satellite Positions in 3 Seconds. No GPU Required

https://atempleton.bearblog.dev/i-made-zig-compute-33-million-satellite-positions-in-3-seconds-no...
131•signa11•11h ago•16 comments

RTS for Agents

https://www.getagentcraft.com/
94•summoned•5d ago•40 comments

EU–INC – A new pan-European legal entity

https://www.eu-inc.org/
666•tilt•10h ago•628 comments

Deaths Linked to AI Chatbots

https://en.wikipedia.org/wiki/Deaths_linked_to_chatbots
15•sieep•39m ago•0 comments

EmuDevz: A game about developing emulators

https://afska.github.io/emudevz/
174•ingve•3d ago•39 comments

Show HN: yolo-cage – AI coding agents that can't exfiltrate secrets

https://github.com/borenstein/yolo-cage
41•borenstein•6h ago•60 comments

Ask HN: How are you automating your coding work?

39•manthangupta109•2h ago•37 comments

Batmobile: 10-20x Faster CUDA Kernels for Equivariant Graph Neural Networks

https://elliotarledge.com/blog/batmobile
83•ipnon•3d ago•11 comments