you can skim the transcript but some personal highlights:
- anthropic employees, with unlimited claude, average to $6/day of usage
- headless claude code as a "linux" utility that you use everywhere in CI is pretty compelling
- claude code as a user extensible platform
- future roadmap of claude code: sandboxing, branching, planning
- sonnet 3.7 as a persistent, agentic model
For a tool that radically increases productivity (say 2x), I think it could still make sense for a VC funded startup or an established company (even $100/day or $36k/year is still a lot less than hiring another developer). But for a side project or bootstrap effort, $36k/year obviously significantly increases cash expenses. $100/month does not, however.
So, I'm going to go back and upgrade to Max and try it again. If that keeps my costs to $100/month, thats a really different value proposition.
Edit:
Found the answer to my own questions
> Send approximately 50-200 prompts with Claude Code every 5 hours[1]
Damn. That's a really good deal
[1] https://support.anthropic.com/en/articles/11145838-using-cla...
From the link:
"Apparently, there are some engineers inside of Anthropic that have spent >$1,000 in one day!"
The question is what is the P50, P75, and P95 spend per employee?
† simonw, gwern
The "golden" end state of coding agents is that you give it a Feature Request (EG Jira ticket), and it gives you a PR to review and give feedback on. Cursor, windsurf, etc, are dead ends in that sense as they are local editors, and can not be in CI.
If you are tooling your codebase for optimal AI usage (Rules, MCP, etc), you should target a technology that can bridge the gap to headless usage. The fact Claude Code can trivially be used as part of automation through the tools means it's now the default way I thinking about coding agents (Codex, the npm package, is the same).
Disclaimer, I focus on helping companies tool their codebases for optimal agent usage, so I might have a bias here to easily configurable tools.
I see your point but in the other hand how depressing to be left only with the most soul crushing part of software entering - the Jira ticket.
I understand the craft of code itself is what some people love though!
We will drop the narrow-minded deadweight that can only collect naive requirements, and the coding side that can only implement unambiguous tickets.
That's what I want and look forward one day
I hate using voice for anything. I hate getting voice messages, I hate creating them. I get cold sweats just thinking about having to direct 10 AI Agents via voice. Just give me a keyboard and a bunch of screens, thanks.
And don’t get me started on video vs text for learning purely non-physical stuff like programming…
sending audio = fast
That's exactly what everyone is hoping for. Well, everyone except software engineers, of course
If the cost of writing software goes down, demand for it will presumably go up...
All those skills can be applied to engineering as well. What makes Fabrice Bellard great? Its not just technical skill I think.
I think some of the most successful people will be a subset of engineers but also Steve Jobs types and artists
Isn’t that effectively the promise of the most recently released OpenAI codex?
From the reviews I’ve been able to find so far though, quality of output is ehh.
I bias a bit to wanting the agent to be a pluggable component into a flow I own, rather than a platform in a box.
It'll be interesting to see where the different value props/use cases of a Delvin/v0 vs a Codex Cloud vs Claude Code/Codex CLI vs Cursor land.
Put the Aider CLI into a GitHub action that's triggered by an issue creation and you're good to go.
Like Anthropic and most big tech companies, they don't want to show off the best until they need to. They used to stockpile some cool features, and they have time to think about their strategy. But now I feel like they are in a rush to show off everything and I'm worried whether the management has time to think about the big picture.
The only possible way for this to be a successful offering is if we have just now reached a plateau of model effectiveness and all foundation models will now trend towards having almost identical performance and capabilities, with integrators choosing based on small niceties, like having a familiar SDK.
At this point Claude Code is a software differentiator in the agent coding space.
I am building things related to AI code assistants - we were hacking ways to integrate Claude Code - it was the first thing we wanted to build around.
It's too early to care about lock in.
Need the best, will only build around the best.
Honestly though, CLI tools for accessing LLMs (including piping content in and out of them) is such a clearly good idea I'm glad to see more tools implementing the pattern.
Can somebody please tell me what software product or service doesn’t compete with general intelligence?
Imagine selling intelligence with a legal term that, under strict interpretation, says you’re not allowed to use it for anything.
Is it so vague it’s unenforceable?
How do we own the output if we can’t use it to compete with a general intelligence?
Is it just a “lol nerd no one cares about the legal terms” thing? If no one cares then why would they have a blanket prohibition on using the service ?
We’re supposed to accept liability to lose a lawsuit just to accept their slop? So many questions
[0] https://aider.chat/docs/scripting.html
[1] https://aider.chat/docs/recordings/tree-sitter-language-pack...
Add a file to your repo and you can talk to any model via issues.
I don't really want it committing and stuff, i mostly like the UX of Claude Code. Thoughts?
[0]: https://docs.anthropic.com/en/docs/claude-code/github-action...
As it only accepts an API key as far as I can tell.
andrewstuart•2h ago
barefootford•2h ago
andrewstuart•2h ago
Works for a reasonable chunk of files say 5 to 10 that aren’t too big.
No doubt they’ll get to better file access.
Anyhow I’m quite happy to do the copy and paste because Geminis coding and debugging capability is far better than Claude.
danenania•2h ago
The default planning/coding models are still Sonnet 3.7 for context size under 200k, but you can switch to Gemini with `\set-model gemini-preview`.
1 - https://github.com/plandex-ai/plandex
mickeyp•2h ago
If you (or anyone else reading this) wants to try out the upcoming beta give me a ping. (see profile.)
Sajarin•2h ago
I think Bard (lol) and Gemini got a late start and so lots of folks dismissed it but I feel like they've fully caught up. Definitely excited to see what Gemini 3 vs GPT-5 vs Claude 4 looks like!
Karrot_Kream•1h ago
fallinditch•1h ago
I suspect that I experience some performance throttling with Gemini 2.5 in my Windsurf setup because it's just not as good as anecdotal reports by others, and benchmarks.
I also seem to run up against a kind of LLM laziness sometimes when they seemingly can't be bothered to answer a challenging prompt ... a consequence of load balancing in action perhaps.
mbesto•1h ago
EDIT: Specifically: https://openrouter.ai/rankings/programming?view=week
ChadMoran•1h ago
simonw•1h ago
andrewstuart•1h ago
cube2222•54m ago
Gemini 2.5 Flash on the other hand has excellent. I’ve started using it to rewrite whole files after talking the changes through with Claude, because it’s just so ridiculously fast (and dependable enough for applying already outlined changes).
ramoz•48m ago
The two work really well with Gemini as a planner and Claude Code as an executor.