2025 is the year of the terminal, apparently?
For my prototype purposes, it’s great, and Claude code the most fun I’ve had with tech in a jillion years.
Sure, you can have your LLM code with any JavaScript framework you want, as long as you don't mind it randomly dropping React code and React-isms in the middle of your app.
LLMs are already trained on JavaScript at a deep level; as LLM reasoning and RAG techniques improve, there will be a time in the not-too-distant future when an LLM can be pointed to the website of a new framework and be able to use it.
https://www.reddit.com/r/cursor/comments/1mk8ks5/discussion_...
https://docs.anthropic.com/en/docs/claude-code/llm-gateway#l...
is there a way to make it more verbose?
https://agent.md [redirect -> https://ampcode.com/AGENT.md] https://agent-rules.org
they also suggest using symlinks for now
Claude Code likes to add "attribution" in commit messages, which is just pure spam.
SICP contains the famous quote: “Programs must be written for people to read, and only incidentally for machines to execute.”
The whole agentic coding via CLI experience could be much improved by:
- Making it easy to see what command I last issued, without having to scroll up through reams of output hunting for context - Making it easy to spin up a proper sandbox to run sessions unattended - Etc.
Maybe for code generation, what we actually need is a code generator that is itself deterministic but uses AI, instead of AI that does code generation.
Till then you can also use symlinks
there are issues opened in some repos for this
- Support "AGENT.md" spec + filename · Issue #4970 · google-gemini/gemini-cli
https://github.com/google-gemini/gemini-cli/issues/4970#issu...
Here for Claude
cat AGENT.md | claude
IIRC this saves some tokens.
Why are we purposely creating CLI dialects?
Yesterday, I was writing about a way I found to pass the same guideline documents into Claude, Gemini, and Aider CLI-coders: https://github.com/sutt/agro/blob/master/docs/case-studies/a...
I guess having links to supplementary rules files is an option, but I'm not sure which agents (if any) would work well with that.
How is the tool supposed to merge multiple md files?
However, if you look through the source code or network requests, you’ll see that merging just means naive “concatenation”.
Not really: our AI agents are probably smart enough to even make sense of somewhat bad instructions.
Eg if you write your instructions in a mixture of base 64, traditional Chinese, morse code and Polish, the LLM will still figure it out.
I am talking about LLMs figuring out how to build your project with some bad and incomplete instructions plus educated guessing.
An agent on the other hand, one who is in that sweet spot where they're no longer ignorant, and not yet confused... It's nice to have them dump their understanding to agent_primers/subsystem_foo.md for consumption by the next agent that touches that subsystem. I don't usually even read these until I suspect a problem in one. They're just nuggets of context transfer.
e.g. maybe for CURSOR.md you just want to provide context and best practices without any tool-calling context (because you've found it doesn't do a great job of tool-calling), while for CLAUDE.md (for use with Claude Code) you might want to specify tools that are available to it (because it does a great job with tool calling).
Probably best if you have an AGENT.md that applies to all, and then the tools can also ingest their particular flavor in addition, which (if anything is in conflict) would trump the baseline AGENT file.
AGENT.md
AGENT.CLAUDE.md
They get applied AGENT first, then AGENT.CLAUDE. You are able to specify agent specific instructions in the AGENT.md: @agent-disable CLAUDE
These instructions are not parsed by claude
@agent-enable CLAUDE---
This project uses shared planning documents for collaboration with Claude Code. Please:
1. First read and understand these files:
- PLAN.md - current project roadmap and objectives
- ARCHITECTURE.md - technical decisions and system design
- TODO.md - current tasks and their status
- DECISIONS.md - decision history with rationale
- COLLABORATION.md - handoff notes from other tools
2. Before making any significant changes, check these documents for:
- Existing architectural decisions
- Current sprint priorities
- Tasks already in progress
- Previous context from Claude Code
3. After completing work, update the relevant planning documents with:
- Task completion status
- New decisions made
- Any changes to architecture or approach
- Notes for future collaboration
Always treat these files as the single source of truth for project state.And then it will promptly forget about CLAUDE.md as well (happened to me on several occasions)
@AGENTS.md
Still messy, but at least it means it's using the same content ln -s AGENTS.md CLAUDE.md # repeat for all
?Side remark: CC is very expensive when using API billing (compared to e.g. GPT-5). Once a company adopts CC and all developers start to adapt to it at full scale, the bill will go out of the roof.
> Set your own rules: Customize Cursor's work with rules, AGENTS.md, and MCP.
There's no mention of it in the docs, though. It's also interesting it's AGENTS.md on that page instead of AGENT.md, I wonder if that's a typo.
I'd rather .agents/claude or something so we can get these files out of the root directory, which, at least for typescript projects, is already a confetti-like jumble of json files for every little "simple" tool that needs its own configuration file.
I get why package.json isn't enough. But would a .config directory standard really have hurt us so much?
Old IDEs were built for the same purpose generally, but prioritized different weaknesses.
p.s: looking forward for crush :3
I guess Cursor makes sense for people who only use LLMs for coding.
I pay for Cursor and ChatGPT. I can imagine I’d pay for Gemini if I used an android. The chat bots (1) won’t keep the subscription competitive with APIs because the cost and usage models are different and (2) most chat bots today are more of a UX competition than model quality. And the only winners are ChatGPT and whatever integrated options the user has by default (Gemini, MSFT Copilot, etc).
If Cursor can build the better UX for all the use-cases, mobile/desktop chatbot, assistant, in IDE coding agent, CLI coding agent, web-based container coding agent, etc.
In theory, they can spend all their resourcing on this, so you could assume they could have those be more polished.
If they win the market-share here, than the models are just commodity, Cursor lets you pick which ever is best at any given time.
In a sense, "users" are going to get locked in on the tooling. They learn the commands, configuration, and so on of Cursor, it's a higher cost for them to re-learn a different UX. Uninstalling and re-installing another app, plugin, etc. is annoying.
The models are just going to be fighting performance/cost. And people will choose the best performance for their budget.
And that's ignoring how good local models are getting as well.
It's not that they'll have their launch eaten by Cursor, it's just that they can't be as focused on user experience when they're also laser focused on improving the models to stay competitive.
Models are commodities.
1. with tight integration between cli, background agent, ide, github apps (e.g. bugbot), cursor will accommodate the end-to-end developer experience.
2. as frontier models internalize task routing, there won't be much that feels special about claude code anymore.
3. we should always promote low switching costs between model providers (by supporting independent companies), keeping incentives toward improving the models not ui/data/network lock-in.
Unless they pair up with OpenAI or Meta.
cursor and 3rd party tools will, unless they make their own superior foundation model, will always have to fight the higher marginal cost battle. This is particularly bad insofar that they offer fixed pricing subscriptions. That means they’re going to have to employ more context saving tricks which are at odds with better performance.
If the cost economics result in Cursor holding, say, 20% fewer tokens in context versus model-provider coding agents, they will necessarily get worse performance, all things equal.
Unless Cursor offers something dramatically different outside of the basic agentic coding stack it’s hard to see why the market will converge to cursor.
You’re underestimating the dollars at play here. With cursor routing all your tokens, they will become a foundation model play sooner than you may think
I would _never_ give an LLM access to any disk I own or control if it had anything more than read permissions
I'm a few degrees removed from an air gapped environment so obviously YMMV. Frankly I find the idea of an LLM writing files or being allowed to access databases or similar cases directly distasteful; I have to review the output anyway and I'll decide what goes to the relevant disk locations / gets run.
If you were being really paranoid then I guess they could write a script in the local directory that then runs and accesses other parts of the filesystem.
I've not seen any evidence an agent would just do that randomly (though I suppose they are nondeterministic). In principle maybe a malicious or unlucky prompt found somewhere in the permitted directory could trigger it?
[1]: https://github.com/google-gemini/gemini-cli/blob/main/docs/c...
GPT-5 seems a bit slow so far (in terms of deciding and awareness). I’ve gone from waiting for a compiler, to waiting for assets to build to now waiting for an agent to decide what to do - progress I guess :)
Cursor on the other hand, especially with GPT-5 today but typically with Sonnet 4.1, has been a workhorse at my company for months. I have never had Claude Code complete a meaningful ticket once. Even a small thing like fixing a small bug or updating the documentation on the site.
Would love any tips on how to make Claude Code not a complete waste of electricity.
I think you're right.
People getting really poor results probably don't recognize that their prompts aren't very good.
I think some users make assumptions about what the model can't do before they even try, so their prompts don't take advantage of all the capabilities the model provides.
I don’t doubt I could improve my prompts but I don’t have those same prompting problems with cursor.
You probably mean Opus 4.1; there's no Sonnet 4.1 yet.
I seem to always have better outcomes with Claude code.
The Cursor you used a month ago is not the one you get now.
Just saying that because in this space you should always compare latest X with latest Y.
I too switched weeks ago to Claude Code. Then between the times I am out of tokens I launch Cursor and actually find it...better than I remember if not on par with Claude Code (the model and quality of prompts/context matters more than the IDE/CLI tool used too).
Git
This may not work for everyone, but as a solo dev who wants to keep a real mental model of my work (and not let it get polluted with AI slop), the Claude Code approach just works really well for me. It's like having a coding partner who can iterate and change direction as you talk, not a junior dev who dumps a pile of code on your plate without discussion.
Larger context is a bonus sometimes, but in general you're degrading the quality of the output by a lot.
Precise prompting and context management is still very important.
https://x.com/OpenAIDevs/status/1953559797883891735 (0.19 now)
I don’t visit Twitter links. Why not a link to the GitHub changelog?
Also, as an aside since you are on the team - the organization verification is frustrating in that the docs indicate:
>You must not have recently verified another organization, as each ID can only verify one organization every 90 days.
I champion OpenAI at my work, so naturally I’d be the one to verify there. But I apparently can’t, because I verify for my personal-led org. That gets in the way of me proselytizing gpt-5 based coding tools (such as, possibly, Codex CLI).
loving the animations and todos so far
also gpt-5 is just great at agentic stuff
Codex needs plan mode (shift-tab in Claude Code)
And Codex needs the prompt to always be available. So you can type to the model while it’s working & have it eventually receive the message and act on it, instead of having to Ctrl-C to interrupt it before typing. Claude Code’s prompt is always ready to type at - you can type while it is working. That goes a long way towards it feeling like it cares about the user.
It's asking me to buy credits but I'm already on Plus?
Getting this error:
Error response Error code: 500
Message: Token exchange failed: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1006)>.
Error code explanation: 500 - Server got itself in trouble.
https://github.com/openai/codex/issues/1243
I am and many other similar users would love to move from Claude Code to Codex, but this is preventing us from doing it.
This all just feels profoundly immature. You tell us one thing then two months later you're on to the next thing. There's no sense of mastery here.
Who would turn loose arbitrary commands (content)
generated by an LLM onto their filesystem?
Then I saw the installation instructions, which are: curl https://cursor.com/install -fsS | bash
And it made sense.Only those comfortable with installing software by downloading shell commands from an arbitrary remote web site and immediately executing them would use it.
So what then is the risk of running arbitrary file system modifications generated from a program installed via arbitrary shell commands? None more than what was accepted in order to install it.
Both are opaque, unreviewed, and susceptible to various well known attacks (such as a supply chain attack[0]).
I find in ide they like opening documents/changing tabs too much and it means j can't do other things.
opencode and Crush can use any model, so apart from a nicer visual experience, are there any aspects that actually make you more productive in one vs the other?
Are those for the anonymous accesses of the AI prompts?
If those are for the authenticated AI prompts, how to create a "non-anonymous" account with a noscript/basic (x)html browsers (not to mention I am self-hosted without paying the DNS mafia, namely my emails are with ip literals, ofc I prefer IPv6).
Why shouldn't you be able to use the abilities of this tool as a batch command, connected with all your other basic tools, in addition to interactive sessions?
Cursor's chat being locked in an IDE sidebar has felt like driving with a trailer attached. For some tasks the editor is secondary or unnecessary, and as a papered-over VS Code fork, Cursor has a lot of warts that you just had to accept. Now you can just use your favorite editor.
Companies make apps but want to be platforms, so they try to put everything in one app and help you forget about everything else. VS Code and Figma, for example, make their own extension ecosystems rather than connecting outward, because it makes them platforms-as-apps and harder to leave. But a desktop task workflow spans many apps and windows. You compose it yourself to your needs. We are computer users more than app users.
To me as a computer user, a tool that's compact and has compatible outward extension points feels good.
AI generated startup + AI generated blog (https://www.nxgntools.com/) doesn't have much to do with HN
unsupp0rted•6mo ago
sblawrie•6mo ago
zaphirplane•6mo ago
NitpickLawyer•6mo ago
But if you take speed/availability/cost into account, there might be "better" offers out there. I did some tests w/ windsurf when they announced their swe1 and swe1-lite models, and the -lite could handle easy tasks pretty well. I also tested 4.1-mini and 4.1-nano. There are tasks that I could see them handle reliably enough to make sense (and they're fast, cheap and don't throttle you).
alwillis•6mo ago
[1]: https://github.com/musistudio/claude-code-router
rvnx•6mo ago
jstummbillig•6mo ago
jonplackett•6mo ago
jstummbillig•6mo ago
_betty_•6mo ago
anthonypasq•6mo ago
bangaladore•6mo ago
With the benefit that you can also pull in people who don't like using VSCode such as people who use Jetbrains or terminal based code editors.
gorjusborg•6mo ago
The IDE/editor is for me, the agent doesn't need it. That also means I am not forced to used whatever imperfect forked IDE the agent is implemented against.
worldsayshi•6mo ago
Because the agents aren't yet good enough for a hands off experience. You have to continuously monitor what it does if you want a passable code base.
tsvetkov•6mo ago
imp0cat•6mo ago
apwell23•6mo ago
Following are now stars of my workflow
* Git plugins - Diffview, gitsigns, fugitive
* Claude Code plugin / Terminals with claude code
* Neovim sessions
* Git worktrees
Editing focused workflows have taken an backseat
* LSP
* Vim motion and editing efficiency
* File navigation
* Layouts
ricericerice•6mo ago
worldsayshi•6mo ago
stavros•6mo ago
Xenoamorphous•6mo ago
stavros•6mo ago
Programming has changed from writing code to reviewing/QAing and reprompting, but the tooling hasn't yet caught up with that workflow. We need Gerrit for coding agents, basically.
fooster•6mo ago
Touche•6mo ago
dagss•6mo ago
I would much rather use IntelliJ so perhaps my habits will change at some point, but right now I am stuck with Cursor/vscode for the tab completion.
nojs•6mo ago
ygouzerh•6mo ago
I haven't found however if Cursor cli provides this kind of extension