Whenever I see that I think about whether I can find a good point to compact or clear. I also just try to clear whenever it makes sense to avoid getting there and try to give smaller tasks that can be cleared after they're done when possible.
Oh, I guess one thing I do is sometimes have it write a file with what was done, if I'm not actually sure if I want to clear or might want to come back to it. I also sometimes do this rather than compact during a large task - document status and clear.
I wish there were a way to persist in-memory context in a file automatically, say on each compact or git commit. Yesterday CC crashed and restarting it and feeding it all the context was a pain since my updated Claude.md file was a couple of days old. It literally went from a Sr Engineer to a Jr post crash.
A couple days ago I was getting so many api errors/timeouts I decided to upgrade from the $20 to the $100 plan (as I was also regularly hitting rate limits as well)
It seemed to fix the issue immediately. But today, the errors came back for about half an hour
Pretty rare to get a 529 outside of that time window in my personal experience, at least during the USA day.
Hopefully they work out whatever issue is going on.
(Just to be clear, I have no idea what on this thread to take seriously and not and who is. I'm joking at least.)
A great deal of claude stupidity is due to context engineering, specifically due to the fact that it tries its hardest to pick out just the slice of code it needs to fulfill the task.
A lot of the annoying “you’re absolute right!” come from CC incrementally discovering that you have more than 10 lines of code in that file that pertains to your task.
I don’t believe conspiracies about dumbed down models. Its all context pruning.
The model feels like it has got stupid when you get on a cold streak after a hot hand.
I did a test with a very detailed prompt, exactly specified what to fix and how. Claude did it, but not very well. Gemini? it got stuck in a loop until i told it to stop, gave it a hint and then it got stuck again and gave up after trying the exact same thing three more times…
And while Claude managed to get through it, it couldn’t get it right even with some help. It took me 15 minutes to write the prompt, 15 minutes of claude implementing it & another 10 trying to get it to do it correctly. It would have taken me about half the time to do it myself i think..
I am giving up on it for a while.
For example, a sub-agent for adding a new stat to an RPG. It could know how to integrate with various systems like items, character stats component, metrics, and so on without having to do as much research into the codebase patterns.
the main problem I have is that the agents just aren't used
For example, I set up a code reviewer agent today and then asked claude to review code, and it went off and did it by itself without using the agent
in one of anthropic's own examples they are specifically telling claude which agents to use which is exactly what I don't want to have to do:
> First use the code-analyzer sub agent to find performance issues, then use the optimizer sub agent to fix them
My working theory is that while Claude has been extensively trained on tool use and is often eager to use whatever tools are available, agents are just different enough that they don't quite fit - maybe asking another agent to do something "feels" very close to asking the user to do something, which is counter to their training
but maybe I just haven't spent enough time trying it out and tweaking the descriptions
So if you have a code review agent or a tdd agent checking the current commit if it matches some specs you have, they'll start a separate "subprocess" with its own context and return whatever they find to the main Claude context.
You don't need a full agent library to write LLM workflows.
Rather: A general purpose agent with a custom addition to the system prompt can be instructed to call other such agents.
(Of course explicitly mamaging everything is the better choice depending on your business case. But i think it would be always cheaper to at least build a prototype using this method.)
That's why I've been using agro to launch agents from each of the main LLM vendors and checking their results when I'm stuck: https://github.com/sutt/agro/blob/master/docs/index.md
I'm working on a proper config screen for them that just modifies the agent files directly, and a future release will also give special formatting for agent output.
It contains 100+ specialized agents covering the most requested development tasks - frontend, backend, DevOps, AI/ML, code review, debugging, and more. All subagents follow best practices and are maintained by the open-source framework community. Just copy to .claude/agents/ in your project to start using them.
T0Bi•6mo ago
¹ https://github.com/ruvnet/claude-flow
nazgul17•6mo ago
dchuk•6mo ago
himeexcelanta•6mo ago
SOLAR_FIELDS•6mo ago
lubujackson•6mo ago
AbstractH24•6mo ago
jampa•6mo ago
> [...]
> # 2. Activate Claude Code with permissions
> claude --dangerously-skip-permissions
Bypassing all permissions and connecting with MCPs, can't wait for "Claude flow deleted all my files and leaked my CI credentials" blog post
data-ottawa•6mo ago
I’ve set it up bespoke but the auth flow gets broken.
oarsinsync•6mo ago
T0Bi•6mo ago
¹ https://github.com/anthropics/claude-code/tree/main/.devcont...
SOLAR_FIELDS•6mo ago
beefnugs•6mo ago
The tooling required for any of this to approach actual engineering reliability levels is unbelievable really
T0Bi•6mo ago
I use the .devcontainer¹ from the claude-code repository. It works great with VSC and let's you work in your docker container without any issues. And as long as you use some sort of version control (git) you cannot really lose anything.
¹ https://github.com/anthropics/claude-code/tree/main/.devcont...
dazzaji•6mo ago
lvl155•6mo ago
Bro…