I built Isogen which uses as much memory as a Chrome tab instead of a tool that can spike up to 1GB of RAM. You drag and drop or paste files into an isolated context and do fast generations file-by-file. This approach allows me to keep a strong mental model of the codebase. I also added a snapshot feature that keeps the history of the files and the generated output. File copies are saved locally with SQLite.
Bring Your Own Key for inference which allows for unlimited generations. The only models supported now are the fast ones from Gemini, ChatGPT, Claude and Grok.