We have a similar wrapper for local LLMs on the roadmap.
If you use CLI only - we run claude 4 + gemini on the backend, gemini serving most of the vision tasks (frontend validation) and claude doing core codegen.
We have baseline cursor rules included in case you want to hack on this manually https://github.com/appdotbuild/agent/tree/main/agent/trpc_ag...
Where we are tied is the LLM provider - you will need to supply your own keys for Anthropic / Gemini.
We did a couple runs on top of Ollama + Gemma - expect support for local LLMs. Can't swear on the timeline, but one of our core contributors recently built a water cooled rig with a bunch of 3090s so my guess is "pretty soon".
davidgomes•1d ago
- https://github.com/appdotbuild/agent
- https://github.com/appdotbuild/platform
And we also blogged[1] about how the whole thing works. We're very excited about getting this out but we have a ton of improvements we'd like to make still. Please let us know if you have any questions!
[1]: https://www.app.build/blog/app-build-open-source-ai-agent
ivape•1d ago
davidgomes•1d ago