On top of that, I would add that most async-to-batch libraries force users to learn a new framework or refactor their existing code, which is a huge friction in itself.
I've been in those trenches as a developer and I decided to create a literal 2-liner python lib that gets you from async to batch: https://github.com/vienneraphael/batchling
You don't need to change your code and it supports most providers (Anthropic, Gemini, Groq, Mistral, OpenAI, Together, Vertex, XAI, Doubleword) and all imagineable python frameworks (Langchain, PydanticAI, Instructor, DSPy, LiteLLM, Pydantic Evals, ..)
alexliu79•1h ago
I’m curious how people here are handling this in practice for evals, extraction pipelines, classification jobs, or other multi-step workflows.
Are you using batch APIs already? Building internal tooling? Or just accepting the extra cost because batch workflows are too painful to adopt?
We’ve been building an open-source library called ParaLLeM to make it easier to move agent workflows from sync to batch without rewriting everything, and I’d love to understand how others are approaching this problem.
Repo: https://github.com/parallem-ai/parallem