Right now there is no set of proper benchmarks from us. In general we are faster than llama.cpp and on par with MLX. In a next release we will be faster than both.
Right now there is no set of proper benchmarks from us. In general we are faster than llama.cpp and on par with MLX. In a next release we will be faster than both.