I wrote a small blogpost on a little experiment I did last week-end. The goal was to see if I could get more tok/s performance from llama.cpp running the latest Qwen3.5 models. A 5.5x perf increase was achieved by 1. recompiling with optimization flags 2. using ik_llama.cpp!
The build I'm using is the 750 USD rig from Digital Spaceport.
Very cool to see more than usable speeds on a cheap setup like this.
DAFtwinTurbo•1h ago
I wrote a small blogpost on a little experiment I did last week-end. The goal was to see if I could get more tok/s performance from llama.cpp running the latest Qwen3.5 models. A 5.5x perf increase was achieved by 1. recompiling with optimization flags 2. using ik_llama.cpp!
The build I'm using is the 750 USD rig from Digital Spaceport.
Very cool to see more than usable speeds on a cheap setup like this.