Hey HN! I've been building MLX-GUI as an open-source inference server that turns any Mac into a multi-user AI server. v1.2.4 just shipped with some major additions:
Complete Whisper ecosystem (99+ languages, word timestamps, any audio format)
23 embedding models across 13 families (E5, ModernBERT, Arctic, etc.)
Mistral Small 24B with vision capabilities
OpenAI-compatible API that's actually faster than Ollama on Apple Silicon
The goal was simple: I wanted to use my Mac Mini/Studio as proper inference servers without the complexity of managing Python environments or paying for cloud APIs while keeping data local.
It's packaged as a native macOS app (no Python install needed) with a beautiful web GUI for model management. The API is drop-in compatible with OpenAI, so existing apps like Jan.ai work immediately.
900+ lines of tests ensure production reliability. G
RamboRogers•7h ago
Complete Whisper ecosystem (99+ languages, word timestamps, any audio format) 23 embedding models across 13 families (E5, ModernBERT, Arctic, etc.) Mistral Small 24B with vision capabilities OpenAI-compatible API that's actually faster than Ollama on Apple Silicon
The goal was simple: I wanted to use my Mac Mini/Studio as proper inference servers without the complexity of managing Python environments or paying for cloud APIs while keeping data local. It's packaged as a native macOS app (no Python install needed) with a beautiful web GUI for model management. The API is drop-in compatible with OpenAI, so existing apps like Jan.ai work immediately. 900+ lines of tests ensure production reliability. G
NU GPL v3 licensed and actively maintained. GitHub: https://github.com/RamboRogers/mlx-gui
Would love feedback from the community - especially on the embedding pipeline and audio processing!
jki275•7h ago