alt.hn

4/7/2026 at 3:24:45 AM

MLX-Serve a Native LLM Runtime for Apple Silicon

https://ddalcu.github.io/mlx-serve/

by ddalcu

4/7/2026 at 3:24:45 AM

Native Zig server that runs MLX-format language models on Apple Silicon and exposes an OpenAI-compatible HTTP API. No Python.

Optional app MLX Claw, a macOS menu bar app with built-in chat, agent mode, and model management.

No dependencies 34MB, very low ram usage compared to other LLM runners.

by ddalcu