The GUI AI model manager for your local llama.cpp engine.
A fast, lightweight orchestrator built exclusively for the llama.cpp engine. OMM automates model lifecycles and VRAM unloading, exposing a local, OpenAI-compatible API so your favorite chat GUIs can talk directly to llama.cpp.
;