Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
Pythonjundot/omlx
omlx
LLM inference server with continuous batching & SSD caching for Apple Silicon — managed from the macOS menu bar