DirectoryVllm Mlx MCP
61
Quality Score
OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX backend, 400+ tok/s. Works with Claude Code.
520stars
70forks
Python
Updated March 11, 2026
#anthropic#apple-silicon#audio-processing#claude-code#computer-vision#image-understanding#inference#llm#machine-learning#macos#mllm#mlx#multimodal-ai#speech-to-text#stt#text-to-speech#tts#video-understanding#vision-language-model#vllm
Installation
npx vllm-mlx
Server Info
Categorytools
LanguagePython
LicenseN/A
StatusACTIVE
Submit Your Server
Built an MCP server? Get it listed in front of thousands of developers.
List Your Server Free