DirectoryVllm Mlx MCP
60
Health
good
61
Popularity
OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX backend, 400+ tok/s. Works with Claude Code.
520stars
70forks
Python
Updated March 11, 2026
#anthropic#apple-silicon#audio-processing#claude-code#computer-vision#image-understanding#inference#llm#machine-learning#macos#mllm#mlx#multimodal-ai#speech-to-text#stt#text-to-speech#tts#video-understanding#vision-language-model#vllm
Add to Claude Desktop
Paste this into your claude_desktop_config.json file:
{
"mcpServers": {
"vllm-mlx-mcp": {
"command": "npx",
"args": [
"-y",
"vllm-mlx"
]
}
}
}Config file location: ~/Library/Application Support/Claude/claude_desktop_config.json (macOS) or %APPDATA%\Claude\claude_desktop_config.json (Windows)
Installation
npx vllm-mlx
Maintenance Healthgood
Freshness32/40
Issue response15/30
Org backing10/20
Longevity3/10
› Updated recently
Server Info
Categorytools
LanguagePython
LicenseN/A
StatusACTIVE
Submit Your Server
Built an MCP server? Get it listed in front of thousands of developers.
List Your Server Free