DirectoryVllm Mlx MCP
tools

Vllm Mlx MCP

by waybarrios

60
Health
good
61
Popularity

OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX backend, 400+ tok/s. Works with Claude Code.

520stars
70forks
Python
Updated March 11, 2026
#anthropic#apple-silicon#audio-processing#claude-code#computer-vision#image-understanding#inference#llm#machine-learning#macos#mllm#mlx#multimodal-ai#speech-to-text#stt#text-to-speech#tts#video-understanding#vision-language-model#vllm

Add to Claude Desktop

Paste this into your claude_desktop_config.json file:

{
  "mcpServers": {
    "vllm-mlx-mcp": {
      "command": "npx",
      "args": [
        "-y",
        "vllm-mlx"
      ]
    }
  }
}

Config file location: ~/Library/Application Support/Claude/claude_desktop_config.json (macOS) or %APPDATA%\Claude\claude_desktop_config.json (Windows)

Installation

npx vllm-mlx

Maintenance Healthgood

Freshness32/40
Issue response15/30
Org backing10/20
Longevity3/10
Updated recently

Server Info

Categorytools
LanguagePython
LicenseN/A
StatusACTIVE

Submit Your Server

Built an MCP server? Get it listed in front of thousands of developers.

List Your Server Free
Back to Directory