vllm-mcp-server

micytao/vllm-mcp-server

3.3

If you are the rightful owner of vllm-mcp-server and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to dayong@mcphub.com.

The vLLM MCP Server is a Model Context Protocol server that provides vLLM capabilities to AI assistants and MCP-compatible clients.

Tools

Functions exposed to the LLM to take actions

vllm_chat

Send chat messages to vLLM with multi-turn conversation support.

vllm_complete

Generate text completions.

list_models

List all available models on the vLLM server.

get_model_info

Get detailed information about a specific model.

vllm_status

Check the health and status of the vLLM server.

Prompts

Interactive templates invoked by user choice

No prompts

Resources

Contextual data attached and managed by the client

No resources