micytao/vllm-mcp-server
3.3
If you are the rightful owner of vllm-mcp-server and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to dayong@mcphub.com.
The vLLM MCP Server is a Model Context Protocol server that provides vLLM capabilities to AI assistants and MCP-compatible clients.
Tools
Functions exposed to the LLM to take actions
vllm_chat
Send chat messages to vLLM with multi-turn conversation support.
vllm_complete
Generate text completions.
list_models
List all available models on the vLLM server.
get_model_info
Get detailed information about a specific model.
vllm_status
Check the health and status of the vLLM server.
Prompts
Interactive templates invoked by user choice
No prompts
Resources
Contextual data attached and managed by the client