MCP vLLM Benchmarking Tool
This is proof of concept on how to use MCP to interactively benchmark vLLM.
We are not new to benchmarking, read our blog:
Benchmarking vLLM
This is just an exploration of possibilities with MCP.
Usage
- Clone the repository
- Add it to your MCP servers:
{
"mcpServers": {
"mcp-vllm": {
"command": "uv",
"args": [
"run",
"/Path/TO/mcp-vllm-benchmarking-tool/server.py"
]
}
}
}
Then you can prompt for example like this:
Do a vllm benchmark for this endpoint: http://10.0.101.39:8888
benchmark the following model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B
run the benchmark 3 times with each 32 num prompts, then compare the results, but ignore the first iteration as that is just a warmup.
Todo:
- Due to some random outputs by vllm it may show that it found some invalid json. I have not really looked into it yet.
MCP vLLM Benchmarking Tool
Project Details
- Eliovp-BV/mcp-vllm-benchmark
- Last Updated: 4/7/2025
Recomended MCP Servers
A Model Context Protocol server that provides search capabilities using a Google CSE (custom search engine).
Allow LLMs to control a browser with Browserbase and Stagehand
A Model Context Protocol server for retrieving and analyzing issues from Sentry.io
Claude can perform Web Search | Exa with MCP (Model Context Protocol)
:file_folder: A file manager / web client for SFTP, S3, FTP, WebDAV, Git, Minio, LDAP, CalDAV, CardDAV, Mysql,...
A MCP server for Cappt
A powerful Model Context Protocol (MCP) server providing comprehensive Gmail integration with LLM processing capabilities.
An open source deep research clone. AI Agent that reasons large amounts of web data extracted with Firecrawl
MCP (Model Context Protocol) Server for the PI API
free MCP server hosting using vercel
An MCP server for Astra DB workloads





