Lightweight MCP server to give your Cursor Agent access to the Vercel API.
This is a lightweight Model Control Protocol (MCP) server bootstrapped with create-mcp, and deployed on Cloudflare Workers.
This MCP server allows AI agents (such as Cursor) to interface with the Vercel API.
It’s still under development, I will be adding more tools as I find myself needing them.
See src/index.ts for the current list of tools. Every method in the class is an MCP tool.
bun create mcp --clone https://github.com/zueai/vercel-api-mcp
Open Cursor Settings -> MCP -> Add new MCP server
and paste the command that was copied to your clipboard.
Upload your Vercel API token as a secret:
bunx wrangler secret put VERCEL_API_TOKEN
bun run deploy
To create new MCP tools, add methods to the MyWorker
class in src/index.ts
. Each function will automatically become an MCP tool that your agent can use.
Example:
/**
* A warm, friendly greeting from your MCP server.
* @param name {string} the name of the person we are greeting.
* @return {string} the contents of our greeting.
*/
sayHello(name: string) {
return `Hello from an MCP Worker, ${name}!`;
}
The JSDoc comments are important:
@param
tags define the tool’s parameters with types and descriptions@return
tag specifies the return value and typeCheck out the following resources to learn more:
MCP server connecting Claude to Vercel
Query model running with Ollama from within Claude Desktop or other MCP clients
A Python package enabling LLM models to interact with the Memos server via the MCP interface for searching, creating, retrieving, and managing memos.