Get available models from each LLM provider
Install dependencies:
pnpm install
Build the server:
pnpm run build
For development with auto-rebuild:
pnpm run watch
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"llm-model-providers": {
"command": "/path/to/llm-model-providers/build/index.js"
"env": {
"OPENAI_API_KEY": "",
"ANTHROPIC_API_KEY": ""
}
}
}
}
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
pnpm run inspector
The Inspector will provide a URL to access debugging tools in your browser.
Seamless access to top MCP servers powering the future of AI integration.