A Model Context Protocol server providing LLM Agents a second opinion via AI-powered Deepseek-Reasoning (R1) mentorship capabilities, including code review, design critique, writing feedback, and idea brainstorming through the Deepseek API. Set your LLM Agent up for success with expert second opinions and actionable insights.
The Model Context Protocol (MCP) enables communication between:
# Clone the repository
git clone git@github.com:cyanheads/mentor-mcp-server.git
cd mentor-mcp-server
# Install dependencies
npm install
# Build the project
npm run build
Add to your MCP client settings:
{
"mcpServers": {
"mentor": {
"command": "node",
"args": ["build/index.js"],
"env": {
"DEEPSEEK_API_KEY": "your_api_key",
"DEEPSEEK_MODEL": "deepseek-reasoner",
"DEEPSEEK_MAX_TOKENS": "8192",
"DEEPSEEK_MAX_RETRIES": "3",
"DEEPSEEK_TIMEOUT": "30000"
}
}
}
}
Variable | Required | Default | Description |
---|---|---|---|
DEEPSEEK_API_KEY | Yes | - | Your Deepseek API key |
DEEPSEEK_MODEL | Yes | deepseek-reasoner | Deepseek model name |
DEEPSEEK_MAX_TOKENS | No | 8192 | Maximum tokens per request |
DEEPSEEK_MAX_RETRIES | No | 3 | Number of retry attempts |
DEEPSEEK_TIMEOUT | No | 30000 | Request timeout (ms) |
<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>code_review</tool_name>
<arguments>
{
"file_path": "src/app.ts",
"language": "typescript"
}
</arguments>
</use_mcp_tool>
<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>design_critique</tool_name>
<arguments>
{
"design_document": "path/to/design.fig",
"design_type": "web UI"
}
</arguments>
</use_mcp_tool>
<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>writing_feedback</tool_name>
<arguments>
{
"text": "Documentation content...",
"writing_type": "documentation"
}
</arguments>
</use_mcp_tool>
<use_mcp_tool>
<server_name>mentor-mcp-server</server_name>
<tool_name>brainstorm_enhancements</tool_name>
<arguments>
{
"concept": "User authentication system"
}
</arguments>
</use_mcp_tool>
Detailed examples of each tool's usage and output can be found in the examples directory:
Each example includes the request format and sample response, demonstrating the tool's capabilities and output structure.
# Build TypeScript code
npm run build
# Start the server
npm run start
# Development with watch mode
npm run dev
# Clean build artifacts
npm run clean
src/
├── api/ # API integration modules
├── tools/ # Tool implementations
│ ├── second-opinion/
│ ├── code-review/
│ ├── design-critique/
│ ├── writing-feedback/
│ └── brainstorm-enhancements/
├── types/ # TypeScript type definitions
├── utils/ # Utility functions
├── config.ts # Server configuration
├── index.ts # Entry point
└── server.ts # Main server implementation
Apache License 2.0. See LICENSE for more information.
Seamless access to top MCP servers powering the future of AI integration.