@mastra/mcp-docs-server 1.1.32-alpha.3 → 1.1.32-alpha.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/{docs/mcp → guides/guide}/publishing-mcp-server.md +21 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/digitalocean.md +19 -3
- package/.docs/models/providers/llmgateway.md +6 -1
- package/.docs/models/providers/mistral.md +3 -2
- package/.docs/models/providers/nvidia.md +2 -1
- package/.docs/models/providers/togetherai.md +2 -1
- package/.docs/reference/tools/mcp-server.md +1 -3
- package/CHANGELOG.md +7 -0
- package/package.json +4 -4
|
@@ -32,7 +32,7 @@ bun add @mastra/mcp @mastra/core tsup
|
|
|
32
32
|
|
|
33
33
|
## Setting up an MCP server
|
|
34
34
|
|
|
35
|
-
1. Create a file for your stdio server, for example,
|
|
35
|
+
1. Create a file for your stdio server, for example, `src/mastra/stdio.ts`.
|
|
36
36
|
|
|
37
37
|
2. Add the following code to the file. Remember to import your actual Mastra tools and name the server appropriately.
|
|
38
38
|
|
|
@@ -68,10 +68,30 @@ bun add @mastra/mcp @mastra/core tsup
|
|
|
68
68
|
|
|
69
69
|
4. Run the build command:
|
|
70
70
|
|
|
71
|
+
**npm**:
|
|
72
|
+
|
|
73
|
+
```bash
|
|
74
|
+
npm run build:mcp
|
|
75
|
+
```
|
|
76
|
+
|
|
77
|
+
**pnpm**:
|
|
78
|
+
|
|
71
79
|
```bash
|
|
72
80
|
pnpm run build:mcp
|
|
73
81
|
```
|
|
74
82
|
|
|
83
|
+
**Yarn**:
|
|
84
|
+
|
|
85
|
+
```bash
|
|
86
|
+
yarn build:mcp
|
|
87
|
+
```
|
|
88
|
+
|
|
89
|
+
**Bun**:
|
|
90
|
+
|
|
91
|
+
```bash
|
|
92
|
+
bun run build:mcp
|
|
93
|
+
```
|
|
94
|
+
|
|
75
95
|
This will compile your server code into both ESM and CJS formats and make the ESM output file executable. On Unix-like systems, the `chmod +x` step makes the file directly executable. Windows users may need to use WSL or handle execution through Node.js directly.
|
|
76
96
|
|
|
77
97
|
## Publishing to NPM
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3803 models from 106 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# DigitalOcean
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 62 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [DigitalOcean documentation](https://docs.digitalocean.com/products/gradient-ai-platform/details/models/).
|
|
6
6
|
|
|
@@ -46,32 +46,43 @@ for await (const chunk of stream) {
|
|
|
46
46
|
| `digitalocean/anthropic-claude-opus-4.7` | 1.0M | | | | | | $5 | $25 |
|
|
47
47
|
| `digitalocean/anthropic-claude-sonnet-4` | 1.0M | | | | | | $3 | $15 |
|
|
48
48
|
| `digitalocean/arcee-trinity-large-thinking` | 256K | | | | | | $0.25 | $0.90 |
|
|
49
|
+
| `digitalocean/bge-m3` | 8K | | | | | | $0.02 | — |
|
|
50
|
+
| `digitalocean/bge-reranker-v2-m3` | 8K | | | | | | $0.01 | — |
|
|
51
|
+
| `digitalocean/deepseek-3.2` | 128K | | | | | | $0.50 | $2 |
|
|
49
52
|
| `digitalocean/deepseek-r1-distill-llama-70b` | 131K | | | | | | $0.99 | $0.99 |
|
|
53
|
+
| `digitalocean/e5-large-v2` | 512 | | | | | | $0.02 | — |
|
|
50
54
|
| `digitalocean/fal-ai/elevenlabs/tts/multilingual-v2` | — | | | | | | — | — |
|
|
51
55
|
| `digitalocean/fal-ai/fast-sdxl` | — | | | | | | — | — |
|
|
52
56
|
| `digitalocean/fal-ai/flux/schnell` | — | | | | | | — | — |
|
|
53
57
|
| `digitalocean/fal-ai/stable-audio-25/text-to-audio` | — | | | | | | — | — |
|
|
58
|
+
| `digitalocean/gemma-4-31B-it` | 256K | | | | | | $0.18 | $0.50 |
|
|
54
59
|
| `digitalocean/glm-5` | 203K | | | | | | $1 | $3 |
|
|
55
60
|
| `digitalocean/gte-large-en-v1.5` | 8K | | | | | | $0.09 | — |
|
|
56
61
|
| `digitalocean/kimi-k2.5` | 262K | | | | | | $0.50 | $3 |
|
|
62
|
+
| `digitalocean/llama-4-maverick` | 1.0M | | | | | | $0.25 | $0.87 |
|
|
63
|
+
| `digitalocean/llama-guard-4-12b` | 128K | | | | | | — | — |
|
|
57
64
|
| `digitalocean/llama3.3-70b-instruct` | 128K | | | | | | $0.65 | $0.65 |
|
|
58
65
|
| `digitalocean/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
|
|
66
|
+
| `digitalocean/mistral-3-14B` | 262K | | | | | | $0.20 | $0.20 |
|
|
59
67
|
| `digitalocean/multi-qa-mpnet-base-dot-v1` | 512 | | | | | | $0.01 | — |
|
|
68
|
+
| `digitalocean/nemotron-3-nano-omni` | 66K | | | | | | $0.50 | $0.90 |
|
|
69
|
+
| `digitalocean/nemotron-nano-12b-v2-vl` | 128K | | | | | | $0.20 | $0.60 |
|
|
60
70
|
| `digitalocean/nvidia-nemotron-3-super-120b` | 256K | | | | | | $0.30 | $0.65 |
|
|
61
71
|
| `digitalocean/openai-gpt-4.1` | 1.0M | | | | | | $2 | $8 |
|
|
62
72
|
| `digitalocean/openai-gpt-4o` | 128K | | | | | | $3 | $10 |
|
|
63
73
|
| `digitalocean/openai-gpt-4o-mini` | 128K | | | | | | $0.15 | $0.60 |
|
|
64
74
|
| `digitalocean/openai-gpt-5` | 400K | | | | | | $1 | $10 |
|
|
65
|
-
| `digitalocean/openai-gpt-5-2-pro` | 400K | | | | | | $21 | $168 |
|
|
66
75
|
| `digitalocean/openai-gpt-5-mini` | 400K | | | | | | $0.25 | $2 |
|
|
67
76
|
| `digitalocean/openai-gpt-5-nano` | 400K | | | | | | $0.05 | $0.40 |
|
|
68
77
|
| `digitalocean/openai-gpt-5.1-codex-max` | 400K | | | | | | $1 | $10 |
|
|
69
78
|
| `digitalocean/openai-gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
79
|
+
| `digitalocean/openai-gpt-5.2-pro` | 400K | | | | | | $21 | $168 |
|
|
70
80
|
| `digitalocean/openai-gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
|
|
71
81
|
| `digitalocean/openai-gpt-5.4` | 1.0M | | | | | | $3 | $15 |
|
|
72
82
|
| `digitalocean/openai-gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
73
83
|
| `digitalocean/openai-gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
74
84
|
| `digitalocean/openai-gpt-5.4-pro` | 400K | | | | | | $30 | $180 |
|
|
85
|
+
| `digitalocean/openai-gpt-5.5` | 1.0M | | | | | | $5 | $30 |
|
|
75
86
|
| `digitalocean/openai-gpt-image-1` | — | | | | | | $5 | $40 |
|
|
76
87
|
| `digitalocean/openai-gpt-image-1.5` | — | | | | | | $5 | $10 |
|
|
77
88
|
| `digitalocean/openai-gpt-oss-120b` | 131K | | | | | | $0.10 | $0.70 |
|
|
@@ -79,7 +90,12 @@ for await (const chunk of stream) {
|
|
|
79
90
|
| `digitalocean/openai-o1` | 200K | | | | | | $15 | $60 |
|
|
80
91
|
| `digitalocean/openai-o3` | 200K | | | | | | $2 | $8 |
|
|
81
92
|
| `digitalocean/openai-o3-mini` | 200K | | | | | | $1 | $4 |
|
|
93
|
+
| `digitalocean/qwen3-coder-flash` | 262K | | | | | | $0.45 | $2 |
|
|
82
94
|
| `digitalocean/qwen3-embedding-0.6b` | 8K | | | | | | $0.04 | — |
|
|
95
|
+
| `digitalocean/qwen3-tts-voicedesign` | 33K | | | | | | — | — |
|
|
96
|
+
| `digitalocean/qwen3.5-397b-a17b` | 262K | | | | | | $0.55 | $4 |
|
|
97
|
+
| `digitalocean/stable-diffusion-3.5-large` | 256 | | | | | | $0.08 | — |
|
|
98
|
+
| `digitalocean/wan2-2-t2v-a14b` | 100 | | | | | | $0.60 | — |
|
|
83
99
|
|
|
84
100
|
## Advanced configuration
|
|
85
101
|
|
|
@@ -109,7 +125,7 @@ const agent = new Agent({
|
|
|
109
125
|
model: ({ requestContext }) => {
|
|
110
126
|
const useAdvanced = requestContext.task === "complex";
|
|
111
127
|
return useAdvanced
|
|
112
|
-
? "digitalocean/
|
|
128
|
+
? "digitalocean/wan2-2-t2v-a14b"
|
|
113
129
|
: "digitalocean/alibaba-qwen3-32b";
|
|
114
130
|
}
|
|
115
131
|
});
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# LLM Gateway
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 189 LLM Gateway models through Mastra's model router. Authentication is handled automatically using the `LLMGATEWAY_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [LLM Gateway documentation](https://llmgateway.io/docs).
|
|
6
6
|
|
|
@@ -120,6 +120,8 @@ for await (const chunk of stream) {
|
|
|
120
120
|
| `llmgateway/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
121
121
|
| `llmgateway/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
122
122
|
| `llmgateway/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
|
|
123
|
+
| `llmgateway/gpt-5.5` | 1.1M | | | | | | $5 | $30 |
|
|
124
|
+
| `llmgateway/gpt-5.5-pro` | 1.1M | | | | | | $30 | $180 |
|
|
123
125
|
| `llmgateway/gpt-oss-120b` | 131K | | | | | | $0.15 | $0.75 |
|
|
124
126
|
| `llmgateway/gpt-oss-20b` | 131K | | | | | | $0.10 | $0.50 |
|
|
125
127
|
| `llmgateway/grok-3` | 131K | | | | | | $3 | $15 |
|
|
@@ -209,6 +211,9 @@ for await (const chunk of stream) {
|
|
|
209
211
|
| `llmgateway/qwen3-vl-8b-instruct` | 131K | | | | | | $0.10 | $0.10 |
|
|
210
212
|
| `llmgateway/qwen3-vl-flash` | 1.0M | | | | | | $0.05 | $0.40 |
|
|
211
213
|
| `llmgateway/qwen3-vl-plus` | 262K | | | | | | $0.20 | $2 |
|
|
214
|
+
| `llmgateway/qwen3.6-35b-a3b` | 262K | | | | | | $0.25 | $1 |
|
|
215
|
+
| `llmgateway/qwen3.6-max-preview` | 262K | | | | | | $1 | $8 |
|
|
216
|
+
| `llmgateway/qwen3.6-plus` | 1.0M | | | | | | $0.28 | $2 |
|
|
212
217
|
| `llmgateway/qwen35-397b-a17b` | 262K | | | | | | $0.60 | $4 |
|
|
213
218
|
| `llmgateway/qwq-plus` | 131K | | | | | | $0.80 | $2 |
|
|
214
219
|
| `llmgateway/seed-1-6-250615` | 256K | | | | | | $0.25 | $2 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Mistral
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 28 Mistral models through Mastra's model router. Authentication is handled automatically using the `MISTRAL_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Mistral documentation](https://docs.mistral.ai/getting-started/models/).
|
|
6
6
|
|
|
@@ -49,7 +49,8 @@ for await (const chunk of stream) {
|
|
|
49
49
|
| `mistral/mistral-large-latest` | 262K | | | | | | $0.50 | $2 |
|
|
50
50
|
| `mistral/mistral-medium-2505` | 131K | | | | | | $0.40 | $2 |
|
|
51
51
|
| `mistral/mistral-medium-2508` | 262K | | | | | | $0.40 | $2 |
|
|
52
|
-
| `mistral/mistral-medium-
|
|
52
|
+
| `mistral/mistral-medium-2604` | 262K | | | | | | $2 | $8 |
|
|
53
|
+
| `mistral/mistral-medium-latest` | 262K | | | | | | $2 | $8 |
|
|
53
54
|
| `mistral/mistral-nemo` | 128K | | | | | | $0.15 | $0.15 |
|
|
54
55
|
| `mistral/mistral-small-2506` | 128K | | | | | | $0.10 | $0.30 |
|
|
55
56
|
| `mistral/mistral-small-2603` | 256K | | | | | | $0.15 | $0.60 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Nvidia
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 81 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Nvidia documentation](https://docs.api.nvidia.com/nim/).
|
|
6
6
|
|
|
@@ -80,6 +80,7 @@ for await (const chunk of stream) {
|
|
|
80
80
|
| `nvidia/mistralai/ministral-14b-instruct-2512` | 262K | | | | | | — | — |
|
|
81
81
|
| `nvidia/mistralai/mistral-large-2-instruct` | 128K | | | | | | — | — |
|
|
82
82
|
| `nvidia/mistralai/mistral-large-3-675b-instruct-2512` | 262K | | | | | | — | — |
|
|
83
|
+
| `nvidia/mistralai/mistral-medium-3.5-128b` | 262K | | | | | | — | — |
|
|
83
84
|
| `nvidia/mistralai/mistral-small-3.1-24b-instruct-2503` | 128K | | | | | | — | — |
|
|
84
85
|
| `nvidia/moonshotai/kimi-k2-instruct` | 128K | | | | | | — | — |
|
|
85
86
|
| `nvidia/moonshotai/kimi-k2-instruct-0905` | 262K | | | | | | — | — |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Together AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 18 Together AI models through Mastra's model router. Authentication is handled automatically using the `TOGETHER_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Together AI documentation](https://docs.together.ai/docs/serverless-models).
|
|
6
6
|
|
|
@@ -48,6 +48,7 @@ for await (const chunk of stream) {
|
|
|
48
48
|
| `togetherai/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8` | 262K | | | | | | $2 | $2 |
|
|
49
49
|
| `togetherai/Qwen/Qwen3-Coder-Next-FP8` | 262K | | | | | | $0.50 | $1 |
|
|
50
50
|
| `togetherai/Qwen/Qwen3.5-397B-A17B` | 262K | | | | | | $0.60 | $4 |
|
|
51
|
+
| `togetherai/Qwen/Qwen3.6-Plus` | 1.0M | | | | | | $0.50 | $3 |
|
|
51
52
|
| `togetherai/zai-org/GLM-5.1` | 203K | | | | | | $1 | $4 |
|
|
52
53
|
|
|
53
54
|
## Advanced configuration
|
|
@@ -768,11 +768,9 @@ await serverWithPrompts.prompts.notifyListChanged()
|
|
|
768
768
|
- Handle errors with informative messages.
|
|
769
769
|
- Document argument expectations and available versions.
|
|
770
770
|
|
|
771
|
-
***
|
|
772
|
-
|
|
773
771
|
## Examples
|
|
774
772
|
|
|
775
|
-
For practical examples of setting up and deploying an MCPServer, see the [Publishing an MCP Server guide](https://mastra.ai/
|
|
773
|
+
For practical examples of setting up and deploying an MCPServer, see the [Publishing an MCP Server guide](https://mastra.ai/guides/guide/publishing-mcp-server).
|
|
776
774
|
|
|
777
775
|
The example at the beginning of this page also demonstrates how to instantiate `MCPServer` with both tools and agents.
|
|
778
776
|
|
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,12 @@
|
|
|
1
1
|
# @mastra/mcp-docs-server
|
|
2
2
|
|
|
3
|
+
## 1.1.32-alpha.4
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [[`b2deb29`](https://github.com/mastra-ai/mastra/commit/b2deb29412b300c868655b5840463614fbb7962d), [`66644be`](https://github.com/mastra-ai/mastra/commit/66644beac1aa560f0e417956ff007c89341dc382), [`310b953`](https://github.com/mastra-ai/mastra/commit/310b95345f302dcd5ba3ed862bdc96f059d44122), [`43f0e1d`](https://github.com/mastra-ai/mastra/commit/43f0e1d5d5a74ba6fc746f2ad89ebe0c64777a7d), [`da0b9e2`](https://github.com/mastra-ai/mastra/commit/da0b9e2ba7ecc560213b426d6c097fe63946086e)]:
|
|
8
|
+
- @mastra/core@1.31.0-alpha.3
|
|
9
|
+
|
|
3
10
|
## 1.1.32-alpha.3
|
|
4
11
|
|
|
5
12
|
### Patch Changes
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.32-alpha.
|
|
3
|
+
"version": "1.1.32-alpha.4",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,8 +29,8 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/
|
|
33
|
-
"@mastra/
|
|
32
|
+
"@mastra/core": "1.31.0-alpha.3",
|
|
33
|
+
"@mastra/mcp": "^1.6.0"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
36
36
|
"@hono/node-server": "^1.19.11",
|
|
@@ -48,7 +48,7 @@
|
|
|
48
48
|
"vitest": "4.1.5",
|
|
49
49
|
"@internal/lint": "0.0.89",
|
|
50
50
|
"@internal/types-builder": "0.0.64",
|
|
51
|
-
"@mastra/core": "1.31.0-alpha.
|
|
51
|
+
"@mastra/core": "1.31.0-alpha.3"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|