@mastra/mcp-docs-server 1.1.32-alpha.3 → 1.1.32-alpha.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -32,7 +32,7 @@ bun add @mastra/mcp @mastra/core tsup
32
32
 
33
33
  ## Setting up an MCP server
34
34
 
35
- 1. Create a file for your stdio server, for example, `/src/mastra/stdio.ts`.
35
+ 1. Create a file for your stdio server, for example, `src/mastra/stdio.ts`.
36
36
 
37
37
  2. Add the following code to the file. Remember to import your actual Mastra tools and name the server appropriately.
38
38
 
@@ -68,10 +68,30 @@ bun add @mastra/mcp @mastra/core tsup
68
68
 
69
69
  4. Run the build command:
70
70
 
71
+ **npm**:
72
+
73
+ ```bash
74
+ npm run build:mcp
75
+ ```
76
+
77
+ **pnpm**:
78
+
71
79
  ```bash
72
80
  pnpm run build:mcp
73
81
  ```
74
82
 
83
+ **Yarn**:
84
+
85
+ ```bash
86
+ yarn build:mcp
87
+ ```
88
+
89
+ **Bun**:
90
+
91
+ ```bash
92
+ bun run build:mcp
93
+ ```
94
+
75
95
  This will compile your server code into both ESM and CJS formats and make the ESM output file executable. On Unix-like systems, the `chmod +x` step makes the file directly executable. Windows users may need to use WSL or handle execution through Node.js directly.
76
96
 
77
97
  ## Publishing to NPM
@@ -1,6 +1,6 @@
1
1
  # Model Providers
2
2
 
3
- Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3779 models from 106 providers through a single API.
3
+ Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3803 models from 106 providers through a single API.
4
4
 
5
5
  ## Features
6
6
 
@@ -1,6 +1,6 @@
1
1
  # ![DigitalOcean logo](https://models.dev/logos/digitalocean.svg)DigitalOcean
2
2
 
3
- Access 46 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
3
+ Access 62 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
4
4
 
5
5
  Learn more in the [DigitalOcean documentation](https://docs.digitalocean.com/products/gradient-ai-platform/details/models/).
6
6
 
@@ -46,32 +46,43 @@ for await (const chunk of stream) {
46
46
  | `digitalocean/anthropic-claude-opus-4.7` | 1.0M | | | | | | $5 | $25 |
47
47
  | `digitalocean/anthropic-claude-sonnet-4` | 1.0M | | | | | | $3 | $15 |
48
48
  | `digitalocean/arcee-trinity-large-thinking` | 256K | | | | | | $0.25 | $0.90 |
49
+ | `digitalocean/bge-m3` | 8K | | | | | | $0.02 | — |
50
+ | `digitalocean/bge-reranker-v2-m3` | 8K | | | | | | $0.01 | — |
51
+ | `digitalocean/deepseek-3.2` | 128K | | | | | | $0.50 | $2 |
49
52
  | `digitalocean/deepseek-r1-distill-llama-70b` | 131K | | | | | | $0.99 | $0.99 |
53
+ | `digitalocean/e5-large-v2` | 512 | | | | | | $0.02 | — |
50
54
  | `digitalocean/fal-ai/elevenlabs/tts/multilingual-v2` | — | | | | | | — | — |
51
55
  | `digitalocean/fal-ai/fast-sdxl` | — | | | | | | — | — |
52
56
  | `digitalocean/fal-ai/flux/schnell` | — | | | | | | — | — |
53
57
  | `digitalocean/fal-ai/stable-audio-25/text-to-audio` | — | | | | | | — | — |
58
+ | `digitalocean/gemma-4-31B-it` | 256K | | | | | | $0.18 | $0.50 |
54
59
  | `digitalocean/glm-5` | 203K | | | | | | $1 | $3 |
55
60
  | `digitalocean/gte-large-en-v1.5` | 8K | | | | | | $0.09 | — |
56
61
  | `digitalocean/kimi-k2.5` | 262K | | | | | | $0.50 | $3 |
62
+ | `digitalocean/llama-4-maverick` | 1.0M | | | | | | $0.25 | $0.87 |
63
+ | `digitalocean/llama-guard-4-12b` | 128K | | | | | | — | — |
57
64
  | `digitalocean/llama3.3-70b-instruct` | 128K | | | | | | $0.65 | $0.65 |
58
65
  | `digitalocean/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
66
+ | `digitalocean/mistral-3-14B` | 262K | | | | | | $0.20 | $0.20 |
59
67
  | `digitalocean/multi-qa-mpnet-base-dot-v1` | 512 | | | | | | $0.01 | — |
68
+ | `digitalocean/nemotron-3-nano-omni` | 66K | | | | | | $0.50 | $0.90 |
69
+ | `digitalocean/nemotron-nano-12b-v2-vl` | 128K | | | | | | $0.20 | $0.60 |
60
70
  | `digitalocean/nvidia-nemotron-3-super-120b` | 256K | | | | | | $0.30 | $0.65 |
61
71
  | `digitalocean/openai-gpt-4.1` | 1.0M | | | | | | $2 | $8 |
62
72
  | `digitalocean/openai-gpt-4o` | 128K | | | | | | $3 | $10 |
63
73
  | `digitalocean/openai-gpt-4o-mini` | 128K | | | | | | $0.15 | $0.60 |
64
74
  | `digitalocean/openai-gpt-5` | 400K | | | | | | $1 | $10 |
65
- | `digitalocean/openai-gpt-5-2-pro` | 400K | | | | | | $21 | $168 |
66
75
  | `digitalocean/openai-gpt-5-mini` | 400K | | | | | | $0.25 | $2 |
67
76
  | `digitalocean/openai-gpt-5-nano` | 400K | | | | | | $0.05 | $0.40 |
68
77
  | `digitalocean/openai-gpt-5.1-codex-max` | 400K | | | | | | $1 | $10 |
69
78
  | `digitalocean/openai-gpt-5.2` | 400K | | | | | | $2 | $14 |
79
+ | `digitalocean/openai-gpt-5.2-pro` | 400K | | | | | | $21 | $168 |
70
80
  | `digitalocean/openai-gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
71
81
  | `digitalocean/openai-gpt-5.4` | 1.0M | | | | | | $3 | $15 |
72
82
  | `digitalocean/openai-gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
73
83
  | `digitalocean/openai-gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
74
84
  | `digitalocean/openai-gpt-5.4-pro` | 400K | | | | | | $30 | $180 |
85
+ | `digitalocean/openai-gpt-5.5` | 1.0M | | | | | | $5 | $30 |
75
86
  | `digitalocean/openai-gpt-image-1` | — | | | | | | $5 | $40 |
76
87
  | `digitalocean/openai-gpt-image-1.5` | — | | | | | | $5 | $10 |
77
88
  | `digitalocean/openai-gpt-oss-120b` | 131K | | | | | | $0.10 | $0.70 |
@@ -79,7 +90,12 @@ for await (const chunk of stream) {
79
90
  | `digitalocean/openai-o1` | 200K | | | | | | $15 | $60 |
80
91
  | `digitalocean/openai-o3` | 200K | | | | | | $2 | $8 |
81
92
  | `digitalocean/openai-o3-mini` | 200K | | | | | | $1 | $4 |
93
+ | `digitalocean/qwen3-coder-flash` | 262K | | | | | | $0.45 | $2 |
82
94
  | `digitalocean/qwen3-embedding-0.6b` | 8K | | | | | | $0.04 | — |
95
+ | `digitalocean/qwen3-tts-voicedesign` | 33K | | | | | | — | — |
96
+ | `digitalocean/qwen3.5-397b-a17b` | 262K | | | | | | $0.55 | $4 |
97
+ | `digitalocean/stable-diffusion-3.5-large` | 256 | | | | | | $0.08 | — |
98
+ | `digitalocean/wan2-2-t2v-a14b` | 100 | | | | | | $0.60 | — |
83
99
 
84
100
  ## Advanced configuration
85
101
 
@@ -109,7 +125,7 @@ const agent = new Agent({
109
125
  model: ({ requestContext }) => {
110
126
  const useAdvanced = requestContext.task === "complex";
111
127
  return useAdvanced
112
- ? "digitalocean/qwen3-embedding-0.6b"
128
+ ? "digitalocean/wan2-2-t2v-a14b"
113
129
  : "digitalocean/alibaba-qwen3-32b";
114
130
  }
115
131
  });
@@ -1,6 +1,6 @@
1
1
  # ![LLM Gateway logo](https://models.dev/logos/llmgateway.svg)LLM Gateway
2
2
 
3
- Access 184 LLM Gateway models through Mastra's model router. Authentication is handled automatically using the `LLMGATEWAY_API_KEY` environment variable.
3
+ Access 189 LLM Gateway models through Mastra's model router. Authentication is handled automatically using the `LLMGATEWAY_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [LLM Gateway documentation](https://llmgateway.io/docs).
6
6
 
@@ -120,6 +120,8 @@ for await (const chunk of stream) {
120
120
  | `llmgateway/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
121
121
  | `llmgateway/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
122
122
  | `llmgateway/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
123
+ | `llmgateway/gpt-5.5` | 1.1M | | | | | | $5 | $30 |
124
+ | `llmgateway/gpt-5.5-pro` | 1.1M | | | | | | $30 | $180 |
123
125
  | `llmgateway/gpt-oss-120b` | 131K | | | | | | $0.15 | $0.75 |
124
126
  | `llmgateway/gpt-oss-20b` | 131K | | | | | | $0.10 | $0.50 |
125
127
  | `llmgateway/grok-3` | 131K | | | | | | $3 | $15 |
@@ -209,6 +211,9 @@ for await (const chunk of stream) {
209
211
  | `llmgateway/qwen3-vl-8b-instruct` | 131K | | | | | | $0.10 | $0.10 |
210
212
  | `llmgateway/qwen3-vl-flash` | 1.0M | | | | | | $0.05 | $0.40 |
211
213
  | `llmgateway/qwen3-vl-plus` | 262K | | | | | | $0.20 | $2 |
214
+ | `llmgateway/qwen3.6-35b-a3b` | 262K | | | | | | $0.25 | $1 |
215
+ | `llmgateway/qwen3.6-max-preview` | 262K | | | | | | $1 | $8 |
216
+ | `llmgateway/qwen3.6-plus` | 1.0M | | | | | | $0.28 | $2 |
212
217
  | `llmgateway/qwen35-397b-a17b` | 262K | | | | | | $0.60 | $4 |
213
218
  | `llmgateway/qwq-plus` | 131K | | | | | | $0.80 | $2 |
214
219
  | `llmgateway/seed-1-6-250615` | 256K | | | | | | $0.25 | $2 |
@@ -1,6 +1,6 @@
1
1
  # ![Mistral logo](https://models.dev/logos/mistral.svg)Mistral
2
2
 
3
- Access 27 Mistral models through Mastra's model router. Authentication is handled automatically using the `MISTRAL_API_KEY` environment variable.
3
+ Access 28 Mistral models through Mastra's model router. Authentication is handled automatically using the `MISTRAL_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Mistral documentation](https://docs.mistral.ai/getting-started/models/).
6
6
 
@@ -49,7 +49,8 @@ for await (const chunk of stream) {
49
49
  | `mistral/mistral-large-latest` | 262K | | | | | | $0.50 | $2 |
50
50
  | `mistral/mistral-medium-2505` | 131K | | | | | | $0.40 | $2 |
51
51
  | `mistral/mistral-medium-2508` | 262K | | | | | | $0.40 | $2 |
52
- | `mistral/mistral-medium-latest` | 128K | | | | | | $0.40 | $2 |
52
+ | `mistral/mistral-medium-2604` | 262K | | | | | | $2 | $8 |
53
+ | `mistral/mistral-medium-latest` | 262K | | | | | | $2 | $8 |
53
54
  | `mistral/mistral-nemo` | 128K | | | | | | $0.15 | $0.15 |
54
55
  | `mistral/mistral-small-2506` | 128K | | | | | | $0.10 | $0.30 |
55
56
  | `mistral/mistral-small-2603` | 256K | | | | | | $0.15 | $0.60 |
@@ -1,6 +1,6 @@
1
1
  # ![Nvidia logo](https://models.dev/logos/nvidia.svg)Nvidia
2
2
 
3
- Access 80 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
3
+ Access 81 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Nvidia documentation](https://docs.api.nvidia.com/nim/).
6
6
 
@@ -80,6 +80,7 @@ for await (const chunk of stream) {
80
80
  | `nvidia/mistralai/ministral-14b-instruct-2512` | 262K | | | | | | — | — |
81
81
  | `nvidia/mistralai/mistral-large-2-instruct` | 128K | | | | | | — | — |
82
82
  | `nvidia/mistralai/mistral-large-3-675b-instruct-2512` | 262K | | | | | | — | — |
83
+ | `nvidia/mistralai/mistral-medium-3.5-128b` | 262K | | | | | | — | — |
83
84
  | `nvidia/mistralai/mistral-small-3.1-24b-instruct-2503` | 128K | | | | | | — | — |
84
85
  | `nvidia/moonshotai/kimi-k2-instruct` | 128K | | | | | | — | — |
85
86
  | `nvidia/moonshotai/kimi-k2-instruct-0905` | 262K | | | | | | — | — |
@@ -1,6 +1,6 @@
1
1
  # ![Together AI logo](https://models.dev/logos/togetherai.svg)Together AI
2
2
 
3
- Access 17 Together AI models through Mastra's model router. Authentication is handled automatically using the `TOGETHER_API_KEY` environment variable.
3
+ Access 18 Together AI models through Mastra's model router. Authentication is handled automatically using the `TOGETHER_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Together AI documentation](https://docs.together.ai/docs/serverless-models).
6
6
 
@@ -48,6 +48,7 @@ for await (const chunk of stream) {
48
48
  | `togetherai/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8` | 262K | | | | | | $2 | $2 |
49
49
  | `togetherai/Qwen/Qwen3-Coder-Next-FP8` | 262K | | | | | | $0.50 | $1 |
50
50
  | `togetherai/Qwen/Qwen3.5-397B-A17B` | 262K | | | | | | $0.60 | $4 |
51
+ | `togetherai/Qwen/Qwen3.6-Plus` | 1.0M | | | | | | $0.50 | $3 |
51
52
  | `togetherai/zai-org/GLM-5.1` | 203K | | | | | | $1 | $4 |
52
53
 
53
54
  ## Advanced configuration
@@ -768,11 +768,9 @@ await serverWithPrompts.prompts.notifyListChanged()
768
768
  - Handle errors with informative messages.
769
769
  - Document argument expectations and available versions.
770
770
 
771
- ***
772
-
773
771
  ## Examples
774
772
 
775
- For practical examples of setting up and deploying an MCPServer, see the [Publishing an MCP Server guide](https://mastra.ai/docs/mcp/publishing-mcp-server).
773
+ For practical examples of setting up and deploying an MCPServer, see the [Publishing an MCP Server guide](https://mastra.ai/guides/guide/publishing-mcp-server).
776
774
 
777
775
  The example at the beginning of this page also demonstrates how to instantiate `MCPServer` with both tools and agents.
778
776
 
package/CHANGELOG.md CHANGED
@@ -1,5 +1,12 @@
1
1
  # @mastra/mcp-docs-server
2
2
 
3
+ ## 1.1.32-alpha.4
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [[`b2deb29`](https://github.com/mastra-ai/mastra/commit/b2deb29412b300c868655b5840463614fbb7962d), [`66644be`](https://github.com/mastra-ai/mastra/commit/66644beac1aa560f0e417956ff007c89341dc382), [`310b953`](https://github.com/mastra-ai/mastra/commit/310b95345f302dcd5ba3ed862bdc96f059d44122), [`43f0e1d`](https://github.com/mastra-ai/mastra/commit/43f0e1d5d5a74ba6fc746f2ad89ebe0c64777a7d), [`da0b9e2`](https://github.com/mastra-ai/mastra/commit/da0b9e2ba7ecc560213b426d6c097fe63946086e)]:
8
+ - @mastra/core@1.31.0-alpha.3
9
+
3
10
  ## 1.1.32-alpha.3
4
11
 
5
12
  ### Patch Changes
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mastra/mcp-docs-server",
3
- "version": "1.1.32-alpha.3",
3
+ "version": "1.1.32-alpha.4",
4
4
  "description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
@@ -29,8 +29,8 @@
29
29
  "jsdom": "^26.1.0",
30
30
  "local-pkg": "^1.1.2",
31
31
  "zod": "^4.3.6",
32
- "@mastra/mcp": "^1.6.0",
33
- "@mastra/core": "1.31.0-alpha.2"
32
+ "@mastra/core": "1.31.0-alpha.3",
33
+ "@mastra/mcp": "^1.6.0"
34
34
  },
35
35
  "devDependencies": {
36
36
  "@hono/node-server": "^1.19.11",
@@ -48,7 +48,7 @@
48
48
  "vitest": "4.1.5",
49
49
  "@internal/lint": "0.0.89",
50
50
  "@internal/types-builder": "0.0.64",
51
- "@mastra/core": "1.31.0-alpha.2"
51
+ "@mastra/core": "1.31.0-alpha.3"
52
52
  },
53
53
  "homepage": "https://mastra.ai",
54
54
  "repository": {