@mastra/mcp-docs-server 1.1.17 → 1.1.18-alpha.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/models/gateways/netlify.md +65 -66
- package/.docs/models/gateways/openrouter.md +2 -1
- package/.docs/models/gateways/vercel.md +3 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/opencode.md +3 -2
- package/.docs/models/providers/vultr.md +11 -16
- package/.docs/models/providers/zenmux.md +2 -31
- package/CHANGELOG.md +7 -0
- package/package.json +4 -4
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Netlify
|
|
2
2
|
|
|
3
|
-
Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access
|
|
3
|
+
Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access 62 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Netlify documentation](https://docs.netlify.com/build/ai-gateway/overview/).
|
|
6
6
|
|
|
@@ -33,68 +33,67 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
33
33
|
|
|
34
34
|
## Available models
|
|
35
35
|
|
|
36
|
-
| Model
|
|
37
|
-
|
|
|
38
|
-
| `anthropic/claude-3-haiku-20240307`
|
|
39
|
-
| `anthropic/claude-haiku-4-5`
|
|
40
|
-
| `anthropic/claude-haiku-4-5-20251001`
|
|
41
|
-
| `anthropic/claude-opus-4-1-20250805`
|
|
42
|
-
| `anthropic/claude-opus-4-20250514`
|
|
43
|
-
| `anthropic/claude-opus-4-5`
|
|
44
|
-
| `anthropic/claude-opus-4-5-20251101`
|
|
45
|
-
| `anthropic/claude-opus-4-6`
|
|
46
|
-
| `anthropic/claude-sonnet-4-0`
|
|
47
|
-
| `anthropic/claude-sonnet-4-20250514`
|
|
48
|
-
| `anthropic/claude-sonnet-4-5`
|
|
49
|
-
| `anthropic/claude-sonnet-4-5-20250929`
|
|
50
|
-
| `anthropic/claude-sonnet-4-6`
|
|
51
|
-
| `gemini/gemini-2.0-flash`
|
|
52
|
-
| `gemini/gemini-2.0-flash-lite`
|
|
53
|
-
| `gemini/gemini-2.5-flash`
|
|
54
|
-
| `gemini/gemini-2.5-flash-image`
|
|
55
|
-
| `gemini/gemini-2.5-flash-lite`
|
|
56
|
-
| `gemini/gemini-2.5-
|
|
57
|
-
| `gemini/gemini-
|
|
58
|
-
| `gemini/gemini-3-
|
|
59
|
-
| `gemini/gemini-3-
|
|
60
|
-
| `gemini/gemini-3.1-flash-
|
|
61
|
-
| `gemini/gemini-3.1-
|
|
62
|
-
| `gemini/gemini-3.1-pro-preview`
|
|
63
|
-
| `gemini/gemini-
|
|
64
|
-
| `gemini/gemini-flash-latest`
|
|
65
|
-
| `
|
|
66
|
-
| `openai/gpt-4.1`
|
|
67
|
-
| `openai/gpt-4.1-
|
|
68
|
-
| `openai/gpt-
|
|
69
|
-
| `openai/gpt-4o`
|
|
70
|
-
| `openai/gpt-
|
|
71
|
-
| `openai/gpt-5`
|
|
72
|
-
| `openai/gpt-5-
|
|
73
|
-
| `openai/gpt-5-
|
|
74
|
-
| `openai/gpt-5-mini`
|
|
75
|
-
| `openai/gpt-5-
|
|
76
|
-
| `openai/gpt-5-
|
|
77
|
-
| `openai/gpt-5
|
|
78
|
-
| `openai/gpt-5.1`
|
|
79
|
-
| `openai/gpt-5.1-
|
|
80
|
-
| `openai/gpt-5.1-codex`
|
|
81
|
-
| `openai/gpt-5.1-codex-
|
|
82
|
-
| `openai/gpt-5.
|
|
83
|
-
| `openai/gpt-5.2`
|
|
84
|
-
| `openai/gpt-5.2-
|
|
85
|
-
| `openai/gpt-5.2-
|
|
86
|
-
| `openai/gpt-5.2-pro`
|
|
87
|
-
| `openai/gpt-5.
|
|
88
|
-
| `openai/gpt-5.3-
|
|
89
|
-
| `openai/gpt-5.
|
|
90
|
-
| `openai/gpt-5.4`
|
|
91
|
-
| `openai/gpt-5.4-
|
|
92
|
-
| `openai/gpt-5.4-mini`
|
|
93
|
-
| `openai/gpt-5.4-
|
|
94
|
-
| `openai/gpt-5.4-nano`
|
|
95
|
-
| `openai/gpt-5.4-
|
|
96
|
-
| `openai/gpt-5.4-pro`
|
|
97
|
-
| `openai/
|
|
98
|
-
| `openai/o3`
|
|
99
|
-
| `openai/
|
|
100
|
-
| `openai/o4-mini` |
|
|
36
|
+
| Model |
|
|
37
|
+
| ------------------------------------------- |
|
|
38
|
+
| `anthropic/claude-3-haiku-20240307` |
|
|
39
|
+
| `anthropic/claude-haiku-4-5` |
|
|
40
|
+
| `anthropic/claude-haiku-4-5-20251001` |
|
|
41
|
+
| `anthropic/claude-opus-4-1-20250805` |
|
|
42
|
+
| `anthropic/claude-opus-4-20250514` |
|
|
43
|
+
| `anthropic/claude-opus-4-5` |
|
|
44
|
+
| `anthropic/claude-opus-4-5-20251101` |
|
|
45
|
+
| `anthropic/claude-opus-4-6` |
|
|
46
|
+
| `anthropic/claude-sonnet-4-0` |
|
|
47
|
+
| `anthropic/claude-sonnet-4-20250514` |
|
|
48
|
+
| `anthropic/claude-sonnet-4-5` |
|
|
49
|
+
| `anthropic/claude-sonnet-4-5-20250929` |
|
|
50
|
+
| `anthropic/claude-sonnet-4-6` |
|
|
51
|
+
| `gemini/gemini-2.0-flash` |
|
|
52
|
+
| `gemini/gemini-2.0-flash-lite` |
|
|
53
|
+
| `gemini/gemini-2.5-flash` |
|
|
54
|
+
| `gemini/gemini-2.5-flash-image` |
|
|
55
|
+
| `gemini/gemini-2.5-flash-lite` |
|
|
56
|
+
| `gemini/gemini-2.5-pro` |
|
|
57
|
+
| `gemini/gemini-3-flash-preview` |
|
|
58
|
+
| `gemini/gemini-3-pro-image-preview` |
|
|
59
|
+
| `gemini/gemini-3.1-flash-image-preview` |
|
|
60
|
+
| `gemini/gemini-3.1-flash-lite-preview` |
|
|
61
|
+
| `gemini/gemini-3.1-pro-preview` |
|
|
62
|
+
| `gemini/gemini-3.1-pro-preview-customtools` |
|
|
63
|
+
| `gemini/gemini-flash-latest` |
|
|
64
|
+
| `gemini/gemini-flash-lite-latest` |
|
|
65
|
+
| `openai/gpt-4.1` |
|
|
66
|
+
| `openai/gpt-4.1-mini` |
|
|
67
|
+
| `openai/gpt-4.1-nano` |
|
|
68
|
+
| `openai/gpt-4o` |
|
|
69
|
+
| `openai/gpt-4o-mini` |
|
|
70
|
+
| `openai/gpt-5` |
|
|
71
|
+
| `openai/gpt-5-2025-08-07` |
|
|
72
|
+
| `openai/gpt-5-codex` |
|
|
73
|
+
| `openai/gpt-5-mini` |
|
|
74
|
+
| `openai/gpt-5-mini-2025-08-07` |
|
|
75
|
+
| `openai/gpt-5-nano` |
|
|
76
|
+
| `openai/gpt-5-pro` |
|
|
77
|
+
| `openai/gpt-5.1` |
|
|
78
|
+
| `openai/gpt-5.1-2025-11-13` |
|
|
79
|
+
| `openai/gpt-5.1-codex` |
|
|
80
|
+
| `openai/gpt-5.1-codex-max` |
|
|
81
|
+
| `openai/gpt-5.1-codex-mini` |
|
|
82
|
+
| `openai/gpt-5.2` |
|
|
83
|
+
| `openai/gpt-5.2-2025-12-11` |
|
|
84
|
+
| `openai/gpt-5.2-codex` |
|
|
85
|
+
| `openai/gpt-5.2-pro` |
|
|
86
|
+
| `openai/gpt-5.2-pro-2025-12-11` |
|
|
87
|
+
| `openai/gpt-5.3-chat-latest` |
|
|
88
|
+
| `openai/gpt-5.3-codex` |
|
|
89
|
+
| `openai/gpt-5.4` |
|
|
90
|
+
| `openai/gpt-5.4-2026-03-05` |
|
|
91
|
+
| `openai/gpt-5.4-mini` |
|
|
92
|
+
| `openai/gpt-5.4-mini-2026-03-17` |
|
|
93
|
+
| `openai/gpt-5.4-nano` |
|
|
94
|
+
| `openai/gpt-5.4-nano-2026-03-17` |
|
|
95
|
+
| `openai/gpt-5.4-pro` |
|
|
96
|
+
| `openai/gpt-5.4-pro-2026-03-05` |
|
|
97
|
+
| `openai/o3` |
|
|
98
|
+
| `openai/o3-mini` |
|
|
99
|
+
| `openai/o4-mini` |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenRouter
|
|
2
2
|
|
|
3
|
-
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 166 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
|
|
6
6
|
|
|
@@ -172,6 +172,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
172
172
|
| `qwen/qwen3-next-80b-a3b-thinking` |
|
|
173
173
|
| `qwen/qwen3.5-397b-a17b` |
|
|
174
174
|
| `qwen/qwen3.5-plus-02-15` |
|
|
175
|
+
| `qwen/qwen3.6-plus-preview:free` |
|
|
175
176
|
| `sourceful/riverflow-v2-fast-preview` |
|
|
176
177
|
| `sourceful/riverflow-v2-max-preview` |
|
|
177
178
|
| `sourceful/riverflow-v2-standard-preview` |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vercel
|
|
2
2
|
|
|
3
|
-
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 226 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
|
|
6
6
|
|
|
@@ -117,6 +117,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
117
117
|
| `inception/mercury-2` |
|
|
118
118
|
| `inception/mercury-coder-small` |
|
|
119
119
|
| `kwaipilot/kat-coder-pro-v1` |
|
|
120
|
+
| `kwaipilot/kat-coder-pro-v2` |
|
|
120
121
|
| `meituan/longcat-flash-chat` |
|
|
121
122
|
| `meituan/longcat-flash-thinking` |
|
|
122
123
|
| `meituan/longcat-flash-thinking-2601` |
|
|
@@ -162,6 +163,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
162
163
|
| `morph/morph-v3-fast` |
|
|
163
164
|
| `morph/morph-v3-large` |
|
|
164
165
|
| `nvidia/nemotron-3-nano-30b-a3b` |
|
|
166
|
+
| `nvidia/nemotron-3-super-120b-a12b` |
|
|
165
167
|
| `nvidia/nemotron-nano-12b-v2-vl` |
|
|
166
168
|
| `nvidia/nemotron-nano-9b-v2` |
|
|
167
169
|
| `openai/codex-mini` |
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3613 models from 95 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenCode Zen
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 34 OpenCode Zen models through Mastra's model router. Authentication is handled automatically using the `OPENCODE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenCode Zen documentation](https://opencode.ai/docs/zen).
|
|
6
6
|
|
|
@@ -67,6 +67,7 @@ for await (const chunk of stream) {
|
|
|
67
67
|
| `opencode/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
|
|
68
68
|
| `opencode/minimax-m2.5-free` | 205K | | | | | | — | — |
|
|
69
69
|
| `opencode/nemotron-3-super-free` | 1.0M | | | | | | — | — |
|
|
70
|
+
| `opencode/qwen3.6-plus-free` | 1.0M | | | | | | — | — |
|
|
70
71
|
|
|
71
72
|
## Advanced configuration
|
|
72
73
|
|
|
@@ -96,7 +97,7 @@ const agent = new Agent({
|
|
|
96
97
|
model: ({ requestContext }) => {
|
|
97
98
|
const useAdvanced = requestContext.task === "complex";
|
|
98
99
|
return useAdvanced
|
|
99
|
-
? "opencode/
|
|
100
|
+
? "opencode/qwen3.6-plus-free"
|
|
100
101
|
: "opencode/big-pickle";
|
|
101
102
|
}
|
|
102
103
|
});
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vultr
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 4 Vultr models through Mastra's model router. Authentication is handled automatically using the `VULTR_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vultr documentation](https://api.vultrinference.com/).
|
|
6
6
|
|
|
@@ -15,7 +15,7 @@ const agent = new Agent({
|
|
|
15
15
|
id: "my-agent",
|
|
16
16
|
name: "My Agent",
|
|
17
17
|
instructions: "You are a helpful assistant",
|
|
18
|
-
model: "vultr/DeepSeek-
|
|
18
|
+
model: "vultr/DeepSeek-V3.2"
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Generate a response
|
|
@@ -32,17 +32,12 @@ for await (const chunk of stream) {
|
|
|
32
32
|
|
|
33
33
|
## Models
|
|
34
34
|
|
|
35
|
-
| Model
|
|
36
|
-
|
|
|
37
|
-
| `vultr/DeepSeek-
|
|
38
|
-
| `vultr/
|
|
39
|
-
| `vultr/
|
|
40
|
-
| `vultr/
|
|
41
|
-
| `vultr/gpt-oss-120b` | 130K | | | | | | $0.15 | $0.60 |
|
|
42
|
-
| `vultr/Kimi-K2.5` | 261K | | | | | | $0.55 | $3 |
|
|
43
|
-
| `vultr/Llama-3_1-Nemotron-Ultra-253B-v1` | 32K | | | | | | $0.55 | $2 |
|
|
44
|
-
| `vultr/MiniMax-M2.5` | 196K | | | | | | $0.30 | $1 |
|
|
45
|
-
| `vultr/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4` | 260K | | | | | | $0.20 | $0.80 |
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| --------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `vultr/DeepSeek-V3.2` | 163K | | | | | | $0.55 | $2 |
|
|
38
|
+
| `vultr/GLM-5-FP8` | 202K | | | | | | $0.85 | $3 |
|
|
39
|
+
| `vultr/Kimi-K2.5` | 261K | | | | | | $0.55 | $3 |
|
|
40
|
+
| `vultr/MiniMax-M2.5` | 196K | | | | | | $0.30 | $1 |
|
|
46
41
|
|
|
47
42
|
## Advanced configuration
|
|
48
43
|
|
|
@@ -54,7 +49,7 @@ const agent = new Agent({
|
|
|
54
49
|
name: "custom-agent",
|
|
55
50
|
model: {
|
|
56
51
|
url: "https://api.vultrinference.com/v1",
|
|
57
|
-
id: "vultr/DeepSeek-
|
|
52
|
+
id: "vultr/DeepSeek-V3.2",
|
|
58
53
|
apiKey: process.env.VULTR_API_KEY,
|
|
59
54
|
headers: {
|
|
60
55
|
"X-Custom-Header": "value"
|
|
@@ -72,8 +67,8 @@ const agent = new Agent({
|
|
|
72
67
|
model: ({ requestContext }) => {
|
|
73
68
|
const useAdvanced = requestContext.task === "complex";
|
|
74
69
|
return useAdvanced
|
|
75
|
-
? "vultr/
|
|
76
|
-
: "vultr/DeepSeek-
|
|
70
|
+
? "vultr/MiniMax-M2.5"
|
|
71
|
+
: "vultr/DeepSeek-V3.2";
|
|
77
72
|
}
|
|
78
73
|
});
|
|
79
74
|
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# ZenMux
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 85 ZenMux models through Mastra's model router. Authentication is handled automatically using the `ZENMUX_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [ZenMux documentation](https://docs.zenmux.ai).
|
|
6
6
|
|
|
@@ -52,7 +52,6 @@ for await (const chunk of stream) {
|
|
|
52
52
|
| `zenmux/google/gemini-2.5-flash-lite` | 1.0M | | | | | | $0.10 | $0.40 |
|
|
53
53
|
| `zenmux/google/gemini-2.5-pro` | 1.0M | | | | | | $1 | $10 |
|
|
54
54
|
| `zenmux/google/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
|
|
55
|
-
| `zenmux/google/gemini-3-pro-image-preview` | 1.0M | | | | | | $2 | $12 |
|
|
56
55
|
| `zenmux/google/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
57
56
|
| `zenmux/google/gemini-3.1-flash-lite-preview` | 1.1M | | | | | | $0.25 | $2 |
|
|
58
57
|
| `zenmux/google/gemini-3.1-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
@@ -130,7 +129,7 @@ const agent = new Agent({
|
|
|
130
129
|
id: "custom-agent",
|
|
131
130
|
name: "custom-agent",
|
|
132
131
|
model: {
|
|
133
|
-
url: "https://zenmux.ai/api/
|
|
132
|
+
url: "https://zenmux.ai/api/v1",
|
|
134
133
|
id: "zenmux/anthropic/claude-3.5-haiku",
|
|
135
134
|
apiKey: process.env.ZENMUX_API_KEY,
|
|
136
135
|
headers: {
|
|
@@ -153,32 +152,4 @@ const agent = new Agent({
|
|
|
153
152
|
: "zenmux/anthropic/claude-3.5-haiku";
|
|
154
153
|
}
|
|
155
154
|
});
|
|
156
|
-
```
|
|
157
|
-
|
|
158
|
-
## Direct provider installation
|
|
159
|
-
|
|
160
|
-
This provider can also be installed directly as a standalone package, which can be used instead of the Mastra model router string. View the [package documentation](https://www.npmjs.com/package/@ai-sdk/anthropic) for more details.
|
|
161
|
-
|
|
162
|
-
**npm**:
|
|
163
|
-
|
|
164
|
-
```bash
|
|
165
|
-
npm install @ai-sdk/anthropic
|
|
166
|
-
```
|
|
167
|
-
|
|
168
|
-
**pnpm**:
|
|
169
|
-
|
|
170
|
-
```bash
|
|
171
|
-
pnpm add @ai-sdk/anthropic
|
|
172
|
-
```
|
|
173
|
-
|
|
174
|
-
**Yarn**:
|
|
175
|
-
|
|
176
|
-
```bash
|
|
177
|
-
yarn add @ai-sdk/anthropic
|
|
178
|
-
```
|
|
179
|
-
|
|
180
|
-
**Bun**:
|
|
181
|
-
|
|
182
|
-
```bash
|
|
183
|
-
bun add @ai-sdk/anthropic
|
|
184
155
|
```
|
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,12 @@
|
|
|
1
1
|
# @mastra/mcp-docs-server
|
|
2
2
|
|
|
3
|
+
## 1.1.18-alpha.0
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [[`180aaaf`](https://github.com/mastra-ai/mastra/commit/180aaaf4d0903d33a49bc72de2d40ca69a5bc599)]:
|
|
8
|
+
- @mastra/core@1.18.1-alpha.0
|
|
9
|
+
|
|
3
10
|
## 1.1.17
|
|
4
11
|
|
|
5
12
|
### Patch Changes
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.
|
|
3
|
+
"version": "1.1.18-alpha.1",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,7 +29,7 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/core": "1.18.0",
|
|
32
|
+
"@mastra/core": "1.18.1-alpha.0",
|
|
33
33
|
"@mastra/mcp": "^1.3.2"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
@@ -46,9 +46,9 @@
|
|
|
46
46
|
"tsx": "^4.21.0",
|
|
47
47
|
"typescript": "^5.9.3",
|
|
48
48
|
"vitest": "4.0.18",
|
|
49
|
+
"@internal/lint": "0.0.75",
|
|
49
50
|
"@internal/types-builder": "0.0.50",
|
|
50
|
-
"@mastra/core": "1.18.0"
|
|
51
|
-
"@internal/lint": "0.0.75"
|
|
51
|
+
"@mastra/core": "1.18.1-alpha.0"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|