@mastra/mcp-docs-server 1.1.17 → 1.1.18-alpha.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  # Netlify
2
2
 
3
- Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access 63 models through Mastra's model router.
3
+ Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access 62 models through Mastra's model router.
4
4
 
5
5
  Learn more in the [Netlify documentation](https://docs.netlify.com/build/ai-gateway/overview/).
6
6
 
@@ -33,68 +33,67 @@ ANTHROPIC_API_KEY=ant-...
33
33
 
34
34
  ## Available models
35
35
 
36
- | Model |
37
- | ---------------------------------------------- |
38
- | `anthropic/claude-3-haiku-20240307` |
39
- | `anthropic/claude-haiku-4-5` |
40
- | `anthropic/claude-haiku-4-5-20251001` |
41
- | `anthropic/claude-opus-4-1-20250805` |
42
- | `anthropic/claude-opus-4-20250514` |
43
- | `anthropic/claude-opus-4-5` |
44
- | `anthropic/claude-opus-4-5-20251101` |
45
- | `anthropic/claude-opus-4-6` |
46
- | `anthropic/claude-sonnet-4-0` |
47
- | `anthropic/claude-sonnet-4-20250514` |
48
- | `anthropic/claude-sonnet-4-5` |
49
- | `anthropic/claude-sonnet-4-5-20250929` |
50
- | `anthropic/claude-sonnet-4-6` |
51
- | `gemini/gemini-2.0-flash` |
52
- | `gemini/gemini-2.0-flash-lite` |
53
- | `gemini/gemini-2.5-flash` |
54
- | `gemini/gemini-2.5-flash-image` |
55
- | `gemini/gemini-2.5-flash-lite` |
56
- | `gemini/gemini-2.5-flash-lite-preview-09-2025` |
57
- | `gemini/gemini-2.5-pro` |
58
- | `gemini/gemini-3-flash-preview` |
59
- | `gemini/gemini-3-pro-image-preview` |
60
- | `gemini/gemini-3.1-flash-image-preview` |
61
- | `gemini/gemini-3.1-flash-lite-preview` |
62
- | `gemini/gemini-3.1-pro-preview` |
63
- | `gemini/gemini-3.1-pro-preview-customtools` |
64
- | `gemini/gemini-flash-latest` |
65
- | `gemini/gemini-flash-lite-latest` |
66
- | `openai/gpt-4.1` |
67
- | `openai/gpt-4.1-mini` |
68
- | `openai/gpt-4.1-nano` |
69
- | `openai/gpt-4o` |
70
- | `openai/gpt-4o-mini` |
71
- | `openai/gpt-5` |
72
- | `openai/gpt-5-2025-08-07` |
73
- | `openai/gpt-5-codex` |
74
- | `openai/gpt-5-mini` |
75
- | `openai/gpt-5-mini-2025-08-07` |
76
- | `openai/gpt-5-nano` |
77
- | `openai/gpt-5-pro` |
78
- | `openai/gpt-5.1` |
79
- | `openai/gpt-5.1-2025-11-13` |
80
- | `openai/gpt-5.1-codex` |
81
- | `openai/gpt-5.1-codex-max` |
82
- | `openai/gpt-5.1-codex-mini` |
83
- | `openai/gpt-5.2` |
84
- | `openai/gpt-5.2-2025-12-11` |
85
- | `openai/gpt-5.2-codex` |
86
- | `openai/gpt-5.2-pro` |
87
- | `openai/gpt-5.2-pro-2025-12-11` |
88
- | `openai/gpt-5.3-chat-latest` |
89
- | `openai/gpt-5.3-codex` |
90
- | `openai/gpt-5.4` |
91
- | `openai/gpt-5.4-2026-03-05` |
92
- | `openai/gpt-5.4-mini` |
93
- | `openai/gpt-5.4-mini-2026-03-17` |
94
- | `openai/gpt-5.4-nano` |
95
- | `openai/gpt-5.4-nano-2026-03-17` |
96
- | `openai/gpt-5.4-pro` |
97
- | `openai/gpt-5.4-pro-2026-03-05` |
98
- | `openai/o3` |
99
- | `openai/o3-mini` |
100
- | `openai/o4-mini` |
36
+ | Model |
37
+ | ------------------------------------------- |
38
+ | `anthropic/claude-3-haiku-20240307` |
39
+ | `anthropic/claude-haiku-4-5` |
40
+ | `anthropic/claude-haiku-4-5-20251001` |
41
+ | `anthropic/claude-opus-4-1-20250805` |
42
+ | `anthropic/claude-opus-4-20250514` |
43
+ | `anthropic/claude-opus-4-5` |
44
+ | `anthropic/claude-opus-4-5-20251101` |
45
+ | `anthropic/claude-opus-4-6` |
46
+ | `anthropic/claude-sonnet-4-0` |
47
+ | `anthropic/claude-sonnet-4-20250514` |
48
+ | `anthropic/claude-sonnet-4-5` |
49
+ | `anthropic/claude-sonnet-4-5-20250929` |
50
+ | `anthropic/claude-sonnet-4-6` |
51
+ | `gemini/gemini-2.0-flash` |
52
+ | `gemini/gemini-2.0-flash-lite` |
53
+ | `gemini/gemini-2.5-flash` |
54
+ | `gemini/gemini-2.5-flash-image` |
55
+ | `gemini/gemini-2.5-flash-lite` |
56
+ | `gemini/gemini-2.5-pro` |
57
+ | `gemini/gemini-3-flash-preview` |
58
+ | `gemini/gemini-3-pro-image-preview` |
59
+ | `gemini/gemini-3.1-flash-image-preview` |
60
+ | `gemini/gemini-3.1-flash-lite-preview` |
61
+ | `gemini/gemini-3.1-pro-preview` |
62
+ | `gemini/gemini-3.1-pro-preview-customtools` |
63
+ | `gemini/gemini-flash-latest` |
64
+ | `gemini/gemini-flash-lite-latest` |
65
+ | `openai/gpt-4.1` |
66
+ | `openai/gpt-4.1-mini` |
67
+ | `openai/gpt-4.1-nano` |
68
+ | `openai/gpt-4o` |
69
+ | `openai/gpt-4o-mini` |
70
+ | `openai/gpt-5` |
71
+ | `openai/gpt-5-2025-08-07` |
72
+ | `openai/gpt-5-codex` |
73
+ | `openai/gpt-5-mini` |
74
+ | `openai/gpt-5-mini-2025-08-07` |
75
+ | `openai/gpt-5-nano` |
76
+ | `openai/gpt-5-pro` |
77
+ | `openai/gpt-5.1` |
78
+ | `openai/gpt-5.1-2025-11-13` |
79
+ | `openai/gpt-5.1-codex` |
80
+ | `openai/gpt-5.1-codex-max` |
81
+ | `openai/gpt-5.1-codex-mini` |
82
+ | `openai/gpt-5.2` |
83
+ | `openai/gpt-5.2-2025-12-11` |
84
+ | `openai/gpt-5.2-codex` |
85
+ | `openai/gpt-5.2-pro` |
86
+ | `openai/gpt-5.2-pro-2025-12-11` |
87
+ | `openai/gpt-5.3-chat-latest` |
88
+ | `openai/gpt-5.3-codex` |
89
+ | `openai/gpt-5.4` |
90
+ | `openai/gpt-5.4-2026-03-05` |
91
+ | `openai/gpt-5.4-mini` |
92
+ | `openai/gpt-5.4-mini-2026-03-17` |
93
+ | `openai/gpt-5.4-nano` |
94
+ | `openai/gpt-5.4-nano-2026-03-17` |
95
+ | `openai/gpt-5.4-pro` |
96
+ | `openai/gpt-5.4-pro-2026-03-05` |
97
+ | `openai/o3` |
98
+ | `openai/o3-mini` |
99
+ | `openai/o4-mini` |
@@ -1,6 +1,6 @@
1
1
  # ![OpenRouter logo](https://models.dev/logos/openrouter.svg)OpenRouter
2
2
 
3
- OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 165 models through Mastra's model router.
3
+ OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 166 models through Mastra's model router.
4
4
 
5
5
  Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
6
6
 
@@ -172,6 +172,7 @@ ANTHROPIC_API_KEY=ant-...
172
172
  | `qwen/qwen3-next-80b-a3b-thinking` |
173
173
  | `qwen/qwen3.5-397b-a17b` |
174
174
  | `qwen/qwen3.5-plus-02-15` |
175
+ | `qwen/qwen3.6-plus-preview:free` |
175
176
  | `sourceful/riverflow-v2-fast-preview` |
176
177
  | `sourceful/riverflow-v2-max-preview` |
177
178
  | `sourceful/riverflow-v2-standard-preview` |
@@ -1,6 +1,6 @@
1
1
  # ![Vercel logo](https://models.dev/logos/vercel.svg)Vercel
2
2
 
3
- Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 224 models through Mastra's model router.
3
+ Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 226 models through Mastra's model router.
4
4
 
5
5
  Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
6
6
 
@@ -117,6 +117,7 @@ ANTHROPIC_API_KEY=ant-...
117
117
  | `inception/mercury-2` |
118
118
  | `inception/mercury-coder-small` |
119
119
  | `kwaipilot/kat-coder-pro-v1` |
120
+ | `kwaipilot/kat-coder-pro-v2` |
120
121
  | `meituan/longcat-flash-chat` |
121
122
  | `meituan/longcat-flash-thinking` |
122
123
  | `meituan/longcat-flash-thinking-2601` |
@@ -162,6 +163,7 @@ ANTHROPIC_API_KEY=ant-...
162
163
  | `morph/morph-v3-fast` |
163
164
  | `morph/morph-v3-large` |
164
165
  | `nvidia/nemotron-3-nano-30b-a3b` |
166
+ | `nvidia/nemotron-3-super-120b-a12b` |
165
167
  | `nvidia/nemotron-nano-12b-v2-vl` |
166
168
  | `nvidia/nemotron-nano-9b-v2` |
167
169
  | `openai/codex-mini` |
@@ -1,6 +1,6 @@
1
1
  # Model Providers
2
2
 
3
- Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3616 models from 95 providers through a single API.
3
+ Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3613 models from 95 providers through a single API.
4
4
 
5
5
  ## Features
6
6
 
@@ -1,6 +1,6 @@
1
1
  # ![OpenCode Zen logo](https://models.dev/logos/opencode.svg)OpenCode Zen
2
2
 
3
- Access 33 OpenCode Zen models through Mastra's model router. Authentication is handled automatically using the `OPENCODE_API_KEY` environment variable.
3
+ Access 34 OpenCode Zen models through Mastra's model router. Authentication is handled automatically using the `OPENCODE_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [OpenCode Zen documentation](https://opencode.ai/docs/zen).
6
6
 
@@ -67,6 +67,7 @@ for await (const chunk of stream) {
67
67
  | `opencode/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
68
68
  | `opencode/minimax-m2.5-free` | 205K | | | | | | — | — |
69
69
  | `opencode/nemotron-3-super-free` | 1.0M | | | | | | — | — |
70
+ | `opencode/qwen3.6-plus-free` | 1.0M | | | | | | — | — |
70
71
 
71
72
  ## Advanced configuration
72
73
 
@@ -96,7 +97,7 @@ const agent = new Agent({
96
97
  model: ({ requestContext }) => {
97
98
  const useAdvanced = requestContext.task === "complex";
98
99
  return useAdvanced
99
- ? "opencode/nemotron-3-super-free"
100
+ ? "opencode/qwen3.6-plus-free"
100
101
  : "opencode/big-pickle";
101
102
  }
102
103
  });
@@ -1,6 +1,6 @@
1
1
  # ![Vultr logo](https://models.dev/logos/vultr.svg)Vultr
2
2
 
3
- Access 9 Vultr models through Mastra's model router. Authentication is handled automatically using the `VULTR_API_KEY` environment variable.
3
+ Access 4 Vultr models through Mastra's model router. Authentication is handled automatically using the `VULTR_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Vultr documentation](https://api.vultrinference.com/).
6
6
 
@@ -15,7 +15,7 @@ const agent = new Agent({
15
15
  id: "my-agent",
16
16
  name: "My Agent",
17
17
  instructions: "You are a helpful assistant",
18
- model: "vultr/DeepSeek-R1-Distill-Llama-70B"
18
+ model: "vultr/DeepSeek-V3.2"
19
19
  });
20
20
 
21
21
  // Generate a response
@@ -32,17 +32,12 @@ for await (const chunk of stream) {
32
32
 
33
33
  ## Models
34
34
 
35
- | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
- | ----------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
- | `vultr/DeepSeek-R1-Distill-Llama-70B` | 130K | | | | | | $2 | $2 |
38
- | `vultr/DeepSeek-R1-Distill-Qwen-32B` | 130K | | | | | | $0.30 | $0.30 |
39
- | `vultr/DeepSeek-V3.2` | 163K | | | | | | $0.55 | $2 |
40
- | `vultr/GLM-5-FP8` | 202K | | | | | | $0.85 | $3 |
41
- | `vultr/gpt-oss-120b` | 130K | | | | | | $0.15 | $0.60 |
42
- | `vultr/Kimi-K2.5` | 261K | | | | | | $0.55 | $3 |
43
- | `vultr/Llama-3_1-Nemotron-Ultra-253B-v1` | 32K | | | | | | $0.55 | $2 |
44
- | `vultr/MiniMax-M2.5` | 196K | | | | | | $0.30 | $1 |
45
- | `vultr/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4` | 260K | | | | | | $0.20 | $0.80 |
35
+ | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
+ | --------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
+ | `vultr/DeepSeek-V3.2` | 163K | | | | | | $0.55 | $2 |
38
+ | `vultr/GLM-5-FP8` | 202K | | | | | | $0.85 | $3 |
39
+ | `vultr/Kimi-K2.5` | 261K | | | | | | $0.55 | $3 |
40
+ | `vultr/MiniMax-M2.5` | 196K | | | | | | $0.30 | $1 |
46
41
 
47
42
  ## Advanced configuration
48
43
 
@@ -54,7 +49,7 @@ const agent = new Agent({
54
49
  name: "custom-agent",
55
50
  model: {
56
51
  url: "https://api.vultrinference.com/v1",
57
- id: "vultr/DeepSeek-R1-Distill-Llama-70B",
52
+ id: "vultr/DeepSeek-V3.2",
58
53
  apiKey: process.env.VULTR_API_KEY,
59
54
  headers: {
60
55
  "X-Custom-Header": "value"
@@ -72,8 +67,8 @@ const agent = new Agent({
72
67
  model: ({ requestContext }) => {
73
68
  const useAdvanced = requestContext.task === "complex";
74
69
  return useAdvanced
75
- ? "vultr/gpt-oss-120b"
76
- : "vultr/DeepSeek-R1-Distill-Llama-70B";
70
+ ? "vultr/MiniMax-M2.5"
71
+ : "vultr/DeepSeek-V3.2";
77
72
  }
78
73
  });
79
74
  ```
@@ -1,6 +1,6 @@
1
1
  # ![ZenMux logo](https://models.dev/logos/zenmux.svg)ZenMux
2
2
 
3
- Access 86 ZenMux models through Mastra's model router. Authentication is handled automatically using the `ZENMUX_API_KEY` environment variable.
3
+ Access 85 ZenMux models through Mastra's model router. Authentication is handled automatically using the `ZENMUX_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [ZenMux documentation](https://docs.zenmux.ai).
6
6
 
@@ -52,7 +52,6 @@ for await (const chunk of stream) {
52
52
  | `zenmux/google/gemini-2.5-flash-lite` | 1.0M | | | | | | $0.10 | $0.40 |
53
53
  | `zenmux/google/gemini-2.5-pro` | 1.0M | | | | | | $1 | $10 |
54
54
  | `zenmux/google/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
55
- | `zenmux/google/gemini-3-pro-image-preview` | 1.0M | | | | | | $2 | $12 |
56
55
  | `zenmux/google/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
57
56
  | `zenmux/google/gemini-3.1-flash-lite-preview` | 1.1M | | | | | | $0.25 | $2 |
58
57
  | `zenmux/google/gemini-3.1-pro-preview` | 1.0M | | | | | | $2 | $12 |
@@ -130,7 +129,7 @@ const agent = new Agent({
130
129
  id: "custom-agent",
131
130
  name: "custom-agent",
132
131
  model: {
133
- url: "https://zenmux.ai/api/anthropic/v1",
132
+ url: "https://zenmux.ai/api/v1",
134
133
  id: "zenmux/anthropic/claude-3.5-haiku",
135
134
  apiKey: process.env.ZENMUX_API_KEY,
136
135
  headers: {
@@ -153,32 +152,4 @@ const agent = new Agent({
153
152
  : "zenmux/anthropic/claude-3.5-haiku";
154
153
  }
155
154
  });
156
- ```
157
-
158
- ## Direct provider installation
159
-
160
- This provider can also be installed directly as a standalone package, which can be used instead of the Mastra model router string. View the [package documentation](https://www.npmjs.com/package/@ai-sdk/anthropic) for more details.
161
-
162
- **npm**:
163
-
164
- ```bash
165
- npm install @ai-sdk/anthropic
166
- ```
167
-
168
- **pnpm**:
169
-
170
- ```bash
171
- pnpm add @ai-sdk/anthropic
172
- ```
173
-
174
- **Yarn**:
175
-
176
- ```bash
177
- yarn add @ai-sdk/anthropic
178
- ```
179
-
180
- **Bun**:
181
-
182
- ```bash
183
- bun add @ai-sdk/anthropic
184
155
  ```
package/CHANGELOG.md CHANGED
@@ -1,5 +1,12 @@
1
1
  # @mastra/mcp-docs-server
2
2
 
3
+ ## 1.1.18-alpha.0
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [[`180aaaf`](https://github.com/mastra-ai/mastra/commit/180aaaf4d0903d33a49bc72de2d40ca69a5bc599)]:
8
+ - @mastra/core@1.18.1-alpha.0
9
+
3
10
  ## 1.1.17
4
11
 
5
12
  ### Patch Changes
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mastra/mcp-docs-server",
3
- "version": "1.1.17",
3
+ "version": "1.1.18-alpha.1",
4
4
  "description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
@@ -29,7 +29,7 @@
29
29
  "jsdom": "^26.1.0",
30
30
  "local-pkg": "^1.1.2",
31
31
  "zod": "^4.3.6",
32
- "@mastra/core": "1.18.0",
32
+ "@mastra/core": "1.18.1-alpha.0",
33
33
  "@mastra/mcp": "^1.3.2"
34
34
  },
35
35
  "devDependencies": {
@@ -46,9 +46,9 @@
46
46
  "tsx": "^4.21.0",
47
47
  "typescript": "^5.9.3",
48
48
  "vitest": "4.0.18",
49
+ "@internal/lint": "0.0.75",
49
50
  "@internal/types-builder": "0.0.50",
50
- "@mastra/core": "1.18.0",
51
- "@internal/lint": "0.0.75"
51
+ "@mastra/core": "1.18.1-alpha.0"
52
52
  },
53
53
  "homepage": "https://mastra.ai",
54
54
  "repository": {