@mastra/mcp-docs-server 1.1.33-alpha.3 → 1.1.33-alpha.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/models/gateways/openrouter.md +3 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/alibaba.md +2 -1
- package/.docs/models/providers/cortecs.md +16 -2
- package/.docs/models/providers/firmware.md +7 -6
- package/.docs/models/providers/neuralwatt.md +84 -0
- package/.docs/models/providers/nvidia.md +2 -1
- package/.docs/models/providers/perplexity-agent.md +3 -1
- package/.docs/models/providers/synthetic.md +2 -1
- package/.docs/models/providers/zhipuai-coding-plan.md +9 -8
- package/.docs/models/providers.md +1 -0
- package/package.json +5 -5
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenRouter
|
|
2
2
|
|
|
3
|
-
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 184 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
|
|
6
6
|
|
|
@@ -168,6 +168,8 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
168
168
|
| `openrouter/free` |
|
|
169
169
|
| `openrouter/owl-alpha` |
|
|
170
170
|
| `openrouter/pareto-code` |
|
|
171
|
+
| `poolside/laguna-m.1:free` |
|
|
172
|
+
| `poolside/laguna-xs.2:free` |
|
|
171
173
|
| `prime-intellect/intellect-3` |
|
|
172
174
|
| `qwen/qwen-2.5-coder-32b-instruct` |
|
|
173
175
|
| `qwen/qwen2.5-vl-72b-instruct` |
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3869 models from 107 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Alibaba
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 48 Alibaba models through Mastra's model router. Authentication is handled automatically using the `DASHSCOPE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Alibaba documentation](https://www.alibabacloud.com/help/en/model-studio/models).
|
|
6
6
|
|
|
@@ -79,6 +79,7 @@ for await (const chunk of stream) {
|
|
|
79
79
|
| `alibaba/qwen3.5-plus` | 1.0M | | | | | | $0.40 | $2 |
|
|
80
80
|
| `alibaba/qwen3.6-27b` | 262K | | | | | | $0.60 | $4 |
|
|
81
81
|
| `alibaba/qwen3.6-35b-a3b` | 262K | | | | | | $0.25 | $1 |
|
|
82
|
+
| `alibaba/qwen3.6-max-preview` | 262K | | | | | | $1 | $8 |
|
|
82
83
|
| `alibaba/qwen3.6-plus` | 1.0M | | | | | | $0.28 | $2 |
|
|
83
84
|
| `alibaba/qwq-plus` | 131K | | | | | | $0.80 | $2 |
|
|
84
85
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Cortecs
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 49 Cortecs models through Mastra's model router. Authentication is handled automatically using the `CORTECS_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Cortecs documentation](https://cortecs.ai).
|
|
6
6
|
|
|
@@ -41,7 +41,12 @@ for await (const chunk of stream) {
|
|
|
41
41
|
| `cortecs/claude-opus4-6` | 1.0M | | | | | | $6 | $30 |
|
|
42
42
|
| `cortecs/claude-opus4-7` | 1.0M | | | | | | $6 | $28 |
|
|
43
43
|
| `cortecs/claude-sonnet-4` | 200K | | | | | | $3 | $17 |
|
|
44
|
+
| `cortecs/codestral-2508` | 256K | | | | | | $0.30 | $0.90 |
|
|
45
|
+
| `cortecs/deepseek-r1-0528` | 164K | | | | | | $0.58 | $2 |
|
|
44
46
|
| `cortecs/deepseek-v3-0324` | 128K | | | | | | $0.55 | $2 |
|
|
47
|
+
| `cortecs/deepseek-v3.2` | 164K | | | | | | $0.27 | $0.44 |
|
|
48
|
+
| `cortecs/deepseek-v4-flash` | 1.0M | | | | | | $0.13 | $0.27 |
|
|
49
|
+
| `cortecs/deepseek-v4-pro` | 1.0M | | | | | | $2 | $3 |
|
|
45
50
|
| `cortecs/devstral-2512` | 262K | | | | | | — | — |
|
|
46
51
|
| `cortecs/devstral-small-2512` | 262K | | | | | | — | — |
|
|
47
52
|
| `cortecs/gemini-2.5-pro` | 1.0M | | | | | | $2 | $11 |
|
|
@@ -53,22 +58,31 @@ for await (const chunk of stream) {
|
|
|
53
58
|
| `cortecs/glm-5.1` | 205K | | | | | | $1 | $4 |
|
|
54
59
|
| `cortecs/gpt-4.1` | 1.0M | | | | | | $2 | $9 |
|
|
55
60
|
| `cortecs/gpt-oss-120b` | 128K | | | | | | — | — |
|
|
61
|
+
| `cortecs/hermes-4-70b` | 128K | | | | | | $0.12 | $0.36 |
|
|
56
62
|
| `cortecs/intellect-3` | 128K | | | | | | $0.22 | $1 |
|
|
57
63
|
| `cortecs/kimi-k2-instruct` | 131K | | | | | | $0.55 | $3 |
|
|
58
64
|
| `cortecs/kimi-k2-thinking` | 262K | | | | | | $0.66 | $3 |
|
|
59
65
|
| `cortecs/kimi-k2.5` | 256K | | | | | | $0.55 | $3 |
|
|
60
66
|
| `cortecs/kimi-k2.6` | 256K | | | | | | $0.81 | $4 |
|
|
61
67
|
| `cortecs/llama-3.1-405b-instruct` | 128K | | | | | | — | — |
|
|
68
|
+
| `cortecs/llama-3.3-70b-instruct` | 131K | | | | | | $0.09 | $0.28 |
|
|
62
69
|
| `cortecs/minimax-m2` | 400K | | | | | | $0.39 | $2 |
|
|
63
70
|
| `cortecs/minimax-m2.1` | 196K | | | | | | $0.34 | $1 |
|
|
64
71
|
| `cortecs/minimax-m2.5` | 197K | | | | | | $0.32 | $1 |
|
|
65
72
|
| `cortecs/minimax-m2.7` | 203K | | | | | | $0.47 | $1 |
|
|
73
|
+
| `cortecs/mistral-large-2512` | 256K | | | | | | $0.50 | $2 |
|
|
74
|
+
| `cortecs/mixtral-8x7B-instruct-v0.1` | 32K | | | | | | $0.44 | $0.68 |
|
|
75
|
+
| `cortecs/nemotron-3-super-120b-a12b` | 262K | | | | | | $0.27 | $0.80 |
|
|
66
76
|
| `cortecs/nova-pro-v1` | 300K | | | | | | $1 | $4 |
|
|
67
77
|
| `cortecs/qwen-2.5-72b-instruct` | 33K | | | | | | $0.06 | $0.23 |
|
|
78
|
+
| `cortecs/qwen3-235b-a22b-instruct-2507` | 131K | | | | | | $0.06 | $0.41 |
|
|
68
79
|
| `cortecs/qwen3-32b` | 16K | | | | | | $0.10 | $0.33 |
|
|
80
|
+
| `cortecs/qwen3-coder-30b-a3b-instruct` | 262K | | | | | | $0.05 | $0.22 |
|
|
69
81
|
| `cortecs/qwen3-coder-480b-a35b-instruct` | 262K | | | | | | $0.44 | $2 |
|
|
70
82
|
| `cortecs/qwen3-coder-next` | 256K | | | | | | $0.16 | $0.84 |
|
|
71
83
|
| `cortecs/qwen3-next-80b-a3b-thinking` | 128K | | | | | | $0.16 | $1 |
|
|
84
|
+
| `cortecs/qwen3.5-122b-a10b` | 262K | | | | | | $0.44 | $3 |
|
|
85
|
+
| `cortecs/qwen3.5-397b-a17b` | 250K | | | | | | $0.60 | $4 |
|
|
72
86
|
|
|
73
87
|
## Advanced configuration
|
|
74
88
|
|
|
@@ -98,7 +112,7 @@ const agent = new Agent({
|
|
|
98
112
|
model: ({ requestContext }) => {
|
|
99
113
|
const useAdvanced = requestContext.task === "complex";
|
|
100
114
|
return useAdvanced
|
|
101
|
-
? "cortecs/qwen3-
|
|
115
|
+
? "cortecs/qwen3.5-397b-a17b"
|
|
102
116
|
: "cortecs/claude-4-5-sonnet";
|
|
103
117
|
}
|
|
104
118
|
});
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Firmware
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 26 Firmware models through Mastra's model router. Authentication is handled automatically using the `FIRMWARE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Firmware documentation](https://docs.frogbot.ai).
|
|
6
6
|
|
|
@@ -38,25 +38,26 @@ for await (const chunk of stream) {
|
|
|
38
38
|
| `firmware/claude-opus-4-6` | 200K | | | | | | $5 | $25 |
|
|
39
39
|
| `firmware/claude-opus-4-7` | 200K | | | | | | $5 | $25 |
|
|
40
40
|
| `firmware/claude-sonnet-4-6` | 200K | | | | | | $3 | $15 |
|
|
41
|
-
| `firmware/deepseek-
|
|
41
|
+
| `firmware/deepseek-v4-pro` | 128K | | | | | | $2 | $3 |
|
|
42
42
|
| `firmware/gemini-2.5-flash` | 1.0M | | | | | | $0.30 | $3 |
|
|
43
43
|
| `firmware/gemini-2.5-pro` | 1.0M | | | | | | $1 | $10 |
|
|
44
44
|
| `firmware/gemini-3-1-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
45
45
|
| `firmware/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
|
|
46
|
-
| `firmware/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
47
46
|
| `firmware/gpt-4o` | 128K | | | | | | $3 | $10 |
|
|
48
47
|
| `firmware/gpt-5-3-codex` | 400K | | | | | | $2 | $14 |
|
|
49
|
-
| `firmware/gpt-5-4`
|
|
50
|
-
| `firmware/gpt-5-
|
|
51
|
-
| `firmware/gpt-5-
|
|
48
|
+
| `firmware/gpt-5-4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
49
|
+
| `firmware/gpt-5-4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
50
|
+
| `firmware/gpt-5-5` | 272K | | | | | | $3 | $15 |
|
|
52
51
|
| `firmware/gpt-oss-120b` | 131K | | | | | | $0.15 | $0.60 |
|
|
53
52
|
| `firmware/gpt-oss-20b` | 131K | | | | | | $0.07 | $0.20 |
|
|
54
53
|
| `firmware/grok-4-1-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
55
54
|
| `firmware/grok-4-1-fast-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
55
|
+
| `firmware/grok-4-3` | 1.0M | | | | | | $1 | $3 |
|
|
56
56
|
| `firmware/grok-code-fast-1` | 256K | | | | | | $0.20 | $2 |
|
|
57
57
|
| `firmware/kimi-k2-6` | 256K | | | | | | $0.95 | $4 |
|
|
58
58
|
| `firmware/kimi-k2.5` | 256K | | | | | | $0.60 | $3 |
|
|
59
59
|
| `firmware/minimax-m2-5` | 192K | | | | | | $0.30 | $1 |
|
|
60
|
+
| `firmware/minimax-m2-7` | 192K | | | | | | $0.30 | $1 |
|
|
60
61
|
| `firmware/qwen-3-6-plus` | 1.0M | | | | | | $0.50 | $3 |
|
|
61
62
|
| `firmware/zai-glm-5-1` | 198K | | | | | | $1 | $4 |
|
|
62
63
|
|
|
@@ -0,0 +1,84 @@
|
|
|
1
|
+
# Neuralwatt
|
|
2
|
+
|
|
3
|
+
Access 14 Neuralwatt models through Mastra's model router. Authentication is handled automatically using the `NEURALWATT_API_KEY` environment variable.
|
|
4
|
+
|
|
5
|
+
Learn more in the [Neuralwatt documentation](https://portal.neuralwatt.com/docs).
|
|
6
|
+
|
|
7
|
+
```bash
|
|
8
|
+
NEURALWATT_API_KEY=your-api-key
|
|
9
|
+
```
|
|
10
|
+
|
|
11
|
+
```typescript
|
|
12
|
+
import { Agent } from "@mastra/core/agent";
|
|
13
|
+
|
|
14
|
+
const agent = new Agent({
|
|
15
|
+
id: "my-agent",
|
|
16
|
+
name: "My Agent",
|
|
17
|
+
instructions: "You are a helpful assistant",
|
|
18
|
+
model: "neuralwatt/MiniMaxAI/MiniMax-M2.5"
|
|
19
|
+
});
|
|
20
|
+
|
|
21
|
+
// Generate a response
|
|
22
|
+
const response = await agent.generate("Hello!");
|
|
23
|
+
|
|
24
|
+
// Stream a response
|
|
25
|
+
const stream = await agent.stream("Tell me a story");
|
|
26
|
+
for await (const chunk of stream) {
|
|
27
|
+
console.log(chunk);
|
|
28
|
+
}
|
|
29
|
+
```
|
|
30
|
+
|
|
31
|
+
> **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [Neuralwatt documentation](https://portal.neuralwatt.com/docs) for details.
|
|
32
|
+
|
|
33
|
+
## Models
|
|
34
|
+
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| --------------------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `neuralwatt/glm-5-fast` | 200K | | | | | | $1 | $4 |
|
|
38
|
+
| `neuralwatt/glm-5.1-fast` | 200K | | | | | | $1 | $4 |
|
|
39
|
+
| `neuralwatt/kimi-k2.5-fast` | 262K | | | | | | $0.52 | $3 |
|
|
40
|
+
| `neuralwatt/kimi-k2.6-fast` | 262K | | | | | | $0.69 | $3 |
|
|
41
|
+
| `neuralwatt/MiniMaxAI/MiniMax-M2.5` | 197K | | | | | | $0.35 | $1 |
|
|
42
|
+
| `neuralwatt/mistralai/Devstral-Small-2-24B-Instruct-2512` | 262K | | | | | | $0.12 | $0.35 |
|
|
43
|
+
| `neuralwatt/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.52 | $3 |
|
|
44
|
+
| `neuralwatt/moonshotai/Kimi-K2.6` | 262K | | | | | | $0.69 | $3 |
|
|
45
|
+
| `neuralwatt/openai/gpt-oss-20b` | 16K | | | | | | $0.03 | $0.16 |
|
|
46
|
+
| `neuralwatt/Qwen/Qwen3.5-397B-A17B-FP8` | 262K | | | | | | $0.69 | $4 |
|
|
47
|
+
| `neuralwatt/Qwen/Qwen3.6-35B-A3B` | 131K | | | | | | $0.05 | $0.10 |
|
|
48
|
+
| `neuralwatt/qwen3.5-397b-fast` | 262K | | | | | | $0.69 | $4 |
|
|
49
|
+
| `neuralwatt/qwen3.6-35b-fast` | 131K | | | | | | $0.05 | $0.10 |
|
|
50
|
+
| `neuralwatt/zai-org/GLM-5.1-FP8` | 200K | | | | | | $1 | $4 |
|
|
51
|
+
|
|
52
|
+
## Advanced configuration
|
|
53
|
+
|
|
54
|
+
### Custom headers
|
|
55
|
+
|
|
56
|
+
```typescript
|
|
57
|
+
const agent = new Agent({
|
|
58
|
+
id: "custom-agent",
|
|
59
|
+
name: "custom-agent",
|
|
60
|
+
model: {
|
|
61
|
+
url: "https://api.neuralwatt.com/v1",
|
|
62
|
+
id: "neuralwatt/MiniMaxAI/MiniMax-M2.5",
|
|
63
|
+
apiKey: process.env.NEURALWATT_API_KEY,
|
|
64
|
+
headers: {
|
|
65
|
+
"X-Custom-Header": "value"
|
|
66
|
+
}
|
|
67
|
+
}
|
|
68
|
+
});
|
|
69
|
+
```
|
|
70
|
+
|
|
71
|
+
### Dynamic model selection
|
|
72
|
+
|
|
73
|
+
```typescript
|
|
74
|
+
const agent = new Agent({
|
|
75
|
+
id: "dynamic-agent",
|
|
76
|
+
name: "Dynamic Agent",
|
|
77
|
+
model: ({ requestContext }) => {
|
|
78
|
+
const useAdvanced = requestContext.task === "complex";
|
|
79
|
+
return useAdvanced
|
|
80
|
+
? "neuralwatt/zai-org/GLM-5.1-FP8"
|
|
81
|
+
: "neuralwatt/MiniMaxAI/MiniMax-M2.5";
|
|
82
|
+
}
|
|
83
|
+
});
|
|
84
|
+
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Nvidia
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 82 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Nvidia documentation](https://docs.api.nvidia.com/nim/).
|
|
6
6
|
|
|
@@ -86,6 +86,7 @@ for await (const chunk of stream) {
|
|
|
86
86
|
| `nvidia/moonshotai/kimi-k2-instruct-0905` | 262K | | | | | | — | — |
|
|
87
87
|
| `nvidia/moonshotai/kimi-k2-thinking` | 262K | | | | | | — | — |
|
|
88
88
|
| `nvidia/moonshotai/kimi-k2.5` | 262K | | | | | | — | — |
|
|
89
|
+
| `nvidia/moonshotai/kimi-k2.6` | 262K | | | | | | — | — |
|
|
89
90
|
| `nvidia/nvidia/cosmos-nemotron-34b` | 131K | | | | | | — | — |
|
|
90
91
|
| `nvidia/nvidia/llama-3.1-nemotron-51b-instruct` | 128K | | | | | | — | — |
|
|
91
92
|
| `nvidia/nvidia/llama-3.1-nemotron-70b-instruct` | 128K | | | | | | — | — |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Perplexity Agent
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 18 Perplexity Agent models through Mastra's model router. Authentication is handled automatically using the `PERPLEXITY_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Perplexity Agent documentation](https://docs.perplexity.ai/docs/agent-api/models).
|
|
6
6
|
|
|
@@ -37,6 +37,7 @@ for await (const chunk of stream) {
|
|
|
37
37
|
| `perplexity-agent/anthropic/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
|
|
38
38
|
| `perplexity-agent/anthropic/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
|
|
39
39
|
| `perplexity-agent/anthropic/claude-opus-4-6` | 200K | | | | | | $5 | $25 |
|
|
40
|
+
| `perplexity-agent/anthropic/claude-opus-4-7` | 1.0M | | | | | | $5 | $25 |
|
|
40
41
|
| `perplexity-agent/anthropic/claude-sonnet-4-5` | 200K | | | | | | $3 | $15 |
|
|
41
42
|
| `perplexity-agent/anthropic/claude-sonnet-4-6` | 200K | | | | | | $3 | $15 |
|
|
42
43
|
| `perplexity-agent/google/gemini-2.5-flash` | 1.0M | | | | | | $0.30 | $3 |
|
|
@@ -48,6 +49,7 @@ for await (const chunk of stream) {
|
|
|
48
49
|
| `perplexity-agent/openai/gpt-5.1` | 400K | | | | | | $1 | $10 |
|
|
49
50
|
| `perplexity-agent/openai/gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
50
51
|
| `perplexity-agent/openai/gpt-5.4` | 1.1M | | | | | | $3 | $15 |
|
|
52
|
+
| `perplexity-agent/openai/gpt-5.5` | 1.1M | | | | | | $5 | $30 |
|
|
51
53
|
| `perplexity-agent/perplexity/sonar` | 128K | | | | | | $0.25 | $3 |
|
|
52
54
|
| `perplexity-agent/xai/grok-4-1-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
53
55
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Synthetic
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 33 Synthetic models through Mastra's model router. Authentication is handled automatically using the `SYNTHETIC_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Synthetic documentation](https://synthetic.new/pricing).
|
|
6
6
|
|
|
@@ -53,6 +53,7 @@ for await (const chunk of stream) {
|
|
|
53
53
|
| `synthetic/hf:moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $1 | $1 |
|
|
54
54
|
| `synthetic/hf:moonshotai/Kimi-K2-Thinking` | 262K | | | | | | $0.55 | $2 |
|
|
55
55
|
| `synthetic/hf:moonshotai/Kimi-K2.5` | 262K | | | | | | $0.55 | $2 |
|
|
56
|
+
| `synthetic/hf:moonshotai/Kimi-K2.6` | 262K | | | | | | $0.95 | $4 |
|
|
56
57
|
| `synthetic/hf:nvidia/Kimi-K2.5-NVFP4` | 262K | | | | | | $0.55 | $2 |
|
|
57
58
|
| `synthetic/hf:nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4` | 262K | | | | | | $0.30 | $1 |
|
|
58
59
|
| `synthetic/hf:openai/gpt-oss-120b` | 128K | | | | | | $0.10 | $0.10 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Zhipu AI Coding Plan
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 5 Zhipu AI Coding Plan models through Mastra's model router. Authentication is handled automatically using the `ZHIPU_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Zhipu AI Coding Plan documentation](https://docs.bigmodel.cn/cn/coding-plan/overview).
|
|
6
6
|
|
|
@@ -32,12 +32,13 @@ for await (const chunk of stream) {
|
|
|
32
32
|
|
|
33
33
|
## Models
|
|
34
34
|
|
|
35
|
-
| Model
|
|
36
|
-
|
|
|
37
|
-
| `zhipuai-coding-plan/glm-4.5-air`
|
|
38
|
-
| `zhipuai-coding-plan/glm-4.7`
|
|
39
|
-
| `zhipuai-coding-plan/glm-5-turbo`
|
|
40
|
-
| `zhipuai-coding-plan/glm-5.1`
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ---------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `zhipuai-coding-plan/glm-4.5-air` | 131K | | | | | | — | — |
|
|
38
|
+
| `zhipuai-coding-plan/glm-4.7` | 205K | | | | | | — | — |
|
|
39
|
+
| `zhipuai-coding-plan/glm-5-turbo` | 200K | | | | | | — | — |
|
|
40
|
+
| `zhipuai-coding-plan/glm-5.1` | 200K | | | | | | — | — |
|
|
41
|
+
| `zhipuai-coding-plan/glm-5v-turbo` | 200K | | | | | | — | — |
|
|
41
42
|
|
|
42
43
|
## Advanced configuration
|
|
43
44
|
|
|
@@ -67,7 +68,7 @@ const agent = new Agent({
|
|
|
67
68
|
model: ({ requestContext }) => {
|
|
68
69
|
const useAdvanced = requestContext.task === "complex";
|
|
69
70
|
return useAdvanced
|
|
70
|
-
? "zhipuai-coding-plan/glm-
|
|
71
|
+
? "zhipuai-coding-plan/glm-5v-turbo"
|
|
71
72
|
: "zhipuai-coding-plan/glm-4.5-air";
|
|
72
73
|
}
|
|
73
74
|
});
|
|
@@ -63,6 +63,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
|
|
|
63
63
|
- [Morph](https://mastra.ai/models/providers/morph)
|
|
64
64
|
- [NanoGPT](https://mastra.ai/models/providers/nano-gpt)
|
|
65
65
|
- [Nebius Token Factory](https://mastra.ai/models/providers/nebius)
|
|
66
|
+
- [Neuralwatt](https://mastra.ai/models/providers/neuralwatt)
|
|
66
67
|
- [Nova](https://mastra.ai/models/providers/nova)
|
|
67
68
|
- [NovitaAI](https://mastra.ai/models/providers/novita-ai)
|
|
68
69
|
- [Nvidia](https://mastra.ai/models/providers/nvidia)
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.33-alpha.
|
|
3
|
+
"version": "1.1.33-alpha.4",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,8 +29,8 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/
|
|
33
|
-
"@mastra/
|
|
32
|
+
"@mastra/mcp": "^1.6.1-alpha.0",
|
|
33
|
+
"@mastra/core": "1.32.0-alpha.1"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
36
36
|
"@hono/node-server": "^1.19.11",
|
|
@@ -46,9 +46,9 @@
|
|
|
46
46
|
"tsx": "^4.21.0",
|
|
47
47
|
"typescript": "^6.0.3",
|
|
48
48
|
"vitest": "4.1.5",
|
|
49
|
+
"@internal/types-builder": "0.0.65",
|
|
49
50
|
"@internal/lint": "0.0.90",
|
|
50
|
-
"@mastra/core": "1.32.0-alpha.1"
|
|
51
|
-
"@internal/types-builder": "0.0.65"
|
|
51
|
+
"@mastra/core": "1.32.0-alpha.1"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|