@mastra/mcp-docs-server 1.1.26-alpha.15 → 1.1.26-alpha.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/302ai.md +32 -1
- package/.docs/models/providers/berget.md +9 -12
- package/.docs/models/providers/cortecs.md +2 -1
- package/.docs/models/providers/wafer.ai.md +72 -0
- package/.docs/models/providers.md +1 -0
- package/CHANGELOG.md +7 -0
- package/package.json +3 -3
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3667 models from 104 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# 302.AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 95 302.AI models through Mastra's model router. Authentication is handled automatically using the `302AI_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [302.AI documentation](https://doc.302.ai).
|
|
6
6
|
|
|
@@ -35,13 +35,25 @@ for await (const chunk of stream) {
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| --------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
37
|
| `302ai/chatgpt-4o-latest` | 128K | | | | | | $5 | $15 |
|
|
38
|
+
| `302ai/claude-3-5-haiku-20241022` | 200K | | | | | | $0.80 | $4 |
|
|
39
|
+
| `302ai/claude-3-5-haiku-latest` | 200K | | | | | | $0.80 | $4 |
|
|
40
|
+
| `302ai/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
|
|
38
41
|
| `302ai/claude-haiku-4-5-20251001` | 200K | | | | | | $1 | $5 |
|
|
39
42
|
| `302ai/claude-opus-4-1-20250805` | 200K | | | | | | $15 | $75 |
|
|
40
43
|
| `302ai/claude-opus-4-1-20250805-thinking` | 200K | | | | | | $15 | $75 |
|
|
44
|
+
| `302ai/claude-opus-4-20250514` | 200K | | | | | | $15 | $75 |
|
|
45
|
+
| `302ai/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
|
|
41
46
|
| `302ai/claude-opus-4-5-20251101` | 200K | | | | | | $5 | $25 |
|
|
42
47
|
| `302ai/claude-opus-4-5-20251101-thinking` | 200K | | | | | | $5 | $25 |
|
|
48
|
+
| `302ai/claude-opus-4-6` | 1.0M | | | | | | $5 | $25 |
|
|
49
|
+
| `302ai/claude-opus-4-6-thinking` | 1.0M | | | | | | $5 | $25 |
|
|
50
|
+
| `302ai/claude-opus-4-7` | 200K | | | | | | $5 | $25 |
|
|
51
|
+
| `302ai/claude-sonnet-4-20250514` | 200K | | | | | | $3 | $15 |
|
|
52
|
+
| `302ai/claude-sonnet-4-5` | 200K | | | | | | $3 | $15 |
|
|
43
53
|
| `302ai/claude-sonnet-4-5-20250929` | 200K | | | | | | $3 | $15 |
|
|
44
54
|
| `302ai/claude-sonnet-4-5-20250929-thinking` | 200K | | | | | | $3 | $15 |
|
|
55
|
+
| `302ai/claude-sonnet-4-6` | 1.0M | | | | | | $3 | $15 |
|
|
56
|
+
| `302ai/claude-sonnet-4-6-thinking` | 1.0M | | | | | | $3 | $15 |
|
|
45
57
|
| `302ai/deepseek-chat` | 128K | | | | | | $0.29 | $0.43 |
|
|
46
58
|
| `302ai/deepseek-reasoner` | 128K | | | | | | $0.29 | $0.43 |
|
|
47
59
|
| `302ai/deepseek-v3.2` | 128K | | | | | | $0.29 | $0.43 |
|
|
@@ -60,11 +72,21 @@ for await (const chunk of stream) {
|
|
|
60
72
|
| `302ai/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
|
|
61
73
|
| `302ai/gemini-3-pro-image-preview` | 33K | | | | | | $2 | $120 |
|
|
62
74
|
| `302ai/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
75
|
+
| `302ai/gemini-3.1-flash-image-preview` | 131K | | | | | | $0.50 | $60 |
|
|
63
76
|
| `302ai/glm-4.5` | 128K | | | | | | $0.29 | $1 |
|
|
77
|
+
| `302ai/glm-4.5-air` | 128K | | | | | | $0.11 | $0.29 |
|
|
78
|
+
| `302ai/glm-4.5-airx` | 128K | | | | | | $0.57 | $2 |
|
|
79
|
+
| `302ai/glm-4.5-x` | 128K | | | | | | $1 | $2 |
|
|
64
80
|
| `302ai/glm-4.5v` | 64K | | | | | | $0.29 | $0.86 |
|
|
65
81
|
| `302ai/glm-4.6` | 200K | | | | | | $0.29 | $1 |
|
|
66
82
|
| `302ai/glm-4.6v` | 128K | | | | | | $0.14 | $0.43 |
|
|
67
83
|
| `302ai/glm-4.7` | 200K | | | | | | $0.29 | $1 |
|
|
84
|
+
| `302ai/glm-4.7-flashx` | 200K | | | | | | $0.07 | $0.43 |
|
|
85
|
+
| `302ai/glm-5` | 200K | | | | | | $0.60 | $3 |
|
|
86
|
+
| `302ai/glm-5-turbo` | 200K | | | | | | $0.72 | $3 |
|
|
87
|
+
| `302ai/glm-5.1` | 200K | | | | | | $0.86 | $4 |
|
|
88
|
+
| `302ai/glm-5v-turbo` | 200K | | | | | | $0.72 | $3 |
|
|
89
|
+
| `302ai/glm-for-coding` | 200K | | | | | | $0.09 | $0.34 |
|
|
68
90
|
| `302ai/gpt-4.1` | 1.0M | | | | | | $2 | $8 |
|
|
69
91
|
| `302ai/gpt-4.1-mini` | 1.0M | | | | | | $0.40 | $2 |
|
|
70
92
|
| `302ai/gpt-4.1-nano` | 1.0M | | | | | | $0.10 | $0.40 |
|
|
@@ -77,17 +99,26 @@ for await (const chunk of stream) {
|
|
|
77
99
|
| `302ai/gpt-5.1-chat-latest` | 128K | | | | | | $1 | $10 |
|
|
78
100
|
| `302ai/gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
79
101
|
| `302ai/gpt-5.2-chat-latest` | 128K | | | | | | $2 | $14 |
|
|
102
|
+
| `302ai/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
103
|
+
| `302ai/gpt-5.4-mini-2026-03-17` | 400K | | | | | | $0.75 | $5 |
|
|
104
|
+
| `302ai/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
105
|
+
| `302ai/gpt-5.4-nano-2026-03-17` | 400K | | | | | | $0.20 | $1 |
|
|
80
106
|
| `302ai/grok-4-1-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
81
107
|
| `302ai/grok-4-1-fast-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
82
108
|
| `302ai/grok-4-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
83
109
|
| `302ai/grok-4-fast-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
84
110
|
| `302ai/grok-4.1` | 200K | | | | | | $2 | $10 |
|
|
111
|
+
| `302ai/grok-4.20-beta-0309-non-reasoning` | 2.0M | | | | | | $2 | $6 |
|
|
112
|
+
| `302ai/grok-4.20-beta-0309-reasoning` | 2.0M | | | | | | $2 | $6 |
|
|
113
|
+
| `302ai/grok-4.20-multi-agent-beta-0309` | 2.0M | | | | | | $2 | $6 |
|
|
85
114
|
| `302ai/kimi-k2-0905-preview` | 262K | | | | | | $0.63 | $3 |
|
|
86
115
|
| `302ai/kimi-k2-thinking` | 262K | | | | | | $0.57 | $2 |
|
|
87
116
|
| `302ai/kimi-k2-thinking-turbo` | 262K | | | | | | $1 | $9 |
|
|
88
117
|
| `302ai/MiniMax-M1` | 1.0M | | | | | | $0.13 | $1 |
|
|
89
118
|
| `302ai/MiniMax-M2` | 1.0M | | | | | | $0.33 | $1 |
|
|
90
119
|
| `302ai/MiniMax-M2.1` | 1.0M | | | | | | $0.30 | $1 |
|
|
120
|
+
| `302ai/MiniMax-M2.7` | 205K | | | | | | $0.30 | $1 |
|
|
121
|
+
| `302ai/MiniMax-M2.7-highspeed` | 205K | | | | | | $0.60 | $5 |
|
|
91
122
|
| `302ai/ministral-14b-2512` | 128K | | | | | | $0.33 | $0.33 |
|
|
92
123
|
| `302ai/mistral-large-2512` | 128K | | | | | | $1 | $3 |
|
|
93
124
|
| `302ai/qwen-flash` | 1.0M | | | | | | $0.02 | $0.22 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Berget.AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 5 Berget.AI models through Mastra's model router. Authentication is handled automatically using the `BERGET_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Berget.AI documentation](https://api.berget.ai).
|
|
6
6
|
|
|
@@ -15,7 +15,7 @@ const agent = new Agent({
|
|
|
15
15
|
id: "my-agent",
|
|
16
16
|
name: "My Agent",
|
|
17
17
|
instructions: "You are a helpful assistant",
|
|
18
|
-
model: "berget/
|
|
18
|
+
model: "berget/google/gemma-4-31B-it"
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Generate a response
|
|
@@ -34,14 +34,11 @@ for await (const chunk of stream) {
|
|
|
34
34
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| ------------------------------------------------------ | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
-
| `berget/
|
|
38
|
-
| `berget/
|
|
39
|
-
| `berget/
|
|
40
|
-
| `berget/
|
|
41
|
-
| `berget/
|
|
42
|
-
| `berget/mistralai/Mistral-Small-3.2-24B-Instruct-2506` | 32K | | | | | | $0.30 | $0.30 |
|
|
43
|
-
| `berget/openai/gpt-oss-120b` | 128K | | | | | | $0.30 | $0.90 |
|
|
44
|
-
| `berget/zai-org/GLM-4.7` | 128K | | | | | | $0.70 | $2 |
|
|
37
|
+
| `berget/google/gemma-4-31B-it` | 128K | | | | | | $0.28 | $0.55 |
|
|
38
|
+
| `berget/meta-llama/Llama-3.3-70B-Instruct` | 128K | | | | | | $0.99 | $0.99 |
|
|
39
|
+
| `berget/mistralai/Mistral-Small-3.2-24B-Instruct-2506` | 32K | | | | | | $0.33 | $0.33 |
|
|
40
|
+
| `berget/openai/gpt-oss-120b` | 128K | | | | | | $0.44 | $0.99 |
|
|
41
|
+
| `berget/zai-org/GLM-4.7` | 128K | | | | | | $0.77 | $3 |
|
|
45
42
|
|
|
46
43
|
## Advanced configuration
|
|
47
44
|
|
|
@@ -53,7 +50,7 @@ const agent = new Agent({
|
|
|
53
50
|
name: "custom-agent",
|
|
54
51
|
model: {
|
|
55
52
|
url: "https://api.berget.ai/v1",
|
|
56
|
-
id: "berget/
|
|
53
|
+
id: "berget/google/gemma-4-31B-it",
|
|
57
54
|
apiKey: process.env.BERGET_API_KEY,
|
|
58
55
|
headers: {
|
|
59
56
|
"X-Custom-Header": "value"
|
|
@@ -72,7 +69,7 @@ const agent = new Agent({
|
|
|
72
69
|
const useAdvanced = requestContext.task === "complex";
|
|
73
70
|
return useAdvanced
|
|
74
71
|
? "berget/zai-org/GLM-4.7"
|
|
75
|
-
: "berget/
|
|
72
|
+
: "berget/google/gemma-4-31B-it";
|
|
76
73
|
}
|
|
77
74
|
});
|
|
78
75
|
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Cortecs
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 33 Cortecs models through Mastra's model router. Authentication is handled automatically using the `CORTECS_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Cortecs documentation](https://cortecs.ai).
|
|
6
6
|
|
|
@@ -39,6 +39,7 @@ for await (const chunk of stream) {
|
|
|
39
39
|
| `cortecs/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
|
|
40
40
|
| `cortecs/claude-opus4-5` | 200K | | | | | | $6 | $30 |
|
|
41
41
|
| `cortecs/claude-opus4-6` | 1.0M | | | | | | $6 | $30 |
|
|
42
|
+
| `cortecs/claude-opus4-7` | 1.0M | | | | | | $6 | $28 |
|
|
42
43
|
| `cortecs/claude-sonnet-4` | 200K | | | | | | $3 | $17 |
|
|
43
44
|
| `cortecs/deepseek-v3-0324` | 128K | | | | | | $0.55 | $2 |
|
|
44
45
|
| `cortecs/devstral-2512` | 262K | | | | | | — | — |
|
|
@@ -0,0 +1,72 @@
|
|
|
1
|
+
# Wafer
|
|
2
|
+
|
|
3
|
+
Access 2 Wafer models through Mastra's model router. Authentication is handled automatically using the `WAFER_API_KEY` environment variable.
|
|
4
|
+
|
|
5
|
+
Learn more in the [Wafer documentation](https://docs.wafer.ai/wafer-pass).
|
|
6
|
+
|
|
7
|
+
```bash
|
|
8
|
+
WAFER_API_KEY=your-api-key
|
|
9
|
+
```
|
|
10
|
+
|
|
11
|
+
```typescript
|
|
12
|
+
import { Agent } from "@mastra/core/agent";
|
|
13
|
+
|
|
14
|
+
const agent = new Agent({
|
|
15
|
+
id: "my-agent",
|
|
16
|
+
name: "My Agent",
|
|
17
|
+
instructions: "You are a helpful assistant",
|
|
18
|
+
model: "wafer.ai/GLM-5.1"
|
|
19
|
+
});
|
|
20
|
+
|
|
21
|
+
// Generate a response
|
|
22
|
+
const response = await agent.generate("Hello!");
|
|
23
|
+
|
|
24
|
+
// Stream a response
|
|
25
|
+
const stream = await agent.stream("Tell me a story");
|
|
26
|
+
for await (const chunk of stream) {
|
|
27
|
+
console.log(chunk);
|
|
28
|
+
}
|
|
29
|
+
```
|
|
30
|
+
|
|
31
|
+
> **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [Wafer documentation](https://docs.wafer.ai/wafer-pass) for details.
|
|
32
|
+
|
|
33
|
+
## Models
|
|
34
|
+
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ---------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `wafer.ai/GLM-5.1` | 203K | | | | | | — | — |
|
|
38
|
+
| `wafer.ai/Qwen3.5-397B-A17B` | 262K | | | | | | — | — |
|
|
39
|
+
|
|
40
|
+
## Advanced configuration
|
|
41
|
+
|
|
42
|
+
### Custom headers
|
|
43
|
+
|
|
44
|
+
```typescript
|
|
45
|
+
const agent = new Agent({
|
|
46
|
+
id: "custom-agent",
|
|
47
|
+
name: "custom-agent",
|
|
48
|
+
model: {
|
|
49
|
+
url: "https://pass.wafer.ai/v1",
|
|
50
|
+
id: "wafer.ai/GLM-5.1",
|
|
51
|
+
apiKey: process.env.WAFER_API_KEY,
|
|
52
|
+
headers: {
|
|
53
|
+
"X-Custom-Header": "value"
|
|
54
|
+
}
|
|
55
|
+
}
|
|
56
|
+
});
|
|
57
|
+
```
|
|
58
|
+
|
|
59
|
+
### Dynamic model selection
|
|
60
|
+
|
|
61
|
+
```typescript
|
|
62
|
+
const agent = new Agent({
|
|
63
|
+
id: "dynamic-agent",
|
|
64
|
+
name: "Dynamic Agent",
|
|
65
|
+
model: ({ requestContext }) => {
|
|
66
|
+
const useAdvanced = requestContext.task === "complex";
|
|
67
|
+
return useAdvanced
|
|
68
|
+
? "wafer.ai/Qwen3.5-397B-A17B"
|
|
69
|
+
: "wafer.ai/GLM-5.1";
|
|
70
|
+
}
|
|
71
|
+
});
|
|
72
|
+
```
|
|
@@ -91,6 +91,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
|
|
|
91
91
|
- [Upstage](https://mastra.ai/models/providers/upstage)
|
|
92
92
|
- [Vivgrid](https://mastra.ai/models/providers/vivgrid)
|
|
93
93
|
- [Vultr](https://mastra.ai/models/providers/vultr)
|
|
94
|
+
- [Wafer](https://mastra.ai/models/providers/wafer.ai)
|
|
94
95
|
- [Weights & Biases](https://mastra.ai/models/providers/wandb)
|
|
95
96
|
- [Xiaomi](https://mastra.ai/models/providers/xiaomi)
|
|
96
97
|
- [Xiaomi Token Plan (China)](https://mastra.ai/models/providers/xiaomi-token-plan-cn)
|
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,12 @@
|
|
|
1
1
|
# @mastra/mcp-docs-server
|
|
2
2
|
|
|
3
|
+
## 1.1.26-alpha.16
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [[`1bd5104`](https://github.com/mastra-ai/mastra/commit/1bd51048b6da93507276d6623e3fd96a9e1a8944)]:
|
|
8
|
+
- @mastra/core@1.26.0-alpha.8
|
|
9
|
+
|
|
3
10
|
## 1.1.26-alpha.14
|
|
4
11
|
|
|
5
12
|
### Patch Changes
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.26-alpha.
|
|
3
|
+
"version": "1.1.26-alpha.16",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,7 +29,7 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/core": "1.26.0-alpha.
|
|
32
|
+
"@mastra/core": "1.26.0-alpha.8",
|
|
33
33
|
"@mastra/mcp": "^1.5.1-alpha.1"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
@@ -48,7 +48,7 @@
|
|
|
48
48
|
"vitest": "4.0.18",
|
|
49
49
|
"@internal/lint": "0.0.83",
|
|
50
50
|
"@internal/types-builder": "0.0.58",
|
|
51
|
-
"@mastra/core": "1.26.0-alpha.
|
|
51
|
+
"@mastra/core": "1.26.0-alpha.8"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|