@mastra/mcp-docs-server 1.1.26-alpha.15 → 1.1.26-alpha.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -109,10 +109,10 @@ If you prefer to work from the command line, or need to script the export, you c
109
109
 
110
110
  3. Export the database to a SQL dump.
111
111
 
112
- Set the credentials provided by support (or use the dashboard values if you already copied them earlier) as environment variables, then dump the database to a local file:
112
+ Set the credentials provided by support (or use the dashboard values if you already copied them earlier) as environment variables, then dump the database to a local file. If you copied the URL from the dashboard, swap the `libsql://` scheme for `https://` — the Turso CLI expects the HTTPS form when passing the URL with an auth token.
113
113
 
114
114
  ```bash
115
- export MASTRA_STORAGE_URL="libsql://<db-name>-<org>.turso.io"
115
+ export MASTRA_STORAGE_URL="https://<db-name>-<org>.turso.io"
116
116
  export MASTRA_STORAGE_AUTH_TOKEN="<token-from-dashboard-or-support>"
117
117
 
118
118
  turso db shell "$MASTRA_STORAGE_URL?authToken=$MASTRA_STORAGE_AUTH_TOKEN" ".dump" > mastra-cloud-dump.sql
@@ -1,6 +1,6 @@
1
1
  # Model Providers
2
2
 
3
- Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3636 models from 103 providers through a single API.
3
+ Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3667 models from 104 providers through a single API.
4
4
 
5
5
  ## Features
6
6
 
@@ -1,6 +1,6 @@
1
1
  # ![302.AI logo](https://models.dev/logos/302ai.svg)302.AI
2
2
 
3
- Access 64 302.AI models through Mastra's model router. Authentication is handled automatically using the `302AI_API_KEY` environment variable.
3
+ Access 95 302.AI models through Mastra's model router. Authentication is handled automatically using the `302AI_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [302.AI documentation](https://doc.302.ai).
6
6
 
@@ -35,13 +35,25 @@ for await (const chunk of stream) {
35
35
  | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
36
  | --------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
37
  | `302ai/chatgpt-4o-latest` | 128K | | | | | | $5 | $15 |
38
+ | `302ai/claude-3-5-haiku-20241022` | 200K | | | | | | $0.80 | $4 |
39
+ | `302ai/claude-3-5-haiku-latest` | 200K | | | | | | $0.80 | $4 |
40
+ | `302ai/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
38
41
  | `302ai/claude-haiku-4-5-20251001` | 200K | | | | | | $1 | $5 |
39
42
  | `302ai/claude-opus-4-1-20250805` | 200K | | | | | | $15 | $75 |
40
43
  | `302ai/claude-opus-4-1-20250805-thinking` | 200K | | | | | | $15 | $75 |
44
+ | `302ai/claude-opus-4-20250514` | 200K | | | | | | $15 | $75 |
45
+ | `302ai/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
41
46
  | `302ai/claude-opus-4-5-20251101` | 200K | | | | | | $5 | $25 |
42
47
  | `302ai/claude-opus-4-5-20251101-thinking` | 200K | | | | | | $5 | $25 |
48
+ | `302ai/claude-opus-4-6` | 1.0M | | | | | | $5 | $25 |
49
+ | `302ai/claude-opus-4-6-thinking` | 1.0M | | | | | | $5 | $25 |
50
+ | `302ai/claude-opus-4-7` | 200K | | | | | | $5 | $25 |
51
+ | `302ai/claude-sonnet-4-20250514` | 200K | | | | | | $3 | $15 |
52
+ | `302ai/claude-sonnet-4-5` | 200K | | | | | | $3 | $15 |
43
53
  | `302ai/claude-sonnet-4-5-20250929` | 200K | | | | | | $3 | $15 |
44
54
  | `302ai/claude-sonnet-4-5-20250929-thinking` | 200K | | | | | | $3 | $15 |
55
+ | `302ai/claude-sonnet-4-6` | 1.0M | | | | | | $3 | $15 |
56
+ | `302ai/claude-sonnet-4-6-thinking` | 1.0M | | | | | | $3 | $15 |
45
57
  | `302ai/deepseek-chat` | 128K | | | | | | $0.29 | $0.43 |
46
58
  | `302ai/deepseek-reasoner` | 128K | | | | | | $0.29 | $0.43 |
47
59
  | `302ai/deepseek-v3.2` | 128K | | | | | | $0.29 | $0.43 |
@@ -60,11 +72,21 @@ for await (const chunk of stream) {
60
72
  | `302ai/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
61
73
  | `302ai/gemini-3-pro-image-preview` | 33K | | | | | | $2 | $120 |
62
74
  | `302ai/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
75
+ | `302ai/gemini-3.1-flash-image-preview` | 131K | | | | | | $0.50 | $60 |
63
76
  | `302ai/glm-4.5` | 128K | | | | | | $0.29 | $1 |
77
+ | `302ai/glm-4.5-air` | 128K | | | | | | $0.11 | $0.29 |
78
+ | `302ai/glm-4.5-airx` | 128K | | | | | | $0.57 | $2 |
79
+ | `302ai/glm-4.5-x` | 128K | | | | | | $1 | $2 |
64
80
  | `302ai/glm-4.5v` | 64K | | | | | | $0.29 | $0.86 |
65
81
  | `302ai/glm-4.6` | 200K | | | | | | $0.29 | $1 |
66
82
  | `302ai/glm-4.6v` | 128K | | | | | | $0.14 | $0.43 |
67
83
  | `302ai/glm-4.7` | 200K | | | | | | $0.29 | $1 |
84
+ | `302ai/glm-4.7-flashx` | 200K | | | | | | $0.07 | $0.43 |
85
+ | `302ai/glm-5` | 200K | | | | | | $0.60 | $3 |
86
+ | `302ai/glm-5-turbo` | 200K | | | | | | $0.72 | $3 |
87
+ | `302ai/glm-5.1` | 200K | | | | | | $0.86 | $4 |
88
+ | `302ai/glm-5v-turbo` | 200K | | | | | | $0.72 | $3 |
89
+ | `302ai/glm-for-coding` | 200K | | | | | | $0.09 | $0.34 |
68
90
  | `302ai/gpt-4.1` | 1.0M | | | | | | $2 | $8 |
69
91
  | `302ai/gpt-4.1-mini` | 1.0M | | | | | | $0.40 | $2 |
70
92
  | `302ai/gpt-4.1-nano` | 1.0M | | | | | | $0.10 | $0.40 |
@@ -77,17 +99,26 @@ for await (const chunk of stream) {
77
99
  | `302ai/gpt-5.1-chat-latest` | 128K | | | | | | $1 | $10 |
78
100
  | `302ai/gpt-5.2` | 400K | | | | | | $2 | $14 |
79
101
  | `302ai/gpt-5.2-chat-latest` | 128K | | | | | | $2 | $14 |
102
+ | `302ai/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
103
+ | `302ai/gpt-5.4-mini-2026-03-17` | 400K | | | | | | $0.75 | $5 |
104
+ | `302ai/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
105
+ | `302ai/gpt-5.4-nano-2026-03-17` | 400K | | | | | | $0.20 | $1 |
80
106
  | `302ai/grok-4-1-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
81
107
  | `302ai/grok-4-1-fast-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
82
108
  | `302ai/grok-4-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
83
109
  | `302ai/grok-4-fast-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
84
110
  | `302ai/grok-4.1` | 200K | | | | | | $2 | $10 |
111
+ | `302ai/grok-4.20-beta-0309-non-reasoning` | 2.0M | | | | | | $2 | $6 |
112
+ | `302ai/grok-4.20-beta-0309-reasoning` | 2.0M | | | | | | $2 | $6 |
113
+ | `302ai/grok-4.20-multi-agent-beta-0309` | 2.0M | | | | | | $2 | $6 |
85
114
  | `302ai/kimi-k2-0905-preview` | 262K | | | | | | $0.63 | $3 |
86
115
  | `302ai/kimi-k2-thinking` | 262K | | | | | | $0.57 | $2 |
87
116
  | `302ai/kimi-k2-thinking-turbo` | 262K | | | | | | $1 | $9 |
88
117
  | `302ai/MiniMax-M1` | 1.0M | | | | | | $0.13 | $1 |
89
118
  | `302ai/MiniMax-M2` | 1.0M | | | | | | $0.33 | $1 |
90
119
  | `302ai/MiniMax-M2.1` | 1.0M | | | | | | $0.30 | $1 |
120
+ | `302ai/MiniMax-M2.7` | 205K | | | | | | $0.30 | $1 |
121
+ | `302ai/MiniMax-M2.7-highspeed` | 205K | | | | | | $0.60 | $5 |
91
122
  | `302ai/ministral-14b-2512` | 128K | | | | | | $0.33 | $0.33 |
92
123
  | `302ai/mistral-large-2512` | 128K | | | | | | $1 | $3 |
93
124
  | `302ai/qwen-flash` | 1.0M | | | | | | $0.02 | $0.22 |
@@ -1,6 +1,6 @@
1
1
  # ![Berget.AI logo](https://models.dev/logos/berget.svg)Berget.AI
2
2
 
3
- Access 8 Berget.AI models through Mastra's model router. Authentication is handled automatically using the `BERGET_API_KEY` environment variable.
3
+ Access 5 Berget.AI models through Mastra's model router. Authentication is handled automatically using the `BERGET_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Berget.AI documentation](https://api.berget.ai).
6
6
 
@@ -15,7 +15,7 @@ const agent = new Agent({
15
15
  id: "my-agent",
16
16
  name: "My Agent",
17
17
  instructions: "You are a helpful assistant",
18
- model: "berget/BAAI/bge-reranker-v2-m3"
18
+ model: "berget/google/gemma-4-31B-it"
19
19
  });
20
20
 
21
21
  // Generate a response
@@ -34,14 +34,11 @@ for await (const chunk of stream) {
34
34
 
35
35
  | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
36
  | ------------------------------------------------------ | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
- | `berget/BAAI/bge-reranker-v2-m3` | 512 | | | | | | $0.10 | $0.10 |
38
- | `berget/intfloat/multilingual-e5-large` | 512 | | | | | | $0.02 | |
39
- | `berget/intfloat/multilingual-e5-large-instruct` | 512 | | | | | | $0.02 | |
40
- | `berget/KBLab/kb-whisper-large` | 480K | | | | | | $3 | $3 |
41
- | `berget/meta-llama/Llama-3.3-70B-Instruct` | 128K | | | | | | $0.90 | $0.90 |
42
- | `berget/mistralai/Mistral-Small-3.2-24B-Instruct-2506` | 32K | | | | | | $0.30 | $0.30 |
43
- | `berget/openai/gpt-oss-120b` | 128K | | | | | | $0.30 | $0.90 |
44
- | `berget/zai-org/GLM-4.7` | 128K | | | | | | $0.70 | $2 |
37
+ | `berget/google/gemma-4-31B-it` | 128K | | | | | | $0.28 | $0.55 |
38
+ | `berget/meta-llama/Llama-3.3-70B-Instruct` | 128K | | | | | | $0.99 | $0.99 |
39
+ | `berget/mistralai/Mistral-Small-3.2-24B-Instruct-2506` | 32K | | | | | | $0.33 | $0.33 |
40
+ | `berget/openai/gpt-oss-120b` | 128K | | | | | | $0.44 | $0.99 |
41
+ | `berget/zai-org/GLM-4.7` | 128K | | | | | | $0.77 | $3 |
45
42
 
46
43
  ## Advanced configuration
47
44
 
@@ -53,7 +50,7 @@ const agent = new Agent({
53
50
  name: "custom-agent",
54
51
  model: {
55
52
  url: "https://api.berget.ai/v1",
56
- id: "berget/BAAI/bge-reranker-v2-m3",
53
+ id: "berget/google/gemma-4-31B-it",
57
54
  apiKey: process.env.BERGET_API_KEY,
58
55
  headers: {
59
56
  "X-Custom-Header": "value"
@@ -72,7 +69,7 @@ const agent = new Agent({
72
69
  const useAdvanced = requestContext.task === "complex";
73
70
  return useAdvanced
74
71
  ? "berget/zai-org/GLM-4.7"
75
- : "berget/BAAI/bge-reranker-v2-m3";
72
+ : "berget/google/gemma-4-31B-it";
76
73
  }
77
74
  });
78
75
  ```
@@ -1,6 +1,6 @@
1
1
  # ![Cortecs logo](https://models.dev/logos/cortecs.svg)Cortecs
2
2
 
3
- Access 32 Cortecs models through Mastra's model router. Authentication is handled automatically using the `CORTECS_API_KEY` environment variable.
3
+ Access 33 Cortecs models through Mastra's model router. Authentication is handled automatically using the `CORTECS_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Cortecs documentation](https://cortecs.ai).
6
6
 
@@ -39,6 +39,7 @@ for await (const chunk of stream) {
39
39
  | `cortecs/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
40
40
  | `cortecs/claude-opus4-5` | 200K | | | | | | $6 | $30 |
41
41
  | `cortecs/claude-opus4-6` | 1.0M | | | | | | $6 | $30 |
42
+ | `cortecs/claude-opus4-7` | 1.0M | | | | | | $6 | $28 |
42
43
  | `cortecs/claude-sonnet-4` | 200K | | | | | | $3 | $17 |
43
44
  | `cortecs/deepseek-v3-0324` | 128K | | | | | | $0.55 | $2 |
44
45
  | `cortecs/devstral-2512` | 262K | | | | | | — | — |
@@ -0,0 +1,72 @@
1
+ # ![Wafer logo](https://models.dev/logos/wafer.ai.svg)Wafer
2
+
3
+ Access 2 Wafer models through Mastra's model router. Authentication is handled automatically using the `WAFER_API_KEY` environment variable.
4
+
5
+ Learn more in the [Wafer documentation](https://docs.wafer.ai/wafer-pass).
6
+
7
+ ```bash
8
+ WAFER_API_KEY=your-api-key
9
+ ```
10
+
11
+ ```typescript
12
+ import { Agent } from "@mastra/core/agent";
13
+
14
+ const agent = new Agent({
15
+ id: "my-agent",
16
+ name: "My Agent",
17
+ instructions: "You are a helpful assistant",
18
+ model: "wafer.ai/GLM-5.1"
19
+ });
20
+
21
+ // Generate a response
22
+ const response = await agent.generate("Hello!");
23
+
24
+ // Stream a response
25
+ const stream = await agent.stream("Tell me a story");
26
+ for await (const chunk of stream) {
27
+ console.log(chunk);
28
+ }
29
+ ```
30
+
31
+ > **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [Wafer documentation](https://docs.wafer.ai/wafer-pass) for details.
32
+
33
+ ## Models
34
+
35
+ | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
+ | ---------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
+ | `wafer.ai/GLM-5.1` | 203K | | | | | | — | — |
38
+ | `wafer.ai/Qwen3.5-397B-A17B` | 262K | | | | | | — | — |
39
+
40
+ ## Advanced configuration
41
+
42
+ ### Custom headers
43
+
44
+ ```typescript
45
+ const agent = new Agent({
46
+ id: "custom-agent",
47
+ name: "custom-agent",
48
+ model: {
49
+ url: "https://pass.wafer.ai/v1",
50
+ id: "wafer.ai/GLM-5.1",
51
+ apiKey: process.env.WAFER_API_KEY,
52
+ headers: {
53
+ "X-Custom-Header": "value"
54
+ }
55
+ }
56
+ });
57
+ ```
58
+
59
+ ### Dynamic model selection
60
+
61
+ ```typescript
62
+ const agent = new Agent({
63
+ id: "dynamic-agent",
64
+ name: "Dynamic Agent",
65
+ model: ({ requestContext }) => {
66
+ const useAdvanced = requestContext.task === "complex";
67
+ return useAdvanced
68
+ ? "wafer.ai/Qwen3.5-397B-A17B"
69
+ : "wafer.ai/GLM-5.1";
70
+ }
71
+ });
72
+ ```
@@ -91,6 +91,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
91
91
  - [Upstage](https://mastra.ai/models/providers/upstage)
92
92
  - [Vivgrid](https://mastra.ai/models/providers/vivgrid)
93
93
  - [Vultr](https://mastra.ai/models/providers/vultr)
94
+ - [Wafer](https://mastra.ai/models/providers/wafer.ai)
94
95
  - [Weights & Biases](https://mastra.ai/models/providers/wandb)
95
96
  - [Xiaomi](https://mastra.ai/models/providers/xiaomi)
96
97
  - [Xiaomi Token Plan (China)](https://mastra.ai/models/providers/xiaomi-token-plan-cn)
package/CHANGELOG.md CHANGED
@@ -1,5 +1,19 @@
1
1
  # @mastra/mcp-docs-server
2
2
 
3
+ ## 1.1.26-alpha.17
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [[`16e34ca`](https://github.com/mastra-ai/mastra/commit/16e34caa98b9a114b17a6125e4e3fd87f169d0d0)]:
8
+ - @mastra/core@1.26.0-alpha.9
9
+
10
+ ## 1.1.26-alpha.16
11
+
12
+ ### Patch Changes
13
+
14
+ - Updated dependencies [[`1bd5104`](https://github.com/mastra-ai/mastra/commit/1bd51048b6da93507276d6623e3fd96a9e1a8944)]:
15
+ - @mastra/core@1.26.0-alpha.8
16
+
3
17
  ## 1.1.26-alpha.14
4
18
 
5
19
  ### Patch Changes
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mastra/mcp-docs-server",
3
- "version": "1.1.26-alpha.15",
3
+ "version": "1.1.26-alpha.18",
4
4
  "description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
@@ -29,7 +29,7 @@
29
29
  "jsdom": "^26.1.0",
30
30
  "local-pkg": "^1.1.2",
31
31
  "zod": "^4.3.6",
32
- "@mastra/core": "1.26.0-alpha.7",
32
+ "@mastra/core": "1.26.0-alpha.9",
33
33
  "@mastra/mcp": "^1.5.1-alpha.1"
34
34
  },
35
35
  "devDependencies": {
@@ -48,7 +48,7 @@
48
48
  "vitest": "4.0.18",
49
49
  "@internal/lint": "0.0.83",
50
50
  "@internal/types-builder": "0.0.58",
51
- "@mastra/core": "1.26.0-alpha.7"
51
+ "@mastra/core": "1.26.0-alpha.9"
52
52
  },
53
53
  "homepage": "https://mastra.ai",
54
54
  "repository": {