@mastra/mcp-docs-server 1.1.14-alpha.3 → 1.1.14

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -268,7 +268,7 @@ Schema-based working memory uses **merge semantics**, meaning the agent only nee
268
268
  ## Choosing between template and schema
269
269
 
270
270
  - Use a **template** (Markdown) if you want the agent to maintain memory as a free-form text block, such as a user profile or scratchpad. Templates use **replace semantics** — the agent must provide the complete memory content on each update.
271
- - Use a **schema** if you need structured, type-safe data that can be validated and programmatically accessed as JSON. Schemas use **merge semantics** — the agent only provides fields to update, and existing fields are preserved.
271
+ - Use a **schema** if you need structured, type-safe data that can be validated and programmatically accessed as JSON. The `workingMemory.schema` field accepts any `PublicSchema`-compatible schema (including Zod v3, Zod v4, JSON Schema, or already-standard schemas). Schemas use **merge semantics** — the agent only provides fields to update, and existing fields are preserved.
272
272
  - Only one mode can be active at a time: setting both `template` and `schema` isn't supported.
273
273
 
274
274
  ## Example: Multi-step retention
@@ -1,6 +1,6 @@
1
1
  # ![Vercel logo](https://models.dev/logos/vercel.svg)Vercel
2
2
 
3
- Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 214 models through Mastra's model router.
3
+ Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 217 models through Mastra's model router.
4
4
 
5
5
  Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
6
6
 
@@ -189,6 +189,8 @@ ANTHROPIC_API_KEY=ant-...
189
189
  | `openai/gpt-5.3-chat` |
190
190
  | `openai/gpt-5.3-codex` |
191
191
  | `openai/gpt-5.4` |
192
+ | `openai/gpt-5.4-mini` |
193
+ | `openai/gpt-5.4-nano` |
192
194
  | `openai/gpt-5.4-pro` |
193
195
  | `openai/gpt-oss-120b` |
194
196
  | `openai/gpt-oss-20b` |
@@ -247,4 +249,5 @@ ANTHROPIC_API_KEY=ant-...
247
249
  | `zai/glm-4.7` |
248
250
  | `zai/glm-4.7-flash` |
249
251
  | `zai/glm-4.7-flashx` |
250
- | `zai/glm-5` |
252
+ | `zai/glm-5` |
253
+ | `zai/glm-5-turbo` |
@@ -1,6 +1,6 @@
1
1
  # Model Providers
2
2
 
3
- Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3353 models from 92 providers through a single API.
3
+ Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3371 models from 94 providers through a single API.
4
4
 
5
5
  ## Features
6
6
 
@@ -15,7 +15,7 @@ const agent = new Agent({
15
15
  id: "my-agent",
16
16
  name: "My Agent",
17
17
  instructions: "You are a helpful assistant",
18
- model: "alibaba-cn/MiniMax/MiniMax-M2.5"
18
+ model: "alibaba-cn/MiniMax-M2.5"
19
19
  });
20
20
 
21
21
  // Generate a response
@@ -49,7 +49,7 @@ for await (const chunk of stream) {
49
49
  | `alibaba-cn/kimi-k2-thinking` | 262K | | | | | | $0.57 | $2 |
50
50
  | `alibaba-cn/kimi-k2.5` | 262K | | | | | | $0.57 | $2 |
51
51
  | `alibaba-cn/kimi/kimi-k2.5` | 262K | | | | | | $0.60 | $3 |
52
- | `alibaba-cn/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
52
+ | `alibaba-cn/MiniMax-M2.5` | 205K | | | | | | $0.30 | $1 |
53
53
  | `alibaba-cn/MiniMax/MiniMax-M2.5` | 205K | | | | | | $0.30 | $1 |
54
54
  | `alibaba-cn/moonshot-kimi-k2-instruct` | 131K | | | | | | $0.57 | $2 |
55
55
  | `alibaba-cn/qvq-max` | 131K | | | | | | $1 | $5 |
@@ -119,7 +119,7 @@ const agent = new Agent({
119
119
  name: "custom-agent",
120
120
  model: {
121
121
  url: "https://dashscope.aliyuncs.com/compatible-mode/v1",
122
- id: "alibaba-cn/MiniMax/MiniMax-M2.5",
122
+ id: "alibaba-cn/MiniMax-M2.5",
123
123
  apiKey: process.env.DASHSCOPE_API_KEY,
124
124
  headers: {
125
125
  "X-Custom-Header": "value"
@@ -138,7 +138,7 @@ const agent = new Agent({
138
138
  const useAdvanced = requestContext.task === "complex";
139
139
  return useAdvanced
140
140
  ? "alibaba-cn/tongyi-intent-detect-v3"
141
- : "alibaba-cn/MiniMax/MiniMax-M2.5";
141
+ : "alibaba-cn/MiniMax-M2.5";
142
142
  }
143
143
  });
144
144
  ```
@@ -0,0 +1,73 @@
1
+ # ![DInference logo](https://models.dev/logos/dinference.svg)DInference
2
+
3
+ Access 3 DInference models through Mastra's model router. Authentication is handled automatically using the `DINFERENCE_API_KEY` environment variable.
4
+
5
+ Learn more in the [DInference documentation](https://dinference.com).
6
+
7
+ ```bash
8
+ DINFERENCE_API_KEY=your-api-key
9
+ ```
10
+
11
+ ```typescript
12
+ import { Agent } from "@mastra/core/agent";
13
+
14
+ const agent = new Agent({
15
+ id: "my-agent",
16
+ name: "My Agent",
17
+ instructions: "You are a helpful assistant",
18
+ model: "dinference/glm-4.7"
19
+ });
20
+
21
+ // Generate a response
22
+ const response = await agent.generate("Hello!");
23
+
24
+ // Stream a response
25
+ const stream = await agent.stream("Tell me a story");
26
+ for await (const chunk of stream) {
27
+ console.log(chunk);
28
+ }
29
+ ```
30
+
31
+ > **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [DInference documentation](https://dinference.com) for details.
32
+
33
+ ## Models
34
+
35
+ | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
+ | ------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
+ | `dinference/glm-4.7` | 200K | | | | | | $0.45 | $2 |
38
+ | `dinference/glm-5` | 200K | | | | | | $0.75 | $2 |
39
+ | `dinference/gpt-oss-120b` | 131K | | | | | | $0.07 | $0.27 |
40
+
41
+ ## Advanced configuration
42
+
43
+ ### Custom headers
44
+
45
+ ```typescript
46
+ const agent = new Agent({
47
+ id: "custom-agent",
48
+ name: "custom-agent",
49
+ model: {
50
+ url: "https://api.dinference.com/v1",
51
+ id: "dinference/glm-4.7",
52
+ apiKey: process.env.DINFERENCE_API_KEY,
53
+ headers: {
54
+ "X-Custom-Header": "value"
55
+ }
56
+ }
57
+ });
58
+ ```
59
+
60
+ ### Dynamic model selection
61
+
62
+ ```typescript
63
+ const agent = new Agent({
64
+ id: "dynamic-agent",
65
+ name: "Dynamic Agent",
66
+ model: ({ requestContext }) => {
67
+ const useAdvanced = requestContext.task === "complex";
68
+ return useAdvanced
69
+ ? "dinference/gpt-oss-120b"
70
+ : "dinference/glm-4.7";
71
+ }
72
+ });
73
+ ```
@@ -1,6 +1,6 @@
1
1
  # ![OpenAI logo](https://models.dev/logos/openai.svg)OpenAI
2
2
 
3
- Access 44 OpenAI models through Mastra's model router. Authentication is handled automatically using the `OPENAI_API_KEY` environment variable.
3
+ Access 46 OpenAI models through Mastra's model router. Authentication is handled automatically using the `OPENAI_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [OpenAI documentation](https://platform.openai.com/docs/models).
6
6
 
@@ -62,6 +62,8 @@ for await (const chunk of stream) {
62
62
  | `openai/gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
63
63
  | `openai/gpt-5.3-codex-spark` | 128K | | | | | | $2 | $14 |
64
64
  | `openai/gpt-5.4` | 1.1M | | | | | | $3 | $15 |
65
+ | `openai/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
66
+ | `openai/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
65
67
  | `openai/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
66
68
  | `openai/o1` | 200K | | | | | | $15 | $60 |
67
69
  | `openai/o1-mini` | 128K | | | | | | $1 | $4 |
@@ -1,6 +1,6 @@
1
1
  # ![Scaleway logo](https://models.dev/logos/scaleway.svg)Scaleway
2
2
 
3
- Access 14 Scaleway models through Mastra's model router. Authentication is handled automatically using the `SCALEWAY_API_KEY` environment variable.
3
+ Access 16 Scaleway models through Mastra's model router. Authentication is handled automatically using the `SCALEWAY_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Scaleway documentation](https://www.scaleway.com/en/docs/generative-apis/).
6
6
 
@@ -34,7 +34,7 @@ for await (const chunk of stream) {
34
34
 
35
35
  | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
36
  | ---------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
- | `scaleway/bge-multilingual-gemma2` | 8K | | | | | | $0.13 | — |
37
+ | `scaleway/bge-multilingual-gemma2` | 8K | | | | | | $0.10 | — |
38
38
  | `scaleway/deepseek-r1-distill-llama-70b` | 32K | | | | | | $0.90 | $0.90 |
39
39
  | `scaleway/devstral-2-123b-instruct-2512` | 256K | | | | | | $0.40 | $2 |
40
40
  | `scaleway/gemma-3-27b-it` | 40K | | | | | | $0.25 | $0.50 |
@@ -46,6 +46,8 @@ for await (const chunk of stream) {
46
46
  | `scaleway/pixtral-12b-2409` | 128K | | | | | | $0.20 | $0.20 |
47
47
  | `scaleway/qwen3-235b-a22b-instruct-2507` | 260K | | | | | | $0.75 | $2 |
48
48
  | `scaleway/qwen3-coder-30b-a3b-instruct` | 128K | | | | | | $0.20 | $0.80 |
49
+ | `scaleway/qwen3-embedding-8b` | 33K | | | | | | $0.10 | — |
50
+ | `scaleway/qwen3.5-397b-a17b` | 256K | | | | | | $0.60 | $4 |
49
51
  | `scaleway/voxtral-small-24b-2507` | 32K | | | | | | $0.15 | $0.35 |
50
52
  | `scaleway/whisper-large-v3` | — | | | | | | $0.00 | — |
51
53
 
@@ -0,0 +1,78 @@
1
+ # ![Tencent Coding Plan (China) logo](https://models.dev/logos/tencent-coding-plan.svg)Tencent Coding Plan (China)
2
+
3
+ Access 8 Tencent Coding Plan (China) models through Mastra's model router. Authentication is handled automatically using the `TENCENT_CODING_PLAN_API_KEY` environment variable.
4
+
5
+ Learn more in the [Tencent Coding Plan (China) documentation](https://cloud.tencent.com/document/product/1772/128947).
6
+
7
+ ```bash
8
+ TENCENT_CODING_PLAN_API_KEY=your-api-key
9
+ ```
10
+
11
+ ```typescript
12
+ import { Agent } from "@mastra/core/agent";
13
+
14
+ const agent = new Agent({
15
+ id: "my-agent",
16
+ name: "My Agent",
17
+ instructions: "You are a helpful assistant",
18
+ model: "tencent-coding-plan/glm-5"
19
+ });
20
+
21
+ // Generate a response
22
+ const response = await agent.generate("Hello!");
23
+
24
+ // Stream a response
25
+ const stream = await agent.stream("Tell me a story");
26
+ for await (const chunk of stream) {
27
+ console.log(chunk);
28
+ }
29
+ ```
30
+
31
+ > **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [Tencent Coding Plan (China) documentation](https://cloud.tencent.com/document/product/1772/128947) for details.
32
+
33
+ ## Models
34
+
35
+ | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
+ | ------------------------------------------ | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
+ | `tencent-coding-plan/glm-5` | 203K | | | | | | — | — |
38
+ | `tencent-coding-plan/hunyuan-2.0-instruct` | 131K | | | | | | — | — |
39
+ | `tencent-coding-plan/hunyuan-2.0-thinking` | 131K | | | | | | — | — |
40
+ | `tencent-coding-plan/hunyuan-t1` | 131K | | | | | | — | — |
41
+ | `tencent-coding-plan/hunyuan-turbos` | 131K | | | | | | — | — |
42
+ | `tencent-coding-plan/kimi-k2.5` | 262K | | | | | | — | — |
43
+ | `tencent-coding-plan/minimax-m2.5` | 205K | | | | | | — | — |
44
+ | `tencent-coding-plan/tc-code-latest` | 131K | | | | | | — | — |
45
+
46
+ ## Advanced configuration
47
+
48
+ ### Custom headers
49
+
50
+ ```typescript
51
+ const agent = new Agent({
52
+ id: "custom-agent",
53
+ name: "custom-agent",
54
+ model: {
55
+ url: "https://api.lkeap.cloud.tencent.com/coding/v3",
56
+ id: "tencent-coding-plan/glm-5",
57
+ apiKey: process.env.TENCENT_CODING_PLAN_API_KEY,
58
+ headers: {
59
+ "X-Custom-Header": "value"
60
+ }
61
+ }
62
+ });
63
+ ```
64
+
65
+ ### Dynamic model selection
66
+
67
+ ```typescript
68
+ const agent = new Agent({
69
+ id: "dynamic-agent",
70
+ name: "Dynamic Agent",
71
+ model: ({ requestContext }) => {
72
+ const useAdvanced = requestContext.task === "complex";
73
+ return useAdvanced
74
+ ? "tencent-coding-plan/tc-code-latest"
75
+ : "tencent-coding-plan/glm-5";
76
+ }
77
+ });
78
+ ```
@@ -27,6 +27,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
27
27
  - [Cortecs](https://mastra.ai/models/providers/cortecs)
28
28
  - [D.Run (China)](https://mastra.ai/models/providers/drun)
29
29
  - [Deep Infra](https://mastra.ai/models/providers/deepinfra)
30
+ - [DInference](https://mastra.ai/models/providers/dinference)
30
31
  - [evroc](https://mastra.ai/models/providers/evroc)
31
32
  - [FastRouter](https://mastra.ai/models/providers/fastrouter)
32
33
  - [Fireworks AI](https://mastra.ai/models/providers/fireworks-ai)
@@ -79,6 +80,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
79
80
  - [StepFun](https://mastra.ai/models/providers/stepfun)
80
81
  - [submodel](https://mastra.ai/models/providers/submodel)
81
82
  - [Synthetic](https://mastra.ai/models/providers/synthetic)
83
+ - [Tencent Coding Plan (China)](https://mastra.ai/models/providers/tencent-coding-plan)
82
84
  - [Together AI](https://mastra.ai/models/providers/togetherai)
83
85
  - [Upstage](https://mastra.ai/models/providers/upstage)
84
86
  - [Vivgrid](https://mastra.ai/models/providers/vivgrid)
package/CHANGELOG.md CHANGED
@@ -1,5 +1,20 @@
1
1
  # @mastra/mcp-docs-server
2
2
 
3
+ ## 1.1.14
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [[`51970b3`](https://github.com/mastra-ai/mastra/commit/51970b3828494d59a8dd4df143b194d37d31e3f5), [`bbcbbce`](https://github.com/mastra-ai/mastra/commit/bbcbbce4f0e268053cbb11ca58350f5ceba15498), [`4444280`](https://github.com/mastra-ai/mastra/commit/444428094253e916ec077e66284e685fde67021e), [`085e371`](https://github.com/mastra-ai/mastra/commit/085e3718a7d0fe9a210fe7dd1c867b9bdfe8d16b), [`b77aa19`](https://github.com/mastra-ai/mastra/commit/b77aa1981361c021f2c881bee8f0c703687f00da), [`dbb879a`](https://github.com/mastra-ai/mastra/commit/dbb879af0b809c668e9b3a9d8bac97d806caa267), [`8b4ce84`](https://github.com/mastra-ai/mastra/commit/8b4ce84aed0808b9805cc4fd7147c1f8a2ef7a36), [`8d4cfe6`](https://github.com/mastra-ai/mastra/commit/8d4cfe6b9a7157d3876206227ec9f04cde6dbc4a), [`dd6ca1c`](https://github.com/mastra-ai/mastra/commit/dd6ca1cdea3b8b6182f4cf61df41070ba0cc0deb), [`ce26fe2`](https://github.com/mastra-ai/mastra/commit/ce26fe2166dd90254f8bee5776e55977143e97de), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a), [`4cb4edf`](https://github.com/mastra-ai/mastra/commit/4cb4edf3c909d197ec356c1790d13270514ffef6), [`8de3555`](https://github.com/mastra-ai/mastra/commit/8de355572c6fd838f863a3e7e6fe24d0947b774f), [`b26307f`](https://github.com/mastra-ai/mastra/commit/b26307f050df39629511b0e831b8fc26973ce8b1), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a), [`133ef20`](https://github.com/mastra-ai/mastra/commit/133ef20c39c696eb0dbbee26e77c8acfec14b8c6)]:
8
+ - @mastra/core@1.14.0
9
+ - @mastra/mcp@1.3.0
10
+
11
+ ## 1.1.14-alpha.4
12
+
13
+ ### Patch Changes
14
+
15
+ - Updated dependencies [[`8b4ce84`](https://github.com/mastra-ai/mastra/commit/8b4ce84aed0808b9805cc4fd7147c1f8a2ef7a36), [`8d4cfe6`](https://github.com/mastra-ai/mastra/commit/8d4cfe6b9a7157d3876206227ec9f04cde6dbc4a), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a)]:
16
+ - @mastra/core@1.14.0-alpha.3
17
+
3
18
  ## 1.1.14-alpha.3
4
19
 
5
20
  ### Patch Changes
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mastra/mcp-docs-server",
3
- "version": "1.1.14-alpha.3",
3
+ "version": "1.1.14",
4
4
  "description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
@@ -29,8 +29,8 @@
29
29
  "jsdom": "^26.1.0",
30
30
  "local-pkg": "^1.1.2",
31
31
  "zod": "^4.3.6",
32
- "@mastra/core": "1.14.0-alpha.2",
33
- "@mastra/mcp": "^1.3.0-alpha.1"
32
+ "@mastra/core": "1.14.0",
33
+ "@mastra/mcp": "^1.3.0"
34
34
  },
35
35
  "devDependencies": {
36
36
  "@hono/node-server": "^1.19.11",
@@ -46,9 +46,9 @@
46
46
  "tsx": "^4.21.0",
47
47
  "typescript": "^5.9.3",
48
48
  "vitest": "4.0.18",
49
- "@internal/lint": "0.0.71",
50
- "@internal/types-builder": "0.0.46",
51
- "@mastra/core": "1.14.0-alpha.2"
49
+ "@internal/lint": "0.0.72",
50
+ "@mastra/core": "1.14.0",
51
+ "@internal/types-builder": "0.0.47"
52
52
  },
53
53
  "homepage": "https://mastra.ai",
54
54
  "repository": {