@mastra/mcp-docs-server 1.1.14-alpha.3 → 1.1.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/memory/working-memory.md +1 -1
- package/.docs/models/gateways/vercel.md +5 -2
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/alibaba-cn.md +4 -4
- package/.docs/models/providers/dinference.md +73 -0
- package/.docs/models/providers/openai.md +3 -1
- package/.docs/models/providers/scaleway.md +4 -2
- package/.docs/models/providers/tencent-coding-plan.md +78 -0
- package/.docs/models/providers.md +2 -0
- package/CHANGELOG.md +15 -0
- package/package.json +6 -6
|
@@ -268,7 +268,7 @@ Schema-based working memory uses **merge semantics**, meaning the agent only nee
|
|
|
268
268
|
## Choosing between template and schema
|
|
269
269
|
|
|
270
270
|
- Use a **template** (Markdown) if you want the agent to maintain memory as a free-form text block, such as a user profile or scratchpad. Templates use **replace semantics** — the agent must provide the complete memory content on each update.
|
|
271
|
-
- Use a **schema** if you need structured, type-safe data that can be validated and programmatically accessed as JSON. Schemas use **merge semantics** — the agent only provides fields to update, and existing fields are preserved.
|
|
271
|
+
- Use a **schema** if you need structured, type-safe data that can be validated and programmatically accessed as JSON. The `workingMemory.schema` field accepts any `PublicSchema`-compatible schema (including Zod v3, Zod v4, JSON Schema, or already-standard schemas). Schemas use **merge semantics** — the agent only provides fields to update, and existing fields are preserved.
|
|
272
272
|
- Only one mode can be active at a time: setting both `template` and `schema` isn't supported.
|
|
273
273
|
|
|
274
274
|
## Example: Multi-step retention
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vercel
|
|
2
2
|
|
|
3
|
-
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 217 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
|
|
6
6
|
|
|
@@ -189,6 +189,8 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
189
189
|
| `openai/gpt-5.3-chat` |
|
|
190
190
|
| `openai/gpt-5.3-codex` |
|
|
191
191
|
| `openai/gpt-5.4` |
|
|
192
|
+
| `openai/gpt-5.4-mini` |
|
|
193
|
+
| `openai/gpt-5.4-nano` |
|
|
192
194
|
| `openai/gpt-5.4-pro` |
|
|
193
195
|
| `openai/gpt-oss-120b` |
|
|
194
196
|
| `openai/gpt-oss-20b` |
|
|
@@ -247,4 +249,5 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
247
249
|
| `zai/glm-4.7` |
|
|
248
250
|
| `zai/glm-4.7-flash` |
|
|
249
251
|
| `zai/glm-4.7-flashx` |
|
|
250
|
-
| `zai/glm-5` |
|
|
252
|
+
| `zai/glm-5` |
|
|
253
|
+
| `zai/glm-5-turbo` |
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3371 models from 94 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -15,7 +15,7 @@ const agent = new Agent({
|
|
|
15
15
|
id: "my-agent",
|
|
16
16
|
name: "My Agent",
|
|
17
17
|
instructions: "You are a helpful assistant",
|
|
18
|
-
model: "alibaba-cn/MiniMax
|
|
18
|
+
model: "alibaba-cn/MiniMax-M2.5"
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Generate a response
|
|
@@ -49,7 +49,7 @@ for await (const chunk of stream) {
|
|
|
49
49
|
| `alibaba-cn/kimi-k2-thinking` | 262K | | | | | | $0.57 | $2 |
|
|
50
50
|
| `alibaba-cn/kimi-k2.5` | 262K | | | | | | $0.57 | $2 |
|
|
51
51
|
| `alibaba-cn/kimi/kimi-k2.5` | 262K | | | | | | $0.60 | $3 |
|
|
52
|
-
| `alibaba-cn/
|
|
52
|
+
| `alibaba-cn/MiniMax-M2.5` | 205K | | | | | | $0.30 | $1 |
|
|
53
53
|
| `alibaba-cn/MiniMax/MiniMax-M2.5` | 205K | | | | | | $0.30 | $1 |
|
|
54
54
|
| `alibaba-cn/moonshot-kimi-k2-instruct` | 131K | | | | | | $0.57 | $2 |
|
|
55
55
|
| `alibaba-cn/qvq-max` | 131K | | | | | | $1 | $5 |
|
|
@@ -119,7 +119,7 @@ const agent = new Agent({
|
|
|
119
119
|
name: "custom-agent",
|
|
120
120
|
model: {
|
|
121
121
|
url: "https://dashscope.aliyuncs.com/compatible-mode/v1",
|
|
122
|
-
id: "alibaba-cn/MiniMax
|
|
122
|
+
id: "alibaba-cn/MiniMax-M2.5",
|
|
123
123
|
apiKey: process.env.DASHSCOPE_API_KEY,
|
|
124
124
|
headers: {
|
|
125
125
|
"X-Custom-Header": "value"
|
|
@@ -138,7 +138,7 @@ const agent = new Agent({
|
|
|
138
138
|
const useAdvanced = requestContext.task === "complex";
|
|
139
139
|
return useAdvanced
|
|
140
140
|
? "alibaba-cn/tongyi-intent-detect-v3"
|
|
141
|
-
: "alibaba-cn/MiniMax
|
|
141
|
+
: "alibaba-cn/MiniMax-M2.5";
|
|
142
142
|
}
|
|
143
143
|
});
|
|
144
144
|
```
|
|
@@ -0,0 +1,73 @@
|
|
|
1
|
+
# DInference
|
|
2
|
+
|
|
3
|
+
Access 3 DInference models through Mastra's model router. Authentication is handled automatically using the `DINFERENCE_API_KEY` environment variable.
|
|
4
|
+
|
|
5
|
+
Learn more in the [DInference documentation](https://dinference.com).
|
|
6
|
+
|
|
7
|
+
```bash
|
|
8
|
+
DINFERENCE_API_KEY=your-api-key
|
|
9
|
+
```
|
|
10
|
+
|
|
11
|
+
```typescript
|
|
12
|
+
import { Agent } from "@mastra/core/agent";
|
|
13
|
+
|
|
14
|
+
const agent = new Agent({
|
|
15
|
+
id: "my-agent",
|
|
16
|
+
name: "My Agent",
|
|
17
|
+
instructions: "You are a helpful assistant",
|
|
18
|
+
model: "dinference/glm-4.7"
|
|
19
|
+
});
|
|
20
|
+
|
|
21
|
+
// Generate a response
|
|
22
|
+
const response = await agent.generate("Hello!");
|
|
23
|
+
|
|
24
|
+
// Stream a response
|
|
25
|
+
const stream = await agent.stream("Tell me a story");
|
|
26
|
+
for await (const chunk of stream) {
|
|
27
|
+
console.log(chunk);
|
|
28
|
+
}
|
|
29
|
+
```
|
|
30
|
+
|
|
31
|
+
> **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [DInference documentation](https://dinference.com) for details.
|
|
32
|
+
|
|
33
|
+
## Models
|
|
34
|
+
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `dinference/glm-4.7` | 200K | | | | | | $0.45 | $2 |
|
|
38
|
+
| `dinference/glm-5` | 200K | | | | | | $0.75 | $2 |
|
|
39
|
+
| `dinference/gpt-oss-120b` | 131K | | | | | | $0.07 | $0.27 |
|
|
40
|
+
|
|
41
|
+
## Advanced configuration
|
|
42
|
+
|
|
43
|
+
### Custom headers
|
|
44
|
+
|
|
45
|
+
```typescript
|
|
46
|
+
const agent = new Agent({
|
|
47
|
+
id: "custom-agent",
|
|
48
|
+
name: "custom-agent",
|
|
49
|
+
model: {
|
|
50
|
+
url: "https://api.dinference.com/v1",
|
|
51
|
+
id: "dinference/glm-4.7",
|
|
52
|
+
apiKey: process.env.DINFERENCE_API_KEY,
|
|
53
|
+
headers: {
|
|
54
|
+
"X-Custom-Header": "value"
|
|
55
|
+
}
|
|
56
|
+
}
|
|
57
|
+
});
|
|
58
|
+
```
|
|
59
|
+
|
|
60
|
+
### Dynamic model selection
|
|
61
|
+
|
|
62
|
+
```typescript
|
|
63
|
+
const agent = new Agent({
|
|
64
|
+
id: "dynamic-agent",
|
|
65
|
+
name: "Dynamic Agent",
|
|
66
|
+
model: ({ requestContext }) => {
|
|
67
|
+
const useAdvanced = requestContext.task === "complex";
|
|
68
|
+
return useAdvanced
|
|
69
|
+
? "dinference/gpt-oss-120b"
|
|
70
|
+
: "dinference/glm-4.7";
|
|
71
|
+
}
|
|
72
|
+
});
|
|
73
|
+
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenAI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 46 OpenAI models through Mastra's model router. Authentication is handled automatically using the `OPENAI_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenAI documentation](https://platform.openai.com/docs/models).
|
|
6
6
|
|
|
@@ -62,6 +62,8 @@ for await (const chunk of stream) {
|
|
|
62
62
|
| `openai/gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
|
|
63
63
|
| `openai/gpt-5.3-codex-spark` | 128K | | | | | | $2 | $14 |
|
|
64
64
|
| `openai/gpt-5.4` | 1.1M | | | | | | $3 | $15 |
|
|
65
|
+
| `openai/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
66
|
+
| `openai/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
65
67
|
| `openai/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
|
|
66
68
|
| `openai/o1` | 200K | | | | | | $15 | $60 |
|
|
67
69
|
| `openai/o1-mini` | 128K | | | | | | $1 | $4 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Scaleway
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 16 Scaleway models through Mastra's model router. Authentication is handled automatically using the `SCALEWAY_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Scaleway documentation](https://www.scaleway.com/en/docs/generative-apis/).
|
|
6
6
|
|
|
@@ -34,7 +34,7 @@ for await (const chunk of stream) {
|
|
|
34
34
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| ---------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
-
| `scaleway/bge-multilingual-gemma2` | 8K | | | | | | $0.
|
|
37
|
+
| `scaleway/bge-multilingual-gemma2` | 8K | | | | | | $0.10 | — |
|
|
38
38
|
| `scaleway/deepseek-r1-distill-llama-70b` | 32K | | | | | | $0.90 | $0.90 |
|
|
39
39
|
| `scaleway/devstral-2-123b-instruct-2512` | 256K | | | | | | $0.40 | $2 |
|
|
40
40
|
| `scaleway/gemma-3-27b-it` | 40K | | | | | | $0.25 | $0.50 |
|
|
@@ -46,6 +46,8 @@ for await (const chunk of stream) {
|
|
|
46
46
|
| `scaleway/pixtral-12b-2409` | 128K | | | | | | $0.20 | $0.20 |
|
|
47
47
|
| `scaleway/qwen3-235b-a22b-instruct-2507` | 260K | | | | | | $0.75 | $2 |
|
|
48
48
|
| `scaleway/qwen3-coder-30b-a3b-instruct` | 128K | | | | | | $0.20 | $0.80 |
|
|
49
|
+
| `scaleway/qwen3-embedding-8b` | 33K | | | | | | $0.10 | — |
|
|
50
|
+
| `scaleway/qwen3.5-397b-a17b` | 256K | | | | | | $0.60 | $4 |
|
|
49
51
|
| `scaleway/voxtral-small-24b-2507` | 32K | | | | | | $0.15 | $0.35 |
|
|
50
52
|
| `scaleway/whisper-large-v3` | — | | | | | | $0.00 | — |
|
|
51
53
|
|
|
@@ -0,0 +1,78 @@
|
|
|
1
|
+
# Tencent Coding Plan (China)
|
|
2
|
+
|
|
3
|
+
Access 8 Tencent Coding Plan (China) models through Mastra's model router. Authentication is handled automatically using the `TENCENT_CODING_PLAN_API_KEY` environment variable.
|
|
4
|
+
|
|
5
|
+
Learn more in the [Tencent Coding Plan (China) documentation](https://cloud.tencent.com/document/product/1772/128947).
|
|
6
|
+
|
|
7
|
+
```bash
|
|
8
|
+
TENCENT_CODING_PLAN_API_KEY=your-api-key
|
|
9
|
+
```
|
|
10
|
+
|
|
11
|
+
```typescript
|
|
12
|
+
import { Agent } from "@mastra/core/agent";
|
|
13
|
+
|
|
14
|
+
const agent = new Agent({
|
|
15
|
+
id: "my-agent",
|
|
16
|
+
name: "My Agent",
|
|
17
|
+
instructions: "You are a helpful assistant",
|
|
18
|
+
model: "tencent-coding-plan/glm-5"
|
|
19
|
+
});
|
|
20
|
+
|
|
21
|
+
// Generate a response
|
|
22
|
+
const response = await agent.generate("Hello!");
|
|
23
|
+
|
|
24
|
+
// Stream a response
|
|
25
|
+
const stream = await agent.stream("Tell me a story");
|
|
26
|
+
for await (const chunk of stream) {
|
|
27
|
+
console.log(chunk);
|
|
28
|
+
}
|
|
29
|
+
```
|
|
30
|
+
|
|
31
|
+
> **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [Tencent Coding Plan (China) documentation](https://cloud.tencent.com/document/product/1772/128947) for details.
|
|
32
|
+
|
|
33
|
+
## Models
|
|
34
|
+
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ------------------------------------------ | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `tencent-coding-plan/glm-5` | 203K | | | | | | — | — |
|
|
38
|
+
| `tencent-coding-plan/hunyuan-2.0-instruct` | 131K | | | | | | — | — |
|
|
39
|
+
| `tencent-coding-plan/hunyuan-2.0-thinking` | 131K | | | | | | — | — |
|
|
40
|
+
| `tencent-coding-plan/hunyuan-t1` | 131K | | | | | | — | — |
|
|
41
|
+
| `tencent-coding-plan/hunyuan-turbos` | 131K | | | | | | — | — |
|
|
42
|
+
| `tencent-coding-plan/kimi-k2.5` | 262K | | | | | | — | — |
|
|
43
|
+
| `tencent-coding-plan/minimax-m2.5` | 205K | | | | | | — | — |
|
|
44
|
+
| `tencent-coding-plan/tc-code-latest` | 131K | | | | | | — | — |
|
|
45
|
+
|
|
46
|
+
## Advanced configuration
|
|
47
|
+
|
|
48
|
+
### Custom headers
|
|
49
|
+
|
|
50
|
+
```typescript
|
|
51
|
+
const agent = new Agent({
|
|
52
|
+
id: "custom-agent",
|
|
53
|
+
name: "custom-agent",
|
|
54
|
+
model: {
|
|
55
|
+
url: "https://api.lkeap.cloud.tencent.com/coding/v3",
|
|
56
|
+
id: "tencent-coding-plan/glm-5",
|
|
57
|
+
apiKey: process.env.TENCENT_CODING_PLAN_API_KEY,
|
|
58
|
+
headers: {
|
|
59
|
+
"X-Custom-Header": "value"
|
|
60
|
+
}
|
|
61
|
+
}
|
|
62
|
+
});
|
|
63
|
+
```
|
|
64
|
+
|
|
65
|
+
### Dynamic model selection
|
|
66
|
+
|
|
67
|
+
```typescript
|
|
68
|
+
const agent = new Agent({
|
|
69
|
+
id: "dynamic-agent",
|
|
70
|
+
name: "Dynamic Agent",
|
|
71
|
+
model: ({ requestContext }) => {
|
|
72
|
+
const useAdvanced = requestContext.task === "complex";
|
|
73
|
+
return useAdvanced
|
|
74
|
+
? "tencent-coding-plan/tc-code-latest"
|
|
75
|
+
: "tencent-coding-plan/glm-5";
|
|
76
|
+
}
|
|
77
|
+
});
|
|
78
|
+
```
|
|
@@ -27,6 +27,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
|
|
|
27
27
|
- [Cortecs](https://mastra.ai/models/providers/cortecs)
|
|
28
28
|
- [D.Run (China)](https://mastra.ai/models/providers/drun)
|
|
29
29
|
- [Deep Infra](https://mastra.ai/models/providers/deepinfra)
|
|
30
|
+
- [DInference](https://mastra.ai/models/providers/dinference)
|
|
30
31
|
- [evroc](https://mastra.ai/models/providers/evroc)
|
|
31
32
|
- [FastRouter](https://mastra.ai/models/providers/fastrouter)
|
|
32
33
|
- [Fireworks AI](https://mastra.ai/models/providers/fireworks-ai)
|
|
@@ -79,6 +80,7 @@ Direct access to individual AI model providers. Each provider offers unique mode
|
|
|
79
80
|
- [StepFun](https://mastra.ai/models/providers/stepfun)
|
|
80
81
|
- [submodel](https://mastra.ai/models/providers/submodel)
|
|
81
82
|
- [Synthetic](https://mastra.ai/models/providers/synthetic)
|
|
83
|
+
- [Tencent Coding Plan (China)](https://mastra.ai/models/providers/tencent-coding-plan)
|
|
82
84
|
- [Together AI](https://mastra.ai/models/providers/togetherai)
|
|
83
85
|
- [Upstage](https://mastra.ai/models/providers/upstage)
|
|
84
86
|
- [Vivgrid](https://mastra.ai/models/providers/vivgrid)
|
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,20 @@
|
|
|
1
1
|
# @mastra/mcp-docs-server
|
|
2
2
|
|
|
3
|
+
## 1.1.14
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [[`51970b3`](https://github.com/mastra-ai/mastra/commit/51970b3828494d59a8dd4df143b194d37d31e3f5), [`bbcbbce`](https://github.com/mastra-ai/mastra/commit/bbcbbce4f0e268053cbb11ca58350f5ceba15498), [`4444280`](https://github.com/mastra-ai/mastra/commit/444428094253e916ec077e66284e685fde67021e), [`085e371`](https://github.com/mastra-ai/mastra/commit/085e3718a7d0fe9a210fe7dd1c867b9bdfe8d16b), [`b77aa19`](https://github.com/mastra-ai/mastra/commit/b77aa1981361c021f2c881bee8f0c703687f00da), [`dbb879a`](https://github.com/mastra-ai/mastra/commit/dbb879af0b809c668e9b3a9d8bac97d806caa267), [`8b4ce84`](https://github.com/mastra-ai/mastra/commit/8b4ce84aed0808b9805cc4fd7147c1f8a2ef7a36), [`8d4cfe6`](https://github.com/mastra-ai/mastra/commit/8d4cfe6b9a7157d3876206227ec9f04cde6dbc4a), [`dd6ca1c`](https://github.com/mastra-ai/mastra/commit/dd6ca1cdea3b8b6182f4cf61df41070ba0cc0deb), [`ce26fe2`](https://github.com/mastra-ai/mastra/commit/ce26fe2166dd90254f8bee5776e55977143e97de), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a), [`4cb4edf`](https://github.com/mastra-ai/mastra/commit/4cb4edf3c909d197ec356c1790d13270514ffef6), [`8de3555`](https://github.com/mastra-ai/mastra/commit/8de355572c6fd838f863a3e7e6fe24d0947b774f), [`b26307f`](https://github.com/mastra-ai/mastra/commit/b26307f050df39629511b0e831b8fc26973ce8b1), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a), [`133ef20`](https://github.com/mastra-ai/mastra/commit/133ef20c39c696eb0dbbee26e77c8acfec14b8c6)]:
|
|
8
|
+
- @mastra/core@1.14.0
|
|
9
|
+
- @mastra/mcp@1.3.0
|
|
10
|
+
|
|
11
|
+
## 1.1.14-alpha.4
|
|
12
|
+
|
|
13
|
+
### Patch Changes
|
|
14
|
+
|
|
15
|
+
- Updated dependencies [[`8b4ce84`](https://github.com/mastra-ai/mastra/commit/8b4ce84aed0808b9805cc4fd7147c1f8a2ef7a36), [`8d4cfe6`](https://github.com/mastra-ai/mastra/commit/8d4cfe6b9a7157d3876206227ec9f04cde6dbc4a), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a), [`68a019d`](https://github.com/mastra-ai/mastra/commit/68a019d30d22251ddd628a2947d60215c03c350a)]:
|
|
16
|
+
- @mastra/core@1.14.0-alpha.3
|
|
17
|
+
|
|
3
18
|
## 1.1.14-alpha.3
|
|
4
19
|
|
|
5
20
|
### Patch Changes
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.14
|
|
3
|
+
"version": "1.1.14",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,8 +29,8 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/core": "1.14.0
|
|
33
|
-
"@mastra/mcp": "^1.3.0
|
|
32
|
+
"@mastra/core": "1.14.0",
|
|
33
|
+
"@mastra/mcp": "^1.3.0"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
36
36
|
"@hono/node-server": "^1.19.11",
|
|
@@ -46,9 +46,9 @@
|
|
|
46
46
|
"tsx": "^4.21.0",
|
|
47
47
|
"typescript": "^5.9.3",
|
|
48
48
|
"vitest": "4.0.18",
|
|
49
|
-
"@internal/lint": "0.0.
|
|
50
|
-
"@
|
|
51
|
-
"@
|
|
49
|
+
"@internal/lint": "0.0.72",
|
|
50
|
+
"@mastra/core": "1.14.0",
|
|
51
|
+
"@internal/types-builder": "0.0.47"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|