@mastra/mcp-docs-server 1.1.29-alpha.1 → 1.1.29-alpha.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/workspace/overview.md +1 -1
- package/.docs/docs/workspace/search.md +2 -2
- package/.docs/docs/workspace/skills.md +16 -16
- package/.docs/guides/guide/code-review-bot.md +2 -2
- package/.docs/guides/guide/dev-assistant.md +4 -4
- package/.docs/models/gateways/netlify.md +5 -1
- package/.docs/models/gateways/openrouter.md +4 -1
- package/.docs/models/gateways/vercel.md +7 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/nvidia.md +2 -1
- package/.docs/models/providers/opencode-go.md +20 -18
- package/.docs/models/providers/opencode.md +3 -1
- package/.docs/reference/processors/skill-search-processor.md +1 -1
- package/.docs/reference/workspace/workspace-class.md +1 -1
- package/CHANGELOG.md +7 -0
- package/package.json +4 -4
|
@@ -143,7 +143,7 @@ Configure `autoIndexPaths` to automatically index files when the workspace initi
|
|
|
143
143
|
const workspace = new Workspace({
|
|
144
144
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
145
145
|
bm25: true,
|
|
146
|
-
autoIndexPaths: ['
|
|
146
|
+
autoIndexPaths: ['docs', 'support/faq'],
|
|
147
147
|
})
|
|
148
148
|
|
|
149
149
|
await workspace.init()
|
|
@@ -157,7 +157,7 @@ Glob patterns let you index specific file types:
|
|
|
157
157
|
const workspace = new Workspace({
|
|
158
158
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
159
159
|
bm25: true,
|
|
160
|
-
autoIndexPaths: ['
|
|
160
|
+
autoIndexPaths: ['docs/**/*.md', 'support/**/*.txt'],
|
|
161
161
|
})
|
|
162
162
|
```
|
|
163
163
|
|
|
@@ -12,13 +12,13 @@ A skill is a folder containing:
|
|
|
12
12
|
- `assets/`: Images and other files (optional)
|
|
13
13
|
|
|
14
14
|
```plaintext
|
|
15
|
-
/
|
|
16
|
-
|
|
15
|
+
skills/
|
|
16
|
+
code-review/
|
|
17
17
|
SKILL.md
|
|
18
|
-
/
|
|
18
|
+
references/
|
|
19
19
|
style-guide.md
|
|
20
20
|
pr-checklist.md
|
|
21
|
-
/
|
|
21
|
+
scripts/
|
|
22
22
|
lint.ts
|
|
23
23
|
```
|
|
24
24
|
|
|
@@ -64,18 +64,18 @@ import { Workspace, LocalFilesystem } from '@mastra/core/workspace'
|
|
|
64
64
|
|
|
65
65
|
const workspace = new Workspace({
|
|
66
66
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
67
|
-
skills: ['
|
|
67
|
+
skills: ['skills'],
|
|
68
68
|
})
|
|
69
69
|
```
|
|
70
70
|
|
|
71
|
-
|
|
71
|
+
You can specify multiple skill directories:
|
|
72
72
|
|
|
73
73
|
```typescript
|
|
74
74
|
const workspace = new Workspace({
|
|
75
75
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
76
76
|
skills: [
|
|
77
|
-
'
|
|
78
|
-
'
|
|
77
|
+
'skills', // Project skills
|
|
78
|
+
'team-skills', // Shared team skills
|
|
79
79
|
],
|
|
80
80
|
})
|
|
81
81
|
```
|
|
@@ -85,7 +85,7 @@ You can also pass a direct path to a skill directory or `SKILL.md` file:
|
|
|
85
85
|
```typescript
|
|
86
86
|
const workspace = new Workspace({
|
|
87
87
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
88
|
-
skills: ['
|
|
88
|
+
skills: ['path/to/my-skill'],
|
|
89
89
|
})
|
|
90
90
|
```
|
|
91
91
|
|
|
@@ -105,10 +105,10 @@ For dynamic skill paths based on context, pass a function:
|
|
|
105
105
|
```typescript
|
|
106
106
|
const workspace = new Workspace({
|
|
107
107
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
108
|
-
skills:
|
|
109
|
-
const paths = ['
|
|
110
|
-
if (
|
|
111
|
-
paths.push('
|
|
108
|
+
skills: ctx => {
|
|
109
|
+
const paths = ['skills']
|
|
110
|
+
if (ctx.requestContext?.get('userRole') === 'developer') {
|
|
111
|
+
paths.push('dev-skills')
|
|
112
112
|
}
|
|
113
113
|
return paths
|
|
114
114
|
},
|
|
@@ -142,7 +142,7 @@ const workspace = new Workspace({
|
|
|
142
142
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
143
143
|
skills: [
|
|
144
144
|
'node_modules/@myorg/skills', // external: provides "brand-guidelines"
|
|
145
|
-
'
|
|
145
|
+
'skills', // local: also provides "brand-guidelines"
|
|
146
146
|
],
|
|
147
147
|
})
|
|
148
148
|
|
|
@@ -157,7 +157,7 @@ If BM25 or vector search is enabled on the workspace, skills are automatically i
|
|
|
157
157
|
```typescript
|
|
158
158
|
const workspace = new Workspace({
|
|
159
159
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
160
|
-
skills: ['
|
|
160
|
+
skills: ['skills'],
|
|
161
161
|
bm25: true,
|
|
162
162
|
})
|
|
163
163
|
```
|
|
@@ -174,7 +174,7 @@ import { VersionedSkillSource } from '@mastra/core/workspace'
|
|
|
174
174
|
|
|
175
175
|
const workspace = new Workspace({
|
|
176
176
|
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
177
|
-
skills: ['
|
|
177
|
+
skills: ['skills'],
|
|
178
178
|
skillSource: new VersionedSkillSource(versionTree, blobStore, versionCreatedAt),
|
|
179
179
|
})
|
|
180
180
|
```
|
|
@@ -21,7 +21,7 @@ const workspace = new Workspace({
|
|
|
21
21
|
filesystem: new LocalFilesystem({
|
|
22
22
|
basePath: resolve(import.meta.dirname, '../../workspace'),
|
|
23
23
|
}),
|
|
24
|
-
skills: ['
|
|
24
|
+
skills: ['skills'],
|
|
25
25
|
})
|
|
26
26
|
|
|
27
27
|
export const mastra = new Mastra({
|
|
@@ -138,7 +138,7 @@ const workspace = new Workspace({
|
|
|
138
138
|
filesystem: new LocalFilesystem({
|
|
139
139
|
basePath: resolve(import.meta.dirname, '../../workspace'),
|
|
140
140
|
}),
|
|
141
|
-
skills: ['
|
|
141
|
+
skills: ['skills'],
|
|
142
142
|
})
|
|
143
143
|
|
|
144
144
|
export const mastra = new Mastra({
|
|
@@ -57,9 +57,9 @@ import { Workspace, LocalFilesystem, LocalSandbox } from '@mastra/core/workspace
|
|
|
57
57
|
const workspace = new Workspace({
|
|
58
58
|
filesystem: new LocalFilesystem({ basePath: resolve(import.meta.dirname, '../../workspace') }),
|
|
59
59
|
sandbox: new LocalSandbox({ workingDirectory: resolve(import.meta.dirname, '../../workspace') }),
|
|
60
|
-
skills: ['
|
|
60
|
+
skills: ['skills'],
|
|
61
61
|
bm25: true,
|
|
62
|
-
autoIndexPaths: ['
|
|
62
|
+
autoIndexPaths: ['docs', 'src'],
|
|
63
63
|
})
|
|
64
64
|
|
|
65
65
|
export const mastra = new Mastra({
|
|
@@ -210,9 +210,9 @@ import { devAssistant } from './agents/dev-assistant'
|
|
|
210
210
|
const workspace = new Workspace({
|
|
211
211
|
filesystem: new LocalFilesystem({ basePath: resolve(import.meta.dirname, '../../workspace') }),
|
|
212
212
|
sandbox: new LocalSandbox({ workingDirectory: resolve(import.meta.dirname, '../../workspace') }),
|
|
213
|
-
skills: ['
|
|
213
|
+
skills: ['skills'],
|
|
214
214
|
bm25: true,
|
|
215
|
-
autoIndexPaths: ['
|
|
215
|
+
autoIndexPaths: ['docs', 'src'],
|
|
216
216
|
})
|
|
217
217
|
|
|
218
218
|
export const mastra = new Mastra({
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Netlify
|
|
2
2
|
|
|
3
|
-
Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access
|
|
3
|
+
Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access 66 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Netlify documentation](https://docs.netlify.com/build/ai-gateway/overview/).
|
|
6
6
|
|
|
@@ -94,6 +94,10 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
94
94
|
| `openai/gpt-5.4-nano-2026-03-17` |
|
|
95
95
|
| `openai/gpt-5.4-pro` |
|
|
96
96
|
| `openai/gpt-5.4-pro-2026-03-05` |
|
|
97
|
+
| `openai/gpt-5.5` |
|
|
98
|
+
| `openai/gpt-5.5-2026-04-23` |
|
|
99
|
+
| `openai/gpt-5.5-pro` |
|
|
100
|
+
| `openai/gpt-5.5-pro-2026-04-23` |
|
|
97
101
|
| `openai/o3` |
|
|
98
102
|
| `openai/o3-mini` |
|
|
99
103
|
| `openai/o4-mini` |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenRouter
|
|
2
2
|
|
|
3
|
-
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 175 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
|
|
6
6
|
|
|
@@ -61,6 +61,8 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
61
61
|
| `deepseek/deepseek-v3.1-terminus:exacto` |
|
|
62
62
|
| `deepseek/deepseek-v3.2` |
|
|
63
63
|
| `deepseek/deepseek-v3.2-speciale` |
|
|
64
|
+
| `deepseek/deepseek-v4-flash` |
|
|
65
|
+
| `deepseek/deepseek-v4-pro` |
|
|
64
66
|
| `google/gemini-2.0-flash-001` |
|
|
65
67
|
| `google/gemini-2.5-flash` |
|
|
66
68
|
| `google/gemini-2.5-flash-lite` |
|
|
@@ -151,6 +153,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
151
153
|
| `openai/gpt-5.4-mini` |
|
|
152
154
|
| `openai/gpt-5.4-nano` |
|
|
153
155
|
| `openai/gpt-5.4-pro` |
|
|
156
|
+
| `openai/gpt-5.5` |
|
|
154
157
|
| `openai/gpt-oss-120b` |
|
|
155
158
|
| `openai/gpt-oss-120b:exacto` |
|
|
156
159
|
| `openai/gpt-oss-120b:free` |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vercel
|
|
2
2
|
|
|
3
|
-
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 240 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
|
|
6
6
|
|
|
@@ -38,6 +38,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
38
38
|
| `alibaba/qwen-3-235b` |
|
|
39
39
|
| `alibaba/qwen-3-30b` |
|
|
40
40
|
| `alibaba/qwen-3-32b` |
|
|
41
|
+
| `alibaba/qwen-3.6-max-preview` |
|
|
41
42
|
| `alibaba/qwen3-235b-a22b-thinking` |
|
|
42
43
|
| `alibaba/qwen3-coder` |
|
|
43
44
|
| `alibaba/qwen3-coder-30b-a3b` |
|
|
@@ -95,6 +96,8 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
95
96
|
| `deepseek/deepseek-v3.2` |
|
|
96
97
|
| `deepseek/deepseek-v3.2-exp` |
|
|
97
98
|
| `deepseek/deepseek-v3.2-thinking` |
|
|
99
|
+
| `deepseek/deepseek-v4-flash` |
|
|
100
|
+
| `deepseek/deepseek-v4-pro` |
|
|
98
101
|
| `google/gemini-2.0-flash` |
|
|
99
102
|
| `google/gemini-2.0-flash-lite` |
|
|
100
103
|
| `google/gemini-2.5-flash` |
|
|
@@ -166,6 +169,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
166
169
|
| `moonshotai/kimi-k2-thinking-turbo` |
|
|
167
170
|
| `moonshotai/kimi-k2-turbo` |
|
|
168
171
|
| `moonshotai/kimi-k2.5` |
|
|
172
|
+
| `moonshotai/kimi-k2.6` |
|
|
169
173
|
| `morph/morph-v3-fast` |
|
|
170
174
|
| `morph/morph-v3-large` |
|
|
171
175
|
| `nvidia/nemotron-3-nano-30b-a3b` |
|
|
@@ -203,6 +207,8 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
203
207
|
| `openai/gpt-5.4-mini` |
|
|
204
208
|
| `openai/gpt-5.4-nano` |
|
|
205
209
|
| `openai/gpt-5.4-pro` |
|
|
210
|
+
| `openai/gpt-5.5` |
|
|
211
|
+
| `openai/gpt-5.5-pro` |
|
|
206
212
|
| `openai/gpt-oss-120b` |
|
|
207
213
|
| `openai/gpt-oss-20b` |
|
|
208
214
|
| `openai/gpt-oss-safeguard-20b` |
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3743 models from 104 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Nvidia
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 78 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Nvidia documentation](https://docs.api.nvidia.com/nim/).
|
|
6
6
|
|
|
@@ -41,6 +41,7 @@ for await (const chunk of stream) {
|
|
|
41
41
|
| `nvidia/deepseek-ai/deepseek-v3.1` | 128K | | | | | | — | — |
|
|
42
42
|
| `nvidia/deepseek-ai/deepseek-v3.1-terminus` | 128K | | | | | | — | — |
|
|
43
43
|
| `nvidia/deepseek-ai/deepseek-v3.2` | 164K | | | | | | — | — |
|
|
44
|
+
| `nvidia/deepseek-ai/deepseek-v4-pro` | 1.0M | | | | | | $2 | $3 |
|
|
44
45
|
| `nvidia/google/codegemma-1.1-7b` | 128K | | | | | | — | — |
|
|
45
46
|
| `nvidia/google/codegemma-7b` | 128K | | | | | | — | — |
|
|
46
47
|
| `nvidia/google/gemma-2-27b-it` | 128K | | | | | | — | — |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenCode Go
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 14 OpenCode Go models through Mastra's model router. Authentication is handled automatically using the `OPENCODE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenCode Go documentation](https://opencode.ai/docs/zen).
|
|
6
6
|
|
|
@@ -15,7 +15,7 @@ const agent = new Agent({
|
|
|
15
15
|
id: "my-agent",
|
|
16
16
|
name: "My Agent",
|
|
17
17
|
instructions: "You are a helpful assistant",
|
|
18
|
-
model: "opencode-go/
|
|
18
|
+
model: "opencode-go/deepseek-v4-flash"
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Generate a response
|
|
@@ -32,20 +32,22 @@ for await (const chunk of stream) {
|
|
|
32
32
|
|
|
33
33
|
## Models
|
|
34
34
|
|
|
35
|
-
| Model
|
|
36
|
-
|
|
|
37
|
-
| `opencode-go/
|
|
38
|
-
| `opencode-go/
|
|
39
|
-
| `opencode-go/
|
|
40
|
-
| `opencode-go/
|
|
41
|
-
| `opencode-go/
|
|
42
|
-
| `opencode-go/
|
|
43
|
-
| `opencode-go/mimo-v2
|
|
44
|
-
| `opencode-go/mimo-v2
|
|
45
|
-
| `opencode-go/
|
|
46
|
-
| `opencode-go/
|
|
47
|
-
| `opencode-go/
|
|
48
|
-
| `opencode-go/
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `opencode-go/deepseek-v4-flash` | 1.0M | | | | | | $0.14 | $0.28 |
|
|
38
|
+
| `opencode-go/deepseek-v4-pro` | 1.0M | | | | | | $2 | $3 |
|
|
39
|
+
| `opencode-go/glm-5` | 205K | | | | | | $1 | $3 |
|
|
40
|
+
| `opencode-go/glm-5.1` | 205K | | | | | | $1 | $4 |
|
|
41
|
+
| `opencode-go/kimi-k2.5` | 262K | | | | | | $0.60 | $3 |
|
|
42
|
+
| `opencode-go/kimi-k2.6` | 262K | | | | | | $0.32 | $1 |
|
|
43
|
+
| `opencode-go/mimo-v2-omni` | 262K | | | | | | $0.40 | $2 |
|
|
44
|
+
| `opencode-go/mimo-v2-pro` | 1.0M | | | | | | $1 | $3 |
|
|
45
|
+
| `opencode-go/mimo-v2.5` | 262K | | | | | | $0.40 | $2 |
|
|
46
|
+
| `opencode-go/mimo-v2.5-pro` | 1.0M | | | | | | $1 | $3 |
|
|
47
|
+
| `opencode-go/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
|
|
48
|
+
| `opencode-go/minimax-m2.7` | 205K | | | | | | $0.30 | $1 |
|
|
49
|
+
| `opencode-go/qwen3.5-plus` | 262K | | | | | | $0.20 | $1 |
|
|
50
|
+
| `opencode-go/qwen3.6-plus` | 262K | | | | | | $0.50 | $3 |
|
|
49
51
|
|
|
50
52
|
## Advanced configuration
|
|
51
53
|
|
|
@@ -57,7 +59,7 @@ const agent = new Agent({
|
|
|
57
59
|
name: "custom-agent",
|
|
58
60
|
model: {
|
|
59
61
|
url: "https://opencode.ai/zen/go/v1",
|
|
60
|
-
id: "opencode-go/
|
|
62
|
+
id: "opencode-go/deepseek-v4-flash",
|
|
61
63
|
apiKey: process.env.OPENCODE_API_KEY,
|
|
62
64
|
headers: {
|
|
63
65
|
"X-Custom-Header": "value"
|
|
@@ -76,7 +78,7 @@ const agent = new Agent({
|
|
|
76
78
|
const useAdvanced = requestContext.task === "complex";
|
|
77
79
|
return useAdvanced
|
|
78
80
|
? "opencode-go/qwen3.6-plus"
|
|
79
|
-
: "opencode-go/
|
|
81
|
+
: "opencode-go/deepseek-v4-flash";
|
|
80
82
|
}
|
|
81
83
|
});
|
|
82
84
|
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenCode Zen
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 41 OpenCode Zen models through Mastra's model router. Authentication is handled automatically using the `OPENCODE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenCode Zen documentation](https://opencode.ai/docs/zen).
|
|
6
6
|
|
|
@@ -63,6 +63,8 @@ for await (const chunk of stream) {
|
|
|
63
63
|
| `opencode/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
64
64
|
| `opencode/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
65
65
|
| `opencode/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
|
|
66
|
+
| `opencode/gpt-5.5` | 1.1M | | | | | | $5 | $30 |
|
|
67
|
+
| `opencode/gpt-5.5-pro` | 1.1M | | | | | | $30 | $180 |
|
|
66
68
|
| `opencode/hy3-preview-free` | 256K | | | | | | — | — |
|
|
67
69
|
| `opencode/kimi-k2.5` | 262K | | | | | | $0.60 | $3 |
|
|
68
70
|
| `opencode/kimi-k2.6` | 262K | | | | | | $0.95 | $4 |
|
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,12 @@
|
|
|
1
1
|
# @mastra/mcp-docs-server
|
|
2
2
|
|
|
3
|
+
## 1.1.29-alpha.2
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [[`7a7b313`](https://github.com/mastra-ai/mastra/commit/7a7b3138fb3bcf0b0c740eaea07971e43d330ef3), [`a6dac0a`](https://github.com/mastra-ai/mastra/commit/a6dac0a40c7181161b1add4e8534f962bcbc9aa7), [`9cef83b`](https://github.com/mastra-ai/mastra/commit/9cef83b8a642b8098747772921e3523b492bafbc), [`d30e215`](https://github.com/mastra-ai/mastra/commit/d30e2156c746bc9fd791745cec1cc24377b66789), [`73b45fa`](https://github.com/mastra-ai/mastra/commit/73b45facdef4fbcb8af710c50f0646f18619dbaa), [`7a7b313`](https://github.com/mastra-ai/mastra/commit/7a7b3138fb3bcf0b0c740eaea07971e43d330ef3)]:
|
|
8
|
+
- @mastra/core@1.29.0-alpha.1
|
|
9
|
+
|
|
3
10
|
## 1.1.29-alpha.0
|
|
4
11
|
|
|
5
12
|
### Patch Changes
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.29-alpha.
|
|
3
|
+
"version": "1.1.29-alpha.3",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,8 +29,8 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/
|
|
33
|
-
"@mastra/
|
|
32
|
+
"@mastra/mcp": "^1.5.2",
|
|
33
|
+
"@mastra/core": "1.29.0-alpha.1"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
36
36
|
"@hono/node-server": "^1.19.11",
|
|
@@ -48,7 +48,7 @@
|
|
|
48
48
|
"vitest": "4.1.5",
|
|
49
49
|
"@internal/lint": "0.0.86",
|
|
50
50
|
"@internal/types-builder": "0.0.61",
|
|
51
|
-
"@mastra/core": "1.29.0-alpha.
|
|
51
|
+
"@mastra/core": "1.29.0-alpha.1"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|