@mastra/mcp-docs-server 1.1.16-alpha.7 → 1.1.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/memory/observational-memory.md +36 -0
- package/.docs/docs/observability/tracing/exporters/datadog.md +132 -2
- package/.docs/docs/workspace/skills.md +23 -0
- package/.docs/models/providers/vivgrid.md +13 -12
- package/.docs/models/providers/vultr.md +1 -2
- package/.docs/reference/memory/observational-memory.md +42 -3
- package/.docs/reference/tools/create-tool.md +1 -1
- package/.docs/reference/workspace/workspace-class.md +13 -1
- package/CHANGELOG.md +30 -0
- package/package.json +5 -5
|
@@ -137,6 +137,42 @@ const memory = new Memory({
|
|
|
137
137
|
|
|
138
138
|
See [model configuration](https://mastra.ai/reference/memory/observational-memory) for using different models per agent.
|
|
139
139
|
|
|
140
|
+
### Token-tiered model selection
|
|
141
|
+
|
|
142
|
+
You can use `ModelByInputTokens` to specify different Observer or Reflector models based on input token count. OM selects the matching model tier at runtime from the configured `upTo` thresholds.
|
|
143
|
+
|
|
144
|
+
```typescript
|
|
145
|
+
import { Memory, ModelByInputTokens } from '@mastra/memory'
|
|
146
|
+
|
|
147
|
+
const memory = new Memory({
|
|
148
|
+
options: {
|
|
149
|
+
observationalMemory: {
|
|
150
|
+
observation: {
|
|
151
|
+
model: new ModelByInputTokens({
|
|
152
|
+
upTo: {
|
|
153
|
+
10_000: 'google/gemini-2.5-flash', // Fast and cheap for small inputs
|
|
154
|
+
40_000: 'openai/gpt-4o', // Stronger for medium inputs
|
|
155
|
+
1_000_000: 'openai/gpt-4.5', // Most capable for very large inputs
|
|
156
|
+
},
|
|
157
|
+
}),
|
|
158
|
+
},
|
|
159
|
+
reflection: {
|
|
160
|
+
model: new ModelByInputTokens({
|
|
161
|
+
upTo: {
|
|
162
|
+
20_000: 'google/gemini-2.5-flash',
|
|
163
|
+
80_000: 'openai/gpt-4o',
|
|
164
|
+
},
|
|
165
|
+
}),
|
|
166
|
+
},
|
|
167
|
+
},
|
|
168
|
+
},
|
|
169
|
+
})
|
|
170
|
+
```
|
|
171
|
+
|
|
172
|
+
The `upTo` keys are inclusive upper bounds. OM computes the actual input token count for the Observer or Reflector call, resolves the matching tier directly, and uses that concrete model for the run.
|
|
173
|
+
|
|
174
|
+
If the input exceeds the largest configured threshold, an error is thrown — ensure your thresholds cover the full range of possible input sizes, or use a model with a sufficiently large context window at the highest tier.
|
|
175
|
+
|
|
140
176
|
## Scopes
|
|
141
177
|
|
|
142
178
|
### Thread scope (default)
|
|
@@ -145,6 +145,135 @@ Mastra span types are automatically mapped to Datadog LLMObs span kinds:
|
|
|
145
145
|
|
|
146
146
|
Other/future Mastra span types will default to 'task' when mapped unless specified.
|
|
147
147
|
|
|
148
|
+
## Application Performance Monitoring
|
|
149
|
+
|
|
150
|
+
The sections above cover Mastra's [LLM Observability](https://docs.datadoghq.com/llm_observability/) integration. To trace your Mastra HTTP server routes (request latency, error tracking, service maps), use `dd-trace` directly for Datadog Application Performance Monitoring (APM).
|
|
151
|
+
|
|
152
|
+
### Prerequisites
|
|
153
|
+
|
|
154
|
+
1. **Datadog Agent**: Install a [Datadog Agent](https://docs.datadoghq.com/agent/) on the same host or accessible via network. The agent receives traces from `dd-trace` on `localhost:8126` and forwards them to Datadog. Follow the [agent installation guide](https://docs.datadoghq.com/agent/) to set it up.
|
|
155
|
+
|
|
156
|
+
2. **dd-trace package**: Install the tracing library in your project:
|
|
157
|
+
|
|
158
|
+
**npm**:
|
|
159
|
+
|
|
160
|
+
```bash
|
|
161
|
+
npm install dd-trace
|
|
162
|
+
```
|
|
163
|
+
|
|
164
|
+
**pnpm**:
|
|
165
|
+
|
|
166
|
+
```bash
|
|
167
|
+
pnpm add dd-trace
|
|
168
|
+
```
|
|
169
|
+
|
|
170
|
+
**Yarn**:
|
|
171
|
+
|
|
172
|
+
```bash
|
|
173
|
+
yarn add dd-trace
|
|
174
|
+
```
|
|
175
|
+
|
|
176
|
+
**Bun**:
|
|
177
|
+
|
|
178
|
+
```bash
|
|
179
|
+
bun add dd-trace
|
|
180
|
+
```
|
|
181
|
+
|
|
182
|
+
> **Note:** APM traces always route through the Datadog Agent. This is different from LLM Observability, which supports agentless mode (direct HTTPS to Datadog).
|
|
183
|
+
|
|
184
|
+
### APM only
|
|
185
|
+
|
|
186
|
+
Import and initialize `dd-trace` at the top of your entry file, before any other imports:
|
|
187
|
+
|
|
188
|
+
```typescript
|
|
189
|
+
import tracer from 'dd-trace'
|
|
190
|
+
|
|
191
|
+
tracer.init({
|
|
192
|
+
service: process.env.DD_SERVICE || 'my-mastra-app',
|
|
193
|
+
env: process.env.DD_ENV || 'production',
|
|
194
|
+
version: process.env.DD_VERSION,
|
|
195
|
+
})
|
|
196
|
+
|
|
197
|
+
import { Mastra } from '@mastra/core'
|
|
198
|
+
|
|
199
|
+
export const mastra = new Mastra({
|
|
200
|
+
bundler: {
|
|
201
|
+
externals: [
|
|
202
|
+
'dd-trace',
|
|
203
|
+
'@datadog/native-metrics',
|
|
204
|
+
'@datadog/native-appsec',
|
|
205
|
+
'@datadog/native-iast-taint-tracking',
|
|
206
|
+
'@datadog/pprof',
|
|
207
|
+
],
|
|
208
|
+
},
|
|
209
|
+
})
|
|
210
|
+
```
|
|
211
|
+
|
|
212
|
+
Set the tracer metadata environment variables:
|
|
213
|
+
|
|
214
|
+
```bash
|
|
215
|
+
DD_SERVICE=my-mastra-app
|
|
216
|
+
DD_ENV=production
|
|
217
|
+
DD_VERSION=1.0.0
|
|
218
|
+
```
|
|
219
|
+
|
|
220
|
+
`dd-trace` auto-instruments popular HTTP frameworks, including those supported by Mastra's [server adapters](https://mastra.ai/docs/server/server-adapters). Inbound requests, outbound HTTP calls, and database queries appear as APM traces in Datadog.
|
|
221
|
+
|
|
222
|
+
### APM and LLM Observability
|
|
223
|
+
|
|
224
|
+
Import and initialize `dd-trace` before creating the Mastra instance. The `DatadogExporter` detects the existing tracer and skips re-initialization, adding LLM Observability on top of your APM setup:
|
|
225
|
+
|
|
226
|
+
```typescript
|
|
227
|
+
import tracer from 'dd-trace'
|
|
228
|
+
|
|
229
|
+
tracer.init({
|
|
230
|
+
service: process.env.DD_SERVICE || 'my-mastra-app',
|
|
231
|
+
env: process.env.DD_ENV || 'production',
|
|
232
|
+
version: process.env.DD_VERSION,
|
|
233
|
+
})
|
|
234
|
+
|
|
235
|
+
import { Mastra } from '@mastra/core'
|
|
236
|
+
import { Observability } from '@mastra/observability'
|
|
237
|
+
import { DatadogExporter } from '@mastra/datadog'
|
|
238
|
+
|
|
239
|
+
export const mastra = new Mastra({
|
|
240
|
+
observability: new Observability({
|
|
241
|
+
configs: {
|
|
242
|
+
datadog: {
|
|
243
|
+
serviceName: 'my-mastra-app',
|
|
244
|
+
exporters: [
|
|
245
|
+
new DatadogExporter({
|
|
246
|
+
mlApp: process.env.DD_LLMOBS_ML_APP!,
|
|
247
|
+
apiKey: process.env.DD_API_KEY!,
|
|
248
|
+
}),
|
|
249
|
+
],
|
|
250
|
+
},
|
|
251
|
+
},
|
|
252
|
+
}),
|
|
253
|
+
bundler: {
|
|
254
|
+
externals: [
|
|
255
|
+
'dd-trace',
|
|
256
|
+
'@datadog/native-metrics',
|
|
257
|
+
'@datadog/native-appsec',
|
|
258
|
+
'@datadog/native-iast-taint-tracking',
|
|
259
|
+
'@datadog/pprof',
|
|
260
|
+
],
|
|
261
|
+
},
|
|
262
|
+
})
|
|
263
|
+
```
|
|
264
|
+
|
|
265
|
+
```bash
|
|
266
|
+
DD_SERVICE=my-mastra-app
|
|
267
|
+
DD_ENV=production
|
|
268
|
+
DD_VERSION=1.0.0
|
|
269
|
+
DD_API_KEY=your-datadog-api-key
|
|
270
|
+
DD_LLMOBS_ML_APP=my-llm-app
|
|
271
|
+
```
|
|
272
|
+
|
|
273
|
+
Server routes appear as APM traces and LLM calls appear as LLM Observability spans, all under the same service in Datadog.
|
|
274
|
+
|
|
275
|
+
> **Note:** Import and initialize `dd-trace` before all other modules. This allows its auto-instrumentation to patch HTTP, database, and framework libraries at load time.
|
|
276
|
+
|
|
148
277
|
## Troubleshooting
|
|
149
278
|
|
|
150
279
|
### Native module ABI mismatch
|
|
@@ -183,5 +312,6 @@ export const mastra = new Mastra({
|
|
|
183
312
|
|
|
184
313
|
## Related
|
|
185
314
|
|
|
186
|
-
- [Tracing
|
|
187
|
-
- [Datadog LLM Observability
|
|
315
|
+
- [Tracing overview](https://mastra.ai/docs/observability/tracing/overview)
|
|
316
|
+
- [Datadog LLM Observability documentation](https://docs.datadoghq.com/llm_observability/)
|
|
317
|
+
- [Datadog APM documentation](https://docs.datadoghq.com/tracing/)
|
|
@@ -127,6 +127,29 @@ The agent has three skill tools:
|
|
|
127
127
|
|
|
128
128
|
This design is stateless — there is no activation state to track. If the skill instructions leave the conversation context (due to context window limits or compaction), the agent can call `skill` again to reload them.
|
|
129
129
|
|
|
130
|
+
## Same-named skills
|
|
131
|
+
|
|
132
|
+
When multiple skill directories contain a skill with the same name, all of them are discovered and listed. The agent sees every skill in its system message, along with each skill's path and source type, so it can tell them apart.
|
|
133
|
+
|
|
134
|
+
When the agent activates a skill by name, tie-breaking determines which one is returned:
|
|
135
|
+
|
|
136
|
+
1. **Source-type priority**: local skills take precedence over managed (`.mastra/`) skills, which take precedence over external (`node_modules/`) skills.
|
|
137
|
+
2. **Unresolvable conflicts throw**: if two skills share the same name _and_ the same source type (for example, two local skills both named `brand-guidelines`), `get()` throws an error. Rename one or move it to a different source type to resolve the conflict.
|
|
138
|
+
3. **Path escape hatch**: the agent can pass a skill's full path instead of its name to activate a specific skill, bypassing tie-breaking entirely.
|
|
139
|
+
|
|
140
|
+
```typescript
|
|
141
|
+
const workspace = new Workspace({
|
|
142
|
+
filesystem: new LocalFilesystem({ basePath: './workspace' }),
|
|
143
|
+
skills: [
|
|
144
|
+
'node_modules/@myorg/skills', // external: provides "brand-guidelines"
|
|
145
|
+
'/skills', // local: also provides "brand-guidelines"
|
|
146
|
+
],
|
|
147
|
+
})
|
|
148
|
+
|
|
149
|
+
// get('brand-guidelines') returns the local copy (local > external)
|
|
150
|
+
// get('node_modules/@myorg/skills/brand-guidelines') returns the external copy
|
|
151
|
+
```
|
|
152
|
+
|
|
130
153
|
## Skill search
|
|
131
154
|
|
|
132
155
|
If BM25 or vector search is enabled on the workspace, skills are automatically indexed. Agents can search across skill content to find relevant instructions.
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vivgrid
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 9 Vivgrid models through Mastra's model router. Authentication is handled automatically using the `VIVGRID_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vivgrid documentation](https://docs.vivgrid.com/models).
|
|
6
6
|
|
|
@@ -32,16 +32,17 @@ for await (const chunk of stream) {
|
|
|
32
32
|
|
|
33
33
|
## Models
|
|
34
34
|
|
|
35
|
-
| Model
|
|
36
|
-
|
|
|
37
|
-
| `vivgrid/deepseek-v3.2`
|
|
38
|
-
| `vivgrid/gemini-3-flash-preview` | 1.0M | | | | | | $0.
|
|
39
|
-
| `vivgrid/gemini-3-pro-preview`
|
|
40
|
-
| `vivgrid/glm-5`
|
|
41
|
-
| `vivgrid/gpt-5-mini`
|
|
42
|
-
| `vivgrid/gpt-5.1-codex`
|
|
43
|
-
| `vivgrid/gpt-5.1-codex-max`
|
|
44
|
-
| `vivgrid/gpt-5.2-codex`
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| --------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `vivgrid/deepseek-v3.2` | 128K | | | | | | $0.28 | $0.42 |
|
|
38
|
+
| `vivgrid/gemini-3.1-flash-lite-preview` | 1.0M | | | | | | $0.25 | $2 |
|
|
39
|
+
| `vivgrid/gemini-3.1-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
40
|
+
| `vivgrid/glm-5` | 203K | | | | | | $1 | $3 |
|
|
41
|
+
| `vivgrid/gpt-5-mini` | 272K | | | | | | $0.25 | $2 |
|
|
42
|
+
| `vivgrid/gpt-5.1-codex` | 400K | | | | | | $1 | $10 |
|
|
43
|
+
| `vivgrid/gpt-5.1-codex-max` | 400K | | | | | | $1 | $10 |
|
|
44
|
+
| `vivgrid/gpt-5.2-codex` | 400K | | | | | | $2 | $14 |
|
|
45
|
+
| `vivgrid/gpt-5.4` | 400K | | | | | | $3 | $15 |
|
|
45
46
|
|
|
46
47
|
## Advanced configuration
|
|
47
48
|
|
|
@@ -71,7 +72,7 @@ const agent = new Agent({
|
|
|
71
72
|
model: ({ requestContext }) => {
|
|
72
73
|
const useAdvanced = requestContext.task === "complex";
|
|
73
74
|
return useAdvanced
|
|
74
|
-
? "vivgrid/gpt-5.
|
|
75
|
+
? "vivgrid/gpt-5.4"
|
|
75
76
|
: "vivgrid/deepseek-v3.2";
|
|
76
77
|
}
|
|
77
78
|
});
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vultr
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 9 Vultr models through Mastra's model router. Authentication is handled automatically using the `VULTR_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vultr documentation](https://api.vultrinference.com/).
|
|
6
6
|
|
|
@@ -43,7 +43,6 @@ for await (const chunk of stream) {
|
|
|
43
43
|
| `vultr/Llama-3_1-Nemotron-Ultra-253B-v1` | 32K | | | | | | $0.55 | $2 |
|
|
44
44
|
| `vultr/MiniMax-M2.5` | 196K | | | | | | $0.30 | $1 |
|
|
45
45
|
| `vultr/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4` | 260K | | | | | | $0.20 | $0.80 |
|
|
46
|
-
| `vultr/Qwen2.5-Coder-32B-Instruct` | 15K | | | | | | $0.20 | $0.60 |
|
|
47
46
|
|
|
48
47
|
## Advanced configuration
|
|
49
48
|
|
|
@@ -32,7 +32,7 @@ OM performs thresholding with fast local token estimation. Text uses `tokenx`, a
|
|
|
32
32
|
|
|
33
33
|
**enabled** (`boolean`): Enable or disable Observational Memory. When omitted from a config object, defaults to \`true\`. Only \`enabled: false\` explicitly disables it. (Default: `true`)
|
|
34
34
|
|
|
35
|
-
**model** (`string | LanguageModel | DynamicModel | ModelWithRetries[]`): Model for both the Observer and Reflector agents. Sets the model for both at once. Cannot be used together with \`observation.model\` or \`reflection.model\` — an error will be thrown if both are set. When using \`observationalMemory: true\`, defaults to \`google/gemini-2.5-flash\`. When passing a config object, this or \`observation.model\`/\`reflection.model\` must be set. Use \`"default"\` to explicitly use the default model (\`google/gemini-2.5-flash\`). (Default: `'google/gemini-2.5-flash' (when using observationalMemory: true)`)
|
|
35
|
+
**model** (`string | LanguageModel | DynamicModel | ModelByInputTokens | ModelWithRetries[]`): Model for both the Observer and Reflector agents. Sets the model for both at once. Cannot be used together with \`observation.model\` or \`reflection.model\` — an error will be thrown if both are set. When using \`observationalMemory: true\`, defaults to \`google/gemini-2.5-flash\`. When passing a config object, this or \`observation.model\`/\`reflection.model\` must be set. Use \`"default"\` to explicitly use the default model (\`google/gemini-2.5-flash\`). (Default: `'google/gemini-2.5-flash' (when using observationalMemory: true)`)
|
|
36
36
|
|
|
37
37
|
**scope** (`'resource' | 'thread'`): Memory scope for observations. \`'thread'\` keeps observations per-thread. \`'resource'\` (experimental) shares observations across all threads for a resource, enabling cross-conversation memory. (Default: `'thread'`)
|
|
38
38
|
|
|
@@ -42,7 +42,7 @@ OM performs thresholding with fast local token estimation. Text uses `tokenx`, a
|
|
|
42
42
|
|
|
43
43
|
**observation** (`ObservationalMemoryObservationConfig`): Configuration for the observation step. Controls when the Observer agent runs and how it behaves.
|
|
44
44
|
|
|
45
|
-
**observation.model** (`string | LanguageModel | DynamicModel | ModelWithRetries[]`): Model for the Observer agent. Cannot be set if a top-level \`model\` is also provided. If neither this nor the top-level \`model\` is set, falls back to \`reflection.model\`.
|
|
45
|
+
**observation.model** (`string | LanguageModel | DynamicModel | ModelByInputTokens | ModelWithRetries[]`): Model for the Observer agent. Cannot be set if a top-level \`model\` is also provided. If neither this nor the top-level \`model\` is set, falls back to \`reflection.model\`.
|
|
46
46
|
|
|
47
47
|
**observation.instruction** (`string`): Custom instruction appended to the Observer's system prompt. Use this to customize what the Observer focuses on, such as domain-specific preferences or priorities.
|
|
48
48
|
|
|
@@ -68,7 +68,7 @@ OM performs thresholding with fast local token estimation. Text uses `tokenx`, a
|
|
|
68
68
|
|
|
69
69
|
**reflection** (`ObservationalMemoryReflectionConfig`): Configuration for the reflection step. Controls when the Reflector agent runs and how it behaves.
|
|
70
70
|
|
|
71
|
-
**reflection.model** (`string | LanguageModel | DynamicModel | ModelWithRetries[]`): Model for the Reflector agent. Cannot be set if a top-level \`model\` is also provided. If neither this nor the top-level \`model\` is set, falls back to \`observation.model\`.
|
|
71
|
+
**reflection.model** (`string | LanguageModel | DynamicModel | ModelByInputTokens | ModelWithRetries[]`): Model for the Reflector agent. Cannot be set if a top-level \`model\` is also provided. If neither this nor the top-level \`model\` is set, falls back to \`observation.model\`.
|
|
72
72
|
|
|
73
73
|
**reflection.instruction** (`string`): Custom instruction appended to the Reflector's system prompt. Use this to customize how the Reflector consolidates observations, such as prioritizing certain types of information.
|
|
74
74
|
|
|
@@ -612,6 +612,45 @@ When `retrieval: true` is set with `scope: 'thread'`, OM registers a `recall` to
|
|
|
612
612
|
|
|
613
613
|
**tokenOffset** (`number`): Approximate number of tokens that were trimmed when \`truncated\` is true.
|
|
614
614
|
|
|
615
|
+
### ModelByInputTokens
|
|
616
|
+
|
|
617
|
+
`ModelByInputTokens` selects a model based on the input token count. It chooses the model for the smallest threshold that covers the actual input size.
|
|
618
|
+
|
|
619
|
+
#### Constructor
|
|
620
|
+
|
|
621
|
+
```typescript
|
|
622
|
+
new ModelByInputTokens(config)
|
|
623
|
+
```
|
|
624
|
+
|
|
625
|
+
Where `config` is an object with `upTo` keys that map token thresholds (numbers) to model targets.
|
|
626
|
+
|
|
627
|
+
#### Example
|
|
628
|
+
|
|
629
|
+
```typescript
|
|
630
|
+
import { ModelByInputTokens } from '@mastra/memory'
|
|
631
|
+
|
|
632
|
+
const selector = new ModelByInputTokens({
|
|
633
|
+
upTo: {
|
|
634
|
+
10_000: 'google/gemini-2.5-flash', // Fast for small inputs
|
|
635
|
+
40_000: 'openai/gpt-4o', // Stronger for medium inputs
|
|
636
|
+
1_000_000: 'openai/gpt-4.5', // Most capable for large inputs
|
|
637
|
+
},
|
|
638
|
+
})
|
|
639
|
+
```
|
|
640
|
+
|
|
641
|
+
#### Behavior
|
|
642
|
+
|
|
643
|
+
- Thresholds are sorted internally, so the order in the config object does not matter.
|
|
644
|
+
- `inputTokens ≤ smallest threshold` → uses that threshold's model
|
|
645
|
+
- `inputTokens > largest threshold` → `resolve()` throws an error. If this happens during an OM Observer or Reflector run, OM aborts via TripWire, so callers receive an empty `text` result or streamed `tripwire` instead of a normal assistant response.
|
|
646
|
+
- OM computes the input token count for the Observer or Reflector call and resolves the matching model tier directly
|
|
647
|
+
|
|
648
|
+
#### Methods
|
|
649
|
+
|
|
650
|
+
**resolve** (`(inputTokens: number) => MastraModelConfig`): Returns the model for the given input token count. Throws if inputTokens exceeds the largest configured threshold. When this happens during an OM run, callers receive a TripWire/empty-text outcome instead of a normal assistant response.
|
|
651
|
+
|
|
652
|
+
**getThresholds** (`() => number[]`): Returns the configured thresholds in ascending order. Useful for introspection.
|
|
653
|
+
|
|
615
654
|
### Related
|
|
616
655
|
|
|
617
656
|
- [Observational Memory](https://mastra.ai/docs/memory/observational-memory)
|
|
@@ -144,7 +144,7 @@ export const weatherTool = createTool({
|
|
|
144
144
|
|
|
145
145
|
**execute.context.abortSignal** (`AbortSignal`): Signal for aborting the tool execution
|
|
146
146
|
|
|
147
|
-
**execute.context.agent** (`AgentToolExecutionContext`): Agent-specific context
|
|
147
|
+
**execute.context.agent** (`AgentToolExecutionContext`): Agent-specific context, available when the tool is executed by an agent.
|
|
148
148
|
|
|
149
149
|
**execute.context.workflow** (`WorkflowToolExecutionContext`): Workflow-specific context (state, setState, suspend, etc.)
|
|
150
150
|
|
|
@@ -305,4 +305,16 @@ Added when BM25 or vector search is configured:
|
|
|
305
305
|
| `mastra_workspace_search` | Search indexed content using keyword (BM25), semantic (vector), or hybrid search. Returns ranked results with scores. |
|
|
306
306
|
| `mastra_workspace_index` | Index content for search. Associates content with a path for later retrieval. |
|
|
307
307
|
|
|
308
|
-
The `index` tool is excluded when the filesystem is in read-only mode.
|
|
308
|
+
The `index` tool is excluded when the filesystem is in read-only mode.
|
|
309
|
+
|
|
310
|
+
### Skill tools
|
|
311
|
+
|
|
312
|
+
Added when skills are configured:
|
|
313
|
+
|
|
314
|
+
| Tool | Description |
|
|
315
|
+
| -------------- | --------------------------------------------------------------------------------------------------------- |
|
|
316
|
+
| `skill` | Activate a skill by name or path. Returns the skill's full instructions, references, scripts, and assets. |
|
|
317
|
+
| `skill_search` | Search across skill content. Accepts an optional list of skill names to filter and a `topK` parameter. |
|
|
318
|
+
| `skill_read` | Read a specific file (reference, script, or asset) from a skill directory. |
|
|
319
|
+
|
|
320
|
+
When multiple skills share the same name, `list()` returns all of them. `get()` with a name applies tie-breaking (local > managed > external). If two skills share the same name _and_ source type, `get()` throws an error. Pass a skill's full path to `get()` to bypass tie-breaking. See [Same-named skills](https://mastra.ai/docs/workspace/skills) for details.
|
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,35 @@
|
|
|
1
1
|
# @mastra/mcp-docs-server
|
|
2
2
|
|
|
3
|
+
## 1.1.16
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [[`68ed4e9`](https://github.com/mastra-ai/mastra/commit/68ed4e9f118e8646b60a6112dabe854d0ef53902), [`085c1da`](https://github.com/mastra-ai/mastra/commit/085c1daf71b55a97b8ebad26623089e40055021c), [`be37de4`](https://github.com/mastra-ai/mastra/commit/be37de4391bd1d5486ce38efacbf00ca51637262), [`7dbd611`](https://github.com/mastra-ai/mastra/commit/7dbd611a85cb1e0c0a1581c57564268cb183d86e), [`f14604c`](https://github.com/mastra-ai/mastra/commit/f14604c7ef01ba794e1a8d5c7bae5415852aacec), [`4a75e10`](https://github.com/mastra-ai/mastra/commit/4a75e106bd31c283a1b3fe74c923610dcc46415b), [`f3ce603`](https://github.com/mastra-ai/mastra/commit/f3ce603fd76180f4a5be90b6dc786d389b6b3e98), [`423aa6f`](https://github.com/mastra-ai/mastra/commit/423aa6fd12406de6a1cc6b68e463d30af1d790fb), [`f21c626`](https://github.com/mastra-ai/mastra/commit/f21c6263789903ab9720b4d11373093298e97f15), [`41aee84`](https://github.com/mastra-ai/mastra/commit/41aee84561ceebe28bad1ecba8702d92838f67f0), [`2871451`](https://github.com/mastra-ai/mastra/commit/2871451703829aefa06c4a5d6eca7fd3731222ef), [`085c1da`](https://github.com/mastra-ai/mastra/commit/085c1daf71b55a97b8ebad26623089e40055021c), [`4bb5adc`](https://github.com/mastra-ai/mastra/commit/4bb5adc05c88e3a83fe1ea5ecb9eae6e17313124), [`4bb5adc`](https://github.com/mastra-ai/mastra/commit/4bb5adc05c88e3a83fe1ea5ecb9eae6e17313124), [`e06b520`](https://github.com/mastra-ai/mastra/commit/e06b520bdd5fdef844760c5e692c7852cbc5c240), [`d3930ea`](https://github.com/mastra-ai/mastra/commit/d3930eac51c30b0ecf7eaa54bb9430758b399777), [`dd9c4e0`](https://github.com/mastra-ai/mastra/commit/dd9c4e0a47962f1413e9b72114fcad912e19a0a6)]:
|
|
8
|
+
- @mastra/core@1.16.0
|
|
9
|
+
- @mastra/mcp@1.3.1
|
|
10
|
+
|
|
11
|
+
## 1.1.16-alpha.11
|
|
12
|
+
|
|
13
|
+
### Patch Changes
|
|
14
|
+
|
|
15
|
+
- Updated dependencies [[`f21c626`](https://github.com/mastra-ai/mastra/commit/f21c6263789903ab9720b4d11373093298e97f15)]:
|
|
16
|
+
- @mastra/core@1.16.0-alpha.5
|
|
17
|
+
|
|
18
|
+
## 1.1.16-alpha.10
|
|
19
|
+
|
|
20
|
+
### Patch Changes
|
|
21
|
+
|
|
22
|
+
- Updated dependencies [[`f14604c`](https://github.com/mastra-ai/mastra/commit/f14604c7ef01ba794e1a8d5c7bae5415852aacec), [`e06b520`](https://github.com/mastra-ai/mastra/commit/e06b520bdd5fdef844760c5e692c7852cbc5c240), [`dd9c4e0`](https://github.com/mastra-ai/mastra/commit/dd9c4e0a47962f1413e9b72114fcad912e19a0a6)]:
|
|
23
|
+
- @mastra/core@1.16.0-alpha.4
|
|
24
|
+
|
|
25
|
+
## 1.1.16-alpha.8
|
|
26
|
+
|
|
27
|
+
### Patch Changes
|
|
28
|
+
|
|
29
|
+
- Updated dependencies [[`423aa6f`](https://github.com/mastra-ai/mastra/commit/423aa6fd12406de6a1cc6b68e463d30af1d790fb), [`4bb5adc`](https://github.com/mastra-ai/mastra/commit/4bb5adc05c88e3a83fe1ea5ecb9eae6e17313124), [`4bb5adc`](https://github.com/mastra-ai/mastra/commit/4bb5adc05c88e3a83fe1ea5ecb9eae6e17313124)]:
|
|
30
|
+
- @mastra/core@1.16.0-alpha.3
|
|
31
|
+
- @mastra/mcp@1.3.1
|
|
32
|
+
|
|
3
33
|
## 1.1.16-alpha.6
|
|
4
34
|
|
|
5
35
|
### Patch Changes
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.16
|
|
3
|
+
"version": "1.1.16",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -29,7 +29,7 @@
|
|
|
29
29
|
"jsdom": "^26.1.0",
|
|
30
30
|
"local-pkg": "^1.1.2",
|
|
31
31
|
"zod": "^4.3.6",
|
|
32
|
-
"@mastra/core": "1.16.0
|
|
32
|
+
"@mastra/core": "1.16.0",
|
|
33
33
|
"@mastra/mcp": "^1.3.1"
|
|
34
34
|
},
|
|
35
35
|
"devDependencies": {
|
|
@@ -46,9 +46,9 @@
|
|
|
46
46
|
"tsx": "^4.21.0",
|
|
47
47
|
"typescript": "^5.9.3",
|
|
48
48
|
"vitest": "4.0.18",
|
|
49
|
-
"@
|
|
50
|
-
"@internal/types-builder": "0.0.
|
|
51
|
-
"@
|
|
49
|
+
"@internal/lint": "0.0.74",
|
|
50
|
+
"@internal/types-builder": "0.0.49",
|
|
51
|
+
"@mastra/core": "1.16.0"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|