@mastra/mcp-docs-server 1.1.32-alpha.2 → 1.1.32-alpha.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -32,7 +32,7 @@ bun add @mastra/mcp @mastra/core tsup
32
32
 
33
33
  ## Setting up an MCP server
34
34
 
35
- 1. Create a file for your stdio server, for example, `/src/mastra/stdio.ts`.
35
+ 1. Create a file for your stdio server, for example, `src/mastra/stdio.ts`.
36
36
 
37
37
  2. Add the following code to the file. Remember to import your actual Mastra tools and name the server appropriately.
38
38
 
@@ -68,10 +68,30 @@ bun add @mastra/mcp @mastra/core tsup
68
68
 
69
69
  4. Run the build command:
70
70
 
71
+ **npm**:
72
+
73
+ ```bash
74
+ npm run build:mcp
75
+ ```
76
+
77
+ **pnpm**:
78
+
71
79
  ```bash
72
80
  pnpm run build:mcp
73
81
  ```
74
82
 
83
+ **Yarn**:
84
+
85
+ ```bash
86
+ yarn build:mcp
87
+ ```
88
+
89
+ **Bun**:
90
+
91
+ ```bash
92
+ bun run build:mcp
93
+ ```
94
+
75
95
  This will compile your server code into both ESM and CJS formats and make the ESM output file executable. On Unix-like systems, the `chmod +x` step makes the file directly executable. Windows users may need to use WSL or handle execution through Node.js directly.
76
96
 
77
97
  ## Publishing to NPM
@@ -1,6 +1,6 @@
1
1
  # Model Providers
2
2
 
3
- Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3779 models from 106 providers through a single API.
3
+ Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3803 models from 106 providers through a single API.
4
4
 
5
5
  ## Features
6
6
 
@@ -1,6 +1,6 @@
1
1
  # ![DigitalOcean logo](https://models.dev/logos/digitalocean.svg)DigitalOcean
2
2
 
3
- Access 46 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
3
+ Access 62 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
4
4
 
5
5
  Learn more in the [DigitalOcean documentation](https://docs.digitalocean.com/products/gradient-ai-platform/details/models/).
6
6
 
@@ -46,32 +46,43 @@ for await (const chunk of stream) {
46
46
  | `digitalocean/anthropic-claude-opus-4.7` | 1.0M | | | | | | $5 | $25 |
47
47
  | `digitalocean/anthropic-claude-sonnet-4` | 1.0M | | | | | | $3 | $15 |
48
48
  | `digitalocean/arcee-trinity-large-thinking` | 256K | | | | | | $0.25 | $0.90 |
49
+ | `digitalocean/bge-m3` | 8K | | | | | | $0.02 | — |
50
+ | `digitalocean/bge-reranker-v2-m3` | 8K | | | | | | $0.01 | — |
51
+ | `digitalocean/deepseek-3.2` | 128K | | | | | | $0.50 | $2 |
49
52
  | `digitalocean/deepseek-r1-distill-llama-70b` | 131K | | | | | | $0.99 | $0.99 |
53
+ | `digitalocean/e5-large-v2` | 512 | | | | | | $0.02 | — |
50
54
  | `digitalocean/fal-ai/elevenlabs/tts/multilingual-v2` | — | | | | | | — | — |
51
55
  | `digitalocean/fal-ai/fast-sdxl` | — | | | | | | — | — |
52
56
  | `digitalocean/fal-ai/flux/schnell` | — | | | | | | — | — |
53
57
  | `digitalocean/fal-ai/stable-audio-25/text-to-audio` | — | | | | | | — | — |
58
+ | `digitalocean/gemma-4-31B-it` | 256K | | | | | | $0.18 | $0.50 |
54
59
  | `digitalocean/glm-5` | 203K | | | | | | $1 | $3 |
55
60
  | `digitalocean/gte-large-en-v1.5` | 8K | | | | | | $0.09 | — |
56
61
  | `digitalocean/kimi-k2.5` | 262K | | | | | | $0.50 | $3 |
62
+ | `digitalocean/llama-4-maverick` | 1.0M | | | | | | $0.25 | $0.87 |
63
+ | `digitalocean/llama-guard-4-12b` | 128K | | | | | | — | — |
57
64
  | `digitalocean/llama3.3-70b-instruct` | 128K | | | | | | $0.65 | $0.65 |
58
65
  | `digitalocean/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
66
+ | `digitalocean/mistral-3-14B` | 262K | | | | | | $0.20 | $0.20 |
59
67
  | `digitalocean/multi-qa-mpnet-base-dot-v1` | 512 | | | | | | $0.01 | — |
68
+ | `digitalocean/nemotron-3-nano-omni` | 66K | | | | | | $0.50 | $0.90 |
69
+ | `digitalocean/nemotron-nano-12b-v2-vl` | 128K | | | | | | $0.20 | $0.60 |
60
70
  | `digitalocean/nvidia-nemotron-3-super-120b` | 256K | | | | | | $0.30 | $0.65 |
61
71
  | `digitalocean/openai-gpt-4.1` | 1.0M | | | | | | $2 | $8 |
62
72
  | `digitalocean/openai-gpt-4o` | 128K | | | | | | $3 | $10 |
63
73
  | `digitalocean/openai-gpt-4o-mini` | 128K | | | | | | $0.15 | $0.60 |
64
74
  | `digitalocean/openai-gpt-5` | 400K | | | | | | $1 | $10 |
65
- | `digitalocean/openai-gpt-5-2-pro` | 400K | | | | | | $21 | $168 |
66
75
  | `digitalocean/openai-gpt-5-mini` | 400K | | | | | | $0.25 | $2 |
67
76
  | `digitalocean/openai-gpt-5-nano` | 400K | | | | | | $0.05 | $0.40 |
68
77
  | `digitalocean/openai-gpt-5.1-codex-max` | 400K | | | | | | $1 | $10 |
69
78
  | `digitalocean/openai-gpt-5.2` | 400K | | | | | | $2 | $14 |
79
+ | `digitalocean/openai-gpt-5.2-pro` | 400K | | | | | | $21 | $168 |
70
80
  | `digitalocean/openai-gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
71
81
  | `digitalocean/openai-gpt-5.4` | 1.0M | | | | | | $3 | $15 |
72
82
  | `digitalocean/openai-gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
73
83
  | `digitalocean/openai-gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
74
84
  | `digitalocean/openai-gpt-5.4-pro` | 400K | | | | | | $30 | $180 |
85
+ | `digitalocean/openai-gpt-5.5` | 1.0M | | | | | | $5 | $30 |
75
86
  | `digitalocean/openai-gpt-image-1` | — | | | | | | $5 | $40 |
76
87
  | `digitalocean/openai-gpt-image-1.5` | — | | | | | | $5 | $10 |
77
88
  | `digitalocean/openai-gpt-oss-120b` | 131K | | | | | | $0.10 | $0.70 |
@@ -79,7 +90,12 @@ for await (const chunk of stream) {
79
90
  | `digitalocean/openai-o1` | 200K | | | | | | $15 | $60 |
80
91
  | `digitalocean/openai-o3` | 200K | | | | | | $2 | $8 |
81
92
  | `digitalocean/openai-o3-mini` | 200K | | | | | | $1 | $4 |
93
+ | `digitalocean/qwen3-coder-flash` | 262K | | | | | | $0.45 | $2 |
82
94
  | `digitalocean/qwen3-embedding-0.6b` | 8K | | | | | | $0.04 | — |
95
+ | `digitalocean/qwen3-tts-voicedesign` | 33K | | | | | | — | — |
96
+ | `digitalocean/qwen3.5-397b-a17b` | 262K | | | | | | $0.55 | $4 |
97
+ | `digitalocean/stable-diffusion-3.5-large` | 256 | | | | | | $0.08 | — |
98
+ | `digitalocean/wan2-2-t2v-a14b` | 100 | | | | | | $0.60 | — |
83
99
 
84
100
  ## Advanced configuration
85
101
 
@@ -109,7 +125,7 @@ const agent = new Agent({
109
125
  model: ({ requestContext }) => {
110
126
  const useAdvanced = requestContext.task === "complex";
111
127
  return useAdvanced
112
- ? "digitalocean/qwen3-embedding-0.6b"
128
+ ? "digitalocean/wan2-2-t2v-a14b"
113
129
  : "digitalocean/alibaba-qwen3-32b";
114
130
  }
115
131
  });
@@ -1,6 +1,6 @@
1
1
  # ![LLM Gateway logo](https://models.dev/logos/llmgateway.svg)LLM Gateway
2
2
 
3
- Access 184 LLM Gateway models through Mastra's model router. Authentication is handled automatically using the `LLMGATEWAY_API_KEY` environment variable.
3
+ Access 189 LLM Gateway models through Mastra's model router. Authentication is handled automatically using the `LLMGATEWAY_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [LLM Gateway documentation](https://llmgateway.io/docs).
6
6
 
@@ -120,6 +120,8 @@ for await (const chunk of stream) {
120
120
  | `llmgateway/gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
121
121
  | `llmgateway/gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
122
122
  | `llmgateway/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
123
+ | `llmgateway/gpt-5.5` | 1.1M | | | | | | $5 | $30 |
124
+ | `llmgateway/gpt-5.5-pro` | 1.1M | | | | | | $30 | $180 |
123
125
  | `llmgateway/gpt-oss-120b` | 131K | | | | | | $0.15 | $0.75 |
124
126
  | `llmgateway/gpt-oss-20b` | 131K | | | | | | $0.10 | $0.50 |
125
127
  | `llmgateway/grok-3` | 131K | | | | | | $3 | $15 |
@@ -209,6 +211,9 @@ for await (const chunk of stream) {
209
211
  | `llmgateway/qwen3-vl-8b-instruct` | 131K | | | | | | $0.10 | $0.10 |
210
212
  | `llmgateway/qwen3-vl-flash` | 1.0M | | | | | | $0.05 | $0.40 |
211
213
  | `llmgateway/qwen3-vl-plus` | 262K | | | | | | $0.20 | $2 |
214
+ | `llmgateway/qwen3.6-35b-a3b` | 262K | | | | | | $0.25 | $1 |
215
+ | `llmgateway/qwen3.6-max-preview` | 262K | | | | | | $1 | $8 |
216
+ | `llmgateway/qwen3.6-plus` | 1.0M | | | | | | $0.28 | $2 |
212
217
  | `llmgateway/qwen35-397b-a17b` | 262K | | | | | | $0.60 | $4 |
213
218
  | `llmgateway/qwq-plus` | 131K | | | | | | $0.80 | $2 |
214
219
  | `llmgateway/seed-1-6-250615` | 256K | | | | | | $0.25 | $2 |
@@ -1,6 +1,6 @@
1
1
  # ![Mistral logo](https://models.dev/logos/mistral.svg)Mistral
2
2
 
3
- Access 27 Mistral models through Mastra's model router. Authentication is handled automatically using the `MISTRAL_API_KEY` environment variable.
3
+ Access 28 Mistral models through Mastra's model router. Authentication is handled automatically using the `MISTRAL_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Mistral documentation](https://docs.mistral.ai/getting-started/models/).
6
6
 
@@ -49,7 +49,8 @@ for await (const chunk of stream) {
49
49
  | `mistral/mistral-large-latest` | 262K | | | | | | $0.50 | $2 |
50
50
  | `mistral/mistral-medium-2505` | 131K | | | | | | $0.40 | $2 |
51
51
  | `mistral/mistral-medium-2508` | 262K | | | | | | $0.40 | $2 |
52
- | `mistral/mistral-medium-latest` | 128K | | | | | | $0.40 | $2 |
52
+ | `mistral/mistral-medium-2604` | 262K | | | | | | $2 | $8 |
53
+ | `mistral/mistral-medium-latest` | 262K | | | | | | $2 | $8 |
53
54
  | `mistral/mistral-nemo` | 128K | | | | | | $0.15 | $0.15 |
54
55
  | `mistral/mistral-small-2506` | 128K | | | | | | $0.10 | $0.30 |
55
56
  | `mistral/mistral-small-2603` | 256K | | | | | | $0.15 | $0.60 |
@@ -1,6 +1,6 @@
1
1
  # ![Nvidia logo](https://models.dev/logos/nvidia.svg)Nvidia
2
2
 
3
- Access 80 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
3
+ Access 81 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Nvidia documentation](https://docs.api.nvidia.com/nim/).
6
6
 
@@ -80,6 +80,7 @@ for await (const chunk of stream) {
80
80
  | `nvidia/mistralai/ministral-14b-instruct-2512` | 262K | | | | | | — | — |
81
81
  | `nvidia/mistralai/mistral-large-2-instruct` | 128K | | | | | | — | — |
82
82
  | `nvidia/mistralai/mistral-large-3-675b-instruct-2512` | 262K | | | | | | — | — |
83
+ | `nvidia/mistralai/mistral-medium-3.5-128b` | 262K | | | | | | — | — |
83
84
  | `nvidia/mistralai/mistral-small-3.1-24b-instruct-2503` | 128K | | | | | | — | — |
84
85
  | `nvidia/moonshotai/kimi-k2-instruct` | 128K | | | | | | — | — |
85
86
  | `nvidia/moonshotai/kimi-k2-instruct-0905` | 262K | | | | | | — | — |
@@ -1,6 +1,6 @@
1
1
  # ![Together AI logo](https://models.dev/logos/togetherai.svg)Together AI
2
2
 
3
- Access 17 Together AI models through Mastra's model router. Authentication is handled automatically using the `TOGETHER_API_KEY` environment variable.
3
+ Access 18 Together AI models through Mastra's model router. Authentication is handled automatically using the `TOGETHER_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Together AI documentation](https://docs.together.ai/docs/serverless-models).
6
6
 
@@ -48,6 +48,7 @@ for await (const chunk of stream) {
48
48
  | `togetherai/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8` | 262K | | | | | | $2 | $2 |
49
49
  | `togetherai/Qwen/Qwen3-Coder-Next-FP8` | 262K | | | | | | $0.50 | $1 |
50
50
  | `togetherai/Qwen/Qwen3.5-397B-A17B` | 262K | | | | | | $0.60 | $4 |
51
+ | `togetherai/Qwen/Qwen3.6-Plus` | 1.0M | | | | | | $0.50 | $3 |
51
52
  | `togetherai/zai-org/GLM-5.1` | 203K | | | | | | $1 | $4 |
52
53
 
53
54
  ## Advanced configuration
@@ -224,6 +224,7 @@ The Reference section provides documentation of Mastra's API, including paramete
224
224
  - [createVectorQueryTool()](https://mastra.ai/reference/tools/vector-query-tool)
225
225
  - [MCPClient](https://mastra.ai/reference/tools/mcp-client)
226
226
  - [MCPServer](https://mastra.ai/reference/tools/mcp-server)
227
+ - [Perplexity Tools](https://mastra.ai/reference/tools/perplexity)
227
228
  - [Tavily Tools](https://mastra.ai/reference/tools/tavily)
228
229
  - [Amazon S3 Vector Store](https://mastra.ai/reference/vectors/s3vectors)
229
230
  - [Astra Vector Store](https://mastra.ai/reference/vectors/astra)
@@ -768,11 +768,9 @@ await serverWithPrompts.prompts.notifyListChanged()
768
768
  - Handle errors with informative messages.
769
769
  - Document argument expectations and available versions.
770
770
 
771
- ***
772
-
773
771
  ## Examples
774
772
 
775
- For practical examples of setting up and deploying an MCPServer, see the [Publishing an MCP Server guide](https://mastra.ai/docs/mcp/publishing-mcp-server).
773
+ For practical examples of setting up and deploying an MCPServer, see the [Publishing an MCP Server guide](https://mastra.ai/guides/guide/publishing-mcp-server).
776
774
 
777
775
  The example at the beginning of this page also demonstrates how to instantiate `MCPServer` with both tools and agents.
778
776
 
@@ -0,0 +1,155 @@
1
+ # Perplexity tools
2
+
3
+ Added in: `@mastra/perplexity@0.1.0-alpha.0`
4
+
5
+ The `@mastra/perplexity` package wraps the [Perplexity Search API](https://docs.perplexity.ai/docs/search/quickstart) as a Mastra-compatible tool. It exposes a factory function that returns a tool created with [`createTool()`](https://mastra.ai/reference/tools/create-tool) and full Zod input/output schemas.
6
+
7
+ For chat completions or agentic workflows powered by Perplexity, use Mastra's built-in [Perplexity model provider](https://mastra.ai/models/providers/perplexity) or [Perplexity Agent provider](https://mastra.ai/models/providers/perplexity-agent). Those are separate from this Search tool.
8
+
9
+ ## Installation
10
+
11
+ Install the package alongside Zod:
12
+
13
+ **npm**:
14
+
15
+ ```sh
16
+ npm install @mastra/perplexity zod
17
+ ```
18
+
19
+ **pnpm**:
20
+
21
+ ```sh
22
+ pnpm add @mastra/perplexity zod
23
+ ```
24
+
25
+ **Yarn**:
26
+
27
+ ```sh
28
+ yarn add @mastra/perplexity zod
29
+ ```
30
+
31
+ **Bun**:
32
+
33
+ ```sh
34
+ bun add @mastra/perplexity zod
35
+ ```
36
+
37
+ ## Usage example
38
+
39
+ The following example creates the search tool with the default configuration. By default, the tool reads `PERPLEXITY_API_KEY` (with `PPLX_API_KEY` as a fallback) from the environment. Pass `{ apiKey }` explicitly to override.
40
+
41
+ ```typescript
42
+ import { createPerplexitySearchTool } from '@mastra/perplexity'
43
+
44
+ const searchTool = createPerplexitySearchTool()
45
+ ```
46
+
47
+ To pass an API key explicitly:
48
+
49
+ ```typescript
50
+ import { createPerplexitySearchTool } from '@mastra/perplexity'
51
+
52
+ const searchTool = createPerplexitySearchTool({ apiKey: 'pplx-...' })
53
+ ```
54
+
55
+ ## Configuration
56
+
57
+ All factory functions accept a `PerplexityClientOptions` object:
58
+
59
+ **apiKey** (`string`): Perplexity API key. Falls back to the \`PERPLEXITY\_API\_KEY\` then \`PPLX\_API\_KEY\` environment variables.
60
+
61
+ **baseUrl** (`string`): Override the API base URL. (Default: `'https://api.perplexity.ai'`)
62
+
63
+ **fetch** (`typeof fetch`): Custom \`fetch\` implementation. Useful for tests, retries, or instrumentation.
64
+
65
+ ## Methods
66
+
67
+ ### Factory functions
68
+
69
+ #### `createPerplexityTools(config?)`
70
+
71
+ Returns an object containing all Perplexity tools that share the supplied configuration.
72
+
73
+ ```typescript
74
+ import { createPerplexityTools } from '@mastra/perplexity'
75
+
76
+ const tools = createPerplexityTools({ apiKey: 'pplx-...' })
77
+ // tools.perplexitySearch
78
+ ```
79
+
80
+ Returns: `{ perplexitySearch }`
81
+
82
+ #### `createPerplexitySearchTool(config?)`
83
+
84
+ Creates a tool that searches the web using the Perplexity Search API. Returns ranked results with titles, URLs, snippets, and optional publication dates.
85
+
86
+ The tool is registered with the ID `perplexity-search`.
87
+
88
+ ```typescript
89
+ import { createPerplexitySearchTool } from '@mastra/perplexity'
90
+
91
+ const searchTool = createPerplexitySearchTool()
92
+ ```
93
+
94
+ ##### Input
95
+
96
+ **query** (`string`): The search query.
97
+
98
+ **maxResults** (`number`): Maximum number of results to return (1-20).
99
+
100
+ **searchDomainFilter** (`string[]`): Restrict (or exclude) results by domain. Prefix a domain with \`-\` to exclude it (for example, \`-pinterest.com\`). Do not mix allow- and deny-list entries in the same call.
101
+
102
+ **searchRecencyFilter** (`'hour' | 'day' | 'week' | 'month' | 'year'`): Only return results from within the given recency window.
103
+
104
+ **searchAfterDateFilter** (`string`): Only return results published on or after this date. Format: m/d/yyyy.
105
+
106
+ **searchBeforeDateFilter** (`string`): Only return results published on or before this date. Format: m/d/yyyy.
107
+
108
+ ##### Output
109
+
110
+ **query** (`string`): The original search query.
111
+
112
+ **results** (`SearchResult[]`): Array of search results.
113
+
114
+ **results.title** (`string`): Result title.
115
+
116
+ **results.url** (`string`): Result URL.
117
+
118
+ **results.snippet** (`string`): Content snippet.
119
+
120
+ **results.date** (`string`): Publication date when available.
121
+
122
+ ## Agent example
123
+
124
+ The following example registers the search tool on an agent so it can fetch fresh web results before answering.
125
+
126
+ ```typescript
127
+ import { Agent } from '@mastra/core/agent'
128
+ import { createPerplexitySearchTool } from '@mastra/perplexity'
129
+
130
+ const agent = new Agent({
131
+ id: 'research-agent',
132
+ name: 'Research Agent',
133
+ model: 'anthropic/claude-sonnet-4-6',
134
+ instructions:
135
+ 'You are a research assistant. Use the perplexity-search tool to find up-to-date information from the web before answering.',
136
+ tools: {
137
+ search: createPerplexitySearchTool(),
138
+ },
139
+ })
140
+ ```
141
+
142
+ ## Environment variables
143
+
144
+ | Variable | Description |
145
+ | -------------------- | ----------------------------------------------------------------------------------------------- |
146
+ | `PERPLEXITY_API_KEY` | Your Perplexity API key. Used as the default when `apiKey` is not passed to a factory function. |
147
+ | `PPLX_API_KEY` | Fallback used when `PERPLEXITY_API_KEY` is unset. |
148
+
149
+ ## Related
150
+
151
+ - [`createTool()`](https://mastra.ai/reference/tools/create-tool)
152
+ - [Perplexity Search quickstart](https://docs.perplexity.ai/docs/search/quickstart)
153
+ - [Perplexity Agent quickstart](https://docs.perplexity.ai/docs/agent/quickstart)
154
+ - [Perplexity model provider](https://mastra.ai/models/providers/perplexity)
155
+ - [Perplexity Agent provider](https://mastra.ai/models/providers/perplexity-agent)
package/CHANGELOG.md CHANGED
@@ -1,5 +1,19 @@
1
1
  # @mastra/mcp-docs-server
2
2
 
3
+ ## 1.1.32-alpha.4
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [[`b2deb29`](https://github.com/mastra-ai/mastra/commit/b2deb29412b300c868655b5840463614fbb7962d), [`66644be`](https://github.com/mastra-ai/mastra/commit/66644beac1aa560f0e417956ff007c89341dc382), [`310b953`](https://github.com/mastra-ai/mastra/commit/310b95345f302dcd5ba3ed862bdc96f059d44122), [`43f0e1d`](https://github.com/mastra-ai/mastra/commit/43f0e1d5d5a74ba6fc746f2ad89ebe0c64777a7d), [`da0b9e2`](https://github.com/mastra-ai/mastra/commit/da0b9e2ba7ecc560213b426d6c097fe63946086e)]:
8
+ - @mastra/core@1.31.0-alpha.3
9
+
10
+ ## 1.1.32-alpha.3
11
+
12
+ ### Patch Changes
13
+
14
+ - Updated dependencies [[`2b0f355`](https://github.com/mastra-ai/mastra/commit/2b0f3553be3e9e5524da539a66e5cf82668440a4)]:
15
+ - @mastra/core@1.31.0-alpha.2
16
+
3
17
  ## 1.1.32-alpha.2
4
18
 
5
19
  ### Patch Changes
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mastra/mcp-docs-server",
3
- "version": "1.1.32-alpha.2",
3
+ "version": "1.1.32-alpha.4",
4
4
  "description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
@@ -29,7 +29,7 @@
29
29
  "jsdom": "^26.1.0",
30
30
  "local-pkg": "^1.1.2",
31
31
  "zod": "^4.3.6",
32
- "@mastra/core": "1.31.0-alpha.1",
32
+ "@mastra/core": "1.31.0-alpha.3",
33
33
  "@mastra/mcp": "^1.6.0"
34
34
  },
35
35
  "devDependencies": {
@@ -48,7 +48,7 @@
48
48
  "vitest": "4.1.5",
49
49
  "@internal/lint": "0.0.89",
50
50
  "@internal/types-builder": "0.0.64",
51
- "@mastra/core": "1.31.0-alpha.1"
51
+ "@mastra/core": "1.31.0-alpha.3"
52
52
  },
53
53
  "homepage": "https://mastra.ai",
54
54
  "repository": {