@mastra/mcp-docs-server 1.1.35-alpha.2 → 1.1.35-alpha.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/guides/deployment/inngest.md +6 -8
- package/.docs/models/gateways/netlify.md +2 -1
- package/.docs/models/gateways/openrouter.md +2 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/deepinfra.md +2 -1
- package/.docs/models/providers/digitalocean.md +2 -1
- package/package.json +3 -3
|
@@ -21,27 +21,29 @@ Install the required packages:
|
|
|
21
21
|
**npm**:
|
|
22
22
|
|
|
23
23
|
```bash
|
|
24
|
-
npm install @mastra/inngest@latest inngest
|
|
24
|
+
npm install @mastra/inngest@latest inngest
|
|
25
25
|
```
|
|
26
26
|
|
|
27
27
|
**pnpm**:
|
|
28
28
|
|
|
29
29
|
```bash
|
|
30
|
-
pnpm add @mastra/inngest@latest inngest
|
|
30
|
+
pnpm add @mastra/inngest@latest inngest
|
|
31
31
|
```
|
|
32
32
|
|
|
33
33
|
**Yarn**:
|
|
34
34
|
|
|
35
35
|
```bash
|
|
36
|
-
yarn add @mastra/inngest@latest inngest
|
|
36
|
+
yarn add @mastra/inngest@latest inngest
|
|
37
37
|
```
|
|
38
38
|
|
|
39
39
|
**Bun**:
|
|
40
40
|
|
|
41
41
|
```bash
|
|
42
|
-
bun add @mastra/inngest@latest inngest
|
|
42
|
+
bun add @mastra/inngest@latest inngest
|
|
43
43
|
```
|
|
44
44
|
|
|
45
|
+
> **Note:** Requires `inngest@^4` and Inngest Dev Server `v1.18.0` or later. Realtime is built into the SDK in v4, so `@inngest/realtime` and `realtimeMiddleware` are no longer used.
|
|
46
|
+
|
|
45
47
|
## Building an Inngest workflow
|
|
46
48
|
|
|
47
49
|
This guide walks through creating a workflow with Inngest and Mastra, demonstrating a counter application that increments a value until it reaches 10.
|
|
@@ -54,13 +56,11 @@ In development:
|
|
|
54
56
|
|
|
55
57
|
```ts
|
|
56
58
|
import { Inngest } from 'inngest'
|
|
57
|
-
import { realtimeMiddleware } from '@inngest/realtime/middleware'
|
|
58
59
|
|
|
59
60
|
export const inngest = new Inngest({
|
|
60
61
|
id: 'mastra',
|
|
61
62
|
baseUrl: 'http://localhost:8288',
|
|
62
63
|
isDev: true,
|
|
63
|
-
middleware: [realtimeMiddleware()],
|
|
64
64
|
})
|
|
65
65
|
```
|
|
66
66
|
|
|
@@ -68,11 +68,9 @@ In production:
|
|
|
68
68
|
|
|
69
69
|
```ts
|
|
70
70
|
import { Inngest } from 'inngest'
|
|
71
|
-
import { realtimeMiddleware } from '@inngest/realtime/middleware'
|
|
72
71
|
|
|
73
72
|
export const inngest = new Inngest({
|
|
74
73
|
id: 'mastra',
|
|
75
|
-
middleware: [realtimeMiddleware()],
|
|
76
74
|
})
|
|
77
75
|
```
|
|
78
76
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Netlify
|
|
2
2
|
|
|
3
|
-
Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access
|
|
3
|
+
Netlify AI Gateway provides unified access to multiple providers with built-in caching and observability. Access 67 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Netlify documentation](https://docs.netlify.com/build/ai-gateway/overview/).
|
|
6
6
|
|
|
@@ -62,6 +62,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
62
62
|
| `gemini/gemini-3.1-pro-preview-customtools` |
|
|
63
63
|
| `gemini/gemini-flash-latest` |
|
|
64
64
|
| `gemini/gemini-flash-lite-latest` |
|
|
65
|
+
| `openai/chat-latest` |
|
|
65
66
|
| `openai/gpt-4.1` |
|
|
66
67
|
| `openai/gpt-4.1-mini` |
|
|
67
68
|
| `openai/gpt-4.1-nano` |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenRouter
|
|
2
2
|
|
|
3
|
-
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 186 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
|
|
6
6
|
|
|
@@ -172,6 +172,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
172
172
|
| `poolside/laguna-xs.2:free` |
|
|
173
173
|
| `prime-intellect/intellect-3` |
|
|
174
174
|
| `qwen/qwen-2.5-coder-32b-instruct` |
|
|
175
|
+
| `qwen/qwen-3.6-27b` |
|
|
175
176
|
| `qwen/qwen2.5-vl-72b-instruct` |
|
|
176
177
|
| `qwen/qwen3-235b-a22b-07-25` |
|
|
177
178
|
| `qwen/qwen3-235b-a22b-thinking-2507` |
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3879 models from 107 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Deep Infra
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 36 Deep Infra models through Mastra's model router. Authentication is handled automatically using the `DEEPINFRA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Deep Infra documentation](https://deepinfra.com/models).
|
|
6
6
|
|
|
@@ -36,6 +36,7 @@ for await (const chunk of stream) {
|
|
|
36
36
|
| `deepinfra/anthropic/claude-4-opus` | 200K | | | | | | $17 | $83 |
|
|
37
37
|
| `deepinfra/deepseek-ai/DeepSeek-R1-0528` | 164K | | | | | | $0.50 | $2 |
|
|
38
38
|
| `deepinfra/deepseek-ai/DeepSeek-V3.2` | 164K | | | | | | $0.26 | $0.38 |
|
|
39
|
+
| `deepinfra/deepseek-ai/DeepSeek-V4-Flash` | 1.0M | | | | | | $0.14 | $0.28 |
|
|
39
40
|
| `deepinfra/deepseek-ai/DeepSeek-V4-Pro` | 66K | | | | | | $2 | $3 |
|
|
40
41
|
| `deepinfra/google/gemma-4-26B-A4B-it` | 256K | | | | | | $0.07 | $0.34 |
|
|
41
42
|
| `deepinfra/google/gemma-4-31B-it` | 256K | | | | | | $0.13 | $0.38 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# DigitalOcean
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 64 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [DigitalOcean documentation](https://docs.digitalocean.com/products/gradient-ai-platform/details/models/).
|
|
6
6
|
|
|
@@ -60,6 +60,7 @@ for await (const chunk of stream) {
|
|
|
60
60
|
| `digitalocean/glm-5` | 203K | | | | | | $1 | $3 |
|
|
61
61
|
| `digitalocean/gte-large-en-v1.5` | 8K | | | | | | $0.09 | — |
|
|
62
62
|
| `digitalocean/kimi-k2.5` | 262K | | | | | | $0.50 | $3 |
|
|
63
|
+
| `digitalocean/kimi-k2.6` | 262K | | | | | | $0.95 | $4 |
|
|
63
64
|
| `digitalocean/llama-4-maverick` | 1.0M | | | | | | $0.25 | $0.87 |
|
|
64
65
|
| `digitalocean/llama-guard-4-12b` | 128K | | | | | | — | — |
|
|
65
66
|
| `digitalocean/llama3.3-70b-instruct` | 128K | | | | | | $0.65 | $0.65 |
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@mastra/mcp-docs-server",
|
|
3
|
-
"version": "1.1.35-alpha.
|
|
3
|
+
"version": "1.1.35-alpha.3",
|
|
4
4
|
"description": "MCP server for accessing Mastra.ai documentation, changelogs, and news.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "dist/index.js",
|
|
@@ -46,9 +46,9 @@
|
|
|
46
46
|
"tsx": "^4.21.0",
|
|
47
47
|
"typescript": "^6.0.3",
|
|
48
48
|
"vitest": "4.1.5",
|
|
49
|
-
"@internal/lint": "0.0.92",
|
|
50
49
|
"@internal/types-builder": "0.0.67",
|
|
51
|
-
"@mastra/core": "1.33.0-alpha.0"
|
|
50
|
+
"@mastra/core": "1.33.0-alpha.0",
|
|
51
|
+
"@internal/lint": "0.0.92"
|
|
52
52
|
},
|
|
53
53
|
"homepage": "https://mastra.ai",
|
|
54
54
|
"repository": {
|