@mastra/mcp-docs-server 1.1.13 → 1.1.14-alpha.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/deployment/studio.md +9 -24
- package/.docs/docs/getting-started/studio.md +28 -16
- package/.docs/docs/observability/tracing/exporters/braintrust.md +15 -0
- package/.docs/docs/server/auth.md +6 -7
- package/.docs/docs/server/custom-api-routes.md +56 -0
- package/.docs/docs/server/mastra-server.md +2 -2
- package/.docs/guides/deployment/cloudflare.md +1 -1
- package/.docs/models/gateways/openrouter.md +4 -1
- package/.docs/models/gateways/vercel.md +7 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/anthropic.md +2 -2
- package/.docs/models/providers/baseten.md +12 -13
- package/.docs/models/providers/chutes.md +5 -5
- package/.docs/models/providers/deepinfra.md +30 -23
- package/.docs/models/providers/google.md +1 -1
- package/.docs/models/providers/kilo.md +342 -272
- package/.docs/models/providers/nano-gpt.md +36 -36
- package/.docs/models/providers/nebius.md +3 -2
- package/.docs/models/providers/perplexity-agent.md +19 -18
- package/.docs/models/providers/synthetic.md +1 -1
- package/.docs/models/providers/vultr.md +17 -12
- package/.docs/models/providers/zai-coding-plan.md +3 -2
- package/.docs/models/providers/zai.md +3 -2
- package/.docs/reference/agents/generate.md +2 -0
- package/.docs/reference/agents/network.md +2 -0
- package/.docs/reference/ai-sdk/chat-route.md +4 -0
- package/.docs/reference/configuration.md +4 -2
- package/.docs/reference/deployer/cloudflare.md +12 -1
- package/.docs/reference/processors/unicode-normalizer.md +1 -1
- package/.docs/reference/streaming/agents/stream.md +2 -0
- package/.docs/reference/workflows/run-methods/restart.md +2 -0
- package/.docs/reference/workflows/run-methods/resume.md +2 -0
- package/.docs/reference/workflows/run-methods/start.md +2 -0
- package/.docs/reference/workflows/run-methods/timeTravel.md +2 -0
- package/CHANGELOG.md +15 -0
- package/dist/prompts/migration.d.ts.map +1 -1
- package/dist/stdio.js.map +1 -1
- package/package.json +8 -8
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
# Deploying
|
|
1
|
+
# Deploying Studio
|
|
2
2
|
|
|
3
3
|
[Studio](https://mastra.ai/docs/getting-started/studio) provides an interactive UI for building and testing your agents. It's a React-based Single Page Application (SPA) that runs in the browser and connects to a running [Mastra server](https://mastra.ai/docs/deployment/mastra-server).
|
|
4
4
|
|
|
@@ -213,35 +213,20 @@ Follow the example below to create a SPA using Vite.
|
|
|
213
213
|
MASTRA_STUDIO_BASE_PATH=
|
|
214
214
|
MASTRA_TELEMETRY_DISABLED=true
|
|
215
215
|
MASTRA_HIDE_CLOUD_CTA=false
|
|
216
|
+
MASTRA_TEMPLATES=false
|
|
216
217
|
MASTRA_CLOUD_API_ENDPOINT=
|
|
217
218
|
MASTRA_EXPERIMENTAL_FEATURES=false
|
|
218
219
|
MASTRA_REQUEST_CONTEXT_PRESETS=
|
|
219
220
|
```
|
|
220
221
|
|
|
221
|
-
|
|
222
|
+
````text
|
|
223
|
+
</StepItem>
|
|
222
224
|
|
|
223
|
-
|
|
225
|
+
<StepItem>
|
|
226
|
+
Run the build script to generate the static files in the `dist` folder:
|
|
224
227
|
|
|
225
|
-
```bash
|
|
228
|
+
```bash npm2yarn
|
|
226
229
|
npm run build
|
|
227
|
-
|
|
228
|
-
|
|
229
|
-
**pnpm**:
|
|
230
|
-
|
|
231
|
-
```bash
|
|
232
|
-
pnpm run build
|
|
233
|
-
```
|
|
234
|
-
|
|
235
|
-
**Yarn**:
|
|
236
|
-
|
|
237
|
-
```bash
|
|
238
|
-
yarn build
|
|
239
|
-
```
|
|
240
|
-
|
|
241
|
-
**Bun**:
|
|
242
|
-
|
|
243
|
-
```bash
|
|
244
|
-
bun run build
|
|
245
|
-
```
|
|
230
|
+
````
|
|
246
231
|
|
|
247
|
-
|
|
232
|
+
7. Point your hosting provider to the `dist` folder and deploy!
|
|
@@ -1,16 +1,14 @@
|
|
|
1
1
|
# Studio
|
|
2
2
|
|
|
3
|
-
Studio provides an interactive UI for building and
|
|
3
|
+
Studio provides an interactive UI for building, testing, and managing your agents, workflows, and tools. Run it locally during development, or [deploy it](https://mastra.ai/docs/deployment/studio) to production so your team can manage agents, monitor performance, and gain insights through built-in observability.
|
|
4
4
|
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
To get started, run Studio locally using the instructions below, or [create a project in Mastra Cloud](https://mastra.ai/docs/mastra-cloud/setup) to collaborate with your team.
|
|
5
|
+
Add [authentication](https://mastra.ai/docs/server/auth) to protect your deployed Studio with login screens, role-based access control, and permission-based UI rendering so you can control what each team member can see and do. You can also [create a project in Mastra Cloud](https://mastra.ai/docs/mastra-cloud/setup) for a hosted option.
|
|
8
6
|
|
|
9
7
|
[YouTube video player](https://www.youtube-nocookie.com/embed/ojGu6Bi4wYk)
|
|
10
8
|
|
|
11
9
|
## Start Studio
|
|
12
10
|
|
|
13
|
-
If you created your application with `create mastra`, start the
|
|
11
|
+
If you created your application with `create mastra`, start the development server using the `dev` script. You can also run it directly with `mastra dev`.
|
|
14
12
|
|
|
15
13
|
**npm**:
|
|
16
14
|
|
|
@@ -36,14 +34,16 @@ yarn dev
|
|
|
36
34
|
bun run dev
|
|
37
35
|
```
|
|
38
36
|
|
|
39
|
-
Once the server
|
|
37
|
+
Once the server is running, you can:
|
|
40
38
|
|
|
41
|
-
- Open the Studio UI at <http://localhost:4111/> to
|
|
39
|
+
- Open the Studio UI at <http://localhost:4111/> to interact with your agents, workflows, and tools.
|
|
42
40
|
- Visit <http://localhost:4111/swagger-ui> to discover and interact with the underlying REST API.
|
|
43
41
|
|
|
42
|
+
To run Studio in production, see [Deploy Studio](https://mastra.ai/docs/deployment/studio).
|
|
43
|
+
|
|
44
44
|
## Studio UI
|
|
45
45
|
|
|
46
|
-
The Studio UI
|
|
46
|
+
The Studio UI lets you interact with your agents, workflows, and tools, observe exactly what happens under the hood with each interaction, and tweak things as you go.
|
|
47
47
|
|
|
48
48
|
### Agents
|
|
49
49
|
|
|
@@ -61,39 +61,51 @@ When running a workflow, you can also view detailed traces showing tool calls, r
|
|
|
61
61
|
|
|
62
62
|
Run tools in isolation to observe their behavior. Test them before assigning them to your agent, or isolate them to debug issues should something go wrong.
|
|
63
63
|
|
|
64
|
+
### Processors
|
|
65
|
+
|
|
66
|
+
View the input and output processors attached to each agent. The agent detail panel lists every processor by name and type, so you can verify your guardrails, token limiters, and custom processors are wired up correctly before testing.
|
|
67
|
+
|
|
68
|
+
See [Processors](https://mastra.ai/docs/agents/processors) and [Guardrails](https://mastra.ai/docs/agents/guardrails) for configuration details.
|
|
69
|
+
|
|
64
70
|
### MCP
|
|
65
71
|
|
|
66
72
|
List the MCP servers attached to your Mastra instance and explore their available tools.
|
|
67
73
|
|
|
68
|
-

|
|
69
|
-
|
|
70
74
|
### Observability
|
|
71
75
|
|
|
72
76
|
When you run an agent or workflow, the Observability tab displays traces that highlight the key AI operations such as model calls, tool executions, and workflow steps. Follow these traces to see how data moves, where time is spent, and what's happening under the hood.
|
|
73
77
|
|
|
74
|
-

|
|
75
|
-
|
|
76
78
|
Tracing filters out low-level framework details so your traces stay focused and readable.
|
|
77
79
|
|
|
78
80
|
### Scorers
|
|
79
81
|
|
|
80
82
|
The Scorers tab displays the results of your agent's scorers as they run. When messages pass through your agent, the defined scorers evaluate each output asynchronously and render their results here. This allows you to understand how your scorers respond to different interactions, compare performance across test cases, and identify areas for improvement.
|
|
81
83
|
|
|
84
|
+
### Datasets
|
|
85
|
+
|
|
86
|
+
Create and manage collections of test cases to evaluate your agents and workflows. Import items from CSV or JSON, define input and ground-truth schemas, and pin to specific versions so you can reproduce experiments exactly. Run experiments with [scorers](https://mastra.ai/docs/evals/overview) to compare quality across prompts, models, or code changes.
|
|
87
|
+
|
|
88
|
+
See [Datasets overview](https://mastra.ai/docs/observability/datasets/overview) for the full API and versioning details.
|
|
89
|
+
|
|
82
90
|
## REST API
|
|
83
91
|
|
|
84
|
-
|
|
92
|
+
Studio is backed by a complete set of REST API routes that let you programmatically interact with your agents, workflows, and tools. These are the same routes exposed by the [Mastra Server](https://mastra.ai/docs/server/mastra-server), so everything you build against locally works identically in production.
|
|
85
93
|
|
|
86
94
|
You can explore all available endpoints in the OpenAPI specification at <http://localhost:4111/api/openapi.json>, which details every endpoint and its request and response schemas.
|
|
87
95
|
|
|
88
96
|
To explore the API interactively, visit the Swagger UI at <http://localhost:4111/swagger-ui>. Here, you can discover endpoints and test them directly from your browser.
|
|
89
97
|
|
|
90
|
-
> **
|
|
98
|
+
> **Note:** The OpenAPI and Swagger endpoints are disabled in production by default. To enable them, set [`server.build.openAPIDocs`](https://mastra.ai/reference/configuration) and [`server.build.swaggerUI`](https://mastra.ai/reference/configuration) to `true` respectively.
|
|
91
99
|
|
|
92
100
|
## Configuration
|
|
93
101
|
|
|
94
|
-
By default, Studio runs at <http://localhost:4111>. You can change the [`host`](https://mastra.ai/reference/configuration), [`port`](https://mastra.ai/reference/configuration), and [`studioBase`](https://mastra.ai/reference/configuration) in the Mastra server configuration.
|
|
102
|
+
By default, Studio runs at <http://localhost:4111>. You can change the [`host`](https://mastra.ai/reference/configuration), [`port`](https://mastra.ai/reference/configuration), and [`studioBase`](https://mastra.ai/reference/configuration) in the Mastra server configuration.
|
|
103
|
+
|
|
104
|
+
For production deployments, see [Deploy Studio](https://mastra.ai/docs/deployment/studio) to learn about hosting Studio alongside your server, as a standalone SPA, or on a CDN.
|
|
105
|
+
|
|
106
|
+
Add [authentication](https://mastra.ai/docs/server/auth) to control who can access Studio in production. Studio displays the appropriate login UI, which can be an SSO button, an email/password form, or both. All API routes require authentication. This applies to any request made to your Mastra API, whether from Studio or a direct API call.
|
|
95
107
|
|
|
96
|
-
|
|
108
|
+
Mastra also supports HTTPS development through the [`--https`](https://mastra.ai/reference/cli/mastra) flag, which automatically creates and manages certificates for your project. When you run `mastra dev --https`, a private key and certificate are generated for localhost (or your configured host). Visit the [HTTPS reference](https://mastra.ai/reference/configuration) to learn more.
|
|
97
109
|
|
|
98
110
|
## Next steps
|
|
99
111
|
|
|
@@ -105,6 +105,21 @@ new BraintrustExporter({
|
|
|
105
105
|
})
|
|
106
106
|
```
|
|
107
107
|
|
|
108
|
+
## Querying Braintrust with returned `spanId`
|
|
109
|
+
|
|
110
|
+
For Braintrust, use `spanId` as the root span identifier when searching for traces because Braintrust root-span queries are typically faster than trace-id queries.
|
|
111
|
+
|
|
112
|
+
```typescript
|
|
113
|
+
const result = await agent.stream('Summarize this ticket')
|
|
114
|
+
|
|
115
|
+
console.log('Mastra trace ID:', result.traceId)
|
|
116
|
+
console.log('Braintrust root span ID:', result.spanId)
|
|
117
|
+
|
|
118
|
+
// Use result.spanId in your Braintrust lookup/query path
|
|
119
|
+
```
|
|
120
|
+
|
|
121
|
+
The same applies to `agent.generate()` and workflow run results (`run.start()`, `run.stream()` final state, `run.resume()`).
|
|
122
|
+
|
|
108
123
|
## Related
|
|
109
124
|
|
|
110
125
|
- [Tracing Overview](https://mastra.ai/docs/observability/tracing/overview)
|
|
@@ -1,18 +1,17 @@
|
|
|
1
1
|
# Auth overview
|
|
2
2
|
|
|
3
|
-
Mastra lets you choose how you handle authentication, so you can secure access to your
|
|
3
|
+
Mastra lets you choose how you handle authentication, so you can secure access to your API and [Studio](https://mastra.ai/docs/getting-started/studio) using the identity system that fits your stack.
|
|
4
4
|
|
|
5
5
|
You can start with basic shared secret JWT authentication and switch to providers like Supabase, Firebase Auth, Auth0, Clerk, or WorkOS when you need more advanced identity features.
|
|
6
6
|
|
|
7
|
-
##
|
|
7
|
+
## What auth secures
|
|
8
8
|
|
|
9
|
-
|
|
9
|
+
Configuring authentication locks down two things at once:
|
|
10
10
|
|
|
11
|
-
- **
|
|
12
|
-
- **
|
|
13
|
-
- **Public access** can be enabled on custom routes using `requiresAuth: false`
|
|
11
|
+
- **Studio UI**: Studio displays a login screen (SSO, email/password, or both) and enforces role-based access control to determine what each team member can see and do.
|
|
12
|
+
- **API routes**: All built-in routes (`/api/agents/*`, `/api/workflows/*`, etc.) and custom routes require authentication, whether requests come from Studio or direct API calls.
|
|
14
13
|
|
|
15
|
-
If no
|
|
14
|
+
Authentication is optional. If no auth is configured, all routes and Studio are publicly accessible. Public access can be enabled on individual custom routes using `requiresAuth: false`.
|
|
16
15
|
|
|
17
16
|
See [Custom API Routes](https://mastra.ai/docs/server/custom-api-routes) for controlling authentication on custom endpoints.
|
|
18
17
|
|
|
@@ -260,6 +260,62 @@ registerApiRoute('/user-profile', {
|
|
|
260
260
|
|
|
261
261
|
For more information about authentication providers, see the [Auth documentation](https://mastra.ai/docs/server/auth).
|
|
262
262
|
|
|
263
|
+
## Continue generation after client disconnect
|
|
264
|
+
|
|
265
|
+
Built-in streaming helpers such as [`chatRoute()`](https://mastra.ai/reference/ai-sdk/chat-route) forward the incoming request's `AbortSignal` to `agent.stream()`. That is the right default when a browser disconnect should cancel the model call.
|
|
266
|
+
|
|
267
|
+
If you want the server to keep generating and persist the final response even after the client disconnects, build a custom route around the underlying `MastraModelOutput`. Start the agent stream without forwarding `c.req.raw.signal`, then call `consumeStream()` in the background so generation continues server-side.
|
|
268
|
+
|
|
269
|
+
```typescript
|
|
270
|
+
import {
|
|
271
|
+
createUIMessageStream,
|
|
272
|
+
createUIMessageStreamResponse,
|
|
273
|
+
InferUIMessageChunk,
|
|
274
|
+
UIMessage,
|
|
275
|
+
} from 'ai'
|
|
276
|
+
import { toAISdkStream } from '@mastra/ai-sdk'
|
|
277
|
+
import { Mastra } from '@mastra/core'
|
|
278
|
+
import { registerApiRoute } from '@mastra/core/server'
|
|
279
|
+
|
|
280
|
+
export const mastra = new Mastra({
|
|
281
|
+
server: {
|
|
282
|
+
apiRoutes: [
|
|
283
|
+
registerApiRoute('/chat/persist/:agentId', {
|
|
284
|
+
method: 'POST',
|
|
285
|
+
handler: async c => {
|
|
286
|
+
const { messages, memory } = await c.req.json()
|
|
287
|
+
const mastra = c.get('mastra')
|
|
288
|
+
const agent = mastra.getAgent(c.req.param('agentId'))
|
|
289
|
+
|
|
290
|
+
const stream = await agent.stream(messages, {
|
|
291
|
+
memory,
|
|
292
|
+
// Do not pass c.req.raw.signal if this route should keep running
|
|
293
|
+
// after the client disconnects.
|
|
294
|
+
})
|
|
295
|
+
|
|
296
|
+
void stream.consumeStream().catch(error => {
|
|
297
|
+
mastra.getLogger()?.error('Background stream consumption failed', { error })
|
|
298
|
+
})
|
|
299
|
+
|
|
300
|
+
const uiStream = createUIMessageStream({
|
|
301
|
+
originalMessages: messages,
|
|
302
|
+
execute: async ({ writer }) => {
|
|
303
|
+
for await (const part of toAISdkStream(stream, { from: 'agent' })) {
|
|
304
|
+
writer.write(part as InferUIMessageChunk<UIMessage>)
|
|
305
|
+
}
|
|
306
|
+
},
|
|
307
|
+
})
|
|
308
|
+
|
|
309
|
+
return createUIMessageStreamResponse({ stream: uiStream })
|
|
310
|
+
},
|
|
311
|
+
}),
|
|
312
|
+
],
|
|
313
|
+
},
|
|
314
|
+
})
|
|
315
|
+
```
|
|
316
|
+
|
|
317
|
+
> **Note:** Use this pattern only when you intentionally want work to continue after the HTTP client is gone. If you want disconnects to cancel generation, keep using `chatRoute()` or forward the request `AbortSignal` yourself.
|
|
318
|
+
|
|
263
319
|
## Related
|
|
264
320
|
|
|
265
321
|
- [registerApiRoute() Reference](https://mastra.ai/reference/server/register-api-route) - Full API reference
|
|
@@ -25,8 +25,8 @@ import { Mastra } from '@mastra/core'
|
|
|
25
25
|
|
|
26
26
|
export const mastra = new Mastra({
|
|
27
27
|
server: {
|
|
28
|
-
port: 3000, // Defaults to 4111
|
|
29
|
-
host: '0.0.0.0', // Defaults to 'localhost'
|
|
28
|
+
port: 3000, // Defaults to PORT env var or 4111
|
|
29
|
+
host: '0.0.0.0', // Defaults to MASTRA_HOST env var or 'localhost'
|
|
30
30
|
},
|
|
31
31
|
})
|
|
32
32
|
```
|
|
@@ -86,7 +86,7 @@ After setting up your project, push it to your remote Git provider of choice (e.
|
|
|
86
86
|
|
|
87
87
|
1. Connect your repository to Cloudflare. On the "Workers & Pages" dashboard, select **Create application** and choose your Git provider in the next step. Continue with the setup process and select the repository you want to deploy.
|
|
88
88
|
|
|
89
|
-
> **Note:** Remember to set your environment variables needed to run your application (e.g. your [model provider](https://mastra.ai/models/providers) API key).
|
|
89
|
+
> **Note:** Remember to set your environment variables needed to run your application (e.g. your [model provider](https://mastra.ai/models/providers) API key). You can upload secrets from your `.env` file using `npx wrangler secret bulk .env`. See [Secrets](https://mastra.ai/reference/deployer/cloudflare) for details.
|
|
90
90
|
|
|
91
91
|
2. Once you're ready, click the **Deploy** button and wait for the first deployment to complete.
|
|
92
92
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenRouter
|
|
2
2
|
|
|
3
|
-
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 198 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
|
|
6
6
|
|
|
@@ -79,6 +79,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
79
79
|
| `google/gemini-2.5-pro-preview-06-05` |
|
|
80
80
|
| `google/gemini-3-flash-preview` |
|
|
81
81
|
| `google/gemini-3-pro-preview` |
|
|
82
|
+
| `google/gemini-3.1-flash-lite-preview` |
|
|
82
83
|
| `google/gemini-3.1-pro-preview` |
|
|
83
84
|
| `google/gemini-3.1-pro-preview-customtools` |
|
|
84
85
|
| `google/gemma-2-9b-it` |
|
|
@@ -218,6 +219,8 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
218
219
|
| `x-ai/grok-4` |
|
|
219
220
|
| `x-ai/grok-4-fast` |
|
|
220
221
|
| `x-ai/grok-4.1-fast` |
|
|
222
|
+
| `x-ai/grok-4.20-beta` |
|
|
223
|
+
| `x-ai/grok-4.20-multi-agent-beta` |
|
|
221
224
|
| `x-ai/grok-code-fast-1` |
|
|
222
225
|
| `xiaomi/mimo-v2-flash` |
|
|
223
226
|
| `z-ai/glm-4.5` |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Vercel
|
|
2
2
|
|
|
3
|
-
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access
|
|
3
|
+
Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 214 models through Mastra's model router.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
|
|
6
6
|
|
|
@@ -118,6 +118,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
118
118
|
| `kwaipilot/kat-coder-pro-v1` |
|
|
119
119
|
| `meituan/longcat-flash-chat` |
|
|
120
120
|
| `meituan/longcat-flash-thinking` |
|
|
121
|
+
| `meituan/longcat-flash-thinking-2601` |
|
|
121
122
|
| `meta/llama-3.1-70b` |
|
|
122
123
|
| `meta/llama-3.1-8b` |
|
|
123
124
|
| `meta/llama-3.2-11b` |
|
|
@@ -131,6 +132,7 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
131
132
|
| `minimax/minimax-m2.1` |
|
|
132
133
|
| `minimax/minimax-m2.1-lightning` |
|
|
133
134
|
| `minimax/minimax-m2.5` |
|
|
135
|
+
| `minimax/minimax-m2.5-highspeed` |
|
|
134
136
|
| `mistral/codestral` |
|
|
135
137
|
| `mistral/codestral-embed` |
|
|
136
138
|
| `mistral/devstral-2` |
|
|
@@ -229,6 +231,9 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
229
231
|
| `xai/grok-4-fast-reasoning` |
|
|
230
232
|
| `xai/grok-4.1-fast-non-reasoning` |
|
|
231
233
|
| `xai/grok-4.1-fast-reasoning` |
|
|
234
|
+
| `xai/grok-4.20-multi-agent-beta` |
|
|
235
|
+
| `xai/grok-4.20-non-reasoning-beta` |
|
|
236
|
+
| `xai/grok-4.20-reasoning-beta` |
|
|
232
237
|
| `xai/grok-code-fast-1` |
|
|
233
238
|
| `xai/grok-imagine-image` |
|
|
234
239
|
| `xai/grok-imagine-image-pro` |
|
|
@@ -240,5 +245,6 @@ ANTHROPIC_API_KEY=ant-...
|
|
|
240
245
|
| `zai/glm-4.6v` |
|
|
241
246
|
| `zai/glm-4.6v-flash` |
|
|
242
247
|
| `zai/glm-4.7` |
|
|
248
|
+
| `zai/glm-4.7-flash` |
|
|
243
249
|
| `zai/glm-4.7-flashx` |
|
|
244
250
|
| `zai/glm-5` |
|
package/.docs/models/index.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Model Providers
|
|
2
2
|
|
|
3
|
-
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to
|
|
3
|
+
Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3353 models from 92 providers through a single API.
|
|
4
4
|
|
|
5
5
|
## Features
|
|
6
6
|
|
|
@@ -49,12 +49,12 @@ for await (const chunk of stream) {
|
|
|
49
49
|
| `anthropic/claude-opus-4-20250514` | 200K | | | | | | $15 | $75 |
|
|
50
50
|
| `anthropic/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
|
|
51
51
|
| `anthropic/claude-opus-4-5-20251101` | 200K | | | | | | $5 | $25 |
|
|
52
|
-
| `anthropic/claude-opus-4-6` |
|
|
52
|
+
| `anthropic/claude-opus-4-6` | 1.0M | | | | | | $5 | $25 |
|
|
53
53
|
| `anthropic/claude-sonnet-4-0` | 200K | | | | | | $3 | $15 |
|
|
54
54
|
| `anthropic/claude-sonnet-4-20250514` | 200K | | | | | | $3 | $15 |
|
|
55
55
|
| `anthropic/claude-sonnet-4-5` | 200K | | | | | | $3 | $15 |
|
|
56
56
|
| `anthropic/claude-sonnet-4-5-20250929` | 200K | | | | | | $3 | $15 |
|
|
57
|
-
| `anthropic/claude-sonnet-4-6` |
|
|
57
|
+
| `anthropic/claude-sonnet-4-6` | 1.0M | | | | | | $3 | $15 |
|
|
58
58
|
|
|
59
59
|
## Advanced configuration
|
|
60
60
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Baseten
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 9 Baseten models through Mastra's model router. Authentication is handled automatically using the `BASETEN_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Baseten documentation](https://docs.baseten.co/development/model-apis/overview).
|
|
6
6
|
|
|
@@ -32,18 +32,17 @@ for await (const chunk of stream) {
|
|
|
32
32
|
|
|
33
33
|
## Models
|
|
34
34
|
|
|
35
|
-
| Model
|
|
36
|
-
|
|
|
37
|
-
| `baseten/deepseek-ai/DeepSeek-V3
|
|
38
|
-
| `baseten/
|
|
39
|
-
| `baseten/
|
|
40
|
-
| `baseten/moonshotai/Kimi-K2
|
|
41
|
-
| `baseten/
|
|
42
|
-
| `baseten/
|
|
43
|
-
| `baseten/
|
|
44
|
-
| `baseten/zai-org/GLM-4.
|
|
45
|
-
| `baseten/zai-org/GLM-
|
|
46
|
-
| `baseten/zai-org/GLM-5` | 203K | | | | | | $0.95 | $3 |
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| -------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `baseten/deepseek-ai/DeepSeek-V3-0324` | 164K | | | | | | $0.77 | $0.77 |
|
|
38
|
+
| `baseten/deepseek-ai/DeepSeek-V3.1` | 164K | | | | | | $0.50 | $2 |
|
|
39
|
+
| `baseten/MiniMaxAI/MiniMax-M2.5` | 204K | | | | | | $0.30 | $1 |
|
|
40
|
+
| `baseten/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.60 | $3 |
|
|
41
|
+
| `baseten/nvidia/Nemotron-3-Super` | 262K | | | | | | $0.30 | $0.75 |
|
|
42
|
+
| `baseten/openai/gpt-oss-120b` | 128K | | | | | | $0.10 | $0.50 |
|
|
43
|
+
| `baseten/zai-org/GLM-4.6` | 200K | | | | | | $0.60 | $2 |
|
|
44
|
+
| `baseten/zai-org/GLM-4.7` | 205K | | | | | | $0.60 | $2 |
|
|
45
|
+
| `baseten/zai-org/GLM-5` | 203K | | | | | | $0.95 | $3 |
|
|
47
46
|
|
|
48
47
|
## Advanced configuration
|
|
49
48
|
|
|
@@ -44,9 +44,9 @@ for await (const chunk of stream) {
|
|
|
44
44
|
| `chutes/deepseek-ai/DeepSeek-V3.1-TEE` | 164K | | | | | | $0.20 | $0.80 |
|
|
45
45
|
| `chutes/deepseek-ai/DeepSeek-V3.1-Terminus-TEE` | 164K | | | | | | $0.23 | $0.90 |
|
|
46
46
|
| `chutes/deepseek-ai/DeepSeek-V3.2-Speciale-TEE` | 164K | | | | | | $0.27 | $0.41 |
|
|
47
|
-
| `chutes/deepseek-ai/DeepSeek-V3.2-TEE` |
|
|
47
|
+
| `chutes/deepseek-ai/DeepSeek-V3.2-TEE` | 131K | | | | | | $0.28 | $0.42 |
|
|
48
48
|
| `chutes/MiniMaxAI/MiniMax-M2.1-TEE` | 197K | | | | | | $0.27 | $1 |
|
|
49
|
-
| `chutes/MiniMaxAI/MiniMax-M2.5-TEE` | 197K | | | | | | $0.
|
|
49
|
+
| `chutes/MiniMaxAI/MiniMax-M2.5-TEE` | 197K | | | | | | $0.30 | $1 |
|
|
50
50
|
| `chutes/miromind-ai/MiroThinker-v1.5-235B` | 262K | | | | | | $0.30 | $1 |
|
|
51
51
|
| `chutes/mistralai/Devstral-2-123B-Instruct-2512-TEE` | 262K | | | | | | $0.05 | $0.22 |
|
|
52
52
|
| `chutes/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $0.39 | $2 |
|
|
@@ -76,7 +76,7 @@ for await (const chunk of stream) {
|
|
|
76
76
|
| `chutes/Qwen/Qwen3-Coder-Next` | 262K | | | | | | $0.07 | $0.30 |
|
|
77
77
|
| `chutes/Qwen/Qwen3-Next-80B-A3B-Instruct` | 262K | | | | | | $0.10 | $0.80 |
|
|
78
78
|
| `chutes/Qwen/Qwen3-VL-235B-A22B-Instruct` | 262K | | | | | | $0.30 | $1 |
|
|
79
|
-
| `chutes/Qwen/Qwen3.5-397B-A17B-TEE` | 262K | | | | | | $0.
|
|
79
|
+
| `chutes/Qwen/Qwen3.5-397B-A17B-TEE` | 262K | | | | | | $0.39 | $2 |
|
|
80
80
|
| `chutes/Qwen/Qwen3Guard-Gen-0.6B` | 33K | | | | | | $0.01 | $0.01 |
|
|
81
81
|
| `chutes/rednote-hilab/dots.ocr` | 131K | | | | | | $0.01 | $0.01 |
|
|
82
82
|
| `chutes/tngtech/DeepSeek-R1T-Chimera` | 164K | | | | | | $0.30 | $1 |
|
|
@@ -95,12 +95,12 @@ for await (const chunk of stream) {
|
|
|
95
95
|
| `chutes/zai-org/GLM-4.5-FP8` | 131K | | | | | | $0.30 | $1 |
|
|
96
96
|
| `chutes/zai-org/GLM-4.5-TEE` | 131K | | | | | | $0.35 | $2 |
|
|
97
97
|
| `chutes/zai-org/GLM-4.6-FP8` | 203K | | | | | | $0.30 | $1 |
|
|
98
|
-
| `chutes/zai-org/GLM-4.6-TEE` | 203K | | | | | | $0.
|
|
98
|
+
| `chutes/zai-org/GLM-4.6-TEE` | 203K | | | | | | $0.40 | $2 |
|
|
99
99
|
| `chutes/zai-org/GLM-4.6V` | 131K | | | | | | $0.30 | $0.90 |
|
|
100
100
|
| `chutes/zai-org/GLM-4.7-Flash` | 203K | | | | | | $0.06 | $0.35 |
|
|
101
101
|
| `chutes/zai-org/GLM-4.7-FP8` | 203K | | | | | | $0.30 | $1 |
|
|
102
102
|
| `chutes/zai-org/GLM-4.7-TEE` | 203K | | | | | | $0.40 | $2 |
|
|
103
|
-
| `chutes/zai-org/GLM-5-TEE` | 203K | | | | | | $0.
|
|
103
|
+
| `chutes/zai-org/GLM-5-TEE` | 203K | | | | | | $0.95 | $3 |
|
|
104
104
|
| `chutes/zai-org/GLM-5-Turbo` | 203K | | | | | | $0.49 | $2 |
|
|
105
105
|
|
|
106
106
|
## Advanced configuration
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Deep Infra
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 27 Deep Infra models through Mastra's model router. Authentication is handled automatically using the `DEEPINFRA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Deep Infra documentation](https://deepinfra.com/models).
|
|
6
6
|
|
|
@@ -30,28 +30,35 @@ for await (const chunk of stream) {
|
|
|
30
30
|
|
|
31
31
|
## Models
|
|
32
32
|
|
|
33
|
-
| Model
|
|
34
|
-
|
|
|
35
|
-
| `deepinfra/anthropic/claude-3-7-sonnet-latest`
|
|
36
|
-
| `deepinfra/anthropic/claude-4-opus`
|
|
37
|
-
| `deepinfra/deepseek-ai/DeepSeek-R1-0528`
|
|
38
|
-
| `deepinfra/deepseek-ai/DeepSeek-V3.2`
|
|
39
|
-
| `deepinfra/
|
|
40
|
-
| `deepinfra/
|
|
41
|
-
| `deepinfra/
|
|
42
|
-
| `deepinfra/
|
|
43
|
-
| `deepinfra/
|
|
44
|
-
| `deepinfra/
|
|
45
|
-
| `deepinfra/
|
|
46
|
-
| `deepinfra/
|
|
47
|
-
| `deepinfra/
|
|
48
|
-
| `deepinfra/
|
|
49
|
-
| `deepinfra/
|
|
50
|
-
| `deepinfra/
|
|
51
|
-
| `deepinfra/
|
|
52
|
-
| `deepinfra/
|
|
53
|
-
| `deepinfra/
|
|
54
|
-
| `deepinfra/
|
|
33
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
34
|
+
| ------------------------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
35
|
+
| `deepinfra/anthropic/claude-3-7-sonnet-latest` | 200K | | | | | | $3 | $17 |
|
|
36
|
+
| `deepinfra/anthropic/claude-4-opus` | 200K | | | | | | $17 | $83 |
|
|
37
|
+
| `deepinfra/deepseek-ai/DeepSeek-R1-0528` | 164K | | | | | | $0.50 | $2 |
|
|
38
|
+
| `deepinfra/deepseek-ai/DeepSeek-V3.2` | 164K | | | | | | $0.26 | $0.38 |
|
|
39
|
+
| `deepinfra/meta-llama/Llama-3.1-70B-Instruct` | 131K | | | | | | $0.40 | $0.40 |
|
|
40
|
+
| `deepinfra/meta-llama/Llama-3.1-70B-Instruct-Turbo` | 131K | | | | | | $0.40 | $0.40 |
|
|
41
|
+
| `deepinfra/meta-llama/Llama-3.1-8B-Instruct` | 131K | | | | | | $0.02 | $0.05 |
|
|
42
|
+
| `deepinfra/meta-llama/Llama-3.1-8B-Instruct-Turbo` | 131K | | | | | | $0.02 | $0.03 |
|
|
43
|
+
| `deepinfra/meta-llama/Llama-3.3-70B-Instruct-Turbo` | 131K | | | | | | $0.10 | $0.32 |
|
|
44
|
+
| `deepinfra/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | 1.0M | | | | | | $0.15 | $0.60 |
|
|
45
|
+
| `deepinfra/meta-llama/Llama-4-Scout-17B-16E-Instruct` | 10.0M | | | | | | $0.08 | $0.30 |
|
|
46
|
+
| `deepinfra/MiniMaxAI/MiniMax-M2` | 262K | | | | | | $0.25 | $1 |
|
|
47
|
+
| `deepinfra/MiniMaxAI/MiniMax-M2.1` | 197K | | | | | | $0.28 | $1 |
|
|
48
|
+
| `deepinfra/MiniMaxAI/MiniMax-M2.5` | 205K | | | | | | $0.27 | $0.95 |
|
|
49
|
+
| `deepinfra/moonshotai/Kimi-K2-Instruct` | 131K | | | | | | $0.50 | $2 |
|
|
50
|
+
| `deepinfra/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $0.40 | $2 |
|
|
51
|
+
| `deepinfra/moonshotai/Kimi-K2-Thinking` | 131K | | | | | | $0.47 | $2 |
|
|
52
|
+
| `deepinfra/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.50 | $3 |
|
|
53
|
+
| `deepinfra/openai/gpt-oss-120b` | 131K | | | | | | $0.05 | $0.24 |
|
|
54
|
+
| `deepinfra/openai/gpt-oss-20b` | 131K | | | | | | $0.03 | $0.14 |
|
|
55
|
+
| `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $0.40 | $2 |
|
|
56
|
+
| `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo` | 262K | | | | | | $0.30 | $1 |
|
|
57
|
+
| `deepinfra/zai-org/GLM-4.6` | 205K | | | | | | $0.43 | $2 |
|
|
58
|
+
| `deepinfra/zai-org/GLM-4.6V` | 205K | | | | | | $0.30 | $0.90 |
|
|
59
|
+
| `deepinfra/zai-org/GLM-4.7` | 203K | | | | | | $0.43 | $2 |
|
|
60
|
+
| `deepinfra/zai-org/GLM-4.7-Flash` | 203K | | | | | | $0.06 | $0.40 |
|
|
61
|
+
| `deepinfra/zai-org/GLM-5` | 203K | | | | | | $0.80 | $3 |
|
|
55
62
|
|
|
56
63
|
## Advanced configuration
|
|
57
64
|
|
|
@@ -54,7 +54,7 @@ for await (const chunk of stream) {
|
|
|
54
54
|
| `google/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
|
|
55
55
|
| `google/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
56
56
|
| `google/gemini-3.1-flash-image-preview` | 131K | | | | | | $0.25 | $60 |
|
|
57
|
-
| `google/gemini-3.1-flash-lite-preview` | 1.0M | | | | | | $0.
|
|
57
|
+
| `google/gemini-3.1-flash-lite-preview` | 1.0M | | | | | | $0.25 | $2 |
|
|
58
58
|
| `google/gemini-3.1-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
59
59
|
| `google/gemini-3.1-pro-preview-customtools` | 1.0M | | | | | | $2 | $12 |
|
|
60
60
|
| `google/gemini-embedding-001` | 2K | | | | | | $0.15 | — |
|