@mastra/mcp-docs-server 1.1.13 → 1.1.14-alpha.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/.docs/docs/deployment/studio.md +9 -24
  2. package/.docs/docs/getting-started/studio.md +28 -16
  3. package/.docs/docs/observability/tracing/exporters/braintrust.md +15 -0
  4. package/.docs/docs/server/auth.md +6 -7
  5. package/.docs/docs/server/custom-api-routes.md +56 -0
  6. package/.docs/docs/server/mastra-server.md +2 -2
  7. package/.docs/guides/deployment/cloudflare.md +1 -1
  8. package/.docs/models/gateways/openrouter.md +4 -1
  9. package/.docs/models/gateways/vercel.md +7 -1
  10. package/.docs/models/index.md +1 -1
  11. package/.docs/models/providers/anthropic.md +2 -2
  12. package/.docs/models/providers/baseten.md +12 -13
  13. package/.docs/models/providers/chutes.md +5 -5
  14. package/.docs/models/providers/deepinfra.md +30 -23
  15. package/.docs/models/providers/google.md +1 -1
  16. package/.docs/models/providers/kilo.md +342 -272
  17. package/.docs/models/providers/nano-gpt.md +36 -36
  18. package/.docs/models/providers/nebius.md +3 -2
  19. package/.docs/models/providers/perplexity-agent.md +19 -18
  20. package/.docs/models/providers/synthetic.md +1 -1
  21. package/.docs/models/providers/vultr.md +17 -12
  22. package/.docs/models/providers/zai-coding-plan.md +3 -2
  23. package/.docs/models/providers/zai.md +3 -2
  24. package/.docs/reference/agents/generate.md +2 -0
  25. package/.docs/reference/agents/network.md +2 -0
  26. package/.docs/reference/ai-sdk/chat-route.md +4 -0
  27. package/.docs/reference/configuration.md +4 -2
  28. package/.docs/reference/deployer/cloudflare.md +12 -1
  29. package/.docs/reference/processors/unicode-normalizer.md +1 -1
  30. package/.docs/reference/streaming/agents/stream.md +2 -0
  31. package/.docs/reference/workflows/run-methods/restart.md +2 -0
  32. package/.docs/reference/workflows/run-methods/resume.md +2 -0
  33. package/.docs/reference/workflows/run-methods/start.md +2 -0
  34. package/.docs/reference/workflows/run-methods/timeTravel.md +2 -0
  35. package/CHANGELOG.md +15 -0
  36. package/dist/prompts/migration.d.ts.map +1 -1
  37. package/dist/stdio.js.map +1 -1
  38. package/package.json +8 -8
@@ -1,4 +1,4 @@
1
- # Deploying studio
1
+ # Deploying Studio
2
2
 
3
3
  [Studio](https://mastra.ai/docs/getting-started/studio) provides an interactive UI for building and testing your agents. It's a React-based Single Page Application (SPA) that runs in the browser and connects to a running [Mastra server](https://mastra.ai/docs/deployment/mastra-server).
4
4
 
@@ -213,35 +213,20 @@ Follow the example below to create a SPA using Vite.
213
213
  MASTRA_STUDIO_BASE_PATH=
214
214
  MASTRA_TELEMETRY_DISABLED=true
215
215
  MASTRA_HIDE_CLOUD_CTA=false
216
+ MASTRA_TEMPLATES=false
216
217
  MASTRA_CLOUD_API_ENDPOINT=
217
218
  MASTRA_EXPERIMENTAL_FEATURES=false
218
219
  MASTRA_REQUEST_CONTEXT_PRESETS=
219
220
  ```
220
221
 
221
- 7. Run the build script to generate the static files in the `dist` folder:
222
+ ````text
223
+ </StepItem>
222
224
 
223
- **npm**:
225
+ <StepItem>
226
+ Run the build script to generate the static files in the `dist` folder:
224
227
 
225
- ```bash
228
+ ```bash npm2yarn
226
229
  npm run build
227
- ```
228
-
229
- **pnpm**:
230
-
231
- ```bash
232
- pnpm run build
233
- ```
234
-
235
- **Yarn**:
236
-
237
- ```bash
238
- yarn build
239
- ```
240
-
241
- **Bun**:
242
-
243
- ```bash
244
- bun run build
245
- ```
230
+ ````
246
231
 
247
- 8. Point your hosting provider to the `dist` folder and deploy!
232
+ 7. Point your hosting provider to the `dist` folder and deploy!
@@ -1,16 +1,14 @@
1
1
  # Studio
2
2
 
3
- Studio provides an interactive UI for building and testing your agents, along with a REST API that exposes your Mastra application as a local service. This lets you start building without worrying about integration right away.
3
+ Studio provides an interactive UI for building, testing, and managing your agents, workflows, and tools. Run it locally during development, or [deploy it](https://mastra.ai/docs/deployment/studio) to production so your team can manage agents, monitor performance, and gain insights through built-in observability.
4
4
 
5
- As your project evolves, Studio's development environment helps you iterate on your agent quickly. Meanwhile, Observability and Scorer features give you visibility into performance at every stage.
6
-
7
- To get started, run Studio locally using the instructions below, or [create a project in Mastra Cloud](https://mastra.ai/docs/mastra-cloud/setup) to collaborate with your team.
5
+ Add [authentication](https://mastra.ai/docs/server/auth) to protect your deployed Studio with login screens, role-based access control, and permission-based UI rendering so you can control what each team member can see and do. You can also [create a project in Mastra Cloud](https://mastra.ai/docs/mastra-cloud/setup) for a hosted option.
8
6
 
9
7
  [YouTube video player](https://www.youtube-nocookie.com/embed/ojGu6Bi4wYk)
10
8
 
11
9
  ## Start Studio
12
10
 
13
- If you created your application with `create mastra`, start the local development server using the `dev` script. You can also run it directly with `mastra dev`.
11
+ If you created your application with `create mastra`, start the development server using the `dev` script. You can also run it directly with `mastra dev`.
14
12
 
15
13
  **npm**:
16
14
 
@@ -36,14 +34,16 @@ yarn dev
36
34
  bun run dev
37
35
  ```
38
36
 
39
- Once the server's running, you can:
37
+ Once the server is running, you can:
40
38
 
41
- - Open the Studio UI at <http://localhost:4111/> to test your agent interactively.
39
+ - Open the Studio UI at <http://localhost:4111/> to interact with your agents, workflows, and tools.
42
40
  - Visit <http://localhost:4111/swagger-ui> to discover and interact with the underlying REST API.
43
41
 
42
+ To run Studio in production, see [Deploy Studio](https://mastra.ai/docs/deployment/studio).
43
+
44
44
  ## Studio UI
45
45
 
46
- The Studio UI provides an interactive development environment for you to test your agents, workflows, and tools, observe exactly what happens under the hood with each interaction, and tweak things as you go.
46
+ The Studio UI lets you interact with your agents, workflows, and tools, observe exactly what happens under the hood with each interaction, and tweak things as you go.
47
47
 
48
48
  ### Agents
49
49
 
@@ -61,39 +61,51 @@ When running a workflow, you can also view detailed traces showing tool calls, r
61
61
 
62
62
  Run tools in isolation to observe their behavior. Test them before assigning them to your agent, or isolate them to debug issues should something go wrong.
63
63
 
64
+ ### Processors
65
+
66
+ View the input and output processors attached to each agent. The agent detail panel lists every processor by name and type, so you can verify your guardrails, token limiters, and custom processors are wired up correctly before testing.
67
+
68
+ See [Processors](https://mastra.ai/docs/agents/processors) and [Guardrails](https://mastra.ai/docs/agents/guardrails) for configuration details.
69
+
64
70
  ### MCP
65
71
 
66
72
  List the MCP servers attached to your Mastra instance and explore their available tools.
67
73
 
68
- ![MCP Servers Studio](/assets/images/local-dev-mcp-server-playground-8551b0af59838b2ef0bf4756ce94dcf5.jpg)
69
-
70
74
  ### Observability
71
75
 
72
76
  When you run an agent or workflow, the Observability tab displays traces that highlight the key AI operations such as model calls, tool executions, and workflow steps. Follow these traces to see how data moves, where time is spent, and what's happening under the hood.
73
77
 
74
- ![](https://mastra.ai/_next/image?url=%2Ftracingafter.png\&w=1920\&q=75)
75
-
76
78
  Tracing filters out low-level framework details so your traces stay focused and readable.
77
79
 
78
80
  ### Scorers
79
81
 
80
82
  The Scorers tab displays the results of your agent's scorers as they run. When messages pass through your agent, the defined scorers evaluate each output asynchronously and render their results here. This allows you to understand how your scorers respond to different interactions, compare performance across test cases, and identify areas for improvement.
81
83
 
84
+ ### Datasets
85
+
86
+ Create and manage collections of test cases to evaluate your agents and workflows. Import items from CSV or JSON, define input and ground-truth schemas, and pin to specific versions so you can reproduce experiments exactly. Run experiments with [scorers](https://mastra.ai/docs/evals/overview) to compare quality across prompts, models, or code changes.
87
+
88
+ See [Datasets overview](https://mastra.ai/docs/observability/datasets/overview) for the full API and versioning details.
89
+
82
90
  ## REST API
83
91
 
84
- The local development server exposes a complete set of REST API routes, allowing you to programmatically interact with your agents, workflows, and tools during development. This is particularly helpful if you plan to deploy the Mastra server, since the local development server uses the exact same API routes as the [Mastra Server](https://mastra.ai/docs/server/mastra-server), allowing you to develop and test against it with full parity.
92
+ Studio is backed by a complete set of REST API routes that let you programmatically interact with your agents, workflows, and tools. These are the same routes exposed by the [Mastra Server](https://mastra.ai/docs/server/mastra-server), so everything you build against locally works identically in production.
85
93
 
86
94
  You can explore all available endpoints in the OpenAPI specification at <http://localhost:4111/api/openapi.json>, which details every endpoint and its request and response schemas.
87
95
 
88
96
  To explore the API interactively, visit the Swagger UI at <http://localhost:4111/swagger-ui>. Here, you can discover endpoints and test them directly from your browser.
89
97
 
90
- > **Info:** The OpenAPI and Swagger endpoints are disabled in production by default. To enable them, set [`server.build.openAPIDocs`](https://mastra.ai/reference/configuration) and [`server.build.swaggerUI`](https://mastra.ai/reference/configuration) to `true` respectively.
98
+ > **Note:** The OpenAPI and Swagger endpoints are disabled in production by default. To enable them, set [`server.build.openAPIDocs`](https://mastra.ai/reference/configuration) and [`server.build.swaggerUI`](https://mastra.ai/reference/configuration) to `true` respectively.
91
99
 
92
100
  ## Configuration
93
101
 
94
- By default, Studio runs at <http://localhost:4111>. You can change the [`host`](https://mastra.ai/reference/configuration), [`port`](https://mastra.ai/reference/configuration), and [`studioBase`](https://mastra.ai/reference/configuration) in the Mastra server configuration. This allows you to customize where and how Studio is hosted.
102
+ By default, Studio runs at <http://localhost:4111>. You can change the [`host`](https://mastra.ai/reference/configuration), [`port`](https://mastra.ai/reference/configuration), and [`studioBase`](https://mastra.ai/reference/configuration) in the Mastra server configuration.
103
+
104
+ For production deployments, see [Deploy Studio](https://mastra.ai/docs/deployment/studio) to learn about hosting Studio alongside your server, as a standalone SPA, or on a CDN.
105
+
106
+ Add [authentication](https://mastra.ai/docs/server/auth) to control who can access Studio in production. Studio displays the appropriate login UI, which can be an SSO button, an email/password form, or both. All API routes require authentication. This applies to any request made to your Mastra API, whether from Studio or a direct API call.
95
107
 
96
- Furthermore, Mastra supports local HTTPS development through the [`--https`](https://mastra.ai/reference/cli/mastra) flag, which automatically creates and manages certificates for your project. When you run `mastra dev --https`, a private key and certificate are generated for localhost (or your configured host). Visit the [HTTPS reference](https://mastra.ai/reference/configuration) to learn more.
108
+ Mastra also supports HTTPS development through the [`--https`](https://mastra.ai/reference/cli/mastra) flag, which automatically creates and manages certificates for your project. When you run `mastra dev --https`, a private key and certificate are generated for localhost (or your configured host). Visit the [HTTPS reference](https://mastra.ai/reference/configuration) to learn more.
97
109
 
98
110
  ## Next steps
99
111
 
@@ -105,6 +105,21 @@ new BraintrustExporter({
105
105
  })
106
106
  ```
107
107
 
108
+ ## Querying Braintrust with returned `spanId`
109
+
110
+ For Braintrust, use `spanId` as the root span identifier when searching for traces because Braintrust root-span queries are typically faster than trace-id queries.
111
+
112
+ ```typescript
113
+ const result = await agent.stream('Summarize this ticket')
114
+
115
+ console.log('Mastra trace ID:', result.traceId)
116
+ console.log('Braintrust root span ID:', result.spanId)
117
+
118
+ // Use result.spanId in your Braintrust lookup/query path
119
+ ```
120
+
121
+ The same applies to `agent.generate()` and workflow run results (`run.start()`, `run.stream()` final state, `run.resume()`).
122
+
108
123
  ## Related
109
124
 
110
125
  - [Tracing Overview](https://mastra.ai/docs/observability/tracing/overview)
@@ -1,18 +1,17 @@
1
1
  # Auth overview
2
2
 
3
- Mastra lets you choose how you handle authentication, so you can secure access to your application's endpoints using the identity system that fits your stack.
3
+ Mastra lets you choose how you handle authentication, so you can secure access to your API and [Studio](https://mastra.ai/docs/getting-started/studio) using the identity system that fits your stack.
4
4
 
5
5
  You can start with basic shared secret JWT authentication and switch to providers like Supabase, Firebase Auth, Auth0, Clerk, or WorkOS when you need more advanced identity features.
6
6
 
7
- ## Default behavior
7
+ ## What auth secures
8
8
 
9
- Authentication is optional in Mastra. When you configure authentication:
9
+ Configuring authentication locks down two things at once:
10
10
 
11
- - **All built-in API routes** (`/api/agents/*`, `/api/workflows/*`, etc.) require authentication by default
12
- - **Custom API routes** also require authentication by default
13
- - **Public access** can be enabled on custom routes using `requiresAuth: false`
11
+ - **Studio UI**: Studio displays a login screen (SSO, email/password, or both) and enforces role-based access control to determine what each team member can see and do.
12
+ - **API routes**: All built-in routes (`/api/agents/*`, `/api/workflows/*`, etc.) and custom routes require authentication, whether requests come from Studio or direct API calls.
14
13
 
15
- If no authentication is configured, all routes are publicly accessible.
14
+ Authentication is optional. If no auth is configured, all routes and Studio are publicly accessible. Public access can be enabled on individual custom routes using `requiresAuth: false`.
16
15
 
17
16
  See [Custom API Routes](https://mastra.ai/docs/server/custom-api-routes) for controlling authentication on custom endpoints.
18
17
 
@@ -260,6 +260,62 @@ registerApiRoute('/user-profile', {
260
260
 
261
261
  For more information about authentication providers, see the [Auth documentation](https://mastra.ai/docs/server/auth).
262
262
 
263
+ ## Continue generation after client disconnect
264
+
265
+ Built-in streaming helpers such as [`chatRoute()`](https://mastra.ai/reference/ai-sdk/chat-route) forward the incoming request's `AbortSignal` to `agent.stream()`. That is the right default when a browser disconnect should cancel the model call.
266
+
267
+ If you want the server to keep generating and persist the final response even after the client disconnects, build a custom route around the underlying `MastraModelOutput`. Start the agent stream without forwarding `c.req.raw.signal`, then call `consumeStream()` in the background so generation continues server-side.
268
+
269
+ ```typescript
270
+ import {
271
+ createUIMessageStream,
272
+ createUIMessageStreamResponse,
273
+ InferUIMessageChunk,
274
+ UIMessage,
275
+ } from 'ai'
276
+ import { toAISdkStream } from '@mastra/ai-sdk'
277
+ import { Mastra } from '@mastra/core'
278
+ import { registerApiRoute } from '@mastra/core/server'
279
+
280
+ export const mastra = new Mastra({
281
+ server: {
282
+ apiRoutes: [
283
+ registerApiRoute('/chat/persist/:agentId', {
284
+ method: 'POST',
285
+ handler: async c => {
286
+ const { messages, memory } = await c.req.json()
287
+ const mastra = c.get('mastra')
288
+ const agent = mastra.getAgent(c.req.param('agentId'))
289
+
290
+ const stream = await agent.stream(messages, {
291
+ memory,
292
+ // Do not pass c.req.raw.signal if this route should keep running
293
+ // after the client disconnects.
294
+ })
295
+
296
+ void stream.consumeStream().catch(error => {
297
+ mastra.getLogger()?.error('Background stream consumption failed', { error })
298
+ })
299
+
300
+ const uiStream = createUIMessageStream({
301
+ originalMessages: messages,
302
+ execute: async ({ writer }) => {
303
+ for await (const part of toAISdkStream(stream, { from: 'agent' })) {
304
+ writer.write(part as InferUIMessageChunk<UIMessage>)
305
+ }
306
+ },
307
+ })
308
+
309
+ return createUIMessageStreamResponse({ stream: uiStream })
310
+ },
311
+ }),
312
+ ],
313
+ },
314
+ })
315
+ ```
316
+
317
+ > **Note:** Use this pattern only when you intentionally want work to continue after the HTTP client is gone. If you want disconnects to cancel generation, keep using `chatRoute()` or forward the request `AbortSignal` yourself.
318
+
263
319
  ## Related
264
320
 
265
321
  - [registerApiRoute() Reference](https://mastra.ai/reference/server/register-api-route) - Full API reference
@@ -25,8 +25,8 @@ import { Mastra } from '@mastra/core'
25
25
 
26
26
  export const mastra = new Mastra({
27
27
  server: {
28
- port: 3000, // Defaults to 4111
29
- host: '0.0.0.0', // Defaults to 'localhost'
28
+ port: 3000, // Defaults to PORT env var or 4111
29
+ host: '0.0.0.0', // Defaults to MASTRA_HOST env var or 'localhost'
30
30
  },
31
31
  })
32
32
  ```
@@ -86,7 +86,7 @@ After setting up your project, push it to your remote Git provider of choice (e.
86
86
 
87
87
  1. Connect your repository to Cloudflare. On the "Workers & Pages" dashboard, select **Create application** and choose your Git provider in the next step. Continue with the setup process and select the repository you want to deploy.
88
88
 
89
- > **Note:** Remember to set your environment variables needed to run your application (e.g. your [model provider](https://mastra.ai/models/providers) API key).
89
+ > **Note:** Remember to set your environment variables needed to run your application (e.g. your [model provider](https://mastra.ai/models/providers) API key). You can upload secrets from your `.env` file using `npx wrangler secret bulk .env`. See [Secrets](https://mastra.ai/reference/deployer/cloudflare) for details.
90
90
 
91
91
  2. Once you're ready, click the **Deploy** button and wait for the first deployment to complete.
92
92
 
@@ -1,6 +1,6 @@
1
1
  # ![OpenRouter logo](https://models.dev/logos/openrouter.svg)OpenRouter
2
2
 
3
- OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 195 models through Mastra's model router.
3
+ OpenRouter aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 198 models through Mastra's model router.
4
4
 
5
5
  Learn more in the [OpenRouter documentation](https://openrouter.ai/models).
6
6
 
@@ -79,6 +79,7 @@ ANTHROPIC_API_KEY=ant-...
79
79
  | `google/gemini-2.5-pro-preview-06-05` |
80
80
  | `google/gemini-3-flash-preview` |
81
81
  | `google/gemini-3-pro-preview` |
82
+ | `google/gemini-3.1-flash-lite-preview` |
82
83
  | `google/gemini-3.1-pro-preview` |
83
84
  | `google/gemini-3.1-pro-preview-customtools` |
84
85
  | `google/gemma-2-9b-it` |
@@ -218,6 +219,8 @@ ANTHROPIC_API_KEY=ant-...
218
219
  | `x-ai/grok-4` |
219
220
  | `x-ai/grok-4-fast` |
220
221
  | `x-ai/grok-4.1-fast` |
222
+ | `x-ai/grok-4.20-beta` |
223
+ | `x-ai/grok-4.20-multi-agent-beta` |
221
224
  | `x-ai/grok-code-fast-1` |
222
225
  | `xiaomi/mimo-v2-flash` |
223
226
  | `z-ai/glm-4.5` |
@@ -1,6 +1,6 @@
1
1
  # ![Vercel logo](https://models.dev/logos/vercel.svg)Vercel
2
2
 
3
- Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 208 models through Mastra's model router.
3
+ Vercel aggregates models from multiple providers with enhanced features like rate limiting and failover. Access 214 models through Mastra's model router.
4
4
 
5
5
  Learn more in the [Vercel documentation](https://ai-sdk.dev/providers/ai-sdk-providers).
6
6
 
@@ -118,6 +118,7 @@ ANTHROPIC_API_KEY=ant-...
118
118
  | `kwaipilot/kat-coder-pro-v1` |
119
119
  | `meituan/longcat-flash-chat` |
120
120
  | `meituan/longcat-flash-thinking` |
121
+ | `meituan/longcat-flash-thinking-2601` |
121
122
  | `meta/llama-3.1-70b` |
122
123
  | `meta/llama-3.1-8b` |
123
124
  | `meta/llama-3.2-11b` |
@@ -131,6 +132,7 @@ ANTHROPIC_API_KEY=ant-...
131
132
  | `minimax/minimax-m2.1` |
132
133
  | `minimax/minimax-m2.1-lightning` |
133
134
  | `minimax/minimax-m2.5` |
135
+ | `minimax/minimax-m2.5-highspeed` |
134
136
  | `mistral/codestral` |
135
137
  | `mistral/codestral-embed` |
136
138
  | `mistral/devstral-2` |
@@ -229,6 +231,9 @@ ANTHROPIC_API_KEY=ant-...
229
231
  | `xai/grok-4-fast-reasoning` |
230
232
  | `xai/grok-4.1-fast-non-reasoning` |
231
233
  | `xai/grok-4.1-fast-reasoning` |
234
+ | `xai/grok-4.20-multi-agent-beta` |
235
+ | `xai/grok-4.20-non-reasoning-beta` |
236
+ | `xai/grok-4.20-reasoning-beta` |
232
237
  | `xai/grok-code-fast-1` |
233
238
  | `xai/grok-imagine-image` |
234
239
  | `xai/grok-imagine-image-pro` |
@@ -240,5 +245,6 @@ ANTHROPIC_API_KEY=ant-...
240
245
  | `zai/glm-4.6v` |
241
246
  | `zai/glm-4.6v-flash` |
242
247
  | `zai/glm-4.7` |
248
+ | `zai/glm-4.7-flash` |
243
249
  | `zai/glm-4.7-flashx` |
244
250
  | `zai/glm-5` |
@@ -1,6 +1,6 @@
1
1
  # Model Providers
2
2
 
3
- Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3259 models from 92 providers through a single API.
3
+ Mastra provides a unified interface for working with LLMs across multiple providers, giving you access to 3353 models from 92 providers through a single API.
4
4
 
5
5
  ## Features
6
6
 
@@ -49,12 +49,12 @@ for await (const chunk of stream) {
49
49
  | `anthropic/claude-opus-4-20250514` | 200K | | | | | | $15 | $75 |
50
50
  | `anthropic/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
51
51
  | `anthropic/claude-opus-4-5-20251101` | 200K | | | | | | $5 | $25 |
52
- | `anthropic/claude-opus-4-6` | 200K | | | | | | $5 | $25 |
52
+ | `anthropic/claude-opus-4-6` | 1.0M | | | | | | $5 | $25 |
53
53
  | `anthropic/claude-sonnet-4-0` | 200K | | | | | | $3 | $15 |
54
54
  | `anthropic/claude-sonnet-4-20250514` | 200K | | | | | | $3 | $15 |
55
55
  | `anthropic/claude-sonnet-4-5` | 200K | | | | | | $3 | $15 |
56
56
  | `anthropic/claude-sonnet-4-5-20250929` | 200K | | | | | | $3 | $15 |
57
- | `anthropic/claude-sonnet-4-6` | 200K | | | | | | $3 | $15 |
57
+ | `anthropic/claude-sonnet-4-6` | 1.0M | | | | | | $3 | $15 |
58
58
 
59
59
  ## Advanced configuration
60
60
 
@@ -1,6 +1,6 @@
1
1
  # ![Baseten logo](https://models.dev/logos/baseten.svg)Baseten
2
2
 
3
- Access 10 Baseten models through Mastra's model router. Authentication is handled automatically using the `BASETEN_API_KEY` environment variable.
3
+ Access 9 Baseten models through Mastra's model router. Authentication is handled automatically using the `BASETEN_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Baseten documentation](https://docs.baseten.co/development/model-apis/overview).
6
6
 
@@ -32,18 +32,17 @@ for await (const chunk of stream) {
32
32
 
33
33
  ## Models
34
34
 
35
- | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
- | --------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
- | `baseten/deepseek-ai/DeepSeek-V3.2` | 164K | | | | | | $0.30 | $0.45 |
38
- | `baseten/MiniMaxAI/MiniMax-M2.5` | 204K | | | | | | $0.30 | $1 |
39
- | `baseten/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $0.60 | $3 |
40
- | `baseten/moonshotai/Kimi-K2-Thinking` | 262K | | | | | | $0.60 | $3 |
41
- | `baseten/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.60 | $3 |
42
- | `baseten/nvidia/Nemotron-3-Super` | 262K | | | | | | $0.30 | $0.75 |
43
- | `baseten/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $0.38 | $2 |
44
- | `baseten/zai-org/GLM-4.6` | 200K | | | | | | $0.60 | $2 |
45
- | `baseten/zai-org/GLM-4.7` | 205K | | | | | | $0.60 | $2 |
46
- | `baseten/zai-org/GLM-5` | 203K | | | | | | $0.95 | $3 |
35
+ | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
36
+ | -------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
37
+ | `baseten/deepseek-ai/DeepSeek-V3-0324` | 164K | | | | | | $0.77 | $0.77 |
38
+ | `baseten/deepseek-ai/DeepSeek-V3.1` | 164K | | | | | | $0.50 | $2 |
39
+ | `baseten/MiniMaxAI/MiniMax-M2.5` | 204K | | | | | | $0.30 | $1 |
40
+ | `baseten/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.60 | $3 |
41
+ | `baseten/nvidia/Nemotron-3-Super` | 262K | | | | | | $0.30 | $0.75 |
42
+ | `baseten/openai/gpt-oss-120b` | 128K | | | | | | $0.10 | $0.50 |
43
+ | `baseten/zai-org/GLM-4.6` | 200K | | | | | | $0.60 | $2 |
44
+ | `baseten/zai-org/GLM-4.7` | 205K | | | | | | $0.60 | $2 |
45
+ | `baseten/zai-org/GLM-5` | 203K | | | | | | $0.95 | $3 |
47
46
 
48
47
  ## Advanced configuration
49
48
 
@@ -44,9 +44,9 @@ for await (const chunk of stream) {
44
44
  | `chutes/deepseek-ai/DeepSeek-V3.1-TEE` | 164K | | | | | | $0.20 | $0.80 |
45
45
  | `chutes/deepseek-ai/DeepSeek-V3.1-Terminus-TEE` | 164K | | | | | | $0.23 | $0.90 |
46
46
  | `chutes/deepseek-ai/DeepSeek-V3.2-Speciale-TEE` | 164K | | | | | | $0.27 | $0.41 |
47
- | `chutes/deepseek-ai/DeepSeek-V3.2-TEE` | 164K | | | | | | $0.25 | $0.38 |
47
+ | `chutes/deepseek-ai/DeepSeek-V3.2-TEE` | 131K | | | | | | $0.28 | $0.42 |
48
48
  | `chutes/MiniMaxAI/MiniMax-M2.1-TEE` | 197K | | | | | | $0.27 | $1 |
49
- | `chutes/MiniMaxAI/MiniMax-M2.5-TEE` | 197K | | | | | | $0.15 | $0.60 |
49
+ | `chutes/MiniMaxAI/MiniMax-M2.5-TEE` | 197K | | | | | | $0.30 | $1 |
50
50
  | `chutes/miromind-ai/MiroThinker-v1.5-235B` | 262K | | | | | | $0.30 | $1 |
51
51
  | `chutes/mistralai/Devstral-2-123B-Instruct-2512-TEE` | 262K | | | | | | $0.05 | $0.22 |
52
52
  | `chutes/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $0.39 | $2 |
@@ -76,7 +76,7 @@ for await (const chunk of stream) {
76
76
  | `chutes/Qwen/Qwen3-Coder-Next` | 262K | | | | | | $0.07 | $0.30 |
77
77
  | `chutes/Qwen/Qwen3-Next-80B-A3B-Instruct` | 262K | | | | | | $0.10 | $0.80 |
78
78
  | `chutes/Qwen/Qwen3-VL-235B-A22B-Instruct` | 262K | | | | | | $0.30 | $1 |
79
- | `chutes/Qwen/Qwen3.5-397B-A17B-TEE` | 262K | | | | | | $0.30 | $1 |
79
+ | `chutes/Qwen/Qwen3.5-397B-A17B-TEE` | 262K | | | | | | $0.39 | $2 |
80
80
  | `chutes/Qwen/Qwen3Guard-Gen-0.6B` | 33K | | | | | | $0.01 | $0.01 |
81
81
  | `chutes/rednote-hilab/dots.ocr` | 131K | | | | | | $0.01 | $0.01 |
82
82
  | `chutes/tngtech/DeepSeek-R1T-Chimera` | 164K | | | | | | $0.30 | $1 |
@@ -95,12 +95,12 @@ for await (const chunk of stream) {
95
95
  | `chutes/zai-org/GLM-4.5-FP8` | 131K | | | | | | $0.30 | $1 |
96
96
  | `chutes/zai-org/GLM-4.5-TEE` | 131K | | | | | | $0.35 | $2 |
97
97
  | `chutes/zai-org/GLM-4.6-FP8` | 203K | | | | | | $0.30 | $1 |
98
- | `chutes/zai-org/GLM-4.6-TEE` | 203K | | | | | | $0.35 | $2 |
98
+ | `chutes/zai-org/GLM-4.6-TEE` | 203K | | | | | | $0.40 | $2 |
99
99
  | `chutes/zai-org/GLM-4.6V` | 131K | | | | | | $0.30 | $0.90 |
100
100
  | `chutes/zai-org/GLM-4.7-Flash` | 203K | | | | | | $0.06 | $0.35 |
101
101
  | `chutes/zai-org/GLM-4.7-FP8` | 203K | | | | | | $0.30 | $1 |
102
102
  | `chutes/zai-org/GLM-4.7-TEE` | 203K | | | | | | $0.40 | $2 |
103
- | `chutes/zai-org/GLM-5-TEE` | 203K | | | | | | $0.75 | $3 |
103
+ | `chutes/zai-org/GLM-5-TEE` | 203K | | | | | | $0.95 | $3 |
104
104
  | `chutes/zai-org/GLM-5-Turbo` | 203K | | | | | | $0.49 | $2 |
105
105
 
106
106
  ## Advanced configuration
@@ -1,6 +1,6 @@
1
1
  # ![Deep Infra logo](https://models.dev/logos/deepinfra.svg)Deep Infra
2
2
 
3
- Access 20 Deep Infra models through Mastra's model router. Authentication is handled automatically using the `DEEPINFRA_API_KEY` environment variable.
3
+ Access 27 Deep Infra models through Mastra's model router. Authentication is handled automatically using the `DEEPINFRA_API_KEY` environment variable.
4
4
 
5
5
  Learn more in the [Deep Infra documentation](https://deepinfra.com/models).
6
6
 
@@ -30,28 +30,35 @@ for await (const chunk of stream) {
30
30
 
31
31
  ## Models
32
32
 
33
- | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
34
- | ----------------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
35
- | `deepinfra/anthropic/claude-3-7-sonnet-latest` | 200K | | | | | | $3 | $17 |
36
- | `deepinfra/anthropic/claude-4-opus` | 200K | | | | | | $17 | $83 |
37
- | `deepinfra/deepseek-ai/DeepSeek-R1-0528` | 164K | | | | | | $0.50 | $2 |
38
- | `deepinfra/deepseek-ai/DeepSeek-V3.2` | 164K | | | | | | $0.26 | $0.38 |
39
- | `deepinfra/MiniMaxAI/MiniMax-M2` | 262K | | | | | | $0.25 | $1 |
40
- | `deepinfra/MiniMaxAI/MiniMax-M2.1` | 197K | | | | | | $0.28 | $1 |
41
- | `deepinfra/MiniMaxAI/MiniMax-M2.5` | 205K | | | | | | $0.27 | $0.95 |
42
- | `deepinfra/moonshotai/Kimi-K2-Instruct` | 131K | | | | | | $0.50 | $2 |
43
- | `deepinfra/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $0.40 | $2 |
44
- | `deepinfra/moonshotai/Kimi-K2-Thinking` | 131K | | | | | | $0.47 | $2 |
45
- | `deepinfra/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.50 | $3 |
46
- | `deepinfra/openai/gpt-oss-120b` | 131K | | | | | | $0.05 | $0.24 |
47
- | `deepinfra/openai/gpt-oss-20b` | 131K | | | | | | $0.03 | $0.14 |
48
- | `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $0.40 | $2 |
49
- | `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo` | 262K | | | | | | $0.30 | $1 |
50
- | `deepinfra/zai-org/GLM-4.6` | 205K | | | | | | $0.43 | $2 |
51
- | `deepinfra/zai-org/GLM-4.6V` | 205K | | | | | | $0.30 | $0.90 |
52
- | `deepinfra/zai-org/GLM-4.7` | 203K | | | | | | $0.43 | $2 |
53
- | `deepinfra/zai-org/GLM-4.7-Flash` | 203K | | | | | | $0.06 | $0.40 |
54
- | `deepinfra/zai-org/GLM-5` | 203K | | | | | | $0.80 | $3 |
33
+ | Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
34
+ | ------------------------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
35
+ | `deepinfra/anthropic/claude-3-7-sonnet-latest` | 200K | | | | | | $3 | $17 |
36
+ | `deepinfra/anthropic/claude-4-opus` | 200K | | | | | | $17 | $83 |
37
+ | `deepinfra/deepseek-ai/DeepSeek-R1-0528` | 164K | | | | | | $0.50 | $2 |
38
+ | `deepinfra/deepseek-ai/DeepSeek-V3.2` | 164K | | | | | | $0.26 | $0.38 |
39
+ | `deepinfra/meta-llama/Llama-3.1-70B-Instruct` | 131K | | | | | | $0.40 | $0.40 |
40
+ | `deepinfra/meta-llama/Llama-3.1-70B-Instruct-Turbo` | 131K | | | | | | $0.40 | $0.40 |
41
+ | `deepinfra/meta-llama/Llama-3.1-8B-Instruct` | 131K | | | | | | $0.02 | $0.05 |
42
+ | `deepinfra/meta-llama/Llama-3.1-8B-Instruct-Turbo` | 131K | | | | | | $0.02 | $0.03 |
43
+ | `deepinfra/meta-llama/Llama-3.3-70B-Instruct-Turbo` | 131K | | | | | | $0.10 | $0.32 |
44
+ | `deepinfra/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | 1.0M | | | | | | $0.15 | $0.60 |
45
+ | `deepinfra/meta-llama/Llama-4-Scout-17B-16E-Instruct` | 10.0M | | | | | | $0.08 | $0.30 |
46
+ | `deepinfra/MiniMaxAI/MiniMax-M2` | 262K | | | | | | $0.25 | $1 |
47
+ | `deepinfra/MiniMaxAI/MiniMax-M2.1` | 197K | | | | | | $0.28 | $1 |
48
+ | `deepinfra/MiniMaxAI/MiniMax-M2.5` | 205K | | | | | | $0.27 | $0.95 |
49
+ | `deepinfra/moonshotai/Kimi-K2-Instruct` | 131K | | | | | | $0.50 | $2 |
50
+ | `deepinfra/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $0.40 | $2 |
51
+ | `deepinfra/moonshotai/Kimi-K2-Thinking` | 131K | | | | | | $0.47 | $2 |
52
+ | `deepinfra/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.50 | $3 |
53
+ | `deepinfra/openai/gpt-oss-120b` | 131K | | | | | | $0.05 | $0.24 |
54
+ | `deepinfra/openai/gpt-oss-20b` | 131K | | | | | | $0.03 | $0.14 |
55
+ | `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $0.40 | $2 |
56
+ | `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo` | 262K | | | | | | $0.30 | $1 |
57
+ | `deepinfra/zai-org/GLM-4.6` | 205K | | | | | | $0.43 | $2 |
58
+ | `deepinfra/zai-org/GLM-4.6V` | 205K | | | | | | $0.30 | $0.90 |
59
+ | `deepinfra/zai-org/GLM-4.7` | 203K | | | | | | $0.43 | $2 |
60
+ | `deepinfra/zai-org/GLM-4.7-Flash` | 203K | | | | | | $0.06 | $0.40 |
61
+ | `deepinfra/zai-org/GLM-5` | 203K | | | | | | $0.80 | $3 |
55
62
 
56
63
  ## Advanced configuration
57
64
 
@@ -54,7 +54,7 @@ for await (const chunk of stream) {
54
54
  | `google/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
55
55
  | `google/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
56
56
  | `google/gemini-3.1-flash-image-preview` | 131K | | | | | | $0.25 | $60 |
57
- | `google/gemini-3.1-flash-lite-preview` | 1.0M | | | | | | $0.50 | $3 |
57
+ | `google/gemini-3.1-flash-lite-preview` | 1.0M | | | | | | $0.25 | $2 |
58
58
  | `google/gemini-3.1-pro-preview` | 1.0M | | | | | | $2 | $12 |
59
59
  | `google/gemini-3.1-pro-preview-customtools` | 1.0M | | | | | | $2 | $12 |
60
60
  | `google/gemini-embedding-001` | 2K | | | | | | $0.15 | — |