@mastra/mcp-docs-server 1.1.26-alpha.6 → 1.1.26
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/agents/structured-output.md +22 -0
- package/.docs/docs/agents/supervisor-agents.md +18 -0
- package/.docs/docs/editor/overview.md +69 -0
- package/.docs/docs/memory/observational-memory.md +4 -0
- package/.docs/docs/memory/storage.md +1 -0
- package/.docs/docs/observability/tracing/exporters/langfuse.md +31 -0
- package/.docs/guides/deployment/netlify.md +16 -1
- package/.docs/guides/getting-started/next-js.md +0 -4
- package/.docs/guides/migrations/mastra-cloud.md +128 -3
- package/.docs/models/gateways/netlify.md +2 -2
- package/.docs/models/gateways/openrouter.md +3 -1
- package/.docs/models/gateways/vercel.md +4 -1
- package/.docs/models/index.md +36 -1
- package/.docs/models/providers/302ai.md +32 -1
- package/.docs/models/providers/alibaba-cn.md +2 -1
- package/.docs/models/providers/anthropic.md +2 -1
- package/.docs/models/providers/berget.md +9 -12
- package/.docs/models/providers/cloudflare-workers-ai.md +2 -1
- package/.docs/models/providers/cortecs.md +4 -1
- package/.docs/models/providers/deepinfra.md +4 -1
- package/.docs/models/providers/digitalocean.md +116 -0
- package/.docs/models/providers/fireworks-ai.md +2 -1
- package/.docs/models/providers/firmware.md +2 -3
- package/.docs/models/providers/helicone.md +1 -2
- package/.docs/models/providers/huggingface.md +2 -1
- package/.docs/models/providers/kilo.md +2 -1
- package/.docs/models/providers/kimi-for-coding.md +2 -1
- package/.docs/models/providers/llmgateway.md +59 -77
- package/.docs/models/providers/moonshotai-cn.md +3 -2
- package/.docs/models/providers/moonshotai.md +3 -2
- package/.docs/models/providers/nano-gpt.md +8 -1
- package/.docs/models/providers/nvidia.md +2 -1
- package/.docs/models/providers/ollama-cloud.md +2 -1
- package/.docs/models/providers/openai.md +1 -2
- package/.docs/models/providers/opencode-go.md +2 -1
- package/.docs/models/providers/opencode.md +5 -1
- package/.docs/models/providers/ovhcloud.md +4 -7
- package/.docs/models/providers/poe.md +2 -1
- package/.docs/models/providers/tencent-token-plan.md +71 -0
- package/.docs/models/providers/tencent-tokenhub.md +71 -0
- package/.docs/models/providers/wafer.ai.md +72 -0
- package/.docs/models/providers/zenmux.md +3 -1
- package/.docs/models/providers.md +4 -0
- package/.docs/reference/agents/generate.md +8 -0
- package/.docs/reference/client-js/workflows.md +12 -0
- package/.docs/reference/core/mastra-class.md +9 -1
- package/.docs/reference/deployer/cloudflare.md +14 -1
- package/.docs/reference/deployer/netlify.md +50 -2
- package/.docs/reference/harness/harness-class.md +72 -49
- package/.docs/reference/index.md +3 -0
- package/.docs/reference/observability/tracing/exporters/langfuse.md +2 -0
- package/.docs/reference/processors/prefill-error-handler.md +5 -5
- package/.docs/reference/storage/cloudflare-d1.md +42 -42
- package/.docs/reference/storage/redis.md +266 -0
- package/.docs/reference/streaming/agents/stream.md +8 -0
- package/.docs/reference/streaming/workflows/resumeStream.md +2 -0
- package/.docs/reference/tools/tavily.md +307 -0
- package/.docs/reference/workflows/run-methods/resume.md +24 -0
- package/.docs/reference/workflows/workflow-methods/foreach.md +14 -1
- package/.docs/reference/workspace/docker-sandbox.md +196 -0
- package/CHANGELOG.md +78 -0
- package/package.json +10 -10
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Anthropic
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 23 Anthropic models through Mastra's model router. Authentication is handled automatically using the `ANTHROPIC_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Anthropic documentation](https://docs.anthropic.com/en/docs/about-claude/models).
|
|
6
6
|
|
|
@@ -49,6 +49,7 @@ for await (const chunk of stream) {
|
|
|
49
49
|
| `anthropic/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
|
|
50
50
|
| `anthropic/claude-opus-4-5-20251101` | 200K | | | | | | $5 | $25 |
|
|
51
51
|
| `anthropic/claude-opus-4-6` | 1.0M | | | | | | $5 | $25 |
|
|
52
|
+
| `anthropic/claude-opus-4-7` | 1.0M | | | | | | $5 | $25 |
|
|
52
53
|
| `anthropic/claude-sonnet-4-0` | 200K | | | | | | $3 | $15 |
|
|
53
54
|
| `anthropic/claude-sonnet-4-20250514` | 200K | | | | | | $3 | $15 |
|
|
54
55
|
| `anthropic/claude-sonnet-4-5` | 200K | | | | | | $3 | $15 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Berget.AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 5 Berget.AI models through Mastra's model router. Authentication is handled automatically using the `BERGET_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Berget.AI documentation](https://api.berget.ai).
|
|
6
6
|
|
|
@@ -15,7 +15,7 @@ const agent = new Agent({
|
|
|
15
15
|
id: "my-agent",
|
|
16
16
|
name: "My Agent",
|
|
17
17
|
instructions: "You are a helpful assistant",
|
|
18
|
-
model: "berget/
|
|
18
|
+
model: "berget/google/gemma-4-31B-it"
|
|
19
19
|
});
|
|
20
20
|
|
|
21
21
|
// Generate a response
|
|
@@ -34,14 +34,11 @@ for await (const chunk of stream) {
|
|
|
34
34
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| ------------------------------------------------------ | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
-
| `berget/
|
|
38
|
-
| `berget/
|
|
39
|
-
| `berget/
|
|
40
|
-
| `berget/
|
|
41
|
-
| `berget/
|
|
42
|
-
| `berget/mistralai/Mistral-Small-3.2-24B-Instruct-2506` | 32K | | | | | | $0.30 | $0.30 |
|
|
43
|
-
| `berget/openai/gpt-oss-120b` | 128K | | | | | | $0.30 | $0.90 |
|
|
44
|
-
| `berget/zai-org/GLM-4.7` | 128K | | | | | | $0.70 | $2 |
|
|
37
|
+
| `berget/google/gemma-4-31B-it` | 128K | | | | | | $0.28 | $0.55 |
|
|
38
|
+
| `berget/meta-llama/Llama-3.3-70B-Instruct` | 128K | | | | | | $0.99 | $0.99 |
|
|
39
|
+
| `berget/mistralai/Mistral-Small-3.2-24B-Instruct-2506` | 32K | | | | | | $0.33 | $0.33 |
|
|
40
|
+
| `berget/openai/gpt-oss-120b` | 128K | | | | | | $0.44 | $0.99 |
|
|
41
|
+
| `berget/zai-org/GLM-4.7` | 128K | | | | | | $0.77 | $3 |
|
|
45
42
|
|
|
46
43
|
## Advanced configuration
|
|
47
44
|
|
|
@@ -53,7 +50,7 @@ const agent = new Agent({
|
|
|
53
50
|
name: "custom-agent",
|
|
54
51
|
model: {
|
|
55
52
|
url: "https://api.berget.ai/v1",
|
|
56
|
-
id: "berget/
|
|
53
|
+
id: "berget/google/gemma-4-31B-it",
|
|
57
54
|
apiKey: process.env.BERGET_API_KEY,
|
|
58
55
|
headers: {
|
|
59
56
|
"X-Custom-Header": "value"
|
|
@@ -72,7 +69,7 @@ const agent = new Agent({
|
|
|
72
69
|
const useAdvanced = requestContext.task === "complex";
|
|
73
70
|
return useAdvanced
|
|
74
71
|
? "berget/zai-org/GLM-4.7"
|
|
75
|
-
: "berget/
|
|
72
|
+
: "berget/google/gemma-4-31B-it";
|
|
76
73
|
}
|
|
77
74
|
});
|
|
78
75
|
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Cloudflare Workers AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 8 Cloudflare Workers AI models through Mastra's model router. Authentication is handled automatically using the `CLOUDFLARE_API_KEY` environment variable. Configure `CLOUDFLARE_ACCOUNT_ID` as well.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Cloudflare Workers AI documentation](https://developers.cloudflare.com/workers-ai/models/).
|
|
6
6
|
|
|
@@ -38,6 +38,7 @@ for await (const chunk of stream) {
|
|
|
38
38
|
| `cloudflare-workers-ai/@cf/google/gemma-4-26b-a4b-it` | 256K | | | | | | $0.10 | $0.30 |
|
|
39
39
|
| `cloudflare-workers-ai/@cf/meta/llama-4-scout-17b-16e-instruct` | 128K | | | | | | $0.27 | $0.85 |
|
|
40
40
|
| `cloudflare-workers-ai/@cf/moonshotai/kimi-k2.5` | 256K | | | | | | $0.60 | $3 |
|
|
41
|
+
| `cloudflare-workers-ai/@cf/moonshotai/kimi-k2.6` | 256K | | | | | | $0.95 | $4 |
|
|
41
42
|
| `cloudflare-workers-ai/@cf/nvidia/nemotron-3-120b-a12b` | 256K | | | | | | $0.50 | $2 |
|
|
42
43
|
| `cloudflare-workers-ai/@cf/openai/gpt-oss-120b` | 128K | | | | | | $0.35 | $0.75 |
|
|
43
44
|
| `cloudflare-workers-ai/@cf/openai/gpt-oss-20b` | 128K | | | | | | $0.20 | $0.30 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Cortecs
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 33 Cortecs models through Mastra's model router. Authentication is handled automatically using the `CORTECS_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Cortecs documentation](https://cortecs.ai).
|
|
6
6
|
|
|
@@ -39,6 +39,7 @@ for await (const chunk of stream) {
|
|
|
39
39
|
| `cortecs/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
|
|
40
40
|
| `cortecs/claude-opus4-5` | 200K | | | | | | $6 | $30 |
|
|
41
41
|
| `cortecs/claude-opus4-6` | 1.0M | | | | | | $6 | $30 |
|
|
42
|
+
| `cortecs/claude-opus4-7` | 1.0M | | | | | | $6 | $28 |
|
|
42
43
|
| `cortecs/claude-sonnet-4` | 200K | | | | | | $3 | $17 |
|
|
43
44
|
| `cortecs/deepseek-v3-0324` | 128K | | | | | | $0.55 | $2 |
|
|
44
45
|
| `cortecs/devstral-2512` | 262K | | | | | | — | — |
|
|
@@ -49,6 +50,7 @@ for await (const chunk of stream) {
|
|
|
49
50
|
| `cortecs/glm-4.7` | 198K | | | | | | $0.45 | $2 |
|
|
50
51
|
| `cortecs/glm-4.7-flash` | 203K | | | | | | $0.09 | $0.53 |
|
|
51
52
|
| `cortecs/glm-5` | 203K | | | | | | $1 | $3 |
|
|
53
|
+
| `cortecs/glm-5.1` | 205K | | | | | | $1 | $4 |
|
|
52
54
|
| `cortecs/gpt-4.1` | 1.0M | | | | | | $2 | $9 |
|
|
53
55
|
| `cortecs/gpt-oss-120b` | 128K | | | | | | — | — |
|
|
54
56
|
| `cortecs/intellect-3` | 128K | | | | | | $0.22 | $1 |
|
|
@@ -59,6 +61,7 @@ for await (const chunk of stream) {
|
|
|
59
61
|
| `cortecs/minimax-m2` | 400K | | | | | | $0.39 | $2 |
|
|
60
62
|
| `cortecs/minimax-m2.1` | 196K | | | | | | $0.34 | $1 |
|
|
61
63
|
| `cortecs/minimax-m2.5` | 197K | | | | | | $0.32 | $1 |
|
|
64
|
+
| `cortecs/minimax-M2.7` | 203K | | | | | | $0.47 | $1 |
|
|
62
65
|
| `cortecs/nova-pro-v1` | 300K | | | | | | $1 | $4 |
|
|
63
66
|
| `cortecs/qwen3-32b` | 16K | | | | | | $0.10 | $0.33 |
|
|
64
67
|
| `cortecs/qwen3-coder-480b-a35b-instruct` | 262K | | | | | | $0.44 | $2 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Deep Infra
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 31 Deep Infra models through Mastra's model router. Authentication is handled automatically using the `DEEPINFRA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Deep Infra documentation](https://deepinfra.com/models).
|
|
6
6
|
|
|
@@ -54,6 +54,9 @@ for await (const chunk of stream) {
|
|
|
54
54
|
| `deepinfra/openai/gpt-oss-20b` | 131K | | | | | | $0.03 | $0.14 |
|
|
55
55
|
| `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $0.40 | $2 |
|
|
56
56
|
| `deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo` | 262K | | | | | | $0.30 | $1 |
|
|
57
|
+
| `deepinfra/Qwen/Qwen3.5-35B-A3B` | 262K | | | | | | $0.20 | $0.95 |
|
|
58
|
+
| `deepinfra/Qwen/Qwen3.5-397B-A17B` | 262K | | | | | | $0.54 | $3 |
|
|
59
|
+
| `deepinfra/Qwen/Qwen3.6-35B-A3B` | 262K | | | | | | $0.20 | $1 |
|
|
57
60
|
| `deepinfra/zai-org/GLM-4.6` | 205K | | | | | | $0.43 | $2 |
|
|
58
61
|
| `deepinfra/zai-org/GLM-4.6V` | 205K | | | | | | $0.30 | $0.90 |
|
|
59
62
|
| `deepinfra/zai-org/GLM-4.7` | 203K | | | | | | $0.43 | $2 |
|
|
@@ -0,0 +1,116 @@
|
|
|
1
|
+
# DigitalOcean
|
|
2
|
+
|
|
3
|
+
Access 46 DigitalOcean models through Mastra's model router. Authentication is handled automatically using the `DIGITALOCEAN_ACCESS_TOKEN` environment variable.
|
|
4
|
+
|
|
5
|
+
Learn more in the [DigitalOcean documentation](https://docs.digitalocean.com/products/gradient-ai-platform/details/models/).
|
|
6
|
+
|
|
7
|
+
```bash
|
|
8
|
+
DIGITALOCEAN_ACCESS_TOKEN=your-api-token
|
|
9
|
+
```
|
|
10
|
+
|
|
11
|
+
```typescript
|
|
12
|
+
import { Agent } from "@mastra/core/agent";
|
|
13
|
+
|
|
14
|
+
const agent = new Agent({
|
|
15
|
+
id: "my-agent",
|
|
16
|
+
name: "My Agent",
|
|
17
|
+
instructions: "You are a helpful assistant",
|
|
18
|
+
model: "digitalocean/alibaba-qwen3-32b"
|
|
19
|
+
});
|
|
20
|
+
|
|
21
|
+
// Generate a response
|
|
22
|
+
const response = await agent.generate("Hello!");
|
|
23
|
+
|
|
24
|
+
// Stream a response
|
|
25
|
+
const stream = await agent.stream("Tell me a story");
|
|
26
|
+
for await (const chunk of stream) {
|
|
27
|
+
console.log(chunk);
|
|
28
|
+
}
|
|
29
|
+
```
|
|
30
|
+
|
|
31
|
+
> **Info:** Mastra uses the OpenAI-compatible `/chat/completions` endpoint. Some provider-specific features may not be available. Check the [DigitalOcean documentation](https://docs.digitalocean.com/products/gradient-ai-platform/details/models/) for details.
|
|
32
|
+
|
|
33
|
+
## Models
|
|
34
|
+
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ---------------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `digitalocean/alibaba-qwen3-32b` | 131K | | | | | | $0.25 | $0.55 |
|
|
38
|
+
| `digitalocean/all-mini-lm-l6-v2` | 256 | | | | | | $0.01 | — |
|
|
39
|
+
| `digitalocean/anthropic-claude-4.1-opus` | 200K | | | | | | $15 | $75 |
|
|
40
|
+
| `digitalocean/anthropic-claude-4.5-sonnet` | 1.0M | | | | | | $3 | $15 |
|
|
41
|
+
| `digitalocean/anthropic-claude-4.6-sonnet` | 1.0M | | | | | | $3 | $15 |
|
|
42
|
+
| `digitalocean/anthropic-claude-haiku-4.5` | 200K | | | | | | $1 | $5 |
|
|
43
|
+
| `digitalocean/anthropic-claude-opus-4` | 200K | | | | | | $15 | $75 |
|
|
44
|
+
| `digitalocean/anthropic-claude-opus-4.5` | 200K | | | | | | $5 | $25 |
|
|
45
|
+
| `digitalocean/anthropic-claude-opus-4.6` | 1.0M | | | | | | $5 | $25 |
|
|
46
|
+
| `digitalocean/anthropic-claude-opus-4.7` | 1.0M | | | | | | $5 | $25 |
|
|
47
|
+
| `digitalocean/anthropic-claude-sonnet-4` | 1.0M | | | | | | $3 | $15 |
|
|
48
|
+
| `digitalocean/arcee-trinity-large-thinking` | 256K | | | | | | $0.25 | $0.90 |
|
|
49
|
+
| `digitalocean/deepseek-r1-distill-llama-70b` | 131K | | | | | | $0.99 | $0.99 |
|
|
50
|
+
| `digitalocean/fal-ai/elevenlabs/tts/multilingual-v2` | — | | | | | | — | — |
|
|
51
|
+
| `digitalocean/fal-ai/fast-sdxl` | — | | | | | | — | — |
|
|
52
|
+
| `digitalocean/fal-ai/flux/schnell` | — | | | | | | — | — |
|
|
53
|
+
| `digitalocean/fal-ai/stable-audio-25/text-to-audio` | — | | | | | | — | — |
|
|
54
|
+
| `digitalocean/glm-5` | 203K | | | | | | $1 | $3 |
|
|
55
|
+
| `digitalocean/gte-large-en-v1.5` | 8K | | | | | | $0.09 | — |
|
|
56
|
+
| `digitalocean/kimi-k2.5` | 262K | | | | | | $0.50 | $3 |
|
|
57
|
+
| `digitalocean/llama3.3-70b-instruct` | 128K | | | | | | $0.65 | $0.65 |
|
|
58
|
+
| `digitalocean/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
|
|
59
|
+
| `digitalocean/multi-qa-mpnet-base-dot-v1` | 512 | | | | | | $0.01 | — |
|
|
60
|
+
| `digitalocean/nvidia-nemotron-3-super-120b` | 256K | | | | | | $0.30 | $0.65 |
|
|
61
|
+
| `digitalocean/openai-gpt-4.1` | 1.0M | | | | | | $2 | $8 |
|
|
62
|
+
| `digitalocean/openai-gpt-4o` | 128K | | | | | | $3 | $10 |
|
|
63
|
+
| `digitalocean/openai-gpt-4o-mini` | 128K | | | | | | $0.15 | $0.60 |
|
|
64
|
+
| `digitalocean/openai-gpt-5` | 400K | | | | | | $1 | $10 |
|
|
65
|
+
| `digitalocean/openai-gpt-5-2-pro` | 400K | | | | | | $21 | $168 |
|
|
66
|
+
| `digitalocean/openai-gpt-5-mini` | 400K | | | | | | $0.25 | $2 |
|
|
67
|
+
| `digitalocean/openai-gpt-5-nano` | 400K | | | | | | $0.05 | $0.40 |
|
|
68
|
+
| `digitalocean/openai-gpt-5.1-codex-max` | 400K | | | | | | $1 | $10 |
|
|
69
|
+
| `digitalocean/openai-gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
70
|
+
| `digitalocean/openai-gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
|
|
71
|
+
| `digitalocean/openai-gpt-5.4` | 1.0M | | | | | | $3 | $15 |
|
|
72
|
+
| `digitalocean/openai-gpt-5.4-mini` | 400K | | | | | | $0.75 | $5 |
|
|
73
|
+
| `digitalocean/openai-gpt-5.4-nano` | 400K | | | | | | $0.20 | $1 |
|
|
74
|
+
| `digitalocean/openai-gpt-5.4-pro` | 400K | | | | | | $30 | $180 |
|
|
75
|
+
| `digitalocean/openai-gpt-image-1` | — | | | | | | $5 | $40 |
|
|
76
|
+
| `digitalocean/openai-gpt-image-1.5` | — | | | | | | $5 | $10 |
|
|
77
|
+
| `digitalocean/openai-gpt-oss-120b` | 131K | | | | | | $0.10 | $0.70 |
|
|
78
|
+
| `digitalocean/openai-gpt-oss-20b` | 131K | | | | | | $0.05 | $0.45 |
|
|
79
|
+
| `digitalocean/openai-o1` | 200K | | | | | | $15 | $60 |
|
|
80
|
+
| `digitalocean/openai-o3` | 200K | | | | | | $2 | $8 |
|
|
81
|
+
| `digitalocean/openai-o3-mini` | 200K | | | | | | $1 | $4 |
|
|
82
|
+
| `digitalocean/qwen3-embedding-0.6b` | 8K | | | | | | $0.04 | — |
|
|
83
|
+
|
|
84
|
+
## Advanced configuration
|
|
85
|
+
|
|
86
|
+
### Custom headers
|
|
87
|
+
|
|
88
|
+
```typescript
|
|
89
|
+
const agent = new Agent({
|
|
90
|
+
id: "custom-agent",
|
|
91
|
+
name: "custom-agent",
|
|
92
|
+
model: {
|
|
93
|
+
url: "https://inference.do-ai.run/v1",
|
|
94
|
+
id: "digitalocean/alibaba-qwen3-32b",
|
|
95
|
+
apiKey: process.env.DIGITALOCEAN_ACCESS_TOKEN,
|
|
96
|
+
headers: {
|
|
97
|
+
"X-Custom-Header": "value"
|
|
98
|
+
}
|
|
99
|
+
}
|
|
100
|
+
});
|
|
101
|
+
```
|
|
102
|
+
|
|
103
|
+
### Dynamic model selection
|
|
104
|
+
|
|
105
|
+
```typescript
|
|
106
|
+
const agent = new Agent({
|
|
107
|
+
id: "dynamic-agent",
|
|
108
|
+
name: "Dynamic Agent",
|
|
109
|
+
model: ({ requestContext }) => {
|
|
110
|
+
const useAdvanced = requestContext.task === "complex";
|
|
111
|
+
return useAdvanced
|
|
112
|
+
? "digitalocean/qwen3-embedding-0.6b"
|
|
113
|
+
: "digitalocean/alibaba-qwen3-32b";
|
|
114
|
+
}
|
|
115
|
+
});
|
|
116
|
+
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Fireworks AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 18 Fireworks AI models through Mastra's model router. Authentication is handled automatically using the `FIREWORKS_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Fireworks AI documentation](https://fireworks.ai/docs/).
|
|
6
6
|
|
|
@@ -46,6 +46,7 @@ for await (const chunk of stream) {
|
|
|
46
46
|
| `fireworks-ai/accounts/fireworks/models/kimi-k2-instruct` | 128K | | | | | | $1 | $3 |
|
|
47
47
|
| `fireworks-ai/accounts/fireworks/models/kimi-k2-thinking` | 256K | | | | | | $0.60 | $3 |
|
|
48
48
|
| `fireworks-ai/accounts/fireworks/models/kimi-k2p5` | 256K | | | | | | $0.60 | $3 |
|
|
49
|
+
| `fireworks-ai/accounts/fireworks/models/kimi-k2p6` | 262K | | | | | | $0.95 | $4 |
|
|
49
50
|
| `fireworks-ai/accounts/fireworks/models/minimax-m2p1` | 200K | | | | | | $0.30 | $1 |
|
|
50
51
|
| `fireworks-ai/accounts/fireworks/models/minimax-m2p5` | 197K | | | | | | $0.30 | $1 |
|
|
51
52
|
| `fireworks-ai/accounts/fireworks/models/minimax-m2p7` | 197K | | | | | | $0.30 | $1 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Firmware
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 24 Firmware models through Mastra's model router. Authentication is handled automatically using the `FIRMWARE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Firmware documentation](https://docs.frogbot.ai).
|
|
6
6
|
|
|
@@ -35,9 +35,8 @@ for await (const chunk of stream) {
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| -------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
37
|
| `firmware/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
|
|
38
|
-
| `firmware/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
|
|
39
38
|
| `firmware/claude-opus-4-6` | 200K | | | | | | $5 | $25 |
|
|
40
|
-
| `firmware/claude-
|
|
39
|
+
| `firmware/claude-opus-4-7` | 200K | | | | | | $5 | $25 |
|
|
41
40
|
| `firmware/claude-sonnet-4-6` | 200K | | | | | | $3 | $15 |
|
|
42
41
|
| `firmware/deepseek-v3-2` | 128K | | | | | | $0.58 | $2 |
|
|
43
42
|
| `firmware/gemini-2.5-flash` | 1.0M | | | | | | $0.30 | $3 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Helicone
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 90 Helicone models through Mastra's model router. Authentication is handled automatically using the `HELICONE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Helicone documentation](https://helicone.ai/models).
|
|
6
6
|
|
|
@@ -48,7 +48,6 @@ for await (const chunk of stream) {
|
|
|
48
48
|
| `helicone/claude-opus-4-1-20250805` | 200K | | | | | | $15 | $75 |
|
|
49
49
|
| `helicone/claude-sonnet-4` | 200K | | | | | | $3 | $15 |
|
|
50
50
|
| `helicone/claude-sonnet-4-5-20250929` | 200K | | | | | | $3 | $15 |
|
|
51
|
-
| `helicone/codex-mini-latest` | 200K | | | | | | $2 | $6 |
|
|
52
51
|
| `helicone/deepseek-r1-distill-llama-70b` | 128K | | | | | | $0.03 | $0.13 |
|
|
53
52
|
| `helicone/deepseek-reasoner` | 128K | | | | | | $0.56 | $2 |
|
|
54
53
|
| `helicone/deepseek-tng-r1t2-chimera` | 130K | | | | | | $0.30 | $1 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Hugging Face
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 23 Hugging Face models through Mastra's model router. Authentication is handled automatically using the `HF_TOKEN` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Hugging Face documentation](https://huggingface.co).
|
|
6
6
|
|
|
@@ -43,6 +43,7 @@ for await (const chunk of stream) {
|
|
|
43
43
|
| `huggingface/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $1 | $3 |
|
|
44
44
|
| `huggingface/moonshotai/Kimi-K2-Thinking` | 262K | | | | | | $0.60 | $3 |
|
|
45
45
|
| `huggingface/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.60 | $3 |
|
|
46
|
+
| `huggingface/moonshotai/Kimi-K2.6` | 262K | | | | | | $0.95 | $4 |
|
|
46
47
|
| `huggingface/Qwen/Qwen3-235B-A22B-Thinking-2507` | 262K | | | | | | $0.30 | $3 |
|
|
47
48
|
| `huggingface/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $2 | $2 |
|
|
48
49
|
| `huggingface/Qwen/Qwen3-Coder-Next` | 262K | | | | | | $0.20 | $2 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Kilo Gateway
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 335 Kilo Gateway models through Mastra's model router. Authentication is handled automatically using the `KILO_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Kilo Gateway documentation](https://kilo.ai).
|
|
6
6
|
|
|
@@ -189,6 +189,7 @@ for await (const chunk of stream) {
|
|
|
189
189
|
| `kilo/moonshotai/kimi-k2-0905` | 131K | | | | | | $0.40 | $2 |
|
|
190
190
|
| `kilo/moonshotai/kimi-k2-thinking` | 131K | | | | | | $0.47 | $2 |
|
|
191
191
|
| `kilo/moonshotai/kimi-k2.5` | 262K | | | | | | $0.45 | $2 |
|
|
192
|
+
| `kilo/moonshotai/kimi-k2.6` | 262K | | | | | | $0.95 | $4 |
|
|
192
193
|
| `kilo/morph/morph-v3-fast` | 82K | | | | | | $0.80 | $1 |
|
|
193
194
|
| `kilo/morph/morph-v3-large` | 262K | | | | | | $0.90 | $2 |
|
|
194
195
|
| `kilo/nex-agi/deepseek-v3.1-nex-n1` | 131K | | | | | | $0.27 | $1 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Kimi For Coding
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 3 Kimi For Coding models through Mastra's model router. Authentication is handled automatically using the `KIMI_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Kimi For Coding documentation](https://www.kimi.com/coding/docs/en/third-party-agents.html).
|
|
6
6
|
|
|
@@ -35,6 +35,7 @@ for await (const chunk of stream) {
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| ---------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
37
|
| `kimi-for-coding/k2p5` | 262K | | | | | | — | — |
|
|
38
|
+
| `kimi-for-coding/k2p6` | 262K | | | | | | — | — |
|
|
38
39
|
| `kimi-for-coding/kimi-k2-thinking` | 262K | | | | | | — | — |
|
|
39
40
|
|
|
40
41
|
## Advanced configuration
|