@mastra/mcp-docs-server 1.1.8 → 1.1.9-alpha.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.docs/docs/agents/agent-memory.md +2 -2
- package/.docs/docs/agents/guardrails.md +3 -3
- package/.docs/docs/agents/network-approval.md +4 -1
- package/.docs/docs/agents/networks.md +1 -1
- package/.docs/docs/agents/overview.md +1 -1
- package/.docs/docs/agents/processors.md +35 -17
- package/.docs/docs/agents/structured-output.md +1 -1
- package/.docs/docs/agents/using-tools.md +2 -2
- package/.docs/docs/build-with-ai/mcp-docs-server.md +4 -4
- package/.docs/docs/build-with-ai/skills.md +1 -1
- package/.docs/docs/community/discord.md +1 -1
- package/.docs/docs/community/licensing.md +1 -1
- package/.docs/docs/deployment/mastra-server.md +1 -1
- package/.docs/docs/deployment/studio.md +2 -2
- package/.docs/docs/deployment/web-framework.md +1 -1
- package/.docs/docs/evals/overview.md +1 -1
- package/.docs/docs/getting-started/build-with-ai.md +1 -1
- package/.docs/docs/getting-started/project-structure.md +1 -1
- package/.docs/docs/index.md +62 -16
- package/.docs/docs/mastra-cloud/deployment.md +1 -1
- package/.docs/docs/mastra-cloud/studio.md +1 -1
- package/.docs/docs/mcp/publishing-mcp-server.md +1 -1
- package/.docs/docs/memory/memory-processors.md +1 -1
- package/.docs/docs/memory/message-history.md +2 -2
- package/.docs/docs/memory/observational-memory.md +6 -2
- package/.docs/docs/memory/semantic-recall.md +2 -2
- package/.docs/docs/memory/storage.md +1 -1
- package/.docs/docs/memory/working-memory.md +6 -6
- package/.docs/docs/observability/tracing/bridges/otel.md +2 -2
- package/.docs/docs/observability/tracing/exporters/default.md +1 -1
- package/.docs/docs/observability/tracing/overview.md +4 -4
- package/.docs/docs/observability/tracing/processors/sensitive-data-filter.md +2 -2
- package/.docs/docs/rag/chunking-and-embedding.md +1 -1
- package/.docs/docs/rag/vector-databases.md +2 -2
- package/.docs/docs/server/auth/auth0.md +1 -1
- package/.docs/docs/server/auth/firebase.md +1 -1
- package/.docs/docs/server/auth/simple-auth.md +1 -1
- package/.docs/docs/server/auth.md +1 -1
- package/.docs/docs/server/mastra-client.md +1 -1
- package/.docs/docs/server/mastra-server.md +1 -1
- package/.docs/docs/server/server-adapters.md +2 -2
- package/.docs/docs/streaming/events.md +1 -1
- package/.docs/docs/streaming/overview.md +1 -1
- package/.docs/docs/streaming/tool-streaming.md +44 -30
- package/.docs/docs/streaming/workflow-streaming.md +1 -1
- package/.docs/docs/workflows/control-flow.md +44 -2
- package/.docs/docs/workflows/error-handling.md +1 -1
- package/.docs/docs/workflows/overview.md +3 -3
- package/.docs/docs/workflows/snapshots.md +1 -1
- package/.docs/docs/workflows/time-travel.md +2 -2
- package/.docs/docs/workspace/filesystem.md +2 -2
- package/.docs/docs/workspace/overview.md +52 -7
- package/.docs/docs/workspace/sandbox.md +72 -13
- package/.docs/docs/workspace/skills.md +2 -2
- package/.docs/guides/build-your-ui/copilotkit.md +1 -1
- package/.docs/guides/deployment/inngest.md +4 -4
- package/.docs/guides/guide/ai-recruiter.md +1 -1
- package/.docs/guides/guide/github-actions-pr-description.md +2 -2
- package/.docs/guides/guide/notes-mcp-server.md +1 -1
- package/.docs/guides/guide/stock-agent.md +2 -2
- package/.docs/guides/migrations/agentnetwork.md +1 -1
- package/.docs/guides/migrations/upgrade-to-v1/client.md +2 -2
- package/.docs/guides/migrations/upgrade-to-v1/deployment.md +1 -1
- package/.docs/guides/migrations/upgrade-to-v1/memory.md +2 -2
- package/.docs/guides/migrations/upgrade-to-v1/storage.md +1 -1
- package/.docs/guides/migrations/upgrade-to-v1/tools.md +2 -2
- package/.docs/guides/migrations/upgrade-to-v1/workflows.md +5 -5
- package/.docs/guides/migrations/vnext-to-standard-apis.md +2 -2
- package/.docs/models/gateways/netlify.md +1 -2
- package/.docs/models/gateways/openrouter.md +8 -1
- package/.docs/models/gateways/vercel.md +3 -1
- package/.docs/models/index.md +1 -1
- package/.docs/models/providers/abacus.md +21 -11
- package/.docs/models/providers/aihubmix.md +7 -2
- package/.docs/models/providers/alibaba-cn.md +80 -71
- package/.docs/models/providers/alibaba-coding-plan-cn.md +78 -0
- package/.docs/models/providers/alibaba-coding-plan.md +78 -0
- package/.docs/models/providers/chutes.md +1 -1
- package/.docs/models/providers/clarifai.md +81 -0
- package/.docs/models/providers/cloudferro-sherlock.md +5 -4
- package/.docs/models/providers/cloudflare-workers-ai.md +3 -2
- package/.docs/models/providers/cortecs.md +7 -5
- package/.docs/models/providers/deepinfra.md +7 -2
- package/.docs/models/providers/deepseek.md +1 -1
- package/.docs/models/providers/drun.md +73 -0
- package/.docs/models/providers/firmware.md +28 -20
- package/.docs/models/providers/google.md +3 -1
- package/.docs/models/providers/inception.md +4 -2
- package/.docs/models/providers/kilo.md +3 -1
- package/.docs/models/providers/nano-gpt.md +519 -40
- package/.docs/models/providers/nebius.md +34 -34
- package/.docs/models/providers/nvidia.md +4 -2
- package/.docs/models/providers/ollama-cloud.md +1 -2
- package/.docs/models/providers/openai.md +3 -1
- package/.docs/models/providers/opencode.md +36 -33
- package/.docs/models/providers/poe.md +8 -2
- package/.docs/models/providers/qiniu-ai.md +20 -5
- package/.docs/models/providers/requesty.md +17 -1
- package/.docs/models/providers/siliconflow-cn.md +7 -1
- package/.docs/models/providers/togetherai.md +1 -3
- package/.docs/models/providers/xai.md +28 -25
- package/.docs/models/providers/xiaomi.md +1 -1
- package/.docs/models/providers/zenmux.md +3 -1
- package/.docs/models/providers.md +4 -0
- package/.docs/reference/agents/getDefaultGenerateOptions.md +1 -1
- package/.docs/reference/agents/getDefaultOptions.md +1 -1
- package/.docs/reference/agents/getDefaultStreamOptions.md +1 -1
- package/.docs/reference/agents/getDescription.md +1 -1
- package/.docs/reference/agents/network.md +3 -1
- package/.docs/reference/ai-sdk/handle-chat-stream.md +2 -0
- package/.docs/reference/ai-sdk/handle-network-stream.md +2 -0
- package/.docs/reference/ai-sdk/network-route.md +2 -0
- package/.docs/reference/ai-sdk/to-ai-sdk-v4-messages.md +1 -1
- package/.docs/reference/ai-sdk/to-ai-sdk-v5-messages.md +1 -1
- package/.docs/reference/auth/auth0.md +3 -3
- package/.docs/reference/auth/firebase.md +1 -1
- package/.docs/reference/auth/workos.md +2 -2
- package/.docs/reference/cli/mastra.md +4 -4
- package/.docs/reference/client-js/mastra-client.md +1 -1
- package/.docs/reference/configuration.md +62 -6
- package/.docs/reference/core/getDeployer.md +1 -1
- package/.docs/reference/core/getLogger.md +1 -1
- package/.docs/reference/core/getScorer.md +2 -2
- package/.docs/reference/core/getServer.md +1 -1
- package/.docs/reference/core/getStorage.md +1 -1
- package/.docs/reference/core/getStoredAgentById.md +1 -1
- package/.docs/reference/core/getTelemetry.md +1 -1
- package/.docs/reference/core/getWorkflow.md +1 -1
- package/.docs/reference/core/listAgents.md +1 -1
- package/.docs/reference/core/listMCPServers.md +1 -1
- package/.docs/reference/core/listStoredAgents.md +1 -1
- package/.docs/reference/core/listVectors.md +1 -1
- package/.docs/reference/core/mastra-class.md +1 -1
- package/.docs/reference/core/setLogger.md +1 -1
- package/.docs/reference/core/setStorage.md +1 -1
- package/.docs/reference/datasets/dataset.md +1 -1
- package/.docs/reference/datasets/datasets-manager.md +1 -1
- package/.docs/reference/datasets/get.md +2 -2
- package/.docs/reference/datasets/getDetails.md +1 -1
- package/.docs/reference/datasets/listItems.md +1 -1
- package/.docs/reference/deployer/vercel.md +1 -1
- package/.docs/reference/evals/answer-relevancy.md +1 -1
- package/.docs/reference/evals/completeness.md +1 -1
- package/.docs/reference/evals/context-precision.md +3 -3
- package/.docs/reference/evals/context-relevance.md +1 -1
- package/.docs/reference/evals/hallucination.md +3 -9
- package/.docs/reference/evals/keyword-coverage.md +1 -1
- package/.docs/reference/evals/mastra-scorer.md +1 -1
- package/.docs/reference/evals/noise-sensitivity.md +2 -2
- package/.docs/reference/evals/textual-difference.md +1 -1
- package/.docs/reference/evals/tone-consistency.md +1 -1
- package/.docs/reference/evals/tool-call-accuracy.md +1 -1
- package/.docs/reference/harness/harness-class.md +4 -2
- package/.docs/reference/index.md +2 -0
- package/.docs/reference/memory/cloneThread.md +1 -1
- package/.docs/reference/memory/observational-memory.md +7 -5
- package/.docs/reference/observability/tracing/bridges/otel.md +1 -1
- package/.docs/reference/observability/tracing/processors/sensitive-data-filter.md +1 -1
- package/.docs/reference/observability/tracing/spans.md +2 -0
- package/.docs/reference/processors/message-history-processor.md +1 -1
- package/.docs/reference/processors/processor-interface.md +6 -2
- package/.docs/reference/processors/token-limiter-processor.md +2 -2
- package/.docs/reference/rag/metadata-filters.md +10 -10
- package/.docs/reference/server/create-route.md +2 -0
- package/.docs/reference/server/koa-adapter.md +1 -1
- package/.docs/reference/server/register-api-route.md +2 -2
- package/.docs/reference/storage/cloudflare-d1.md +3 -3
- package/.docs/reference/storage/cloudflare.md +1 -1
- package/.docs/reference/storage/convex.md +1 -1
- package/.docs/reference/storage/dynamodb.md +2 -2
- package/.docs/reference/storage/lance.md +2 -2
- package/.docs/reference/storage/mongodb.md +1 -1
- package/.docs/reference/storage/mssql.md +1 -1
- package/.docs/reference/storage/postgresql.md +2 -2
- package/.docs/reference/storage/upstash.md +1 -1
- package/.docs/reference/streaming/workflows/observeStream.md +1 -1
- package/.docs/reference/templates/overview.md +1 -1
- package/.docs/reference/tools/create-tool.md +1 -1
- package/.docs/reference/tools/mcp-server.md +4 -4
- package/.docs/reference/vectors/chroma.md +2 -2
- package/.docs/reference/vectors/couchbase.md +6 -6
- package/.docs/reference/vectors/pg.md +2 -0
- package/.docs/reference/vectors/s3vectors.md +5 -5
- package/.docs/reference/voice/azure.md +4 -2
- package/.docs/reference/voice/cloudflare.md +4 -2
- package/.docs/reference/voice/elevenlabs.md +1 -1
- package/.docs/reference/voice/google-gemini-live.md +2 -2
- package/.docs/reference/voice/google.md +3 -3
- package/.docs/reference/voice/mastra-voice.md +1 -1
- package/.docs/reference/voice/murf.md +2 -2
- package/.docs/reference/voice/openai-realtime.md +3 -1
- package/.docs/reference/voice/openai.md +7 -3
- package/.docs/reference/voice/playai.md +4 -2
- package/.docs/reference/voice/sarvam.md +3 -1
- package/.docs/reference/voice/speechify.md +6 -4
- package/.docs/reference/voice/voice.addInstructions.md +2 -2
- package/.docs/reference/voice/voice.addTools.md +1 -1
- package/.docs/reference/voice/voice.close.md +2 -2
- package/.docs/reference/voice/voice.connect.md +4 -2
- package/.docs/reference/voice/voice.events.md +2 -2
- package/.docs/reference/voice/voice.getSpeakers.md +1 -1
- package/.docs/reference/voice/voice.listen.md +11 -5
- package/.docs/reference/voice/voice.off.md +2 -2
- package/.docs/reference/voice/voice.on.md +2 -2
- package/.docs/reference/voice/voice.speak.md +14 -4
- package/.docs/reference/voice/voice.updateConfig.md +1 -1
- package/.docs/reference/workflows/run-methods/timeTravel.md +1 -1
- package/.docs/reference/workspace/blaxel-sandbox.md +164 -0
- package/.docs/reference/workspace/daytona-sandbox.md +48 -139
- package/.docs/reference/workspace/e2b-sandbox.md +39 -75
- package/.docs/reference/workspace/filesystem.md +24 -10
- package/.docs/reference/workspace/gcs-filesystem.md +20 -0
- package/.docs/reference/workspace/local-filesystem.md +23 -9
- package/.docs/reference/workspace/local-sandbox.md +23 -98
- package/.docs/reference/workspace/process-manager.md +296 -0
- package/.docs/reference/workspace/s3-filesystem.md +20 -0
- package/.docs/reference/workspace/sandbox.md +9 -1
- package/.docs/reference/workspace/workspace-class.md +93 -25
- package/CHANGELOG.md +8 -0
- package/dist/tools/course.d.ts +7 -27
- package/dist/tools/course.d.ts.map +1 -1
- package/dist/tools/docs.d.ts +6 -18
- package/dist/tools/docs.d.ts.map +1 -1
- package/dist/tools/embedded-docs.d.ts +12 -112
- package/dist/tools/embedded-docs.d.ts.map +1 -1
- package/dist/tools/migration.d.ts +6 -26
- package/dist/tools/migration.d.ts.map +1 -1
- package/package.json +7 -7
|
@@ -38,48 +38,48 @@ for await (const chunk of stream) {
|
|
|
38
38
|
| `nebius/BAAI/bge-multilingual-gemma2` | 8K | | | | | | $0.01 | — |
|
|
39
39
|
| `nebius/black-forest-labs/flux-dev` | 77 | | | | | | — | — |
|
|
40
40
|
| `nebius/black-forest-labs/flux-schnell` | 77 | | | | | | — | — |
|
|
41
|
-
| `nebius/deepseek-ai/
|
|
42
|
-
| `nebius/deepseek-ai/
|
|
43
|
-
| `nebius/deepseek-ai/
|
|
44
|
-
| `nebius/deepseek-ai/
|
|
45
|
-
| `nebius/deepseek-ai/
|
|
41
|
+
| `nebius/deepseek-ai/DeepSeek-R1-0528` | 128K | | | | | | $0.80 | $2 |
|
|
42
|
+
| `nebius/deepseek-ai/DeepSeek-R1-0528-fast` | 131K | | | | | | $2 | $6 |
|
|
43
|
+
| `nebius/deepseek-ai/DeepSeek-V3-0324` | 128K | | | | | | $0.50 | $2 |
|
|
44
|
+
| `nebius/deepseek-ai/DeepSeek-V3-0324-fast` | 128K | | | | | | $0.75 | $2 |
|
|
45
|
+
| `nebius/deepseek-ai/DeepSeek-V3.2` | 128K | | | | | | $0.30 | $0.45 |
|
|
46
46
|
| `nebius/google/gemma-2-2b-it` | 8K | | | | | | $0.02 | $0.06 |
|
|
47
47
|
| `nebius/google/gemma-2-9b-it-fast` | 8K | | | | | | $0.03 | $0.09 |
|
|
48
48
|
| `nebius/google/gemma-3-27b-it` | 128K | | | | | | $0.10 | $0.30 |
|
|
49
49
|
| `nebius/google/gemma-3-27b-it-fast` | 128K | | | | | | $0.20 | $0.60 |
|
|
50
50
|
| `nebius/intfloat/e5-mistral-7b-instruct` | 33K | | | | | | $0.01 | — |
|
|
51
51
|
| `nebius/meta-llama/Llama-3.3-70B-Instruct` | 128K | | | | | | $0.13 | $0.40 |
|
|
52
|
-
| `nebius/meta-llama/
|
|
53
|
-
| `nebius/meta-llama/
|
|
54
|
-
| `nebius/meta-llama/
|
|
55
|
-
| `nebius/meta-llama/
|
|
56
|
-
| `nebius/MiniMaxAI/
|
|
57
|
-
| `nebius/moonshotai/
|
|
58
|
-
| `nebius/moonshotai/
|
|
52
|
+
| `nebius/meta-llama/Llama-3.3-70B-Instruct-fast` | 128K | | | | | | $0.25 | $0.75 |
|
|
53
|
+
| `nebius/meta-llama/Llama-Guard-3-8B` | 8K | | | | | | $0.02 | $0.06 |
|
|
54
|
+
| `nebius/meta-llama/Meta-Llama-3.1-8B-Instruct` | 128K | | | | | | $0.02 | $0.06 |
|
|
55
|
+
| `nebius/meta-llama/Meta-Llama-3.1-8B-Instruct-fast` | 128K | | | | | | $0.03 | $0.09 |
|
|
56
|
+
| `nebius/MiniMaxAI/MiniMax-M2.1` | 128K | | | | | | $0.30 | $1 |
|
|
57
|
+
| `nebius/moonshotai/Kimi-K2-Instruct` | 200K | | | | | | $0.50 | $2 |
|
|
58
|
+
| `nebius/moonshotai/Kimi-K2-Thinking` | 128K | | | | | | $0.60 | $3 |
|
|
59
59
|
| `nebius/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.50 | $3 |
|
|
60
|
-
| `nebius/NousResearch/
|
|
61
|
-
| `nebius/NousResearch/
|
|
62
|
-
| `nebius/nvidia/
|
|
63
|
-
| `nebius/nvidia/
|
|
64
|
-
| `nebius/nvidia/
|
|
60
|
+
| `nebius/NousResearch/Hermes-4-405B` | 128K | | | | | | $1 | $3 |
|
|
61
|
+
| `nebius/NousResearch/Hermes-4-70B` | 128K | | | | | | $0.13 | $0.40 |
|
|
62
|
+
| `nebius/nvidia/Llama-3_1-Nemotron-Ultra-253B-v1` | 128K | | | | | | $0.60 | $2 |
|
|
63
|
+
| `nebius/nvidia/Nemotron-Nano-V2-12b` | 32K | | | | | | $0.07 | $0.20 |
|
|
64
|
+
| `nebius/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B` | 32K | | | | | | $0.06 | $0.24 |
|
|
65
65
|
| `nebius/openai/gpt-oss-120b` | 128K | | | | | | $0.15 | $0.60 |
|
|
66
66
|
| `nebius/openai/gpt-oss-20b` | 128K | | | | | | $0.05 | $0.20 |
|
|
67
|
-
| `nebius/PrimeIntellect/
|
|
68
|
-
| `nebius/
|
|
69
|
-
| `nebius/
|
|
70
|
-
| `nebius/
|
|
71
|
-
| `nebius/
|
|
72
|
-
| `nebius/
|
|
73
|
-
| `nebius/
|
|
74
|
-
| `nebius/
|
|
75
|
-
| `nebius/
|
|
76
|
-
| `nebius/
|
|
77
|
-
| `nebius/
|
|
78
|
-
| `nebius/
|
|
79
|
-
| `nebius/
|
|
80
|
-
| `nebius/zai-org/
|
|
81
|
-
| `nebius/zai-org/
|
|
82
|
-
| `nebius/zai-org/
|
|
67
|
+
| `nebius/PrimeIntellect/INTELLECT-3` | 128K | | | | | | $0.20 | $1 |
|
|
68
|
+
| `nebius/Qwen/Qwen2.5-Coder-7B-fast` | 128K | | | | | | $0.03 | $0.09 |
|
|
69
|
+
| `nebius/Qwen/Qwen2.5-VL-72B-Instruct` | 128K | | | | | | $0.25 | $0.75 |
|
|
70
|
+
| `nebius/Qwen/Qwen3-235B-A22B-Instruct-2507` | 262K | | | | | | $0.20 | $0.60 |
|
|
71
|
+
| `nebius/Qwen/Qwen3-235B-A22B-Thinking-2507` | 262K | | | | | | $0.20 | $0.80 |
|
|
72
|
+
| `nebius/Qwen/Qwen3-30B-A3B-Instruct-2507` | 128K | | | | | | $0.10 | $0.30 |
|
|
73
|
+
| `nebius/Qwen/Qwen3-30B-A3B-Thinking-2507` | 128K | | | | | | $0.10 | $0.30 |
|
|
74
|
+
| `nebius/Qwen/Qwen3-32B` | 128K | | | | | | $0.10 | $0.30 |
|
|
75
|
+
| `nebius/Qwen/Qwen3-32B-fast` | 128K | | | | | | $0.20 | $0.60 |
|
|
76
|
+
| `nebius/Qwen/Qwen3-Coder-30B-A3B-Instruct` | 128K | | | | | | $0.10 | $0.30 |
|
|
77
|
+
| `nebius/Qwen/Qwen3-Coder-480B-A35B-Instruct` | 262K | | | | | | $0.40 | $2 |
|
|
78
|
+
| `nebius/Qwen/Qwen3-Embedding-8B` | 33K | | | | | | $0.01 | — |
|
|
79
|
+
| `nebius/Qwen/Qwen3-Next-80B-A3B-Thinking` | 128K | | | | | | $0.15 | $1 |
|
|
80
|
+
| `nebius/zai-org/GLM-4.5` | 128K | | | | | | $0.60 | $2 |
|
|
81
|
+
| `nebius/zai-org/GLM-4.5-Air` | 128K | | | | | | $0.20 | $1 |
|
|
82
|
+
| `nebius/zai-org/GLM-4.7-FP8` | 128K | | | | | | $0.40 | $2 |
|
|
83
83
|
|
|
84
84
|
## Advanced Configuration
|
|
85
85
|
|
|
@@ -109,7 +109,7 @@ const agent = new Agent({
|
|
|
109
109
|
model: ({ requestContext }) => {
|
|
110
110
|
const useAdvanced = requestContext.task === "complex";
|
|
111
111
|
return useAdvanced
|
|
112
|
-
? "nebius/zai-org/
|
|
112
|
+
? "nebius/zai-org/GLM-4.7-FP8"
|
|
113
113
|
: "nebius/BAAI/bge-en-icl";
|
|
114
114
|
}
|
|
115
115
|
});
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Nvidia
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 73 Nvidia models through Mastra's model router. Authentication is handled automatically using the `NVIDIA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Nvidia documentation](https://docs.api.nvidia.com/nim/).
|
|
6
6
|
|
|
@@ -68,8 +68,8 @@ for await (const chunk of stream) {
|
|
|
68
68
|
| `nvidia/microsoft/phi-3.5-moe-instruct` | 128K | | | | | | — | — |
|
|
69
69
|
| `nvidia/microsoft/phi-3.5-vision-instruct` | 128K | | | | | | — | — |
|
|
70
70
|
| `nvidia/microsoft/phi-4-mini-instruct` | 131K | | | | | | — | — |
|
|
71
|
-
| `nvidia/minimaxai/minimax-m2` | 128K | | | | | | — | — |
|
|
72
71
|
| `nvidia/minimaxai/minimax-m2.1` | 205K | | | | | | — | — |
|
|
72
|
+
| `nvidia/minimaxai/minimax-m2.5` | 205K | | | | | | — | — |
|
|
73
73
|
| `nvidia/mistralai/codestral-22b-instruct-v0.1` | 128K | | | | | | — | — |
|
|
74
74
|
| `nvidia/mistralai/devstral-2-123b-instruct-2512` | 262K | | | | | | — | — |
|
|
75
75
|
| `nvidia/mistralai/mamba-codestral-7b-v0.1` | 128K | | | | | | — | — |
|
|
@@ -102,7 +102,9 @@ for await (const chunk of stream) {
|
|
|
102
102
|
| `nvidia/qwen/qwen3-coder-480b-a35b-instruct` | 262K | | | | | | — | — |
|
|
103
103
|
| `nvidia/qwen/qwen3-next-80b-a3b-instruct` | 262K | | | | | | — | — |
|
|
104
104
|
| `nvidia/qwen/qwen3-next-80b-a3b-thinking` | 262K | | | | | | — | — |
|
|
105
|
+
| `nvidia/qwen/qwen3.5-397b-a17b` | 262K | | | | | | — | — |
|
|
105
106
|
| `nvidia/qwen/qwq-32b` | 128K | | | | | | — | — |
|
|
107
|
+
| `nvidia/stepfun-ai/step-3-5-flash` | 256K | | | | | | — | — |
|
|
106
108
|
| `nvidia/z-ai/glm4.7` | 205K | | | | | | — | — |
|
|
107
109
|
| `nvidia/z-ai/glm5` | 203K | | | | | | — | — |
|
|
108
110
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Ollama Cloud
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 32 Ollama Cloud models through Mastra's model router. Authentication is handled automatically using the `OLLAMA_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Ollama Cloud documentation](https://docs.ollama.com/cloud).
|
|
6
6
|
|
|
@@ -40,7 +40,6 @@ for await (const chunk of stream) {
|
|
|
40
40
|
| `ollama-cloud/devstral-2:123b` | 262K | | | | | | — | — |
|
|
41
41
|
| `ollama-cloud/devstral-small-2:24b` | 262K | | | | | | — | — |
|
|
42
42
|
| `ollama-cloud/gemini-3-flash-preview` | 1.0M | | | | | | — | — |
|
|
43
|
-
| `ollama-cloud/gemini-3-pro-preview` | 1.0M | | | | | | — | — |
|
|
44
43
|
| `ollama-cloud/gemma3:12b` | 131K | | | | | | — | — |
|
|
45
44
|
| `ollama-cloud/gemma3:27b` | 131K | | | | | | — | — |
|
|
46
45
|
| `ollama-cloud/gemma3:4b` | 131K | | | | | | — | — |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenAI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 44 OpenAI models through Mastra's model router. Authentication is handled automatically using the `OPENAI_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenAI documentation](https://platform.openai.com/docs/models).
|
|
6
6
|
|
|
@@ -61,6 +61,8 @@ for await (const chunk of stream) {
|
|
|
61
61
|
| `openai/gpt-5.2-pro` | 400K | | | | | | $21 | $168 |
|
|
62
62
|
| `openai/gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
|
|
63
63
|
| `openai/gpt-5.3-codex-spark` | 128K | | | | | | $2 | $14 |
|
|
64
|
+
| `openai/gpt-5.4` | 1.1M | | | | | | $3 | $15 |
|
|
65
|
+
| `openai/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
|
|
64
66
|
| `openai/o1` | 200K | | | | | | $15 | $60 |
|
|
65
67
|
| `openai/o1-mini` | 128K | | | | | | $1 | $4 |
|
|
66
68
|
| `openai/o1-preview` | 128K | | | | | | $15 | $60 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# OpenCode Zen
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 33 OpenCode Zen models through Mastra's model router. Authentication is handled automatically using the `OPENCODE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [OpenCode Zen documentation](https://opencode.ai/docs/zen).
|
|
6
6
|
|
|
@@ -32,38 +32,41 @@ for await (const chunk of stream) {
|
|
|
32
32
|
|
|
33
33
|
## Models
|
|
34
34
|
|
|
35
|
-
| Model
|
|
36
|
-
|
|
|
37
|
-
| `opencode/big-pickle`
|
|
38
|
-
| `opencode/claude-3-5-haiku`
|
|
39
|
-
| `opencode/claude-haiku-4-5`
|
|
40
|
-
| `opencode/claude-opus-4-1`
|
|
41
|
-
| `opencode/claude-opus-4-5`
|
|
42
|
-
| `opencode/claude-opus-4-6`
|
|
43
|
-
| `opencode/claude-sonnet-4`
|
|
44
|
-
| `opencode/claude-sonnet-4-5`
|
|
45
|
-
| `opencode/claude-sonnet-4-6`
|
|
46
|
-
| `opencode/gemini-3-flash`
|
|
47
|
-
| `opencode/gemini-3-pro`
|
|
48
|
-
| `opencode/gemini-3.1-pro`
|
|
49
|
-
| `opencode/glm-4.6`
|
|
50
|
-
| `opencode/glm-4.7`
|
|
51
|
-
| `opencode/glm-5`
|
|
52
|
-
| `opencode/gpt-5`
|
|
53
|
-
| `opencode/gpt-5-codex`
|
|
54
|
-
| `opencode/gpt-5-nano`
|
|
55
|
-
| `opencode/gpt-5.1`
|
|
56
|
-
| `opencode/gpt-5.1-codex`
|
|
57
|
-
| `opencode/gpt-5.1-codex-max`
|
|
58
|
-
| `opencode/gpt-5.1-codex-mini`
|
|
59
|
-
| `opencode/gpt-5.2`
|
|
60
|
-
| `opencode/gpt-5.2-codex`
|
|
61
|
-
| `opencode/gpt-5.3-codex`
|
|
62
|
-
| `opencode/gpt-5.
|
|
63
|
-
| `opencode/
|
|
64
|
-
| `opencode/
|
|
65
|
-
| `opencode/
|
|
66
|
-
| `opencode/
|
|
35
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
|
+
| ------------------------------ | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
+
| `opencode/big-pickle` | 200K | | | | | | — | — |
|
|
38
|
+
| `opencode/claude-3-5-haiku` | 200K | | | | | | $0.80 | $4 |
|
|
39
|
+
| `opencode/claude-haiku-4-5` | 200K | | | | | | $1 | $5 |
|
|
40
|
+
| `opencode/claude-opus-4-1` | 200K | | | | | | $15 | $75 |
|
|
41
|
+
| `opencode/claude-opus-4-5` | 200K | | | | | | $5 | $25 |
|
|
42
|
+
| `opencode/claude-opus-4-6` | 1.0M | | | | | | $5 | $25 |
|
|
43
|
+
| `opencode/claude-sonnet-4` | 1.0M | | | | | | $3 | $15 |
|
|
44
|
+
| `opencode/claude-sonnet-4-5` | 1.0M | | | | | | $3 | $15 |
|
|
45
|
+
| `opencode/claude-sonnet-4-6` | 1.0M | | | | | | $3 | $15 |
|
|
46
|
+
| `opencode/gemini-3-flash` | 1.0M | | | | | | $0.50 | $3 |
|
|
47
|
+
| `opencode/gemini-3-pro` | 1.0M | | | | | | $2 | $12 |
|
|
48
|
+
| `opencode/gemini-3.1-pro` | 1.0M | | | | | | $2 | $12 |
|
|
49
|
+
| `opencode/glm-4.6` | 205K | | | | | | $0.60 | $2 |
|
|
50
|
+
| `opencode/glm-4.7` | 205K | | | | | | $0.60 | $2 |
|
|
51
|
+
| `opencode/glm-5` | 205K | | | | | | $1 | $3 |
|
|
52
|
+
| `opencode/gpt-5` | 400K | | | | | | $1 | $9 |
|
|
53
|
+
| `opencode/gpt-5-codex` | 400K | | | | | | $1 | $9 |
|
|
54
|
+
| `opencode/gpt-5-nano` | 400K | | | | | | — | — |
|
|
55
|
+
| `opencode/gpt-5.1` | 400K | | | | | | $1 | $9 |
|
|
56
|
+
| `opencode/gpt-5.1-codex` | 400K | | | | | | $1 | $9 |
|
|
57
|
+
| `opencode/gpt-5.1-codex-max` | 400K | | | | | | $1 | $10 |
|
|
58
|
+
| `opencode/gpt-5.1-codex-mini` | 400K | | | | | | $0.25 | $2 |
|
|
59
|
+
| `opencode/gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
60
|
+
| `opencode/gpt-5.2-codex` | 400K | | | | | | $2 | $14 |
|
|
61
|
+
| `opencode/gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
|
|
62
|
+
| `opencode/gpt-5.3-codex-spark` | 128K | | | | | | $2 | $14 |
|
|
63
|
+
| `opencode/gpt-5.4` | 1.1M | | | | | | $3 | $15 |
|
|
64
|
+
| `opencode/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
|
|
65
|
+
| `opencode/kimi-k2.5` | 262K | | | | | | $0.60 | $3 |
|
|
66
|
+
| `opencode/mimo-v2-flash-free` | 262K | | | | | | — | — |
|
|
67
|
+
| `opencode/minimax-m2.1` | 205K | | | | | | $0.30 | $1 |
|
|
68
|
+
| `opencode/minimax-m2.5` | 205K | | | | | | $0.30 | $1 |
|
|
69
|
+
| `opencode/minimax-m2.5-free` | 205K | | | | | | — | — |
|
|
67
70
|
|
|
68
71
|
## Advanced Configuration
|
|
69
72
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Poe
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 120 Poe models through Mastra's model router. Authentication is handled automatically using the `POE_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Poe documentation](https://creator.poe.com/docs/external-applications/openai-compatible-api).
|
|
6
6
|
|
|
@@ -62,6 +62,8 @@ for await (const chunk of stream) {
|
|
|
62
62
|
| `poe/google/gemini-2.5-pro` | 1.1M | | | | | | $0.87 | $7 |
|
|
63
63
|
| `poe/google/gemini-3-flash` | 1.0M | | | | | | $0.40 | $2 |
|
|
64
64
|
| `poe/google/gemini-3-pro` | 1.0M | | | | | | $2 | $10 |
|
|
65
|
+
| `poe/google/gemini-3.1-flash-lite` | 1.0M | | | | | | $0.25 | $2 |
|
|
66
|
+
| `poe/google/gemini-3.1-pro` | 1.0M | | | | | | $2 | $12 |
|
|
65
67
|
| `poe/google/gemini-deep-research` | 1.0M | | | | | | $2 | $10 |
|
|
66
68
|
| `poe/google/imagen-3` | 480 | | | | | | — | — |
|
|
67
69
|
| `poe/google/imagen-3-fast` | 480 | | | | | | — | — |
|
|
@@ -102,7 +104,7 @@ for await (const chunk of stream) {
|
|
|
102
104
|
| `poe/openai/gpt-4.1-nano` | 1.0M | | | | | | $0.09 | $0.36 |
|
|
103
105
|
| `poe/openai/gpt-4o` | 128K | | | | | | — | — |
|
|
104
106
|
| `poe/openai/gpt-4o-aug` | 128K | | | | | | $2 | $9 |
|
|
105
|
-
| `poe/openai/gpt-4o-mini` |
|
|
107
|
+
| `poe/openai/gpt-4o-mini` | 124K | | | | | | $0.14 | $0.54 |
|
|
106
108
|
| `poe/openai/gpt-4o-mini-search` | 128K | | | | | | $0.14 | $0.54 |
|
|
107
109
|
| `poe/openai/gpt-4o-search` | 128K | | | | | | $2 | $9 |
|
|
108
110
|
| `poe/openai/gpt-5` | 400K | | | | | | $1 | $9 |
|
|
@@ -120,6 +122,10 @@ for await (const chunk of stream) {
|
|
|
120
122
|
| `poe/openai/gpt-5.2-codex` | 400K | | | | | | $2 | $13 |
|
|
121
123
|
| `poe/openai/gpt-5.2-instant` | 128K | | | | | | $2 | $13 |
|
|
122
124
|
| `poe/openai/gpt-5.2-pro` | 400K | | | | | | $19 | $150 |
|
|
125
|
+
| `poe/openai/gpt-5.3-codex` | 400K | | | | | | $2 | $13 |
|
|
126
|
+
| `poe/openai/gpt-5.3-instant` | 128K | | | | | | $2 | $13 |
|
|
127
|
+
| `poe/openai/gpt-5.4` | 1.1M | | | | | | $2 | $14 |
|
|
128
|
+
| `poe/openai/gpt-5.4-pro` | 1.1M | | | | | | $27 | $160 |
|
|
123
129
|
| `poe/openai/gpt-image-1` | 128K | | | | | | — | — |
|
|
124
130
|
| `poe/openai/gpt-image-1-mini` | — | | | | | | — | — |
|
|
125
131
|
| `poe/openai/gpt-image-1.5` | 128K | | | | | | — | — |
|
|
@@ -1,11 +1,11 @@
|
|
|
1
1
|
# Qiniu
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 91 Qiniu models through Mastra's model router. Authentication is handled automatically using the `QINIU_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Qiniu documentation](https://developer.qiniu.com/aitokenapi).
|
|
6
6
|
|
|
7
7
|
```bash
|
|
8
|
-
|
|
8
|
+
QINIU_API_KEY=your-api-key
|
|
9
9
|
```
|
|
10
10
|
|
|
11
11
|
```typescript
|
|
@@ -60,6 +60,10 @@ for await (const chunk of stream) {
|
|
|
60
60
|
| `qiniu-ai/doubao-seed-1.6` | 256K | | | | | | — | — |
|
|
61
61
|
| `qiniu-ai/doubao-seed-1.6-flash` | 256K | | | | | | — | — |
|
|
62
62
|
| `qiniu-ai/doubao-seed-1.6-thinking` | 256K | | | | | | — | — |
|
|
63
|
+
| `qiniu-ai/doubao-seed-2.0-code` | 256K | | | | | | — | — |
|
|
64
|
+
| `qiniu-ai/doubao-seed-2.0-lite` | 256K | | | | | | — | — |
|
|
65
|
+
| `qiniu-ai/doubao-seed-2.0-mini` | 256K | | | | | | — | — |
|
|
66
|
+
| `qiniu-ai/doubao-seed-2.0-pro` | 256K | | | | | | — | — |
|
|
63
67
|
| `qiniu-ai/gemini-2.0-flash` | 1.0M | | | | | | — | — |
|
|
64
68
|
| `qiniu-ai/gemini-2.0-flash-lite` | 1.0M | | | | | | — | — |
|
|
65
69
|
| `qiniu-ai/gemini-2.5-flash` | 1.0M | | | | | | — | — |
|
|
@@ -76,12 +80,16 @@ for await (const chunk of stream) {
|
|
|
76
80
|
| `qiniu-ai/kimi-k2` | 128K | | | | | | — | — |
|
|
77
81
|
| `qiniu-ai/kling-v2-6` | 100.0M | | | | | | — | — |
|
|
78
82
|
| `qiniu-ai/meituan/longcat-flash-chat` | 131K | | | | | | — | — |
|
|
83
|
+
| `qiniu-ai/meituan/longcat-flash-lite` | 256K | | | | | | — | — |
|
|
79
84
|
| `qiniu-ai/mimo-v2-flash` | 256K | | | | | | — | — |
|
|
80
85
|
| `qiniu-ai/MiniMax-M1` | 1.0M | | | | | | — | — |
|
|
81
86
|
| `qiniu-ai/minimax/minimax-m2` | 200K | | | | | | — | — |
|
|
82
87
|
| `qiniu-ai/minimax/minimax-m2.1` | 205K | | | | | | — | — |
|
|
88
|
+
| `qiniu-ai/minimax/minimax-m2.5` | 205K | | | | | | — | — |
|
|
89
|
+
| `qiniu-ai/minimax/minimax-m2.5-highspeed` | 205K | | | | | | — | — |
|
|
83
90
|
| `qiniu-ai/moonshotai/kimi-k2-0905` | 256K | | | | | | — | — |
|
|
84
91
|
| `qiniu-ai/moonshotai/kimi-k2-thinking` | 256K | | | | | | — | — |
|
|
92
|
+
| `qiniu-ai/moonshotai/kimi-k2.5` | 256K | | | | | | — | — |
|
|
85
93
|
| `qiniu-ai/openai/gpt-5` | 400K | | | | | | — | — |
|
|
86
94
|
| `qiniu-ai/openai/gpt-5.2` | 400K | | | | | | — | — |
|
|
87
95
|
| `qiniu-ai/qwen-max-2025-01-25` | 128K | | | | | | — | — |
|
|
@@ -93,13 +101,18 @@ for await (const chunk of stream) {
|
|
|
93
101
|
| `qiniu-ai/qwen3-235b-a22b-instruct-2507` | 262K | | | | | | — | — |
|
|
94
102
|
| `qiniu-ai/qwen3-235b-a22b-thinking-2507` | 262K | | | | | | — | — |
|
|
95
103
|
| `qiniu-ai/qwen3-30b-a3b` | 40K | | | | | | — | — |
|
|
104
|
+
| `qiniu-ai/qwen3-30b-a3b-instruct-2507` | 128K | | | | | | — | — |
|
|
105
|
+
| `qiniu-ai/qwen3-30b-a3b-thinking-2507` | 126K | | | | | | — | — |
|
|
96
106
|
| `qiniu-ai/qwen3-32b` | 40K | | | | | | — | — |
|
|
97
107
|
| `qiniu-ai/qwen3-coder-480b-a35b-instruct` | 262K | | | | | | — | — |
|
|
98
108
|
| `qiniu-ai/qwen3-max` | 262K | | | | | | — | — |
|
|
99
109
|
| `qiniu-ai/qwen3-max-preview` | 256K | | | | | | — | — |
|
|
100
110
|
| `qiniu-ai/qwen3-next-80b-a3b-instruct` | 131K | | | | | | — | — |
|
|
101
111
|
| `qiniu-ai/qwen3-next-80b-a3b-thinking` | 131K | | | | | | — | — |
|
|
112
|
+
| `qiniu-ai/qwen3-vl-30b-a3b-thinking` | 128K | | | | | | — | — |
|
|
113
|
+
| `qiniu-ai/qwen3.5-397b-a17b` | 256K | | | | | | — | — |
|
|
102
114
|
| `qiniu-ai/stepfun-ai/gelab-zero-4b-preview` | 8K | | | | | | — | — |
|
|
115
|
+
| `qiniu-ai/stepfun/step-3.5-flash` | 64K | | | | | | — | — |
|
|
103
116
|
| `qiniu-ai/x-ai/grok-4-fast` | 2.0M | | | | | | — | — |
|
|
104
117
|
| `qiniu-ai/x-ai/grok-4-fast-non-reasoning` | 2.0M | | | | | | — | — |
|
|
105
118
|
| `qiniu-ai/x-ai/grok-4-fast-reasoning` | 2.0M | | | | | | — | — |
|
|
@@ -107,9 +120,11 @@ for await (const chunk of stream) {
|
|
|
107
120
|
| `qiniu-ai/x-ai/grok-4.1-fast-non-reasoning` | 2.0M | | | | | | — | — |
|
|
108
121
|
| `qiniu-ai/x-ai/grok-4.1-fast-reasoning` | 20.0M | | | | | | — | — |
|
|
109
122
|
| `qiniu-ai/x-ai/grok-code-fast-1` | 256K | | | | | | — | — |
|
|
123
|
+
| `qiniu-ai/xiaomi/mimo-v2-flash` | 256K | | | | | | — | — |
|
|
110
124
|
| `qiniu-ai/z-ai/autoglm-phone-9b` | 13K | | | | | | — | — |
|
|
111
125
|
| `qiniu-ai/z-ai/glm-4.6` | 200K | | | | | | — | — |
|
|
112
126
|
| `qiniu-ai/z-ai/glm-4.7` | 200K | | | | | | — | — |
|
|
127
|
+
| `qiniu-ai/z-ai/glm-5` | 200K | | | | | | — | — |
|
|
113
128
|
|
|
114
129
|
## Advanced Configuration
|
|
115
130
|
|
|
@@ -120,9 +135,9 @@ const agent = new Agent({
|
|
|
120
135
|
id: "custom-agent",
|
|
121
136
|
name: "custom-agent",
|
|
122
137
|
model: {
|
|
123
|
-
url: "https://api.qnaigc.com
|
|
138
|
+
url: "https://api.qnaigc.com/v1",
|
|
124
139
|
id: "qiniu-ai/MiniMax-M1",
|
|
125
|
-
apiKey: process.env.
|
|
140
|
+
apiKey: process.env.QINIU_API_KEY,
|
|
126
141
|
headers: {
|
|
127
142
|
"X-Custom-Header": "value"
|
|
128
143
|
}
|
|
@@ -139,7 +154,7 @@ const agent = new Agent({
|
|
|
139
154
|
model: ({ requestContext }) => {
|
|
140
155
|
const useAdvanced = requestContext.task === "complex";
|
|
141
156
|
return useAdvanced
|
|
142
|
-
? "qiniu-ai/z-ai/glm-
|
|
157
|
+
? "qiniu-ai/z-ai/glm-5"
|
|
143
158
|
: "qiniu-ai/MiniMax-M1";
|
|
144
159
|
}
|
|
145
160
|
});
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Requesty
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 36 Requesty models through Mastra's model router. Authentication is handled automatically using the `REQUESTY_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Requesty documentation](https://requesty.ai/solution/llm-routing/models).
|
|
6
6
|
|
|
@@ -49,8 +49,24 @@ for await (const chunk of stream) {
|
|
|
49
49
|
| `requesty/openai/gpt-4.1-mini` | 1.0M | | | | | | $0.40 | $2 |
|
|
50
50
|
| `requesty/openai/gpt-4o-mini` | 128K | | | | | | $0.15 | $0.60 |
|
|
51
51
|
| `requesty/openai/gpt-5` | 400K | | | | | | $1 | $10 |
|
|
52
|
+
| `requesty/openai/gpt-5-chat` | 400K | | | | | | $1 | $10 |
|
|
53
|
+
| `requesty/openai/gpt-5-codex` | 400K | | | | | | $1 | $10 |
|
|
54
|
+
| `requesty/openai/gpt-5-image` | 400K | | | | | | $5 | $10 |
|
|
52
55
|
| `requesty/openai/gpt-5-mini` | 128K | | | | | | $0.25 | $2 |
|
|
53
56
|
| `requesty/openai/gpt-5-nano` | 16K | | | | | | $0.05 | $0.40 |
|
|
57
|
+
| `requesty/openai/gpt-5-pro` | 400K | | | | | | $15 | $120 |
|
|
58
|
+
| `requesty/openai/gpt-5.1` | 400K | | | | | | $1 | $10 |
|
|
59
|
+
| `requesty/openai/gpt-5.1-chat` | 128K | | | | | | $1 | $10 |
|
|
60
|
+
| `requesty/openai/gpt-5.1-codex` | 400K | | | | | | $1 | $10 |
|
|
61
|
+
| `requesty/openai/gpt-5.1-codex-max` | 400K | | | | | | $1 | $9 |
|
|
62
|
+
| `requesty/openai/gpt-5.1-codex-mini` | 400K | | | | | | $0.25 | $2 |
|
|
63
|
+
| `requesty/openai/gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
64
|
+
| `requesty/openai/gpt-5.2-chat` | 128K | | | | | | $2 | $14 |
|
|
65
|
+
| `requesty/openai/gpt-5.2-codex` | 400K | | | | | | $2 | $14 |
|
|
66
|
+
| `requesty/openai/gpt-5.2-pro` | 400K | | | | | | $21 | $168 |
|
|
67
|
+
| `requesty/openai/gpt-5.3-codex` | 400K | | | | | | $2 | $14 |
|
|
68
|
+
| `requesty/openai/gpt-5.4` | 1.1M | | | | | | $3 | $15 |
|
|
69
|
+
| `requesty/openai/gpt-5.4-pro` | 1.1M | | | | | | $30 | $180 |
|
|
54
70
|
| `requesty/openai/o4-mini` | 200K | | | | | | $1 | $4 |
|
|
55
71
|
| `requesty/xai/grok-4` | 256K | | | | | | $3 | $15 |
|
|
56
72
|
| `requesty/xai/grok-4-fast` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# SiliconFlow (China)
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 78 SiliconFlow (China) models through Mastra's model router. Authentication is handled automatically using the `SILICONFLOW_CN_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [SiliconFlow (China) documentation](https://cloud.siliconflow.com/models).
|
|
6
6
|
|
|
@@ -94,6 +94,12 @@ for await (const chunk of stream) {
|
|
|
94
94
|
| `siliconflow-cn/Qwen/Qwen3-VL-32B-Thinking` | 262K | | | | | | $0.20 | $2 |
|
|
95
95
|
| `siliconflow-cn/Qwen/Qwen3-VL-8B-Instruct` | 262K | | | | | | $0.18 | $0.68 |
|
|
96
96
|
| `siliconflow-cn/Qwen/Qwen3-VL-8B-Thinking` | 262K | | | | | | $0.18 | $2 |
|
|
97
|
+
| `siliconflow-cn/Qwen/Qwen3.5-122B-A10B` | 262K | | | | | | $0.29 | $2 |
|
|
98
|
+
| `siliconflow-cn/Qwen/Qwen3.5-27B` | 262K | | | | | | $0.26 | $2 |
|
|
99
|
+
| `siliconflow-cn/Qwen/Qwen3.5-35B-A3B` | 262K | | | | | | $0.23 | $2 |
|
|
100
|
+
| `siliconflow-cn/Qwen/Qwen3.5-397B-A17B` | 262K | | | | | | $0.29 | $2 |
|
|
101
|
+
| `siliconflow-cn/Qwen/Qwen3.5-4B` | 262K | | | | | | — | — |
|
|
102
|
+
| `siliconflow-cn/Qwen/Qwen3.5-9B` | 262K | | | | | | $0.22 | $2 |
|
|
97
103
|
| `siliconflow-cn/Qwen/QwQ-32B` | 131K | | | | | | $0.15 | $0.58 |
|
|
98
104
|
| `siliconflow-cn/stepfun-ai/Step-3.5-Flash` | 262K | | | | | | $0.10 | $0.30 |
|
|
99
105
|
| `siliconflow-cn/tencent/Hunyuan-A13B-Instruct` | 131K | | | | | | $0.14 | $0.57 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Together AI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 17 Together AI models through Mastra's model router. Authentication is handled automatically using the `TOGETHER_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [Together AI documentation](https://docs.together.ai/docs/serverless-models).
|
|
6
6
|
|
|
@@ -39,8 +39,6 @@ for await (const chunk of stream) {
|
|
|
39
39
|
| `togetherai/meta-llama/Llama-3.3-70B-Instruct-Turbo` | 131K | | | | | | $0.88 | $0.88 |
|
|
40
40
|
| `togetherai/MiniMaxAI/MiniMax-M2.5` | 205K | | | | | | $0.30 | $1 |
|
|
41
41
|
| `togetherai/moonshotai/Kimi-K2-Instruct` | 131K | | | | | | $1 | $3 |
|
|
42
|
-
| `togetherai/moonshotai/Kimi-K2-Instruct-0905` | 262K | | | | | | $1 | $3 |
|
|
43
|
-
| `togetherai/moonshotai/Kimi-K2-Thinking` | 262K | | | | | | $1 | $4 |
|
|
44
42
|
| `togetherai/moonshotai/Kimi-K2.5` | 262K | | | | | | $0.50 | $3 |
|
|
45
43
|
| `togetherai/openai/gpt-oss-120b` | 131K | | | | | | $0.15 | $0.60 |
|
|
46
44
|
| `togetherai/Qwen/Qwen3-235B-A22B-Instruct-2507-tput` | 262K | | | | | | $0.20 | $0.60 |
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# xAI
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 25 xAI models through Mastra's model router. Authentication is handled automatically using the `XAI_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [xAI documentation](https://docs.x.ai/docs/models).
|
|
6
6
|
|
|
@@ -30,30 +30,33 @@ for await (const chunk of stream) {
|
|
|
30
30
|
|
|
31
31
|
## Models
|
|
32
32
|
|
|
33
|
-
| Model
|
|
34
|
-
|
|
|
35
|
-
| `xai/grok-2`
|
|
36
|
-
| `xai/grok-2-1212`
|
|
37
|
-
| `xai/grok-2-latest`
|
|
38
|
-
| `xai/grok-2-vision`
|
|
39
|
-
| `xai/grok-2-vision-1212`
|
|
40
|
-
| `xai/grok-2-vision-latest`
|
|
41
|
-
| `xai/grok-3`
|
|
42
|
-
| `xai/grok-3-fast`
|
|
43
|
-
| `xai/grok-3-fast-latest`
|
|
44
|
-
| `xai/grok-3-latest`
|
|
45
|
-
| `xai/grok-3-mini`
|
|
46
|
-
| `xai/grok-3-mini-fast`
|
|
47
|
-
| `xai/grok-3-mini-fast-latest`
|
|
48
|
-
| `xai/grok-3-mini-latest`
|
|
49
|
-
| `xai/grok-4`
|
|
50
|
-
| `xai/grok-4-1-fast`
|
|
51
|
-
| `xai/grok-4-1-fast-non-reasoning`
|
|
52
|
-
| `xai/grok-4-fast`
|
|
53
|
-
| `xai/grok-4-fast-non-reasoning`
|
|
54
|
-
| `xai/grok-beta`
|
|
55
|
-
| `xai/grok-
|
|
56
|
-
| `xai/grok-
|
|
33
|
+
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
34
|
+
| ---------------------------------------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
35
|
+
| `xai/grok-2` | 131K | | | | | | $2 | $10 |
|
|
36
|
+
| `xai/grok-2-1212` | 131K | | | | | | $2 | $10 |
|
|
37
|
+
| `xai/grok-2-latest` | 131K | | | | | | $2 | $10 |
|
|
38
|
+
| `xai/grok-2-vision` | 8K | | | | | | $2 | $10 |
|
|
39
|
+
| `xai/grok-2-vision-1212` | 8K | | | | | | $2 | $10 |
|
|
40
|
+
| `xai/grok-2-vision-latest` | 8K | | | | | | $2 | $10 |
|
|
41
|
+
| `xai/grok-3` | 131K | | | | | | $3 | $15 |
|
|
42
|
+
| `xai/grok-3-fast` | 131K | | | | | | $5 | $25 |
|
|
43
|
+
| `xai/grok-3-fast-latest` | 131K | | | | | | $5 | $25 |
|
|
44
|
+
| `xai/grok-3-latest` | 131K | | | | | | $3 | $15 |
|
|
45
|
+
| `xai/grok-3-mini` | 131K | | | | | | $0.30 | $0.50 |
|
|
46
|
+
| `xai/grok-3-mini-fast` | 131K | | | | | | $0.60 | $4 |
|
|
47
|
+
| `xai/grok-3-mini-fast-latest` | 131K | | | | | | $0.60 | $4 |
|
|
48
|
+
| `xai/grok-3-mini-latest` | 131K | | | | | | $0.30 | $0.50 |
|
|
49
|
+
| `xai/grok-4` | 256K | | | | | | $3 | $15 |
|
|
50
|
+
| `xai/grok-4-1-fast` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
51
|
+
| `xai/grok-4-1-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
52
|
+
| `xai/grok-4-fast` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
53
|
+
| `xai/grok-4-fast-non-reasoning` | 2.0M | | | | | | $0.20 | $0.50 |
|
|
54
|
+
| `xai/grok-4.20-experimental-beta-0304-non-reasoning` | 2.0M | | | | | | $2 | $6 |
|
|
55
|
+
| `xai/grok-4.20-experimental-beta-0304-reasoning` | 2.0M | | | | | | $2 | $6 |
|
|
56
|
+
| `xai/grok-4.20-multi-agent-experimental-beta-0304` | 2.0M | | | | | | $2 | $6 |
|
|
57
|
+
| `xai/grok-beta` | 131K | | | | | | $5 | $15 |
|
|
58
|
+
| `xai/grok-code-fast-1` | 256K | | | | | | $0.20 | $2 |
|
|
59
|
+
| `xai/grok-vision-beta` | 8K | | | | | | $5 | $15 |
|
|
57
60
|
|
|
58
61
|
## Advanced Configuration
|
|
59
62
|
|
|
@@ -34,7 +34,7 @@ for await (const chunk of stream) {
|
|
|
34
34
|
|
|
35
35
|
| Model | Context | Tools | Reasoning | Image | Audio | Video | Input $/1M | Output $/1M |
|
|
36
36
|
| ---------------------- | ------- | ----- | --------- | ----- | ----- | ----- | ---------- | ----------- |
|
|
37
|
-
| `xiaomi/mimo-v2-flash` | 256K | | | | | | $0.
|
|
37
|
+
| `xiaomi/mimo-v2-flash` | 256K | | | | | | $0.10 | $0.30 |
|
|
38
38
|
|
|
39
39
|
## Advanced Configuration
|
|
40
40
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# ZenMux
|
|
2
2
|
|
|
3
|
-
Access
|
|
3
|
+
Access 69 ZenMux models through Mastra's model router. Authentication is handled automatically using the `ZENMUX_API_KEY` environment variable.
|
|
4
4
|
|
|
5
5
|
Learn more in the [ZenMux documentation](https://docs.zenmux.ai).
|
|
6
6
|
|
|
@@ -54,6 +54,7 @@ for await (const chunk of stream) {
|
|
|
54
54
|
| `zenmux/google/gemini-2.5-pro` | 1.0M | | | | | | $1 | $10 |
|
|
55
55
|
| `zenmux/google/gemini-3-flash-preview` | 1.0M | | | | | | $0.50 | $3 |
|
|
56
56
|
| `zenmux/google/gemini-3-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
57
|
+
| `zenmux/google/gemini-3.1-pro-preview` | 1.0M | | | | | | $2 | $12 |
|
|
57
58
|
| `zenmux/inclusionai/ling-1t` | 128K | | | | | | $0.56 | $2 |
|
|
58
59
|
| `zenmux/inclusionai/ring-1t` | 128K | | | | | | $0.56 | $2 |
|
|
59
60
|
| `zenmux/kuaishou/kat-coder-pro-v1` | 256K | | | | | | $0.30 | $1 |
|
|
@@ -74,6 +75,7 @@ for await (const chunk of stream) {
|
|
|
74
75
|
| `zenmux/openai/gpt-5.1-codex-mini` | 400K | | | | | | $0.25 | $2 |
|
|
75
76
|
| `zenmux/openai/gpt-5.2` | 400K | | | | | | $2 | $14 |
|
|
76
77
|
| `zenmux/openai/gpt-5.2-codex` | 400K | | | | | | $2 | $14 |
|
|
78
|
+
| `zenmux/openai/gpt-5.2-pro` | 400K | | | | | | $21 | $168 |
|
|
77
79
|
| `zenmux/qwen/qwen3-coder-plus` | 1.0M | | | | | | $1 | $5 |
|
|
78
80
|
| `zenmux/qwen/qwen3-max` | 256K | | | | | | $1 | $6 |
|
|
79
81
|
| `zenmux/stepfun/step-3` | 66K | | | | | | $0.21 | $0.57 |
|
|
@@ -14,14 +14,18 @@ Direct access to individual AI model providers. Each provider offers unique mode
|
|
|
14
14
|
- [AIHubMix](https://mastra.ai/models/providers/aihubmix)
|
|
15
15
|
- [Alibaba](https://mastra.ai/models/providers/alibaba)
|
|
16
16
|
- [Alibaba (China)](https://mastra.ai/models/providers/alibaba-cn)
|
|
17
|
+
- [Alibaba Coding Plan](https://mastra.ai/models/providers/alibaba-coding-plan)
|
|
18
|
+
- [Alibaba Coding Plan (China)](https://mastra.ai/models/providers/alibaba-coding-plan-cn)
|
|
17
19
|
- [Bailing](https://mastra.ai/models/providers/bailing)
|
|
18
20
|
- [Baseten](https://mastra.ai/models/providers/baseten)
|
|
19
21
|
- [Berget.AI](https://mastra.ai/models/providers/berget)
|
|
20
22
|
- [Cerebras](https://mastra.ai/models/providers/cerebras)
|
|
21
23
|
- [Chutes](https://mastra.ai/models/providers/chutes)
|
|
24
|
+
- [Clarifai](https://mastra.ai/models/providers/clarifai)
|
|
22
25
|
- [CloudFerro Sherlock](https://mastra.ai/models/providers/cloudferro-sherlock)
|
|
23
26
|
- [Cloudflare Workers AI](https://mastra.ai/models/providers/cloudflare-workers-ai)
|
|
24
27
|
- [Cortecs](https://mastra.ai/models/providers/cortecs)
|
|
28
|
+
- [D.Run (China)](https://mastra.ai/models/providers/drun)
|
|
25
29
|
- [Deep Infra](https://mastra.ai/models/providers/deepinfra)
|
|
26
30
|
- [evroc](https://mastra.ai/models/providers/evroc)
|
|
27
31
|
- [FastRouter](https://mastra.ai/models/providers/fastrouter)
|