gsd-pi 2.41.0 → 2.42.0-dev.1df898f
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +92 -29
- package/dist/cli-web-branch.d.ts +6 -0
- package/dist/cli-web-branch.js +17 -0
- package/dist/cli.js +18 -3
- package/dist/loader.js +3 -1
- package/dist/onboarding.js +2 -1
- package/dist/resource-loader.js +39 -6
- package/dist/resources/extensions/async-jobs/async-bash-tool.js +52 -4
- package/dist/resources/extensions/async-jobs/await-tool.js +5 -0
- package/dist/resources/extensions/async-jobs/index.js +2 -0
- package/dist/resources/extensions/gsd/auto/loop.js +89 -1
- package/dist/resources/extensions/gsd/auto/phases.js +29 -13
- package/dist/resources/extensions/gsd/auto/session.js +6 -0
- package/dist/resources/extensions/gsd/auto-dashboard.js +8 -2
- package/dist/resources/extensions/gsd/auto-dispatch.js +19 -2
- package/dist/resources/extensions/gsd/auto-post-unit.js +7 -0
- package/dist/resources/extensions/gsd/auto-prompts.js +3 -16
- package/dist/resources/extensions/gsd/auto-recovery.js +12 -4
- package/dist/resources/extensions/gsd/auto-start.js +16 -14
- package/dist/resources/extensions/gsd/auto-worktree.js +147 -13
- package/dist/resources/extensions/gsd/auto.js +64 -2
- package/dist/resources/extensions/gsd/bootstrap/db-tools.js +199 -164
- package/dist/resources/extensions/gsd/bootstrap/journal-tools.js +62 -0
- package/dist/resources/extensions/gsd/bootstrap/register-extension.js +2 -0
- package/dist/resources/extensions/gsd/bootstrap/register-hooks.js +25 -3
- package/dist/resources/extensions/gsd/bootstrap/tool-call-loop-guard.js +7 -2
- package/dist/resources/extensions/gsd/commands/catalog.js +40 -1
- package/dist/resources/extensions/gsd/commands/handlers/core.js +1 -0
- package/dist/resources/extensions/gsd/commands/handlers/ops.js +5 -0
- package/dist/resources/extensions/gsd/commands/handlers/workflow.js +146 -0
- package/dist/resources/extensions/gsd/context-injector.js +74 -0
- package/dist/resources/extensions/gsd/context-store.js +4 -3
- package/dist/resources/extensions/gsd/custom-execution-policy.js +47 -0
- package/dist/resources/extensions/gsd/custom-verification.js +145 -0
- package/dist/resources/extensions/gsd/custom-workflow-engine.js +164 -0
- package/dist/resources/extensions/gsd/dashboard-overlay.js +1 -0
- package/dist/resources/extensions/gsd/db-writer.js +5 -2
- package/dist/resources/extensions/gsd/definition-loader.js +352 -0
- package/dist/resources/extensions/gsd/detection.js +20 -1
- package/dist/resources/extensions/gsd/dev-execution-policy.js +24 -0
- package/dist/resources/extensions/gsd/dev-workflow-engine.js +82 -0
- package/dist/resources/extensions/gsd/doctor-checks.js +31 -1
- package/dist/resources/extensions/gsd/doctor-providers.js +10 -0
- package/dist/resources/extensions/gsd/doctor.js +11 -1
- package/dist/resources/extensions/gsd/engine-resolver.js +40 -0
- package/dist/resources/extensions/gsd/engine-types.js +8 -0
- package/dist/resources/extensions/gsd/execution-policy.js +8 -0
- package/dist/resources/extensions/gsd/exit-command.js +12 -2
- package/dist/resources/extensions/gsd/export.js +9 -13
- package/dist/resources/extensions/gsd/extension-manifest.json +2 -2
- package/dist/resources/extensions/gsd/files.js +28 -11
- package/dist/resources/extensions/gsd/forensics.js +94 -3
- package/dist/resources/extensions/gsd/git-constants.js +1 -0
- package/dist/resources/extensions/gsd/git-service.js +6 -2
- package/dist/resources/extensions/gsd/graph.js +225 -0
- package/dist/resources/extensions/gsd/gsd-db.js +25 -8
- package/dist/resources/extensions/gsd/guided-flow-queue.js +1 -1
- package/dist/resources/extensions/gsd/guided-flow.js +7 -3
- package/dist/resources/extensions/gsd/journal.js +85 -0
- package/dist/resources/extensions/gsd/md-importer.js +5 -0
- package/dist/resources/extensions/gsd/milestone-ids.js +1 -1
- package/dist/resources/extensions/gsd/native-git-bridge.js +3 -2
- package/dist/resources/extensions/gsd/post-unit-hooks.js +24 -412
- package/dist/resources/extensions/gsd/preferences-types.js +2 -0
- package/dist/resources/extensions/gsd/preferences.js +60 -8
- package/dist/resources/extensions/gsd/prompt-loader.js +34 -4
- package/dist/resources/extensions/gsd/prompts/complete-milestone.md +11 -10
- package/dist/resources/extensions/gsd/prompts/discuss-headless.md +2 -2
- package/dist/resources/extensions/gsd/prompts/discuss.md +1 -1
- package/dist/resources/extensions/gsd/prompts/forensics.md +12 -5
- package/dist/resources/extensions/gsd/prompts/queue.md +1 -1
- package/dist/resources/extensions/gsd/repo-identity.js +92 -7
- package/dist/resources/extensions/gsd/rule-registry.js +489 -0
- package/dist/resources/extensions/gsd/rule-types.js +6 -0
- package/dist/resources/extensions/gsd/run-manager.js +134 -0
- package/dist/resources/extensions/gsd/service-tier.js +147 -0
- package/dist/resources/extensions/gsd/session-lock.js +2 -2
- package/dist/resources/extensions/gsd/structured-data-formatter.js +2 -1
- package/dist/resources/extensions/gsd/templates/decisions.md +2 -2
- package/dist/resources/extensions/gsd/workflow-engine.js +7 -0
- package/dist/resources/extensions/gsd/workflow-templates.js +13 -1
- package/dist/resources/extensions/gsd/worktree-manager.js +20 -6
- package/dist/resources/extensions/gsd/worktree-resolver.js +21 -4
- package/dist/resources/extensions/gsd/worktree.js +2 -2
- package/dist/resources/extensions/mcp-client/index.js +2 -1
- package/dist/resources/extensions/search-the-web/tool-search.js +3 -3
- package/dist/resources/extensions/subagent/index.js +7 -3
- package/dist/resources/extensions/voice/index.js +4 -4
- package/dist/resources/skills/create-workflow/SKILL.md +103 -0
- package/dist/resources/skills/create-workflow/references/feature-patterns.md +128 -0
- package/dist/resources/skills/create-workflow/references/verification-policies.md +76 -0
- package/dist/resources/skills/create-workflow/references/yaml-schema-v1.md +46 -0
- package/dist/resources/skills/create-workflow/templates/blog-post-pipeline.yaml +60 -0
- package/dist/resources/skills/create-workflow/templates/code-audit.yaml +60 -0
- package/dist/resources/skills/create-workflow/templates/release-checklist.yaml +66 -0
- package/dist/resources/skills/create-workflow/templates/workflow-definition.yaml +32 -0
- package/dist/resources/skills/create-workflow/workflows/create-from-scratch.md +104 -0
- package/dist/resources/skills/create-workflow/workflows/create-from-template.md +72 -0
- package/dist/web/standalone/.next/BUILD_ID +1 -1
- package/dist/web/standalone/.next/app-path-routes-manifest.json +9 -9
- package/dist/web/standalone/.next/build-manifest.json +4 -4
- package/dist/web/standalone/.next/prerender-manifest.json +3 -3
- package/dist/web/standalone/.next/react-loadable-manifest.json +1 -1
- package/dist/web/standalone/.next/required-server-files.json +3 -3
- package/dist/web/standalone/.next/server/app/_global-error/page.js +3 -3
- package/dist/web/standalone/.next/server/app/_global-error/page_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.html +2 -2
- package/dist/web/standalone/.next/server/app/_global-error.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_full.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_global-error/__PAGE__.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_global-error.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_head.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_index.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_tree.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found/page.js +2 -2
- package/dist/web/standalone/.next/server/app/_not-found/page_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.html +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.rsc +3 -3
- package/dist/web/standalone/.next/server/app/_not-found.segments/_full.segment.rsc +3 -3
- package/dist/web/standalone/.next/server/app/_not-found.segments/_head.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_index.segment.rsc +3 -3
- package/dist/web/standalone/.next/server/app/_not-found.segments/_not-found/__PAGE__.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_not-found.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_tree.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/api/boot/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/boot/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/bridge-terminal/input/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/bridge-terminal/input/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/bridge-terminal/resize/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/bridge-terminal/resize/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/bridge-terminal/stream/route.js +2 -2
- package/dist/web/standalone/.next/server/app/api/bridge-terminal/stream/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/browse-directories/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/browse-directories/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/captures/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/captures/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/cleanup/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/cleanup/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/dev-mode/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/dev-mode/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/doctor/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/doctor/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/export-data/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/export-data/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/files/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/files/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/forensics/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/forensics/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/git/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/git/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/history/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/history/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/hooks/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/hooks/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/inspect/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/inspect/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/knowledge/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/knowledge/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/live-state/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/live-state/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/onboarding/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/onboarding/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/preferences/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/preferences/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/projects/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/projects/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/recovery/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/recovery/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/remote-questions/route.js +5 -5
- package/dist/web/standalone/.next/server/app/api/remote-questions/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/browser/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/browser/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/command/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/command/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/events/route.js +2 -2
- package/dist/web/standalone/.next/server/app/api/session/events/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/manage/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/session/manage/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/settings-data/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/settings-data/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/shutdown/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/shutdown/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/skill-health/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/skill-health/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/steer/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/steer/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/input/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/input/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/resize/route.js +2 -2
- package/dist/web/standalone/.next/server/app/api/terminal/resize/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/sessions/route.js +2 -2
- package/dist/web/standalone/.next/server/app/api/terminal/sessions/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/stream/route.js +4 -4
- package/dist/web/standalone/.next/server/app/api/terminal/stream/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/upload/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/terminal/upload/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/undo/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/undo/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/update/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/update/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/api/visualizer/route.js +1 -1
- package/dist/web/standalone/.next/server/app/api/visualizer/route_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app/index.html +1 -1
- package/dist/web/standalone/.next/server/app/index.rsc +4 -4
- package/dist/web/standalone/.next/server/app/index.segments/__PAGE__.segment.rsc +2 -2
- package/dist/web/standalone/.next/server/app/index.segments/_full.segment.rsc +4 -4
- package/dist/web/standalone/.next/server/app/index.segments/_head.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/index.segments/_index.segment.rsc +3 -3
- package/dist/web/standalone/.next/server/app/index.segments/_tree.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/page.js +2 -2
- package/dist/web/standalone/.next/server/app/page_client-reference-manifest.js +1 -1
- package/dist/web/standalone/.next/server/app-paths-manifest.json +9 -9
- package/dist/web/standalone/.next/server/chunks/229.js +3 -3
- package/dist/web/standalone/.next/server/chunks/471.js +3 -3
- package/dist/web/standalone/.next/server/middleware-build-manifest.js +1 -1
- package/dist/web/standalone/.next/server/middleware-react-loadable-manifest.js +1 -1
- package/dist/web/standalone/.next/server/middleware.js +2 -2
- package/dist/web/standalone/.next/server/next-font-manifest.js +1 -1
- package/dist/web/standalone/.next/server/next-font-manifest.json +1 -1
- package/dist/web/standalone/.next/server/pages/404.html +1 -1
- package/dist/web/standalone/.next/server/pages/500.html +2 -2
- package/dist/web/standalone/.next/server/server-reference-manifest.json +1 -1
- package/dist/web/standalone/.next/static/chunks/4024.c195dc1fdd2adbea.js +9 -0
- package/dist/web/standalone/.next/static/chunks/app/_not-found/page-f2a7482d42a5614b.js +1 -0
- package/dist/web/standalone/.next/static/chunks/app/layout-a16c7a7ecdf0c2cf.js +1 -0
- package/dist/web/standalone/.next/static/chunks/app/page-b9367c5ae13b99c6.js +1 -0
- package/dist/web/standalone/.next/static/chunks/{main-app-2f2ee7b85712c2bd.js → main-app-fdab67f7802d7832.js} +1 -1
- package/dist/web/standalone/.next/static/chunks/next/dist/client/components/builtin/global-error-459824ffb8c323dd.js +1 -0
- package/dist/web/standalone/.next/static/chunks/{webpack-9afaaebf6042a1d7.js → webpack-fa307370fcf9fb2c.js} +1 -1
- package/dist/web/standalone/node_modules/node-pty/build/Makefile +2 -2
- package/dist/web/standalone/node_modules/node-pty/build/Release/pty.node +0 -0
- package/dist/web/standalone/node_modules/node-pty/build/pty.target.mk +14 -14
- package/dist/web/standalone/node_modules/node-pty/node-addon-api/node_addon_api.target.mk +14 -14
- package/dist/web/standalone/node_modules/node-pty/node-addon-api/node_addon_api_except.target.mk +14 -14
- package/dist/web/standalone/node_modules/node-pty/node-addon-api/node_addon_api_maybe.target.mk +14 -14
- package/dist/web/standalone/server.js +1 -1
- package/dist/web-mode.d.ts +4 -0
- package/dist/web-mode.js +69 -11
- package/package.json +1 -1
- package/packages/native/src/__tests__/text.test.mjs +33 -0
- package/packages/pi-agent-core/dist/agent.d.ts.map +1 -1
- package/packages/pi-agent-core/dist/agent.js +2 -0
- package/packages/pi-agent-core/dist/agent.js.map +1 -1
- package/packages/pi-agent-core/dist/types.d.ts +6 -0
- package/packages/pi-agent-core/dist/types.d.ts.map +1 -1
- package/packages/pi-agent-core/dist/types.js.map +1 -1
- package/packages/pi-agent-core/src/agent.test.ts +53 -0
- package/packages/pi-agent-core/src/agent.ts +3 -0
- package/packages/pi-agent-core/src/types.ts +6 -0
- package/packages/pi-agent-core/tsconfig.json +1 -1
- package/packages/pi-ai/dist/models.d.ts +5 -3
- package/packages/pi-ai/dist/models.d.ts.map +1 -1
- package/packages/pi-ai/dist/models.generated.d.ts +801 -1468
- package/packages/pi-ai/dist/models.generated.d.ts.map +1 -1
- package/packages/pi-ai/dist/models.generated.js +1135 -1588
- package/packages/pi-ai/dist/models.generated.js.map +1 -1
- package/packages/pi-ai/dist/models.js.map +1 -1
- package/packages/pi-ai/dist/utils/oauth/github-copilot.d.ts.map +1 -1
- package/packages/pi-ai/dist/utils/oauth/github-copilot.js +60 -2
- package/packages/pi-ai/dist/utils/oauth/github-copilot.js.map +1 -1
- package/packages/pi-ai/scripts/generate-models.ts +1543 -0
- package/packages/pi-ai/src/models.generated.ts +1140 -1593
- package/packages/pi-ai/src/models.ts +7 -4
- package/packages/pi-ai/src/utils/oauth/github-copilot.ts +74 -2
- package/packages/pi-coding-agent/dist/core/agent-session.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/agent-session.js +8 -1
- package/packages/pi-coding-agent/dist/core/agent-session.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/auth-storage.d.ts +7 -0
- package/packages/pi-coding-agent/dist/core/auth-storage.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/auth-storage.js +29 -2
- package/packages/pi-coding-agent/dist/core/auth-storage.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/auth-storage.test.js +60 -0
- package/packages/pi-coding-agent/dist/core/auth-storage.test.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/discovery-cache.test.js +3 -1
- package/packages/pi-coding-agent/dist/core/discovery-cache.test.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/extensions/loader.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/extensions/loader.js +18 -0
- package/packages/pi-coding-agent/dist/core/extensions/loader.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/lsp/client.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/lsp/client.js +23 -0
- package/packages/pi-coding-agent/dist/core/lsp/client.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/model-registry.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/model-registry.js +2 -0
- package/packages/pi-coding-agent/dist/core/model-registry.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/package-manager.d.ts +6 -0
- package/packages/pi-coding-agent/dist/core/package-manager.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/package-manager.js +63 -11
- package/packages/pi-coding-agent/dist/core/package-manager.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/resource-loader.d.ts +9 -0
- package/packages/pi-coding-agent/dist/core/resource-loader.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/resource-loader.js +20 -6
- package/packages/pi-coding-agent/dist/core/resource-loader.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/system-prompt.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/system-prompt.js +6 -5
- package/packages/pi-coding-agent/dist/core/system-prompt.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.js +3 -0
- package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js +9 -6
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/login-dialog.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/login-dialog.js +10 -7
- package/packages/pi-coding-agent/dist/modes/interactive/components/login-dialog.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js +34 -10
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js.map +1 -1
- package/packages/pi-coding-agent/package.json +1 -1
- package/packages/pi-coding-agent/src/core/agent-session.ts +7 -1
- package/packages/pi-coding-agent/src/core/auth-storage.test.ts +68 -0
- package/packages/pi-coding-agent/src/core/auth-storage.ts +30 -2
- package/packages/pi-coding-agent/src/core/discovery-cache.test.ts +4 -2
- package/packages/pi-coding-agent/src/core/extensions/loader.ts +18 -0
- package/packages/pi-coding-agent/src/core/lsp/client.ts +29 -0
- package/packages/pi-coding-agent/src/core/model-registry.ts +3 -0
- package/packages/pi-coding-agent/src/core/package-manager.ts +99 -58
- package/packages/pi-coding-agent/src/core/resource-loader.ts +24 -6
- package/packages/pi-coding-agent/src/core/system-prompt.ts +6 -5
- package/packages/pi-coding-agent/src/modes/interactive/components/extension-editor.ts +3 -0
- package/packages/pi-coding-agent/src/modes/interactive/components/footer.ts +10 -6
- package/packages/pi-coding-agent/src/modes/interactive/components/login-dialog.ts +11 -7
- package/packages/pi-coding-agent/src/modes/interactive/interactive-mode.ts +36 -11
- package/pkg/package.json +1 -1
- package/src/resources/extensions/async-jobs/async-bash-timeout.test.ts +122 -0
- package/src/resources/extensions/async-jobs/async-bash-tool.ts +40 -4
- package/src/resources/extensions/async-jobs/await-tool.test.ts +47 -0
- package/src/resources/extensions/async-jobs/await-tool.ts +5 -0
- package/src/resources/extensions/async-jobs/index.ts +1 -0
- package/src/resources/extensions/async-jobs/job-manager.ts +2 -0
- package/src/resources/extensions/gsd/auto/loop-deps.ts +5 -2
- package/src/resources/extensions/gsd/auto/loop.ts +101 -1
- package/src/resources/extensions/gsd/auto/phases.ts +31 -13
- package/src/resources/extensions/gsd/auto/session.ts +6 -0
- package/src/resources/extensions/gsd/auto/types.ts +4 -0
- package/src/resources/extensions/gsd/auto-dashboard.ts +9 -2
- package/src/resources/extensions/gsd/auto-dispatch.ts +25 -5
- package/src/resources/extensions/gsd/auto-post-unit.ts +8 -0
- package/src/resources/extensions/gsd/auto-prompts.ts +2 -18
- package/src/resources/extensions/gsd/auto-recovery.ts +12 -4
- package/src/resources/extensions/gsd/auto-start.ts +15 -13
- package/src/resources/extensions/gsd/auto-worktree.ts +162 -18
- package/src/resources/extensions/gsd/auto.ts +71 -2
- package/src/resources/extensions/gsd/bootstrap/db-tools.ts +209 -162
- package/src/resources/extensions/gsd/bootstrap/journal-tools.ts +62 -0
- package/src/resources/extensions/gsd/bootstrap/register-extension.ts +2 -0
- package/src/resources/extensions/gsd/bootstrap/register-hooks.ts +25 -4
- package/src/resources/extensions/gsd/bootstrap/tool-call-loop-guard.ts +9 -2
- package/src/resources/extensions/gsd/commands/catalog.ts +40 -1
- package/src/resources/extensions/gsd/commands/handlers/core.ts +1 -0
- package/src/resources/extensions/gsd/commands/handlers/ops.ts +5 -0
- package/src/resources/extensions/gsd/commands/handlers/workflow.ts +164 -0
- package/src/resources/extensions/gsd/context-injector.ts +100 -0
- package/src/resources/extensions/gsd/context-store.ts +4 -3
- package/src/resources/extensions/gsd/custom-execution-policy.ts +73 -0
- package/src/resources/extensions/gsd/custom-verification.ts +180 -0
- package/src/resources/extensions/gsd/custom-workflow-engine.ts +216 -0
- package/src/resources/extensions/gsd/dashboard-overlay.ts +1 -0
- package/src/resources/extensions/gsd/db-writer.ts +6 -2
- package/src/resources/extensions/gsd/definition-loader.ts +462 -0
- package/src/resources/extensions/gsd/detection.ts +20 -1
- package/src/resources/extensions/gsd/dev-execution-policy.ts +51 -0
- package/src/resources/extensions/gsd/dev-workflow-engine.ts +110 -0
- package/src/resources/extensions/gsd/doctor-checks.ts +32 -1
- package/src/resources/extensions/gsd/doctor-providers.ts +13 -0
- package/src/resources/extensions/gsd/doctor-types.ts +1 -0
- package/src/resources/extensions/gsd/doctor.ts +12 -1
- package/src/resources/extensions/gsd/engine-resolver.ts +57 -0
- package/src/resources/extensions/gsd/engine-types.ts +71 -0
- package/src/resources/extensions/gsd/execution-policy.ts +43 -0
- package/src/resources/extensions/gsd/exit-command.ts +14 -2
- package/src/resources/extensions/gsd/export.ts +8 -15
- package/src/resources/extensions/gsd/extension-manifest.json +2 -2
- package/src/resources/extensions/gsd/files.ts +29 -12
- package/src/resources/extensions/gsd/forensics.ts +101 -3
- package/src/resources/extensions/gsd/git-constants.ts +1 -0
- package/src/resources/extensions/gsd/git-service.ts +5 -5
- package/src/resources/extensions/gsd/gitignore.ts +1 -1
- package/src/resources/extensions/gsd/graph.ts +312 -0
- package/src/resources/extensions/gsd/gsd-db.ts +37 -8
- package/src/resources/extensions/gsd/guided-flow-queue.ts +1 -1
- package/src/resources/extensions/gsd/guided-flow.ts +7 -3
- package/src/resources/extensions/gsd/journal.ts +134 -0
- package/src/resources/extensions/gsd/md-importer.ts +6 -0
- package/src/resources/extensions/gsd/milestone-ids.ts +1 -1
- package/src/resources/extensions/gsd/native-git-bridge.ts +3 -2
- package/src/resources/extensions/gsd/post-unit-hooks.ts +24 -462
- package/src/resources/extensions/gsd/preferences-types.ts +6 -0
- package/src/resources/extensions/gsd/preferences.ts +63 -6
- package/src/resources/extensions/gsd/prompt-loader.ts +35 -4
- package/src/resources/extensions/gsd/prompts/complete-milestone.md +11 -10
- package/src/resources/extensions/gsd/prompts/discuss-headless.md +2 -2
- package/src/resources/extensions/gsd/prompts/discuss.md +1 -1
- package/src/resources/extensions/gsd/prompts/forensics.md +12 -5
- package/src/resources/extensions/gsd/prompts/queue.md +1 -1
- package/src/resources/extensions/gsd/repo-identity.ts +95 -7
- package/src/resources/extensions/gsd/rule-registry.ts +599 -0
- package/src/resources/extensions/gsd/rule-types.ts +68 -0
- package/src/resources/extensions/gsd/run-manager.ts +180 -0
- package/src/resources/extensions/gsd/service-tier.ts +184 -0
- package/src/resources/extensions/gsd/session-lock.ts +2 -2
- package/src/resources/extensions/gsd/structured-data-formatter.ts +3 -1
- package/src/resources/extensions/gsd/templates/decisions.md +2 -2
- package/src/resources/extensions/gsd/tests/activity-log.test.ts +31 -69
- package/src/resources/extensions/gsd/tests/auto-loop.test.ts +103 -120
- package/src/resources/extensions/gsd/tests/auto-recovery.test.ts +85 -0
- package/src/resources/extensions/gsd/tests/auto-secrets-gate.test.ts +2 -2
- package/src/resources/extensions/gsd/tests/auto-worktree-milestone-merge.test.ts +202 -0
- package/src/resources/extensions/gsd/tests/bundled-workflow-defs.test.ts +180 -0
- package/src/resources/extensions/gsd/tests/captures.test.ts +12 -1
- package/src/resources/extensions/gsd/tests/commands-workflow-custom.test.ts +283 -0
- package/src/resources/extensions/gsd/tests/context-injector.test.ts +313 -0
- package/src/resources/extensions/gsd/tests/context-store.test.ts +10 -5
- package/src/resources/extensions/gsd/tests/continue-here.test.ts +20 -20
- package/src/resources/extensions/gsd/tests/custom-engine-loop-integration.test.ts +540 -0
- package/src/resources/extensions/gsd/tests/custom-verification.test.ts +382 -0
- package/src/resources/extensions/gsd/tests/custom-workflow-engine.test.ts +339 -0
- package/src/resources/extensions/gsd/tests/dashboard-custom-engine.test.ts +87 -0
- package/src/resources/extensions/gsd/tests/db-writer.test.ts +10 -0
- package/src/resources/extensions/gsd/tests/definition-loader.test.ts +778 -0
- package/src/resources/extensions/gsd/tests/dev-engine-wrapper.test.ts +318 -0
- package/src/resources/extensions/gsd/tests/doctor-completion-deferral.test.ts +15 -10
- package/src/resources/extensions/gsd/tests/doctor-fixlevel.test.ts +5 -4
- package/src/resources/extensions/gsd/tests/doctor-roadmap-summary-atomicity.test.ts +167 -0
- package/src/resources/extensions/gsd/tests/doctor-task-done-missing-summary-slice-loop.test.ts +174 -0
- package/src/resources/extensions/gsd/tests/e2e-workflow-pipeline-integration.test.ts +476 -0
- package/src/resources/extensions/gsd/tests/engine-interfaces-contract.test.ts +271 -0
- package/src/resources/extensions/gsd/tests/exit-command.test.ts +55 -0
- package/src/resources/extensions/gsd/tests/forensics-dedup.test.ts +48 -0
- package/src/resources/extensions/gsd/tests/forensics-issue-routing.test.ts +43 -0
- package/src/resources/extensions/gsd/tests/git-locale.test.ts +133 -0
- package/src/resources/extensions/gsd/tests/git-service.test.ts +44 -0
- package/src/resources/extensions/gsd/tests/graph-operations.test.ts +599 -0
- package/src/resources/extensions/gsd/tests/gsd-db.test.ts +8 -1
- package/src/resources/extensions/gsd/tests/gsd-tools.test.ts +7 -7
- package/src/resources/extensions/gsd/tests/iterate-engine-integration.test.ts +429 -0
- package/src/resources/extensions/gsd/tests/journal-integration.test.ts +513 -0
- package/src/resources/extensions/gsd/tests/journal-query-tool.test.ts +147 -0
- package/src/resources/extensions/gsd/tests/journal.test.ts +341 -0
- package/src/resources/extensions/gsd/tests/manifest-status.test.ts +73 -82
- package/src/resources/extensions/gsd/tests/md-importer.test.ts +31 -1
- package/src/resources/extensions/gsd/tests/memory-store.test.ts +2 -2
- package/src/resources/extensions/gsd/tests/milestone-id-reservation.test.ts +1 -1
- package/src/resources/extensions/gsd/tests/parsers.test.ts +110 -0
- package/src/resources/extensions/gsd/tests/preferences.test.ts +47 -25
- package/src/resources/extensions/gsd/tests/prompt-db.test.ts +3 -1
- package/src/resources/extensions/gsd/tests/repo-identity-worktree.test.ts +61 -1
- package/src/resources/extensions/gsd/tests/routing-history.test.ts +11 -22
- package/src/resources/extensions/gsd/tests/rule-registry.test.ts +413 -0
- package/src/resources/extensions/gsd/tests/run-manager.test.ts +229 -0
- package/src/resources/extensions/gsd/tests/service-tier.test.ts +127 -0
- package/src/resources/extensions/gsd/tests/skill-activation.test.ts +56 -3
- package/src/resources/extensions/gsd/tests/skill-lifecycle.test.ts +2 -2
- package/src/resources/extensions/gsd/tests/stalled-tool-recovery.test.ts +102 -0
- package/src/resources/extensions/gsd/tests/structured-data-formatter.test.ts +4 -3
- package/src/resources/extensions/gsd/tests/symlink-numbered-variants.test.ts +151 -0
- package/src/resources/extensions/gsd/tests/tool-call-loop-guard.test.ts +45 -0
- package/src/resources/extensions/gsd/tests/tool-naming.test.ts +117 -0
- package/src/resources/extensions/gsd/tests/triage-dispatch.test.ts +6 -1
- package/src/resources/extensions/gsd/tests/verification-gate.test.ts +156 -263
- package/src/resources/extensions/gsd/tests/windows-path-normalization.test.ts +99 -0
- package/src/resources/extensions/gsd/tests/worktree-db-integration.test.ts +1 -0
- package/src/resources/extensions/gsd/tests/worktree-db.test.ts +4 -0
- package/src/resources/extensions/gsd/tests/worktree-health-dispatch.test.ts +135 -0
- package/src/resources/extensions/gsd/tests/worktree-manager.test.ts +203 -106
- package/src/resources/extensions/gsd/tests/worktree-resolver.test.ts +79 -5
- package/src/resources/extensions/gsd/tests/worktree-symlink-removal.test.ts +140 -0
- package/src/resources/extensions/gsd/tests/worktree-sync-milestones.test.ts +74 -0
- package/src/resources/extensions/gsd/types.ts +3 -0
- package/src/resources/extensions/gsd/workflow-engine.ts +38 -0
- package/src/resources/extensions/gsd/workflow-templates.ts +12 -1
- package/src/resources/extensions/gsd/worktree-manager.ts +21 -6
- package/src/resources/extensions/gsd/worktree-resolver.ts +32 -12
- package/src/resources/extensions/gsd/worktree.ts +2 -2
- package/src/resources/extensions/mcp-client/index.ts +5 -1
- package/src/resources/extensions/search-the-web/tool-search.ts +3 -3
- package/src/resources/extensions/subagent/index.ts +7 -3
- package/src/resources/extensions/voice/index.ts +4 -4
- package/src/resources/skills/create-workflow/SKILL.md +103 -0
- package/src/resources/skills/create-workflow/references/feature-patterns.md +128 -0
- package/src/resources/skills/create-workflow/references/verification-policies.md +76 -0
- package/src/resources/skills/create-workflow/references/yaml-schema-v1.md +46 -0
- package/src/resources/skills/create-workflow/templates/blog-post-pipeline.yaml +60 -0
- package/src/resources/skills/create-workflow/templates/code-audit.yaml +60 -0
- package/src/resources/skills/create-workflow/templates/release-checklist.yaml +66 -0
- package/src/resources/skills/create-workflow/templates/workflow-definition.yaml +32 -0
- package/src/resources/skills/create-workflow/workflows/create-from-scratch.md +104 -0
- package/src/resources/skills/create-workflow/workflows/create-from-template.md +72 -0
- package/dist/web/standalone/.next/static/chunks/4024.279c423e4661ece1.js +0 -9
- package/dist/web/standalone/.next/static/chunks/app/_not-found/page-e07acdb7dd069836.js +0 -1
- package/dist/web/standalone/.next/static/chunks/app/layout-745c6ed5fea5fb06.js +0 -1
- package/dist/web/standalone/.next/static/chunks/app/page-801b53eff6e83579.js +0 -1
- package/dist/web/standalone/.next/static/chunks/next/dist/client/components/builtin/global-error-e6255954dccfcf0a.js +0 -1
- /package/dist/web/standalone/.next/static/{Ute3pMouVczQyT15qrBBO → qw8qDHXOTLUXBq1vEknSz}/_buildManifest.js +0 -0
- /package/dist/web/standalone/.next/static/{Ute3pMouVczQyT15qrBBO → qw8qDHXOTLUXBq1vEknSz}/_ssgManifest.js +0 -0
|
@@ -87,40 +87,6 @@ export const MODELS = {
|
|
|
87
87
|
contextWindow: 300000,
|
|
88
88
|
maxTokens: 8192,
|
|
89
89
|
},
|
|
90
|
-
"amazon.titan-text-express-v1": {
|
|
91
|
-
id: "amazon.titan-text-express-v1",
|
|
92
|
-
name: "Titan Text G1 - Express",
|
|
93
|
-
api: "bedrock-converse-stream",
|
|
94
|
-
provider: "amazon-bedrock",
|
|
95
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
96
|
-
reasoning: false,
|
|
97
|
-
input: ["text"],
|
|
98
|
-
cost: {
|
|
99
|
-
input: 0.2,
|
|
100
|
-
output: 0.6,
|
|
101
|
-
cacheRead: 0,
|
|
102
|
-
cacheWrite: 0,
|
|
103
|
-
},
|
|
104
|
-
contextWindow: 128000,
|
|
105
|
-
maxTokens: 4096,
|
|
106
|
-
},
|
|
107
|
-
"amazon.titan-text-express-v1:0:8k": {
|
|
108
|
-
id: "amazon.titan-text-express-v1:0:8k",
|
|
109
|
-
name: "Titan Text G1 - Express",
|
|
110
|
-
api: "bedrock-converse-stream",
|
|
111
|
-
provider: "amazon-bedrock",
|
|
112
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
113
|
-
reasoning: false,
|
|
114
|
-
input: ["text"],
|
|
115
|
-
cost: {
|
|
116
|
-
input: 0.2,
|
|
117
|
-
output: 0.6,
|
|
118
|
-
cacheRead: 0,
|
|
119
|
-
cacheWrite: 0,
|
|
120
|
-
},
|
|
121
|
-
contextWindow: 128000,
|
|
122
|
-
maxTokens: 4096,
|
|
123
|
-
},
|
|
124
90
|
"anthropic.claude-3-5-haiku-20241022-v1:0": {
|
|
125
91
|
id: "anthropic.claude-3-5-haiku-20241022-v1:0",
|
|
126
92
|
name: "Claude Haiku 3.5",
|
|
@@ -206,40 +172,6 @@ export const MODELS = {
|
|
|
206
172
|
contextWindow: 200000,
|
|
207
173
|
maxTokens: 4096,
|
|
208
174
|
},
|
|
209
|
-
"anthropic.claude-3-opus-20240229-v1:0": {
|
|
210
|
-
id: "anthropic.claude-3-opus-20240229-v1:0",
|
|
211
|
-
name: "Claude Opus 3",
|
|
212
|
-
api: "bedrock-converse-stream",
|
|
213
|
-
provider: "amazon-bedrock",
|
|
214
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
215
|
-
reasoning: false,
|
|
216
|
-
input: ["text", "image"],
|
|
217
|
-
cost: {
|
|
218
|
-
input: 15,
|
|
219
|
-
output: 75,
|
|
220
|
-
cacheRead: 0,
|
|
221
|
-
cacheWrite: 0,
|
|
222
|
-
},
|
|
223
|
-
contextWindow: 200000,
|
|
224
|
-
maxTokens: 4096,
|
|
225
|
-
},
|
|
226
|
-
"anthropic.claude-3-sonnet-20240229-v1:0": {
|
|
227
|
-
id: "anthropic.claude-3-sonnet-20240229-v1:0",
|
|
228
|
-
name: "Claude Sonnet 3",
|
|
229
|
-
api: "bedrock-converse-stream",
|
|
230
|
-
provider: "amazon-bedrock",
|
|
231
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
232
|
-
reasoning: false,
|
|
233
|
-
input: ["text", "image"],
|
|
234
|
-
cost: {
|
|
235
|
-
input: 3,
|
|
236
|
-
output: 15,
|
|
237
|
-
cacheRead: 0,
|
|
238
|
-
cacheWrite: 0,
|
|
239
|
-
},
|
|
240
|
-
contextWindow: 200000,
|
|
241
|
-
maxTokens: 4096,
|
|
242
|
-
},
|
|
243
175
|
"anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
244
176
|
id: "anthropic.claude-haiku-4-5-20251001-v1:0",
|
|
245
177
|
name: "Claude Haiku 4.5",
|
|
@@ -322,7 +254,7 @@ export const MODELS = {
|
|
|
322
254
|
cacheRead: 0.5,
|
|
323
255
|
cacheWrite: 6.25,
|
|
324
256
|
},
|
|
325
|
-
contextWindow:
|
|
257
|
+
contextWindow: 1000000,
|
|
326
258
|
maxTokens: 128000,
|
|
327
259
|
},
|
|
328
260
|
"anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -373,43 +305,9 @@ export const MODELS = {
|
|
|
373
305
|
cacheRead: 0.3,
|
|
374
306
|
cacheWrite: 3.75,
|
|
375
307
|
},
|
|
376
|
-
contextWindow:
|
|
308
|
+
contextWindow: 1000000,
|
|
377
309
|
maxTokens: 64000,
|
|
378
310
|
},
|
|
379
|
-
"cohere.command-r-plus-v1:0": {
|
|
380
|
-
id: "cohere.command-r-plus-v1:0",
|
|
381
|
-
name: "Command R+",
|
|
382
|
-
api: "bedrock-converse-stream",
|
|
383
|
-
provider: "amazon-bedrock",
|
|
384
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
385
|
-
reasoning: false,
|
|
386
|
-
input: ["text"],
|
|
387
|
-
cost: {
|
|
388
|
-
input: 3,
|
|
389
|
-
output: 15,
|
|
390
|
-
cacheRead: 0,
|
|
391
|
-
cacheWrite: 0,
|
|
392
|
-
},
|
|
393
|
-
contextWindow: 128000,
|
|
394
|
-
maxTokens: 4096,
|
|
395
|
-
},
|
|
396
|
-
"cohere.command-r-v1:0": {
|
|
397
|
-
id: "cohere.command-r-v1:0",
|
|
398
|
-
name: "Command R",
|
|
399
|
-
api: "bedrock-converse-stream",
|
|
400
|
-
provider: "amazon-bedrock",
|
|
401
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
402
|
-
reasoning: false,
|
|
403
|
-
input: ["text"],
|
|
404
|
-
cost: {
|
|
405
|
-
input: 0.5,
|
|
406
|
-
output: 1.5,
|
|
407
|
-
cacheRead: 0,
|
|
408
|
-
cacheWrite: 0,
|
|
409
|
-
},
|
|
410
|
-
contextWindow: 128000,
|
|
411
|
-
maxTokens: 4096,
|
|
412
|
-
},
|
|
413
311
|
"deepseek.r1-v1:0": {
|
|
414
312
|
id: "deepseek.r1-v1:0",
|
|
415
313
|
name: "DeepSeek-R1",
|
|
@@ -444,8 +342,8 @@ export const MODELS = {
|
|
|
444
342
|
contextWindow: 163840,
|
|
445
343
|
maxTokens: 81920,
|
|
446
344
|
},
|
|
447
|
-
"deepseek.v3.2
|
|
448
|
-
id: "deepseek.v3.2
|
|
345
|
+
"deepseek.v3.2": {
|
|
346
|
+
id: "deepseek.v3.2",
|
|
449
347
|
name: "DeepSeek-V3.2",
|
|
450
348
|
api: "bedrock-converse-stream",
|
|
451
349
|
provider: "amazon-bedrock",
|
|
@@ -509,7 +407,7 @@ export const MODELS = {
|
|
|
509
407
|
cacheRead: 0.5,
|
|
510
408
|
cacheWrite: 6.25,
|
|
511
409
|
},
|
|
512
|
-
contextWindow:
|
|
410
|
+
contextWindow: 1000000,
|
|
513
411
|
maxTokens: 128000,
|
|
514
412
|
},
|
|
515
413
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -560,7 +458,7 @@ export const MODELS = {
|
|
|
560
458
|
cacheRead: 0.3,
|
|
561
459
|
cacheWrite: 3.75,
|
|
562
460
|
},
|
|
563
|
-
contextWindow:
|
|
461
|
+
contextWindow: 1000000,
|
|
564
462
|
maxTokens: 64000,
|
|
565
463
|
},
|
|
566
464
|
"global.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
@@ -611,7 +509,7 @@ export const MODELS = {
|
|
|
611
509
|
cacheRead: 0.5,
|
|
612
510
|
cacheWrite: 6.25,
|
|
613
511
|
},
|
|
614
|
-
contextWindow:
|
|
512
|
+
contextWindow: 1000000,
|
|
615
513
|
maxTokens: 128000,
|
|
616
514
|
},
|
|
617
515
|
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -662,7 +560,7 @@ export const MODELS = {
|
|
|
662
560
|
cacheRead: 0.3,
|
|
663
561
|
cacheWrite: 3.75,
|
|
664
562
|
},
|
|
665
|
-
contextWindow:
|
|
563
|
+
contextWindow: 1000000,
|
|
666
564
|
maxTokens: 64000,
|
|
667
565
|
},
|
|
668
566
|
"google.gemma-3-27b-it": {
|
|
@@ -699,6 +597,23 @@ export const MODELS = {
|
|
|
699
597
|
contextWindow: 128000,
|
|
700
598
|
maxTokens: 4096,
|
|
701
599
|
},
|
|
600
|
+
"meta.llama3-1-405b-instruct-v1:0": {
|
|
601
|
+
id: "meta.llama3-1-405b-instruct-v1:0",
|
|
602
|
+
name: "Llama 3.1 405B Instruct",
|
|
603
|
+
api: "bedrock-converse-stream",
|
|
604
|
+
provider: "amazon-bedrock",
|
|
605
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
606
|
+
reasoning: false,
|
|
607
|
+
input: ["text"],
|
|
608
|
+
cost: {
|
|
609
|
+
input: 2.4,
|
|
610
|
+
output: 2.4,
|
|
611
|
+
cacheRead: 0,
|
|
612
|
+
cacheWrite: 0,
|
|
613
|
+
},
|
|
614
|
+
contextWindow: 128000,
|
|
615
|
+
maxTokens: 4096,
|
|
616
|
+
},
|
|
702
617
|
"meta.llama3-1-70b-instruct-v1:0": {
|
|
703
618
|
id: "meta.llama3-1-70b-instruct-v1:0",
|
|
704
619
|
name: "Llama 3.1 70B Instruct",
|
|
@@ -886,6 +801,40 @@ export const MODELS = {
|
|
|
886
801
|
contextWindow: 204800,
|
|
887
802
|
maxTokens: 131072,
|
|
888
803
|
},
|
|
804
|
+
"mistral.devstral-2-123b": {
|
|
805
|
+
id: "mistral.devstral-2-123b",
|
|
806
|
+
name: "Devstral 2 123B",
|
|
807
|
+
api: "bedrock-converse-stream",
|
|
808
|
+
provider: "amazon-bedrock",
|
|
809
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
810
|
+
reasoning: false,
|
|
811
|
+
input: ["text"],
|
|
812
|
+
cost: {
|
|
813
|
+
input: 0.4,
|
|
814
|
+
output: 2,
|
|
815
|
+
cacheRead: 0,
|
|
816
|
+
cacheWrite: 0,
|
|
817
|
+
},
|
|
818
|
+
contextWindow: 256000,
|
|
819
|
+
maxTokens: 8192,
|
|
820
|
+
},
|
|
821
|
+
"mistral.magistral-small-2509": {
|
|
822
|
+
id: "mistral.magistral-small-2509",
|
|
823
|
+
name: "Magistral Small 1.2",
|
|
824
|
+
api: "bedrock-converse-stream",
|
|
825
|
+
provider: "amazon-bedrock",
|
|
826
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
827
|
+
reasoning: true,
|
|
828
|
+
input: ["text", "image"],
|
|
829
|
+
cost: {
|
|
830
|
+
input: 0.5,
|
|
831
|
+
output: 1.5,
|
|
832
|
+
cacheRead: 0,
|
|
833
|
+
cacheWrite: 0,
|
|
834
|
+
},
|
|
835
|
+
contextWindow: 128000,
|
|
836
|
+
maxTokens: 40000,
|
|
837
|
+
},
|
|
889
838
|
"mistral.ministral-3-14b-instruct": {
|
|
890
839
|
id: "mistral.ministral-3-14b-instruct",
|
|
891
840
|
name: "Ministral 14B 3.0",
|
|
@@ -903,6 +852,23 @@ export const MODELS = {
|
|
|
903
852
|
contextWindow: 128000,
|
|
904
853
|
maxTokens: 4096,
|
|
905
854
|
},
|
|
855
|
+
"mistral.ministral-3-3b-instruct": {
|
|
856
|
+
id: "mistral.ministral-3-3b-instruct",
|
|
857
|
+
name: "Ministral 3 3B",
|
|
858
|
+
api: "bedrock-converse-stream",
|
|
859
|
+
provider: "amazon-bedrock",
|
|
860
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
861
|
+
reasoning: false,
|
|
862
|
+
input: ["text", "image"],
|
|
863
|
+
cost: {
|
|
864
|
+
input: 0.1,
|
|
865
|
+
output: 0.1,
|
|
866
|
+
cacheRead: 0,
|
|
867
|
+
cacheWrite: 0,
|
|
868
|
+
},
|
|
869
|
+
contextWindow: 256000,
|
|
870
|
+
maxTokens: 8192,
|
|
871
|
+
},
|
|
906
872
|
"mistral.ministral-3-8b-instruct": {
|
|
907
873
|
id: "mistral.ministral-3-8b-instruct",
|
|
908
874
|
name: "Ministral 3 8B",
|
|
@@ -920,22 +886,39 @@ export const MODELS = {
|
|
|
920
886
|
contextWindow: 128000,
|
|
921
887
|
maxTokens: 4096,
|
|
922
888
|
},
|
|
923
|
-
"mistral.mistral-large-
|
|
924
|
-
id: "mistral.mistral-large-
|
|
925
|
-
name: "Mistral Large
|
|
889
|
+
"mistral.mistral-large-3-675b-instruct": {
|
|
890
|
+
id: "mistral.mistral-large-3-675b-instruct",
|
|
891
|
+
name: "Mistral Large 3",
|
|
926
892
|
api: "bedrock-converse-stream",
|
|
927
893
|
provider: "amazon-bedrock",
|
|
928
894
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
929
895
|
reasoning: false,
|
|
930
|
-
input: ["text"],
|
|
896
|
+
input: ["text", "image"],
|
|
931
897
|
cost: {
|
|
932
898
|
input: 0.5,
|
|
933
899
|
output: 1.5,
|
|
934
900
|
cacheRead: 0,
|
|
935
901
|
cacheWrite: 0,
|
|
936
902
|
},
|
|
903
|
+
contextWindow: 256000,
|
|
904
|
+
maxTokens: 8192,
|
|
905
|
+
},
|
|
906
|
+
"mistral.pixtral-large-2502-v1:0": {
|
|
907
|
+
id: "mistral.pixtral-large-2502-v1:0",
|
|
908
|
+
name: "Pixtral Large (25.02)",
|
|
909
|
+
api: "bedrock-converse-stream",
|
|
910
|
+
provider: "amazon-bedrock",
|
|
911
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
912
|
+
reasoning: false,
|
|
913
|
+
input: ["text", "image"],
|
|
914
|
+
cost: {
|
|
915
|
+
input: 2,
|
|
916
|
+
output: 6,
|
|
917
|
+
cacheRead: 0,
|
|
918
|
+
cacheWrite: 0,
|
|
919
|
+
},
|
|
937
920
|
contextWindow: 128000,
|
|
938
|
-
maxTokens:
|
|
921
|
+
maxTokens: 8192,
|
|
939
922
|
},
|
|
940
923
|
"mistral.voxtral-mini-3b-2507": {
|
|
941
924
|
id: "mistral.voxtral-mini-3b-2507",
|
|
@@ -1022,6 +1005,23 @@ export const MODELS = {
|
|
|
1022
1005
|
contextWindow: 128000,
|
|
1023
1006
|
maxTokens: 4096,
|
|
1024
1007
|
},
|
|
1008
|
+
"nvidia.nemotron-nano-3-30b": {
|
|
1009
|
+
id: "nvidia.nemotron-nano-3-30b",
|
|
1010
|
+
name: "NVIDIA Nemotron Nano 3 30B",
|
|
1011
|
+
api: "bedrock-converse-stream",
|
|
1012
|
+
provider: "amazon-bedrock",
|
|
1013
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1014
|
+
reasoning: true,
|
|
1015
|
+
input: ["text"],
|
|
1016
|
+
cost: {
|
|
1017
|
+
input: 0.06,
|
|
1018
|
+
output: 0.24,
|
|
1019
|
+
cacheRead: 0,
|
|
1020
|
+
cacheWrite: 0,
|
|
1021
|
+
},
|
|
1022
|
+
contextWindow: 128000,
|
|
1023
|
+
maxTokens: 4096,
|
|
1024
|
+
},
|
|
1025
1025
|
"nvidia.nemotron-nano-9b-v2": {
|
|
1026
1026
|
id: "nvidia.nemotron-nano-9b-v2",
|
|
1027
1027
|
name: "NVIDIA Nemotron Nano 9B v2",
|
|
@@ -1291,7 +1291,7 @@ export const MODELS = {
|
|
|
1291
1291
|
cacheRead: 0.5,
|
|
1292
1292
|
cacheWrite: 6.25,
|
|
1293
1293
|
},
|
|
1294
|
-
contextWindow:
|
|
1294
|
+
contextWindow: 1000000,
|
|
1295
1295
|
maxTokens: 128000,
|
|
1296
1296
|
},
|
|
1297
1297
|
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -1342,7 +1342,7 @@ export const MODELS = {
|
|
|
1342
1342
|
cacheRead: 0.3,
|
|
1343
1343
|
cacheWrite: 3.75,
|
|
1344
1344
|
},
|
|
1345
|
-
contextWindow:
|
|
1345
|
+
contextWindow: 1000000,
|
|
1346
1346
|
maxTokens: 64000,
|
|
1347
1347
|
},
|
|
1348
1348
|
"writer.palmyra-x4-v1:0": {
|
|
@@ -1718,23 +1718,6 @@ export const MODELS = {
|
|
|
1718
1718
|
cacheRead: 0.5,
|
|
1719
1719
|
cacheWrite: 6.25,
|
|
1720
1720
|
},
|
|
1721
|
-
contextWindow: 200000,
|
|
1722
|
-
maxTokens: 128000,
|
|
1723
|
-
},
|
|
1724
|
-
"claude-opus-4-6[1m]": {
|
|
1725
|
-
id: "claude-opus-4-6[1m]",
|
|
1726
|
-
name: "Claude Opus 4.6 (1M)",
|
|
1727
|
-
api: "anthropic-messages",
|
|
1728
|
-
provider: "anthropic",
|
|
1729
|
-
baseUrl: "https://api.anthropic.com",
|
|
1730
|
-
reasoning: true,
|
|
1731
|
-
input: ["text", "image"],
|
|
1732
|
-
cost: {
|
|
1733
|
-
input: 5,
|
|
1734
|
-
output: 25,
|
|
1735
|
-
cacheRead: 0.5,
|
|
1736
|
-
cacheWrite: 6.25,
|
|
1737
|
-
},
|
|
1738
1721
|
contextWindow: 1000000,
|
|
1739
1722
|
maxTokens: 128000,
|
|
1740
1723
|
},
|
|
@@ -1820,216 +1803,44 @@ export const MODELS = {
|
|
|
1820
1803
|
cacheRead: 0.3,
|
|
1821
1804
|
cacheWrite: 3.75,
|
|
1822
1805
|
},
|
|
1823
|
-
contextWindow:
|
|
1806
|
+
contextWindow: 1000000,
|
|
1824
1807
|
maxTokens: 64000,
|
|
1825
1808
|
},
|
|
1826
1809
|
},
|
|
1827
|
-
"
|
|
1828
|
-
"
|
|
1829
|
-
id: "
|
|
1830
|
-
name: "
|
|
1831
|
-
api: "
|
|
1832
|
-
provider: "
|
|
1833
|
-
baseUrl: "
|
|
1810
|
+
"azure-openai-responses": {
|
|
1811
|
+
"codex-mini-latest": {
|
|
1812
|
+
id: "codex-mini-latest",
|
|
1813
|
+
name: "Codex Mini",
|
|
1814
|
+
api: "azure-openai-responses",
|
|
1815
|
+
provider: "azure-openai-responses",
|
|
1816
|
+
baseUrl: "",
|
|
1834
1817
|
reasoning: true,
|
|
1835
|
-
input: ["text"
|
|
1818
|
+
input: ["text"],
|
|
1836
1819
|
cost: {
|
|
1837
|
-
input:
|
|
1838
|
-
output:
|
|
1839
|
-
cacheRead:
|
|
1840
|
-
cacheWrite:
|
|
1820
|
+
input: 1.5,
|
|
1821
|
+
output: 6,
|
|
1822
|
+
cacheRead: 0.375,
|
|
1823
|
+
cacheWrite: 0,
|
|
1841
1824
|
},
|
|
1842
1825
|
contextWindow: 200000,
|
|
1843
|
-
maxTokens:
|
|
1826
|
+
maxTokens: 100000,
|
|
1844
1827
|
},
|
|
1845
|
-
"
|
|
1846
|
-
id: "
|
|
1847
|
-
name: "
|
|
1848
|
-
api: "
|
|
1849
|
-
provider: "
|
|
1850
|
-
baseUrl: "
|
|
1851
|
-
reasoning:
|
|
1852
|
-
input: ["text"
|
|
1828
|
+
"gpt-4": {
|
|
1829
|
+
id: "gpt-4",
|
|
1830
|
+
name: "GPT-4",
|
|
1831
|
+
api: "azure-openai-responses",
|
|
1832
|
+
provider: "azure-openai-responses",
|
|
1833
|
+
baseUrl: "",
|
|
1834
|
+
reasoning: false,
|
|
1835
|
+
input: ["text"],
|
|
1853
1836
|
cost: {
|
|
1854
|
-
input:
|
|
1855
|
-
output:
|
|
1856
|
-
cacheRead:
|
|
1857
|
-
cacheWrite:
|
|
1837
|
+
input: 30,
|
|
1838
|
+
output: 60,
|
|
1839
|
+
cacheRead: 0,
|
|
1840
|
+
cacheWrite: 0,
|
|
1858
1841
|
},
|
|
1859
|
-
contextWindow:
|
|
1860
|
-
maxTokens:
|
|
1861
|
-
},
|
|
1862
|
-
"claude-sonnet-4-6": {
|
|
1863
|
-
id: "claude-sonnet-4-6",
|
|
1864
|
-
name: "Claude Sonnet 4.6 (Vertex)",
|
|
1865
|
-
api: "anthropic-vertex",
|
|
1866
|
-
provider: "anthropic-vertex",
|
|
1867
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1868
|
-
reasoning: true,
|
|
1869
|
-
input: ["text", "image"],
|
|
1870
|
-
cost: {
|
|
1871
|
-
input: 3,
|
|
1872
|
-
output: 15,
|
|
1873
|
-
cacheRead: 0.3,
|
|
1874
|
-
cacheWrite: 3.75,
|
|
1875
|
-
},
|
|
1876
|
-
contextWindow: 200000,
|
|
1877
|
-
maxTokens: 64000,
|
|
1878
|
-
},
|
|
1879
|
-
"claude-sonnet-4-6[1m]": {
|
|
1880
|
-
id: "claude-sonnet-4-6[1m]",
|
|
1881
|
-
name: "Claude Sonnet 4.6 1M (Vertex)",
|
|
1882
|
-
api: "anthropic-vertex",
|
|
1883
|
-
provider: "anthropic-vertex",
|
|
1884
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1885
|
-
reasoning: true,
|
|
1886
|
-
input: ["text", "image"],
|
|
1887
|
-
cost: {
|
|
1888
|
-
input: 3,
|
|
1889
|
-
output: 15,
|
|
1890
|
-
cacheRead: 0.3,
|
|
1891
|
-
cacheWrite: 3.75,
|
|
1892
|
-
},
|
|
1893
|
-
contextWindow: 1000000,
|
|
1894
|
-
maxTokens: 64000,
|
|
1895
|
-
},
|
|
1896
|
-
"claude-sonnet-4-5@20250929": {
|
|
1897
|
-
id: "claude-sonnet-4-5@20250929",
|
|
1898
|
-
name: "Claude Sonnet 4.5 (Vertex)",
|
|
1899
|
-
api: "anthropic-vertex",
|
|
1900
|
-
provider: "anthropic-vertex",
|
|
1901
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1902
|
-
reasoning: true,
|
|
1903
|
-
input: ["text", "image"],
|
|
1904
|
-
cost: {
|
|
1905
|
-
input: 3,
|
|
1906
|
-
output: 15,
|
|
1907
|
-
cacheRead: 0.3,
|
|
1908
|
-
cacheWrite: 3.75,
|
|
1909
|
-
},
|
|
1910
|
-
contextWindow: 200000,
|
|
1911
|
-
maxTokens: 64000,
|
|
1912
|
-
},
|
|
1913
|
-
"claude-sonnet-4@20250514": {
|
|
1914
|
-
id: "claude-sonnet-4@20250514",
|
|
1915
|
-
name: "Claude Sonnet 4 (Vertex)",
|
|
1916
|
-
api: "anthropic-vertex",
|
|
1917
|
-
provider: "anthropic-vertex",
|
|
1918
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1919
|
-
reasoning: true,
|
|
1920
|
-
input: ["text", "image"],
|
|
1921
|
-
cost: {
|
|
1922
|
-
input: 3,
|
|
1923
|
-
output: 15,
|
|
1924
|
-
cacheRead: 0.3,
|
|
1925
|
-
cacheWrite: 3.75,
|
|
1926
|
-
},
|
|
1927
|
-
contextWindow: 200000,
|
|
1928
|
-
maxTokens: 64000,
|
|
1929
|
-
},
|
|
1930
|
-
"claude-opus-4-5@20251101": {
|
|
1931
|
-
id: "claude-opus-4-5@20251101",
|
|
1932
|
-
name: "Claude Opus 4.5 (Vertex)",
|
|
1933
|
-
api: "anthropic-vertex",
|
|
1934
|
-
provider: "anthropic-vertex",
|
|
1935
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1936
|
-
reasoning: true,
|
|
1937
|
-
input: ["text", "image"],
|
|
1938
|
-
cost: {
|
|
1939
|
-
input: 15,
|
|
1940
|
-
output: 75,
|
|
1941
|
-
cacheRead: 1.5,
|
|
1942
|
-
cacheWrite: 18.75,
|
|
1943
|
-
},
|
|
1944
|
-
contextWindow: 200000,
|
|
1945
|
-
maxTokens: 32000,
|
|
1946
|
-
},
|
|
1947
|
-
"claude-opus-4-1@20250805": {
|
|
1948
|
-
id: "claude-opus-4-1@20250805",
|
|
1949
|
-
name: "Claude Opus 4.1 (Vertex)",
|
|
1950
|
-
api: "anthropic-vertex",
|
|
1951
|
-
provider: "anthropic-vertex",
|
|
1952
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1953
|
-
reasoning: true,
|
|
1954
|
-
input: ["text", "image"],
|
|
1955
|
-
cost: {
|
|
1956
|
-
input: 15,
|
|
1957
|
-
output: 75,
|
|
1958
|
-
cacheRead: 1.5,
|
|
1959
|
-
cacheWrite: 18.75,
|
|
1960
|
-
},
|
|
1961
|
-
contextWindow: 200000,
|
|
1962
|
-
maxTokens: 32000,
|
|
1963
|
-
},
|
|
1964
|
-
"claude-opus-4@20250514": {
|
|
1965
|
-
id: "claude-opus-4@20250514",
|
|
1966
|
-
name: "Claude Opus 4 (Vertex)",
|
|
1967
|
-
api: "anthropic-vertex",
|
|
1968
|
-
provider: "anthropic-vertex",
|
|
1969
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1970
|
-
reasoning: true,
|
|
1971
|
-
input: ["text", "image"],
|
|
1972
|
-
cost: {
|
|
1973
|
-
input: 15,
|
|
1974
|
-
output: 75,
|
|
1975
|
-
cacheRead: 1.5,
|
|
1976
|
-
cacheWrite: 18.75,
|
|
1977
|
-
},
|
|
1978
|
-
contextWindow: 200000,
|
|
1979
|
-
maxTokens: 32000,
|
|
1980
|
-
},
|
|
1981
|
-
"claude-haiku-4-5@20251001": {
|
|
1982
|
-
id: "claude-haiku-4-5@20251001",
|
|
1983
|
-
name: "Claude Haiku 4.5 (Vertex)",
|
|
1984
|
-
api: "anthropic-vertex",
|
|
1985
|
-
provider: "anthropic-vertex",
|
|
1986
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1987
|
-
reasoning: false,
|
|
1988
|
-
input: ["text", "image"],
|
|
1989
|
-
cost: {
|
|
1990
|
-
input: 0.8,
|
|
1991
|
-
output: 4,
|
|
1992
|
-
cacheRead: 0.08,
|
|
1993
|
-
cacheWrite: 1,
|
|
1994
|
-
},
|
|
1995
|
-
contextWindow: 200000,
|
|
1996
|
-
maxTokens: 8192,
|
|
1997
|
-
},
|
|
1998
|
-
},
|
|
1999
|
-
"azure-openai-responses": {
|
|
2000
|
-
"codex-mini-latest": {
|
|
2001
|
-
id: "codex-mini-latest",
|
|
2002
|
-
name: "Codex Mini",
|
|
2003
|
-
api: "azure-openai-responses",
|
|
2004
|
-
provider: "azure-openai-responses",
|
|
2005
|
-
baseUrl: "",
|
|
2006
|
-
reasoning: true,
|
|
2007
|
-
input: ["text"],
|
|
2008
|
-
cost: {
|
|
2009
|
-
input: 1.5,
|
|
2010
|
-
output: 6,
|
|
2011
|
-
cacheRead: 0.375,
|
|
2012
|
-
cacheWrite: 0,
|
|
2013
|
-
},
|
|
2014
|
-
contextWindow: 200000,
|
|
2015
|
-
maxTokens: 100000,
|
|
2016
|
-
},
|
|
2017
|
-
"gpt-4": {
|
|
2018
|
-
id: "gpt-4",
|
|
2019
|
-
name: "GPT-4",
|
|
2020
|
-
api: "azure-openai-responses",
|
|
2021
|
-
provider: "azure-openai-responses",
|
|
2022
|
-
baseUrl: "",
|
|
2023
|
-
reasoning: false,
|
|
2024
|
-
input: ["text"],
|
|
2025
|
-
cost: {
|
|
2026
|
-
input: 30,
|
|
2027
|
-
output: 60,
|
|
2028
|
-
cacheRead: 0,
|
|
2029
|
-
cacheWrite: 0,
|
|
2030
|
-
},
|
|
2031
|
-
contextWindow: 8192,
|
|
2032
|
-
maxTokens: 8192,
|
|
1842
|
+
contextWindow: 8192,
|
|
1843
|
+
maxTokens: 8192,
|
|
2033
1844
|
},
|
|
2034
1845
|
"gpt-4-turbo": {
|
|
2035
1846
|
id: "gpt-4-turbo",
|
|
@@ -2490,6 +2301,40 @@ export const MODELS = {
|
|
|
2490
2301
|
contextWindow: 272000,
|
|
2491
2302
|
maxTokens: 128000,
|
|
2492
2303
|
},
|
|
2304
|
+
"gpt-5.4-mini": {
|
|
2305
|
+
id: "gpt-5.4-mini",
|
|
2306
|
+
name: "GPT-5.4 mini",
|
|
2307
|
+
api: "azure-openai-responses",
|
|
2308
|
+
provider: "azure-openai-responses",
|
|
2309
|
+
baseUrl: "",
|
|
2310
|
+
reasoning: true,
|
|
2311
|
+
input: ["text", "image"],
|
|
2312
|
+
cost: {
|
|
2313
|
+
input: 0.75,
|
|
2314
|
+
output: 4.5,
|
|
2315
|
+
cacheRead: 0.075,
|
|
2316
|
+
cacheWrite: 0,
|
|
2317
|
+
},
|
|
2318
|
+
contextWindow: 400000,
|
|
2319
|
+
maxTokens: 128000,
|
|
2320
|
+
},
|
|
2321
|
+
"gpt-5.4-nano": {
|
|
2322
|
+
id: "gpt-5.4-nano",
|
|
2323
|
+
name: "GPT-5.4 nano",
|
|
2324
|
+
api: "azure-openai-responses",
|
|
2325
|
+
provider: "azure-openai-responses",
|
|
2326
|
+
baseUrl: "",
|
|
2327
|
+
reasoning: true,
|
|
2328
|
+
input: ["text", "image"],
|
|
2329
|
+
cost: {
|
|
2330
|
+
input: 0.2,
|
|
2331
|
+
output: 1.25,
|
|
2332
|
+
cacheRead: 0.02,
|
|
2333
|
+
cacheWrite: 0,
|
|
2334
|
+
},
|
|
2335
|
+
contextWindow: 400000,
|
|
2336
|
+
maxTokens: 128000,
|
|
2337
|
+
},
|
|
2493
2338
|
"gpt-5.4-pro": {
|
|
2494
2339
|
id: "gpt-5.4-pro",
|
|
2495
2340
|
name: "GPT-5.4 Pro",
|
|
@@ -2730,7 +2575,7 @@ export const MODELS = {
|
|
|
2730
2575
|
cacheRead: 0,
|
|
2731
2576
|
cacheWrite: 0,
|
|
2732
2577
|
},
|
|
2733
|
-
contextWindow:
|
|
2578
|
+
contextWindow: 144000,
|
|
2734
2579
|
maxTokens: 32000,
|
|
2735
2580
|
},
|
|
2736
2581
|
"claude-opus-4.5": {
|
|
@@ -2748,7 +2593,7 @@ export const MODELS = {
|
|
|
2748
2593
|
cacheRead: 0,
|
|
2749
2594
|
cacheWrite: 0,
|
|
2750
2595
|
},
|
|
2751
|
-
contextWindow:
|
|
2596
|
+
contextWindow: 160000,
|
|
2752
2597
|
maxTokens: 32000,
|
|
2753
2598
|
},
|
|
2754
2599
|
"claude-opus-4.6": {
|
|
@@ -2766,7 +2611,7 @@ export const MODELS = {
|
|
|
2766
2611
|
cacheRead: 0,
|
|
2767
2612
|
cacheWrite: 0,
|
|
2768
2613
|
},
|
|
2769
|
-
contextWindow:
|
|
2614
|
+
contextWindow: 144000,
|
|
2770
2615
|
maxTokens: 64000,
|
|
2771
2616
|
},
|
|
2772
2617
|
"claude-sonnet-4": {
|
|
@@ -2784,7 +2629,7 @@ export const MODELS = {
|
|
|
2784
2629
|
cacheRead: 0,
|
|
2785
2630
|
cacheWrite: 0,
|
|
2786
2631
|
},
|
|
2787
|
-
contextWindow:
|
|
2632
|
+
contextWindow: 216000,
|
|
2788
2633
|
maxTokens: 16000,
|
|
2789
2634
|
},
|
|
2790
2635
|
"claude-sonnet-4.5": {
|
|
@@ -2802,7 +2647,7 @@ export const MODELS = {
|
|
|
2802
2647
|
cacheRead: 0,
|
|
2803
2648
|
cacheWrite: 0,
|
|
2804
2649
|
},
|
|
2805
|
-
contextWindow:
|
|
2650
|
+
contextWindow: 144000,
|
|
2806
2651
|
maxTokens: 32000,
|
|
2807
2652
|
},
|
|
2808
2653
|
"claude-sonnet-4.6": {
|
|
@@ -2820,7 +2665,7 @@ export const MODELS = {
|
|
|
2820
2665
|
cacheRead: 0,
|
|
2821
2666
|
cacheWrite: 0,
|
|
2822
2667
|
},
|
|
2823
|
-
contextWindow:
|
|
2668
|
+
contextWindow: 200000,
|
|
2824
2669
|
maxTokens: 32000,
|
|
2825
2670
|
},
|
|
2826
2671
|
"gemini-2.5-pro": {
|
|
@@ -2915,7 +2760,7 @@ export const MODELS = {
|
|
|
2915
2760
|
cacheRead: 0,
|
|
2916
2761
|
cacheWrite: 0,
|
|
2917
2762
|
},
|
|
2918
|
-
contextWindow:
|
|
2763
|
+
contextWindow: 128000,
|
|
2919
2764
|
maxTokens: 16384,
|
|
2920
2765
|
},
|
|
2921
2766
|
"gpt-4o": {
|
|
@@ -2934,8 +2779,8 @@ export const MODELS = {
|
|
|
2934
2779
|
cacheRead: 0,
|
|
2935
2780
|
cacheWrite: 0,
|
|
2936
2781
|
},
|
|
2937
|
-
contextWindow:
|
|
2938
|
-
maxTokens:
|
|
2782
|
+
contextWindow: 128000,
|
|
2783
|
+
maxTokens: 4096,
|
|
2939
2784
|
},
|
|
2940
2785
|
"gpt-5": {
|
|
2941
2786
|
id: "gpt-5",
|
|
@@ -2970,7 +2815,7 @@ export const MODELS = {
|
|
|
2970
2815
|
cacheRead: 0,
|
|
2971
2816
|
cacheWrite: 0,
|
|
2972
2817
|
},
|
|
2973
|
-
contextWindow:
|
|
2818
|
+
contextWindow: 264000,
|
|
2974
2819
|
maxTokens: 64000,
|
|
2975
2820
|
},
|
|
2976
2821
|
"gpt-5.1": {
|
|
@@ -2988,7 +2833,7 @@ export const MODELS = {
|
|
|
2988
2833
|
cacheRead: 0,
|
|
2989
2834
|
cacheWrite: 0,
|
|
2990
2835
|
},
|
|
2991
|
-
contextWindow:
|
|
2836
|
+
contextWindow: 264000,
|
|
2992
2837
|
maxTokens: 64000,
|
|
2993
2838
|
},
|
|
2994
2839
|
"gpt-5.1-codex": {
|
|
@@ -3006,7 +2851,7 @@ export const MODELS = {
|
|
|
3006
2851
|
cacheRead: 0,
|
|
3007
2852
|
cacheWrite: 0,
|
|
3008
2853
|
},
|
|
3009
|
-
contextWindow:
|
|
2854
|
+
contextWindow: 400000,
|
|
3010
2855
|
maxTokens: 128000,
|
|
3011
2856
|
},
|
|
3012
2857
|
"gpt-5.1-codex-max": {
|
|
@@ -3024,7 +2869,7 @@ export const MODELS = {
|
|
|
3024
2869
|
cacheRead: 0,
|
|
3025
2870
|
cacheWrite: 0,
|
|
3026
2871
|
},
|
|
3027
|
-
contextWindow:
|
|
2872
|
+
contextWindow: 400000,
|
|
3028
2873
|
maxTokens: 128000,
|
|
3029
2874
|
},
|
|
3030
2875
|
"gpt-5.1-codex-mini": {
|
|
@@ -3042,7 +2887,7 @@ export const MODELS = {
|
|
|
3042
2887
|
cacheRead: 0,
|
|
3043
2888
|
cacheWrite: 0,
|
|
3044
2889
|
},
|
|
3045
|
-
contextWindow:
|
|
2890
|
+
contextWindow: 400000,
|
|
3046
2891
|
maxTokens: 128000,
|
|
3047
2892
|
},
|
|
3048
2893
|
"gpt-5.2": {
|
|
@@ -3060,7 +2905,7 @@ export const MODELS = {
|
|
|
3060
2905
|
cacheRead: 0,
|
|
3061
2906
|
cacheWrite: 0,
|
|
3062
2907
|
},
|
|
3063
|
-
contextWindow:
|
|
2908
|
+
contextWindow: 264000,
|
|
3064
2909
|
maxTokens: 64000,
|
|
3065
2910
|
},
|
|
3066
2911
|
"gpt-5.2-codex": {
|
|
@@ -3078,7 +2923,7 @@ export const MODELS = {
|
|
|
3078
2923
|
cacheRead: 0,
|
|
3079
2924
|
cacheWrite: 0,
|
|
3080
2925
|
},
|
|
3081
|
-
contextWindow:
|
|
2926
|
+
contextWindow: 400000,
|
|
3082
2927
|
maxTokens: 128000,
|
|
3083
2928
|
},
|
|
3084
2929
|
"gpt-5.3-codex": {
|
|
@@ -3117,6 +2962,24 @@ export const MODELS = {
|
|
|
3117
2962
|
contextWindow: 400000,
|
|
3118
2963
|
maxTokens: 128000,
|
|
3119
2964
|
},
|
|
2965
|
+
"gpt-5.4-mini": {
|
|
2966
|
+
id: "gpt-5.4-mini",
|
|
2967
|
+
name: "GPT-5.4 mini",
|
|
2968
|
+
api: "openai-responses",
|
|
2969
|
+
provider: "github-copilot",
|
|
2970
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
2971
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
2972
|
+
reasoning: true,
|
|
2973
|
+
input: ["text", "image"],
|
|
2974
|
+
cost: {
|
|
2975
|
+
input: 0,
|
|
2976
|
+
output: 0,
|
|
2977
|
+
cacheRead: 0,
|
|
2978
|
+
cacheWrite: 0,
|
|
2979
|
+
},
|
|
2980
|
+
contextWindow: 400000,
|
|
2981
|
+
maxTokens: 128000,
|
|
2982
|
+
},
|
|
3120
2983
|
"grok-code-fast-1": {
|
|
3121
2984
|
id: "grok-code-fast-1",
|
|
3122
2985
|
name: "Grok Code Fast 1",
|
|
@@ -3436,10 +3299,10 @@ export const MODELS = {
|
|
|
3436
3299
|
reasoning: true,
|
|
3437
3300
|
input: ["text", "image"],
|
|
3438
3301
|
cost: {
|
|
3439
|
-
input: 0,
|
|
3440
|
-
output:
|
|
3441
|
-
cacheRead: 0,
|
|
3442
|
-
cacheWrite:
|
|
3302
|
+
input: 0.25,
|
|
3303
|
+
output: 1.5,
|
|
3304
|
+
cacheRead: 0.025,
|
|
3305
|
+
cacheWrite: 1,
|
|
3443
3306
|
},
|
|
3444
3307
|
contextWindow: 1048576,
|
|
3445
3308
|
maxTokens: 65536,
|
|
@@ -4700,6 +4563,40 @@ export const MODELS = {
|
|
|
4700
4563
|
contextWindow: 204800,
|
|
4701
4564
|
maxTokens: 131072,
|
|
4702
4565
|
},
|
|
4566
|
+
"MiniMax-M2.7": {
|
|
4567
|
+
id: "MiniMax-M2.7",
|
|
4568
|
+
name: "MiniMax-M2.7",
|
|
4569
|
+
api: "anthropic-messages",
|
|
4570
|
+
provider: "minimax",
|
|
4571
|
+
baseUrl: "https://api.minimax.io/anthropic",
|
|
4572
|
+
reasoning: true,
|
|
4573
|
+
input: ["text"],
|
|
4574
|
+
cost: {
|
|
4575
|
+
input: 0.3,
|
|
4576
|
+
output: 1.2,
|
|
4577
|
+
cacheRead: 0.06,
|
|
4578
|
+
cacheWrite: 0.375,
|
|
4579
|
+
},
|
|
4580
|
+
contextWindow: 204800,
|
|
4581
|
+
maxTokens: 131072,
|
|
4582
|
+
},
|
|
4583
|
+
"MiniMax-M2.7-highspeed": {
|
|
4584
|
+
id: "MiniMax-M2.7-highspeed",
|
|
4585
|
+
name: "MiniMax-M2.7-highspeed",
|
|
4586
|
+
api: "anthropic-messages",
|
|
4587
|
+
provider: "minimax",
|
|
4588
|
+
baseUrl: "https://api.minimax.io/anthropic",
|
|
4589
|
+
reasoning: true,
|
|
4590
|
+
input: ["text"],
|
|
4591
|
+
cost: {
|
|
4592
|
+
input: 0.6,
|
|
4593
|
+
output: 2.4,
|
|
4594
|
+
cacheRead: 0.06,
|
|
4595
|
+
cacheWrite: 0.375,
|
|
4596
|
+
},
|
|
4597
|
+
contextWindow: 204800,
|
|
4598
|
+
maxTokens: 131072,
|
|
4599
|
+
},
|
|
4703
4600
|
},
|
|
4704
4601
|
"minimax-cn": {
|
|
4705
4602
|
"MiniMax-M2": {
|
|
@@ -4770,11 +4667,45 @@ export const MODELS = {
|
|
|
4770
4667
|
contextWindow: 204800,
|
|
4771
4668
|
maxTokens: 131072,
|
|
4772
4669
|
},
|
|
4670
|
+
"MiniMax-M2.7": {
|
|
4671
|
+
id: "MiniMax-M2.7",
|
|
4672
|
+
name: "MiniMax-M2.7",
|
|
4673
|
+
api: "anthropic-messages",
|
|
4674
|
+
provider: "minimax-cn",
|
|
4675
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4676
|
+
reasoning: true,
|
|
4677
|
+
input: ["text"],
|
|
4678
|
+
cost: {
|
|
4679
|
+
input: 0.3,
|
|
4680
|
+
output: 1.2,
|
|
4681
|
+
cacheRead: 0.06,
|
|
4682
|
+
cacheWrite: 0.375,
|
|
4683
|
+
},
|
|
4684
|
+
contextWindow: 204800,
|
|
4685
|
+
maxTokens: 131072,
|
|
4686
|
+
},
|
|
4687
|
+
"MiniMax-M2.7-highspeed": {
|
|
4688
|
+
id: "MiniMax-M2.7-highspeed",
|
|
4689
|
+
name: "MiniMax-M2.7-highspeed",
|
|
4690
|
+
api: "anthropic-messages",
|
|
4691
|
+
provider: "minimax-cn",
|
|
4692
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4693
|
+
reasoning: true,
|
|
4694
|
+
input: ["text"],
|
|
4695
|
+
cost: {
|
|
4696
|
+
input: 0.6,
|
|
4697
|
+
output: 2.4,
|
|
4698
|
+
cacheRead: 0.06,
|
|
4699
|
+
cacheWrite: 0.375,
|
|
4700
|
+
},
|
|
4701
|
+
contextWindow: 204800,
|
|
4702
|
+
maxTokens: 131072,
|
|
4703
|
+
},
|
|
4773
4704
|
},
|
|
4774
4705
|
"mistral": {
|
|
4775
4706
|
"codestral-latest": {
|
|
4776
4707
|
id: "codestral-latest",
|
|
4777
|
-
name: "Codestral",
|
|
4708
|
+
name: "Codestral (latest)",
|
|
4778
4709
|
api: "mistral-conversations",
|
|
4779
4710
|
provider: "mistral",
|
|
4780
4711
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4825,7 +4756,7 @@ export const MODELS = {
|
|
|
4825
4756
|
},
|
|
4826
4757
|
"devstral-medium-latest": {
|
|
4827
4758
|
id: "devstral-medium-latest",
|
|
4828
|
-
name: "Devstral 2",
|
|
4759
|
+
name: "Devstral 2 (latest)",
|
|
4829
4760
|
api: "mistral-conversations",
|
|
4830
4761
|
provider: "mistral",
|
|
4831
4762
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4893,7 +4824,7 @@ export const MODELS = {
|
|
|
4893
4824
|
},
|
|
4894
4825
|
"magistral-medium-latest": {
|
|
4895
4826
|
id: "magistral-medium-latest",
|
|
4896
|
-
name: "Magistral Medium",
|
|
4827
|
+
name: "Magistral Medium (latest)",
|
|
4897
4828
|
api: "mistral-conversations",
|
|
4898
4829
|
provider: "mistral",
|
|
4899
4830
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4927,7 +4858,7 @@ export const MODELS = {
|
|
|
4927
4858
|
},
|
|
4928
4859
|
"ministral-3b-latest": {
|
|
4929
4860
|
id: "ministral-3b-latest",
|
|
4930
|
-
name: "Ministral 3B",
|
|
4861
|
+
name: "Ministral 3B (latest)",
|
|
4931
4862
|
api: "mistral-conversations",
|
|
4932
4863
|
provider: "mistral",
|
|
4933
4864
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4944,7 +4875,7 @@ export const MODELS = {
|
|
|
4944
4875
|
},
|
|
4945
4876
|
"ministral-8b-latest": {
|
|
4946
4877
|
id: "ministral-8b-latest",
|
|
4947
|
-
name: "Ministral 8B",
|
|
4878
|
+
name: "Ministral 8B (latest)",
|
|
4948
4879
|
api: "mistral-conversations",
|
|
4949
4880
|
provider: "mistral",
|
|
4950
4881
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4995,7 +4926,7 @@ export const MODELS = {
|
|
|
4995
4926
|
},
|
|
4996
4927
|
"mistral-large-latest": {
|
|
4997
4928
|
id: "mistral-large-latest",
|
|
4998
|
-
name: "Mistral Large",
|
|
4929
|
+
name: "Mistral Large (latest)",
|
|
4999
4930
|
api: "mistral-conversations",
|
|
5000
4931
|
provider: "mistral",
|
|
5001
4932
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5046,7 +4977,7 @@ export const MODELS = {
|
|
|
5046
4977
|
},
|
|
5047
4978
|
"mistral-medium-latest": {
|
|
5048
4979
|
id: "mistral-medium-latest",
|
|
5049
|
-
name: "Mistral Medium",
|
|
4980
|
+
name: "Mistral Medium (latest)",
|
|
5050
4981
|
api: "mistral-conversations",
|
|
5051
4982
|
provider: "mistral",
|
|
5052
4983
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5097,7 +5028,7 @@ export const MODELS = {
|
|
|
5097
5028
|
},
|
|
5098
5029
|
"mistral-small-latest": {
|
|
5099
5030
|
id: "mistral-small-latest",
|
|
5100
|
-
name: "Mistral Small",
|
|
5031
|
+
name: "Mistral Small (latest)",
|
|
5101
5032
|
api: "mistral-conversations",
|
|
5102
5033
|
provider: "mistral",
|
|
5103
5034
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5182,7 +5113,7 @@ export const MODELS = {
|
|
|
5182
5113
|
},
|
|
5183
5114
|
"pixtral-large-latest": {
|
|
5184
5115
|
id: "pixtral-large-latest",
|
|
5185
|
-
name: "Pixtral Large",
|
|
5116
|
+
name: "Pixtral Large (latest)",
|
|
5186
5117
|
api: "mistral-conversations",
|
|
5187
5118
|
provider: "mistral",
|
|
5188
5119
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5692,6 +5623,40 @@ export const MODELS = {
|
|
|
5692
5623
|
contextWindow: 272000,
|
|
5693
5624
|
maxTokens: 128000,
|
|
5694
5625
|
},
|
|
5626
|
+
"gpt-5.4-mini": {
|
|
5627
|
+
id: "gpt-5.4-mini",
|
|
5628
|
+
name: "GPT-5.4 mini",
|
|
5629
|
+
api: "openai-responses",
|
|
5630
|
+
provider: "openai",
|
|
5631
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5632
|
+
reasoning: true,
|
|
5633
|
+
input: ["text", "image"],
|
|
5634
|
+
cost: {
|
|
5635
|
+
input: 0.75,
|
|
5636
|
+
output: 4.5,
|
|
5637
|
+
cacheRead: 0.075,
|
|
5638
|
+
cacheWrite: 0,
|
|
5639
|
+
},
|
|
5640
|
+
contextWindow: 400000,
|
|
5641
|
+
maxTokens: 128000,
|
|
5642
|
+
},
|
|
5643
|
+
"gpt-5.4-nano": {
|
|
5644
|
+
id: "gpt-5.4-nano",
|
|
5645
|
+
name: "GPT-5.4 nano",
|
|
5646
|
+
api: "openai-responses",
|
|
5647
|
+
provider: "openai",
|
|
5648
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5649
|
+
reasoning: true,
|
|
5650
|
+
input: ["text", "image"],
|
|
5651
|
+
cost: {
|
|
5652
|
+
input: 0.2,
|
|
5653
|
+
output: 1.25,
|
|
5654
|
+
cacheRead: 0.02,
|
|
5655
|
+
cacheWrite: 0,
|
|
5656
|
+
},
|
|
5657
|
+
contextWindow: 400000,
|
|
5658
|
+
maxTokens: 128000,
|
|
5659
|
+
},
|
|
5695
5660
|
"gpt-5.4-pro": {
|
|
5696
5661
|
id: "gpt-5.4-pro",
|
|
5697
5662
|
name: "GPT-5.4 Pro",
|
|
@@ -6084,7 +6049,7 @@ export const MODELS = {
|
|
|
6084
6049
|
cacheRead: 0.5,
|
|
6085
6050
|
cacheWrite: 6.25,
|
|
6086
6051
|
},
|
|
6087
|
-
contextWindow:
|
|
6052
|
+
contextWindow: 1000000,
|
|
6088
6053
|
maxTokens: 128000,
|
|
6089
6054
|
},
|
|
6090
6055
|
"claude-sonnet-4": {
|
|
@@ -6155,23 +6120,6 @@ export const MODELS = {
|
|
|
6155
6120
|
contextWindow: 1048576,
|
|
6156
6121
|
maxTokens: 65536,
|
|
6157
6122
|
},
|
|
6158
|
-
"gemini-3-pro": {
|
|
6159
|
-
id: "gemini-3-pro",
|
|
6160
|
-
name: "Gemini 3 Pro",
|
|
6161
|
-
api: "google-generative-ai",
|
|
6162
|
-
provider: "opencode",
|
|
6163
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6164
|
-
reasoning: true,
|
|
6165
|
-
input: ["text", "image"],
|
|
6166
|
-
cost: {
|
|
6167
|
-
input: 2,
|
|
6168
|
-
output: 12,
|
|
6169
|
-
cacheRead: 0.2,
|
|
6170
|
-
cacheWrite: 0,
|
|
6171
|
-
},
|
|
6172
|
-
contextWindow: 1048576,
|
|
6173
|
-
maxTokens: 65536,
|
|
6174
|
-
},
|
|
6175
6123
|
"gemini-3.1-pro": {
|
|
6176
6124
|
id: "gemini-3.1-pro",
|
|
6177
6125
|
name: "Gemini 3.1 Pro Preview",
|
|
@@ -6189,40 +6137,6 @@ export const MODELS = {
|
|
|
6189
6137
|
contextWindow: 1048576,
|
|
6190
6138
|
maxTokens: 65536,
|
|
6191
6139
|
},
|
|
6192
|
-
"glm-4.6": {
|
|
6193
|
-
id: "glm-4.6",
|
|
6194
|
-
name: "GLM-4.6",
|
|
6195
|
-
api: "openai-completions",
|
|
6196
|
-
provider: "opencode",
|
|
6197
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6198
|
-
reasoning: true,
|
|
6199
|
-
input: ["text"],
|
|
6200
|
-
cost: {
|
|
6201
|
-
input: 0.6,
|
|
6202
|
-
output: 2.2,
|
|
6203
|
-
cacheRead: 0.1,
|
|
6204
|
-
cacheWrite: 0,
|
|
6205
|
-
},
|
|
6206
|
-
contextWindow: 204800,
|
|
6207
|
-
maxTokens: 131072,
|
|
6208
|
-
},
|
|
6209
|
-
"glm-4.7": {
|
|
6210
|
-
id: "glm-4.7",
|
|
6211
|
-
name: "GLM-4.7",
|
|
6212
|
-
api: "openai-completions",
|
|
6213
|
-
provider: "opencode",
|
|
6214
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6215
|
-
reasoning: true,
|
|
6216
|
-
input: ["text"],
|
|
6217
|
-
cost: {
|
|
6218
|
-
input: 0.6,
|
|
6219
|
-
output: 2.2,
|
|
6220
|
-
cacheRead: 0.1,
|
|
6221
|
-
cacheWrite: 0,
|
|
6222
|
-
},
|
|
6223
|
-
contextWindow: 204800,
|
|
6224
|
-
maxTokens: 131072,
|
|
6225
|
-
},
|
|
6226
6140
|
"glm-5": {
|
|
6227
6141
|
id: "glm-5",
|
|
6228
6142
|
name: "GLM-5",
|
|
@@ -6427,6 +6341,40 @@ export const MODELS = {
|
|
|
6427
6341
|
contextWindow: 272000,
|
|
6428
6342
|
maxTokens: 128000,
|
|
6429
6343
|
},
|
|
6344
|
+
"gpt-5.4-mini": {
|
|
6345
|
+
id: "gpt-5.4-mini",
|
|
6346
|
+
name: "GPT-5.4 Mini",
|
|
6347
|
+
api: "openai-responses",
|
|
6348
|
+
provider: "opencode",
|
|
6349
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6350
|
+
reasoning: true,
|
|
6351
|
+
input: ["text", "image"],
|
|
6352
|
+
cost: {
|
|
6353
|
+
input: 0.75,
|
|
6354
|
+
output: 4.5,
|
|
6355
|
+
cacheRead: 0.075,
|
|
6356
|
+
cacheWrite: 0,
|
|
6357
|
+
},
|
|
6358
|
+
contextWindow: 400000,
|
|
6359
|
+
maxTokens: 128000,
|
|
6360
|
+
},
|
|
6361
|
+
"gpt-5.4-nano": {
|
|
6362
|
+
id: "gpt-5.4-nano",
|
|
6363
|
+
name: "GPT-5.4 Nano",
|
|
6364
|
+
api: "openai-responses",
|
|
6365
|
+
provider: "opencode",
|
|
6366
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6367
|
+
reasoning: true,
|
|
6368
|
+
input: ["text", "image"],
|
|
6369
|
+
cost: {
|
|
6370
|
+
input: 0.2,
|
|
6371
|
+
output: 1.25,
|
|
6372
|
+
cacheRead: 0.02,
|
|
6373
|
+
cacheWrite: 0,
|
|
6374
|
+
},
|
|
6375
|
+
contextWindow: 400000,
|
|
6376
|
+
maxTokens: 128000,
|
|
6377
|
+
},
|
|
6430
6378
|
"gpt-5.4-pro": {
|
|
6431
6379
|
id: "gpt-5.4-pro",
|
|
6432
6380
|
name: "GPT-5.4 Pro",
|
|
@@ -6461,22 +6409,39 @@ export const MODELS = {
|
|
|
6461
6409
|
contextWindow: 262144,
|
|
6462
6410
|
maxTokens: 65536,
|
|
6463
6411
|
},
|
|
6464
|
-
"
|
|
6465
|
-
id: "
|
|
6466
|
-
name: "
|
|
6412
|
+
"mimo-v2-omni-free": {
|
|
6413
|
+
id: "mimo-v2-omni-free",
|
|
6414
|
+
name: "MiMo V2 Omni Free",
|
|
6415
|
+
api: "openai-completions",
|
|
6416
|
+
provider: "opencode",
|
|
6417
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6418
|
+
reasoning: true,
|
|
6419
|
+
input: ["text", "image"],
|
|
6420
|
+
cost: {
|
|
6421
|
+
input: 0,
|
|
6422
|
+
output: 0,
|
|
6423
|
+
cacheRead: 0,
|
|
6424
|
+
cacheWrite: 0,
|
|
6425
|
+
},
|
|
6426
|
+
contextWindow: 262144,
|
|
6427
|
+
maxTokens: 64000,
|
|
6428
|
+
},
|
|
6429
|
+
"mimo-v2-pro-free": {
|
|
6430
|
+
id: "mimo-v2-pro-free",
|
|
6431
|
+
name: "MiMo V2 Pro Free",
|
|
6467
6432
|
api: "openai-completions",
|
|
6468
6433
|
provider: "opencode",
|
|
6469
6434
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
6470
6435
|
reasoning: true,
|
|
6471
6436
|
input: ["text"],
|
|
6472
6437
|
cost: {
|
|
6473
|
-
input: 0
|
|
6474
|
-
output:
|
|
6475
|
-
cacheRead: 0
|
|
6438
|
+
input: 0,
|
|
6439
|
+
output: 0,
|
|
6440
|
+
cacheRead: 0,
|
|
6476
6441
|
cacheWrite: 0,
|
|
6477
6442
|
},
|
|
6478
|
-
contextWindow:
|
|
6479
|
-
maxTokens:
|
|
6443
|
+
contextWindow: 1048576,
|
|
6444
|
+
maxTokens: 64000,
|
|
6480
6445
|
},
|
|
6481
6446
|
"minimax-m2.5": {
|
|
6482
6447
|
id: "minimax-m2.5",
|
|
@@ -6512,6 +6477,23 @@ export const MODELS = {
|
|
|
6512
6477
|
contextWindow: 204800,
|
|
6513
6478
|
maxTokens: 131072,
|
|
6514
6479
|
},
|
|
6480
|
+
"nemotron-3-super-free": {
|
|
6481
|
+
id: "nemotron-3-super-free",
|
|
6482
|
+
name: "Nemotron 3 Super Free",
|
|
6483
|
+
api: "openai-completions",
|
|
6484
|
+
provider: "opencode",
|
|
6485
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6486
|
+
reasoning: true,
|
|
6487
|
+
input: ["text"],
|
|
6488
|
+
cost: {
|
|
6489
|
+
input: 0,
|
|
6490
|
+
output: 0,
|
|
6491
|
+
cacheRead: 0,
|
|
6492
|
+
cacheWrite: 0,
|
|
6493
|
+
},
|
|
6494
|
+
contextWindow: 1000000,
|
|
6495
|
+
maxTokens: 128000,
|
|
6496
|
+
},
|
|
6515
6497
|
},
|
|
6516
6498
|
"opencode-go": {
|
|
6517
6499
|
"glm-5": {
|
|
@@ -6565,6 +6547,23 @@ export const MODELS = {
|
|
|
6565
6547
|
contextWindow: 204800,
|
|
6566
6548
|
maxTokens: 131072,
|
|
6567
6549
|
},
|
|
6550
|
+
"minimax-m2.7": {
|
|
6551
|
+
id: "minimax-m2.7",
|
|
6552
|
+
name: "MiniMax M2.7",
|
|
6553
|
+
api: "anthropic-messages",
|
|
6554
|
+
provider: "opencode-go",
|
|
6555
|
+
baseUrl: "https://opencode.ai/zen/go",
|
|
6556
|
+
reasoning: true,
|
|
6557
|
+
input: ["text"],
|
|
6558
|
+
cost: {
|
|
6559
|
+
input: 0.3,
|
|
6560
|
+
output: 1.2,
|
|
6561
|
+
cacheRead: 0.06,
|
|
6562
|
+
cacheWrite: 0,
|
|
6563
|
+
},
|
|
6564
|
+
contextWindow: 204800,
|
|
6565
|
+
maxTokens: 131072,
|
|
6566
|
+
},
|
|
6568
6567
|
},
|
|
6569
6568
|
"openrouter": {
|
|
6570
6569
|
"ai21/jamba-large-1.7": {
|
|
@@ -7077,6 +7076,23 @@ export const MODELS = {
|
|
|
7077
7076
|
contextWindow: 262144,
|
|
7078
7077
|
maxTokens: 32768,
|
|
7079
7078
|
},
|
|
7079
|
+
"bytedance-seed/seed-2.0-lite": {
|
|
7080
|
+
id: "bytedance-seed/seed-2.0-lite",
|
|
7081
|
+
name: "ByteDance Seed: Seed-2.0-Lite",
|
|
7082
|
+
api: "openai-completions",
|
|
7083
|
+
provider: "openrouter",
|
|
7084
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7085
|
+
reasoning: true,
|
|
7086
|
+
input: ["text", "image"],
|
|
7087
|
+
cost: {
|
|
7088
|
+
input: 0.25,
|
|
7089
|
+
output: 2,
|
|
7090
|
+
cacheRead: 0,
|
|
7091
|
+
cacheWrite: 0,
|
|
7092
|
+
},
|
|
7093
|
+
contextWindow: 262144,
|
|
7094
|
+
maxTokens: 131072,
|
|
7095
|
+
},
|
|
7080
7096
|
"bytedance-seed/seed-2.0-mini": {
|
|
7081
7097
|
id: "bytedance-seed/seed-2.0-mini",
|
|
7082
7098
|
name: "ByteDance Seed: Seed-2.0-Mini",
|
|
@@ -7156,11 +7172,11 @@ export const MODELS = {
|
|
|
7156
7172
|
cost: {
|
|
7157
7173
|
input: 0.19999999999999998,
|
|
7158
7174
|
output: 0.77,
|
|
7159
|
-
cacheRead: 0.
|
|
7175
|
+
cacheRead: 0.135,
|
|
7160
7176
|
cacheWrite: 0,
|
|
7161
7177
|
},
|
|
7162
7178
|
contextWindow: 163840,
|
|
7163
|
-
maxTokens:
|
|
7179
|
+
maxTokens: 4096,
|
|
7164
7180
|
},
|
|
7165
7181
|
"deepseek/deepseek-chat-v3.1": {
|
|
7166
7182
|
id: "deepseek/deepseek-chat-v3.1",
|
|
@@ -7230,23 +7246,6 @@ export const MODELS = {
|
|
|
7230
7246
|
contextWindow: 163840,
|
|
7231
7247
|
maxTokens: 4096,
|
|
7232
7248
|
},
|
|
7233
|
-
"deepseek/deepseek-v3.1-terminus:exacto": {
|
|
7234
|
-
id: "deepseek/deepseek-v3.1-terminus:exacto",
|
|
7235
|
-
name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
|
|
7236
|
-
api: "openai-completions",
|
|
7237
|
-
provider: "openrouter",
|
|
7238
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7239
|
-
reasoning: true,
|
|
7240
|
-
input: ["text"],
|
|
7241
|
-
cost: {
|
|
7242
|
-
input: 0.21,
|
|
7243
|
-
output: 0.7899999999999999,
|
|
7244
|
-
cacheRead: 0.16799999999999998,
|
|
7245
|
-
cacheWrite: 0,
|
|
7246
|
-
},
|
|
7247
|
-
contextWindow: 163840,
|
|
7248
|
-
maxTokens: 4096,
|
|
7249
|
-
},
|
|
7250
7249
|
"deepseek/deepseek-v3.2": {
|
|
7251
7250
|
id: "deepseek/deepseek-v3.2",
|
|
7252
7251
|
name: "DeepSeek: DeepSeek V3.2",
|
|
@@ -7256,13 +7255,13 @@ export const MODELS = {
|
|
|
7256
7255
|
reasoning: true,
|
|
7257
7256
|
input: ["text"],
|
|
7258
7257
|
cost: {
|
|
7259
|
-
input: 0.
|
|
7260
|
-
output: 0.
|
|
7261
|
-
cacheRead: 0,
|
|
7258
|
+
input: 0.26,
|
|
7259
|
+
output: 0.38,
|
|
7260
|
+
cacheRead: 0.13,
|
|
7262
7261
|
cacheWrite: 0,
|
|
7263
7262
|
},
|
|
7264
7263
|
contextWindow: 163840,
|
|
7265
|
-
maxTokens:
|
|
7264
|
+
maxTokens: 4096,
|
|
7266
7265
|
},
|
|
7267
7266
|
"deepseek/deepseek-v3.2-exp": {
|
|
7268
7267
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7519,40 +7518,6 @@ export const MODELS = {
|
|
|
7519
7518
|
contextWindow: 1048576,
|
|
7520
7519
|
maxTokens: 65536,
|
|
7521
7520
|
},
|
|
7522
|
-
"google/gemma-3-27b-it": {
|
|
7523
|
-
id: "google/gemma-3-27b-it",
|
|
7524
|
-
name: "Google: Gemma 3 27B",
|
|
7525
|
-
api: "openai-completions",
|
|
7526
|
-
provider: "openrouter",
|
|
7527
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7528
|
-
reasoning: false,
|
|
7529
|
-
input: ["text", "image"],
|
|
7530
|
-
cost: {
|
|
7531
|
-
input: 0.04,
|
|
7532
|
-
output: 0.15,
|
|
7533
|
-
cacheRead: 0.02,
|
|
7534
|
-
cacheWrite: 0,
|
|
7535
|
-
},
|
|
7536
|
-
contextWindow: 128000,
|
|
7537
|
-
maxTokens: 65536,
|
|
7538
|
-
},
|
|
7539
|
-
"google/gemma-3-27b-it:free": {
|
|
7540
|
-
id: "google/gemma-3-27b-it:free",
|
|
7541
|
-
name: "Google: Gemma 3 27B (free)",
|
|
7542
|
-
api: "openai-completions",
|
|
7543
|
-
provider: "openrouter",
|
|
7544
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7545
|
-
reasoning: false,
|
|
7546
|
-
input: ["text", "image"],
|
|
7547
|
-
cost: {
|
|
7548
|
-
input: 0,
|
|
7549
|
-
output: 0,
|
|
7550
|
-
cacheRead: 0,
|
|
7551
|
-
cacheWrite: 0,
|
|
7552
|
-
},
|
|
7553
|
-
contextWindow: 131072,
|
|
7554
|
-
maxTokens: 8192,
|
|
7555
|
-
},
|
|
7556
7521
|
"inception/mercury": {
|
|
7557
7522
|
id: "inception/mercury",
|
|
7558
7523
|
name: "Inception: Mercury",
|
|
@@ -7655,23 +7620,6 @@ export const MODELS = {
|
|
|
7655
7620
|
contextWindow: 8192,
|
|
7656
7621
|
maxTokens: 16384,
|
|
7657
7622
|
},
|
|
7658
|
-
"meta-llama/llama-3.1-405b-instruct": {
|
|
7659
|
-
id: "meta-llama/llama-3.1-405b-instruct",
|
|
7660
|
-
name: "Meta: Llama 3.1 405B Instruct",
|
|
7661
|
-
api: "openai-completions",
|
|
7662
|
-
provider: "openrouter",
|
|
7663
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7664
|
-
reasoning: false,
|
|
7665
|
-
input: ["text"],
|
|
7666
|
-
cost: {
|
|
7667
|
-
input: 4,
|
|
7668
|
-
output: 4,
|
|
7669
|
-
cacheRead: 0,
|
|
7670
|
-
cacheWrite: 0,
|
|
7671
|
-
},
|
|
7672
|
-
contextWindow: 131000,
|
|
7673
|
-
maxTokens: 4096,
|
|
7674
|
-
},
|
|
7675
7623
|
"meta-llama/llama-3.1-70b-instruct": {
|
|
7676
7624
|
id: "meta-llama/llama-3.1-70b-instruct",
|
|
7677
7625
|
name: "Meta: Llama 3.1 70B Instruct",
|
|
@@ -7737,8 +7685,8 @@ export const MODELS = {
|
|
|
7737
7685
|
cacheRead: 0,
|
|
7738
7686
|
cacheWrite: 0,
|
|
7739
7687
|
},
|
|
7740
|
-
contextWindow:
|
|
7741
|
-
maxTokens:
|
|
7688
|
+
contextWindow: 65536,
|
|
7689
|
+
maxTokens: 4096,
|
|
7742
7690
|
},
|
|
7743
7691
|
"meta-llama/llama-4-maverick": {
|
|
7744
7692
|
id: "meta-llama/llama-4-maverick",
|
|
@@ -7834,28 +7782,62 @@ export const MODELS = {
|
|
|
7834
7782
|
reasoning: true,
|
|
7835
7783
|
input: ["text"],
|
|
7836
7784
|
cost: {
|
|
7837
|
-
input: 0.
|
|
7838
|
-
output: 1.
|
|
7839
|
-
cacheRead: 0.
|
|
7785
|
+
input: 0.19999999999999998,
|
|
7786
|
+
output: 1.17,
|
|
7787
|
+
cacheRead: 0.09999999999999999,
|
|
7840
7788
|
cacheWrite: 0,
|
|
7841
7789
|
},
|
|
7842
7790
|
contextWindow: 196608,
|
|
7843
|
-
maxTokens:
|
|
7791
|
+
maxTokens: 65536,
|
|
7844
7792
|
},
|
|
7845
|
-
"
|
|
7846
|
-
id: "
|
|
7847
|
-
name: "
|
|
7793
|
+
"minimax/minimax-m2.5:free": {
|
|
7794
|
+
id: "minimax/minimax-m2.5:free",
|
|
7795
|
+
name: "MiniMax: MiniMax M2.5 (free)",
|
|
7848
7796
|
api: "openai-completions",
|
|
7849
7797
|
provider: "openrouter",
|
|
7850
7798
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7851
|
-
reasoning:
|
|
7799
|
+
reasoning: true,
|
|
7852
7800
|
input: ["text"],
|
|
7853
7801
|
cost: {
|
|
7854
|
-
input: 0
|
|
7855
|
-
output: 0
|
|
7802
|
+
input: 0,
|
|
7803
|
+
output: 0,
|
|
7856
7804
|
cacheRead: 0,
|
|
7857
7805
|
cacheWrite: 0,
|
|
7858
7806
|
},
|
|
7807
|
+
contextWindow: 196608,
|
|
7808
|
+
maxTokens: 196608,
|
|
7809
|
+
},
|
|
7810
|
+
"minimax/minimax-m2.7": {
|
|
7811
|
+
id: "minimax/minimax-m2.7",
|
|
7812
|
+
name: "MiniMax: MiniMax M2.7",
|
|
7813
|
+
api: "openai-completions",
|
|
7814
|
+
provider: "openrouter",
|
|
7815
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7816
|
+
reasoning: true,
|
|
7817
|
+
input: ["text"],
|
|
7818
|
+
cost: {
|
|
7819
|
+
input: 0.3,
|
|
7820
|
+
output: 1.2,
|
|
7821
|
+
cacheRead: 0.06,
|
|
7822
|
+
cacheWrite: 0,
|
|
7823
|
+
},
|
|
7824
|
+
contextWindow: 204800,
|
|
7825
|
+
maxTokens: 131072,
|
|
7826
|
+
},
|
|
7827
|
+
"mistralai/codestral-2508": {
|
|
7828
|
+
id: "mistralai/codestral-2508",
|
|
7829
|
+
name: "Mistral: Codestral 2508",
|
|
7830
|
+
api: "openai-completions",
|
|
7831
|
+
provider: "openrouter",
|
|
7832
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7833
|
+
reasoning: false,
|
|
7834
|
+
input: ["text"],
|
|
7835
|
+
cost: {
|
|
7836
|
+
input: 0.3,
|
|
7837
|
+
output: 0.8999999999999999,
|
|
7838
|
+
cacheRead: 0.03,
|
|
7839
|
+
cacheWrite: 0,
|
|
7840
|
+
},
|
|
7859
7841
|
contextWindow: 256000,
|
|
7860
7842
|
maxTokens: 4096,
|
|
7861
7843
|
},
|
|
@@ -7870,7 +7852,7 @@ export const MODELS = {
|
|
|
7870
7852
|
cost: {
|
|
7871
7853
|
input: 0.39999999999999997,
|
|
7872
7854
|
output: 2,
|
|
7873
|
-
cacheRead: 0,
|
|
7855
|
+
cacheRead: 0.04,
|
|
7874
7856
|
cacheWrite: 0,
|
|
7875
7857
|
},
|
|
7876
7858
|
contextWindow: 262144,
|
|
@@ -7887,7 +7869,7 @@ export const MODELS = {
|
|
|
7887
7869
|
cost: {
|
|
7888
7870
|
input: 0.39999999999999997,
|
|
7889
7871
|
output: 2,
|
|
7890
|
-
cacheRead: 0,
|
|
7872
|
+
cacheRead: 0.04,
|
|
7891
7873
|
cacheWrite: 0,
|
|
7892
7874
|
},
|
|
7893
7875
|
contextWindow: 131072,
|
|
@@ -7904,7 +7886,7 @@ export const MODELS = {
|
|
|
7904
7886
|
cost: {
|
|
7905
7887
|
input: 0.09999999999999999,
|
|
7906
7888
|
output: 0.3,
|
|
7907
|
-
cacheRead: 0,
|
|
7889
|
+
cacheRead: 0.01,
|
|
7908
7890
|
cacheWrite: 0,
|
|
7909
7891
|
},
|
|
7910
7892
|
contextWindow: 131072,
|
|
@@ -7921,7 +7903,7 @@ export const MODELS = {
|
|
|
7921
7903
|
cost: {
|
|
7922
7904
|
input: 0.19999999999999998,
|
|
7923
7905
|
output: 0.19999999999999998,
|
|
7924
|
-
cacheRead: 0,
|
|
7906
|
+
cacheRead: 0.02,
|
|
7925
7907
|
cacheWrite: 0,
|
|
7926
7908
|
},
|
|
7927
7909
|
contextWindow: 262144,
|
|
@@ -7938,7 +7920,7 @@ export const MODELS = {
|
|
|
7938
7920
|
cost: {
|
|
7939
7921
|
input: 0.09999999999999999,
|
|
7940
7922
|
output: 0.09999999999999999,
|
|
7941
|
-
cacheRead: 0,
|
|
7923
|
+
cacheRead: 0.01,
|
|
7942
7924
|
cacheWrite: 0,
|
|
7943
7925
|
},
|
|
7944
7926
|
contextWindow: 131072,
|
|
@@ -7955,7 +7937,7 @@ export const MODELS = {
|
|
|
7955
7937
|
cost: {
|
|
7956
7938
|
input: 0.15,
|
|
7957
7939
|
output: 0.15,
|
|
7958
|
-
cacheRead: 0,
|
|
7940
|
+
cacheRead: 0.015,
|
|
7959
7941
|
cacheWrite: 0,
|
|
7960
7942
|
},
|
|
7961
7943
|
contextWindow: 262144,
|
|
@@ -7972,7 +7954,7 @@ export const MODELS = {
|
|
|
7972
7954
|
cost: {
|
|
7973
7955
|
input: 2,
|
|
7974
7956
|
output: 6,
|
|
7975
|
-
cacheRead: 0,
|
|
7957
|
+
cacheRead: 0.19999999999999998,
|
|
7976
7958
|
cacheWrite: 0,
|
|
7977
7959
|
},
|
|
7978
7960
|
contextWindow: 128000,
|
|
@@ -7989,7 +7971,7 @@ export const MODELS = {
|
|
|
7989
7971
|
cost: {
|
|
7990
7972
|
input: 2,
|
|
7991
7973
|
output: 6,
|
|
7992
|
-
cacheRead: 0,
|
|
7974
|
+
cacheRead: 0.19999999999999998,
|
|
7993
7975
|
cacheWrite: 0,
|
|
7994
7976
|
},
|
|
7995
7977
|
contextWindow: 131072,
|
|
@@ -8006,7 +7988,7 @@ export const MODELS = {
|
|
|
8006
7988
|
cost: {
|
|
8007
7989
|
input: 2,
|
|
8008
7990
|
output: 6,
|
|
8009
|
-
cacheRead: 0,
|
|
7991
|
+
cacheRead: 0.19999999999999998,
|
|
8010
7992
|
cacheWrite: 0,
|
|
8011
7993
|
},
|
|
8012
7994
|
contextWindow: 131072,
|
|
@@ -8023,7 +8005,7 @@ export const MODELS = {
|
|
|
8023
8005
|
cost: {
|
|
8024
8006
|
input: 0.5,
|
|
8025
8007
|
output: 1.5,
|
|
8026
|
-
cacheRead: 0,
|
|
8008
|
+
cacheRead: 0.049999999999999996,
|
|
8027
8009
|
cacheWrite: 0,
|
|
8028
8010
|
},
|
|
8029
8011
|
contextWindow: 262144,
|
|
@@ -8040,7 +8022,7 @@ export const MODELS = {
|
|
|
8040
8022
|
cost: {
|
|
8041
8023
|
input: 0.39999999999999997,
|
|
8042
8024
|
output: 2,
|
|
8043
|
-
cacheRead: 0,
|
|
8025
|
+
cacheRead: 0.04,
|
|
8044
8026
|
cacheWrite: 0,
|
|
8045
8027
|
},
|
|
8046
8028
|
contextWindow: 131072,
|
|
@@ -8057,7 +8039,7 @@ export const MODELS = {
|
|
|
8057
8039
|
cost: {
|
|
8058
8040
|
input: 0.39999999999999997,
|
|
8059
8041
|
output: 2,
|
|
8060
|
-
cacheRead: 0,
|
|
8042
|
+
cacheRead: 0.04,
|
|
8061
8043
|
cacheWrite: 0,
|
|
8062
8044
|
},
|
|
8063
8045
|
contextWindow: 131072,
|
|
@@ -8091,7 +8073,7 @@ export const MODELS = {
|
|
|
8091
8073
|
cost: {
|
|
8092
8074
|
input: 0.19999999999999998,
|
|
8093
8075
|
output: 0.6,
|
|
8094
|
-
cacheRead: 0,
|
|
8076
|
+
cacheRead: 0.02,
|
|
8095
8077
|
cacheWrite: 0,
|
|
8096
8078
|
},
|
|
8097
8079
|
contextWindow: 32768,
|
|
@@ -8114,6 +8096,23 @@ export const MODELS = {
|
|
|
8114
8096
|
contextWindow: 32768,
|
|
8115
8097
|
maxTokens: 16384,
|
|
8116
8098
|
},
|
|
8099
|
+
"mistralai/mistral-small-2603": {
|
|
8100
|
+
id: "mistralai/mistral-small-2603",
|
|
8101
|
+
name: "Mistral: Mistral Small 4",
|
|
8102
|
+
api: "openai-completions",
|
|
8103
|
+
provider: "openrouter",
|
|
8104
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8105
|
+
reasoning: true,
|
|
8106
|
+
input: ["text", "image"],
|
|
8107
|
+
cost: {
|
|
8108
|
+
input: 0.15,
|
|
8109
|
+
output: 0.6,
|
|
8110
|
+
cacheRead: 0.015,
|
|
8111
|
+
cacheWrite: 0,
|
|
8112
|
+
},
|
|
8113
|
+
contextWindow: 262144,
|
|
8114
|
+
maxTokens: 4096,
|
|
8115
|
+
},
|
|
8117
8116
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8118
8117
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
8119
8118
|
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
@@ -8140,13 +8139,13 @@ export const MODELS = {
|
|
|
8140
8139
|
reasoning: false,
|
|
8141
8140
|
input: ["text", "image"],
|
|
8142
8141
|
cost: {
|
|
8143
|
-
input: 0.
|
|
8144
|
-
output: 0.
|
|
8145
|
-
cacheRead: 0
|
|
8142
|
+
input: 0.075,
|
|
8143
|
+
output: 0.19999999999999998,
|
|
8144
|
+
cacheRead: 0,
|
|
8146
8145
|
cacheWrite: 0,
|
|
8147
8146
|
},
|
|
8148
|
-
contextWindow:
|
|
8149
|
-
maxTokens:
|
|
8147
|
+
contextWindow: 128000,
|
|
8148
|
+
maxTokens: 4096,
|
|
8150
8149
|
},
|
|
8151
8150
|
"mistralai/mistral-small-creative": {
|
|
8152
8151
|
id: "mistralai/mistral-small-creative",
|
|
@@ -8159,7 +8158,7 @@ export const MODELS = {
|
|
|
8159
8158
|
cost: {
|
|
8160
8159
|
input: 0.09999999999999999,
|
|
8161
8160
|
output: 0.3,
|
|
8162
|
-
cacheRead: 0,
|
|
8161
|
+
cacheRead: 0.01,
|
|
8163
8162
|
cacheWrite: 0,
|
|
8164
8163
|
},
|
|
8165
8164
|
contextWindow: 32768,
|
|
@@ -8176,7 +8175,7 @@ export const MODELS = {
|
|
|
8176
8175
|
cost: {
|
|
8177
8176
|
input: 2,
|
|
8178
8177
|
output: 6,
|
|
8179
|
-
cacheRead: 0,
|
|
8178
|
+
cacheRead: 0.19999999999999998,
|
|
8180
8179
|
cacheWrite: 0,
|
|
8181
8180
|
},
|
|
8182
8181
|
contextWindow: 65536,
|
|
@@ -8210,7 +8209,7 @@ export const MODELS = {
|
|
|
8210
8209
|
cost: {
|
|
8211
8210
|
input: 2,
|
|
8212
8211
|
output: 6,
|
|
8213
|
-
cacheRead: 0,
|
|
8212
|
+
cacheRead: 0.19999999999999998,
|
|
8214
8213
|
cacheWrite: 0,
|
|
8215
8214
|
},
|
|
8216
8215
|
contextWindow: 131072,
|
|
@@ -8227,7 +8226,7 @@ export const MODELS = {
|
|
|
8227
8226
|
cost: {
|
|
8228
8227
|
input: 0.09999999999999999,
|
|
8229
8228
|
output: 0.3,
|
|
8230
|
-
cacheRead: 0,
|
|
8229
|
+
cacheRead: 0.01,
|
|
8231
8230
|
cacheWrite: 0,
|
|
8232
8231
|
},
|
|
8233
8232
|
contextWindow: 32000,
|
|
@@ -8267,23 +8266,6 @@ export const MODELS = {
|
|
|
8267
8266
|
contextWindow: 131072,
|
|
8268
8267
|
maxTokens: 4096,
|
|
8269
8268
|
},
|
|
8270
|
-
"moonshotai/kimi-k2-0905:exacto": {
|
|
8271
|
-
id: "moonshotai/kimi-k2-0905:exacto",
|
|
8272
|
-
name: "MoonshotAI: Kimi K2 0905 (exacto)",
|
|
8273
|
-
api: "openai-completions",
|
|
8274
|
-
provider: "openrouter",
|
|
8275
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8276
|
-
reasoning: false,
|
|
8277
|
-
input: ["text"],
|
|
8278
|
-
cost: {
|
|
8279
|
-
input: 0.6,
|
|
8280
|
-
output: 2.5,
|
|
8281
|
-
cacheRead: 0,
|
|
8282
|
-
cacheWrite: 0,
|
|
8283
|
-
},
|
|
8284
|
-
contextWindow: 262144,
|
|
8285
|
-
maxTokens: 4096,
|
|
8286
|
-
},
|
|
8287
8269
|
"moonshotai/kimi-k2-thinking": {
|
|
8288
8270
|
id: "moonshotai/kimi-k2-thinking",
|
|
8289
8271
|
name: "MoonshotAI: Kimi K2 Thinking",
|
|
@@ -8403,6 +8385,40 @@ export const MODELS = {
|
|
|
8403
8385
|
contextWindow: 256000,
|
|
8404
8386
|
maxTokens: 4096,
|
|
8405
8387
|
},
|
|
8388
|
+
"nvidia/nemotron-3-super-120b-a12b": {
|
|
8389
|
+
id: "nvidia/nemotron-3-super-120b-a12b",
|
|
8390
|
+
name: "NVIDIA: Nemotron 3 Super",
|
|
8391
|
+
api: "openai-completions",
|
|
8392
|
+
provider: "openrouter",
|
|
8393
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8394
|
+
reasoning: true,
|
|
8395
|
+
input: ["text"],
|
|
8396
|
+
cost: {
|
|
8397
|
+
input: 0.09999999999999999,
|
|
8398
|
+
output: 0.5,
|
|
8399
|
+
cacheRead: 0.04,
|
|
8400
|
+
cacheWrite: 0,
|
|
8401
|
+
},
|
|
8402
|
+
contextWindow: 262144,
|
|
8403
|
+
maxTokens: 4096,
|
|
8404
|
+
},
|
|
8405
|
+
"nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8406
|
+
id: "nvidia/nemotron-3-super-120b-a12b:free",
|
|
8407
|
+
name: "NVIDIA: Nemotron 3 Super (free)",
|
|
8408
|
+
api: "openai-completions",
|
|
8409
|
+
provider: "openrouter",
|
|
8410
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8411
|
+
reasoning: true,
|
|
8412
|
+
input: ["text"],
|
|
8413
|
+
cost: {
|
|
8414
|
+
input: 0,
|
|
8415
|
+
output: 0,
|
|
8416
|
+
cacheRead: 0,
|
|
8417
|
+
cacheWrite: 0,
|
|
8418
|
+
},
|
|
8419
|
+
contextWindow: 262144,
|
|
8420
|
+
maxTokens: 262144,
|
|
8421
|
+
},
|
|
8406
8422
|
"nvidia/nemotron-nano-12b-v2-vl:free": {
|
|
8407
8423
|
id: "nvidia/nemotron-nano-12b-v2-vl:free",
|
|
8408
8424
|
name: "NVIDIA: Nemotron Nano 12B 2 VL (free)",
|
|
@@ -9100,6 +9116,40 @@ export const MODELS = {
|
|
|
9100
9116
|
contextWindow: 1050000,
|
|
9101
9117
|
maxTokens: 128000,
|
|
9102
9118
|
},
|
|
9119
|
+
"openai/gpt-5.4-mini": {
|
|
9120
|
+
id: "openai/gpt-5.4-mini",
|
|
9121
|
+
name: "OpenAI: GPT-5.4 Mini",
|
|
9122
|
+
api: "openai-completions",
|
|
9123
|
+
provider: "openrouter",
|
|
9124
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9125
|
+
reasoning: true,
|
|
9126
|
+
input: ["text", "image"],
|
|
9127
|
+
cost: {
|
|
9128
|
+
input: 0.75,
|
|
9129
|
+
output: 4.5,
|
|
9130
|
+
cacheRead: 0.075,
|
|
9131
|
+
cacheWrite: 0,
|
|
9132
|
+
},
|
|
9133
|
+
contextWindow: 400000,
|
|
9134
|
+
maxTokens: 128000,
|
|
9135
|
+
},
|
|
9136
|
+
"openai/gpt-5.4-nano": {
|
|
9137
|
+
id: "openai/gpt-5.4-nano",
|
|
9138
|
+
name: "OpenAI: GPT-5.4 Nano",
|
|
9139
|
+
api: "openai-completions",
|
|
9140
|
+
provider: "openrouter",
|
|
9141
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9142
|
+
reasoning: true,
|
|
9143
|
+
input: ["text", "image"],
|
|
9144
|
+
cost: {
|
|
9145
|
+
input: 0.19999999999999998,
|
|
9146
|
+
output: 1.25,
|
|
9147
|
+
cacheRead: 0.02,
|
|
9148
|
+
cacheWrite: 0,
|
|
9149
|
+
},
|
|
9150
|
+
contextWindow: 400000,
|
|
9151
|
+
maxTokens: 128000,
|
|
9152
|
+
},
|
|
9103
9153
|
"openai/gpt-5.4-pro": {
|
|
9104
9154
|
id: "openai/gpt-5.4-pro",
|
|
9105
9155
|
name: "OpenAI: GPT-5.4 Pro",
|
|
@@ -9134,23 +9184,6 @@ export const MODELS = {
|
|
|
9134
9184
|
contextWindow: 131072,
|
|
9135
9185
|
maxTokens: 4096,
|
|
9136
9186
|
},
|
|
9137
|
-
"openai/gpt-oss-120b:exacto": {
|
|
9138
|
-
id: "openai/gpt-oss-120b:exacto",
|
|
9139
|
-
name: "OpenAI: gpt-oss-120b (exacto)",
|
|
9140
|
-
api: "openai-completions",
|
|
9141
|
-
provider: "openrouter",
|
|
9142
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9143
|
-
reasoning: true,
|
|
9144
|
-
input: ["text"],
|
|
9145
|
-
cost: {
|
|
9146
|
-
input: 0.039,
|
|
9147
|
-
output: 0.19,
|
|
9148
|
-
cacheRead: 0,
|
|
9149
|
-
cacheWrite: 0,
|
|
9150
|
-
},
|
|
9151
|
-
contextWindow: 131072,
|
|
9152
|
-
maxTokens: 4096,
|
|
9153
|
-
},
|
|
9154
9187
|
"openai/gpt-oss-120b:free": {
|
|
9155
9188
|
id: "openai/gpt-oss-120b:free",
|
|
9156
9189
|
name: "OpenAI: gpt-oss-120b (free)",
|
|
@@ -9178,12 +9211,12 @@ export const MODELS = {
|
|
|
9178
9211
|
input: ["text"],
|
|
9179
9212
|
cost: {
|
|
9180
9213
|
input: 0.03,
|
|
9181
|
-
output: 0.
|
|
9182
|
-
cacheRead: 0,
|
|
9214
|
+
output: 0.11,
|
|
9215
|
+
cacheRead: 0.015,
|
|
9183
9216
|
cacheWrite: 0,
|
|
9184
9217
|
},
|
|
9185
9218
|
contextWindow: 131072,
|
|
9186
|
-
maxTokens:
|
|
9219
|
+
maxTokens: 131072,
|
|
9187
9220
|
},
|
|
9188
9221
|
"openai/gpt-oss-20b:free": {
|
|
9189
9222
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -9225,7 +9258,7 @@ export const MODELS = {
|
|
|
9225
9258
|
api: "openai-completions",
|
|
9226
9259
|
provider: "openrouter",
|
|
9227
9260
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9228
|
-
reasoning:
|
|
9261
|
+
reasoning: true,
|
|
9229
9262
|
input: ["text", "image"],
|
|
9230
9263
|
cost: {
|
|
9231
9264
|
input: 15,
|
|
@@ -9276,7 +9309,7 @@ export const MODELS = {
|
|
|
9276
9309
|
api: "openai-completions",
|
|
9277
9310
|
provider: "openrouter",
|
|
9278
9311
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9279
|
-
reasoning:
|
|
9312
|
+
reasoning: true,
|
|
9280
9313
|
input: ["text"],
|
|
9281
9314
|
cost: {
|
|
9282
9315
|
input: 1.1,
|
|
@@ -9293,7 +9326,7 @@ export const MODELS = {
|
|
|
9293
9326
|
api: "openai-completions",
|
|
9294
9327
|
provider: "openrouter",
|
|
9295
9328
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9296
|
-
reasoning:
|
|
9329
|
+
reasoning: true,
|
|
9297
9330
|
input: ["text"],
|
|
9298
9331
|
cost: {
|
|
9299
9332
|
input: 1.1,
|
|
@@ -9483,9 +9516,9 @@ export const MODELS = {
|
|
|
9483
9516
|
reasoning: false,
|
|
9484
9517
|
input: ["text"],
|
|
9485
9518
|
cost: {
|
|
9486
|
-
input: 0.
|
|
9487
|
-
output:
|
|
9488
|
-
cacheRead: 0.
|
|
9519
|
+
input: 0.26,
|
|
9520
|
+
output: 0.78,
|
|
9521
|
+
cacheRead: 0.052000000000000005,
|
|
9489
9522
|
cacheWrite: 0,
|
|
9490
9523
|
},
|
|
9491
9524
|
contextWindow: 1000000,
|
|
@@ -9551,8 +9584,8 @@ export const MODELS = {
|
|
|
9551
9584
|
reasoning: false,
|
|
9552
9585
|
input: ["text", "image"],
|
|
9553
9586
|
cost: {
|
|
9554
|
-
input: 0.
|
|
9555
|
-
output:
|
|
9587
|
+
input: 0.52,
|
|
9588
|
+
output: 2.08,
|
|
9556
9589
|
cacheRead: 0,
|
|
9557
9590
|
cacheWrite: 0,
|
|
9558
9591
|
},
|
|
@@ -9619,13 +9652,13 @@ export const MODELS = {
|
|
|
9619
9652
|
reasoning: true,
|
|
9620
9653
|
input: ["text"],
|
|
9621
9654
|
cost: {
|
|
9622
|
-
input: 0.
|
|
9623
|
-
output:
|
|
9624
|
-
cacheRead: 0
|
|
9655
|
+
input: 0.14950000000000002,
|
|
9656
|
+
output: 1.495,
|
|
9657
|
+
cacheRead: 0,
|
|
9625
9658
|
cacheWrite: 0,
|
|
9626
9659
|
},
|
|
9627
|
-
contextWindow:
|
|
9628
|
-
maxTokens:
|
|
9660
|
+
contextWindow: 131072,
|
|
9661
|
+
maxTokens: 4096,
|
|
9629
9662
|
},
|
|
9630
9663
|
"qwen/qwen3-30b-a3b": {
|
|
9631
9664
|
id: "qwen/qwen3-30b-a3b",
|
|
@@ -9670,13 +9703,13 @@ export const MODELS = {
|
|
|
9670
9703
|
reasoning: true,
|
|
9671
9704
|
input: ["text"],
|
|
9672
9705
|
cost: {
|
|
9673
|
-
input: 0.
|
|
9674
|
-
output: 0.
|
|
9675
|
-
cacheRead: 0,
|
|
9706
|
+
input: 0.08,
|
|
9707
|
+
output: 0.39999999999999997,
|
|
9708
|
+
cacheRead: 0.08,
|
|
9676
9709
|
cacheWrite: 0,
|
|
9677
9710
|
},
|
|
9678
|
-
contextWindow:
|
|
9679
|
-
maxTokens:
|
|
9711
|
+
contextWindow: 131072,
|
|
9712
|
+
maxTokens: 131072,
|
|
9680
9713
|
},
|
|
9681
9714
|
"qwen/qwen3-32b": {
|
|
9682
9715
|
id: "qwen/qwen3-32b",
|
|
@@ -9814,23 +9847,6 @@ export const MODELS = {
|
|
|
9814
9847
|
contextWindow: 1000000,
|
|
9815
9848
|
maxTokens: 65536,
|
|
9816
9849
|
},
|
|
9817
|
-
"qwen/qwen3-coder:exacto": {
|
|
9818
|
-
id: "qwen/qwen3-coder:exacto",
|
|
9819
|
-
name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
|
|
9820
|
-
api: "openai-completions",
|
|
9821
|
-
provider: "openrouter",
|
|
9822
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9823
|
-
reasoning: false,
|
|
9824
|
-
input: ["text"],
|
|
9825
|
-
cost: {
|
|
9826
|
-
input: 0.22,
|
|
9827
|
-
output: 1.7999999999999998,
|
|
9828
|
-
cacheRead: 0.022,
|
|
9829
|
-
cacheWrite: 0,
|
|
9830
|
-
},
|
|
9831
|
-
contextWindow: 262144,
|
|
9832
|
-
maxTokens: 65536,
|
|
9833
|
-
},
|
|
9834
9850
|
"qwen/qwen3-coder:free": {
|
|
9835
9851
|
id: "qwen/qwen3-coder:free",
|
|
9836
9852
|
name: "Qwen: Qwen3 Coder 480B A35B (free)",
|
|
@@ -9857,9 +9873,9 @@ export const MODELS = {
|
|
|
9857
9873
|
reasoning: false,
|
|
9858
9874
|
input: ["text"],
|
|
9859
9875
|
cost: {
|
|
9860
|
-
input:
|
|
9861
|
-
output:
|
|
9862
|
-
cacheRead: 0.
|
|
9876
|
+
input: 0.78,
|
|
9877
|
+
output: 3.9,
|
|
9878
|
+
cacheRead: 0.156,
|
|
9863
9879
|
cacheWrite: 0,
|
|
9864
9880
|
},
|
|
9865
9881
|
contextWindow: 262144,
|
|
@@ -9925,13 +9941,13 @@ export const MODELS = {
|
|
|
9925
9941
|
reasoning: true,
|
|
9926
9942
|
input: ["text"],
|
|
9927
9943
|
cost: {
|
|
9928
|
-
input: 0.
|
|
9929
|
-
output:
|
|
9944
|
+
input: 0.0975,
|
|
9945
|
+
output: 0.78,
|
|
9930
9946
|
cacheRead: 0,
|
|
9931
9947
|
cacheWrite: 0,
|
|
9932
9948
|
},
|
|
9933
|
-
contextWindow:
|
|
9934
|
-
maxTokens:
|
|
9949
|
+
contextWindow: 131072,
|
|
9950
|
+
maxTokens: 32768,
|
|
9935
9951
|
},
|
|
9936
9952
|
"qwen/qwen3-vl-235b-a22b-instruct": {
|
|
9937
9953
|
id: "qwen/qwen3-vl-235b-a22b-instruct",
|
|
@@ -9959,8 +9975,8 @@ export const MODELS = {
|
|
|
9959
9975
|
reasoning: true,
|
|
9960
9976
|
input: ["text", "image"],
|
|
9961
9977
|
cost: {
|
|
9962
|
-
input: 0,
|
|
9963
|
-
output:
|
|
9978
|
+
input: 0.26,
|
|
9979
|
+
output: 2.6,
|
|
9964
9980
|
cacheRead: 0,
|
|
9965
9981
|
cacheWrite: 0,
|
|
9966
9982
|
},
|
|
@@ -9993,8 +10009,8 @@ export const MODELS = {
|
|
|
9993
10009
|
reasoning: true,
|
|
9994
10010
|
input: ["text", "image"],
|
|
9995
10011
|
cost: {
|
|
9996
|
-
input: 0,
|
|
9997
|
-
output:
|
|
10012
|
+
input: 0.13,
|
|
10013
|
+
output: 1.56,
|
|
9998
10014
|
cacheRead: 0,
|
|
9999
10015
|
cacheWrite: 0,
|
|
10000
10016
|
},
|
|
@@ -10120,6 +10136,23 @@ export const MODELS = {
|
|
|
10120
10136
|
contextWindow: 262144,
|
|
10121
10137
|
maxTokens: 65536,
|
|
10122
10138
|
},
|
|
10139
|
+
"qwen/qwen3.5-9b": {
|
|
10140
|
+
id: "qwen/qwen3.5-9b",
|
|
10141
|
+
name: "Qwen: Qwen3.5-9B",
|
|
10142
|
+
api: "openai-completions",
|
|
10143
|
+
provider: "openrouter",
|
|
10144
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10145
|
+
reasoning: true,
|
|
10146
|
+
input: ["text", "image"],
|
|
10147
|
+
cost: {
|
|
10148
|
+
input: 0.049999999999999996,
|
|
10149
|
+
output: 0.15,
|
|
10150
|
+
cacheRead: 0,
|
|
10151
|
+
cacheWrite: 0,
|
|
10152
|
+
},
|
|
10153
|
+
contextWindow: 256000,
|
|
10154
|
+
maxTokens: 4096,
|
|
10155
|
+
},
|
|
10123
10156
|
"qwen/qwen3.5-flash-02-23": {
|
|
10124
10157
|
id: "qwen/qwen3.5-flash-02-23",
|
|
10125
10158
|
name: "Qwen: Qwen3.5-Flash",
|
|
@@ -10129,8 +10162,8 @@ export const MODELS = {
|
|
|
10129
10162
|
reasoning: true,
|
|
10130
10163
|
input: ["text", "image"],
|
|
10131
10164
|
cost: {
|
|
10132
|
-
input: 0.
|
|
10133
|
-
output: 0.
|
|
10165
|
+
input: 0.065,
|
|
10166
|
+
output: 0.26,
|
|
10134
10167
|
cacheRead: 0,
|
|
10135
10168
|
cacheWrite: 0,
|
|
10136
10169
|
},
|
|
@@ -10164,12 +10197,12 @@ export const MODELS = {
|
|
|
10164
10197
|
input: ["text"],
|
|
10165
10198
|
cost: {
|
|
10166
10199
|
input: 0.15,
|
|
10167
|
-
output: 0.
|
|
10200
|
+
output: 0.58,
|
|
10168
10201
|
cacheRead: 0,
|
|
10169
10202
|
cacheWrite: 0,
|
|
10170
10203
|
},
|
|
10171
|
-
contextWindow:
|
|
10172
|
-
maxTokens:
|
|
10204
|
+
contextWindow: 131072,
|
|
10205
|
+
maxTokens: 131072,
|
|
10173
10206
|
},
|
|
10174
10207
|
"relace/relace-search": {
|
|
10175
10208
|
id: "relace/relace-search",
|
|
@@ -10214,13 +10247,13 @@ export const MODELS = {
|
|
|
10214
10247
|
reasoning: false,
|
|
10215
10248
|
input: ["text"],
|
|
10216
10249
|
cost: {
|
|
10217
|
-
input: 0.
|
|
10218
|
-
output: 0.
|
|
10250
|
+
input: 0.85,
|
|
10251
|
+
output: 0.85,
|
|
10219
10252
|
cacheRead: 0,
|
|
10220
10253
|
cacheWrite: 0,
|
|
10221
10254
|
},
|
|
10222
|
-
contextWindow:
|
|
10223
|
-
maxTokens:
|
|
10255
|
+
contextWindow: 131072,
|
|
10256
|
+
maxTokens: 16384,
|
|
10224
10257
|
},
|
|
10225
10258
|
"stepfun/step-3.5-flash": {
|
|
10226
10259
|
id: "stepfun/step-3.5-flash",
|
|
@@ -10299,9 +10332,9 @@ export const MODELS = {
|
|
|
10299
10332
|
reasoning: true,
|
|
10300
10333
|
input: ["text"],
|
|
10301
10334
|
cost: {
|
|
10302
|
-
input: 0.
|
|
10303
|
-
output:
|
|
10304
|
-
cacheRead: 0.
|
|
10335
|
+
input: 0.3,
|
|
10336
|
+
output: 1.1,
|
|
10337
|
+
cacheRead: 0.15,
|
|
10305
10338
|
cacheWrite: 0,
|
|
10306
10339
|
},
|
|
10307
10340
|
contextWindow: 163840,
|
|
@@ -10443,6 +10476,23 @@ export const MODELS = {
|
|
|
10443
10476
|
contextWindow: 2000000,
|
|
10444
10477
|
maxTokens: 30000,
|
|
10445
10478
|
},
|
|
10479
|
+
"x-ai/grok-4.20-beta": {
|
|
10480
|
+
id: "x-ai/grok-4.20-beta",
|
|
10481
|
+
name: "xAI: Grok 4.20 Beta",
|
|
10482
|
+
api: "openai-completions",
|
|
10483
|
+
provider: "openrouter",
|
|
10484
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10485
|
+
reasoning: true,
|
|
10486
|
+
input: ["text", "image"],
|
|
10487
|
+
cost: {
|
|
10488
|
+
input: 2,
|
|
10489
|
+
output: 6,
|
|
10490
|
+
cacheRead: 0.19999999999999998,
|
|
10491
|
+
cacheWrite: 0,
|
|
10492
|
+
},
|
|
10493
|
+
contextWindow: 2000000,
|
|
10494
|
+
maxTokens: 4096,
|
|
10495
|
+
},
|
|
10446
10496
|
"x-ai/grok-code-fast-1": {
|
|
10447
10497
|
id: "x-ai/grok-code-fast-1",
|
|
10448
10498
|
name: "xAI: Grok Code Fast 1",
|
|
@@ -10477,6 +10527,40 @@ export const MODELS = {
|
|
|
10477
10527
|
contextWindow: 262144,
|
|
10478
10528
|
maxTokens: 65536,
|
|
10479
10529
|
},
|
|
10530
|
+
"xiaomi/mimo-v2-omni": {
|
|
10531
|
+
id: "xiaomi/mimo-v2-omni",
|
|
10532
|
+
name: "Xiaomi: MiMo-V2-Omni",
|
|
10533
|
+
api: "openai-completions",
|
|
10534
|
+
provider: "openrouter",
|
|
10535
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10536
|
+
reasoning: true,
|
|
10537
|
+
input: ["text", "image"],
|
|
10538
|
+
cost: {
|
|
10539
|
+
input: 0.39999999999999997,
|
|
10540
|
+
output: 2,
|
|
10541
|
+
cacheRead: 0.08,
|
|
10542
|
+
cacheWrite: 0,
|
|
10543
|
+
},
|
|
10544
|
+
contextWindow: 262144,
|
|
10545
|
+
maxTokens: 65536,
|
|
10546
|
+
},
|
|
10547
|
+
"xiaomi/mimo-v2-pro": {
|
|
10548
|
+
id: "xiaomi/mimo-v2-pro",
|
|
10549
|
+
name: "Xiaomi: MiMo-V2-Pro",
|
|
10550
|
+
api: "openai-completions",
|
|
10551
|
+
provider: "openrouter",
|
|
10552
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10553
|
+
reasoning: true,
|
|
10554
|
+
input: ["text"],
|
|
10555
|
+
cost: {
|
|
10556
|
+
input: 1,
|
|
10557
|
+
output: 3,
|
|
10558
|
+
cacheRead: 0.19999999999999998,
|
|
10559
|
+
cacheWrite: 0,
|
|
10560
|
+
},
|
|
10561
|
+
contextWindow: 1048576,
|
|
10562
|
+
maxTokens: 131072,
|
|
10563
|
+
},
|
|
10480
10564
|
"z-ai/glm-4-32b": {
|
|
10481
10565
|
id: "z-ai/glm-4-32b",
|
|
10482
10566
|
name: "Z.ai: GLM 4 32B ",
|
|
@@ -10579,23 +10663,6 @@ export const MODELS = {
|
|
|
10579
10663
|
contextWindow: 204800,
|
|
10580
10664
|
maxTokens: 204800,
|
|
10581
10665
|
},
|
|
10582
|
-
"z-ai/glm-4.6:exacto": {
|
|
10583
|
-
id: "z-ai/glm-4.6:exacto",
|
|
10584
|
-
name: "Z.ai: GLM 4.6 (exacto)",
|
|
10585
|
-
api: "openai-completions",
|
|
10586
|
-
provider: "openrouter",
|
|
10587
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
10588
|
-
reasoning: true,
|
|
10589
|
-
input: ["text"],
|
|
10590
|
-
cost: {
|
|
10591
|
-
input: 0.44,
|
|
10592
|
-
output: 1.76,
|
|
10593
|
-
cacheRead: 0.11,
|
|
10594
|
-
cacheWrite: 0,
|
|
10595
|
-
},
|
|
10596
|
-
contextWindow: 204800,
|
|
10597
|
-
maxTokens: 131072,
|
|
10598
|
-
},
|
|
10599
10666
|
"z-ai/glm-4.6v": {
|
|
10600
10667
|
id: "z-ai/glm-4.6v",
|
|
10601
10668
|
name: "Z.ai: GLM 4.6V",
|
|
@@ -10622,13 +10689,13 @@ export const MODELS = {
|
|
|
10622
10689
|
reasoning: true,
|
|
10623
10690
|
input: ["text"],
|
|
10624
10691
|
cost: {
|
|
10625
|
-
input: 0.
|
|
10626
|
-
output: 1.
|
|
10627
|
-
cacheRead: 0.
|
|
10692
|
+
input: 0.39,
|
|
10693
|
+
output: 1.75,
|
|
10694
|
+
cacheRead: 0.195,
|
|
10628
10695
|
cacheWrite: 0,
|
|
10629
10696
|
},
|
|
10630
10697
|
contextWindow: 202752,
|
|
10631
|
-
maxTokens:
|
|
10698
|
+
maxTokens: 65535,
|
|
10632
10699
|
},
|
|
10633
10700
|
"z-ai/glm-4.7-flash": {
|
|
10634
10701
|
id: "z-ai/glm-4.7-flash",
|
|
@@ -10661,8 +10728,25 @@ export const MODELS = {
|
|
|
10661
10728
|
cacheRead: 0.119,
|
|
10662
10729
|
cacheWrite: 0,
|
|
10663
10730
|
},
|
|
10731
|
+
contextWindow: 80000,
|
|
10732
|
+
maxTokens: 131072,
|
|
10733
|
+
},
|
|
10734
|
+
"z-ai/glm-5-turbo": {
|
|
10735
|
+
id: "z-ai/glm-5-turbo",
|
|
10736
|
+
name: "Z.ai: GLM 5 Turbo",
|
|
10737
|
+
api: "openai-completions",
|
|
10738
|
+
provider: "openrouter",
|
|
10739
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10740
|
+
reasoning: true,
|
|
10741
|
+
input: ["text"],
|
|
10742
|
+
cost: {
|
|
10743
|
+
input: 0.96,
|
|
10744
|
+
output: 3.1999999999999997,
|
|
10745
|
+
cacheRead: 0.192,
|
|
10746
|
+
cacheWrite: 0,
|
|
10747
|
+
},
|
|
10664
10748
|
contextWindow: 202752,
|
|
10665
|
-
maxTokens:
|
|
10749
|
+
maxTokens: 131072,
|
|
10666
10750
|
},
|
|
10667
10751
|
},
|
|
10668
10752
|
"vercel-ai-gateway": {
|
|
@@ -10675,7 +10759,7 @@ export const MODELS = {
|
|
|
10675
10759
|
reasoning: true,
|
|
10676
10760
|
input: ["text"],
|
|
10677
10761
|
cost: {
|
|
10678
|
-
input: 0.
|
|
10762
|
+
input: 0.12,
|
|
10679
10763
|
output: 0.24,
|
|
10680
10764
|
cacheRead: 0,
|
|
10681
10765
|
cacheWrite: 0,
|
|
@@ -10726,13 +10810,13 @@ export const MODELS = {
|
|
|
10726
10810
|
reasoning: true,
|
|
10727
10811
|
input: ["text"],
|
|
10728
10812
|
cost: {
|
|
10729
|
-
input: 0.
|
|
10730
|
-
output: 0.
|
|
10731
|
-
cacheRead: 0,
|
|
10813
|
+
input: 0.29,
|
|
10814
|
+
output: 0.59,
|
|
10815
|
+
cacheRead: 0.145,
|
|
10732
10816
|
cacheWrite: 0,
|
|
10733
10817
|
},
|
|
10734
|
-
contextWindow:
|
|
10735
|
-
maxTokens:
|
|
10818
|
+
contextWindow: 131072,
|
|
10819
|
+
maxTokens: 40960,
|
|
10736
10820
|
},
|
|
10737
10821
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10738
10822
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10743,9 +10827,9 @@ export const MODELS = {
|
|
|
10743
10827
|
reasoning: true,
|
|
10744
10828
|
input: ["text", "image"],
|
|
10745
10829
|
cost: {
|
|
10746
|
-
input: 0.
|
|
10747
|
-
output: 2.
|
|
10748
|
-
cacheRead: 0,
|
|
10830
|
+
input: 0.22999999999999998,
|
|
10831
|
+
output: 2.3,
|
|
10832
|
+
cacheRead: 0.19999999999999998,
|
|
10749
10833
|
cacheWrite: 0,
|
|
10750
10834
|
},
|
|
10751
10835
|
contextWindow: 262114,
|
|
@@ -10762,7 +10846,7 @@ export const MODELS = {
|
|
|
10762
10846
|
cost: {
|
|
10763
10847
|
input: 0.39999999999999997,
|
|
10764
10848
|
output: 1.5999999999999999,
|
|
10765
|
-
cacheRead: 0,
|
|
10849
|
+
cacheRead: 0.022,
|
|
10766
10850
|
cacheWrite: 0,
|
|
10767
10851
|
},
|
|
10768
10852
|
contextWindow: 262144,
|
|
@@ -10777,13 +10861,13 @@ export const MODELS = {
|
|
|
10777
10861
|
reasoning: true,
|
|
10778
10862
|
input: ["text"],
|
|
10779
10863
|
cost: {
|
|
10780
|
-
input: 0.
|
|
10781
|
-
output: 0.
|
|
10864
|
+
input: 0.15,
|
|
10865
|
+
output: 0.6,
|
|
10782
10866
|
cacheRead: 0,
|
|
10783
10867
|
cacheWrite: 0,
|
|
10784
10868
|
},
|
|
10785
|
-
contextWindow:
|
|
10786
|
-
maxTokens:
|
|
10869
|
+
contextWindow: 262144,
|
|
10870
|
+
maxTokens: 8192,
|
|
10787
10871
|
},
|
|
10788
10872
|
"alibaba/qwen3-coder-next": {
|
|
10789
10873
|
id: "alibaba/qwen3-coder-next",
|
|
@@ -10791,7 +10875,7 @@ export const MODELS = {
|
|
|
10791
10875
|
api: "anthropic-messages",
|
|
10792
10876
|
provider: "vercel-ai-gateway",
|
|
10793
10877
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10794
|
-
reasoning:
|
|
10878
|
+
reasoning: false,
|
|
10795
10879
|
input: ["text"],
|
|
10796
10880
|
cost: {
|
|
10797
10881
|
input: 0.5,
|
|
@@ -10819,6 +10903,23 @@ export const MODELS = {
|
|
|
10819
10903
|
contextWindow: 1000000,
|
|
10820
10904
|
maxTokens: 65536,
|
|
10821
10905
|
},
|
|
10906
|
+
"alibaba/qwen3-max": {
|
|
10907
|
+
id: "alibaba/qwen3-max",
|
|
10908
|
+
name: "Qwen3 Max",
|
|
10909
|
+
api: "anthropic-messages",
|
|
10910
|
+
provider: "vercel-ai-gateway",
|
|
10911
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10912
|
+
reasoning: false,
|
|
10913
|
+
input: ["text"],
|
|
10914
|
+
cost: {
|
|
10915
|
+
input: 1.2,
|
|
10916
|
+
output: 6,
|
|
10917
|
+
cacheRead: 0.24,
|
|
10918
|
+
cacheWrite: 0,
|
|
10919
|
+
},
|
|
10920
|
+
contextWindow: 262144,
|
|
10921
|
+
maxTokens: 32768,
|
|
10922
|
+
},
|
|
10822
10923
|
"alibaba/qwen3-max-preview": {
|
|
10823
10924
|
id: "alibaba/qwen3-max-preview",
|
|
10824
10925
|
name: "Qwen3 Max Preview",
|
|
@@ -10966,8 +11067,8 @@ export const MODELS = {
|
|
|
10966
11067
|
cost: {
|
|
10967
11068
|
input: 3,
|
|
10968
11069
|
output: 15,
|
|
10969
|
-
cacheRead: 0,
|
|
10970
|
-
cacheWrite:
|
|
11070
|
+
cacheRead: 0.3,
|
|
11071
|
+
cacheWrite: 3.75,
|
|
10971
11072
|
},
|
|
10972
11073
|
contextWindow: 200000,
|
|
10973
11074
|
maxTokens: 8192,
|
|
@@ -11176,6 +11277,23 @@ export const MODELS = {
|
|
|
11176
11277
|
contextWindow: 256000,
|
|
11177
11278
|
maxTokens: 8000,
|
|
11178
11279
|
},
|
|
11280
|
+
"deepseek/deepseek-r1": {
|
|
11281
|
+
id: "deepseek/deepseek-r1",
|
|
11282
|
+
name: "DeepSeek-R1",
|
|
11283
|
+
api: "anthropic-messages",
|
|
11284
|
+
provider: "vercel-ai-gateway",
|
|
11285
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11286
|
+
reasoning: true,
|
|
11287
|
+
input: ["text"],
|
|
11288
|
+
cost: {
|
|
11289
|
+
input: 1.35,
|
|
11290
|
+
output: 5.4,
|
|
11291
|
+
cacheRead: 0,
|
|
11292
|
+
cacheWrite: 0,
|
|
11293
|
+
},
|
|
11294
|
+
contextWindow: 128000,
|
|
11295
|
+
maxTokens: 8192,
|
|
11296
|
+
},
|
|
11179
11297
|
"deepseek/deepseek-v3": {
|
|
11180
11298
|
id: "deepseek/deepseek-v3",
|
|
11181
11299
|
name: "DeepSeek V3 0324",
|
|
@@ -11202,13 +11320,13 @@ export const MODELS = {
|
|
|
11202
11320
|
reasoning: true,
|
|
11203
11321
|
input: ["text"],
|
|
11204
11322
|
cost: {
|
|
11205
|
-
input: 0.
|
|
11206
|
-
output:
|
|
11323
|
+
input: 0.5,
|
|
11324
|
+
output: 1.5,
|
|
11207
11325
|
cacheRead: 0,
|
|
11208
11326
|
cacheWrite: 0,
|
|
11209
11327
|
},
|
|
11210
11328
|
contextWindow: 163840,
|
|
11211
|
-
maxTokens:
|
|
11329
|
+
maxTokens: 16384,
|
|
11212
11330
|
},
|
|
11213
11331
|
"deepseek/deepseek-v3.1-terminus": {
|
|
11214
11332
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11221,7 +11339,7 @@ export const MODELS = {
|
|
|
11221
11339
|
cost: {
|
|
11222
11340
|
input: 0.27,
|
|
11223
11341
|
output: 1,
|
|
11224
|
-
cacheRead: 0,
|
|
11342
|
+
cacheRead: 0.135,
|
|
11225
11343
|
cacheWrite: 0,
|
|
11226
11344
|
},
|
|
11227
11345
|
contextWindow: 131072,
|
|
@@ -11236,9 +11354,9 @@ export const MODELS = {
|
|
|
11236
11354
|
reasoning: false,
|
|
11237
11355
|
input: ["text"],
|
|
11238
11356
|
cost: {
|
|
11239
|
-
input: 0.
|
|
11240
|
-
output: 0.
|
|
11241
|
-
cacheRead: 0.
|
|
11357
|
+
input: 0.28,
|
|
11358
|
+
output: 0.42,
|
|
11359
|
+
cacheRead: 0.028,
|
|
11242
11360
|
cacheWrite: 0,
|
|
11243
11361
|
},
|
|
11244
11362
|
contextWindow: 128000,
|
|
@@ -11261,72 +11379,72 @@ export const MODELS = {
|
|
|
11261
11379
|
contextWindow: 128000,
|
|
11262
11380
|
maxTokens: 64000,
|
|
11263
11381
|
},
|
|
11264
|
-
"google/gemini-2.
|
|
11265
|
-
id: "google/gemini-2.
|
|
11266
|
-
name: "Gemini 2.
|
|
11382
|
+
"google/gemini-2.0-flash": {
|
|
11383
|
+
id: "google/gemini-2.0-flash",
|
|
11384
|
+
name: "Gemini 2.0 Flash",
|
|
11267
11385
|
api: "anthropic-messages",
|
|
11268
11386
|
provider: "vercel-ai-gateway",
|
|
11269
11387
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11270
|
-
reasoning:
|
|
11271
|
-
input: ["text"],
|
|
11388
|
+
reasoning: false,
|
|
11389
|
+
input: ["text", "image"],
|
|
11272
11390
|
cost: {
|
|
11273
|
-
input: 0.
|
|
11274
|
-
output:
|
|
11275
|
-
cacheRead: 0,
|
|
11391
|
+
input: 0.15,
|
|
11392
|
+
output: 0.6,
|
|
11393
|
+
cacheRead: 0.024999999999999998,
|
|
11276
11394
|
cacheWrite: 0,
|
|
11277
11395
|
},
|
|
11278
|
-
contextWindow:
|
|
11279
|
-
maxTokens:
|
|
11396
|
+
contextWindow: 1048576,
|
|
11397
|
+
maxTokens: 8192,
|
|
11280
11398
|
},
|
|
11281
|
-
"google/gemini-2.
|
|
11282
|
-
id: "google/gemini-2.
|
|
11283
|
-
name: "Gemini 2.
|
|
11399
|
+
"google/gemini-2.0-flash-lite": {
|
|
11400
|
+
id: "google/gemini-2.0-flash-lite",
|
|
11401
|
+
name: "Gemini 2.0 Flash Lite",
|
|
11284
11402
|
api: "anthropic-messages",
|
|
11285
11403
|
provider: "vercel-ai-gateway",
|
|
11286
11404
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11287
|
-
reasoning:
|
|
11405
|
+
reasoning: false,
|
|
11288
11406
|
input: ["text", "image"],
|
|
11289
11407
|
cost: {
|
|
11290
|
-
input: 0.
|
|
11291
|
-
output: 0.
|
|
11292
|
-
cacheRead: 0.
|
|
11408
|
+
input: 0.075,
|
|
11409
|
+
output: 0.3,
|
|
11410
|
+
cacheRead: 0.02,
|
|
11293
11411
|
cacheWrite: 0,
|
|
11294
11412
|
},
|
|
11295
11413
|
contextWindow: 1048576,
|
|
11296
|
-
maxTokens:
|
|
11414
|
+
maxTokens: 8192,
|
|
11297
11415
|
},
|
|
11298
|
-
"google/gemini-2.5-flash
|
|
11299
|
-
id: "google/gemini-2.5-flash
|
|
11300
|
-
name: "Gemini 2.5 Flash
|
|
11416
|
+
"google/gemini-2.5-flash": {
|
|
11417
|
+
id: "google/gemini-2.5-flash",
|
|
11418
|
+
name: "Gemini 2.5 Flash",
|
|
11301
11419
|
api: "anthropic-messages",
|
|
11302
11420
|
provider: "vercel-ai-gateway",
|
|
11303
11421
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11304
11422
|
reasoning: true,
|
|
11305
11423
|
input: ["text", "image"],
|
|
11306
11424
|
cost: {
|
|
11307
|
-
input: 0.
|
|
11308
|
-
output:
|
|
11309
|
-
cacheRead: 0.
|
|
11425
|
+
input: 0.3,
|
|
11426
|
+
output: 2.5,
|
|
11427
|
+
cacheRead: 0.03,
|
|
11310
11428
|
cacheWrite: 0,
|
|
11311
11429
|
},
|
|
11312
|
-
contextWindow:
|
|
11430
|
+
contextWindow: 1000000,
|
|
11313
11431
|
maxTokens: 65536,
|
|
11314
11432
|
},
|
|
11315
|
-
"google/gemini-2.5-flash-
|
|
11316
|
-
id: "google/gemini-2.5-flash-
|
|
11317
|
-
name: "Gemini 2.5 Flash
|
|
11433
|
+
"google/gemini-2.5-flash-lite": {
|
|
11434
|
+
id: "google/gemini-2.5-flash-lite",
|
|
11435
|
+
name: "Gemini 2.5 Flash Lite",
|
|
11318
11436
|
api: "anthropic-messages",
|
|
11319
11437
|
provider: "vercel-ai-gateway",
|
|
11320
11438
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11321
11439
|
reasoning: true,
|
|
11322
11440
|
input: ["text", "image"],
|
|
11323
11441
|
cost: {
|
|
11324
|
-
input: 0.
|
|
11325
|
-
output:
|
|
11326
|
-
cacheRead: 0.
|
|
11442
|
+
input: 0.09999999999999999,
|
|
11443
|
+
output: 0.39999999999999997,
|
|
11444
|
+
cacheRead: 0.01,
|
|
11327
11445
|
cacheWrite: 0,
|
|
11328
11446
|
},
|
|
11329
|
-
contextWindow:
|
|
11447
|
+
contextWindow: 1048576,
|
|
11330
11448
|
maxTokens: 65536,
|
|
11331
11449
|
},
|
|
11332
11450
|
"google/gemini-2.5-pro": {
|
|
@@ -11336,11 +11454,11 @@ export const MODELS = {
|
|
|
11336
11454
|
provider: "vercel-ai-gateway",
|
|
11337
11455
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11338
11456
|
reasoning: true,
|
|
11339
|
-
input: ["text"],
|
|
11457
|
+
input: ["text", "image"],
|
|
11340
11458
|
cost: {
|
|
11341
11459
|
input: 1.25,
|
|
11342
11460
|
output: 10,
|
|
11343
|
-
cacheRead: 0,
|
|
11461
|
+
cacheRead: 0.125,
|
|
11344
11462
|
cacheWrite: 0,
|
|
11345
11463
|
},
|
|
11346
11464
|
contextWindow: 1048576,
|
|
@@ -11361,7 +11479,7 @@ export const MODELS = {
|
|
|
11361
11479
|
cacheWrite: 0,
|
|
11362
11480
|
},
|
|
11363
11481
|
contextWindow: 1000000,
|
|
11364
|
-
maxTokens:
|
|
11482
|
+
maxTokens: 65000,
|
|
11365
11483
|
},
|
|
11366
11484
|
"google/gemini-3-pro-preview": {
|
|
11367
11485
|
id: "google/gemini-3-pro-preview",
|
|
@@ -11463,7 +11581,7 @@ export const MODELS = {
|
|
|
11463
11581
|
cacheWrite: 0,
|
|
11464
11582
|
},
|
|
11465
11583
|
contextWindow: 128000,
|
|
11466
|
-
maxTokens:
|
|
11584
|
+
maxTokens: 100000,
|
|
11467
11585
|
},
|
|
11468
11586
|
"meituan/longcat-flash-thinking": {
|
|
11469
11587
|
id: "meituan/longcat-flash-thinking",
|
|
@@ -11491,13 +11609,13 @@ export const MODELS = {
|
|
|
11491
11609
|
reasoning: false,
|
|
11492
11610
|
input: ["text"],
|
|
11493
11611
|
cost: {
|
|
11494
|
-
input: 0.
|
|
11495
|
-
output: 0.
|
|
11612
|
+
input: 0.72,
|
|
11613
|
+
output: 0.72,
|
|
11496
11614
|
cacheRead: 0,
|
|
11497
11615
|
cacheWrite: 0,
|
|
11498
11616
|
},
|
|
11499
|
-
contextWindow:
|
|
11500
|
-
maxTokens:
|
|
11617
|
+
contextWindow: 128000,
|
|
11618
|
+
maxTokens: 8192,
|
|
11501
11619
|
},
|
|
11502
11620
|
"meta/llama-3.1-8b": {
|
|
11503
11621
|
id: "meta/llama-3.1-8b",
|
|
@@ -11508,12 +11626,12 @@ export const MODELS = {
|
|
|
11508
11626
|
reasoning: false,
|
|
11509
11627
|
input: ["text"],
|
|
11510
11628
|
cost: {
|
|
11511
|
-
input: 0.
|
|
11512
|
-
output: 0.
|
|
11513
|
-
cacheRead: 0,
|
|
11629
|
+
input: 0.09999999999999999,
|
|
11630
|
+
output: 0.09999999999999999,
|
|
11631
|
+
cacheRead: 0.09999999999999999,
|
|
11514
11632
|
cacheWrite: 0,
|
|
11515
11633
|
},
|
|
11516
|
-
contextWindow:
|
|
11634
|
+
contextWindow: 128000,
|
|
11517
11635
|
maxTokens: 16384,
|
|
11518
11636
|
},
|
|
11519
11637
|
"meta/llama-3.2-11b": {
|
|
@@ -11576,12 +11694,12 @@ export const MODELS = {
|
|
|
11576
11694
|
reasoning: false,
|
|
11577
11695
|
input: ["text", "image"],
|
|
11578
11696
|
cost: {
|
|
11579
|
-
input: 0.
|
|
11580
|
-
output: 0.
|
|
11697
|
+
input: 0.24,
|
|
11698
|
+
output: 0.9700000000000001,
|
|
11581
11699
|
cacheRead: 0,
|
|
11582
11700
|
cacheWrite: 0,
|
|
11583
11701
|
},
|
|
11584
|
-
contextWindow:
|
|
11702
|
+
contextWindow: 128000,
|
|
11585
11703
|
maxTokens: 8192,
|
|
11586
11704
|
},
|
|
11587
11705
|
"meta/llama-4-scout": {
|
|
@@ -11593,12 +11711,12 @@ export const MODELS = {
|
|
|
11593
11711
|
reasoning: false,
|
|
11594
11712
|
input: ["text", "image"],
|
|
11595
11713
|
cost: {
|
|
11596
|
-
input: 0.
|
|
11597
|
-
output: 0.
|
|
11714
|
+
input: 0.16999999999999998,
|
|
11715
|
+
output: 0.66,
|
|
11598
11716
|
cacheRead: 0,
|
|
11599
11717
|
cacheWrite: 0,
|
|
11600
11718
|
},
|
|
11601
|
-
contextWindow:
|
|
11719
|
+
contextWindow: 128000,
|
|
11602
11720
|
maxTokens: 8192,
|
|
11603
11721
|
},
|
|
11604
11722
|
"minimax/minimax-m2": {
|
|
@@ -11629,8 +11747,8 @@ export const MODELS = {
|
|
|
11629
11747
|
cost: {
|
|
11630
11748
|
input: 0.3,
|
|
11631
11749
|
output: 1.2,
|
|
11632
|
-
cacheRead: 0.
|
|
11633
|
-
cacheWrite: 0,
|
|
11750
|
+
cacheRead: 0.03,
|
|
11751
|
+
cacheWrite: 0.375,
|
|
11634
11752
|
},
|
|
11635
11753
|
contextWindow: 204800,
|
|
11636
11754
|
maxTokens: 131072,
|
|
@@ -11683,8 +11801,42 @@ export const MODELS = {
|
|
|
11683
11801
|
cacheRead: 0.03,
|
|
11684
11802
|
cacheWrite: 0.375,
|
|
11685
11803
|
},
|
|
11686
|
-
contextWindow:
|
|
11687
|
-
maxTokens:
|
|
11804
|
+
contextWindow: 204800,
|
|
11805
|
+
maxTokens: 131000,
|
|
11806
|
+
},
|
|
11807
|
+
"minimax/minimax-m2.7": {
|
|
11808
|
+
id: "minimax/minimax-m2.7",
|
|
11809
|
+
name: "Minimax M2.7",
|
|
11810
|
+
api: "anthropic-messages",
|
|
11811
|
+
provider: "vercel-ai-gateway",
|
|
11812
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11813
|
+
reasoning: true,
|
|
11814
|
+
input: ["text", "image"],
|
|
11815
|
+
cost: {
|
|
11816
|
+
input: 0.3,
|
|
11817
|
+
output: 1.2,
|
|
11818
|
+
cacheRead: 0.06,
|
|
11819
|
+
cacheWrite: 0.375,
|
|
11820
|
+
},
|
|
11821
|
+
contextWindow: 204800,
|
|
11822
|
+
maxTokens: 131000,
|
|
11823
|
+
},
|
|
11824
|
+
"minimax/minimax-m2.7-highspeed": {
|
|
11825
|
+
id: "minimax/minimax-m2.7-highspeed",
|
|
11826
|
+
name: "MiniMax M2.7 High Speed",
|
|
11827
|
+
api: "anthropic-messages",
|
|
11828
|
+
provider: "vercel-ai-gateway",
|
|
11829
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11830
|
+
reasoning: true,
|
|
11831
|
+
input: ["text", "image"],
|
|
11832
|
+
cost: {
|
|
11833
|
+
input: 0.6,
|
|
11834
|
+
output: 2.4,
|
|
11835
|
+
cacheRead: 0.06,
|
|
11836
|
+
cacheWrite: 0.375,
|
|
11837
|
+
},
|
|
11838
|
+
contextWindow: 204800,
|
|
11839
|
+
maxTokens: 131100,
|
|
11688
11840
|
},
|
|
11689
11841
|
"mistral/codestral": {
|
|
11690
11842
|
id: "mistral/codestral",
|
|
@@ -11712,8 +11864,8 @@ export const MODELS = {
|
|
|
11712
11864
|
reasoning: false,
|
|
11713
11865
|
input: ["text"],
|
|
11714
11866
|
cost: {
|
|
11715
|
-
input: 0,
|
|
11716
|
-
output:
|
|
11867
|
+
input: 0.39999999999999997,
|
|
11868
|
+
output: 2,
|
|
11717
11869
|
cacheRead: 0,
|
|
11718
11870
|
cacheWrite: 0,
|
|
11719
11871
|
},
|
|
@@ -11746,8 +11898,8 @@ export const MODELS = {
|
|
|
11746
11898
|
reasoning: false,
|
|
11747
11899
|
input: ["text"],
|
|
11748
11900
|
cost: {
|
|
11749
|
-
input: 0,
|
|
11750
|
-
output: 0,
|
|
11901
|
+
input: 0.09999999999999999,
|
|
11902
|
+
output: 0.3,
|
|
11751
11903
|
cacheRead: 0,
|
|
11752
11904
|
cacheWrite: 0,
|
|
11753
11905
|
},
|
|
@@ -11763,8 +11915,8 @@ export const MODELS = {
|
|
|
11763
11915
|
reasoning: false,
|
|
11764
11916
|
input: ["text"],
|
|
11765
11917
|
cost: {
|
|
11766
|
-
input: 0.
|
|
11767
|
-
output: 0.
|
|
11918
|
+
input: 0.09999999999999999,
|
|
11919
|
+
output: 0.09999999999999999,
|
|
11768
11920
|
cacheRead: 0,
|
|
11769
11921
|
cacheWrite: 0,
|
|
11770
11922
|
},
|
|
@@ -11780,8 +11932,8 @@ export const MODELS = {
|
|
|
11780
11932
|
reasoning: false,
|
|
11781
11933
|
input: ["text"],
|
|
11782
11934
|
cost: {
|
|
11783
|
-
input: 0.
|
|
11784
|
-
output: 0.
|
|
11935
|
+
input: 0.15,
|
|
11936
|
+
output: 0.15,
|
|
11785
11937
|
cacheRead: 0,
|
|
11786
11938
|
cacheWrite: 0,
|
|
11787
11939
|
},
|
|
@@ -11865,14 +12017,31 @@ export const MODELS = {
|
|
|
11865
12017
|
reasoning: false,
|
|
11866
12018
|
input: ["text"],
|
|
11867
12019
|
cost: {
|
|
11868
|
-
input: 0.
|
|
11869
|
-
output: 2,
|
|
11870
|
-
cacheRead: 0,
|
|
12020
|
+
input: 0.6,
|
|
12021
|
+
output: 2.5,
|
|
12022
|
+
cacheRead: 0.15,
|
|
11871
12023
|
cacheWrite: 0,
|
|
11872
12024
|
},
|
|
11873
12025
|
contextWindow: 131072,
|
|
11874
12026
|
maxTokens: 16384,
|
|
11875
12027
|
},
|
|
12028
|
+
"moonshotai/kimi-k2-0905": {
|
|
12029
|
+
id: "moonshotai/kimi-k2-0905",
|
|
12030
|
+
name: "Kimi K2 0905",
|
|
12031
|
+
api: "anthropic-messages",
|
|
12032
|
+
provider: "vercel-ai-gateway",
|
|
12033
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12034
|
+
reasoning: false,
|
|
12035
|
+
input: ["text"],
|
|
12036
|
+
cost: {
|
|
12037
|
+
input: 0.6,
|
|
12038
|
+
output: 2.5,
|
|
12039
|
+
cacheRead: 0.15,
|
|
12040
|
+
cacheWrite: 0,
|
|
12041
|
+
},
|
|
12042
|
+
contextWindow: 256000,
|
|
12043
|
+
maxTokens: 16384,
|
|
12044
|
+
},
|
|
11876
12045
|
"moonshotai/kimi-k2-thinking": {
|
|
11877
12046
|
id: "moonshotai/kimi-k2-thinking",
|
|
11878
12047
|
name: "Kimi K2 Thinking",
|
|
@@ -11882,13 +12051,13 @@ export const MODELS = {
|
|
|
11882
12051
|
reasoning: true,
|
|
11883
12052
|
input: ["text"],
|
|
11884
12053
|
cost: {
|
|
11885
|
-
input: 0.
|
|
11886
|
-
output: 2,
|
|
11887
|
-
cacheRead: 0.
|
|
12054
|
+
input: 0.6,
|
|
12055
|
+
output: 2.5,
|
|
12056
|
+
cacheRead: 0.15,
|
|
11888
12057
|
cacheWrite: 0,
|
|
11889
12058
|
},
|
|
11890
|
-
contextWindow:
|
|
11891
|
-
maxTokens:
|
|
12059
|
+
contextWindow: 262114,
|
|
12060
|
+
maxTokens: 262114,
|
|
11892
12061
|
},
|
|
11893
12062
|
"moonshotai/kimi-k2-thinking-turbo": {
|
|
11894
12063
|
id: "moonshotai/kimi-k2-thinking-turbo",
|
|
@@ -11916,9 +12085,9 @@ export const MODELS = {
|
|
|
11916
12085
|
reasoning: false,
|
|
11917
12086
|
input: ["text"],
|
|
11918
12087
|
cost: {
|
|
11919
|
-
input:
|
|
11920
|
-
output:
|
|
11921
|
-
cacheRead: 0,
|
|
12088
|
+
input: 1.15,
|
|
12089
|
+
output: 8,
|
|
12090
|
+
cacheRead: 0.15,
|
|
11922
12091
|
cacheWrite: 0,
|
|
11923
12092
|
},
|
|
11924
12093
|
contextWindow: 256000,
|
|
@@ -11933,13 +12102,13 @@ export const MODELS = {
|
|
|
11933
12102
|
reasoning: true,
|
|
11934
12103
|
input: ["text", "image"],
|
|
11935
12104
|
cost: {
|
|
11936
|
-
input: 0.
|
|
11937
|
-
output:
|
|
11938
|
-
cacheRead: 0,
|
|
12105
|
+
input: 0.6,
|
|
12106
|
+
output: 3,
|
|
12107
|
+
cacheRead: 0.09999999999999999,
|
|
11939
12108
|
cacheWrite: 0,
|
|
11940
12109
|
},
|
|
11941
|
-
contextWindow:
|
|
11942
|
-
maxTokens:
|
|
12110
|
+
contextWindow: 262114,
|
|
12111
|
+
maxTokens: 262114,
|
|
11943
12112
|
},
|
|
11944
12113
|
"nvidia/nemotron-nano-12b-v2-vl": {
|
|
11945
12114
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
@@ -11967,31 +12136,14 @@ export const MODELS = {
|
|
|
11967
12136
|
reasoning: true,
|
|
11968
12137
|
input: ["text"],
|
|
11969
12138
|
cost: {
|
|
11970
|
-
input: 0.
|
|
11971
|
-
output: 0.
|
|
12139
|
+
input: 0.06,
|
|
12140
|
+
output: 0.22999999999999998,
|
|
11972
12141
|
cacheRead: 0,
|
|
11973
12142
|
cacheWrite: 0,
|
|
11974
12143
|
},
|
|
11975
12144
|
contextWindow: 131072,
|
|
11976
12145
|
maxTokens: 131072,
|
|
11977
12146
|
},
|
|
11978
|
-
"openai/codex-mini": {
|
|
11979
|
-
id: "openai/codex-mini",
|
|
11980
|
-
name: "Codex Mini",
|
|
11981
|
-
api: "anthropic-messages",
|
|
11982
|
-
provider: "vercel-ai-gateway",
|
|
11983
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11984
|
-
reasoning: true,
|
|
11985
|
-
input: ["text", "image"],
|
|
11986
|
-
cost: {
|
|
11987
|
-
input: 1.5,
|
|
11988
|
-
output: 6,
|
|
11989
|
-
cacheRead: 0.375,
|
|
11990
|
-
cacheWrite: 0,
|
|
11991
|
-
},
|
|
11992
|
-
contextWindow: 200000,
|
|
11993
|
-
maxTokens: 100000,
|
|
11994
|
-
},
|
|
11995
12147
|
"openai/gpt-4-turbo": {
|
|
11996
12148
|
id: "openai/gpt-4-turbo",
|
|
11997
12149
|
name: "GPT-4 Turbo",
|
|
@@ -12054,7 +12206,7 @@ export const MODELS = {
|
|
|
12054
12206
|
cost: {
|
|
12055
12207
|
input: 0.09999999999999999,
|
|
12056
12208
|
output: 0.39999999999999997,
|
|
12057
|
-
cacheRead: 0.
|
|
12209
|
+
cacheRead: 0.024999999999999998,
|
|
12058
12210
|
cacheWrite: 0,
|
|
12059
12211
|
},
|
|
12060
12212
|
contextWindow: 1047576,
|
|
@@ -12105,7 +12257,7 @@ export const MODELS = {
|
|
|
12105
12257
|
cost: {
|
|
12106
12258
|
input: 1.25,
|
|
12107
12259
|
output: 10,
|
|
12108
|
-
cacheRead: 0.
|
|
12260
|
+
cacheRead: 0.125,
|
|
12109
12261
|
cacheWrite: 0,
|
|
12110
12262
|
},
|
|
12111
12263
|
contextWindow: 400000,
|
|
@@ -12135,11 +12287,11 @@ export const MODELS = {
|
|
|
12135
12287
|
provider: "vercel-ai-gateway",
|
|
12136
12288
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12137
12289
|
reasoning: true,
|
|
12138
|
-
input: ["text"
|
|
12290
|
+
input: ["text"],
|
|
12139
12291
|
cost: {
|
|
12140
12292
|
input: 1.25,
|
|
12141
12293
|
output: 10,
|
|
12142
|
-
cacheRead: 0.
|
|
12294
|
+
cacheRead: 0.125,
|
|
12143
12295
|
cacheWrite: 0,
|
|
12144
12296
|
},
|
|
12145
12297
|
contextWindow: 400000,
|
|
@@ -12156,7 +12308,7 @@ export const MODELS = {
|
|
|
12156
12308
|
cost: {
|
|
12157
12309
|
input: 0.25,
|
|
12158
12310
|
output: 2,
|
|
12159
|
-
cacheRead: 0.
|
|
12311
|
+
cacheRead: 0.024999999999999998,
|
|
12160
12312
|
cacheWrite: 0,
|
|
12161
12313
|
},
|
|
12162
12314
|
contextWindow: 400000,
|
|
@@ -12173,7 +12325,7 @@ export const MODELS = {
|
|
|
12173
12325
|
cost: {
|
|
12174
12326
|
input: 0.049999999999999996,
|
|
12175
12327
|
output: 0.39999999999999997,
|
|
12176
|
-
cacheRead: 0.
|
|
12328
|
+
cacheRead: 0.005,
|
|
12177
12329
|
cacheWrite: 0,
|
|
12178
12330
|
},
|
|
12179
12331
|
contextWindow: 400000,
|
|
@@ -12207,7 +12359,7 @@ export const MODELS = {
|
|
|
12207
12359
|
cost: {
|
|
12208
12360
|
input: 1.25,
|
|
12209
12361
|
output: 10,
|
|
12210
|
-
cacheRead: 0.
|
|
12362
|
+
cacheRead: 0.125,
|
|
12211
12363
|
cacheWrite: 0,
|
|
12212
12364
|
},
|
|
12213
12365
|
contextWindow: 400000,
|
|
@@ -12258,7 +12410,7 @@ export const MODELS = {
|
|
|
12258
12410
|
cost: {
|
|
12259
12411
|
input: 1.25,
|
|
12260
12412
|
output: 10,
|
|
12261
|
-
cacheRead: 0.
|
|
12413
|
+
cacheRead: 0.125,
|
|
12262
12414
|
cacheWrite: 0,
|
|
12263
12415
|
},
|
|
12264
12416
|
contextWindow: 128000,
|
|
@@ -12275,7 +12427,7 @@ export const MODELS = {
|
|
|
12275
12427
|
cost: {
|
|
12276
12428
|
input: 1.25,
|
|
12277
12429
|
output: 10,
|
|
12278
|
-
cacheRead: 0.
|
|
12430
|
+
cacheRead: 0.125,
|
|
12279
12431
|
cacheWrite: 0,
|
|
12280
12432
|
},
|
|
12281
12433
|
contextWindow: 400000,
|
|
@@ -12292,7 +12444,7 @@ export const MODELS = {
|
|
|
12292
12444
|
cost: {
|
|
12293
12445
|
input: 1.75,
|
|
12294
12446
|
output: 14,
|
|
12295
|
-
cacheRead: 0.
|
|
12447
|
+
cacheRead: 0.175,
|
|
12296
12448
|
cacheWrite: 0,
|
|
12297
12449
|
},
|
|
12298
12450
|
contextWindow: 400000,
|
|
@@ -12397,42 +12549,59 @@ export const MODELS = {
|
|
|
12397
12549
|
cacheRead: 0.25,
|
|
12398
12550
|
cacheWrite: 0,
|
|
12399
12551
|
},
|
|
12400
|
-
contextWindow:
|
|
12552
|
+
contextWindow: 1050000,
|
|
12401
12553
|
maxTokens: 128000,
|
|
12402
12554
|
},
|
|
12403
|
-
"openai/gpt-5.4-
|
|
12404
|
-
id: "openai/gpt-5.4-
|
|
12405
|
-
name: "GPT 5.4
|
|
12555
|
+
"openai/gpt-5.4-mini": {
|
|
12556
|
+
id: "openai/gpt-5.4-mini",
|
|
12557
|
+
name: "GPT 5.4 Mini",
|
|
12406
12558
|
api: "anthropic-messages",
|
|
12407
12559
|
provider: "vercel-ai-gateway",
|
|
12408
12560
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12409
12561
|
reasoning: true,
|
|
12410
12562
|
input: ["text", "image"],
|
|
12411
12563
|
cost: {
|
|
12412
|
-
input:
|
|
12413
|
-
output:
|
|
12414
|
-
cacheRead: 0,
|
|
12564
|
+
input: 0.75,
|
|
12565
|
+
output: 4.5,
|
|
12566
|
+
cacheRead: 0.075,
|
|
12415
12567
|
cacheWrite: 0,
|
|
12416
12568
|
},
|
|
12417
|
-
contextWindow:
|
|
12569
|
+
contextWindow: 400000,
|
|
12418
12570
|
maxTokens: 128000,
|
|
12419
12571
|
},
|
|
12420
|
-
"openai/gpt-
|
|
12421
|
-
id: "openai/gpt-
|
|
12422
|
-
name: "
|
|
12572
|
+
"openai/gpt-5.4-nano": {
|
|
12573
|
+
id: "openai/gpt-5.4-nano",
|
|
12574
|
+
name: "GPT 5.4 Nano",
|
|
12423
12575
|
api: "anthropic-messages",
|
|
12424
12576
|
provider: "vercel-ai-gateway",
|
|
12425
12577
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12426
12578
|
reasoning: true,
|
|
12427
|
-
input: ["text"],
|
|
12579
|
+
input: ["text", "image"],
|
|
12428
12580
|
cost: {
|
|
12429
|
-
input: 0.
|
|
12430
|
-
output:
|
|
12581
|
+
input: 0.19999999999999998,
|
|
12582
|
+
output: 1.25,
|
|
12583
|
+
cacheRead: 0.02,
|
|
12584
|
+
cacheWrite: 0,
|
|
12585
|
+
},
|
|
12586
|
+
contextWindow: 400000,
|
|
12587
|
+
maxTokens: 128000,
|
|
12588
|
+
},
|
|
12589
|
+
"openai/gpt-5.4-pro": {
|
|
12590
|
+
id: "openai/gpt-5.4-pro",
|
|
12591
|
+
name: "GPT 5.4 Pro",
|
|
12592
|
+
api: "anthropic-messages",
|
|
12593
|
+
provider: "vercel-ai-gateway",
|
|
12594
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12595
|
+
reasoning: true,
|
|
12596
|
+
input: ["text", "image"],
|
|
12597
|
+
cost: {
|
|
12598
|
+
input: 30,
|
|
12599
|
+
output: 180,
|
|
12431
12600
|
cacheRead: 0,
|
|
12432
12601
|
cacheWrite: 0,
|
|
12433
12602
|
},
|
|
12434
|
-
contextWindow:
|
|
12435
|
-
maxTokens:
|
|
12603
|
+
contextWindow: 1050000,
|
|
12604
|
+
maxTokens: 128000,
|
|
12436
12605
|
},
|
|
12437
12606
|
"openai/gpt-oss-20b": {
|
|
12438
12607
|
id: "openai/gpt-oss-20b",
|
|
@@ -12621,40 +12790,6 @@ export const MODELS = {
|
|
|
12621
12790
|
contextWindow: 131072,
|
|
12622
12791
|
maxTokens: 131072,
|
|
12623
12792
|
},
|
|
12624
|
-
"vercel/v0-1.0-md": {
|
|
12625
|
-
id: "vercel/v0-1.0-md",
|
|
12626
|
-
name: "v0-1.0-md",
|
|
12627
|
-
api: "anthropic-messages",
|
|
12628
|
-
provider: "vercel-ai-gateway",
|
|
12629
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12630
|
-
reasoning: false,
|
|
12631
|
-
input: ["text", "image"],
|
|
12632
|
-
cost: {
|
|
12633
|
-
input: 3,
|
|
12634
|
-
output: 15,
|
|
12635
|
-
cacheRead: 0,
|
|
12636
|
-
cacheWrite: 0,
|
|
12637
|
-
},
|
|
12638
|
-
contextWindow: 128000,
|
|
12639
|
-
maxTokens: 32000,
|
|
12640
|
-
},
|
|
12641
|
-
"vercel/v0-1.5-md": {
|
|
12642
|
-
id: "vercel/v0-1.5-md",
|
|
12643
|
-
name: "v0-1.5-md",
|
|
12644
|
-
api: "anthropic-messages",
|
|
12645
|
-
provider: "vercel-ai-gateway",
|
|
12646
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12647
|
-
reasoning: false,
|
|
12648
|
-
input: ["text", "image"],
|
|
12649
|
-
cost: {
|
|
12650
|
-
input: 3,
|
|
12651
|
-
output: 15,
|
|
12652
|
-
cacheRead: 0,
|
|
12653
|
-
cacheWrite: 0,
|
|
12654
|
-
},
|
|
12655
|
-
contextWindow: 128000,
|
|
12656
|
-
maxTokens: 32768,
|
|
12657
|
-
},
|
|
12658
12793
|
"xai/grok-2-vision": {
|
|
12659
12794
|
id: "xai/grok-2-vision",
|
|
12660
12795
|
name: "Grok 2 Vision",
|
|
@@ -12683,7 +12818,7 @@ export const MODELS = {
|
|
|
12683
12818
|
cost: {
|
|
12684
12819
|
input: 3,
|
|
12685
12820
|
output: 15,
|
|
12686
|
-
cacheRead: 0,
|
|
12821
|
+
cacheRead: 0.75,
|
|
12687
12822
|
cacheWrite: 0,
|
|
12688
12823
|
},
|
|
12689
12824
|
contextWindow: 131072,
|
|
@@ -12700,7 +12835,7 @@ export const MODELS = {
|
|
|
12700
12835
|
cost: {
|
|
12701
12836
|
input: 5,
|
|
12702
12837
|
output: 25,
|
|
12703
|
-
cacheRead:
|
|
12838
|
+
cacheRead: 1.25,
|
|
12704
12839
|
cacheWrite: 0,
|
|
12705
12840
|
},
|
|
12706
12841
|
contextWindow: 131072,
|
|
@@ -12717,7 +12852,7 @@ export const MODELS = {
|
|
|
12717
12852
|
cost: {
|
|
12718
12853
|
input: 0.3,
|
|
12719
12854
|
output: 0.5,
|
|
12720
|
-
cacheRead: 0,
|
|
12855
|
+
cacheRead: 0.075,
|
|
12721
12856
|
cacheWrite: 0,
|
|
12722
12857
|
},
|
|
12723
12858
|
contextWindow: 131072,
|
|
@@ -12751,7 +12886,7 @@ export const MODELS = {
|
|
|
12751
12886
|
cost: {
|
|
12752
12887
|
input: 3,
|
|
12753
12888
|
output: 15,
|
|
12754
|
-
cacheRead: 0,
|
|
12889
|
+
cacheRead: 0.75,
|
|
12755
12890
|
cacheWrite: 0,
|
|
12756
12891
|
},
|
|
12757
12892
|
contextWindow: 256000,
|
|
@@ -12825,6 +12960,57 @@ export const MODELS = {
|
|
|
12825
12960
|
contextWindow: 2000000,
|
|
12826
12961
|
maxTokens: 30000,
|
|
12827
12962
|
},
|
|
12963
|
+
"xai/grok-4.20-multi-agent-beta": {
|
|
12964
|
+
id: "xai/grok-4.20-multi-agent-beta",
|
|
12965
|
+
name: "Grok 4.20 Multi Agent Beta",
|
|
12966
|
+
api: "anthropic-messages",
|
|
12967
|
+
provider: "vercel-ai-gateway",
|
|
12968
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12969
|
+
reasoning: true,
|
|
12970
|
+
input: ["text"],
|
|
12971
|
+
cost: {
|
|
12972
|
+
input: 2,
|
|
12973
|
+
output: 6,
|
|
12974
|
+
cacheRead: 0.19999999999999998,
|
|
12975
|
+
cacheWrite: 0,
|
|
12976
|
+
},
|
|
12977
|
+
contextWindow: 2000000,
|
|
12978
|
+
maxTokens: 2000000,
|
|
12979
|
+
},
|
|
12980
|
+
"xai/grok-4.20-non-reasoning-beta": {
|
|
12981
|
+
id: "xai/grok-4.20-non-reasoning-beta",
|
|
12982
|
+
name: "Grok 4.20 Beta Non-Reasoning",
|
|
12983
|
+
api: "anthropic-messages",
|
|
12984
|
+
provider: "vercel-ai-gateway",
|
|
12985
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12986
|
+
reasoning: false,
|
|
12987
|
+
input: ["text", "image"],
|
|
12988
|
+
cost: {
|
|
12989
|
+
input: 2,
|
|
12990
|
+
output: 6,
|
|
12991
|
+
cacheRead: 0.19999999999999998,
|
|
12992
|
+
cacheWrite: 0,
|
|
12993
|
+
},
|
|
12994
|
+
contextWindow: 2000000,
|
|
12995
|
+
maxTokens: 2000000,
|
|
12996
|
+
},
|
|
12997
|
+
"xai/grok-4.20-reasoning-beta": {
|
|
12998
|
+
id: "xai/grok-4.20-reasoning-beta",
|
|
12999
|
+
name: "Grok 4.20 Beta Reasoning",
|
|
13000
|
+
api: "anthropic-messages",
|
|
13001
|
+
provider: "vercel-ai-gateway",
|
|
13002
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13003
|
+
reasoning: true,
|
|
13004
|
+
input: ["text", "image"],
|
|
13005
|
+
cost: {
|
|
13006
|
+
input: 2,
|
|
13007
|
+
output: 6,
|
|
13008
|
+
cacheRead: 0.19999999999999998,
|
|
13009
|
+
cacheWrite: 0,
|
|
13010
|
+
},
|
|
13011
|
+
contextWindow: 2000000,
|
|
13012
|
+
maxTokens: 2000000,
|
|
13013
|
+
},
|
|
12828
13014
|
"xai/grok-code-fast-1": {
|
|
12829
13015
|
id: "xai/grok-code-fast-1",
|
|
12830
13016
|
name: "Grok Code Fast 1",
|
|
@@ -12851,17 +13037,34 @@ export const MODELS = {
|
|
|
12851
13037
|
reasoning: true,
|
|
12852
13038
|
input: ["text"],
|
|
12853
13039
|
cost: {
|
|
12854
|
-
input: 0.
|
|
12855
|
-
output: 0.
|
|
12856
|
-
cacheRead: 0,
|
|
13040
|
+
input: 0.09999999999999999,
|
|
13041
|
+
output: 0.3,
|
|
13042
|
+
cacheRead: 0.02,
|
|
12857
13043
|
cacheWrite: 0,
|
|
12858
13044
|
},
|
|
12859
13045
|
contextWindow: 262144,
|
|
12860
13046
|
maxTokens: 32000,
|
|
12861
13047
|
},
|
|
12862
|
-
"
|
|
12863
|
-
id: "
|
|
12864
|
-
name: "
|
|
13048
|
+
"xiaomi/mimo-v2-pro": {
|
|
13049
|
+
id: "xiaomi/mimo-v2-pro",
|
|
13050
|
+
name: "MiMo V2 Pro",
|
|
13051
|
+
api: "anthropic-messages",
|
|
13052
|
+
provider: "vercel-ai-gateway",
|
|
13053
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13054
|
+
reasoning: true,
|
|
13055
|
+
input: ["text"],
|
|
13056
|
+
cost: {
|
|
13057
|
+
input: 1,
|
|
13058
|
+
output: 3,
|
|
13059
|
+
cacheRead: 0.19999999999999998,
|
|
13060
|
+
cacheWrite: 0,
|
|
13061
|
+
},
|
|
13062
|
+
contextWindow: 1000000,
|
|
13063
|
+
maxTokens: 128000,
|
|
13064
|
+
},
|
|
13065
|
+
"zai/glm-4.5": {
|
|
13066
|
+
id: "zai/glm-4.5",
|
|
13067
|
+
name: "GLM-4.5",
|
|
12865
13068
|
api: "anthropic-messages",
|
|
12866
13069
|
provider: "vercel-ai-gateway",
|
|
12867
13070
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -12870,11 +13073,11 @@ export const MODELS = {
|
|
|
12870
13073
|
cost: {
|
|
12871
13074
|
input: 0.6,
|
|
12872
13075
|
output: 2.2,
|
|
12873
|
-
cacheRead: 0,
|
|
13076
|
+
cacheRead: 0.11,
|
|
12874
13077
|
cacheWrite: 0,
|
|
12875
13078
|
},
|
|
12876
|
-
contextWindow:
|
|
12877
|
-
maxTokens:
|
|
13079
|
+
contextWindow: 128000,
|
|
13080
|
+
maxTokens: 96000,
|
|
12878
13081
|
},
|
|
12879
13082
|
"zai/glm-4.5-air": {
|
|
12880
13083
|
id: "zai/glm-4.5-air",
|
|
@@ -12899,16 +13102,16 @@ export const MODELS = {
|
|
|
12899
13102
|
api: "anthropic-messages",
|
|
12900
13103
|
provider: "vercel-ai-gateway",
|
|
12901
13104
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12902
|
-
reasoning:
|
|
13105
|
+
reasoning: false,
|
|
12903
13106
|
input: ["text", "image"],
|
|
12904
13107
|
cost: {
|
|
12905
13108
|
input: 0.6,
|
|
12906
13109
|
output: 1.7999999999999998,
|
|
12907
|
-
cacheRead: 0,
|
|
13110
|
+
cacheRead: 0.11,
|
|
12908
13111
|
cacheWrite: 0,
|
|
12909
13112
|
},
|
|
12910
|
-
contextWindow:
|
|
12911
|
-
maxTokens:
|
|
13113
|
+
contextWindow: 66000,
|
|
13114
|
+
maxTokens: 16000,
|
|
12912
13115
|
},
|
|
12913
13116
|
"zai/glm-4.6": {
|
|
12914
13117
|
id: "zai/glm-4.6",
|
|
@@ -12919,8 +13122,8 @@ export const MODELS = {
|
|
|
12919
13122
|
reasoning: true,
|
|
12920
13123
|
input: ["text"],
|
|
12921
13124
|
cost: {
|
|
12922
|
-
input: 0.
|
|
12923
|
-
output:
|
|
13125
|
+
input: 0.6,
|
|
13126
|
+
output: 2.2,
|
|
12924
13127
|
cacheRead: 0.11,
|
|
12925
13128
|
cacheWrite: 0,
|
|
12926
13129
|
},
|
|
@@ -12970,14 +13173,31 @@ export const MODELS = {
|
|
|
12970
13173
|
reasoning: true,
|
|
12971
13174
|
input: ["text"],
|
|
12972
13175
|
cost: {
|
|
12973
|
-
input: 0.
|
|
12974
|
-
output:
|
|
12975
|
-
cacheRead: 0
|
|
13176
|
+
input: 0.6,
|
|
13177
|
+
output: 2.2,
|
|
13178
|
+
cacheRead: 0,
|
|
12976
13179
|
cacheWrite: 0,
|
|
12977
13180
|
},
|
|
12978
|
-
contextWindow:
|
|
13181
|
+
contextWindow: 200000,
|
|
12979
13182
|
maxTokens: 120000,
|
|
12980
13183
|
},
|
|
13184
|
+
"zai/glm-4.7-flash": {
|
|
13185
|
+
id: "zai/glm-4.7-flash",
|
|
13186
|
+
name: "GLM 4.7 Flash",
|
|
13187
|
+
api: "anthropic-messages",
|
|
13188
|
+
provider: "vercel-ai-gateway",
|
|
13189
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13190
|
+
reasoning: true,
|
|
13191
|
+
input: ["text"],
|
|
13192
|
+
cost: {
|
|
13193
|
+
input: 0.07,
|
|
13194
|
+
output: 0.39999999999999997,
|
|
13195
|
+
cacheRead: 0,
|
|
13196
|
+
cacheWrite: 0,
|
|
13197
|
+
},
|
|
13198
|
+
contextWindow: 200000,
|
|
13199
|
+
maxTokens: 131000,
|
|
13200
|
+
},
|
|
12981
13201
|
"zai/glm-4.7-flashx": {
|
|
12982
13202
|
id: "zai/glm-4.7-flashx",
|
|
12983
13203
|
name: "GLM 4.7 FlashX",
|
|
@@ -12997,7 +13217,7 @@ export const MODELS = {
|
|
|
12997
13217
|
},
|
|
12998
13218
|
"zai/glm-5": {
|
|
12999
13219
|
id: "zai/glm-5",
|
|
13000
|
-
name: "GLM
|
|
13220
|
+
name: "GLM 5",
|
|
13001
13221
|
api: "anthropic-messages",
|
|
13002
13222
|
provider: "vercel-ai-gateway",
|
|
13003
13223
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -13010,7 +13230,24 @@ export const MODELS = {
|
|
|
13010
13230
|
cacheWrite: 0,
|
|
13011
13231
|
},
|
|
13012
13232
|
contextWindow: 202800,
|
|
13013
|
-
maxTokens:
|
|
13233
|
+
maxTokens: 131100,
|
|
13234
|
+
},
|
|
13235
|
+
"zai/glm-5-turbo": {
|
|
13236
|
+
id: "zai/glm-5-turbo",
|
|
13237
|
+
name: "GLM 5 Turbo",
|
|
13238
|
+
api: "anthropic-messages",
|
|
13239
|
+
provider: "vercel-ai-gateway",
|
|
13240
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13241
|
+
reasoning: true,
|
|
13242
|
+
input: ["text"],
|
|
13243
|
+
cost: {
|
|
13244
|
+
input: 1.2,
|
|
13245
|
+
output: 4,
|
|
13246
|
+
cacheRead: 0.24,
|
|
13247
|
+
cacheWrite: 0,
|
|
13248
|
+
},
|
|
13249
|
+
contextWindow: 202800,
|
|
13250
|
+
maxTokens: 131100,
|
|
13014
13251
|
},
|
|
13015
13252
|
},
|
|
13016
13253
|
"xai": {
|
|
@@ -13337,6 +13574,40 @@ export const MODELS = {
|
|
|
13337
13574
|
contextWindow: 2000000,
|
|
13338
13575
|
maxTokens: 30000,
|
|
13339
13576
|
},
|
|
13577
|
+
"grok-4.20-0309-non-reasoning": {
|
|
13578
|
+
id: "grok-4.20-0309-non-reasoning",
|
|
13579
|
+
name: "Grok 4.20 (Non-Reasoning)",
|
|
13580
|
+
api: "openai-completions",
|
|
13581
|
+
provider: "xai",
|
|
13582
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13583
|
+
reasoning: false,
|
|
13584
|
+
input: ["text", "image"],
|
|
13585
|
+
cost: {
|
|
13586
|
+
input: 2,
|
|
13587
|
+
output: 6,
|
|
13588
|
+
cacheRead: 0.2,
|
|
13589
|
+
cacheWrite: 0,
|
|
13590
|
+
},
|
|
13591
|
+
contextWindow: 2000000,
|
|
13592
|
+
maxTokens: 30000,
|
|
13593
|
+
},
|
|
13594
|
+
"grok-4.20-0309-reasoning": {
|
|
13595
|
+
id: "grok-4.20-0309-reasoning",
|
|
13596
|
+
name: "Grok 4.20 (Reasoning)",
|
|
13597
|
+
api: "openai-completions",
|
|
13598
|
+
provider: "xai",
|
|
13599
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13600
|
+
reasoning: true,
|
|
13601
|
+
input: ["text", "image"],
|
|
13602
|
+
cost: {
|
|
13603
|
+
input: 2,
|
|
13604
|
+
output: 6,
|
|
13605
|
+
cacheRead: 0.2,
|
|
13606
|
+
cacheWrite: 0,
|
|
13607
|
+
},
|
|
13608
|
+
contextWindow: 2000000,
|
|
13609
|
+
maxTokens: 30000,
|
|
13610
|
+
},
|
|
13340
13611
|
"grok-beta": {
|
|
13341
13612
|
id: "grok-beta",
|
|
13342
13613
|
name: "Grok Beta",
|
|
@@ -13552,748 +13823,24 @@ export const MODELS = {
|
|
|
13552
13823
|
contextWindow: 204800,
|
|
13553
13824
|
maxTokens: 131072,
|
|
13554
13825
|
},
|
|
13555
|
-
|
|
13556
|
-
|
|
13557
|
-
|
|
13558
|
-
id: "qwen3.5-plus",
|
|
13559
|
-
name: "Qwen3.5 Plus",
|
|
13560
|
-
api: "openai-completions",
|
|
13561
|
-
provider: "alibaba-coding-plan",
|
|
13562
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13563
|
-
reasoning: true,
|
|
13564
|
-
input: ["text"],
|
|
13565
|
-
cost: {
|
|
13566
|
-
input: 0,
|
|
13567
|
-
output: 0,
|
|
13568
|
-
cacheRead: 0,
|
|
13569
|
-
cacheWrite: 0,
|
|
13570
|
-
},
|
|
13571
|
-
contextWindow: 983616,
|
|
13572
|
-
maxTokens: 65536,
|
|
13573
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13574
|
-
},
|
|
13575
|
-
"qwen3-max-2026-01-23": {
|
|
13576
|
-
id: "qwen3-max-2026-01-23",
|
|
13577
|
-
name: "Qwen3 Max 2026-01-23",
|
|
13578
|
-
api: "openai-completions",
|
|
13579
|
-
provider: "alibaba-coding-plan",
|
|
13580
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13581
|
-
reasoning: true,
|
|
13582
|
-
input: ["text"],
|
|
13583
|
-
cost: {
|
|
13584
|
-
input: 0,
|
|
13585
|
-
output: 0,
|
|
13586
|
-
cacheRead: 0,
|
|
13587
|
-
cacheWrite: 0,
|
|
13588
|
-
},
|
|
13589
|
-
contextWindow: 258048,
|
|
13590
|
-
maxTokens: 32768,
|
|
13591
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13592
|
-
},
|
|
13593
|
-
"qwen3-coder-next": {
|
|
13594
|
-
id: "qwen3-coder-next",
|
|
13595
|
-
name: "Qwen3 Coder Next",
|
|
13596
|
-
api: "openai-completions",
|
|
13597
|
-
provider: "alibaba-coding-plan",
|
|
13598
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13599
|
-
reasoning: false,
|
|
13600
|
-
input: ["text"],
|
|
13601
|
-
cost: {
|
|
13602
|
-
input: 0,
|
|
13603
|
-
output: 0,
|
|
13604
|
-
cacheRead: 0,
|
|
13605
|
-
cacheWrite: 0,
|
|
13606
|
-
},
|
|
13607
|
-
contextWindow: 204800,
|
|
13608
|
-
maxTokens: 65536,
|
|
13609
|
-
compat: { supportsDeveloperRole: false },
|
|
13610
|
-
},
|
|
13611
|
-
"qwen3-coder-plus": {
|
|
13612
|
-
id: "qwen3-coder-plus",
|
|
13613
|
-
name: "Qwen3 Coder Plus",
|
|
13614
|
-
api: "openai-completions",
|
|
13615
|
-
provider: "alibaba-coding-plan",
|
|
13616
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13617
|
-
reasoning: false,
|
|
13618
|
-
input: ["text"],
|
|
13619
|
-
cost: {
|
|
13620
|
-
input: 0,
|
|
13621
|
-
output: 0,
|
|
13622
|
-
cacheRead: 0,
|
|
13623
|
-
cacheWrite: 0,
|
|
13624
|
-
},
|
|
13625
|
-
contextWindow: 997952,
|
|
13626
|
-
maxTokens: 65536,
|
|
13627
|
-
compat: { supportsDeveloperRole: false },
|
|
13628
|
-
},
|
|
13629
|
-
"MiniMax-M2.5": {
|
|
13630
|
-
id: "MiniMax-M2.5",
|
|
13631
|
-
name: "MiniMax M2.5",
|
|
13632
|
-
api: "openai-completions",
|
|
13633
|
-
provider: "alibaba-coding-plan",
|
|
13634
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13635
|
-
reasoning: true,
|
|
13636
|
-
input: ["text"],
|
|
13637
|
-
cost: {
|
|
13638
|
-
input: 0,
|
|
13639
|
-
output: 0,
|
|
13640
|
-
cacheRead: 0,
|
|
13641
|
-
cacheWrite: 0,
|
|
13642
|
-
},
|
|
13643
|
-
contextWindow: 196608,
|
|
13644
|
-
maxTokens: 65536,
|
|
13645
|
-
compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: true, maxTokensField: "max_tokens" },
|
|
13646
|
-
},
|
|
13647
|
-
"glm-5": {
|
|
13648
|
-
id: "glm-5",
|
|
13649
|
-
name: "GLM-5",
|
|
13650
|
-
api: "openai-completions",
|
|
13651
|
-
provider: "alibaba-coding-plan",
|
|
13652
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13653
|
-
reasoning: true,
|
|
13654
|
-
input: ["text"],
|
|
13655
|
-
cost: {
|
|
13656
|
-
input: 0,
|
|
13657
|
-
output: 0,
|
|
13658
|
-
cacheRead: 0,
|
|
13659
|
-
cacheWrite: 0,
|
|
13660
|
-
},
|
|
13661
|
-
contextWindow: 202752,
|
|
13662
|
-
maxTokens: 16384,
|
|
13663
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13664
|
-
},
|
|
13665
|
-
"glm-4.7": {
|
|
13666
|
-
id: "glm-4.7",
|
|
13667
|
-
name: "GLM-4.7",
|
|
13668
|
-
api: "openai-completions",
|
|
13669
|
-
provider: "alibaba-coding-plan",
|
|
13670
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13671
|
-
reasoning: true,
|
|
13672
|
-
input: ["text"],
|
|
13673
|
-
cost: {
|
|
13674
|
-
input: 0,
|
|
13675
|
-
output: 0,
|
|
13676
|
-
cacheRead: 0,
|
|
13677
|
-
cacheWrite: 0,
|
|
13678
|
-
},
|
|
13679
|
-
contextWindow: 169984,
|
|
13680
|
-
maxTokens: 16384,
|
|
13681
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13682
|
-
},
|
|
13683
|
-
"kimi-k2.5": {
|
|
13684
|
-
id: "kimi-k2.5",
|
|
13685
|
-
name: "Kimi K2.5",
|
|
13686
|
-
api: "openai-completions",
|
|
13687
|
-
provider: "alibaba-coding-plan",
|
|
13688
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13689
|
-
reasoning: true,
|
|
13690
|
-
input: ["text"],
|
|
13691
|
-
cost: {
|
|
13692
|
-
input: 0,
|
|
13693
|
-
output: 0,
|
|
13694
|
-
cacheRead: 0,
|
|
13695
|
-
cacheWrite: 0,
|
|
13696
|
-
},
|
|
13697
|
-
contextWindow: 258048,
|
|
13698
|
-
maxTokens: 32768,
|
|
13699
|
-
compat: { thinkingFormat: "zai", supportsDeveloperRole: false },
|
|
13700
|
-
},
|
|
13701
|
-
},
|
|
13702
|
-
"ollama-cloud": {
|
|
13703
|
-
"cogito-2.1:671b": {
|
|
13704
|
-
id: "cogito-2.1:671b",
|
|
13705
|
-
name: "Cogito 2.1 671B",
|
|
13706
|
-
api: "openai-completions",
|
|
13707
|
-
provider: "ollama-cloud",
|
|
13708
|
-
baseUrl: "https://ollama.com/v1",
|
|
13709
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13710
|
-
reasoning: true,
|
|
13711
|
-
input: ["text"],
|
|
13712
|
-
cost: {
|
|
13713
|
-
input: 0,
|
|
13714
|
-
output: 0,
|
|
13715
|
-
cacheRead: 0,
|
|
13716
|
-
cacheWrite: 0,
|
|
13717
|
-
},
|
|
13718
|
-
contextWindow: 163840,
|
|
13719
|
-
maxTokens: 32000,
|
|
13720
|
-
},
|
|
13721
|
-
"deepseek-v3.1:671b": {
|
|
13722
|
-
id: "deepseek-v3.1:671b",
|
|
13723
|
-
name: "DeepSeek V3.1 671B",
|
|
13724
|
-
api: "openai-completions",
|
|
13725
|
-
provider: "ollama-cloud",
|
|
13726
|
-
baseUrl: "https://ollama.com/v1",
|
|
13727
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13728
|
-
reasoning: true,
|
|
13729
|
-
input: ["text"],
|
|
13730
|
-
cost: {
|
|
13731
|
-
input: 0,
|
|
13732
|
-
output: 0,
|
|
13733
|
-
cacheRead: 0,
|
|
13734
|
-
cacheWrite: 0,
|
|
13735
|
-
},
|
|
13736
|
-
contextWindow: 163840,
|
|
13737
|
-
maxTokens: 163840,
|
|
13738
|
-
},
|
|
13739
|
-
"deepseek-v3.2": {
|
|
13740
|
-
id: "deepseek-v3.2",
|
|
13741
|
-
name: "DeepSeek V3.2",
|
|
13826
|
+
"glm-5-turbo": {
|
|
13827
|
+
id: "glm-5-turbo",
|
|
13828
|
+
name: "GLM-5-Turbo",
|
|
13742
13829
|
api: "openai-completions",
|
|
13743
|
-
provider: "
|
|
13744
|
-
baseUrl: "https://
|
|
13745
|
-
compat: { "
|
|
13830
|
+
provider: "zai",
|
|
13831
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13832
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
13746
13833
|
reasoning: true,
|
|
13747
13834
|
input: ["text"],
|
|
13748
13835
|
cost: {
|
|
13749
|
-
input:
|
|
13750
|
-
output:
|
|
13751
|
-
cacheRead: 0,
|
|
13752
|
-
cacheWrite: 0,
|
|
13753
|
-
},
|
|
13754
|
-
contextWindow: 163840,
|
|
13755
|
-
maxTokens: 65536,
|
|
13756
|
-
},
|
|
13757
|
-
"devstral-2:123b": {
|
|
13758
|
-
id: "devstral-2:123b",
|
|
13759
|
-
name: "Devstral 2 123B",
|
|
13760
|
-
api: "openai-completions",
|
|
13761
|
-
provider: "ollama-cloud",
|
|
13762
|
-
baseUrl: "https://ollama.com/v1",
|
|
13763
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13764
|
-
reasoning: false,
|
|
13765
|
-
input: ["text"],
|
|
13766
|
-
cost: {
|
|
13767
|
-
input: 0,
|
|
13768
|
-
output: 0,
|
|
13769
|
-
cacheRead: 0,
|
|
13836
|
+
input: 1.2,
|
|
13837
|
+
output: 4,
|
|
13838
|
+
cacheRead: 0.24,
|
|
13770
13839
|
cacheWrite: 0,
|
|
13771
13840
|
},
|
|
13772
|
-
contextWindow:
|
|
13773
|
-
maxTokens: 262144,
|
|
13774
|
-
},
|
|
13775
|
-
"devstral-small-2:24b": {
|
|
13776
|
-
id: "devstral-small-2:24b",
|
|
13777
|
-
name: "Devstral Small 2 24B",
|
|
13778
|
-
api: "openai-completions",
|
|
13779
|
-
provider: "ollama-cloud",
|
|
13780
|
-
baseUrl: "https://ollama.com/v1",
|
|
13781
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13782
|
-
reasoning: false,
|
|
13783
|
-
input: ["text", "image"],
|
|
13784
|
-
cost: {
|
|
13785
|
-
input: 0,
|
|
13786
|
-
output: 0,
|
|
13787
|
-
cacheRead: 0,
|
|
13788
|
-
cacheWrite: 0,
|
|
13789
|
-
},
|
|
13790
|
-
contextWindow: 262144,
|
|
13791
|
-
maxTokens: 262144,
|
|
13792
|
-
},
|
|
13793
|
-
"gemini-3-flash-preview": {
|
|
13794
|
-
id: "gemini-3-flash-preview",
|
|
13795
|
-
name: "Gemini 3 Flash Preview",
|
|
13796
|
-
api: "openai-completions",
|
|
13797
|
-
provider: "ollama-cloud",
|
|
13798
|
-
baseUrl: "https://ollama.com/v1",
|
|
13799
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13800
|
-
reasoning: true,
|
|
13801
|
-
input: ["text"],
|
|
13802
|
-
cost: {
|
|
13803
|
-
input: 0,
|
|
13804
|
-
output: 0,
|
|
13805
|
-
cacheRead: 0,
|
|
13806
|
-
cacheWrite: 0,
|
|
13807
|
-
},
|
|
13808
|
-
contextWindow: 1048576,
|
|
13809
|
-
maxTokens: 65536,
|
|
13810
|
-
},
|
|
13811
|
-
"gemma3:12b": {
|
|
13812
|
-
id: "gemma3:12b",
|
|
13813
|
-
name: "Gemma 3 12B",
|
|
13814
|
-
api: "openai-completions",
|
|
13815
|
-
provider: "ollama-cloud",
|
|
13816
|
-
baseUrl: "https://ollama.com/v1",
|
|
13817
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13818
|
-
reasoning: false,
|
|
13819
|
-
input: ["text", "image"],
|
|
13820
|
-
cost: {
|
|
13821
|
-
input: 0,
|
|
13822
|
-
output: 0,
|
|
13823
|
-
cacheRead: 0,
|
|
13824
|
-
cacheWrite: 0,
|
|
13825
|
-
},
|
|
13826
|
-
contextWindow: 131072,
|
|
13827
|
-
maxTokens: 131072,
|
|
13828
|
-
},
|
|
13829
|
-
"gemma3:27b": {
|
|
13830
|
-
id: "gemma3:27b",
|
|
13831
|
-
name: "Gemma 3 27B",
|
|
13832
|
-
api: "openai-completions",
|
|
13833
|
-
provider: "ollama-cloud",
|
|
13834
|
-
baseUrl: "https://ollama.com/v1",
|
|
13835
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13836
|
-
reasoning: false,
|
|
13837
|
-
input: ["text", "image"],
|
|
13838
|
-
cost: {
|
|
13839
|
-
input: 0,
|
|
13840
|
-
output: 0,
|
|
13841
|
-
cacheRead: 0,
|
|
13842
|
-
cacheWrite: 0,
|
|
13843
|
-
},
|
|
13844
|
-
contextWindow: 131072,
|
|
13845
|
-
maxTokens: 131072,
|
|
13846
|
-
},
|
|
13847
|
-
"gemma3:4b": {
|
|
13848
|
-
id: "gemma3:4b",
|
|
13849
|
-
name: "Gemma 3 4B",
|
|
13850
|
-
api: "openai-completions",
|
|
13851
|
-
provider: "ollama-cloud",
|
|
13852
|
-
baseUrl: "https://ollama.com/v1",
|
|
13853
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13854
|
-
reasoning: false,
|
|
13855
|
-
input: ["text", "image"],
|
|
13856
|
-
cost: {
|
|
13857
|
-
input: 0,
|
|
13858
|
-
output: 0,
|
|
13859
|
-
cacheRead: 0,
|
|
13860
|
-
cacheWrite: 0,
|
|
13861
|
-
},
|
|
13862
|
-
contextWindow: 131072,
|
|
13863
|
-
maxTokens: 131072,
|
|
13864
|
-
},
|
|
13865
|
-
"glm-4.6": {
|
|
13866
|
-
id: "glm-4.6",
|
|
13867
|
-
name: "GLM 4.6",
|
|
13868
|
-
api: "openai-completions",
|
|
13869
|
-
provider: "ollama-cloud",
|
|
13870
|
-
baseUrl: "https://ollama.com/v1",
|
|
13871
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13872
|
-
reasoning: true,
|
|
13873
|
-
input: ["text"],
|
|
13874
|
-
cost: {
|
|
13875
|
-
input: 0,
|
|
13876
|
-
output: 0,
|
|
13877
|
-
cacheRead: 0,
|
|
13878
|
-
cacheWrite: 0,
|
|
13879
|
-
},
|
|
13880
|
-
contextWindow: 202752,
|
|
13881
|
-
maxTokens: 131072,
|
|
13882
|
-
},
|
|
13883
|
-
"glm-4.7": {
|
|
13884
|
-
id: "glm-4.7",
|
|
13885
|
-
name: "GLM 4.7",
|
|
13886
|
-
api: "openai-completions",
|
|
13887
|
-
provider: "ollama-cloud",
|
|
13888
|
-
baseUrl: "https://ollama.com/v1",
|
|
13889
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13890
|
-
reasoning: true,
|
|
13891
|
-
input: ["text"],
|
|
13892
|
-
cost: {
|
|
13893
|
-
input: 0,
|
|
13894
|
-
output: 0,
|
|
13895
|
-
cacheRead: 0,
|
|
13896
|
-
cacheWrite: 0,
|
|
13897
|
-
},
|
|
13898
|
-
contextWindow: 202752,
|
|
13899
|
-
maxTokens: 131072,
|
|
13900
|
-
},
|
|
13901
|
-
"glm-5": {
|
|
13902
|
-
id: "glm-5",
|
|
13903
|
-
name: "GLM 5",
|
|
13904
|
-
api: "openai-completions",
|
|
13905
|
-
provider: "ollama-cloud",
|
|
13906
|
-
baseUrl: "https://ollama.com/v1",
|
|
13907
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13908
|
-
reasoning: true,
|
|
13909
|
-
input: ["text"],
|
|
13910
|
-
cost: {
|
|
13911
|
-
input: 0,
|
|
13912
|
-
output: 0,
|
|
13913
|
-
cacheRead: 0,
|
|
13914
|
-
cacheWrite: 0,
|
|
13915
|
-
},
|
|
13916
|
-
contextWindow: 202752,
|
|
13917
|
-
maxTokens: 131072,
|
|
13918
|
-
},
|
|
13919
|
-
"gpt-oss:120b": {
|
|
13920
|
-
id: "gpt-oss:120b",
|
|
13921
|
-
name: "GPT-OSS 120B",
|
|
13922
|
-
api: "openai-completions",
|
|
13923
|
-
provider: "ollama-cloud",
|
|
13924
|
-
baseUrl: "https://ollama.com/v1",
|
|
13925
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13926
|
-
reasoning: true,
|
|
13927
|
-
input: ["text"],
|
|
13928
|
-
cost: {
|
|
13929
|
-
input: 0,
|
|
13930
|
-
output: 0,
|
|
13931
|
-
cacheRead: 0,
|
|
13932
|
-
cacheWrite: 0,
|
|
13933
|
-
},
|
|
13934
|
-
contextWindow: 131072,
|
|
13935
|
-
maxTokens: 32768,
|
|
13936
|
-
},
|
|
13937
|
-
"gpt-oss:20b": {
|
|
13938
|
-
id: "gpt-oss:20b",
|
|
13939
|
-
name: "GPT-OSS 20B",
|
|
13940
|
-
api: "openai-completions",
|
|
13941
|
-
provider: "ollama-cloud",
|
|
13942
|
-
baseUrl: "https://ollama.com/v1",
|
|
13943
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13944
|
-
reasoning: true,
|
|
13945
|
-
input: ["text"],
|
|
13946
|
-
cost: {
|
|
13947
|
-
input: 0,
|
|
13948
|
-
output: 0,
|
|
13949
|
-
cacheRead: 0,
|
|
13950
|
-
cacheWrite: 0,
|
|
13951
|
-
},
|
|
13952
|
-
contextWindow: 131072,
|
|
13953
|
-
maxTokens: 32768,
|
|
13954
|
-
},
|
|
13955
|
-
"kimi-k2:1t": {
|
|
13956
|
-
id: "kimi-k2:1t",
|
|
13957
|
-
name: "Kimi K2 1T",
|
|
13958
|
-
api: "openai-completions",
|
|
13959
|
-
provider: "ollama-cloud",
|
|
13960
|
-
baseUrl: "https://ollama.com/v1",
|
|
13961
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13962
|
-
reasoning: false,
|
|
13963
|
-
input: ["text"],
|
|
13964
|
-
cost: {
|
|
13965
|
-
input: 0,
|
|
13966
|
-
output: 0,
|
|
13967
|
-
cacheRead: 0,
|
|
13968
|
-
cacheWrite: 0,
|
|
13969
|
-
},
|
|
13970
|
-
contextWindow: 262144,
|
|
13971
|
-
maxTokens: 262144,
|
|
13972
|
-
},
|
|
13973
|
-
"kimi-k2.5": {
|
|
13974
|
-
id: "kimi-k2.5",
|
|
13975
|
-
name: "Kimi K2.5",
|
|
13976
|
-
api: "openai-completions",
|
|
13977
|
-
provider: "ollama-cloud",
|
|
13978
|
-
baseUrl: "https://ollama.com/v1",
|
|
13979
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13980
|
-
reasoning: true,
|
|
13981
|
-
input: ["text", "image"],
|
|
13982
|
-
cost: {
|
|
13983
|
-
input: 0,
|
|
13984
|
-
output: 0,
|
|
13985
|
-
cacheRead: 0,
|
|
13986
|
-
cacheWrite: 0,
|
|
13987
|
-
},
|
|
13988
|
-
contextWindow: 262144,
|
|
13989
|
-
maxTokens: 262144,
|
|
13990
|
-
},
|
|
13991
|
-
"kimi-k2-thinking": {
|
|
13992
|
-
id: "kimi-k2-thinking",
|
|
13993
|
-
name: "Kimi K2 Thinking",
|
|
13994
|
-
api: "openai-completions",
|
|
13995
|
-
provider: "ollama-cloud",
|
|
13996
|
-
baseUrl: "https://ollama.com/v1",
|
|
13997
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13998
|
-
reasoning: true,
|
|
13999
|
-
input: ["text"],
|
|
14000
|
-
cost: {
|
|
14001
|
-
input: 0,
|
|
14002
|
-
output: 0,
|
|
14003
|
-
cacheRead: 0,
|
|
14004
|
-
cacheWrite: 0,
|
|
14005
|
-
},
|
|
14006
|
-
contextWindow: 262144,
|
|
14007
|
-
maxTokens: 262144,
|
|
14008
|
-
},
|
|
14009
|
-
"minimax-m2.1": {
|
|
14010
|
-
id: "minimax-m2.1",
|
|
14011
|
-
name: "Minimax M2.1",
|
|
14012
|
-
api: "openai-completions",
|
|
14013
|
-
provider: "ollama-cloud",
|
|
14014
|
-
baseUrl: "https://ollama.com/v1",
|
|
14015
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14016
|
-
reasoning: true,
|
|
14017
|
-
input: ["text"],
|
|
14018
|
-
cost: {
|
|
14019
|
-
input: 0,
|
|
14020
|
-
output: 0,
|
|
14021
|
-
cacheRead: 0,
|
|
14022
|
-
cacheWrite: 0,
|
|
14023
|
-
},
|
|
14024
|
-
contextWindow: 204800,
|
|
14025
|
-
maxTokens: 131072,
|
|
14026
|
-
},
|
|
14027
|
-
"minimax-m2.5": {
|
|
14028
|
-
id: "minimax-m2.5",
|
|
14029
|
-
name: "Minimax M2.5",
|
|
14030
|
-
api: "openai-completions",
|
|
14031
|
-
provider: "ollama-cloud",
|
|
14032
|
-
baseUrl: "https://ollama.com/v1",
|
|
14033
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14034
|
-
reasoning: true,
|
|
14035
|
-
input: ["text"],
|
|
14036
|
-
cost: {
|
|
14037
|
-
input: 0,
|
|
14038
|
-
output: 0,
|
|
14039
|
-
cacheRead: 0,
|
|
14040
|
-
cacheWrite: 0,
|
|
14041
|
-
},
|
|
14042
|
-
contextWindow: 204800,
|
|
14043
|
-
maxTokens: 131072,
|
|
14044
|
-
},
|
|
14045
|
-
"minimax-m2": {
|
|
14046
|
-
id: "minimax-m2",
|
|
14047
|
-
name: "Minimax M2",
|
|
14048
|
-
api: "openai-completions",
|
|
14049
|
-
provider: "ollama-cloud",
|
|
14050
|
-
baseUrl: "https://ollama.com/v1",
|
|
14051
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14052
|
-
reasoning: false,
|
|
14053
|
-
input: ["text"],
|
|
14054
|
-
cost: {
|
|
14055
|
-
input: 0,
|
|
14056
|
-
output: 0,
|
|
14057
|
-
cacheRead: 0,
|
|
14058
|
-
cacheWrite: 0,
|
|
14059
|
-
},
|
|
14060
|
-
contextWindow: 204800,
|
|
14061
|
-
maxTokens: 128000,
|
|
14062
|
-
},
|
|
14063
|
-
"ministral-3:14b": {
|
|
14064
|
-
id: "ministral-3:14b",
|
|
14065
|
-
name: "Ministral 3 14B",
|
|
14066
|
-
api: "openai-completions",
|
|
14067
|
-
provider: "ollama-cloud",
|
|
14068
|
-
baseUrl: "https://ollama.com/v1",
|
|
14069
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14070
|
-
reasoning: false,
|
|
14071
|
-
input: ["text", "image"],
|
|
14072
|
-
cost: {
|
|
14073
|
-
input: 0,
|
|
14074
|
-
output: 0,
|
|
14075
|
-
cacheRead: 0,
|
|
14076
|
-
cacheWrite: 0,
|
|
14077
|
-
},
|
|
14078
|
-
contextWindow: 262144,
|
|
14079
|
-
maxTokens: 128000,
|
|
14080
|
-
},
|
|
14081
|
-
"ministral-3:3b": {
|
|
14082
|
-
id: "ministral-3:3b",
|
|
14083
|
-
name: "Ministral 3 3B",
|
|
14084
|
-
api: "openai-completions",
|
|
14085
|
-
provider: "ollama-cloud",
|
|
14086
|
-
baseUrl: "https://ollama.com/v1",
|
|
14087
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14088
|
-
reasoning: false,
|
|
14089
|
-
input: ["text", "image"],
|
|
14090
|
-
cost: {
|
|
14091
|
-
input: 0,
|
|
14092
|
-
output: 0,
|
|
14093
|
-
cacheRead: 0,
|
|
14094
|
-
cacheWrite: 0,
|
|
14095
|
-
},
|
|
14096
|
-
contextWindow: 262144,
|
|
14097
|
-
maxTokens: 128000,
|
|
14098
|
-
},
|
|
14099
|
-
"ministral-3:8b": {
|
|
14100
|
-
id: "ministral-3:8b",
|
|
14101
|
-
name: "Ministral 3 8B",
|
|
14102
|
-
api: "openai-completions",
|
|
14103
|
-
provider: "ollama-cloud",
|
|
14104
|
-
baseUrl: "https://ollama.com/v1",
|
|
14105
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14106
|
-
reasoning: false,
|
|
14107
|
-
input: ["text", "image"],
|
|
14108
|
-
cost: {
|
|
14109
|
-
input: 0,
|
|
14110
|
-
output: 0,
|
|
14111
|
-
cacheRead: 0,
|
|
14112
|
-
cacheWrite: 0,
|
|
14113
|
-
},
|
|
14114
|
-
contextWindow: 262144,
|
|
14115
|
-
maxTokens: 128000,
|
|
14116
|
-
},
|
|
14117
|
-
"mistral-large-3:675b": {
|
|
14118
|
-
id: "mistral-large-3:675b",
|
|
14119
|
-
name: "Mistral Large 3 675B",
|
|
14120
|
-
api: "openai-completions",
|
|
14121
|
-
provider: "ollama-cloud",
|
|
14122
|
-
baseUrl: "https://ollama.com/v1",
|
|
14123
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14124
|
-
reasoning: false,
|
|
14125
|
-
input: ["text", "image"],
|
|
14126
|
-
cost: {
|
|
14127
|
-
input: 0,
|
|
14128
|
-
output: 0,
|
|
14129
|
-
cacheRead: 0,
|
|
14130
|
-
cacheWrite: 0,
|
|
14131
|
-
},
|
|
14132
|
-
contextWindow: 262144,
|
|
14133
|
-
maxTokens: 262144,
|
|
14134
|
-
},
|
|
14135
|
-
"nemotron-3-nano:30b": {
|
|
14136
|
-
id: "nemotron-3-nano:30b",
|
|
14137
|
-
name: "Nemotron 3 Nano 30B",
|
|
14138
|
-
api: "openai-completions",
|
|
14139
|
-
provider: "ollama-cloud",
|
|
14140
|
-
baseUrl: "https://ollama.com/v1",
|
|
14141
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14142
|
-
reasoning: true,
|
|
14143
|
-
input: ["text"],
|
|
14144
|
-
cost: {
|
|
14145
|
-
input: 0,
|
|
14146
|
-
output: 0,
|
|
14147
|
-
cacheRead: 0,
|
|
14148
|
-
cacheWrite: 0,
|
|
14149
|
-
},
|
|
14150
|
-
contextWindow: 1048576,
|
|
14151
|
-
maxTokens: 131072,
|
|
14152
|
-
},
|
|
14153
|
-
"nemotron-3-super": {
|
|
14154
|
-
id: "nemotron-3-super",
|
|
14155
|
-
name: "Nemotron 3 Super",
|
|
14156
|
-
api: "openai-completions",
|
|
14157
|
-
provider: "ollama-cloud",
|
|
14158
|
-
baseUrl: "https://ollama.com/v1",
|
|
14159
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14160
|
-
reasoning: true,
|
|
14161
|
-
input: ["text"],
|
|
14162
|
-
cost: {
|
|
14163
|
-
input: 0,
|
|
14164
|
-
output: 0,
|
|
14165
|
-
cacheRead: 0,
|
|
14166
|
-
cacheWrite: 0,
|
|
14167
|
-
},
|
|
14168
|
-
contextWindow: 262144,
|
|
14169
|
-
maxTokens: 65536,
|
|
14170
|
-
},
|
|
14171
|
-
"qwen3.5:397b": {
|
|
14172
|
-
id: "qwen3.5:397b",
|
|
14173
|
-
name: "Qwen 3.5 397B",
|
|
14174
|
-
api: "openai-completions",
|
|
14175
|
-
provider: "ollama-cloud",
|
|
14176
|
-
baseUrl: "https://ollama.com/v1",
|
|
14177
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14178
|
-
reasoning: true,
|
|
14179
|
-
input: ["text", "image"],
|
|
14180
|
-
cost: {
|
|
14181
|
-
input: 0,
|
|
14182
|
-
output: 0,
|
|
14183
|
-
cacheRead: 0,
|
|
14184
|
-
cacheWrite: 0,
|
|
14185
|
-
},
|
|
14186
|
-
contextWindow: 262144,
|
|
14187
|
-
maxTokens: 81920,
|
|
14188
|
-
},
|
|
14189
|
-
"qwen3-coder:480b": {
|
|
14190
|
-
id: "qwen3-coder:480b",
|
|
14191
|
-
name: "Qwen 3 Coder 480B",
|
|
14192
|
-
api: "openai-completions",
|
|
14193
|
-
provider: "ollama-cloud",
|
|
14194
|
-
baseUrl: "https://ollama.com/v1",
|
|
14195
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14196
|
-
reasoning: false,
|
|
14197
|
-
input: ["text"],
|
|
14198
|
-
cost: {
|
|
14199
|
-
input: 0,
|
|
14200
|
-
output: 0,
|
|
14201
|
-
cacheRead: 0,
|
|
14202
|
-
cacheWrite: 0,
|
|
14203
|
-
},
|
|
14204
|
-
contextWindow: 262144,
|
|
14205
|
-
maxTokens: 65536,
|
|
14206
|
-
},
|
|
14207
|
-
"qwen3-coder-next": {
|
|
14208
|
-
id: "qwen3-coder-next",
|
|
14209
|
-
name: "Qwen 3 Coder Next",
|
|
14210
|
-
api: "openai-completions",
|
|
14211
|
-
provider: "ollama-cloud",
|
|
14212
|
-
baseUrl: "https://ollama.com/v1",
|
|
14213
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14214
|
-
reasoning: false,
|
|
14215
|
-
input: ["text"],
|
|
14216
|
-
cost: {
|
|
14217
|
-
input: 0,
|
|
14218
|
-
output: 0,
|
|
14219
|
-
cacheRead: 0,
|
|
14220
|
-
cacheWrite: 0,
|
|
14221
|
-
},
|
|
14222
|
-
contextWindow: 262144,
|
|
14223
|
-
maxTokens: 65536,
|
|
14224
|
-
},
|
|
14225
|
-
"qwen3-next:80b": {
|
|
14226
|
-
id: "qwen3-next:80b",
|
|
14227
|
-
name: "Qwen 3 Next 80B",
|
|
14228
|
-
api: "openai-completions",
|
|
14229
|
-
provider: "ollama-cloud",
|
|
14230
|
-
baseUrl: "https://ollama.com/v1",
|
|
14231
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14232
|
-
reasoning: true,
|
|
14233
|
-
input: ["text"],
|
|
14234
|
-
cost: {
|
|
14235
|
-
input: 0,
|
|
14236
|
-
output: 0,
|
|
14237
|
-
cacheRead: 0,
|
|
14238
|
-
cacheWrite: 0,
|
|
14239
|
-
},
|
|
14240
|
-
contextWindow: 262144,
|
|
14241
|
-
maxTokens: 32768,
|
|
14242
|
-
},
|
|
14243
|
-
"qwen3-vl:235b-instruct": {
|
|
14244
|
-
id: "qwen3-vl:235b-instruct",
|
|
14245
|
-
name: "Qwen 3 VL 235B Instruct",
|
|
14246
|
-
api: "openai-completions",
|
|
14247
|
-
provider: "ollama-cloud",
|
|
14248
|
-
baseUrl: "https://ollama.com/v1",
|
|
14249
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14250
|
-
reasoning: false,
|
|
14251
|
-
input: ["text", "image"],
|
|
14252
|
-
cost: {
|
|
14253
|
-
input: 0,
|
|
14254
|
-
output: 0,
|
|
14255
|
-
cacheRead: 0,
|
|
14256
|
-
cacheWrite: 0,
|
|
14257
|
-
},
|
|
14258
|
-
contextWindow: 262144,
|
|
13841
|
+
contextWindow: 200000,
|
|
14259
13842
|
maxTokens: 131072,
|
|
14260
13843
|
},
|
|
14261
|
-
"qwen3-vl:235b": {
|
|
14262
|
-
id: "qwen3-vl:235b",
|
|
14263
|
-
name: "Qwen 3 VL 235B",
|
|
14264
|
-
api: "openai-completions",
|
|
14265
|
-
provider: "ollama-cloud",
|
|
14266
|
-
baseUrl: "https://ollama.com/v1",
|
|
14267
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14268
|
-
reasoning: true,
|
|
14269
|
-
input: ["text", "image"],
|
|
14270
|
-
cost: {
|
|
14271
|
-
input: 0,
|
|
14272
|
-
output: 0,
|
|
14273
|
-
cacheRead: 0,
|
|
14274
|
-
cacheWrite: 0,
|
|
14275
|
-
},
|
|
14276
|
-
contextWindow: 262144,
|
|
14277
|
-
maxTokens: 32768,
|
|
14278
|
-
},
|
|
14279
|
-
"rnj-1:8b": {
|
|
14280
|
-
id: "rnj-1:8b",
|
|
14281
|
-
name: "RNJ 1 8B",
|
|
14282
|
-
api: "openai-completions",
|
|
14283
|
-
provider: "ollama-cloud",
|
|
14284
|
-
baseUrl: "https://ollama.com/v1",
|
|
14285
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14286
|
-
reasoning: false,
|
|
14287
|
-
input: ["text"],
|
|
14288
|
-
cost: {
|
|
14289
|
-
input: 0,
|
|
14290
|
-
output: 0,
|
|
14291
|
-
cacheRead: 0,
|
|
14292
|
-
cacheWrite: 0,
|
|
14293
|
-
},
|
|
14294
|
-
contextWindow: 32768,
|
|
14295
|
-
maxTokens: 4096,
|
|
14296
|
-
},
|
|
14297
13844
|
},
|
|
14298
13845
|
};
|
|
14299
13846
|
//# sourceMappingURL=models.generated.js.map
|