gsd-pi 2.42.0-dev.97e9e30 → 2.42.0-dev.eedc83f
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +23 -0
- package/dist/cli.js +15 -1
- package/dist/resource-loader.js +39 -6
- package/dist/resources/extensions/async-jobs/async-bash-tool.js +52 -4
- package/dist/resources/extensions/gsd/auto-prompts.js +1 -1
- package/dist/resources/extensions/gsd/bootstrap/register-hooks.js +11 -5
- package/dist/resources/extensions/gsd/detection.js +19 -0
- package/dist/resources/extensions/gsd/doctor-checks.js +31 -1
- package/dist/resources/extensions/gsd/doctor-providers.js +10 -0
- package/dist/resources/extensions/gsd/forensics.js +84 -0
- package/dist/resources/extensions/gsd/git-constants.js +1 -0
- package/dist/resources/extensions/gsd/git-service.js +68 -2
- package/dist/resources/extensions/gsd/native-git-bridge.js +1 -0
- package/dist/resources/extensions/gsd/preferences-types.js +1 -0
- package/dist/resources/extensions/gsd/preferences.js +59 -8
- package/dist/resources/extensions/gsd/prompts/forensics.md +12 -5
- package/dist/resources/extensions/gsd/repo-identity.js +46 -5
- package/dist/resources/extensions/gsd/service-tier.js +13 -4
- package/dist/resources/extensions/gsd/session-lock.js +2 -2
- package/dist/resources/extensions/gsd/worktree-resolver.js +2 -2
- package/dist/resources/extensions/mcp-client/index.js +2 -1
- package/dist/resources/extensions/search-the-web/tool-search.js +3 -3
- package/dist/web/standalone/.next/BUILD_ID +1 -1
- package/dist/web/standalone/.next/app-path-routes-manifest.json +12 -12
- package/dist/web/standalone/.next/build-manifest.json +2 -2
- package/dist/web/standalone/.next/prerender-manifest.json +3 -3
- package/dist/web/standalone/.next/server/app/_global-error.html +2 -2
- package/dist/web/standalone/.next/server/app/_global-error.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_full.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_global-error/__PAGE__.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_global-error.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_head.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_index.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_global-error.segments/_tree.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.html +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_full.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_head.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_index.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_not-found/__PAGE__.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_not-found.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/_not-found.segments/_tree.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/api/git/route.js +1 -1
- package/dist/web/standalone/.next/server/app/index.html +1 -1
- package/dist/web/standalone/.next/server/app/index.rsc +1 -1
- package/dist/web/standalone/.next/server/app/index.segments/__PAGE__.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/index.segments/_full.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/index.segments/_head.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/index.segments/_index.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app/index.segments/_tree.segment.rsc +1 -1
- package/dist/web/standalone/.next/server/app-paths-manifest.json +12 -12
- package/dist/web/standalone/.next/server/chunks/229.js +2 -2
- package/dist/web/standalone/.next/server/pages/404.html +1 -1
- package/dist/web/standalone/.next/server/pages/500.html +2 -2
- package/dist/web/standalone/.next/server/server-reference-manifest.json +1 -1
- package/dist/web-mode.d.ts +2 -0
- package/dist/web-mode.js +40 -4
- package/package.json +1 -1
- package/packages/pi-agent-core/dist/agent.d.ts.map +1 -1
- package/packages/pi-agent-core/dist/agent.js +2 -0
- package/packages/pi-agent-core/dist/agent.js.map +1 -1
- package/packages/pi-agent-core/dist/types.d.ts +6 -0
- package/packages/pi-agent-core/dist/types.d.ts.map +1 -1
- package/packages/pi-agent-core/dist/types.js.map +1 -1
- package/packages/pi-agent-core/src/agent.test.ts +53 -0
- package/packages/pi-agent-core/src/agent.ts +3 -0
- package/packages/pi-agent-core/src/types.ts +6 -0
- package/packages/pi-agent-core/tsconfig.json +1 -1
- package/packages/pi-ai/dist/models.d.ts +5 -3
- package/packages/pi-ai/dist/models.d.ts.map +1 -1
- package/packages/pi-ai/dist/models.generated.d.ts +801 -1468
- package/packages/pi-ai/dist/models.generated.d.ts.map +1 -1
- package/packages/pi-ai/dist/models.generated.js +1135 -1588
- package/packages/pi-ai/dist/models.generated.js.map +1 -1
- package/packages/pi-ai/dist/models.js.map +1 -1
- package/packages/pi-ai/dist/utils/oauth/github-copilot.d.ts.map +1 -1
- package/packages/pi-ai/dist/utils/oauth/github-copilot.js +60 -2
- package/packages/pi-ai/dist/utils/oauth/github-copilot.js.map +1 -1
- package/packages/pi-ai/scripts/generate-models.ts +1543 -0
- package/packages/pi-ai/src/models.generated.ts +1140 -1593
- package/packages/pi-ai/src/models.ts +7 -4
- package/packages/pi-ai/src/utils/oauth/github-copilot.ts +74 -2
- package/packages/pi-coding-agent/dist/core/agent-session.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/agent-session.js +8 -1
- package/packages/pi-coding-agent/dist/core/agent-session.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/auth-storage.d.ts +7 -0
- package/packages/pi-coding-agent/dist/core/auth-storage.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/auth-storage.js +29 -2
- package/packages/pi-coding-agent/dist/core/auth-storage.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/auth-storage.test.js +60 -0
- package/packages/pi-coding-agent/dist/core/auth-storage.test.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/extensions/loader.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/extensions/loader.js +18 -0
- package/packages/pi-coding-agent/dist/core/extensions/loader.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/lsp/client.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/lsp/client.js +23 -0
- package/packages/pi-coding-agent/dist/core/lsp/client.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/model-registry.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/model-registry.js +2 -0
- package/packages/pi-coding-agent/dist/core/model-registry.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/package-manager.d.ts +6 -0
- package/packages/pi-coding-agent/dist/core/package-manager.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/package-manager.js +63 -11
- package/packages/pi-coding-agent/dist/core/package-manager.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/resource-loader.d.ts +9 -0
- package/packages/pi-coding-agent/dist/core/resource-loader.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/resource-loader.js +20 -6
- package/packages/pi-coding-agent/dist/core/resource-loader.js.map +1 -1
- package/packages/pi-coding-agent/dist/core/system-prompt.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/core/system-prompt.js +6 -5
- package/packages/pi-coding-agent/dist/core/system-prompt.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.js +3 -0
- package/packages/pi-coding-agent/dist/modes/interactive/components/extension-editor.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js +9 -6
- package/packages/pi-coding-agent/dist/modes/interactive/components/footer.js.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.d.ts.map +1 -1
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js +30 -10
- package/packages/pi-coding-agent/dist/modes/interactive/interactive-mode.js.map +1 -1
- package/packages/pi-coding-agent/src/core/agent-session.ts +7 -1
- package/packages/pi-coding-agent/src/core/auth-storage.test.ts +68 -0
- package/packages/pi-coding-agent/src/core/auth-storage.ts +30 -2
- package/packages/pi-coding-agent/src/core/extensions/loader.ts +18 -0
- package/packages/pi-coding-agent/src/core/lsp/client.ts +29 -0
- package/packages/pi-coding-agent/src/core/model-registry.ts +3 -0
- package/packages/pi-coding-agent/src/core/package-manager.ts +99 -58
- package/packages/pi-coding-agent/src/core/resource-loader.ts +24 -6
- package/packages/pi-coding-agent/src/core/system-prompt.ts +6 -5
- package/packages/pi-coding-agent/src/modes/interactive/components/extension-editor.ts +3 -0
- package/packages/pi-coding-agent/src/modes/interactive/components/footer.ts +10 -6
- package/packages/pi-coding-agent/src/modes/interactive/interactive-mode.ts +31 -11
- package/src/resources/extensions/async-jobs/async-bash-timeout.test.ts +122 -0
- package/src/resources/extensions/async-jobs/async-bash-tool.ts +40 -4
- package/src/resources/extensions/gsd/auto-prompts.ts +1 -1
- package/src/resources/extensions/gsd/bootstrap/register-hooks.ts +13 -5
- package/src/resources/extensions/gsd/detection.ts +19 -0
- package/src/resources/extensions/gsd/doctor-checks.ts +32 -1
- package/src/resources/extensions/gsd/doctor-providers.ts +13 -0
- package/src/resources/extensions/gsd/doctor-types.ts +1 -0
- package/src/resources/extensions/gsd/forensics.ts +92 -0
- package/src/resources/extensions/gsd/git-constants.ts +1 -0
- package/src/resources/extensions/gsd/git-service.ts +71 -2
- package/src/resources/extensions/gsd/native-git-bridge.ts +1 -0
- package/src/resources/extensions/gsd/preferences-types.ts +3 -0
- package/src/resources/extensions/gsd/preferences.ts +62 -6
- package/src/resources/extensions/gsd/prompts/forensics.md +12 -5
- package/src/resources/extensions/gsd/repo-identity.ts +48 -5
- package/src/resources/extensions/gsd/service-tier.ts +17 -4
- package/src/resources/extensions/gsd/session-lock.ts +2 -2
- package/src/resources/extensions/gsd/tests/activity-log.test.ts +31 -69
- package/src/resources/extensions/gsd/tests/forensics-dedup.test.ts +48 -0
- package/src/resources/extensions/gsd/tests/forensics-issue-routing.test.ts +43 -0
- package/src/resources/extensions/gsd/tests/git-locale.test.ts +133 -0
- package/src/resources/extensions/gsd/tests/git-service.test.ts +49 -0
- package/src/resources/extensions/gsd/tests/journal.test.ts +82 -127
- package/src/resources/extensions/gsd/tests/manifest-status.test.ts +73 -82
- package/src/resources/extensions/gsd/tests/service-tier.test.ts +30 -1
- package/src/resources/extensions/gsd/tests/symlink-numbered-variants.test.ts +151 -0
- package/src/resources/extensions/gsd/tests/verification-gate.test.ts +156 -263
- package/src/resources/extensions/gsd/tests/worktree-health-dispatch.test.ts +35 -78
- package/src/resources/extensions/gsd/tests/worktree-manager.test.ts +81 -74
- package/src/resources/extensions/gsd/worktree-resolver.ts +2 -2
- package/src/resources/extensions/mcp-client/index.ts +5 -1
- package/src/resources/extensions/search-the-web/tool-search.ts +3 -3
- /package/dist/web/standalone/.next/static/{PXrI5DoWsm7rwAVnEU2rD → JUBX5FUR73jiViQU5a-Cx}/_buildManifest.js +0 -0
- /package/dist/web/standalone/.next/static/{PXrI5DoWsm7rwAVnEU2rD → JUBX5FUR73jiViQU5a-Cx}/_ssgManifest.js +0 -0
|
@@ -87,40 +87,6 @@ export const MODELS = {
|
|
|
87
87
|
contextWindow: 300000,
|
|
88
88
|
maxTokens: 8192,
|
|
89
89
|
},
|
|
90
|
-
"amazon.titan-text-express-v1": {
|
|
91
|
-
id: "amazon.titan-text-express-v1",
|
|
92
|
-
name: "Titan Text G1 - Express",
|
|
93
|
-
api: "bedrock-converse-stream",
|
|
94
|
-
provider: "amazon-bedrock",
|
|
95
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
96
|
-
reasoning: false,
|
|
97
|
-
input: ["text"],
|
|
98
|
-
cost: {
|
|
99
|
-
input: 0.2,
|
|
100
|
-
output: 0.6,
|
|
101
|
-
cacheRead: 0,
|
|
102
|
-
cacheWrite: 0,
|
|
103
|
-
},
|
|
104
|
-
contextWindow: 128000,
|
|
105
|
-
maxTokens: 4096,
|
|
106
|
-
},
|
|
107
|
-
"amazon.titan-text-express-v1:0:8k": {
|
|
108
|
-
id: "amazon.titan-text-express-v1:0:8k",
|
|
109
|
-
name: "Titan Text G1 - Express",
|
|
110
|
-
api: "bedrock-converse-stream",
|
|
111
|
-
provider: "amazon-bedrock",
|
|
112
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
113
|
-
reasoning: false,
|
|
114
|
-
input: ["text"],
|
|
115
|
-
cost: {
|
|
116
|
-
input: 0.2,
|
|
117
|
-
output: 0.6,
|
|
118
|
-
cacheRead: 0,
|
|
119
|
-
cacheWrite: 0,
|
|
120
|
-
},
|
|
121
|
-
contextWindow: 128000,
|
|
122
|
-
maxTokens: 4096,
|
|
123
|
-
},
|
|
124
90
|
"anthropic.claude-3-5-haiku-20241022-v1:0": {
|
|
125
91
|
id: "anthropic.claude-3-5-haiku-20241022-v1:0",
|
|
126
92
|
name: "Claude Haiku 3.5",
|
|
@@ -206,40 +172,6 @@ export const MODELS = {
|
|
|
206
172
|
contextWindow: 200000,
|
|
207
173
|
maxTokens: 4096,
|
|
208
174
|
},
|
|
209
|
-
"anthropic.claude-3-opus-20240229-v1:0": {
|
|
210
|
-
id: "anthropic.claude-3-opus-20240229-v1:0",
|
|
211
|
-
name: "Claude Opus 3",
|
|
212
|
-
api: "bedrock-converse-stream",
|
|
213
|
-
provider: "amazon-bedrock",
|
|
214
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
215
|
-
reasoning: false,
|
|
216
|
-
input: ["text", "image"],
|
|
217
|
-
cost: {
|
|
218
|
-
input: 15,
|
|
219
|
-
output: 75,
|
|
220
|
-
cacheRead: 0,
|
|
221
|
-
cacheWrite: 0,
|
|
222
|
-
},
|
|
223
|
-
contextWindow: 200000,
|
|
224
|
-
maxTokens: 4096,
|
|
225
|
-
},
|
|
226
|
-
"anthropic.claude-3-sonnet-20240229-v1:0": {
|
|
227
|
-
id: "anthropic.claude-3-sonnet-20240229-v1:0",
|
|
228
|
-
name: "Claude Sonnet 3",
|
|
229
|
-
api: "bedrock-converse-stream",
|
|
230
|
-
provider: "amazon-bedrock",
|
|
231
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
232
|
-
reasoning: false,
|
|
233
|
-
input: ["text", "image"],
|
|
234
|
-
cost: {
|
|
235
|
-
input: 3,
|
|
236
|
-
output: 15,
|
|
237
|
-
cacheRead: 0,
|
|
238
|
-
cacheWrite: 0,
|
|
239
|
-
},
|
|
240
|
-
contextWindow: 200000,
|
|
241
|
-
maxTokens: 4096,
|
|
242
|
-
},
|
|
243
175
|
"anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
244
176
|
id: "anthropic.claude-haiku-4-5-20251001-v1:0",
|
|
245
177
|
name: "Claude Haiku 4.5",
|
|
@@ -322,7 +254,7 @@ export const MODELS = {
|
|
|
322
254
|
cacheRead: 0.5,
|
|
323
255
|
cacheWrite: 6.25,
|
|
324
256
|
},
|
|
325
|
-
contextWindow:
|
|
257
|
+
contextWindow: 1000000,
|
|
326
258
|
maxTokens: 128000,
|
|
327
259
|
},
|
|
328
260
|
"anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -373,43 +305,9 @@ export const MODELS = {
|
|
|
373
305
|
cacheRead: 0.3,
|
|
374
306
|
cacheWrite: 3.75,
|
|
375
307
|
},
|
|
376
|
-
contextWindow:
|
|
308
|
+
contextWindow: 1000000,
|
|
377
309
|
maxTokens: 64000,
|
|
378
310
|
},
|
|
379
|
-
"cohere.command-r-plus-v1:0": {
|
|
380
|
-
id: "cohere.command-r-plus-v1:0",
|
|
381
|
-
name: "Command R+",
|
|
382
|
-
api: "bedrock-converse-stream",
|
|
383
|
-
provider: "amazon-bedrock",
|
|
384
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
385
|
-
reasoning: false,
|
|
386
|
-
input: ["text"],
|
|
387
|
-
cost: {
|
|
388
|
-
input: 3,
|
|
389
|
-
output: 15,
|
|
390
|
-
cacheRead: 0,
|
|
391
|
-
cacheWrite: 0,
|
|
392
|
-
},
|
|
393
|
-
contextWindow: 128000,
|
|
394
|
-
maxTokens: 4096,
|
|
395
|
-
},
|
|
396
|
-
"cohere.command-r-v1:0": {
|
|
397
|
-
id: "cohere.command-r-v1:0",
|
|
398
|
-
name: "Command R",
|
|
399
|
-
api: "bedrock-converse-stream",
|
|
400
|
-
provider: "amazon-bedrock",
|
|
401
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
402
|
-
reasoning: false,
|
|
403
|
-
input: ["text"],
|
|
404
|
-
cost: {
|
|
405
|
-
input: 0.5,
|
|
406
|
-
output: 1.5,
|
|
407
|
-
cacheRead: 0,
|
|
408
|
-
cacheWrite: 0,
|
|
409
|
-
},
|
|
410
|
-
contextWindow: 128000,
|
|
411
|
-
maxTokens: 4096,
|
|
412
|
-
},
|
|
413
311
|
"deepseek.r1-v1:0": {
|
|
414
312
|
id: "deepseek.r1-v1:0",
|
|
415
313
|
name: "DeepSeek-R1",
|
|
@@ -444,8 +342,8 @@ export const MODELS = {
|
|
|
444
342
|
contextWindow: 163840,
|
|
445
343
|
maxTokens: 81920,
|
|
446
344
|
},
|
|
447
|
-
"deepseek.v3.2
|
|
448
|
-
id: "deepseek.v3.2
|
|
345
|
+
"deepseek.v3.2": {
|
|
346
|
+
id: "deepseek.v3.2",
|
|
449
347
|
name: "DeepSeek-V3.2",
|
|
450
348
|
api: "bedrock-converse-stream",
|
|
451
349
|
provider: "amazon-bedrock",
|
|
@@ -509,7 +407,7 @@ export const MODELS = {
|
|
|
509
407
|
cacheRead: 0.5,
|
|
510
408
|
cacheWrite: 6.25,
|
|
511
409
|
},
|
|
512
|
-
contextWindow:
|
|
410
|
+
contextWindow: 1000000,
|
|
513
411
|
maxTokens: 128000,
|
|
514
412
|
},
|
|
515
413
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -560,7 +458,7 @@ export const MODELS = {
|
|
|
560
458
|
cacheRead: 0.3,
|
|
561
459
|
cacheWrite: 3.75,
|
|
562
460
|
},
|
|
563
|
-
contextWindow:
|
|
461
|
+
contextWindow: 1000000,
|
|
564
462
|
maxTokens: 64000,
|
|
565
463
|
},
|
|
566
464
|
"global.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
@@ -611,7 +509,7 @@ export const MODELS = {
|
|
|
611
509
|
cacheRead: 0.5,
|
|
612
510
|
cacheWrite: 6.25,
|
|
613
511
|
},
|
|
614
|
-
contextWindow:
|
|
512
|
+
contextWindow: 1000000,
|
|
615
513
|
maxTokens: 128000,
|
|
616
514
|
},
|
|
617
515
|
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -662,7 +560,7 @@ export const MODELS = {
|
|
|
662
560
|
cacheRead: 0.3,
|
|
663
561
|
cacheWrite: 3.75,
|
|
664
562
|
},
|
|
665
|
-
contextWindow:
|
|
563
|
+
contextWindow: 1000000,
|
|
666
564
|
maxTokens: 64000,
|
|
667
565
|
},
|
|
668
566
|
"google.gemma-3-27b-it": {
|
|
@@ -699,6 +597,23 @@ export const MODELS = {
|
|
|
699
597
|
contextWindow: 128000,
|
|
700
598
|
maxTokens: 4096,
|
|
701
599
|
},
|
|
600
|
+
"meta.llama3-1-405b-instruct-v1:0": {
|
|
601
|
+
id: "meta.llama3-1-405b-instruct-v1:0",
|
|
602
|
+
name: "Llama 3.1 405B Instruct",
|
|
603
|
+
api: "bedrock-converse-stream",
|
|
604
|
+
provider: "amazon-bedrock",
|
|
605
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
606
|
+
reasoning: false,
|
|
607
|
+
input: ["text"],
|
|
608
|
+
cost: {
|
|
609
|
+
input: 2.4,
|
|
610
|
+
output: 2.4,
|
|
611
|
+
cacheRead: 0,
|
|
612
|
+
cacheWrite: 0,
|
|
613
|
+
},
|
|
614
|
+
contextWindow: 128000,
|
|
615
|
+
maxTokens: 4096,
|
|
616
|
+
},
|
|
702
617
|
"meta.llama3-1-70b-instruct-v1:0": {
|
|
703
618
|
id: "meta.llama3-1-70b-instruct-v1:0",
|
|
704
619
|
name: "Llama 3.1 70B Instruct",
|
|
@@ -886,6 +801,40 @@ export const MODELS = {
|
|
|
886
801
|
contextWindow: 204800,
|
|
887
802
|
maxTokens: 131072,
|
|
888
803
|
},
|
|
804
|
+
"mistral.devstral-2-123b": {
|
|
805
|
+
id: "mistral.devstral-2-123b",
|
|
806
|
+
name: "Devstral 2 123B",
|
|
807
|
+
api: "bedrock-converse-stream",
|
|
808
|
+
provider: "amazon-bedrock",
|
|
809
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
810
|
+
reasoning: false,
|
|
811
|
+
input: ["text"],
|
|
812
|
+
cost: {
|
|
813
|
+
input: 0.4,
|
|
814
|
+
output: 2,
|
|
815
|
+
cacheRead: 0,
|
|
816
|
+
cacheWrite: 0,
|
|
817
|
+
},
|
|
818
|
+
contextWindow: 256000,
|
|
819
|
+
maxTokens: 8192,
|
|
820
|
+
},
|
|
821
|
+
"mistral.magistral-small-2509": {
|
|
822
|
+
id: "mistral.magistral-small-2509",
|
|
823
|
+
name: "Magistral Small 1.2",
|
|
824
|
+
api: "bedrock-converse-stream",
|
|
825
|
+
provider: "amazon-bedrock",
|
|
826
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
827
|
+
reasoning: true,
|
|
828
|
+
input: ["text", "image"],
|
|
829
|
+
cost: {
|
|
830
|
+
input: 0.5,
|
|
831
|
+
output: 1.5,
|
|
832
|
+
cacheRead: 0,
|
|
833
|
+
cacheWrite: 0,
|
|
834
|
+
},
|
|
835
|
+
contextWindow: 128000,
|
|
836
|
+
maxTokens: 40000,
|
|
837
|
+
},
|
|
889
838
|
"mistral.ministral-3-14b-instruct": {
|
|
890
839
|
id: "mistral.ministral-3-14b-instruct",
|
|
891
840
|
name: "Ministral 14B 3.0",
|
|
@@ -903,6 +852,23 @@ export const MODELS = {
|
|
|
903
852
|
contextWindow: 128000,
|
|
904
853
|
maxTokens: 4096,
|
|
905
854
|
},
|
|
855
|
+
"mistral.ministral-3-3b-instruct": {
|
|
856
|
+
id: "mistral.ministral-3-3b-instruct",
|
|
857
|
+
name: "Ministral 3 3B",
|
|
858
|
+
api: "bedrock-converse-stream",
|
|
859
|
+
provider: "amazon-bedrock",
|
|
860
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
861
|
+
reasoning: false,
|
|
862
|
+
input: ["text", "image"],
|
|
863
|
+
cost: {
|
|
864
|
+
input: 0.1,
|
|
865
|
+
output: 0.1,
|
|
866
|
+
cacheRead: 0,
|
|
867
|
+
cacheWrite: 0,
|
|
868
|
+
},
|
|
869
|
+
contextWindow: 256000,
|
|
870
|
+
maxTokens: 8192,
|
|
871
|
+
},
|
|
906
872
|
"mistral.ministral-3-8b-instruct": {
|
|
907
873
|
id: "mistral.ministral-3-8b-instruct",
|
|
908
874
|
name: "Ministral 3 8B",
|
|
@@ -920,22 +886,39 @@ export const MODELS = {
|
|
|
920
886
|
contextWindow: 128000,
|
|
921
887
|
maxTokens: 4096,
|
|
922
888
|
},
|
|
923
|
-
"mistral.mistral-large-
|
|
924
|
-
id: "mistral.mistral-large-
|
|
925
|
-
name: "Mistral Large
|
|
889
|
+
"mistral.mistral-large-3-675b-instruct": {
|
|
890
|
+
id: "mistral.mistral-large-3-675b-instruct",
|
|
891
|
+
name: "Mistral Large 3",
|
|
926
892
|
api: "bedrock-converse-stream",
|
|
927
893
|
provider: "amazon-bedrock",
|
|
928
894
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
929
895
|
reasoning: false,
|
|
930
|
-
input: ["text"],
|
|
896
|
+
input: ["text", "image"],
|
|
931
897
|
cost: {
|
|
932
898
|
input: 0.5,
|
|
933
899
|
output: 1.5,
|
|
934
900
|
cacheRead: 0,
|
|
935
901
|
cacheWrite: 0,
|
|
936
902
|
},
|
|
903
|
+
contextWindow: 256000,
|
|
904
|
+
maxTokens: 8192,
|
|
905
|
+
},
|
|
906
|
+
"mistral.pixtral-large-2502-v1:0": {
|
|
907
|
+
id: "mistral.pixtral-large-2502-v1:0",
|
|
908
|
+
name: "Pixtral Large (25.02)",
|
|
909
|
+
api: "bedrock-converse-stream",
|
|
910
|
+
provider: "amazon-bedrock",
|
|
911
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
912
|
+
reasoning: false,
|
|
913
|
+
input: ["text", "image"],
|
|
914
|
+
cost: {
|
|
915
|
+
input: 2,
|
|
916
|
+
output: 6,
|
|
917
|
+
cacheRead: 0,
|
|
918
|
+
cacheWrite: 0,
|
|
919
|
+
},
|
|
937
920
|
contextWindow: 128000,
|
|
938
|
-
maxTokens:
|
|
921
|
+
maxTokens: 8192,
|
|
939
922
|
},
|
|
940
923
|
"mistral.voxtral-mini-3b-2507": {
|
|
941
924
|
id: "mistral.voxtral-mini-3b-2507",
|
|
@@ -1022,6 +1005,23 @@ export const MODELS = {
|
|
|
1022
1005
|
contextWindow: 128000,
|
|
1023
1006
|
maxTokens: 4096,
|
|
1024
1007
|
},
|
|
1008
|
+
"nvidia.nemotron-nano-3-30b": {
|
|
1009
|
+
id: "nvidia.nemotron-nano-3-30b",
|
|
1010
|
+
name: "NVIDIA Nemotron Nano 3 30B",
|
|
1011
|
+
api: "bedrock-converse-stream",
|
|
1012
|
+
provider: "amazon-bedrock",
|
|
1013
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1014
|
+
reasoning: true,
|
|
1015
|
+
input: ["text"],
|
|
1016
|
+
cost: {
|
|
1017
|
+
input: 0.06,
|
|
1018
|
+
output: 0.24,
|
|
1019
|
+
cacheRead: 0,
|
|
1020
|
+
cacheWrite: 0,
|
|
1021
|
+
},
|
|
1022
|
+
contextWindow: 128000,
|
|
1023
|
+
maxTokens: 4096,
|
|
1024
|
+
},
|
|
1025
1025
|
"nvidia.nemotron-nano-9b-v2": {
|
|
1026
1026
|
id: "nvidia.nemotron-nano-9b-v2",
|
|
1027
1027
|
name: "NVIDIA Nemotron Nano 9B v2",
|
|
@@ -1291,7 +1291,7 @@ export const MODELS = {
|
|
|
1291
1291
|
cacheRead: 0.5,
|
|
1292
1292
|
cacheWrite: 6.25,
|
|
1293
1293
|
},
|
|
1294
|
-
contextWindow:
|
|
1294
|
+
contextWindow: 1000000,
|
|
1295
1295
|
maxTokens: 128000,
|
|
1296
1296
|
},
|
|
1297
1297
|
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -1342,7 +1342,7 @@ export const MODELS = {
|
|
|
1342
1342
|
cacheRead: 0.3,
|
|
1343
1343
|
cacheWrite: 3.75,
|
|
1344
1344
|
},
|
|
1345
|
-
contextWindow:
|
|
1345
|
+
contextWindow: 1000000,
|
|
1346
1346
|
maxTokens: 64000,
|
|
1347
1347
|
},
|
|
1348
1348
|
"writer.palmyra-x4-v1:0": {
|
|
@@ -1718,23 +1718,6 @@ export const MODELS = {
|
|
|
1718
1718
|
cacheRead: 0.5,
|
|
1719
1719
|
cacheWrite: 6.25,
|
|
1720
1720
|
},
|
|
1721
|
-
contextWindow: 200000,
|
|
1722
|
-
maxTokens: 128000,
|
|
1723
|
-
},
|
|
1724
|
-
"claude-opus-4-6[1m]": {
|
|
1725
|
-
id: "claude-opus-4-6[1m]",
|
|
1726
|
-
name: "Claude Opus 4.6 (1M)",
|
|
1727
|
-
api: "anthropic-messages",
|
|
1728
|
-
provider: "anthropic",
|
|
1729
|
-
baseUrl: "https://api.anthropic.com",
|
|
1730
|
-
reasoning: true,
|
|
1731
|
-
input: ["text", "image"],
|
|
1732
|
-
cost: {
|
|
1733
|
-
input: 5,
|
|
1734
|
-
output: 25,
|
|
1735
|
-
cacheRead: 0.5,
|
|
1736
|
-
cacheWrite: 6.25,
|
|
1737
|
-
},
|
|
1738
1721
|
contextWindow: 1000000,
|
|
1739
1722
|
maxTokens: 128000,
|
|
1740
1723
|
},
|
|
@@ -1820,216 +1803,44 @@ export const MODELS = {
|
|
|
1820
1803
|
cacheRead: 0.3,
|
|
1821
1804
|
cacheWrite: 3.75,
|
|
1822
1805
|
},
|
|
1823
|
-
contextWindow:
|
|
1806
|
+
contextWindow: 1000000,
|
|
1824
1807
|
maxTokens: 64000,
|
|
1825
1808
|
},
|
|
1826
1809
|
},
|
|
1827
|
-
"
|
|
1828
|
-
"
|
|
1829
|
-
id: "
|
|
1830
|
-
name: "
|
|
1831
|
-
api: "
|
|
1832
|
-
provider: "
|
|
1833
|
-
baseUrl: "
|
|
1810
|
+
"azure-openai-responses": {
|
|
1811
|
+
"codex-mini-latest": {
|
|
1812
|
+
id: "codex-mini-latest",
|
|
1813
|
+
name: "Codex Mini",
|
|
1814
|
+
api: "azure-openai-responses",
|
|
1815
|
+
provider: "azure-openai-responses",
|
|
1816
|
+
baseUrl: "",
|
|
1834
1817
|
reasoning: true,
|
|
1835
|
-
input: ["text"
|
|
1818
|
+
input: ["text"],
|
|
1836
1819
|
cost: {
|
|
1837
|
-
input:
|
|
1838
|
-
output:
|
|
1839
|
-
cacheRead:
|
|
1840
|
-
cacheWrite:
|
|
1820
|
+
input: 1.5,
|
|
1821
|
+
output: 6,
|
|
1822
|
+
cacheRead: 0.375,
|
|
1823
|
+
cacheWrite: 0,
|
|
1841
1824
|
},
|
|
1842
1825
|
contextWindow: 200000,
|
|
1843
|
-
maxTokens:
|
|
1826
|
+
maxTokens: 100000,
|
|
1844
1827
|
},
|
|
1845
|
-
"
|
|
1846
|
-
id: "
|
|
1847
|
-
name: "
|
|
1848
|
-
api: "
|
|
1849
|
-
provider: "
|
|
1850
|
-
baseUrl: "
|
|
1851
|
-
reasoning:
|
|
1852
|
-
input: ["text"
|
|
1828
|
+
"gpt-4": {
|
|
1829
|
+
id: "gpt-4",
|
|
1830
|
+
name: "GPT-4",
|
|
1831
|
+
api: "azure-openai-responses",
|
|
1832
|
+
provider: "azure-openai-responses",
|
|
1833
|
+
baseUrl: "",
|
|
1834
|
+
reasoning: false,
|
|
1835
|
+
input: ["text"],
|
|
1853
1836
|
cost: {
|
|
1854
|
-
input:
|
|
1855
|
-
output:
|
|
1856
|
-
cacheRead:
|
|
1857
|
-
cacheWrite:
|
|
1837
|
+
input: 30,
|
|
1838
|
+
output: 60,
|
|
1839
|
+
cacheRead: 0,
|
|
1840
|
+
cacheWrite: 0,
|
|
1858
1841
|
},
|
|
1859
|
-
contextWindow:
|
|
1860
|
-
maxTokens:
|
|
1861
|
-
},
|
|
1862
|
-
"claude-sonnet-4-6": {
|
|
1863
|
-
id: "claude-sonnet-4-6",
|
|
1864
|
-
name: "Claude Sonnet 4.6 (Vertex)",
|
|
1865
|
-
api: "anthropic-vertex",
|
|
1866
|
-
provider: "anthropic-vertex",
|
|
1867
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1868
|
-
reasoning: true,
|
|
1869
|
-
input: ["text", "image"],
|
|
1870
|
-
cost: {
|
|
1871
|
-
input: 3,
|
|
1872
|
-
output: 15,
|
|
1873
|
-
cacheRead: 0.3,
|
|
1874
|
-
cacheWrite: 3.75,
|
|
1875
|
-
},
|
|
1876
|
-
contextWindow: 200000,
|
|
1877
|
-
maxTokens: 64000,
|
|
1878
|
-
},
|
|
1879
|
-
"claude-sonnet-4-6[1m]": {
|
|
1880
|
-
id: "claude-sonnet-4-6[1m]",
|
|
1881
|
-
name: "Claude Sonnet 4.6 1M (Vertex)",
|
|
1882
|
-
api: "anthropic-vertex",
|
|
1883
|
-
provider: "anthropic-vertex",
|
|
1884
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1885
|
-
reasoning: true,
|
|
1886
|
-
input: ["text", "image"],
|
|
1887
|
-
cost: {
|
|
1888
|
-
input: 3,
|
|
1889
|
-
output: 15,
|
|
1890
|
-
cacheRead: 0.3,
|
|
1891
|
-
cacheWrite: 3.75,
|
|
1892
|
-
},
|
|
1893
|
-
contextWindow: 1000000,
|
|
1894
|
-
maxTokens: 64000,
|
|
1895
|
-
},
|
|
1896
|
-
"claude-sonnet-4-5@20250929": {
|
|
1897
|
-
id: "claude-sonnet-4-5@20250929",
|
|
1898
|
-
name: "Claude Sonnet 4.5 (Vertex)",
|
|
1899
|
-
api: "anthropic-vertex",
|
|
1900
|
-
provider: "anthropic-vertex",
|
|
1901
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1902
|
-
reasoning: true,
|
|
1903
|
-
input: ["text", "image"],
|
|
1904
|
-
cost: {
|
|
1905
|
-
input: 3,
|
|
1906
|
-
output: 15,
|
|
1907
|
-
cacheRead: 0.3,
|
|
1908
|
-
cacheWrite: 3.75,
|
|
1909
|
-
},
|
|
1910
|
-
contextWindow: 200000,
|
|
1911
|
-
maxTokens: 64000,
|
|
1912
|
-
},
|
|
1913
|
-
"claude-sonnet-4@20250514": {
|
|
1914
|
-
id: "claude-sonnet-4@20250514",
|
|
1915
|
-
name: "Claude Sonnet 4 (Vertex)",
|
|
1916
|
-
api: "anthropic-vertex",
|
|
1917
|
-
provider: "anthropic-vertex",
|
|
1918
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1919
|
-
reasoning: true,
|
|
1920
|
-
input: ["text", "image"],
|
|
1921
|
-
cost: {
|
|
1922
|
-
input: 3,
|
|
1923
|
-
output: 15,
|
|
1924
|
-
cacheRead: 0.3,
|
|
1925
|
-
cacheWrite: 3.75,
|
|
1926
|
-
},
|
|
1927
|
-
contextWindow: 200000,
|
|
1928
|
-
maxTokens: 64000,
|
|
1929
|
-
},
|
|
1930
|
-
"claude-opus-4-5@20251101": {
|
|
1931
|
-
id: "claude-opus-4-5@20251101",
|
|
1932
|
-
name: "Claude Opus 4.5 (Vertex)",
|
|
1933
|
-
api: "anthropic-vertex",
|
|
1934
|
-
provider: "anthropic-vertex",
|
|
1935
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1936
|
-
reasoning: true,
|
|
1937
|
-
input: ["text", "image"],
|
|
1938
|
-
cost: {
|
|
1939
|
-
input: 15,
|
|
1940
|
-
output: 75,
|
|
1941
|
-
cacheRead: 1.5,
|
|
1942
|
-
cacheWrite: 18.75,
|
|
1943
|
-
},
|
|
1944
|
-
contextWindow: 200000,
|
|
1945
|
-
maxTokens: 32000,
|
|
1946
|
-
},
|
|
1947
|
-
"claude-opus-4-1@20250805": {
|
|
1948
|
-
id: "claude-opus-4-1@20250805",
|
|
1949
|
-
name: "Claude Opus 4.1 (Vertex)",
|
|
1950
|
-
api: "anthropic-vertex",
|
|
1951
|
-
provider: "anthropic-vertex",
|
|
1952
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1953
|
-
reasoning: true,
|
|
1954
|
-
input: ["text", "image"],
|
|
1955
|
-
cost: {
|
|
1956
|
-
input: 15,
|
|
1957
|
-
output: 75,
|
|
1958
|
-
cacheRead: 1.5,
|
|
1959
|
-
cacheWrite: 18.75,
|
|
1960
|
-
},
|
|
1961
|
-
contextWindow: 200000,
|
|
1962
|
-
maxTokens: 32000,
|
|
1963
|
-
},
|
|
1964
|
-
"claude-opus-4@20250514": {
|
|
1965
|
-
id: "claude-opus-4@20250514",
|
|
1966
|
-
name: "Claude Opus 4 (Vertex)",
|
|
1967
|
-
api: "anthropic-vertex",
|
|
1968
|
-
provider: "anthropic-vertex",
|
|
1969
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1970
|
-
reasoning: true,
|
|
1971
|
-
input: ["text", "image"],
|
|
1972
|
-
cost: {
|
|
1973
|
-
input: 15,
|
|
1974
|
-
output: 75,
|
|
1975
|
-
cacheRead: 1.5,
|
|
1976
|
-
cacheWrite: 18.75,
|
|
1977
|
-
},
|
|
1978
|
-
contextWindow: 200000,
|
|
1979
|
-
maxTokens: 32000,
|
|
1980
|
-
},
|
|
1981
|
-
"claude-haiku-4-5@20251001": {
|
|
1982
|
-
id: "claude-haiku-4-5@20251001",
|
|
1983
|
-
name: "Claude Haiku 4.5 (Vertex)",
|
|
1984
|
-
api: "anthropic-vertex",
|
|
1985
|
-
provider: "anthropic-vertex",
|
|
1986
|
-
baseUrl: "https://us-central1-aiplatform.googleapis.com",
|
|
1987
|
-
reasoning: false,
|
|
1988
|
-
input: ["text", "image"],
|
|
1989
|
-
cost: {
|
|
1990
|
-
input: 0.8,
|
|
1991
|
-
output: 4,
|
|
1992
|
-
cacheRead: 0.08,
|
|
1993
|
-
cacheWrite: 1,
|
|
1994
|
-
},
|
|
1995
|
-
contextWindow: 200000,
|
|
1996
|
-
maxTokens: 8192,
|
|
1997
|
-
},
|
|
1998
|
-
},
|
|
1999
|
-
"azure-openai-responses": {
|
|
2000
|
-
"codex-mini-latest": {
|
|
2001
|
-
id: "codex-mini-latest",
|
|
2002
|
-
name: "Codex Mini",
|
|
2003
|
-
api: "azure-openai-responses",
|
|
2004
|
-
provider: "azure-openai-responses",
|
|
2005
|
-
baseUrl: "",
|
|
2006
|
-
reasoning: true,
|
|
2007
|
-
input: ["text"],
|
|
2008
|
-
cost: {
|
|
2009
|
-
input: 1.5,
|
|
2010
|
-
output: 6,
|
|
2011
|
-
cacheRead: 0.375,
|
|
2012
|
-
cacheWrite: 0,
|
|
2013
|
-
},
|
|
2014
|
-
contextWindow: 200000,
|
|
2015
|
-
maxTokens: 100000,
|
|
2016
|
-
},
|
|
2017
|
-
"gpt-4": {
|
|
2018
|
-
id: "gpt-4",
|
|
2019
|
-
name: "GPT-4",
|
|
2020
|
-
api: "azure-openai-responses",
|
|
2021
|
-
provider: "azure-openai-responses",
|
|
2022
|
-
baseUrl: "",
|
|
2023
|
-
reasoning: false,
|
|
2024
|
-
input: ["text"],
|
|
2025
|
-
cost: {
|
|
2026
|
-
input: 30,
|
|
2027
|
-
output: 60,
|
|
2028
|
-
cacheRead: 0,
|
|
2029
|
-
cacheWrite: 0,
|
|
2030
|
-
},
|
|
2031
|
-
contextWindow: 8192,
|
|
2032
|
-
maxTokens: 8192,
|
|
1842
|
+
contextWindow: 8192,
|
|
1843
|
+
maxTokens: 8192,
|
|
2033
1844
|
},
|
|
2034
1845
|
"gpt-4-turbo": {
|
|
2035
1846
|
id: "gpt-4-turbo",
|
|
@@ -2490,6 +2301,40 @@ export const MODELS = {
|
|
|
2490
2301
|
contextWindow: 272000,
|
|
2491
2302
|
maxTokens: 128000,
|
|
2492
2303
|
},
|
|
2304
|
+
"gpt-5.4-mini": {
|
|
2305
|
+
id: "gpt-5.4-mini",
|
|
2306
|
+
name: "GPT-5.4 mini",
|
|
2307
|
+
api: "azure-openai-responses",
|
|
2308
|
+
provider: "azure-openai-responses",
|
|
2309
|
+
baseUrl: "",
|
|
2310
|
+
reasoning: true,
|
|
2311
|
+
input: ["text", "image"],
|
|
2312
|
+
cost: {
|
|
2313
|
+
input: 0.75,
|
|
2314
|
+
output: 4.5,
|
|
2315
|
+
cacheRead: 0.075,
|
|
2316
|
+
cacheWrite: 0,
|
|
2317
|
+
},
|
|
2318
|
+
contextWindow: 400000,
|
|
2319
|
+
maxTokens: 128000,
|
|
2320
|
+
},
|
|
2321
|
+
"gpt-5.4-nano": {
|
|
2322
|
+
id: "gpt-5.4-nano",
|
|
2323
|
+
name: "GPT-5.4 nano",
|
|
2324
|
+
api: "azure-openai-responses",
|
|
2325
|
+
provider: "azure-openai-responses",
|
|
2326
|
+
baseUrl: "",
|
|
2327
|
+
reasoning: true,
|
|
2328
|
+
input: ["text", "image"],
|
|
2329
|
+
cost: {
|
|
2330
|
+
input: 0.2,
|
|
2331
|
+
output: 1.25,
|
|
2332
|
+
cacheRead: 0.02,
|
|
2333
|
+
cacheWrite: 0,
|
|
2334
|
+
},
|
|
2335
|
+
contextWindow: 400000,
|
|
2336
|
+
maxTokens: 128000,
|
|
2337
|
+
},
|
|
2493
2338
|
"gpt-5.4-pro": {
|
|
2494
2339
|
id: "gpt-5.4-pro",
|
|
2495
2340
|
name: "GPT-5.4 Pro",
|
|
@@ -2730,7 +2575,7 @@ export const MODELS = {
|
|
|
2730
2575
|
cacheRead: 0,
|
|
2731
2576
|
cacheWrite: 0,
|
|
2732
2577
|
},
|
|
2733
|
-
contextWindow:
|
|
2578
|
+
contextWindow: 144000,
|
|
2734
2579
|
maxTokens: 32000,
|
|
2735
2580
|
},
|
|
2736
2581
|
"claude-opus-4.5": {
|
|
@@ -2748,7 +2593,7 @@ export const MODELS = {
|
|
|
2748
2593
|
cacheRead: 0,
|
|
2749
2594
|
cacheWrite: 0,
|
|
2750
2595
|
},
|
|
2751
|
-
contextWindow:
|
|
2596
|
+
contextWindow: 160000,
|
|
2752
2597
|
maxTokens: 32000,
|
|
2753
2598
|
},
|
|
2754
2599
|
"claude-opus-4.6": {
|
|
@@ -2766,7 +2611,7 @@ export const MODELS = {
|
|
|
2766
2611
|
cacheRead: 0,
|
|
2767
2612
|
cacheWrite: 0,
|
|
2768
2613
|
},
|
|
2769
|
-
contextWindow:
|
|
2614
|
+
contextWindow: 144000,
|
|
2770
2615
|
maxTokens: 64000,
|
|
2771
2616
|
},
|
|
2772
2617
|
"claude-sonnet-4": {
|
|
@@ -2784,7 +2629,7 @@ export const MODELS = {
|
|
|
2784
2629
|
cacheRead: 0,
|
|
2785
2630
|
cacheWrite: 0,
|
|
2786
2631
|
},
|
|
2787
|
-
contextWindow:
|
|
2632
|
+
contextWindow: 216000,
|
|
2788
2633
|
maxTokens: 16000,
|
|
2789
2634
|
},
|
|
2790
2635
|
"claude-sonnet-4.5": {
|
|
@@ -2802,7 +2647,7 @@ export const MODELS = {
|
|
|
2802
2647
|
cacheRead: 0,
|
|
2803
2648
|
cacheWrite: 0,
|
|
2804
2649
|
},
|
|
2805
|
-
contextWindow:
|
|
2650
|
+
contextWindow: 144000,
|
|
2806
2651
|
maxTokens: 32000,
|
|
2807
2652
|
},
|
|
2808
2653
|
"claude-sonnet-4.6": {
|
|
@@ -2820,7 +2665,7 @@ export const MODELS = {
|
|
|
2820
2665
|
cacheRead: 0,
|
|
2821
2666
|
cacheWrite: 0,
|
|
2822
2667
|
},
|
|
2823
|
-
contextWindow:
|
|
2668
|
+
contextWindow: 200000,
|
|
2824
2669
|
maxTokens: 32000,
|
|
2825
2670
|
},
|
|
2826
2671
|
"gemini-2.5-pro": {
|
|
@@ -2915,7 +2760,7 @@ export const MODELS = {
|
|
|
2915
2760
|
cacheRead: 0,
|
|
2916
2761
|
cacheWrite: 0,
|
|
2917
2762
|
},
|
|
2918
|
-
contextWindow:
|
|
2763
|
+
contextWindow: 128000,
|
|
2919
2764
|
maxTokens: 16384,
|
|
2920
2765
|
},
|
|
2921
2766
|
"gpt-4o": {
|
|
@@ -2934,8 +2779,8 @@ export const MODELS = {
|
|
|
2934
2779
|
cacheRead: 0,
|
|
2935
2780
|
cacheWrite: 0,
|
|
2936
2781
|
},
|
|
2937
|
-
contextWindow:
|
|
2938
|
-
maxTokens:
|
|
2782
|
+
contextWindow: 128000,
|
|
2783
|
+
maxTokens: 4096,
|
|
2939
2784
|
},
|
|
2940
2785
|
"gpt-5": {
|
|
2941
2786
|
id: "gpt-5",
|
|
@@ -2970,7 +2815,7 @@ export const MODELS = {
|
|
|
2970
2815
|
cacheRead: 0,
|
|
2971
2816
|
cacheWrite: 0,
|
|
2972
2817
|
},
|
|
2973
|
-
contextWindow:
|
|
2818
|
+
contextWindow: 264000,
|
|
2974
2819
|
maxTokens: 64000,
|
|
2975
2820
|
},
|
|
2976
2821
|
"gpt-5.1": {
|
|
@@ -2988,7 +2833,7 @@ export const MODELS = {
|
|
|
2988
2833
|
cacheRead: 0,
|
|
2989
2834
|
cacheWrite: 0,
|
|
2990
2835
|
},
|
|
2991
|
-
contextWindow:
|
|
2836
|
+
contextWindow: 264000,
|
|
2992
2837
|
maxTokens: 64000,
|
|
2993
2838
|
},
|
|
2994
2839
|
"gpt-5.1-codex": {
|
|
@@ -3006,7 +2851,7 @@ export const MODELS = {
|
|
|
3006
2851
|
cacheRead: 0,
|
|
3007
2852
|
cacheWrite: 0,
|
|
3008
2853
|
},
|
|
3009
|
-
contextWindow:
|
|
2854
|
+
contextWindow: 400000,
|
|
3010
2855
|
maxTokens: 128000,
|
|
3011
2856
|
},
|
|
3012
2857
|
"gpt-5.1-codex-max": {
|
|
@@ -3024,7 +2869,7 @@ export const MODELS = {
|
|
|
3024
2869
|
cacheRead: 0,
|
|
3025
2870
|
cacheWrite: 0,
|
|
3026
2871
|
},
|
|
3027
|
-
contextWindow:
|
|
2872
|
+
contextWindow: 400000,
|
|
3028
2873
|
maxTokens: 128000,
|
|
3029
2874
|
},
|
|
3030
2875
|
"gpt-5.1-codex-mini": {
|
|
@@ -3042,7 +2887,7 @@ export const MODELS = {
|
|
|
3042
2887
|
cacheRead: 0,
|
|
3043
2888
|
cacheWrite: 0,
|
|
3044
2889
|
},
|
|
3045
|
-
contextWindow:
|
|
2890
|
+
contextWindow: 400000,
|
|
3046
2891
|
maxTokens: 128000,
|
|
3047
2892
|
},
|
|
3048
2893
|
"gpt-5.2": {
|
|
@@ -3060,7 +2905,7 @@ export const MODELS = {
|
|
|
3060
2905
|
cacheRead: 0,
|
|
3061
2906
|
cacheWrite: 0,
|
|
3062
2907
|
},
|
|
3063
|
-
contextWindow:
|
|
2908
|
+
contextWindow: 264000,
|
|
3064
2909
|
maxTokens: 64000,
|
|
3065
2910
|
},
|
|
3066
2911
|
"gpt-5.2-codex": {
|
|
@@ -3078,7 +2923,7 @@ export const MODELS = {
|
|
|
3078
2923
|
cacheRead: 0,
|
|
3079
2924
|
cacheWrite: 0,
|
|
3080
2925
|
},
|
|
3081
|
-
contextWindow:
|
|
2926
|
+
contextWindow: 400000,
|
|
3082
2927
|
maxTokens: 128000,
|
|
3083
2928
|
},
|
|
3084
2929
|
"gpt-5.3-codex": {
|
|
@@ -3117,6 +2962,24 @@ export const MODELS = {
|
|
|
3117
2962
|
contextWindow: 400000,
|
|
3118
2963
|
maxTokens: 128000,
|
|
3119
2964
|
},
|
|
2965
|
+
"gpt-5.4-mini": {
|
|
2966
|
+
id: "gpt-5.4-mini",
|
|
2967
|
+
name: "GPT-5.4 mini",
|
|
2968
|
+
api: "openai-responses",
|
|
2969
|
+
provider: "github-copilot",
|
|
2970
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
2971
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
2972
|
+
reasoning: true,
|
|
2973
|
+
input: ["text", "image"],
|
|
2974
|
+
cost: {
|
|
2975
|
+
input: 0,
|
|
2976
|
+
output: 0,
|
|
2977
|
+
cacheRead: 0,
|
|
2978
|
+
cacheWrite: 0,
|
|
2979
|
+
},
|
|
2980
|
+
contextWindow: 400000,
|
|
2981
|
+
maxTokens: 128000,
|
|
2982
|
+
},
|
|
3120
2983
|
"grok-code-fast-1": {
|
|
3121
2984
|
id: "grok-code-fast-1",
|
|
3122
2985
|
name: "Grok Code Fast 1",
|
|
@@ -3436,10 +3299,10 @@ export const MODELS = {
|
|
|
3436
3299
|
reasoning: true,
|
|
3437
3300
|
input: ["text", "image"],
|
|
3438
3301
|
cost: {
|
|
3439
|
-
input: 0,
|
|
3440
|
-
output:
|
|
3441
|
-
cacheRead: 0,
|
|
3442
|
-
cacheWrite:
|
|
3302
|
+
input: 0.25,
|
|
3303
|
+
output: 1.5,
|
|
3304
|
+
cacheRead: 0.025,
|
|
3305
|
+
cacheWrite: 1,
|
|
3443
3306
|
},
|
|
3444
3307
|
contextWindow: 1048576,
|
|
3445
3308
|
maxTokens: 65536,
|
|
@@ -4700,6 +4563,40 @@ export const MODELS = {
|
|
|
4700
4563
|
contextWindow: 204800,
|
|
4701
4564
|
maxTokens: 131072,
|
|
4702
4565
|
},
|
|
4566
|
+
"MiniMax-M2.7": {
|
|
4567
|
+
id: "MiniMax-M2.7",
|
|
4568
|
+
name: "MiniMax-M2.7",
|
|
4569
|
+
api: "anthropic-messages",
|
|
4570
|
+
provider: "minimax",
|
|
4571
|
+
baseUrl: "https://api.minimax.io/anthropic",
|
|
4572
|
+
reasoning: true,
|
|
4573
|
+
input: ["text"],
|
|
4574
|
+
cost: {
|
|
4575
|
+
input: 0.3,
|
|
4576
|
+
output: 1.2,
|
|
4577
|
+
cacheRead: 0.06,
|
|
4578
|
+
cacheWrite: 0.375,
|
|
4579
|
+
},
|
|
4580
|
+
contextWindow: 204800,
|
|
4581
|
+
maxTokens: 131072,
|
|
4582
|
+
},
|
|
4583
|
+
"MiniMax-M2.7-highspeed": {
|
|
4584
|
+
id: "MiniMax-M2.7-highspeed",
|
|
4585
|
+
name: "MiniMax-M2.7-highspeed",
|
|
4586
|
+
api: "anthropic-messages",
|
|
4587
|
+
provider: "minimax",
|
|
4588
|
+
baseUrl: "https://api.minimax.io/anthropic",
|
|
4589
|
+
reasoning: true,
|
|
4590
|
+
input: ["text"],
|
|
4591
|
+
cost: {
|
|
4592
|
+
input: 0.6,
|
|
4593
|
+
output: 2.4,
|
|
4594
|
+
cacheRead: 0.06,
|
|
4595
|
+
cacheWrite: 0.375,
|
|
4596
|
+
},
|
|
4597
|
+
contextWindow: 204800,
|
|
4598
|
+
maxTokens: 131072,
|
|
4599
|
+
},
|
|
4703
4600
|
},
|
|
4704
4601
|
"minimax-cn": {
|
|
4705
4602
|
"MiniMax-M2": {
|
|
@@ -4770,11 +4667,45 @@ export const MODELS = {
|
|
|
4770
4667
|
contextWindow: 204800,
|
|
4771
4668
|
maxTokens: 131072,
|
|
4772
4669
|
},
|
|
4670
|
+
"MiniMax-M2.7": {
|
|
4671
|
+
id: "MiniMax-M2.7",
|
|
4672
|
+
name: "MiniMax-M2.7",
|
|
4673
|
+
api: "anthropic-messages",
|
|
4674
|
+
provider: "minimax-cn",
|
|
4675
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4676
|
+
reasoning: true,
|
|
4677
|
+
input: ["text"],
|
|
4678
|
+
cost: {
|
|
4679
|
+
input: 0.3,
|
|
4680
|
+
output: 1.2,
|
|
4681
|
+
cacheRead: 0.06,
|
|
4682
|
+
cacheWrite: 0.375,
|
|
4683
|
+
},
|
|
4684
|
+
contextWindow: 204800,
|
|
4685
|
+
maxTokens: 131072,
|
|
4686
|
+
},
|
|
4687
|
+
"MiniMax-M2.7-highspeed": {
|
|
4688
|
+
id: "MiniMax-M2.7-highspeed",
|
|
4689
|
+
name: "MiniMax-M2.7-highspeed",
|
|
4690
|
+
api: "anthropic-messages",
|
|
4691
|
+
provider: "minimax-cn",
|
|
4692
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4693
|
+
reasoning: true,
|
|
4694
|
+
input: ["text"],
|
|
4695
|
+
cost: {
|
|
4696
|
+
input: 0.6,
|
|
4697
|
+
output: 2.4,
|
|
4698
|
+
cacheRead: 0.06,
|
|
4699
|
+
cacheWrite: 0.375,
|
|
4700
|
+
},
|
|
4701
|
+
contextWindow: 204800,
|
|
4702
|
+
maxTokens: 131072,
|
|
4703
|
+
},
|
|
4773
4704
|
},
|
|
4774
4705
|
"mistral": {
|
|
4775
4706
|
"codestral-latest": {
|
|
4776
4707
|
id: "codestral-latest",
|
|
4777
|
-
name: "Codestral",
|
|
4708
|
+
name: "Codestral (latest)",
|
|
4778
4709
|
api: "mistral-conversations",
|
|
4779
4710
|
provider: "mistral",
|
|
4780
4711
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4825,7 +4756,7 @@ export const MODELS = {
|
|
|
4825
4756
|
},
|
|
4826
4757
|
"devstral-medium-latest": {
|
|
4827
4758
|
id: "devstral-medium-latest",
|
|
4828
|
-
name: "Devstral 2",
|
|
4759
|
+
name: "Devstral 2 (latest)",
|
|
4829
4760
|
api: "mistral-conversations",
|
|
4830
4761
|
provider: "mistral",
|
|
4831
4762
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4893,7 +4824,7 @@ export const MODELS = {
|
|
|
4893
4824
|
},
|
|
4894
4825
|
"magistral-medium-latest": {
|
|
4895
4826
|
id: "magistral-medium-latest",
|
|
4896
|
-
name: "Magistral Medium",
|
|
4827
|
+
name: "Magistral Medium (latest)",
|
|
4897
4828
|
api: "mistral-conversations",
|
|
4898
4829
|
provider: "mistral",
|
|
4899
4830
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4927,7 +4858,7 @@ export const MODELS = {
|
|
|
4927
4858
|
},
|
|
4928
4859
|
"ministral-3b-latest": {
|
|
4929
4860
|
id: "ministral-3b-latest",
|
|
4930
|
-
name: "Ministral 3B",
|
|
4861
|
+
name: "Ministral 3B (latest)",
|
|
4931
4862
|
api: "mistral-conversations",
|
|
4932
4863
|
provider: "mistral",
|
|
4933
4864
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4944,7 +4875,7 @@ export const MODELS = {
|
|
|
4944
4875
|
},
|
|
4945
4876
|
"ministral-8b-latest": {
|
|
4946
4877
|
id: "ministral-8b-latest",
|
|
4947
|
-
name: "Ministral 8B",
|
|
4878
|
+
name: "Ministral 8B (latest)",
|
|
4948
4879
|
api: "mistral-conversations",
|
|
4949
4880
|
provider: "mistral",
|
|
4950
4881
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -4995,7 +4926,7 @@ export const MODELS = {
|
|
|
4995
4926
|
},
|
|
4996
4927
|
"mistral-large-latest": {
|
|
4997
4928
|
id: "mistral-large-latest",
|
|
4998
|
-
name: "Mistral Large",
|
|
4929
|
+
name: "Mistral Large (latest)",
|
|
4999
4930
|
api: "mistral-conversations",
|
|
5000
4931
|
provider: "mistral",
|
|
5001
4932
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5046,7 +4977,7 @@ export const MODELS = {
|
|
|
5046
4977
|
},
|
|
5047
4978
|
"mistral-medium-latest": {
|
|
5048
4979
|
id: "mistral-medium-latest",
|
|
5049
|
-
name: "Mistral Medium",
|
|
4980
|
+
name: "Mistral Medium (latest)",
|
|
5050
4981
|
api: "mistral-conversations",
|
|
5051
4982
|
provider: "mistral",
|
|
5052
4983
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5097,7 +5028,7 @@ export const MODELS = {
|
|
|
5097
5028
|
},
|
|
5098
5029
|
"mistral-small-latest": {
|
|
5099
5030
|
id: "mistral-small-latest",
|
|
5100
|
-
name: "Mistral Small",
|
|
5031
|
+
name: "Mistral Small (latest)",
|
|
5101
5032
|
api: "mistral-conversations",
|
|
5102
5033
|
provider: "mistral",
|
|
5103
5034
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5182,7 +5113,7 @@ export const MODELS = {
|
|
|
5182
5113
|
},
|
|
5183
5114
|
"pixtral-large-latest": {
|
|
5184
5115
|
id: "pixtral-large-latest",
|
|
5185
|
-
name: "Pixtral Large",
|
|
5116
|
+
name: "Pixtral Large (latest)",
|
|
5186
5117
|
api: "mistral-conversations",
|
|
5187
5118
|
provider: "mistral",
|
|
5188
5119
|
baseUrl: "https://api.mistral.ai",
|
|
@@ -5692,6 +5623,40 @@ export const MODELS = {
|
|
|
5692
5623
|
contextWindow: 272000,
|
|
5693
5624
|
maxTokens: 128000,
|
|
5694
5625
|
},
|
|
5626
|
+
"gpt-5.4-mini": {
|
|
5627
|
+
id: "gpt-5.4-mini",
|
|
5628
|
+
name: "GPT-5.4 mini",
|
|
5629
|
+
api: "openai-responses",
|
|
5630
|
+
provider: "openai",
|
|
5631
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5632
|
+
reasoning: true,
|
|
5633
|
+
input: ["text", "image"],
|
|
5634
|
+
cost: {
|
|
5635
|
+
input: 0.75,
|
|
5636
|
+
output: 4.5,
|
|
5637
|
+
cacheRead: 0.075,
|
|
5638
|
+
cacheWrite: 0,
|
|
5639
|
+
},
|
|
5640
|
+
contextWindow: 400000,
|
|
5641
|
+
maxTokens: 128000,
|
|
5642
|
+
},
|
|
5643
|
+
"gpt-5.4-nano": {
|
|
5644
|
+
id: "gpt-5.4-nano",
|
|
5645
|
+
name: "GPT-5.4 nano",
|
|
5646
|
+
api: "openai-responses",
|
|
5647
|
+
provider: "openai",
|
|
5648
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5649
|
+
reasoning: true,
|
|
5650
|
+
input: ["text", "image"],
|
|
5651
|
+
cost: {
|
|
5652
|
+
input: 0.2,
|
|
5653
|
+
output: 1.25,
|
|
5654
|
+
cacheRead: 0.02,
|
|
5655
|
+
cacheWrite: 0,
|
|
5656
|
+
},
|
|
5657
|
+
contextWindow: 400000,
|
|
5658
|
+
maxTokens: 128000,
|
|
5659
|
+
},
|
|
5695
5660
|
"gpt-5.4-pro": {
|
|
5696
5661
|
id: "gpt-5.4-pro",
|
|
5697
5662
|
name: "GPT-5.4 Pro",
|
|
@@ -6084,7 +6049,7 @@ export const MODELS = {
|
|
|
6084
6049
|
cacheRead: 0.5,
|
|
6085
6050
|
cacheWrite: 6.25,
|
|
6086
6051
|
},
|
|
6087
|
-
contextWindow:
|
|
6052
|
+
contextWindow: 1000000,
|
|
6088
6053
|
maxTokens: 128000,
|
|
6089
6054
|
},
|
|
6090
6055
|
"claude-sonnet-4": {
|
|
@@ -6155,23 +6120,6 @@ export const MODELS = {
|
|
|
6155
6120
|
contextWindow: 1048576,
|
|
6156
6121
|
maxTokens: 65536,
|
|
6157
6122
|
},
|
|
6158
|
-
"gemini-3-pro": {
|
|
6159
|
-
id: "gemini-3-pro",
|
|
6160
|
-
name: "Gemini 3 Pro",
|
|
6161
|
-
api: "google-generative-ai",
|
|
6162
|
-
provider: "opencode",
|
|
6163
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6164
|
-
reasoning: true,
|
|
6165
|
-
input: ["text", "image"],
|
|
6166
|
-
cost: {
|
|
6167
|
-
input: 2,
|
|
6168
|
-
output: 12,
|
|
6169
|
-
cacheRead: 0.2,
|
|
6170
|
-
cacheWrite: 0,
|
|
6171
|
-
},
|
|
6172
|
-
contextWindow: 1048576,
|
|
6173
|
-
maxTokens: 65536,
|
|
6174
|
-
},
|
|
6175
6123
|
"gemini-3.1-pro": {
|
|
6176
6124
|
id: "gemini-3.1-pro",
|
|
6177
6125
|
name: "Gemini 3.1 Pro Preview",
|
|
@@ -6189,40 +6137,6 @@ export const MODELS = {
|
|
|
6189
6137
|
contextWindow: 1048576,
|
|
6190
6138
|
maxTokens: 65536,
|
|
6191
6139
|
},
|
|
6192
|
-
"glm-4.6": {
|
|
6193
|
-
id: "glm-4.6",
|
|
6194
|
-
name: "GLM-4.6",
|
|
6195
|
-
api: "openai-completions",
|
|
6196
|
-
provider: "opencode",
|
|
6197
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6198
|
-
reasoning: true,
|
|
6199
|
-
input: ["text"],
|
|
6200
|
-
cost: {
|
|
6201
|
-
input: 0.6,
|
|
6202
|
-
output: 2.2,
|
|
6203
|
-
cacheRead: 0.1,
|
|
6204
|
-
cacheWrite: 0,
|
|
6205
|
-
},
|
|
6206
|
-
contextWindow: 204800,
|
|
6207
|
-
maxTokens: 131072,
|
|
6208
|
-
},
|
|
6209
|
-
"glm-4.7": {
|
|
6210
|
-
id: "glm-4.7",
|
|
6211
|
-
name: "GLM-4.7",
|
|
6212
|
-
api: "openai-completions",
|
|
6213
|
-
provider: "opencode",
|
|
6214
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6215
|
-
reasoning: true,
|
|
6216
|
-
input: ["text"],
|
|
6217
|
-
cost: {
|
|
6218
|
-
input: 0.6,
|
|
6219
|
-
output: 2.2,
|
|
6220
|
-
cacheRead: 0.1,
|
|
6221
|
-
cacheWrite: 0,
|
|
6222
|
-
},
|
|
6223
|
-
contextWindow: 204800,
|
|
6224
|
-
maxTokens: 131072,
|
|
6225
|
-
},
|
|
6226
6140
|
"glm-5": {
|
|
6227
6141
|
id: "glm-5",
|
|
6228
6142
|
name: "GLM-5",
|
|
@@ -6427,6 +6341,40 @@ export const MODELS = {
|
|
|
6427
6341
|
contextWindow: 272000,
|
|
6428
6342
|
maxTokens: 128000,
|
|
6429
6343
|
},
|
|
6344
|
+
"gpt-5.4-mini": {
|
|
6345
|
+
id: "gpt-5.4-mini",
|
|
6346
|
+
name: "GPT-5.4 Mini",
|
|
6347
|
+
api: "openai-responses",
|
|
6348
|
+
provider: "opencode",
|
|
6349
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6350
|
+
reasoning: true,
|
|
6351
|
+
input: ["text", "image"],
|
|
6352
|
+
cost: {
|
|
6353
|
+
input: 0.75,
|
|
6354
|
+
output: 4.5,
|
|
6355
|
+
cacheRead: 0.075,
|
|
6356
|
+
cacheWrite: 0,
|
|
6357
|
+
},
|
|
6358
|
+
contextWindow: 400000,
|
|
6359
|
+
maxTokens: 128000,
|
|
6360
|
+
},
|
|
6361
|
+
"gpt-5.4-nano": {
|
|
6362
|
+
id: "gpt-5.4-nano",
|
|
6363
|
+
name: "GPT-5.4 Nano",
|
|
6364
|
+
api: "openai-responses",
|
|
6365
|
+
provider: "opencode",
|
|
6366
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6367
|
+
reasoning: true,
|
|
6368
|
+
input: ["text", "image"],
|
|
6369
|
+
cost: {
|
|
6370
|
+
input: 0.2,
|
|
6371
|
+
output: 1.25,
|
|
6372
|
+
cacheRead: 0.02,
|
|
6373
|
+
cacheWrite: 0,
|
|
6374
|
+
},
|
|
6375
|
+
contextWindow: 400000,
|
|
6376
|
+
maxTokens: 128000,
|
|
6377
|
+
},
|
|
6430
6378
|
"gpt-5.4-pro": {
|
|
6431
6379
|
id: "gpt-5.4-pro",
|
|
6432
6380
|
name: "GPT-5.4 Pro",
|
|
@@ -6461,22 +6409,39 @@ export const MODELS = {
|
|
|
6461
6409
|
contextWindow: 262144,
|
|
6462
6410
|
maxTokens: 65536,
|
|
6463
6411
|
},
|
|
6464
|
-
"
|
|
6465
|
-
id: "
|
|
6466
|
-
name: "
|
|
6412
|
+
"mimo-v2-omni-free": {
|
|
6413
|
+
id: "mimo-v2-omni-free",
|
|
6414
|
+
name: "MiMo V2 Omni Free",
|
|
6415
|
+
api: "openai-completions",
|
|
6416
|
+
provider: "opencode",
|
|
6417
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6418
|
+
reasoning: true,
|
|
6419
|
+
input: ["text", "image"],
|
|
6420
|
+
cost: {
|
|
6421
|
+
input: 0,
|
|
6422
|
+
output: 0,
|
|
6423
|
+
cacheRead: 0,
|
|
6424
|
+
cacheWrite: 0,
|
|
6425
|
+
},
|
|
6426
|
+
contextWindow: 262144,
|
|
6427
|
+
maxTokens: 64000,
|
|
6428
|
+
},
|
|
6429
|
+
"mimo-v2-pro-free": {
|
|
6430
|
+
id: "mimo-v2-pro-free",
|
|
6431
|
+
name: "MiMo V2 Pro Free",
|
|
6467
6432
|
api: "openai-completions",
|
|
6468
6433
|
provider: "opencode",
|
|
6469
6434
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
6470
6435
|
reasoning: true,
|
|
6471
6436
|
input: ["text"],
|
|
6472
6437
|
cost: {
|
|
6473
|
-
input: 0
|
|
6474
|
-
output:
|
|
6475
|
-
cacheRead: 0
|
|
6438
|
+
input: 0,
|
|
6439
|
+
output: 0,
|
|
6440
|
+
cacheRead: 0,
|
|
6476
6441
|
cacheWrite: 0,
|
|
6477
6442
|
},
|
|
6478
|
-
contextWindow:
|
|
6479
|
-
maxTokens:
|
|
6443
|
+
contextWindow: 1048576,
|
|
6444
|
+
maxTokens: 64000,
|
|
6480
6445
|
},
|
|
6481
6446
|
"minimax-m2.5": {
|
|
6482
6447
|
id: "minimax-m2.5",
|
|
@@ -6512,6 +6477,23 @@ export const MODELS = {
|
|
|
6512
6477
|
contextWindow: 204800,
|
|
6513
6478
|
maxTokens: 131072,
|
|
6514
6479
|
},
|
|
6480
|
+
"nemotron-3-super-free": {
|
|
6481
|
+
id: "nemotron-3-super-free",
|
|
6482
|
+
name: "Nemotron 3 Super Free",
|
|
6483
|
+
api: "openai-completions",
|
|
6484
|
+
provider: "opencode",
|
|
6485
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6486
|
+
reasoning: true,
|
|
6487
|
+
input: ["text"],
|
|
6488
|
+
cost: {
|
|
6489
|
+
input: 0,
|
|
6490
|
+
output: 0,
|
|
6491
|
+
cacheRead: 0,
|
|
6492
|
+
cacheWrite: 0,
|
|
6493
|
+
},
|
|
6494
|
+
contextWindow: 1000000,
|
|
6495
|
+
maxTokens: 128000,
|
|
6496
|
+
},
|
|
6515
6497
|
},
|
|
6516
6498
|
"opencode-go": {
|
|
6517
6499
|
"glm-5": {
|
|
@@ -6565,6 +6547,23 @@ export const MODELS = {
|
|
|
6565
6547
|
contextWindow: 204800,
|
|
6566
6548
|
maxTokens: 131072,
|
|
6567
6549
|
},
|
|
6550
|
+
"minimax-m2.7": {
|
|
6551
|
+
id: "minimax-m2.7",
|
|
6552
|
+
name: "MiniMax M2.7",
|
|
6553
|
+
api: "anthropic-messages",
|
|
6554
|
+
provider: "opencode-go",
|
|
6555
|
+
baseUrl: "https://opencode.ai/zen/go",
|
|
6556
|
+
reasoning: true,
|
|
6557
|
+
input: ["text"],
|
|
6558
|
+
cost: {
|
|
6559
|
+
input: 0.3,
|
|
6560
|
+
output: 1.2,
|
|
6561
|
+
cacheRead: 0.06,
|
|
6562
|
+
cacheWrite: 0,
|
|
6563
|
+
},
|
|
6564
|
+
contextWindow: 204800,
|
|
6565
|
+
maxTokens: 131072,
|
|
6566
|
+
},
|
|
6568
6567
|
},
|
|
6569
6568
|
"openrouter": {
|
|
6570
6569
|
"ai21/jamba-large-1.7": {
|
|
@@ -7077,6 +7076,23 @@ export const MODELS = {
|
|
|
7077
7076
|
contextWindow: 262144,
|
|
7078
7077
|
maxTokens: 32768,
|
|
7079
7078
|
},
|
|
7079
|
+
"bytedance-seed/seed-2.0-lite": {
|
|
7080
|
+
id: "bytedance-seed/seed-2.0-lite",
|
|
7081
|
+
name: "ByteDance Seed: Seed-2.0-Lite",
|
|
7082
|
+
api: "openai-completions",
|
|
7083
|
+
provider: "openrouter",
|
|
7084
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7085
|
+
reasoning: true,
|
|
7086
|
+
input: ["text", "image"],
|
|
7087
|
+
cost: {
|
|
7088
|
+
input: 0.25,
|
|
7089
|
+
output: 2,
|
|
7090
|
+
cacheRead: 0,
|
|
7091
|
+
cacheWrite: 0,
|
|
7092
|
+
},
|
|
7093
|
+
contextWindow: 262144,
|
|
7094
|
+
maxTokens: 131072,
|
|
7095
|
+
},
|
|
7080
7096
|
"bytedance-seed/seed-2.0-mini": {
|
|
7081
7097
|
id: "bytedance-seed/seed-2.0-mini",
|
|
7082
7098
|
name: "ByteDance Seed: Seed-2.0-Mini",
|
|
@@ -7156,11 +7172,11 @@ export const MODELS = {
|
|
|
7156
7172
|
cost: {
|
|
7157
7173
|
input: 0.19999999999999998,
|
|
7158
7174
|
output: 0.77,
|
|
7159
|
-
cacheRead: 0.
|
|
7175
|
+
cacheRead: 0.135,
|
|
7160
7176
|
cacheWrite: 0,
|
|
7161
7177
|
},
|
|
7162
7178
|
contextWindow: 163840,
|
|
7163
|
-
maxTokens:
|
|
7179
|
+
maxTokens: 4096,
|
|
7164
7180
|
},
|
|
7165
7181
|
"deepseek/deepseek-chat-v3.1": {
|
|
7166
7182
|
id: "deepseek/deepseek-chat-v3.1",
|
|
@@ -7230,23 +7246,6 @@ export const MODELS = {
|
|
|
7230
7246
|
contextWindow: 163840,
|
|
7231
7247
|
maxTokens: 4096,
|
|
7232
7248
|
},
|
|
7233
|
-
"deepseek/deepseek-v3.1-terminus:exacto": {
|
|
7234
|
-
id: "deepseek/deepseek-v3.1-terminus:exacto",
|
|
7235
|
-
name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
|
|
7236
|
-
api: "openai-completions",
|
|
7237
|
-
provider: "openrouter",
|
|
7238
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7239
|
-
reasoning: true,
|
|
7240
|
-
input: ["text"],
|
|
7241
|
-
cost: {
|
|
7242
|
-
input: 0.21,
|
|
7243
|
-
output: 0.7899999999999999,
|
|
7244
|
-
cacheRead: 0.16799999999999998,
|
|
7245
|
-
cacheWrite: 0,
|
|
7246
|
-
},
|
|
7247
|
-
contextWindow: 163840,
|
|
7248
|
-
maxTokens: 4096,
|
|
7249
|
-
},
|
|
7250
7249
|
"deepseek/deepseek-v3.2": {
|
|
7251
7250
|
id: "deepseek/deepseek-v3.2",
|
|
7252
7251
|
name: "DeepSeek: DeepSeek V3.2",
|
|
@@ -7256,13 +7255,13 @@ export const MODELS = {
|
|
|
7256
7255
|
reasoning: true,
|
|
7257
7256
|
input: ["text"],
|
|
7258
7257
|
cost: {
|
|
7259
|
-
input: 0.
|
|
7260
|
-
output: 0.
|
|
7261
|
-
cacheRead: 0,
|
|
7258
|
+
input: 0.26,
|
|
7259
|
+
output: 0.38,
|
|
7260
|
+
cacheRead: 0.13,
|
|
7262
7261
|
cacheWrite: 0,
|
|
7263
7262
|
},
|
|
7264
7263
|
contextWindow: 163840,
|
|
7265
|
-
maxTokens:
|
|
7264
|
+
maxTokens: 4096,
|
|
7266
7265
|
},
|
|
7267
7266
|
"deepseek/deepseek-v3.2-exp": {
|
|
7268
7267
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7519,40 +7518,6 @@ export const MODELS = {
|
|
|
7519
7518
|
contextWindow: 1048576,
|
|
7520
7519
|
maxTokens: 65536,
|
|
7521
7520
|
},
|
|
7522
|
-
"google/gemma-3-27b-it": {
|
|
7523
|
-
id: "google/gemma-3-27b-it",
|
|
7524
|
-
name: "Google: Gemma 3 27B",
|
|
7525
|
-
api: "openai-completions",
|
|
7526
|
-
provider: "openrouter",
|
|
7527
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7528
|
-
reasoning: false,
|
|
7529
|
-
input: ["text", "image"],
|
|
7530
|
-
cost: {
|
|
7531
|
-
input: 0.04,
|
|
7532
|
-
output: 0.15,
|
|
7533
|
-
cacheRead: 0.02,
|
|
7534
|
-
cacheWrite: 0,
|
|
7535
|
-
},
|
|
7536
|
-
contextWindow: 128000,
|
|
7537
|
-
maxTokens: 65536,
|
|
7538
|
-
},
|
|
7539
|
-
"google/gemma-3-27b-it:free": {
|
|
7540
|
-
id: "google/gemma-3-27b-it:free",
|
|
7541
|
-
name: "Google: Gemma 3 27B (free)",
|
|
7542
|
-
api: "openai-completions",
|
|
7543
|
-
provider: "openrouter",
|
|
7544
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7545
|
-
reasoning: false,
|
|
7546
|
-
input: ["text", "image"],
|
|
7547
|
-
cost: {
|
|
7548
|
-
input: 0,
|
|
7549
|
-
output: 0,
|
|
7550
|
-
cacheRead: 0,
|
|
7551
|
-
cacheWrite: 0,
|
|
7552
|
-
},
|
|
7553
|
-
contextWindow: 131072,
|
|
7554
|
-
maxTokens: 8192,
|
|
7555
|
-
},
|
|
7556
7521
|
"inception/mercury": {
|
|
7557
7522
|
id: "inception/mercury",
|
|
7558
7523
|
name: "Inception: Mercury",
|
|
@@ -7655,23 +7620,6 @@ export const MODELS = {
|
|
|
7655
7620
|
contextWindow: 8192,
|
|
7656
7621
|
maxTokens: 16384,
|
|
7657
7622
|
},
|
|
7658
|
-
"meta-llama/llama-3.1-405b-instruct": {
|
|
7659
|
-
id: "meta-llama/llama-3.1-405b-instruct",
|
|
7660
|
-
name: "Meta: Llama 3.1 405B Instruct",
|
|
7661
|
-
api: "openai-completions",
|
|
7662
|
-
provider: "openrouter",
|
|
7663
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7664
|
-
reasoning: false,
|
|
7665
|
-
input: ["text"],
|
|
7666
|
-
cost: {
|
|
7667
|
-
input: 4,
|
|
7668
|
-
output: 4,
|
|
7669
|
-
cacheRead: 0,
|
|
7670
|
-
cacheWrite: 0,
|
|
7671
|
-
},
|
|
7672
|
-
contextWindow: 131000,
|
|
7673
|
-
maxTokens: 4096,
|
|
7674
|
-
},
|
|
7675
7623
|
"meta-llama/llama-3.1-70b-instruct": {
|
|
7676
7624
|
id: "meta-llama/llama-3.1-70b-instruct",
|
|
7677
7625
|
name: "Meta: Llama 3.1 70B Instruct",
|
|
@@ -7737,8 +7685,8 @@ export const MODELS = {
|
|
|
7737
7685
|
cacheRead: 0,
|
|
7738
7686
|
cacheWrite: 0,
|
|
7739
7687
|
},
|
|
7740
|
-
contextWindow:
|
|
7741
|
-
maxTokens:
|
|
7688
|
+
contextWindow: 65536,
|
|
7689
|
+
maxTokens: 4096,
|
|
7742
7690
|
},
|
|
7743
7691
|
"meta-llama/llama-4-maverick": {
|
|
7744
7692
|
id: "meta-llama/llama-4-maverick",
|
|
@@ -7834,28 +7782,62 @@ export const MODELS = {
|
|
|
7834
7782
|
reasoning: true,
|
|
7835
7783
|
input: ["text"],
|
|
7836
7784
|
cost: {
|
|
7837
|
-
input: 0.
|
|
7838
|
-
output: 1.
|
|
7839
|
-
cacheRead: 0.
|
|
7785
|
+
input: 0.19999999999999998,
|
|
7786
|
+
output: 1.17,
|
|
7787
|
+
cacheRead: 0.09999999999999999,
|
|
7840
7788
|
cacheWrite: 0,
|
|
7841
7789
|
},
|
|
7842
7790
|
contextWindow: 196608,
|
|
7843
|
-
maxTokens:
|
|
7791
|
+
maxTokens: 65536,
|
|
7844
7792
|
},
|
|
7845
|
-
"
|
|
7846
|
-
id: "
|
|
7847
|
-
name: "
|
|
7793
|
+
"minimax/minimax-m2.5:free": {
|
|
7794
|
+
id: "minimax/minimax-m2.5:free",
|
|
7795
|
+
name: "MiniMax: MiniMax M2.5 (free)",
|
|
7848
7796
|
api: "openai-completions",
|
|
7849
7797
|
provider: "openrouter",
|
|
7850
7798
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7851
|
-
reasoning:
|
|
7799
|
+
reasoning: true,
|
|
7852
7800
|
input: ["text"],
|
|
7853
7801
|
cost: {
|
|
7854
|
-
input: 0
|
|
7855
|
-
output: 0
|
|
7802
|
+
input: 0,
|
|
7803
|
+
output: 0,
|
|
7856
7804
|
cacheRead: 0,
|
|
7857
7805
|
cacheWrite: 0,
|
|
7858
7806
|
},
|
|
7807
|
+
contextWindow: 196608,
|
|
7808
|
+
maxTokens: 196608,
|
|
7809
|
+
},
|
|
7810
|
+
"minimax/minimax-m2.7": {
|
|
7811
|
+
id: "minimax/minimax-m2.7",
|
|
7812
|
+
name: "MiniMax: MiniMax M2.7",
|
|
7813
|
+
api: "openai-completions",
|
|
7814
|
+
provider: "openrouter",
|
|
7815
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7816
|
+
reasoning: true,
|
|
7817
|
+
input: ["text"],
|
|
7818
|
+
cost: {
|
|
7819
|
+
input: 0.3,
|
|
7820
|
+
output: 1.2,
|
|
7821
|
+
cacheRead: 0.06,
|
|
7822
|
+
cacheWrite: 0,
|
|
7823
|
+
},
|
|
7824
|
+
contextWindow: 204800,
|
|
7825
|
+
maxTokens: 131072,
|
|
7826
|
+
},
|
|
7827
|
+
"mistralai/codestral-2508": {
|
|
7828
|
+
id: "mistralai/codestral-2508",
|
|
7829
|
+
name: "Mistral: Codestral 2508",
|
|
7830
|
+
api: "openai-completions",
|
|
7831
|
+
provider: "openrouter",
|
|
7832
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7833
|
+
reasoning: false,
|
|
7834
|
+
input: ["text"],
|
|
7835
|
+
cost: {
|
|
7836
|
+
input: 0.3,
|
|
7837
|
+
output: 0.8999999999999999,
|
|
7838
|
+
cacheRead: 0.03,
|
|
7839
|
+
cacheWrite: 0,
|
|
7840
|
+
},
|
|
7859
7841
|
contextWindow: 256000,
|
|
7860
7842
|
maxTokens: 4096,
|
|
7861
7843
|
},
|
|
@@ -7870,7 +7852,7 @@ export const MODELS = {
|
|
|
7870
7852
|
cost: {
|
|
7871
7853
|
input: 0.39999999999999997,
|
|
7872
7854
|
output: 2,
|
|
7873
|
-
cacheRead: 0,
|
|
7855
|
+
cacheRead: 0.04,
|
|
7874
7856
|
cacheWrite: 0,
|
|
7875
7857
|
},
|
|
7876
7858
|
contextWindow: 262144,
|
|
@@ -7887,7 +7869,7 @@ export const MODELS = {
|
|
|
7887
7869
|
cost: {
|
|
7888
7870
|
input: 0.39999999999999997,
|
|
7889
7871
|
output: 2,
|
|
7890
|
-
cacheRead: 0,
|
|
7872
|
+
cacheRead: 0.04,
|
|
7891
7873
|
cacheWrite: 0,
|
|
7892
7874
|
},
|
|
7893
7875
|
contextWindow: 131072,
|
|
@@ -7904,7 +7886,7 @@ export const MODELS = {
|
|
|
7904
7886
|
cost: {
|
|
7905
7887
|
input: 0.09999999999999999,
|
|
7906
7888
|
output: 0.3,
|
|
7907
|
-
cacheRead: 0,
|
|
7889
|
+
cacheRead: 0.01,
|
|
7908
7890
|
cacheWrite: 0,
|
|
7909
7891
|
},
|
|
7910
7892
|
contextWindow: 131072,
|
|
@@ -7921,7 +7903,7 @@ export const MODELS = {
|
|
|
7921
7903
|
cost: {
|
|
7922
7904
|
input: 0.19999999999999998,
|
|
7923
7905
|
output: 0.19999999999999998,
|
|
7924
|
-
cacheRead: 0,
|
|
7906
|
+
cacheRead: 0.02,
|
|
7925
7907
|
cacheWrite: 0,
|
|
7926
7908
|
},
|
|
7927
7909
|
contextWindow: 262144,
|
|
@@ -7938,7 +7920,7 @@ export const MODELS = {
|
|
|
7938
7920
|
cost: {
|
|
7939
7921
|
input: 0.09999999999999999,
|
|
7940
7922
|
output: 0.09999999999999999,
|
|
7941
|
-
cacheRead: 0,
|
|
7923
|
+
cacheRead: 0.01,
|
|
7942
7924
|
cacheWrite: 0,
|
|
7943
7925
|
},
|
|
7944
7926
|
contextWindow: 131072,
|
|
@@ -7955,7 +7937,7 @@ export const MODELS = {
|
|
|
7955
7937
|
cost: {
|
|
7956
7938
|
input: 0.15,
|
|
7957
7939
|
output: 0.15,
|
|
7958
|
-
cacheRead: 0,
|
|
7940
|
+
cacheRead: 0.015,
|
|
7959
7941
|
cacheWrite: 0,
|
|
7960
7942
|
},
|
|
7961
7943
|
contextWindow: 262144,
|
|
@@ -7972,7 +7954,7 @@ export const MODELS = {
|
|
|
7972
7954
|
cost: {
|
|
7973
7955
|
input: 2,
|
|
7974
7956
|
output: 6,
|
|
7975
|
-
cacheRead: 0,
|
|
7957
|
+
cacheRead: 0.19999999999999998,
|
|
7976
7958
|
cacheWrite: 0,
|
|
7977
7959
|
},
|
|
7978
7960
|
contextWindow: 128000,
|
|
@@ -7989,7 +7971,7 @@ export const MODELS = {
|
|
|
7989
7971
|
cost: {
|
|
7990
7972
|
input: 2,
|
|
7991
7973
|
output: 6,
|
|
7992
|
-
cacheRead: 0,
|
|
7974
|
+
cacheRead: 0.19999999999999998,
|
|
7993
7975
|
cacheWrite: 0,
|
|
7994
7976
|
},
|
|
7995
7977
|
contextWindow: 131072,
|
|
@@ -8006,7 +7988,7 @@ export const MODELS = {
|
|
|
8006
7988
|
cost: {
|
|
8007
7989
|
input: 2,
|
|
8008
7990
|
output: 6,
|
|
8009
|
-
cacheRead: 0,
|
|
7991
|
+
cacheRead: 0.19999999999999998,
|
|
8010
7992
|
cacheWrite: 0,
|
|
8011
7993
|
},
|
|
8012
7994
|
contextWindow: 131072,
|
|
@@ -8023,7 +8005,7 @@ export const MODELS = {
|
|
|
8023
8005
|
cost: {
|
|
8024
8006
|
input: 0.5,
|
|
8025
8007
|
output: 1.5,
|
|
8026
|
-
cacheRead: 0,
|
|
8008
|
+
cacheRead: 0.049999999999999996,
|
|
8027
8009
|
cacheWrite: 0,
|
|
8028
8010
|
},
|
|
8029
8011
|
contextWindow: 262144,
|
|
@@ -8040,7 +8022,7 @@ export const MODELS = {
|
|
|
8040
8022
|
cost: {
|
|
8041
8023
|
input: 0.39999999999999997,
|
|
8042
8024
|
output: 2,
|
|
8043
|
-
cacheRead: 0,
|
|
8025
|
+
cacheRead: 0.04,
|
|
8044
8026
|
cacheWrite: 0,
|
|
8045
8027
|
},
|
|
8046
8028
|
contextWindow: 131072,
|
|
@@ -8057,7 +8039,7 @@ export const MODELS = {
|
|
|
8057
8039
|
cost: {
|
|
8058
8040
|
input: 0.39999999999999997,
|
|
8059
8041
|
output: 2,
|
|
8060
|
-
cacheRead: 0,
|
|
8042
|
+
cacheRead: 0.04,
|
|
8061
8043
|
cacheWrite: 0,
|
|
8062
8044
|
},
|
|
8063
8045
|
contextWindow: 131072,
|
|
@@ -8091,7 +8073,7 @@ export const MODELS = {
|
|
|
8091
8073
|
cost: {
|
|
8092
8074
|
input: 0.19999999999999998,
|
|
8093
8075
|
output: 0.6,
|
|
8094
|
-
cacheRead: 0,
|
|
8076
|
+
cacheRead: 0.02,
|
|
8095
8077
|
cacheWrite: 0,
|
|
8096
8078
|
},
|
|
8097
8079
|
contextWindow: 32768,
|
|
@@ -8114,6 +8096,23 @@ export const MODELS = {
|
|
|
8114
8096
|
contextWindow: 32768,
|
|
8115
8097
|
maxTokens: 16384,
|
|
8116
8098
|
},
|
|
8099
|
+
"mistralai/mistral-small-2603": {
|
|
8100
|
+
id: "mistralai/mistral-small-2603",
|
|
8101
|
+
name: "Mistral: Mistral Small 4",
|
|
8102
|
+
api: "openai-completions",
|
|
8103
|
+
provider: "openrouter",
|
|
8104
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8105
|
+
reasoning: true,
|
|
8106
|
+
input: ["text", "image"],
|
|
8107
|
+
cost: {
|
|
8108
|
+
input: 0.15,
|
|
8109
|
+
output: 0.6,
|
|
8110
|
+
cacheRead: 0.015,
|
|
8111
|
+
cacheWrite: 0,
|
|
8112
|
+
},
|
|
8113
|
+
contextWindow: 262144,
|
|
8114
|
+
maxTokens: 4096,
|
|
8115
|
+
},
|
|
8117
8116
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8118
8117
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
8119
8118
|
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
@@ -8140,13 +8139,13 @@ export const MODELS = {
|
|
|
8140
8139
|
reasoning: false,
|
|
8141
8140
|
input: ["text", "image"],
|
|
8142
8141
|
cost: {
|
|
8143
|
-
input: 0.
|
|
8144
|
-
output: 0.
|
|
8145
|
-
cacheRead: 0
|
|
8142
|
+
input: 0.075,
|
|
8143
|
+
output: 0.19999999999999998,
|
|
8144
|
+
cacheRead: 0,
|
|
8146
8145
|
cacheWrite: 0,
|
|
8147
8146
|
},
|
|
8148
|
-
contextWindow:
|
|
8149
|
-
maxTokens:
|
|
8147
|
+
contextWindow: 128000,
|
|
8148
|
+
maxTokens: 4096,
|
|
8150
8149
|
},
|
|
8151
8150
|
"mistralai/mistral-small-creative": {
|
|
8152
8151
|
id: "mistralai/mistral-small-creative",
|
|
@@ -8159,7 +8158,7 @@ export const MODELS = {
|
|
|
8159
8158
|
cost: {
|
|
8160
8159
|
input: 0.09999999999999999,
|
|
8161
8160
|
output: 0.3,
|
|
8162
|
-
cacheRead: 0,
|
|
8161
|
+
cacheRead: 0.01,
|
|
8163
8162
|
cacheWrite: 0,
|
|
8164
8163
|
},
|
|
8165
8164
|
contextWindow: 32768,
|
|
@@ -8176,7 +8175,7 @@ export const MODELS = {
|
|
|
8176
8175
|
cost: {
|
|
8177
8176
|
input: 2,
|
|
8178
8177
|
output: 6,
|
|
8179
|
-
cacheRead: 0,
|
|
8178
|
+
cacheRead: 0.19999999999999998,
|
|
8180
8179
|
cacheWrite: 0,
|
|
8181
8180
|
},
|
|
8182
8181
|
contextWindow: 65536,
|
|
@@ -8210,7 +8209,7 @@ export const MODELS = {
|
|
|
8210
8209
|
cost: {
|
|
8211
8210
|
input: 2,
|
|
8212
8211
|
output: 6,
|
|
8213
|
-
cacheRead: 0,
|
|
8212
|
+
cacheRead: 0.19999999999999998,
|
|
8214
8213
|
cacheWrite: 0,
|
|
8215
8214
|
},
|
|
8216
8215
|
contextWindow: 131072,
|
|
@@ -8227,7 +8226,7 @@ export const MODELS = {
|
|
|
8227
8226
|
cost: {
|
|
8228
8227
|
input: 0.09999999999999999,
|
|
8229
8228
|
output: 0.3,
|
|
8230
|
-
cacheRead: 0,
|
|
8229
|
+
cacheRead: 0.01,
|
|
8231
8230
|
cacheWrite: 0,
|
|
8232
8231
|
},
|
|
8233
8232
|
contextWindow: 32000,
|
|
@@ -8267,23 +8266,6 @@ export const MODELS = {
|
|
|
8267
8266
|
contextWindow: 131072,
|
|
8268
8267
|
maxTokens: 4096,
|
|
8269
8268
|
},
|
|
8270
|
-
"moonshotai/kimi-k2-0905:exacto": {
|
|
8271
|
-
id: "moonshotai/kimi-k2-0905:exacto",
|
|
8272
|
-
name: "MoonshotAI: Kimi K2 0905 (exacto)",
|
|
8273
|
-
api: "openai-completions",
|
|
8274
|
-
provider: "openrouter",
|
|
8275
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8276
|
-
reasoning: false,
|
|
8277
|
-
input: ["text"],
|
|
8278
|
-
cost: {
|
|
8279
|
-
input: 0.6,
|
|
8280
|
-
output: 2.5,
|
|
8281
|
-
cacheRead: 0,
|
|
8282
|
-
cacheWrite: 0,
|
|
8283
|
-
},
|
|
8284
|
-
contextWindow: 262144,
|
|
8285
|
-
maxTokens: 4096,
|
|
8286
|
-
},
|
|
8287
8269
|
"moonshotai/kimi-k2-thinking": {
|
|
8288
8270
|
id: "moonshotai/kimi-k2-thinking",
|
|
8289
8271
|
name: "MoonshotAI: Kimi K2 Thinking",
|
|
@@ -8403,6 +8385,40 @@ export const MODELS = {
|
|
|
8403
8385
|
contextWindow: 256000,
|
|
8404
8386
|
maxTokens: 4096,
|
|
8405
8387
|
},
|
|
8388
|
+
"nvidia/nemotron-3-super-120b-a12b": {
|
|
8389
|
+
id: "nvidia/nemotron-3-super-120b-a12b",
|
|
8390
|
+
name: "NVIDIA: Nemotron 3 Super",
|
|
8391
|
+
api: "openai-completions",
|
|
8392
|
+
provider: "openrouter",
|
|
8393
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8394
|
+
reasoning: true,
|
|
8395
|
+
input: ["text"],
|
|
8396
|
+
cost: {
|
|
8397
|
+
input: 0.09999999999999999,
|
|
8398
|
+
output: 0.5,
|
|
8399
|
+
cacheRead: 0.04,
|
|
8400
|
+
cacheWrite: 0,
|
|
8401
|
+
},
|
|
8402
|
+
contextWindow: 262144,
|
|
8403
|
+
maxTokens: 4096,
|
|
8404
|
+
},
|
|
8405
|
+
"nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8406
|
+
id: "nvidia/nemotron-3-super-120b-a12b:free",
|
|
8407
|
+
name: "NVIDIA: Nemotron 3 Super (free)",
|
|
8408
|
+
api: "openai-completions",
|
|
8409
|
+
provider: "openrouter",
|
|
8410
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8411
|
+
reasoning: true,
|
|
8412
|
+
input: ["text"],
|
|
8413
|
+
cost: {
|
|
8414
|
+
input: 0,
|
|
8415
|
+
output: 0,
|
|
8416
|
+
cacheRead: 0,
|
|
8417
|
+
cacheWrite: 0,
|
|
8418
|
+
},
|
|
8419
|
+
contextWindow: 262144,
|
|
8420
|
+
maxTokens: 262144,
|
|
8421
|
+
},
|
|
8406
8422
|
"nvidia/nemotron-nano-12b-v2-vl:free": {
|
|
8407
8423
|
id: "nvidia/nemotron-nano-12b-v2-vl:free",
|
|
8408
8424
|
name: "NVIDIA: Nemotron Nano 12B 2 VL (free)",
|
|
@@ -9100,6 +9116,40 @@ export const MODELS = {
|
|
|
9100
9116
|
contextWindow: 1050000,
|
|
9101
9117
|
maxTokens: 128000,
|
|
9102
9118
|
},
|
|
9119
|
+
"openai/gpt-5.4-mini": {
|
|
9120
|
+
id: "openai/gpt-5.4-mini",
|
|
9121
|
+
name: "OpenAI: GPT-5.4 Mini",
|
|
9122
|
+
api: "openai-completions",
|
|
9123
|
+
provider: "openrouter",
|
|
9124
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9125
|
+
reasoning: true,
|
|
9126
|
+
input: ["text", "image"],
|
|
9127
|
+
cost: {
|
|
9128
|
+
input: 0.75,
|
|
9129
|
+
output: 4.5,
|
|
9130
|
+
cacheRead: 0.075,
|
|
9131
|
+
cacheWrite: 0,
|
|
9132
|
+
},
|
|
9133
|
+
contextWindow: 400000,
|
|
9134
|
+
maxTokens: 128000,
|
|
9135
|
+
},
|
|
9136
|
+
"openai/gpt-5.4-nano": {
|
|
9137
|
+
id: "openai/gpt-5.4-nano",
|
|
9138
|
+
name: "OpenAI: GPT-5.4 Nano",
|
|
9139
|
+
api: "openai-completions",
|
|
9140
|
+
provider: "openrouter",
|
|
9141
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9142
|
+
reasoning: true,
|
|
9143
|
+
input: ["text", "image"],
|
|
9144
|
+
cost: {
|
|
9145
|
+
input: 0.19999999999999998,
|
|
9146
|
+
output: 1.25,
|
|
9147
|
+
cacheRead: 0.02,
|
|
9148
|
+
cacheWrite: 0,
|
|
9149
|
+
},
|
|
9150
|
+
contextWindow: 400000,
|
|
9151
|
+
maxTokens: 128000,
|
|
9152
|
+
},
|
|
9103
9153
|
"openai/gpt-5.4-pro": {
|
|
9104
9154
|
id: "openai/gpt-5.4-pro",
|
|
9105
9155
|
name: "OpenAI: GPT-5.4 Pro",
|
|
@@ -9134,23 +9184,6 @@ export const MODELS = {
|
|
|
9134
9184
|
contextWindow: 131072,
|
|
9135
9185
|
maxTokens: 4096,
|
|
9136
9186
|
},
|
|
9137
|
-
"openai/gpt-oss-120b:exacto": {
|
|
9138
|
-
id: "openai/gpt-oss-120b:exacto",
|
|
9139
|
-
name: "OpenAI: gpt-oss-120b (exacto)",
|
|
9140
|
-
api: "openai-completions",
|
|
9141
|
-
provider: "openrouter",
|
|
9142
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9143
|
-
reasoning: true,
|
|
9144
|
-
input: ["text"],
|
|
9145
|
-
cost: {
|
|
9146
|
-
input: 0.039,
|
|
9147
|
-
output: 0.19,
|
|
9148
|
-
cacheRead: 0,
|
|
9149
|
-
cacheWrite: 0,
|
|
9150
|
-
},
|
|
9151
|
-
contextWindow: 131072,
|
|
9152
|
-
maxTokens: 4096,
|
|
9153
|
-
},
|
|
9154
9187
|
"openai/gpt-oss-120b:free": {
|
|
9155
9188
|
id: "openai/gpt-oss-120b:free",
|
|
9156
9189
|
name: "OpenAI: gpt-oss-120b (free)",
|
|
@@ -9178,12 +9211,12 @@ export const MODELS = {
|
|
|
9178
9211
|
input: ["text"],
|
|
9179
9212
|
cost: {
|
|
9180
9213
|
input: 0.03,
|
|
9181
|
-
output: 0.
|
|
9182
|
-
cacheRead: 0,
|
|
9214
|
+
output: 0.11,
|
|
9215
|
+
cacheRead: 0.015,
|
|
9183
9216
|
cacheWrite: 0,
|
|
9184
9217
|
},
|
|
9185
9218
|
contextWindow: 131072,
|
|
9186
|
-
maxTokens:
|
|
9219
|
+
maxTokens: 131072,
|
|
9187
9220
|
},
|
|
9188
9221
|
"openai/gpt-oss-20b:free": {
|
|
9189
9222
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -9225,7 +9258,7 @@ export const MODELS = {
|
|
|
9225
9258
|
api: "openai-completions",
|
|
9226
9259
|
provider: "openrouter",
|
|
9227
9260
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9228
|
-
reasoning:
|
|
9261
|
+
reasoning: true,
|
|
9229
9262
|
input: ["text", "image"],
|
|
9230
9263
|
cost: {
|
|
9231
9264
|
input: 15,
|
|
@@ -9276,7 +9309,7 @@ export const MODELS = {
|
|
|
9276
9309
|
api: "openai-completions",
|
|
9277
9310
|
provider: "openrouter",
|
|
9278
9311
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9279
|
-
reasoning:
|
|
9312
|
+
reasoning: true,
|
|
9280
9313
|
input: ["text"],
|
|
9281
9314
|
cost: {
|
|
9282
9315
|
input: 1.1,
|
|
@@ -9293,7 +9326,7 @@ export const MODELS = {
|
|
|
9293
9326
|
api: "openai-completions",
|
|
9294
9327
|
provider: "openrouter",
|
|
9295
9328
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9296
|
-
reasoning:
|
|
9329
|
+
reasoning: true,
|
|
9297
9330
|
input: ["text"],
|
|
9298
9331
|
cost: {
|
|
9299
9332
|
input: 1.1,
|
|
@@ -9483,9 +9516,9 @@ export const MODELS = {
|
|
|
9483
9516
|
reasoning: false,
|
|
9484
9517
|
input: ["text"],
|
|
9485
9518
|
cost: {
|
|
9486
|
-
input: 0.
|
|
9487
|
-
output:
|
|
9488
|
-
cacheRead: 0.
|
|
9519
|
+
input: 0.26,
|
|
9520
|
+
output: 0.78,
|
|
9521
|
+
cacheRead: 0.052000000000000005,
|
|
9489
9522
|
cacheWrite: 0,
|
|
9490
9523
|
},
|
|
9491
9524
|
contextWindow: 1000000,
|
|
@@ -9551,8 +9584,8 @@ export const MODELS = {
|
|
|
9551
9584
|
reasoning: false,
|
|
9552
9585
|
input: ["text", "image"],
|
|
9553
9586
|
cost: {
|
|
9554
|
-
input: 0.
|
|
9555
|
-
output:
|
|
9587
|
+
input: 0.52,
|
|
9588
|
+
output: 2.08,
|
|
9556
9589
|
cacheRead: 0,
|
|
9557
9590
|
cacheWrite: 0,
|
|
9558
9591
|
},
|
|
@@ -9619,13 +9652,13 @@ export const MODELS = {
|
|
|
9619
9652
|
reasoning: true,
|
|
9620
9653
|
input: ["text"],
|
|
9621
9654
|
cost: {
|
|
9622
|
-
input: 0.
|
|
9623
|
-
output:
|
|
9624
|
-
cacheRead: 0
|
|
9655
|
+
input: 0.14950000000000002,
|
|
9656
|
+
output: 1.495,
|
|
9657
|
+
cacheRead: 0,
|
|
9625
9658
|
cacheWrite: 0,
|
|
9626
9659
|
},
|
|
9627
|
-
contextWindow:
|
|
9628
|
-
maxTokens:
|
|
9660
|
+
contextWindow: 131072,
|
|
9661
|
+
maxTokens: 4096,
|
|
9629
9662
|
},
|
|
9630
9663
|
"qwen/qwen3-30b-a3b": {
|
|
9631
9664
|
id: "qwen/qwen3-30b-a3b",
|
|
@@ -9670,13 +9703,13 @@ export const MODELS = {
|
|
|
9670
9703
|
reasoning: true,
|
|
9671
9704
|
input: ["text"],
|
|
9672
9705
|
cost: {
|
|
9673
|
-
input: 0.
|
|
9674
|
-
output: 0.
|
|
9675
|
-
cacheRead: 0,
|
|
9706
|
+
input: 0.08,
|
|
9707
|
+
output: 0.39999999999999997,
|
|
9708
|
+
cacheRead: 0.08,
|
|
9676
9709
|
cacheWrite: 0,
|
|
9677
9710
|
},
|
|
9678
|
-
contextWindow:
|
|
9679
|
-
maxTokens:
|
|
9711
|
+
contextWindow: 131072,
|
|
9712
|
+
maxTokens: 131072,
|
|
9680
9713
|
},
|
|
9681
9714
|
"qwen/qwen3-32b": {
|
|
9682
9715
|
id: "qwen/qwen3-32b",
|
|
@@ -9814,23 +9847,6 @@ export const MODELS = {
|
|
|
9814
9847
|
contextWindow: 1000000,
|
|
9815
9848
|
maxTokens: 65536,
|
|
9816
9849
|
},
|
|
9817
|
-
"qwen/qwen3-coder:exacto": {
|
|
9818
|
-
id: "qwen/qwen3-coder:exacto",
|
|
9819
|
-
name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
|
|
9820
|
-
api: "openai-completions",
|
|
9821
|
-
provider: "openrouter",
|
|
9822
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9823
|
-
reasoning: false,
|
|
9824
|
-
input: ["text"],
|
|
9825
|
-
cost: {
|
|
9826
|
-
input: 0.22,
|
|
9827
|
-
output: 1.7999999999999998,
|
|
9828
|
-
cacheRead: 0.022,
|
|
9829
|
-
cacheWrite: 0,
|
|
9830
|
-
},
|
|
9831
|
-
contextWindow: 262144,
|
|
9832
|
-
maxTokens: 65536,
|
|
9833
|
-
},
|
|
9834
9850
|
"qwen/qwen3-coder:free": {
|
|
9835
9851
|
id: "qwen/qwen3-coder:free",
|
|
9836
9852
|
name: "Qwen: Qwen3 Coder 480B A35B (free)",
|
|
@@ -9857,9 +9873,9 @@ export const MODELS = {
|
|
|
9857
9873
|
reasoning: false,
|
|
9858
9874
|
input: ["text"],
|
|
9859
9875
|
cost: {
|
|
9860
|
-
input:
|
|
9861
|
-
output:
|
|
9862
|
-
cacheRead: 0.
|
|
9876
|
+
input: 0.78,
|
|
9877
|
+
output: 3.9,
|
|
9878
|
+
cacheRead: 0.156,
|
|
9863
9879
|
cacheWrite: 0,
|
|
9864
9880
|
},
|
|
9865
9881
|
contextWindow: 262144,
|
|
@@ -9925,13 +9941,13 @@ export const MODELS = {
|
|
|
9925
9941
|
reasoning: true,
|
|
9926
9942
|
input: ["text"],
|
|
9927
9943
|
cost: {
|
|
9928
|
-
input: 0.
|
|
9929
|
-
output:
|
|
9944
|
+
input: 0.0975,
|
|
9945
|
+
output: 0.78,
|
|
9930
9946
|
cacheRead: 0,
|
|
9931
9947
|
cacheWrite: 0,
|
|
9932
9948
|
},
|
|
9933
|
-
contextWindow:
|
|
9934
|
-
maxTokens:
|
|
9949
|
+
contextWindow: 131072,
|
|
9950
|
+
maxTokens: 32768,
|
|
9935
9951
|
},
|
|
9936
9952
|
"qwen/qwen3-vl-235b-a22b-instruct": {
|
|
9937
9953
|
id: "qwen/qwen3-vl-235b-a22b-instruct",
|
|
@@ -9959,8 +9975,8 @@ export const MODELS = {
|
|
|
9959
9975
|
reasoning: true,
|
|
9960
9976
|
input: ["text", "image"],
|
|
9961
9977
|
cost: {
|
|
9962
|
-
input: 0,
|
|
9963
|
-
output:
|
|
9978
|
+
input: 0.26,
|
|
9979
|
+
output: 2.6,
|
|
9964
9980
|
cacheRead: 0,
|
|
9965
9981
|
cacheWrite: 0,
|
|
9966
9982
|
},
|
|
@@ -9993,8 +10009,8 @@ export const MODELS = {
|
|
|
9993
10009
|
reasoning: true,
|
|
9994
10010
|
input: ["text", "image"],
|
|
9995
10011
|
cost: {
|
|
9996
|
-
input: 0,
|
|
9997
|
-
output:
|
|
10012
|
+
input: 0.13,
|
|
10013
|
+
output: 1.56,
|
|
9998
10014
|
cacheRead: 0,
|
|
9999
10015
|
cacheWrite: 0,
|
|
10000
10016
|
},
|
|
@@ -10120,6 +10136,23 @@ export const MODELS = {
|
|
|
10120
10136
|
contextWindow: 262144,
|
|
10121
10137
|
maxTokens: 65536,
|
|
10122
10138
|
},
|
|
10139
|
+
"qwen/qwen3.5-9b": {
|
|
10140
|
+
id: "qwen/qwen3.5-9b",
|
|
10141
|
+
name: "Qwen: Qwen3.5-9B",
|
|
10142
|
+
api: "openai-completions",
|
|
10143
|
+
provider: "openrouter",
|
|
10144
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10145
|
+
reasoning: true,
|
|
10146
|
+
input: ["text", "image"],
|
|
10147
|
+
cost: {
|
|
10148
|
+
input: 0.049999999999999996,
|
|
10149
|
+
output: 0.15,
|
|
10150
|
+
cacheRead: 0,
|
|
10151
|
+
cacheWrite: 0,
|
|
10152
|
+
},
|
|
10153
|
+
contextWindow: 256000,
|
|
10154
|
+
maxTokens: 4096,
|
|
10155
|
+
},
|
|
10123
10156
|
"qwen/qwen3.5-flash-02-23": {
|
|
10124
10157
|
id: "qwen/qwen3.5-flash-02-23",
|
|
10125
10158
|
name: "Qwen: Qwen3.5-Flash",
|
|
@@ -10129,8 +10162,8 @@ export const MODELS = {
|
|
|
10129
10162
|
reasoning: true,
|
|
10130
10163
|
input: ["text", "image"],
|
|
10131
10164
|
cost: {
|
|
10132
|
-
input: 0.
|
|
10133
|
-
output: 0.
|
|
10165
|
+
input: 0.065,
|
|
10166
|
+
output: 0.26,
|
|
10134
10167
|
cacheRead: 0,
|
|
10135
10168
|
cacheWrite: 0,
|
|
10136
10169
|
},
|
|
@@ -10164,12 +10197,12 @@ export const MODELS = {
|
|
|
10164
10197
|
input: ["text"],
|
|
10165
10198
|
cost: {
|
|
10166
10199
|
input: 0.15,
|
|
10167
|
-
output: 0.
|
|
10200
|
+
output: 0.58,
|
|
10168
10201
|
cacheRead: 0,
|
|
10169
10202
|
cacheWrite: 0,
|
|
10170
10203
|
},
|
|
10171
|
-
contextWindow:
|
|
10172
|
-
maxTokens:
|
|
10204
|
+
contextWindow: 131072,
|
|
10205
|
+
maxTokens: 131072,
|
|
10173
10206
|
},
|
|
10174
10207
|
"relace/relace-search": {
|
|
10175
10208
|
id: "relace/relace-search",
|
|
@@ -10214,13 +10247,13 @@ export const MODELS = {
|
|
|
10214
10247
|
reasoning: false,
|
|
10215
10248
|
input: ["text"],
|
|
10216
10249
|
cost: {
|
|
10217
|
-
input: 0.
|
|
10218
|
-
output: 0.
|
|
10250
|
+
input: 0.85,
|
|
10251
|
+
output: 0.85,
|
|
10219
10252
|
cacheRead: 0,
|
|
10220
10253
|
cacheWrite: 0,
|
|
10221
10254
|
},
|
|
10222
|
-
contextWindow:
|
|
10223
|
-
maxTokens:
|
|
10255
|
+
contextWindow: 131072,
|
|
10256
|
+
maxTokens: 16384,
|
|
10224
10257
|
},
|
|
10225
10258
|
"stepfun/step-3.5-flash": {
|
|
10226
10259
|
id: "stepfun/step-3.5-flash",
|
|
@@ -10299,9 +10332,9 @@ export const MODELS = {
|
|
|
10299
10332
|
reasoning: true,
|
|
10300
10333
|
input: ["text"],
|
|
10301
10334
|
cost: {
|
|
10302
|
-
input: 0.
|
|
10303
|
-
output:
|
|
10304
|
-
cacheRead: 0.
|
|
10335
|
+
input: 0.3,
|
|
10336
|
+
output: 1.1,
|
|
10337
|
+
cacheRead: 0.15,
|
|
10305
10338
|
cacheWrite: 0,
|
|
10306
10339
|
},
|
|
10307
10340
|
contextWindow: 163840,
|
|
@@ -10443,6 +10476,23 @@ export const MODELS = {
|
|
|
10443
10476
|
contextWindow: 2000000,
|
|
10444
10477
|
maxTokens: 30000,
|
|
10445
10478
|
},
|
|
10479
|
+
"x-ai/grok-4.20-beta": {
|
|
10480
|
+
id: "x-ai/grok-4.20-beta",
|
|
10481
|
+
name: "xAI: Grok 4.20 Beta",
|
|
10482
|
+
api: "openai-completions",
|
|
10483
|
+
provider: "openrouter",
|
|
10484
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10485
|
+
reasoning: true,
|
|
10486
|
+
input: ["text", "image"],
|
|
10487
|
+
cost: {
|
|
10488
|
+
input: 2,
|
|
10489
|
+
output: 6,
|
|
10490
|
+
cacheRead: 0.19999999999999998,
|
|
10491
|
+
cacheWrite: 0,
|
|
10492
|
+
},
|
|
10493
|
+
contextWindow: 2000000,
|
|
10494
|
+
maxTokens: 4096,
|
|
10495
|
+
},
|
|
10446
10496
|
"x-ai/grok-code-fast-1": {
|
|
10447
10497
|
id: "x-ai/grok-code-fast-1",
|
|
10448
10498
|
name: "xAI: Grok Code Fast 1",
|
|
@@ -10477,6 +10527,40 @@ export const MODELS = {
|
|
|
10477
10527
|
contextWindow: 262144,
|
|
10478
10528
|
maxTokens: 65536,
|
|
10479
10529
|
},
|
|
10530
|
+
"xiaomi/mimo-v2-omni": {
|
|
10531
|
+
id: "xiaomi/mimo-v2-omni",
|
|
10532
|
+
name: "Xiaomi: MiMo-V2-Omni",
|
|
10533
|
+
api: "openai-completions",
|
|
10534
|
+
provider: "openrouter",
|
|
10535
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10536
|
+
reasoning: true,
|
|
10537
|
+
input: ["text", "image"],
|
|
10538
|
+
cost: {
|
|
10539
|
+
input: 0.39999999999999997,
|
|
10540
|
+
output: 2,
|
|
10541
|
+
cacheRead: 0.08,
|
|
10542
|
+
cacheWrite: 0,
|
|
10543
|
+
},
|
|
10544
|
+
contextWindow: 262144,
|
|
10545
|
+
maxTokens: 65536,
|
|
10546
|
+
},
|
|
10547
|
+
"xiaomi/mimo-v2-pro": {
|
|
10548
|
+
id: "xiaomi/mimo-v2-pro",
|
|
10549
|
+
name: "Xiaomi: MiMo-V2-Pro",
|
|
10550
|
+
api: "openai-completions",
|
|
10551
|
+
provider: "openrouter",
|
|
10552
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10553
|
+
reasoning: true,
|
|
10554
|
+
input: ["text"],
|
|
10555
|
+
cost: {
|
|
10556
|
+
input: 1,
|
|
10557
|
+
output: 3,
|
|
10558
|
+
cacheRead: 0.19999999999999998,
|
|
10559
|
+
cacheWrite: 0,
|
|
10560
|
+
},
|
|
10561
|
+
contextWindow: 1048576,
|
|
10562
|
+
maxTokens: 131072,
|
|
10563
|
+
},
|
|
10480
10564
|
"z-ai/glm-4-32b": {
|
|
10481
10565
|
id: "z-ai/glm-4-32b",
|
|
10482
10566
|
name: "Z.ai: GLM 4 32B ",
|
|
@@ -10579,23 +10663,6 @@ export const MODELS = {
|
|
|
10579
10663
|
contextWindow: 204800,
|
|
10580
10664
|
maxTokens: 204800,
|
|
10581
10665
|
},
|
|
10582
|
-
"z-ai/glm-4.6:exacto": {
|
|
10583
|
-
id: "z-ai/glm-4.6:exacto",
|
|
10584
|
-
name: "Z.ai: GLM 4.6 (exacto)",
|
|
10585
|
-
api: "openai-completions",
|
|
10586
|
-
provider: "openrouter",
|
|
10587
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
10588
|
-
reasoning: true,
|
|
10589
|
-
input: ["text"],
|
|
10590
|
-
cost: {
|
|
10591
|
-
input: 0.44,
|
|
10592
|
-
output: 1.76,
|
|
10593
|
-
cacheRead: 0.11,
|
|
10594
|
-
cacheWrite: 0,
|
|
10595
|
-
},
|
|
10596
|
-
contextWindow: 204800,
|
|
10597
|
-
maxTokens: 131072,
|
|
10598
|
-
},
|
|
10599
10666
|
"z-ai/glm-4.6v": {
|
|
10600
10667
|
id: "z-ai/glm-4.6v",
|
|
10601
10668
|
name: "Z.ai: GLM 4.6V",
|
|
@@ -10622,13 +10689,13 @@ export const MODELS = {
|
|
|
10622
10689
|
reasoning: true,
|
|
10623
10690
|
input: ["text"],
|
|
10624
10691
|
cost: {
|
|
10625
|
-
input: 0.
|
|
10626
|
-
output: 1.
|
|
10627
|
-
cacheRead: 0.
|
|
10692
|
+
input: 0.39,
|
|
10693
|
+
output: 1.75,
|
|
10694
|
+
cacheRead: 0.195,
|
|
10628
10695
|
cacheWrite: 0,
|
|
10629
10696
|
},
|
|
10630
10697
|
contextWindow: 202752,
|
|
10631
|
-
maxTokens:
|
|
10698
|
+
maxTokens: 65535,
|
|
10632
10699
|
},
|
|
10633
10700
|
"z-ai/glm-4.7-flash": {
|
|
10634
10701
|
id: "z-ai/glm-4.7-flash",
|
|
@@ -10661,8 +10728,25 @@ export const MODELS = {
|
|
|
10661
10728
|
cacheRead: 0.119,
|
|
10662
10729
|
cacheWrite: 0,
|
|
10663
10730
|
},
|
|
10731
|
+
contextWindow: 80000,
|
|
10732
|
+
maxTokens: 131072,
|
|
10733
|
+
},
|
|
10734
|
+
"z-ai/glm-5-turbo": {
|
|
10735
|
+
id: "z-ai/glm-5-turbo",
|
|
10736
|
+
name: "Z.ai: GLM 5 Turbo",
|
|
10737
|
+
api: "openai-completions",
|
|
10738
|
+
provider: "openrouter",
|
|
10739
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10740
|
+
reasoning: true,
|
|
10741
|
+
input: ["text"],
|
|
10742
|
+
cost: {
|
|
10743
|
+
input: 0.96,
|
|
10744
|
+
output: 3.1999999999999997,
|
|
10745
|
+
cacheRead: 0.192,
|
|
10746
|
+
cacheWrite: 0,
|
|
10747
|
+
},
|
|
10664
10748
|
contextWindow: 202752,
|
|
10665
|
-
maxTokens:
|
|
10749
|
+
maxTokens: 131072,
|
|
10666
10750
|
},
|
|
10667
10751
|
},
|
|
10668
10752
|
"vercel-ai-gateway": {
|
|
@@ -10675,7 +10759,7 @@ export const MODELS = {
|
|
|
10675
10759
|
reasoning: true,
|
|
10676
10760
|
input: ["text"],
|
|
10677
10761
|
cost: {
|
|
10678
|
-
input: 0.
|
|
10762
|
+
input: 0.12,
|
|
10679
10763
|
output: 0.24,
|
|
10680
10764
|
cacheRead: 0,
|
|
10681
10765
|
cacheWrite: 0,
|
|
@@ -10726,13 +10810,13 @@ export const MODELS = {
|
|
|
10726
10810
|
reasoning: true,
|
|
10727
10811
|
input: ["text"],
|
|
10728
10812
|
cost: {
|
|
10729
|
-
input: 0.
|
|
10730
|
-
output: 0.
|
|
10731
|
-
cacheRead: 0,
|
|
10813
|
+
input: 0.29,
|
|
10814
|
+
output: 0.59,
|
|
10815
|
+
cacheRead: 0.145,
|
|
10732
10816
|
cacheWrite: 0,
|
|
10733
10817
|
},
|
|
10734
|
-
contextWindow:
|
|
10735
|
-
maxTokens:
|
|
10818
|
+
contextWindow: 131072,
|
|
10819
|
+
maxTokens: 40960,
|
|
10736
10820
|
},
|
|
10737
10821
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10738
10822
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10743,9 +10827,9 @@ export const MODELS = {
|
|
|
10743
10827
|
reasoning: true,
|
|
10744
10828
|
input: ["text", "image"],
|
|
10745
10829
|
cost: {
|
|
10746
|
-
input: 0.
|
|
10747
|
-
output: 2.
|
|
10748
|
-
cacheRead: 0,
|
|
10830
|
+
input: 0.22999999999999998,
|
|
10831
|
+
output: 2.3,
|
|
10832
|
+
cacheRead: 0.19999999999999998,
|
|
10749
10833
|
cacheWrite: 0,
|
|
10750
10834
|
},
|
|
10751
10835
|
contextWindow: 262114,
|
|
@@ -10762,7 +10846,7 @@ export const MODELS = {
|
|
|
10762
10846
|
cost: {
|
|
10763
10847
|
input: 0.39999999999999997,
|
|
10764
10848
|
output: 1.5999999999999999,
|
|
10765
|
-
cacheRead: 0,
|
|
10849
|
+
cacheRead: 0.022,
|
|
10766
10850
|
cacheWrite: 0,
|
|
10767
10851
|
},
|
|
10768
10852
|
contextWindow: 262144,
|
|
@@ -10777,13 +10861,13 @@ export const MODELS = {
|
|
|
10777
10861
|
reasoning: true,
|
|
10778
10862
|
input: ["text"],
|
|
10779
10863
|
cost: {
|
|
10780
|
-
input: 0.
|
|
10781
|
-
output: 0.
|
|
10864
|
+
input: 0.15,
|
|
10865
|
+
output: 0.6,
|
|
10782
10866
|
cacheRead: 0,
|
|
10783
10867
|
cacheWrite: 0,
|
|
10784
10868
|
},
|
|
10785
|
-
contextWindow:
|
|
10786
|
-
maxTokens:
|
|
10869
|
+
contextWindow: 262144,
|
|
10870
|
+
maxTokens: 8192,
|
|
10787
10871
|
},
|
|
10788
10872
|
"alibaba/qwen3-coder-next": {
|
|
10789
10873
|
id: "alibaba/qwen3-coder-next",
|
|
@@ -10791,7 +10875,7 @@ export const MODELS = {
|
|
|
10791
10875
|
api: "anthropic-messages",
|
|
10792
10876
|
provider: "vercel-ai-gateway",
|
|
10793
10877
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10794
|
-
reasoning:
|
|
10878
|
+
reasoning: false,
|
|
10795
10879
|
input: ["text"],
|
|
10796
10880
|
cost: {
|
|
10797
10881
|
input: 0.5,
|
|
@@ -10819,6 +10903,23 @@ export const MODELS = {
|
|
|
10819
10903
|
contextWindow: 1000000,
|
|
10820
10904
|
maxTokens: 65536,
|
|
10821
10905
|
},
|
|
10906
|
+
"alibaba/qwen3-max": {
|
|
10907
|
+
id: "alibaba/qwen3-max",
|
|
10908
|
+
name: "Qwen3 Max",
|
|
10909
|
+
api: "anthropic-messages",
|
|
10910
|
+
provider: "vercel-ai-gateway",
|
|
10911
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10912
|
+
reasoning: false,
|
|
10913
|
+
input: ["text"],
|
|
10914
|
+
cost: {
|
|
10915
|
+
input: 1.2,
|
|
10916
|
+
output: 6,
|
|
10917
|
+
cacheRead: 0.24,
|
|
10918
|
+
cacheWrite: 0,
|
|
10919
|
+
},
|
|
10920
|
+
contextWindow: 262144,
|
|
10921
|
+
maxTokens: 32768,
|
|
10922
|
+
},
|
|
10822
10923
|
"alibaba/qwen3-max-preview": {
|
|
10823
10924
|
id: "alibaba/qwen3-max-preview",
|
|
10824
10925
|
name: "Qwen3 Max Preview",
|
|
@@ -10966,8 +11067,8 @@ export const MODELS = {
|
|
|
10966
11067
|
cost: {
|
|
10967
11068
|
input: 3,
|
|
10968
11069
|
output: 15,
|
|
10969
|
-
cacheRead: 0,
|
|
10970
|
-
cacheWrite:
|
|
11070
|
+
cacheRead: 0.3,
|
|
11071
|
+
cacheWrite: 3.75,
|
|
10971
11072
|
},
|
|
10972
11073
|
contextWindow: 200000,
|
|
10973
11074
|
maxTokens: 8192,
|
|
@@ -11176,6 +11277,23 @@ export const MODELS = {
|
|
|
11176
11277
|
contextWindow: 256000,
|
|
11177
11278
|
maxTokens: 8000,
|
|
11178
11279
|
},
|
|
11280
|
+
"deepseek/deepseek-r1": {
|
|
11281
|
+
id: "deepseek/deepseek-r1",
|
|
11282
|
+
name: "DeepSeek-R1",
|
|
11283
|
+
api: "anthropic-messages",
|
|
11284
|
+
provider: "vercel-ai-gateway",
|
|
11285
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11286
|
+
reasoning: true,
|
|
11287
|
+
input: ["text"],
|
|
11288
|
+
cost: {
|
|
11289
|
+
input: 1.35,
|
|
11290
|
+
output: 5.4,
|
|
11291
|
+
cacheRead: 0,
|
|
11292
|
+
cacheWrite: 0,
|
|
11293
|
+
},
|
|
11294
|
+
contextWindow: 128000,
|
|
11295
|
+
maxTokens: 8192,
|
|
11296
|
+
},
|
|
11179
11297
|
"deepseek/deepseek-v3": {
|
|
11180
11298
|
id: "deepseek/deepseek-v3",
|
|
11181
11299
|
name: "DeepSeek V3 0324",
|
|
@@ -11202,13 +11320,13 @@ export const MODELS = {
|
|
|
11202
11320
|
reasoning: true,
|
|
11203
11321
|
input: ["text"],
|
|
11204
11322
|
cost: {
|
|
11205
|
-
input: 0.
|
|
11206
|
-
output:
|
|
11323
|
+
input: 0.5,
|
|
11324
|
+
output: 1.5,
|
|
11207
11325
|
cacheRead: 0,
|
|
11208
11326
|
cacheWrite: 0,
|
|
11209
11327
|
},
|
|
11210
11328
|
contextWindow: 163840,
|
|
11211
|
-
maxTokens:
|
|
11329
|
+
maxTokens: 16384,
|
|
11212
11330
|
},
|
|
11213
11331
|
"deepseek/deepseek-v3.1-terminus": {
|
|
11214
11332
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11221,7 +11339,7 @@ export const MODELS = {
|
|
|
11221
11339
|
cost: {
|
|
11222
11340
|
input: 0.27,
|
|
11223
11341
|
output: 1,
|
|
11224
|
-
cacheRead: 0,
|
|
11342
|
+
cacheRead: 0.135,
|
|
11225
11343
|
cacheWrite: 0,
|
|
11226
11344
|
},
|
|
11227
11345
|
contextWindow: 131072,
|
|
@@ -11236,9 +11354,9 @@ export const MODELS = {
|
|
|
11236
11354
|
reasoning: false,
|
|
11237
11355
|
input: ["text"],
|
|
11238
11356
|
cost: {
|
|
11239
|
-
input: 0.
|
|
11240
|
-
output: 0.
|
|
11241
|
-
cacheRead: 0.
|
|
11357
|
+
input: 0.28,
|
|
11358
|
+
output: 0.42,
|
|
11359
|
+
cacheRead: 0.028,
|
|
11242
11360
|
cacheWrite: 0,
|
|
11243
11361
|
},
|
|
11244
11362
|
contextWindow: 128000,
|
|
@@ -11261,72 +11379,72 @@ export const MODELS = {
|
|
|
11261
11379
|
contextWindow: 128000,
|
|
11262
11380
|
maxTokens: 64000,
|
|
11263
11381
|
},
|
|
11264
|
-
"google/gemini-2.
|
|
11265
|
-
id: "google/gemini-2.
|
|
11266
|
-
name: "Gemini 2.
|
|
11382
|
+
"google/gemini-2.0-flash": {
|
|
11383
|
+
id: "google/gemini-2.0-flash",
|
|
11384
|
+
name: "Gemini 2.0 Flash",
|
|
11267
11385
|
api: "anthropic-messages",
|
|
11268
11386
|
provider: "vercel-ai-gateway",
|
|
11269
11387
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11270
|
-
reasoning:
|
|
11271
|
-
input: ["text"],
|
|
11388
|
+
reasoning: false,
|
|
11389
|
+
input: ["text", "image"],
|
|
11272
11390
|
cost: {
|
|
11273
|
-
input: 0.
|
|
11274
|
-
output:
|
|
11275
|
-
cacheRead: 0,
|
|
11391
|
+
input: 0.15,
|
|
11392
|
+
output: 0.6,
|
|
11393
|
+
cacheRead: 0.024999999999999998,
|
|
11276
11394
|
cacheWrite: 0,
|
|
11277
11395
|
},
|
|
11278
|
-
contextWindow:
|
|
11279
|
-
maxTokens:
|
|
11396
|
+
contextWindow: 1048576,
|
|
11397
|
+
maxTokens: 8192,
|
|
11280
11398
|
},
|
|
11281
|
-
"google/gemini-2.
|
|
11282
|
-
id: "google/gemini-2.
|
|
11283
|
-
name: "Gemini 2.
|
|
11399
|
+
"google/gemini-2.0-flash-lite": {
|
|
11400
|
+
id: "google/gemini-2.0-flash-lite",
|
|
11401
|
+
name: "Gemini 2.0 Flash Lite",
|
|
11284
11402
|
api: "anthropic-messages",
|
|
11285
11403
|
provider: "vercel-ai-gateway",
|
|
11286
11404
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11287
|
-
reasoning:
|
|
11405
|
+
reasoning: false,
|
|
11288
11406
|
input: ["text", "image"],
|
|
11289
11407
|
cost: {
|
|
11290
|
-
input: 0.
|
|
11291
|
-
output: 0.
|
|
11292
|
-
cacheRead: 0.
|
|
11408
|
+
input: 0.075,
|
|
11409
|
+
output: 0.3,
|
|
11410
|
+
cacheRead: 0.02,
|
|
11293
11411
|
cacheWrite: 0,
|
|
11294
11412
|
},
|
|
11295
11413
|
contextWindow: 1048576,
|
|
11296
|
-
maxTokens:
|
|
11414
|
+
maxTokens: 8192,
|
|
11297
11415
|
},
|
|
11298
|
-
"google/gemini-2.5-flash
|
|
11299
|
-
id: "google/gemini-2.5-flash
|
|
11300
|
-
name: "Gemini 2.5 Flash
|
|
11416
|
+
"google/gemini-2.5-flash": {
|
|
11417
|
+
id: "google/gemini-2.5-flash",
|
|
11418
|
+
name: "Gemini 2.5 Flash",
|
|
11301
11419
|
api: "anthropic-messages",
|
|
11302
11420
|
provider: "vercel-ai-gateway",
|
|
11303
11421
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11304
11422
|
reasoning: true,
|
|
11305
11423
|
input: ["text", "image"],
|
|
11306
11424
|
cost: {
|
|
11307
|
-
input: 0.
|
|
11308
|
-
output:
|
|
11309
|
-
cacheRead: 0.
|
|
11425
|
+
input: 0.3,
|
|
11426
|
+
output: 2.5,
|
|
11427
|
+
cacheRead: 0.03,
|
|
11310
11428
|
cacheWrite: 0,
|
|
11311
11429
|
},
|
|
11312
|
-
contextWindow:
|
|
11430
|
+
contextWindow: 1000000,
|
|
11313
11431
|
maxTokens: 65536,
|
|
11314
11432
|
},
|
|
11315
|
-
"google/gemini-2.5-flash-
|
|
11316
|
-
id: "google/gemini-2.5-flash-
|
|
11317
|
-
name: "Gemini 2.5 Flash
|
|
11433
|
+
"google/gemini-2.5-flash-lite": {
|
|
11434
|
+
id: "google/gemini-2.5-flash-lite",
|
|
11435
|
+
name: "Gemini 2.5 Flash Lite",
|
|
11318
11436
|
api: "anthropic-messages",
|
|
11319
11437
|
provider: "vercel-ai-gateway",
|
|
11320
11438
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11321
11439
|
reasoning: true,
|
|
11322
11440
|
input: ["text", "image"],
|
|
11323
11441
|
cost: {
|
|
11324
|
-
input: 0.
|
|
11325
|
-
output:
|
|
11326
|
-
cacheRead: 0.
|
|
11442
|
+
input: 0.09999999999999999,
|
|
11443
|
+
output: 0.39999999999999997,
|
|
11444
|
+
cacheRead: 0.01,
|
|
11327
11445
|
cacheWrite: 0,
|
|
11328
11446
|
},
|
|
11329
|
-
contextWindow:
|
|
11447
|
+
contextWindow: 1048576,
|
|
11330
11448
|
maxTokens: 65536,
|
|
11331
11449
|
},
|
|
11332
11450
|
"google/gemini-2.5-pro": {
|
|
@@ -11336,11 +11454,11 @@ export const MODELS = {
|
|
|
11336
11454
|
provider: "vercel-ai-gateway",
|
|
11337
11455
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11338
11456
|
reasoning: true,
|
|
11339
|
-
input: ["text"],
|
|
11457
|
+
input: ["text", "image"],
|
|
11340
11458
|
cost: {
|
|
11341
11459
|
input: 1.25,
|
|
11342
11460
|
output: 10,
|
|
11343
|
-
cacheRead: 0,
|
|
11461
|
+
cacheRead: 0.125,
|
|
11344
11462
|
cacheWrite: 0,
|
|
11345
11463
|
},
|
|
11346
11464
|
contextWindow: 1048576,
|
|
@@ -11361,7 +11479,7 @@ export const MODELS = {
|
|
|
11361
11479
|
cacheWrite: 0,
|
|
11362
11480
|
},
|
|
11363
11481
|
contextWindow: 1000000,
|
|
11364
|
-
maxTokens:
|
|
11482
|
+
maxTokens: 65000,
|
|
11365
11483
|
},
|
|
11366
11484
|
"google/gemini-3-pro-preview": {
|
|
11367
11485
|
id: "google/gemini-3-pro-preview",
|
|
@@ -11463,7 +11581,7 @@ export const MODELS = {
|
|
|
11463
11581
|
cacheWrite: 0,
|
|
11464
11582
|
},
|
|
11465
11583
|
contextWindow: 128000,
|
|
11466
|
-
maxTokens:
|
|
11584
|
+
maxTokens: 100000,
|
|
11467
11585
|
},
|
|
11468
11586
|
"meituan/longcat-flash-thinking": {
|
|
11469
11587
|
id: "meituan/longcat-flash-thinking",
|
|
@@ -11491,13 +11609,13 @@ export const MODELS = {
|
|
|
11491
11609
|
reasoning: false,
|
|
11492
11610
|
input: ["text"],
|
|
11493
11611
|
cost: {
|
|
11494
|
-
input: 0.
|
|
11495
|
-
output: 0.
|
|
11612
|
+
input: 0.72,
|
|
11613
|
+
output: 0.72,
|
|
11496
11614
|
cacheRead: 0,
|
|
11497
11615
|
cacheWrite: 0,
|
|
11498
11616
|
},
|
|
11499
|
-
contextWindow:
|
|
11500
|
-
maxTokens:
|
|
11617
|
+
contextWindow: 128000,
|
|
11618
|
+
maxTokens: 8192,
|
|
11501
11619
|
},
|
|
11502
11620
|
"meta/llama-3.1-8b": {
|
|
11503
11621
|
id: "meta/llama-3.1-8b",
|
|
@@ -11508,12 +11626,12 @@ export const MODELS = {
|
|
|
11508
11626
|
reasoning: false,
|
|
11509
11627
|
input: ["text"],
|
|
11510
11628
|
cost: {
|
|
11511
|
-
input: 0.
|
|
11512
|
-
output: 0.
|
|
11513
|
-
cacheRead: 0,
|
|
11629
|
+
input: 0.09999999999999999,
|
|
11630
|
+
output: 0.09999999999999999,
|
|
11631
|
+
cacheRead: 0.09999999999999999,
|
|
11514
11632
|
cacheWrite: 0,
|
|
11515
11633
|
},
|
|
11516
|
-
contextWindow:
|
|
11634
|
+
contextWindow: 128000,
|
|
11517
11635
|
maxTokens: 16384,
|
|
11518
11636
|
},
|
|
11519
11637
|
"meta/llama-3.2-11b": {
|
|
@@ -11576,12 +11694,12 @@ export const MODELS = {
|
|
|
11576
11694
|
reasoning: false,
|
|
11577
11695
|
input: ["text", "image"],
|
|
11578
11696
|
cost: {
|
|
11579
|
-
input: 0.
|
|
11580
|
-
output: 0.
|
|
11697
|
+
input: 0.24,
|
|
11698
|
+
output: 0.9700000000000001,
|
|
11581
11699
|
cacheRead: 0,
|
|
11582
11700
|
cacheWrite: 0,
|
|
11583
11701
|
},
|
|
11584
|
-
contextWindow:
|
|
11702
|
+
contextWindow: 128000,
|
|
11585
11703
|
maxTokens: 8192,
|
|
11586
11704
|
},
|
|
11587
11705
|
"meta/llama-4-scout": {
|
|
@@ -11593,12 +11711,12 @@ export const MODELS = {
|
|
|
11593
11711
|
reasoning: false,
|
|
11594
11712
|
input: ["text", "image"],
|
|
11595
11713
|
cost: {
|
|
11596
|
-
input: 0.
|
|
11597
|
-
output: 0.
|
|
11714
|
+
input: 0.16999999999999998,
|
|
11715
|
+
output: 0.66,
|
|
11598
11716
|
cacheRead: 0,
|
|
11599
11717
|
cacheWrite: 0,
|
|
11600
11718
|
},
|
|
11601
|
-
contextWindow:
|
|
11719
|
+
contextWindow: 128000,
|
|
11602
11720
|
maxTokens: 8192,
|
|
11603
11721
|
},
|
|
11604
11722
|
"minimax/minimax-m2": {
|
|
@@ -11629,8 +11747,8 @@ export const MODELS = {
|
|
|
11629
11747
|
cost: {
|
|
11630
11748
|
input: 0.3,
|
|
11631
11749
|
output: 1.2,
|
|
11632
|
-
cacheRead: 0.
|
|
11633
|
-
cacheWrite: 0,
|
|
11750
|
+
cacheRead: 0.03,
|
|
11751
|
+
cacheWrite: 0.375,
|
|
11634
11752
|
},
|
|
11635
11753
|
contextWindow: 204800,
|
|
11636
11754
|
maxTokens: 131072,
|
|
@@ -11683,8 +11801,42 @@ export const MODELS = {
|
|
|
11683
11801
|
cacheRead: 0.03,
|
|
11684
11802
|
cacheWrite: 0.375,
|
|
11685
11803
|
},
|
|
11686
|
-
contextWindow:
|
|
11687
|
-
maxTokens:
|
|
11804
|
+
contextWindow: 204800,
|
|
11805
|
+
maxTokens: 131000,
|
|
11806
|
+
},
|
|
11807
|
+
"minimax/minimax-m2.7": {
|
|
11808
|
+
id: "minimax/minimax-m2.7",
|
|
11809
|
+
name: "Minimax M2.7",
|
|
11810
|
+
api: "anthropic-messages",
|
|
11811
|
+
provider: "vercel-ai-gateway",
|
|
11812
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11813
|
+
reasoning: true,
|
|
11814
|
+
input: ["text", "image"],
|
|
11815
|
+
cost: {
|
|
11816
|
+
input: 0.3,
|
|
11817
|
+
output: 1.2,
|
|
11818
|
+
cacheRead: 0.06,
|
|
11819
|
+
cacheWrite: 0.375,
|
|
11820
|
+
},
|
|
11821
|
+
contextWindow: 204800,
|
|
11822
|
+
maxTokens: 131000,
|
|
11823
|
+
},
|
|
11824
|
+
"minimax/minimax-m2.7-highspeed": {
|
|
11825
|
+
id: "minimax/minimax-m2.7-highspeed",
|
|
11826
|
+
name: "MiniMax M2.7 High Speed",
|
|
11827
|
+
api: "anthropic-messages",
|
|
11828
|
+
provider: "vercel-ai-gateway",
|
|
11829
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11830
|
+
reasoning: true,
|
|
11831
|
+
input: ["text", "image"],
|
|
11832
|
+
cost: {
|
|
11833
|
+
input: 0.6,
|
|
11834
|
+
output: 2.4,
|
|
11835
|
+
cacheRead: 0.06,
|
|
11836
|
+
cacheWrite: 0.375,
|
|
11837
|
+
},
|
|
11838
|
+
contextWindow: 204800,
|
|
11839
|
+
maxTokens: 131100,
|
|
11688
11840
|
},
|
|
11689
11841
|
"mistral/codestral": {
|
|
11690
11842
|
id: "mistral/codestral",
|
|
@@ -11712,8 +11864,8 @@ export const MODELS = {
|
|
|
11712
11864
|
reasoning: false,
|
|
11713
11865
|
input: ["text"],
|
|
11714
11866
|
cost: {
|
|
11715
|
-
input: 0,
|
|
11716
|
-
output:
|
|
11867
|
+
input: 0.39999999999999997,
|
|
11868
|
+
output: 2,
|
|
11717
11869
|
cacheRead: 0,
|
|
11718
11870
|
cacheWrite: 0,
|
|
11719
11871
|
},
|
|
@@ -11746,8 +11898,8 @@ export const MODELS = {
|
|
|
11746
11898
|
reasoning: false,
|
|
11747
11899
|
input: ["text"],
|
|
11748
11900
|
cost: {
|
|
11749
|
-
input: 0,
|
|
11750
|
-
output: 0,
|
|
11901
|
+
input: 0.09999999999999999,
|
|
11902
|
+
output: 0.3,
|
|
11751
11903
|
cacheRead: 0,
|
|
11752
11904
|
cacheWrite: 0,
|
|
11753
11905
|
},
|
|
@@ -11763,8 +11915,8 @@ export const MODELS = {
|
|
|
11763
11915
|
reasoning: false,
|
|
11764
11916
|
input: ["text"],
|
|
11765
11917
|
cost: {
|
|
11766
|
-
input: 0.
|
|
11767
|
-
output: 0.
|
|
11918
|
+
input: 0.09999999999999999,
|
|
11919
|
+
output: 0.09999999999999999,
|
|
11768
11920
|
cacheRead: 0,
|
|
11769
11921
|
cacheWrite: 0,
|
|
11770
11922
|
},
|
|
@@ -11780,8 +11932,8 @@ export const MODELS = {
|
|
|
11780
11932
|
reasoning: false,
|
|
11781
11933
|
input: ["text"],
|
|
11782
11934
|
cost: {
|
|
11783
|
-
input: 0.
|
|
11784
|
-
output: 0.
|
|
11935
|
+
input: 0.15,
|
|
11936
|
+
output: 0.15,
|
|
11785
11937
|
cacheRead: 0,
|
|
11786
11938
|
cacheWrite: 0,
|
|
11787
11939
|
},
|
|
@@ -11865,14 +12017,31 @@ export const MODELS = {
|
|
|
11865
12017
|
reasoning: false,
|
|
11866
12018
|
input: ["text"],
|
|
11867
12019
|
cost: {
|
|
11868
|
-
input: 0.
|
|
11869
|
-
output: 2,
|
|
11870
|
-
cacheRead: 0,
|
|
12020
|
+
input: 0.6,
|
|
12021
|
+
output: 2.5,
|
|
12022
|
+
cacheRead: 0.15,
|
|
11871
12023
|
cacheWrite: 0,
|
|
11872
12024
|
},
|
|
11873
12025
|
contextWindow: 131072,
|
|
11874
12026
|
maxTokens: 16384,
|
|
11875
12027
|
},
|
|
12028
|
+
"moonshotai/kimi-k2-0905": {
|
|
12029
|
+
id: "moonshotai/kimi-k2-0905",
|
|
12030
|
+
name: "Kimi K2 0905",
|
|
12031
|
+
api: "anthropic-messages",
|
|
12032
|
+
provider: "vercel-ai-gateway",
|
|
12033
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12034
|
+
reasoning: false,
|
|
12035
|
+
input: ["text"],
|
|
12036
|
+
cost: {
|
|
12037
|
+
input: 0.6,
|
|
12038
|
+
output: 2.5,
|
|
12039
|
+
cacheRead: 0.15,
|
|
12040
|
+
cacheWrite: 0,
|
|
12041
|
+
},
|
|
12042
|
+
contextWindow: 256000,
|
|
12043
|
+
maxTokens: 16384,
|
|
12044
|
+
},
|
|
11876
12045
|
"moonshotai/kimi-k2-thinking": {
|
|
11877
12046
|
id: "moonshotai/kimi-k2-thinking",
|
|
11878
12047
|
name: "Kimi K2 Thinking",
|
|
@@ -11882,13 +12051,13 @@ export const MODELS = {
|
|
|
11882
12051
|
reasoning: true,
|
|
11883
12052
|
input: ["text"],
|
|
11884
12053
|
cost: {
|
|
11885
|
-
input: 0.
|
|
11886
|
-
output: 2,
|
|
11887
|
-
cacheRead: 0.
|
|
12054
|
+
input: 0.6,
|
|
12055
|
+
output: 2.5,
|
|
12056
|
+
cacheRead: 0.15,
|
|
11888
12057
|
cacheWrite: 0,
|
|
11889
12058
|
},
|
|
11890
|
-
contextWindow:
|
|
11891
|
-
maxTokens:
|
|
12059
|
+
contextWindow: 262114,
|
|
12060
|
+
maxTokens: 262114,
|
|
11892
12061
|
},
|
|
11893
12062
|
"moonshotai/kimi-k2-thinking-turbo": {
|
|
11894
12063
|
id: "moonshotai/kimi-k2-thinking-turbo",
|
|
@@ -11916,9 +12085,9 @@ export const MODELS = {
|
|
|
11916
12085
|
reasoning: false,
|
|
11917
12086
|
input: ["text"],
|
|
11918
12087
|
cost: {
|
|
11919
|
-
input:
|
|
11920
|
-
output:
|
|
11921
|
-
cacheRead: 0,
|
|
12088
|
+
input: 1.15,
|
|
12089
|
+
output: 8,
|
|
12090
|
+
cacheRead: 0.15,
|
|
11922
12091
|
cacheWrite: 0,
|
|
11923
12092
|
},
|
|
11924
12093
|
contextWindow: 256000,
|
|
@@ -11933,13 +12102,13 @@ export const MODELS = {
|
|
|
11933
12102
|
reasoning: true,
|
|
11934
12103
|
input: ["text", "image"],
|
|
11935
12104
|
cost: {
|
|
11936
|
-
input: 0.
|
|
11937
|
-
output:
|
|
11938
|
-
cacheRead: 0,
|
|
12105
|
+
input: 0.6,
|
|
12106
|
+
output: 3,
|
|
12107
|
+
cacheRead: 0.09999999999999999,
|
|
11939
12108
|
cacheWrite: 0,
|
|
11940
12109
|
},
|
|
11941
|
-
contextWindow:
|
|
11942
|
-
maxTokens:
|
|
12110
|
+
contextWindow: 262114,
|
|
12111
|
+
maxTokens: 262114,
|
|
11943
12112
|
},
|
|
11944
12113
|
"nvidia/nemotron-nano-12b-v2-vl": {
|
|
11945
12114
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
@@ -11967,31 +12136,14 @@ export const MODELS = {
|
|
|
11967
12136
|
reasoning: true,
|
|
11968
12137
|
input: ["text"],
|
|
11969
12138
|
cost: {
|
|
11970
|
-
input: 0.
|
|
11971
|
-
output: 0.
|
|
12139
|
+
input: 0.06,
|
|
12140
|
+
output: 0.22999999999999998,
|
|
11972
12141
|
cacheRead: 0,
|
|
11973
12142
|
cacheWrite: 0,
|
|
11974
12143
|
},
|
|
11975
12144
|
contextWindow: 131072,
|
|
11976
12145
|
maxTokens: 131072,
|
|
11977
12146
|
},
|
|
11978
|
-
"openai/codex-mini": {
|
|
11979
|
-
id: "openai/codex-mini",
|
|
11980
|
-
name: "Codex Mini",
|
|
11981
|
-
api: "anthropic-messages",
|
|
11982
|
-
provider: "vercel-ai-gateway",
|
|
11983
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11984
|
-
reasoning: true,
|
|
11985
|
-
input: ["text", "image"],
|
|
11986
|
-
cost: {
|
|
11987
|
-
input: 1.5,
|
|
11988
|
-
output: 6,
|
|
11989
|
-
cacheRead: 0.375,
|
|
11990
|
-
cacheWrite: 0,
|
|
11991
|
-
},
|
|
11992
|
-
contextWindow: 200000,
|
|
11993
|
-
maxTokens: 100000,
|
|
11994
|
-
},
|
|
11995
12147
|
"openai/gpt-4-turbo": {
|
|
11996
12148
|
id: "openai/gpt-4-turbo",
|
|
11997
12149
|
name: "GPT-4 Turbo",
|
|
@@ -12054,7 +12206,7 @@ export const MODELS = {
|
|
|
12054
12206
|
cost: {
|
|
12055
12207
|
input: 0.09999999999999999,
|
|
12056
12208
|
output: 0.39999999999999997,
|
|
12057
|
-
cacheRead: 0.
|
|
12209
|
+
cacheRead: 0.024999999999999998,
|
|
12058
12210
|
cacheWrite: 0,
|
|
12059
12211
|
},
|
|
12060
12212
|
contextWindow: 1047576,
|
|
@@ -12105,7 +12257,7 @@ export const MODELS = {
|
|
|
12105
12257
|
cost: {
|
|
12106
12258
|
input: 1.25,
|
|
12107
12259
|
output: 10,
|
|
12108
|
-
cacheRead: 0.
|
|
12260
|
+
cacheRead: 0.125,
|
|
12109
12261
|
cacheWrite: 0,
|
|
12110
12262
|
},
|
|
12111
12263
|
contextWindow: 400000,
|
|
@@ -12135,11 +12287,11 @@ export const MODELS = {
|
|
|
12135
12287
|
provider: "vercel-ai-gateway",
|
|
12136
12288
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12137
12289
|
reasoning: true,
|
|
12138
|
-
input: ["text"
|
|
12290
|
+
input: ["text"],
|
|
12139
12291
|
cost: {
|
|
12140
12292
|
input: 1.25,
|
|
12141
12293
|
output: 10,
|
|
12142
|
-
cacheRead: 0.
|
|
12294
|
+
cacheRead: 0.125,
|
|
12143
12295
|
cacheWrite: 0,
|
|
12144
12296
|
},
|
|
12145
12297
|
contextWindow: 400000,
|
|
@@ -12156,7 +12308,7 @@ export const MODELS = {
|
|
|
12156
12308
|
cost: {
|
|
12157
12309
|
input: 0.25,
|
|
12158
12310
|
output: 2,
|
|
12159
|
-
cacheRead: 0.
|
|
12311
|
+
cacheRead: 0.024999999999999998,
|
|
12160
12312
|
cacheWrite: 0,
|
|
12161
12313
|
},
|
|
12162
12314
|
contextWindow: 400000,
|
|
@@ -12173,7 +12325,7 @@ export const MODELS = {
|
|
|
12173
12325
|
cost: {
|
|
12174
12326
|
input: 0.049999999999999996,
|
|
12175
12327
|
output: 0.39999999999999997,
|
|
12176
|
-
cacheRead: 0.
|
|
12328
|
+
cacheRead: 0.005,
|
|
12177
12329
|
cacheWrite: 0,
|
|
12178
12330
|
},
|
|
12179
12331
|
contextWindow: 400000,
|
|
@@ -12207,7 +12359,7 @@ export const MODELS = {
|
|
|
12207
12359
|
cost: {
|
|
12208
12360
|
input: 1.25,
|
|
12209
12361
|
output: 10,
|
|
12210
|
-
cacheRead: 0.
|
|
12362
|
+
cacheRead: 0.125,
|
|
12211
12363
|
cacheWrite: 0,
|
|
12212
12364
|
},
|
|
12213
12365
|
contextWindow: 400000,
|
|
@@ -12258,7 +12410,7 @@ export const MODELS = {
|
|
|
12258
12410
|
cost: {
|
|
12259
12411
|
input: 1.25,
|
|
12260
12412
|
output: 10,
|
|
12261
|
-
cacheRead: 0.
|
|
12413
|
+
cacheRead: 0.125,
|
|
12262
12414
|
cacheWrite: 0,
|
|
12263
12415
|
},
|
|
12264
12416
|
contextWindow: 128000,
|
|
@@ -12275,7 +12427,7 @@ export const MODELS = {
|
|
|
12275
12427
|
cost: {
|
|
12276
12428
|
input: 1.25,
|
|
12277
12429
|
output: 10,
|
|
12278
|
-
cacheRead: 0.
|
|
12430
|
+
cacheRead: 0.125,
|
|
12279
12431
|
cacheWrite: 0,
|
|
12280
12432
|
},
|
|
12281
12433
|
contextWindow: 400000,
|
|
@@ -12292,7 +12444,7 @@ export const MODELS = {
|
|
|
12292
12444
|
cost: {
|
|
12293
12445
|
input: 1.75,
|
|
12294
12446
|
output: 14,
|
|
12295
|
-
cacheRead: 0.
|
|
12447
|
+
cacheRead: 0.175,
|
|
12296
12448
|
cacheWrite: 0,
|
|
12297
12449
|
},
|
|
12298
12450
|
contextWindow: 400000,
|
|
@@ -12397,42 +12549,59 @@ export const MODELS = {
|
|
|
12397
12549
|
cacheRead: 0.25,
|
|
12398
12550
|
cacheWrite: 0,
|
|
12399
12551
|
},
|
|
12400
|
-
contextWindow:
|
|
12552
|
+
contextWindow: 1050000,
|
|
12401
12553
|
maxTokens: 128000,
|
|
12402
12554
|
},
|
|
12403
|
-
"openai/gpt-5.4-
|
|
12404
|
-
id: "openai/gpt-5.4-
|
|
12405
|
-
name: "GPT 5.4
|
|
12555
|
+
"openai/gpt-5.4-mini": {
|
|
12556
|
+
id: "openai/gpt-5.4-mini",
|
|
12557
|
+
name: "GPT 5.4 Mini",
|
|
12406
12558
|
api: "anthropic-messages",
|
|
12407
12559
|
provider: "vercel-ai-gateway",
|
|
12408
12560
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12409
12561
|
reasoning: true,
|
|
12410
12562
|
input: ["text", "image"],
|
|
12411
12563
|
cost: {
|
|
12412
|
-
input:
|
|
12413
|
-
output:
|
|
12414
|
-
cacheRead: 0,
|
|
12564
|
+
input: 0.75,
|
|
12565
|
+
output: 4.5,
|
|
12566
|
+
cacheRead: 0.075,
|
|
12415
12567
|
cacheWrite: 0,
|
|
12416
12568
|
},
|
|
12417
|
-
contextWindow:
|
|
12569
|
+
contextWindow: 400000,
|
|
12418
12570
|
maxTokens: 128000,
|
|
12419
12571
|
},
|
|
12420
|
-
"openai/gpt-
|
|
12421
|
-
id: "openai/gpt-
|
|
12422
|
-
name: "
|
|
12572
|
+
"openai/gpt-5.4-nano": {
|
|
12573
|
+
id: "openai/gpt-5.4-nano",
|
|
12574
|
+
name: "GPT 5.4 Nano",
|
|
12423
12575
|
api: "anthropic-messages",
|
|
12424
12576
|
provider: "vercel-ai-gateway",
|
|
12425
12577
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12426
12578
|
reasoning: true,
|
|
12427
|
-
input: ["text"],
|
|
12579
|
+
input: ["text", "image"],
|
|
12428
12580
|
cost: {
|
|
12429
|
-
input: 0.
|
|
12430
|
-
output:
|
|
12581
|
+
input: 0.19999999999999998,
|
|
12582
|
+
output: 1.25,
|
|
12583
|
+
cacheRead: 0.02,
|
|
12584
|
+
cacheWrite: 0,
|
|
12585
|
+
},
|
|
12586
|
+
contextWindow: 400000,
|
|
12587
|
+
maxTokens: 128000,
|
|
12588
|
+
},
|
|
12589
|
+
"openai/gpt-5.4-pro": {
|
|
12590
|
+
id: "openai/gpt-5.4-pro",
|
|
12591
|
+
name: "GPT 5.4 Pro",
|
|
12592
|
+
api: "anthropic-messages",
|
|
12593
|
+
provider: "vercel-ai-gateway",
|
|
12594
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12595
|
+
reasoning: true,
|
|
12596
|
+
input: ["text", "image"],
|
|
12597
|
+
cost: {
|
|
12598
|
+
input: 30,
|
|
12599
|
+
output: 180,
|
|
12431
12600
|
cacheRead: 0,
|
|
12432
12601
|
cacheWrite: 0,
|
|
12433
12602
|
},
|
|
12434
|
-
contextWindow:
|
|
12435
|
-
maxTokens:
|
|
12603
|
+
contextWindow: 1050000,
|
|
12604
|
+
maxTokens: 128000,
|
|
12436
12605
|
},
|
|
12437
12606
|
"openai/gpt-oss-20b": {
|
|
12438
12607
|
id: "openai/gpt-oss-20b",
|
|
@@ -12621,40 +12790,6 @@ export const MODELS = {
|
|
|
12621
12790
|
contextWindow: 131072,
|
|
12622
12791
|
maxTokens: 131072,
|
|
12623
12792
|
},
|
|
12624
|
-
"vercel/v0-1.0-md": {
|
|
12625
|
-
id: "vercel/v0-1.0-md",
|
|
12626
|
-
name: "v0-1.0-md",
|
|
12627
|
-
api: "anthropic-messages",
|
|
12628
|
-
provider: "vercel-ai-gateway",
|
|
12629
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12630
|
-
reasoning: false,
|
|
12631
|
-
input: ["text", "image"],
|
|
12632
|
-
cost: {
|
|
12633
|
-
input: 3,
|
|
12634
|
-
output: 15,
|
|
12635
|
-
cacheRead: 0,
|
|
12636
|
-
cacheWrite: 0,
|
|
12637
|
-
},
|
|
12638
|
-
contextWindow: 128000,
|
|
12639
|
-
maxTokens: 32000,
|
|
12640
|
-
},
|
|
12641
|
-
"vercel/v0-1.5-md": {
|
|
12642
|
-
id: "vercel/v0-1.5-md",
|
|
12643
|
-
name: "v0-1.5-md",
|
|
12644
|
-
api: "anthropic-messages",
|
|
12645
|
-
provider: "vercel-ai-gateway",
|
|
12646
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12647
|
-
reasoning: false,
|
|
12648
|
-
input: ["text", "image"],
|
|
12649
|
-
cost: {
|
|
12650
|
-
input: 3,
|
|
12651
|
-
output: 15,
|
|
12652
|
-
cacheRead: 0,
|
|
12653
|
-
cacheWrite: 0,
|
|
12654
|
-
},
|
|
12655
|
-
contextWindow: 128000,
|
|
12656
|
-
maxTokens: 32768,
|
|
12657
|
-
},
|
|
12658
12793
|
"xai/grok-2-vision": {
|
|
12659
12794
|
id: "xai/grok-2-vision",
|
|
12660
12795
|
name: "Grok 2 Vision",
|
|
@@ -12683,7 +12818,7 @@ export const MODELS = {
|
|
|
12683
12818
|
cost: {
|
|
12684
12819
|
input: 3,
|
|
12685
12820
|
output: 15,
|
|
12686
|
-
cacheRead: 0,
|
|
12821
|
+
cacheRead: 0.75,
|
|
12687
12822
|
cacheWrite: 0,
|
|
12688
12823
|
},
|
|
12689
12824
|
contextWindow: 131072,
|
|
@@ -12700,7 +12835,7 @@ export const MODELS = {
|
|
|
12700
12835
|
cost: {
|
|
12701
12836
|
input: 5,
|
|
12702
12837
|
output: 25,
|
|
12703
|
-
cacheRead:
|
|
12838
|
+
cacheRead: 1.25,
|
|
12704
12839
|
cacheWrite: 0,
|
|
12705
12840
|
},
|
|
12706
12841
|
contextWindow: 131072,
|
|
@@ -12717,7 +12852,7 @@ export const MODELS = {
|
|
|
12717
12852
|
cost: {
|
|
12718
12853
|
input: 0.3,
|
|
12719
12854
|
output: 0.5,
|
|
12720
|
-
cacheRead: 0,
|
|
12855
|
+
cacheRead: 0.075,
|
|
12721
12856
|
cacheWrite: 0,
|
|
12722
12857
|
},
|
|
12723
12858
|
contextWindow: 131072,
|
|
@@ -12751,7 +12886,7 @@ export const MODELS = {
|
|
|
12751
12886
|
cost: {
|
|
12752
12887
|
input: 3,
|
|
12753
12888
|
output: 15,
|
|
12754
|
-
cacheRead: 0,
|
|
12889
|
+
cacheRead: 0.75,
|
|
12755
12890
|
cacheWrite: 0,
|
|
12756
12891
|
},
|
|
12757
12892
|
contextWindow: 256000,
|
|
@@ -12825,6 +12960,57 @@ export const MODELS = {
|
|
|
12825
12960
|
contextWindow: 2000000,
|
|
12826
12961
|
maxTokens: 30000,
|
|
12827
12962
|
},
|
|
12963
|
+
"xai/grok-4.20-multi-agent-beta": {
|
|
12964
|
+
id: "xai/grok-4.20-multi-agent-beta",
|
|
12965
|
+
name: "Grok 4.20 Multi Agent Beta",
|
|
12966
|
+
api: "anthropic-messages",
|
|
12967
|
+
provider: "vercel-ai-gateway",
|
|
12968
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12969
|
+
reasoning: true,
|
|
12970
|
+
input: ["text"],
|
|
12971
|
+
cost: {
|
|
12972
|
+
input: 2,
|
|
12973
|
+
output: 6,
|
|
12974
|
+
cacheRead: 0.19999999999999998,
|
|
12975
|
+
cacheWrite: 0,
|
|
12976
|
+
},
|
|
12977
|
+
contextWindow: 2000000,
|
|
12978
|
+
maxTokens: 2000000,
|
|
12979
|
+
},
|
|
12980
|
+
"xai/grok-4.20-non-reasoning-beta": {
|
|
12981
|
+
id: "xai/grok-4.20-non-reasoning-beta",
|
|
12982
|
+
name: "Grok 4.20 Beta Non-Reasoning",
|
|
12983
|
+
api: "anthropic-messages",
|
|
12984
|
+
provider: "vercel-ai-gateway",
|
|
12985
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12986
|
+
reasoning: false,
|
|
12987
|
+
input: ["text", "image"],
|
|
12988
|
+
cost: {
|
|
12989
|
+
input: 2,
|
|
12990
|
+
output: 6,
|
|
12991
|
+
cacheRead: 0.19999999999999998,
|
|
12992
|
+
cacheWrite: 0,
|
|
12993
|
+
},
|
|
12994
|
+
contextWindow: 2000000,
|
|
12995
|
+
maxTokens: 2000000,
|
|
12996
|
+
},
|
|
12997
|
+
"xai/grok-4.20-reasoning-beta": {
|
|
12998
|
+
id: "xai/grok-4.20-reasoning-beta",
|
|
12999
|
+
name: "Grok 4.20 Beta Reasoning",
|
|
13000
|
+
api: "anthropic-messages",
|
|
13001
|
+
provider: "vercel-ai-gateway",
|
|
13002
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13003
|
+
reasoning: true,
|
|
13004
|
+
input: ["text", "image"],
|
|
13005
|
+
cost: {
|
|
13006
|
+
input: 2,
|
|
13007
|
+
output: 6,
|
|
13008
|
+
cacheRead: 0.19999999999999998,
|
|
13009
|
+
cacheWrite: 0,
|
|
13010
|
+
},
|
|
13011
|
+
contextWindow: 2000000,
|
|
13012
|
+
maxTokens: 2000000,
|
|
13013
|
+
},
|
|
12828
13014
|
"xai/grok-code-fast-1": {
|
|
12829
13015
|
id: "xai/grok-code-fast-1",
|
|
12830
13016
|
name: "Grok Code Fast 1",
|
|
@@ -12851,17 +13037,34 @@ export const MODELS = {
|
|
|
12851
13037
|
reasoning: true,
|
|
12852
13038
|
input: ["text"],
|
|
12853
13039
|
cost: {
|
|
12854
|
-
input: 0.
|
|
12855
|
-
output: 0.
|
|
12856
|
-
cacheRead: 0,
|
|
13040
|
+
input: 0.09999999999999999,
|
|
13041
|
+
output: 0.3,
|
|
13042
|
+
cacheRead: 0.02,
|
|
12857
13043
|
cacheWrite: 0,
|
|
12858
13044
|
},
|
|
12859
13045
|
contextWindow: 262144,
|
|
12860
13046
|
maxTokens: 32000,
|
|
12861
13047
|
},
|
|
12862
|
-
"
|
|
12863
|
-
id: "
|
|
12864
|
-
name: "
|
|
13048
|
+
"xiaomi/mimo-v2-pro": {
|
|
13049
|
+
id: "xiaomi/mimo-v2-pro",
|
|
13050
|
+
name: "MiMo V2 Pro",
|
|
13051
|
+
api: "anthropic-messages",
|
|
13052
|
+
provider: "vercel-ai-gateway",
|
|
13053
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13054
|
+
reasoning: true,
|
|
13055
|
+
input: ["text"],
|
|
13056
|
+
cost: {
|
|
13057
|
+
input: 1,
|
|
13058
|
+
output: 3,
|
|
13059
|
+
cacheRead: 0.19999999999999998,
|
|
13060
|
+
cacheWrite: 0,
|
|
13061
|
+
},
|
|
13062
|
+
contextWindow: 1000000,
|
|
13063
|
+
maxTokens: 128000,
|
|
13064
|
+
},
|
|
13065
|
+
"zai/glm-4.5": {
|
|
13066
|
+
id: "zai/glm-4.5",
|
|
13067
|
+
name: "GLM-4.5",
|
|
12865
13068
|
api: "anthropic-messages",
|
|
12866
13069
|
provider: "vercel-ai-gateway",
|
|
12867
13070
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -12870,11 +13073,11 @@ export const MODELS = {
|
|
|
12870
13073
|
cost: {
|
|
12871
13074
|
input: 0.6,
|
|
12872
13075
|
output: 2.2,
|
|
12873
|
-
cacheRead: 0,
|
|
13076
|
+
cacheRead: 0.11,
|
|
12874
13077
|
cacheWrite: 0,
|
|
12875
13078
|
},
|
|
12876
|
-
contextWindow:
|
|
12877
|
-
maxTokens:
|
|
13079
|
+
contextWindow: 128000,
|
|
13080
|
+
maxTokens: 96000,
|
|
12878
13081
|
},
|
|
12879
13082
|
"zai/glm-4.5-air": {
|
|
12880
13083
|
id: "zai/glm-4.5-air",
|
|
@@ -12899,16 +13102,16 @@ export const MODELS = {
|
|
|
12899
13102
|
api: "anthropic-messages",
|
|
12900
13103
|
provider: "vercel-ai-gateway",
|
|
12901
13104
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12902
|
-
reasoning:
|
|
13105
|
+
reasoning: false,
|
|
12903
13106
|
input: ["text", "image"],
|
|
12904
13107
|
cost: {
|
|
12905
13108
|
input: 0.6,
|
|
12906
13109
|
output: 1.7999999999999998,
|
|
12907
|
-
cacheRead: 0,
|
|
13110
|
+
cacheRead: 0.11,
|
|
12908
13111
|
cacheWrite: 0,
|
|
12909
13112
|
},
|
|
12910
|
-
contextWindow:
|
|
12911
|
-
maxTokens:
|
|
13113
|
+
contextWindow: 66000,
|
|
13114
|
+
maxTokens: 16000,
|
|
12912
13115
|
},
|
|
12913
13116
|
"zai/glm-4.6": {
|
|
12914
13117
|
id: "zai/glm-4.6",
|
|
@@ -12919,8 +13122,8 @@ export const MODELS = {
|
|
|
12919
13122
|
reasoning: true,
|
|
12920
13123
|
input: ["text"],
|
|
12921
13124
|
cost: {
|
|
12922
|
-
input: 0.
|
|
12923
|
-
output:
|
|
13125
|
+
input: 0.6,
|
|
13126
|
+
output: 2.2,
|
|
12924
13127
|
cacheRead: 0.11,
|
|
12925
13128
|
cacheWrite: 0,
|
|
12926
13129
|
},
|
|
@@ -12970,14 +13173,31 @@ export const MODELS = {
|
|
|
12970
13173
|
reasoning: true,
|
|
12971
13174
|
input: ["text"],
|
|
12972
13175
|
cost: {
|
|
12973
|
-
input: 0.
|
|
12974
|
-
output:
|
|
12975
|
-
cacheRead: 0
|
|
13176
|
+
input: 0.6,
|
|
13177
|
+
output: 2.2,
|
|
13178
|
+
cacheRead: 0,
|
|
12976
13179
|
cacheWrite: 0,
|
|
12977
13180
|
},
|
|
12978
|
-
contextWindow:
|
|
13181
|
+
contextWindow: 200000,
|
|
12979
13182
|
maxTokens: 120000,
|
|
12980
13183
|
},
|
|
13184
|
+
"zai/glm-4.7-flash": {
|
|
13185
|
+
id: "zai/glm-4.7-flash",
|
|
13186
|
+
name: "GLM 4.7 Flash",
|
|
13187
|
+
api: "anthropic-messages",
|
|
13188
|
+
provider: "vercel-ai-gateway",
|
|
13189
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13190
|
+
reasoning: true,
|
|
13191
|
+
input: ["text"],
|
|
13192
|
+
cost: {
|
|
13193
|
+
input: 0.07,
|
|
13194
|
+
output: 0.39999999999999997,
|
|
13195
|
+
cacheRead: 0,
|
|
13196
|
+
cacheWrite: 0,
|
|
13197
|
+
},
|
|
13198
|
+
contextWindow: 200000,
|
|
13199
|
+
maxTokens: 131000,
|
|
13200
|
+
},
|
|
12981
13201
|
"zai/glm-4.7-flashx": {
|
|
12982
13202
|
id: "zai/glm-4.7-flashx",
|
|
12983
13203
|
name: "GLM 4.7 FlashX",
|
|
@@ -12997,7 +13217,7 @@ export const MODELS = {
|
|
|
12997
13217
|
},
|
|
12998
13218
|
"zai/glm-5": {
|
|
12999
13219
|
id: "zai/glm-5",
|
|
13000
|
-
name: "GLM
|
|
13220
|
+
name: "GLM 5",
|
|
13001
13221
|
api: "anthropic-messages",
|
|
13002
13222
|
provider: "vercel-ai-gateway",
|
|
13003
13223
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -13010,7 +13230,24 @@ export const MODELS = {
|
|
|
13010
13230
|
cacheWrite: 0,
|
|
13011
13231
|
},
|
|
13012
13232
|
contextWindow: 202800,
|
|
13013
|
-
maxTokens:
|
|
13233
|
+
maxTokens: 131100,
|
|
13234
|
+
},
|
|
13235
|
+
"zai/glm-5-turbo": {
|
|
13236
|
+
id: "zai/glm-5-turbo",
|
|
13237
|
+
name: "GLM 5 Turbo",
|
|
13238
|
+
api: "anthropic-messages",
|
|
13239
|
+
provider: "vercel-ai-gateway",
|
|
13240
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13241
|
+
reasoning: true,
|
|
13242
|
+
input: ["text"],
|
|
13243
|
+
cost: {
|
|
13244
|
+
input: 1.2,
|
|
13245
|
+
output: 4,
|
|
13246
|
+
cacheRead: 0.24,
|
|
13247
|
+
cacheWrite: 0,
|
|
13248
|
+
},
|
|
13249
|
+
contextWindow: 202800,
|
|
13250
|
+
maxTokens: 131100,
|
|
13014
13251
|
},
|
|
13015
13252
|
},
|
|
13016
13253
|
"xai": {
|
|
@@ -13337,6 +13574,40 @@ export const MODELS = {
|
|
|
13337
13574
|
contextWindow: 2000000,
|
|
13338
13575
|
maxTokens: 30000,
|
|
13339
13576
|
},
|
|
13577
|
+
"grok-4.20-0309-non-reasoning": {
|
|
13578
|
+
id: "grok-4.20-0309-non-reasoning",
|
|
13579
|
+
name: "Grok 4.20 (Non-Reasoning)",
|
|
13580
|
+
api: "openai-completions",
|
|
13581
|
+
provider: "xai",
|
|
13582
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13583
|
+
reasoning: false,
|
|
13584
|
+
input: ["text", "image"],
|
|
13585
|
+
cost: {
|
|
13586
|
+
input: 2,
|
|
13587
|
+
output: 6,
|
|
13588
|
+
cacheRead: 0.2,
|
|
13589
|
+
cacheWrite: 0,
|
|
13590
|
+
},
|
|
13591
|
+
contextWindow: 2000000,
|
|
13592
|
+
maxTokens: 30000,
|
|
13593
|
+
},
|
|
13594
|
+
"grok-4.20-0309-reasoning": {
|
|
13595
|
+
id: "grok-4.20-0309-reasoning",
|
|
13596
|
+
name: "Grok 4.20 (Reasoning)",
|
|
13597
|
+
api: "openai-completions",
|
|
13598
|
+
provider: "xai",
|
|
13599
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13600
|
+
reasoning: true,
|
|
13601
|
+
input: ["text", "image"],
|
|
13602
|
+
cost: {
|
|
13603
|
+
input: 2,
|
|
13604
|
+
output: 6,
|
|
13605
|
+
cacheRead: 0.2,
|
|
13606
|
+
cacheWrite: 0,
|
|
13607
|
+
},
|
|
13608
|
+
contextWindow: 2000000,
|
|
13609
|
+
maxTokens: 30000,
|
|
13610
|
+
},
|
|
13340
13611
|
"grok-beta": {
|
|
13341
13612
|
id: "grok-beta",
|
|
13342
13613
|
name: "Grok Beta",
|
|
@@ -13552,748 +13823,24 @@ export const MODELS = {
|
|
|
13552
13823
|
contextWindow: 204800,
|
|
13553
13824
|
maxTokens: 131072,
|
|
13554
13825
|
},
|
|
13555
|
-
|
|
13556
|
-
|
|
13557
|
-
|
|
13558
|
-
id: "qwen3.5-plus",
|
|
13559
|
-
name: "Qwen3.5 Plus",
|
|
13560
|
-
api: "openai-completions",
|
|
13561
|
-
provider: "alibaba-coding-plan",
|
|
13562
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13563
|
-
reasoning: true,
|
|
13564
|
-
input: ["text"],
|
|
13565
|
-
cost: {
|
|
13566
|
-
input: 0,
|
|
13567
|
-
output: 0,
|
|
13568
|
-
cacheRead: 0,
|
|
13569
|
-
cacheWrite: 0,
|
|
13570
|
-
},
|
|
13571
|
-
contextWindow: 983616,
|
|
13572
|
-
maxTokens: 65536,
|
|
13573
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13574
|
-
},
|
|
13575
|
-
"qwen3-max-2026-01-23": {
|
|
13576
|
-
id: "qwen3-max-2026-01-23",
|
|
13577
|
-
name: "Qwen3 Max 2026-01-23",
|
|
13578
|
-
api: "openai-completions",
|
|
13579
|
-
provider: "alibaba-coding-plan",
|
|
13580
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13581
|
-
reasoning: true,
|
|
13582
|
-
input: ["text"],
|
|
13583
|
-
cost: {
|
|
13584
|
-
input: 0,
|
|
13585
|
-
output: 0,
|
|
13586
|
-
cacheRead: 0,
|
|
13587
|
-
cacheWrite: 0,
|
|
13588
|
-
},
|
|
13589
|
-
contextWindow: 258048,
|
|
13590
|
-
maxTokens: 32768,
|
|
13591
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13592
|
-
},
|
|
13593
|
-
"qwen3-coder-next": {
|
|
13594
|
-
id: "qwen3-coder-next",
|
|
13595
|
-
name: "Qwen3 Coder Next",
|
|
13596
|
-
api: "openai-completions",
|
|
13597
|
-
provider: "alibaba-coding-plan",
|
|
13598
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13599
|
-
reasoning: false,
|
|
13600
|
-
input: ["text"],
|
|
13601
|
-
cost: {
|
|
13602
|
-
input: 0,
|
|
13603
|
-
output: 0,
|
|
13604
|
-
cacheRead: 0,
|
|
13605
|
-
cacheWrite: 0,
|
|
13606
|
-
},
|
|
13607
|
-
contextWindow: 204800,
|
|
13608
|
-
maxTokens: 65536,
|
|
13609
|
-
compat: { supportsDeveloperRole: false },
|
|
13610
|
-
},
|
|
13611
|
-
"qwen3-coder-plus": {
|
|
13612
|
-
id: "qwen3-coder-plus",
|
|
13613
|
-
name: "Qwen3 Coder Plus",
|
|
13614
|
-
api: "openai-completions",
|
|
13615
|
-
provider: "alibaba-coding-plan",
|
|
13616
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13617
|
-
reasoning: false,
|
|
13618
|
-
input: ["text"],
|
|
13619
|
-
cost: {
|
|
13620
|
-
input: 0,
|
|
13621
|
-
output: 0,
|
|
13622
|
-
cacheRead: 0,
|
|
13623
|
-
cacheWrite: 0,
|
|
13624
|
-
},
|
|
13625
|
-
contextWindow: 997952,
|
|
13626
|
-
maxTokens: 65536,
|
|
13627
|
-
compat: { supportsDeveloperRole: false },
|
|
13628
|
-
},
|
|
13629
|
-
"MiniMax-M2.5": {
|
|
13630
|
-
id: "MiniMax-M2.5",
|
|
13631
|
-
name: "MiniMax M2.5",
|
|
13632
|
-
api: "openai-completions",
|
|
13633
|
-
provider: "alibaba-coding-plan",
|
|
13634
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13635
|
-
reasoning: true,
|
|
13636
|
-
input: ["text"],
|
|
13637
|
-
cost: {
|
|
13638
|
-
input: 0,
|
|
13639
|
-
output: 0,
|
|
13640
|
-
cacheRead: 0,
|
|
13641
|
-
cacheWrite: 0,
|
|
13642
|
-
},
|
|
13643
|
-
contextWindow: 196608,
|
|
13644
|
-
maxTokens: 65536,
|
|
13645
|
-
compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: true, maxTokensField: "max_tokens" },
|
|
13646
|
-
},
|
|
13647
|
-
"glm-5": {
|
|
13648
|
-
id: "glm-5",
|
|
13649
|
-
name: "GLM-5",
|
|
13650
|
-
api: "openai-completions",
|
|
13651
|
-
provider: "alibaba-coding-plan",
|
|
13652
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13653
|
-
reasoning: true,
|
|
13654
|
-
input: ["text"],
|
|
13655
|
-
cost: {
|
|
13656
|
-
input: 0,
|
|
13657
|
-
output: 0,
|
|
13658
|
-
cacheRead: 0,
|
|
13659
|
-
cacheWrite: 0,
|
|
13660
|
-
},
|
|
13661
|
-
contextWindow: 202752,
|
|
13662
|
-
maxTokens: 16384,
|
|
13663
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13664
|
-
},
|
|
13665
|
-
"glm-4.7": {
|
|
13666
|
-
id: "glm-4.7",
|
|
13667
|
-
name: "GLM-4.7",
|
|
13668
|
-
api: "openai-completions",
|
|
13669
|
-
provider: "alibaba-coding-plan",
|
|
13670
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13671
|
-
reasoning: true,
|
|
13672
|
-
input: ["text"],
|
|
13673
|
-
cost: {
|
|
13674
|
-
input: 0,
|
|
13675
|
-
output: 0,
|
|
13676
|
-
cacheRead: 0,
|
|
13677
|
-
cacheWrite: 0,
|
|
13678
|
-
},
|
|
13679
|
-
contextWindow: 169984,
|
|
13680
|
-
maxTokens: 16384,
|
|
13681
|
-
compat: { thinkingFormat: "qwen", supportsDeveloperRole: false },
|
|
13682
|
-
},
|
|
13683
|
-
"kimi-k2.5": {
|
|
13684
|
-
id: "kimi-k2.5",
|
|
13685
|
-
name: "Kimi K2.5",
|
|
13686
|
-
api: "openai-completions",
|
|
13687
|
-
provider: "alibaba-coding-plan",
|
|
13688
|
-
baseUrl: "https://coding-intl.dashscope.aliyuncs.com/v1",
|
|
13689
|
-
reasoning: true,
|
|
13690
|
-
input: ["text"],
|
|
13691
|
-
cost: {
|
|
13692
|
-
input: 0,
|
|
13693
|
-
output: 0,
|
|
13694
|
-
cacheRead: 0,
|
|
13695
|
-
cacheWrite: 0,
|
|
13696
|
-
},
|
|
13697
|
-
contextWindow: 258048,
|
|
13698
|
-
maxTokens: 32768,
|
|
13699
|
-
compat: { thinkingFormat: "zai", supportsDeveloperRole: false },
|
|
13700
|
-
},
|
|
13701
|
-
},
|
|
13702
|
-
"ollama-cloud": {
|
|
13703
|
-
"cogito-2.1:671b": {
|
|
13704
|
-
id: "cogito-2.1:671b",
|
|
13705
|
-
name: "Cogito 2.1 671B",
|
|
13706
|
-
api: "openai-completions",
|
|
13707
|
-
provider: "ollama-cloud",
|
|
13708
|
-
baseUrl: "https://ollama.com/v1",
|
|
13709
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13710
|
-
reasoning: true,
|
|
13711
|
-
input: ["text"],
|
|
13712
|
-
cost: {
|
|
13713
|
-
input: 0,
|
|
13714
|
-
output: 0,
|
|
13715
|
-
cacheRead: 0,
|
|
13716
|
-
cacheWrite: 0,
|
|
13717
|
-
},
|
|
13718
|
-
contextWindow: 163840,
|
|
13719
|
-
maxTokens: 32000,
|
|
13720
|
-
},
|
|
13721
|
-
"deepseek-v3.1:671b": {
|
|
13722
|
-
id: "deepseek-v3.1:671b",
|
|
13723
|
-
name: "DeepSeek V3.1 671B",
|
|
13724
|
-
api: "openai-completions",
|
|
13725
|
-
provider: "ollama-cloud",
|
|
13726
|
-
baseUrl: "https://ollama.com/v1",
|
|
13727
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13728
|
-
reasoning: true,
|
|
13729
|
-
input: ["text"],
|
|
13730
|
-
cost: {
|
|
13731
|
-
input: 0,
|
|
13732
|
-
output: 0,
|
|
13733
|
-
cacheRead: 0,
|
|
13734
|
-
cacheWrite: 0,
|
|
13735
|
-
},
|
|
13736
|
-
contextWindow: 163840,
|
|
13737
|
-
maxTokens: 163840,
|
|
13738
|
-
},
|
|
13739
|
-
"deepseek-v3.2": {
|
|
13740
|
-
id: "deepseek-v3.2",
|
|
13741
|
-
name: "DeepSeek V3.2",
|
|
13826
|
+
"glm-5-turbo": {
|
|
13827
|
+
id: "glm-5-turbo",
|
|
13828
|
+
name: "GLM-5-Turbo",
|
|
13742
13829
|
api: "openai-completions",
|
|
13743
|
-
provider: "
|
|
13744
|
-
baseUrl: "https://
|
|
13745
|
-
compat: { "
|
|
13830
|
+
provider: "zai",
|
|
13831
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13832
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
13746
13833
|
reasoning: true,
|
|
13747
13834
|
input: ["text"],
|
|
13748
13835
|
cost: {
|
|
13749
|
-
input:
|
|
13750
|
-
output:
|
|
13751
|
-
cacheRead: 0,
|
|
13752
|
-
cacheWrite: 0,
|
|
13753
|
-
},
|
|
13754
|
-
contextWindow: 163840,
|
|
13755
|
-
maxTokens: 65536,
|
|
13756
|
-
},
|
|
13757
|
-
"devstral-2:123b": {
|
|
13758
|
-
id: "devstral-2:123b",
|
|
13759
|
-
name: "Devstral 2 123B",
|
|
13760
|
-
api: "openai-completions",
|
|
13761
|
-
provider: "ollama-cloud",
|
|
13762
|
-
baseUrl: "https://ollama.com/v1",
|
|
13763
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13764
|
-
reasoning: false,
|
|
13765
|
-
input: ["text"],
|
|
13766
|
-
cost: {
|
|
13767
|
-
input: 0,
|
|
13768
|
-
output: 0,
|
|
13769
|
-
cacheRead: 0,
|
|
13836
|
+
input: 1.2,
|
|
13837
|
+
output: 4,
|
|
13838
|
+
cacheRead: 0.24,
|
|
13770
13839
|
cacheWrite: 0,
|
|
13771
13840
|
},
|
|
13772
|
-
contextWindow:
|
|
13773
|
-
maxTokens: 262144,
|
|
13774
|
-
},
|
|
13775
|
-
"devstral-small-2:24b": {
|
|
13776
|
-
id: "devstral-small-2:24b",
|
|
13777
|
-
name: "Devstral Small 2 24B",
|
|
13778
|
-
api: "openai-completions",
|
|
13779
|
-
provider: "ollama-cloud",
|
|
13780
|
-
baseUrl: "https://ollama.com/v1",
|
|
13781
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13782
|
-
reasoning: false,
|
|
13783
|
-
input: ["text", "image"],
|
|
13784
|
-
cost: {
|
|
13785
|
-
input: 0,
|
|
13786
|
-
output: 0,
|
|
13787
|
-
cacheRead: 0,
|
|
13788
|
-
cacheWrite: 0,
|
|
13789
|
-
},
|
|
13790
|
-
contextWindow: 262144,
|
|
13791
|
-
maxTokens: 262144,
|
|
13792
|
-
},
|
|
13793
|
-
"gemini-3-flash-preview": {
|
|
13794
|
-
id: "gemini-3-flash-preview",
|
|
13795
|
-
name: "Gemini 3 Flash Preview",
|
|
13796
|
-
api: "openai-completions",
|
|
13797
|
-
provider: "ollama-cloud",
|
|
13798
|
-
baseUrl: "https://ollama.com/v1",
|
|
13799
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13800
|
-
reasoning: true,
|
|
13801
|
-
input: ["text"],
|
|
13802
|
-
cost: {
|
|
13803
|
-
input: 0,
|
|
13804
|
-
output: 0,
|
|
13805
|
-
cacheRead: 0,
|
|
13806
|
-
cacheWrite: 0,
|
|
13807
|
-
},
|
|
13808
|
-
contextWindow: 1048576,
|
|
13809
|
-
maxTokens: 65536,
|
|
13810
|
-
},
|
|
13811
|
-
"gemma3:12b": {
|
|
13812
|
-
id: "gemma3:12b",
|
|
13813
|
-
name: "Gemma 3 12B",
|
|
13814
|
-
api: "openai-completions",
|
|
13815
|
-
provider: "ollama-cloud",
|
|
13816
|
-
baseUrl: "https://ollama.com/v1",
|
|
13817
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13818
|
-
reasoning: false,
|
|
13819
|
-
input: ["text", "image"],
|
|
13820
|
-
cost: {
|
|
13821
|
-
input: 0,
|
|
13822
|
-
output: 0,
|
|
13823
|
-
cacheRead: 0,
|
|
13824
|
-
cacheWrite: 0,
|
|
13825
|
-
},
|
|
13826
|
-
contextWindow: 131072,
|
|
13827
|
-
maxTokens: 131072,
|
|
13828
|
-
},
|
|
13829
|
-
"gemma3:27b": {
|
|
13830
|
-
id: "gemma3:27b",
|
|
13831
|
-
name: "Gemma 3 27B",
|
|
13832
|
-
api: "openai-completions",
|
|
13833
|
-
provider: "ollama-cloud",
|
|
13834
|
-
baseUrl: "https://ollama.com/v1",
|
|
13835
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13836
|
-
reasoning: false,
|
|
13837
|
-
input: ["text", "image"],
|
|
13838
|
-
cost: {
|
|
13839
|
-
input: 0,
|
|
13840
|
-
output: 0,
|
|
13841
|
-
cacheRead: 0,
|
|
13842
|
-
cacheWrite: 0,
|
|
13843
|
-
},
|
|
13844
|
-
contextWindow: 131072,
|
|
13845
|
-
maxTokens: 131072,
|
|
13846
|
-
},
|
|
13847
|
-
"gemma3:4b": {
|
|
13848
|
-
id: "gemma3:4b",
|
|
13849
|
-
name: "Gemma 3 4B",
|
|
13850
|
-
api: "openai-completions",
|
|
13851
|
-
provider: "ollama-cloud",
|
|
13852
|
-
baseUrl: "https://ollama.com/v1",
|
|
13853
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13854
|
-
reasoning: false,
|
|
13855
|
-
input: ["text", "image"],
|
|
13856
|
-
cost: {
|
|
13857
|
-
input: 0,
|
|
13858
|
-
output: 0,
|
|
13859
|
-
cacheRead: 0,
|
|
13860
|
-
cacheWrite: 0,
|
|
13861
|
-
},
|
|
13862
|
-
contextWindow: 131072,
|
|
13863
|
-
maxTokens: 131072,
|
|
13864
|
-
},
|
|
13865
|
-
"glm-4.6": {
|
|
13866
|
-
id: "glm-4.6",
|
|
13867
|
-
name: "GLM 4.6",
|
|
13868
|
-
api: "openai-completions",
|
|
13869
|
-
provider: "ollama-cloud",
|
|
13870
|
-
baseUrl: "https://ollama.com/v1",
|
|
13871
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13872
|
-
reasoning: true,
|
|
13873
|
-
input: ["text"],
|
|
13874
|
-
cost: {
|
|
13875
|
-
input: 0,
|
|
13876
|
-
output: 0,
|
|
13877
|
-
cacheRead: 0,
|
|
13878
|
-
cacheWrite: 0,
|
|
13879
|
-
},
|
|
13880
|
-
contextWindow: 202752,
|
|
13881
|
-
maxTokens: 131072,
|
|
13882
|
-
},
|
|
13883
|
-
"glm-4.7": {
|
|
13884
|
-
id: "glm-4.7",
|
|
13885
|
-
name: "GLM 4.7",
|
|
13886
|
-
api: "openai-completions",
|
|
13887
|
-
provider: "ollama-cloud",
|
|
13888
|
-
baseUrl: "https://ollama.com/v1",
|
|
13889
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13890
|
-
reasoning: true,
|
|
13891
|
-
input: ["text"],
|
|
13892
|
-
cost: {
|
|
13893
|
-
input: 0,
|
|
13894
|
-
output: 0,
|
|
13895
|
-
cacheRead: 0,
|
|
13896
|
-
cacheWrite: 0,
|
|
13897
|
-
},
|
|
13898
|
-
contextWindow: 202752,
|
|
13899
|
-
maxTokens: 131072,
|
|
13900
|
-
},
|
|
13901
|
-
"glm-5": {
|
|
13902
|
-
id: "glm-5",
|
|
13903
|
-
name: "GLM 5",
|
|
13904
|
-
api: "openai-completions",
|
|
13905
|
-
provider: "ollama-cloud",
|
|
13906
|
-
baseUrl: "https://ollama.com/v1",
|
|
13907
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13908
|
-
reasoning: true,
|
|
13909
|
-
input: ["text"],
|
|
13910
|
-
cost: {
|
|
13911
|
-
input: 0,
|
|
13912
|
-
output: 0,
|
|
13913
|
-
cacheRead: 0,
|
|
13914
|
-
cacheWrite: 0,
|
|
13915
|
-
},
|
|
13916
|
-
contextWindow: 202752,
|
|
13917
|
-
maxTokens: 131072,
|
|
13918
|
-
},
|
|
13919
|
-
"gpt-oss:120b": {
|
|
13920
|
-
id: "gpt-oss:120b",
|
|
13921
|
-
name: "GPT-OSS 120B",
|
|
13922
|
-
api: "openai-completions",
|
|
13923
|
-
provider: "ollama-cloud",
|
|
13924
|
-
baseUrl: "https://ollama.com/v1",
|
|
13925
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13926
|
-
reasoning: true,
|
|
13927
|
-
input: ["text"],
|
|
13928
|
-
cost: {
|
|
13929
|
-
input: 0,
|
|
13930
|
-
output: 0,
|
|
13931
|
-
cacheRead: 0,
|
|
13932
|
-
cacheWrite: 0,
|
|
13933
|
-
},
|
|
13934
|
-
contextWindow: 131072,
|
|
13935
|
-
maxTokens: 32768,
|
|
13936
|
-
},
|
|
13937
|
-
"gpt-oss:20b": {
|
|
13938
|
-
id: "gpt-oss:20b",
|
|
13939
|
-
name: "GPT-OSS 20B",
|
|
13940
|
-
api: "openai-completions",
|
|
13941
|
-
provider: "ollama-cloud",
|
|
13942
|
-
baseUrl: "https://ollama.com/v1",
|
|
13943
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13944
|
-
reasoning: true,
|
|
13945
|
-
input: ["text"],
|
|
13946
|
-
cost: {
|
|
13947
|
-
input: 0,
|
|
13948
|
-
output: 0,
|
|
13949
|
-
cacheRead: 0,
|
|
13950
|
-
cacheWrite: 0,
|
|
13951
|
-
},
|
|
13952
|
-
contextWindow: 131072,
|
|
13953
|
-
maxTokens: 32768,
|
|
13954
|
-
},
|
|
13955
|
-
"kimi-k2:1t": {
|
|
13956
|
-
id: "kimi-k2:1t",
|
|
13957
|
-
name: "Kimi K2 1T",
|
|
13958
|
-
api: "openai-completions",
|
|
13959
|
-
provider: "ollama-cloud",
|
|
13960
|
-
baseUrl: "https://ollama.com/v1",
|
|
13961
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13962
|
-
reasoning: false,
|
|
13963
|
-
input: ["text"],
|
|
13964
|
-
cost: {
|
|
13965
|
-
input: 0,
|
|
13966
|
-
output: 0,
|
|
13967
|
-
cacheRead: 0,
|
|
13968
|
-
cacheWrite: 0,
|
|
13969
|
-
},
|
|
13970
|
-
contextWindow: 262144,
|
|
13971
|
-
maxTokens: 262144,
|
|
13972
|
-
},
|
|
13973
|
-
"kimi-k2.5": {
|
|
13974
|
-
id: "kimi-k2.5",
|
|
13975
|
-
name: "Kimi K2.5",
|
|
13976
|
-
api: "openai-completions",
|
|
13977
|
-
provider: "ollama-cloud",
|
|
13978
|
-
baseUrl: "https://ollama.com/v1",
|
|
13979
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13980
|
-
reasoning: true,
|
|
13981
|
-
input: ["text", "image"],
|
|
13982
|
-
cost: {
|
|
13983
|
-
input: 0,
|
|
13984
|
-
output: 0,
|
|
13985
|
-
cacheRead: 0,
|
|
13986
|
-
cacheWrite: 0,
|
|
13987
|
-
},
|
|
13988
|
-
contextWindow: 262144,
|
|
13989
|
-
maxTokens: 262144,
|
|
13990
|
-
},
|
|
13991
|
-
"kimi-k2-thinking": {
|
|
13992
|
-
id: "kimi-k2-thinking",
|
|
13993
|
-
name: "Kimi K2 Thinking",
|
|
13994
|
-
api: "openai-completions",
|
|
13995
|
-
provider: "ollama-cloud",
|
|
13996
|
-
baseUrl: "https://ollama.com/v1",
|
|
13997
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
13998
|
-
reasoning: true,
|
|
13999
|
-
input: ["text"],
|
|
14000
|
-
cost: {
|
|
14001
|
-
input: 0,
|
|
14002
|
-
output: 0,
|
|
14003
|
-
cacheRead: 0,
|
|
14004
|
-
cacheWrite: 0,
|
|
14005
|
-
},
|
|
14006
|
-
contextWindow: 262144,
|
|
14007
|
-
maxTokens: 262144,
|
|
14008
|
-
},
|
|
14009
|
-
"minimax-m2.1": {
|
|
14010
|
-
id: "minimax-m2.1",
|
|
14011
|
-
name: "Minimax M2.1",
|
|
14012
|
-
api: "openai-completions",
|
|
14013
|
-
provider: "ollama-cloud",
|
|
14014
|
-
baseUrl: "https://ollama.com/v1",
|
|
14015
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14016
|
-
reasoning: true,
|
|
14017
|
-
input: ["text"],
|
|
14018
|
-
cost: {
|
|
14019
|
-
input: 0,
|
|
14020
|
-
output: 0,
|
|
14021
|
-
cacheRead: 0,
|
|
14022
|
-
cacheWrite: 0,
|
|
14023
|
-
},
|
|
14024
|
-
contextWindow: 204800,
|
|
14025
|
-
maxTokens: 131072,
|
|
14026
|
-
},
|
|
14027
|
-
"minimax-m2.5": {
|
|
14028
|
-
id: "minimax-m2.5",
|
|
14029
|
-
name: "Minimax M2.5",
|
|
14030
|
-
api: "openai-completions",
|
|
14031
|
-
provider: "ollama-cloud",
|
|
14032
|
-
baseUrl: "https://ollama.com/v1",
|
|
14033
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14034
|
-
reasoning: true,
|
|
14035
|
-
input: ["text"],
|
|
14036
|
-
cost: {
|
|
14037
|
-
input: 0,
|
|
14038
|
-
output: 0,
|
|
14039
|
-
cacheRead: 0,
|
|
14040
|
-
cacheWrite: 0,
|
|
14041
|
-
},
|
|
14042
|
-
contextWindow: 204800,
|
|
14043
|
-
maxTokens: 131072,
|
|
14044
|
-
},
|
|
14045
|
-
"minimax-m2": {
|
|
14046
|
-
id: "minimax-m2",
|
|
14047
|
-
name: "Minimax M2",
|
|
14048
|
-
api: "openai-completions",
|
|
14049
|
-
provider: "ollama-cloud",
|
|
14050
|
-
baseUrl: "https://ollama.com/v1",
|
|
14051
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14052
|
-
reasoning: false,
|
|
14053
|
-
input: ["text"],
|
|
14054
|
-
cost: {
|
|
14055
|
-
input: 0,
|
|
14056
|
-
output: 0,
|
|
14057
|
-
cacheRead: 0,
|
|
14058
|
-
cacheWrite: 0,
|
|
14059
|
-
},
|
|
14060
|
-
contextWindow: 204800,
|
|
14061
|
-
maxTokens: 128000,
|
|
14062
|
-
},
|
|
14063
|
-
"ministral-3:14b": {
|
|
14064
|
-
id: "ministral-3:14b",
|
|
14065
|
-
name: "Ministral 3 14B",
|
|
14066
|
-
api: "openai-completions",
|
|
14067
|
-
provider: "ollama-cloud",
|
|
14068
|
-
baseUrl: "https://ollama.com/v1",
|
|
14069
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14070
|
-
reasoning: false,
|
|
14071
|
-
input: ["text", "image"],
|
|
14072
|
-
cost: {
|
|
14073
|
-
input: 0,
|
|
14074
|
-
output: 0,
|
|
14075
|
-
cacheRead: 0,
|
|
14076
|
-
cacheWrite: 0,
|
|
14077
|
-
},
|
|
14078
|
-
contextWindow: 262144,
|
|
14079
|
-
maxTokens: 128000,
|
|
14080
|
-
},
|
|
14081
|
-
"ministral-3:3b": {
|
|
14082
|
-
id: "ministral-3:3b",
|
|
14083
|
-
name: "Ministral 3 3B",
|
|
14084
|
-
api: "openai-completions",
|
|
14085
|
-
provider: "ollama-cloud",
|
|
14086
|
-
baseUrl: "https://ollama.com/v1",
|
|
14087
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14088
|
-
reasoning: false,
|
|
14089
|
-
input: ["text", "image"],
|
|
14090
|
-
cost: {
|
|
14091
|
-
input: 0,
|
|
14092
|
-
output: 0,
|
|
14093
|
-
cacheRead: 0,
|
|
14094
|
-
cacheWrite: 0,
|
|
14095
|
-
},
|
|
14096
|
-
contextWindow: 262144,
|
|
14097
|
-
maxTokens: 128000,
|
|
14098
|
-
},
|
|
14099
|
-
"ministral-3:8b": {
|
|
14100
|
-
id: "ministral-3:8b",
|
|
14101
|
-
name: "Ministral 3 8B",
|
|
14102
|
-
api: "openai-completions",
|
|
14103
|
-
provider: "ollama-cloud",
|
|
14104
|
-
baseUrl: "https://ollama.com/v1",
|
|
14105
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14106
|
-
reasoning: false,
|
|
14107
|
-
input: ["text", "image"],
|
|
14108
|
-
cost: {
|
|
14109
|
-
input: 0,
|
|
14110
|
-
output: 0,
|
|
14111
|
-
cacheRead: 0,
|
|
14112
|
-
cacheWrite: 0,
|
|
14113
|
-
},
|
|
14114
|
-
contextWindow: 262144,
|
|
14115
|
-
maxTokens: 128000,
|
|
14116
|
-
},
|
|
14117
|
-
"mistral-large-3:675b": {
|
|
14118
|
-
id: "mistral-large-3:675b",
|
|
14119
|
-
name: "Mistral Large 3 675B",
|
|
14120
|
-
api: "openai-completions",
|
|
14121
|
-
provider: "ollama-cloud",
|
|
14122
|
-
baseUrl: "https://ollama.com/v1",
|
|
14123
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14124
|
-
reasoning: false,
|
|
14125
|
-
input: ["text", "image"],
|
|
14126
|
-
cost: {
|
|
14127
|
-
input: 0,
|
|
14128
|
-
output: 0,
|
|
14129
|
-
cacheRead: 0,
|
|
14130
|
-
cacheWrite: 0,
|
|
14131
|
-
},
|
|
14132
|
-
contextWindow: 262144,
|
|
14133
|
-
maxTokens: 262144,
|
|
14134
|
-
},
|
|
14135
|
-
"nemotron-3-nano:30b": {
|
|
14136
|
-
id: "nemotron-3-nano:30b",
|
|
14137
|
-
name: "Nemotron 3 Nano 30B",
|
|
14138
|
-
api: "openai-completions",
|
|
14139
|
-
provider: "ollama-cloud",
|
|
14140
|
-
baseUrl: "https://ollama.com/v1",
|
|
14141
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14142
|
-
reasoning: true,
|
|
14143
|
-
input: ["text"],
|
|
14144
|
-
cost: {
|
|
14145
|
-
input: 0,
|
|
14146
|
-
output: 0,
|
|
14147
|
-
cacheRead: 0,
|
|
14148
|
-
cacheWrite: 0,
|
|
14149
|
-
},
|
|
14150
|
-
contextWindow: 1048576,
|
|
14151
|
-
maxTokens: 131072,
|
|
14152
|
-
},
|
|
14153
|
-
"nemotron-3-super": {
|
|
14154
|
-
id: "nemotron-3-super",
|
|
14155
|
-
name: "Nemotron 3 Super",
|
|
14156
|
-
api: "openai-completions",
|
|
14157
|
-
provider: "ollama-cloud",
|
|
14158
|
-
baseUrl: "https://ollama.com/v1",
|
|
14159
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14160
|
-
reasoning: true,
|
|
14161
|
-
input: ["text"],
|
|
14162
|
-
cost: {
|
|
14163
|
-
input: 0,
|
|
14164
|
-
output: 0,
|
|
14165
|
-
cacheRead: 0,
|
|
14166
|
-
cacheWrite: 0,
|
|
14167
|
-
},
|
|
14168
|
-
contextWindow: 262144,
|
|
14169
|
-
maxTokens: 65536,
|
|
14170
|
-
},
|
|
14171
|
-
"qwen3.5:397b": {
|
|
14172
|
-
id: "qwen3.5:397b",
|
|
14173
|
-
name: "Qwen 3.5 397B",
|
|
14174
|
-
api: "openai-completions",
|
|
14175
|
-
provider: "ollama-cloud",
|
|
14176
|
-
baseUrl: "https://ollama.com/v1",
|
|
14177
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14178
|
-
reasoning: true,
|
|
14179
|
-
input: ["text", "image"],
|
|
14180
|
-
cost: {
|
|
14181
|
-
input: 0,
|
|
14182
|
-
output: 0,
|
|
14183
|
-
cacheRead: 0,
|
|
14184
|
-
cacheWrite: 0,
|
|
14185
|
-
},
|
|
14186
|
-
contextWindow: 262144,
|
|
14187
|
-
maxTokens: 81920,
|
|
14188
|
-
},
|
|
14189
|
-
"qwen3-coder:480b": {
|
|
14190
|
-
id: "qwen3-coder:480b",
|
|
14191
|
-
name: "Qwen 3 Coder 480B",
|
|
14192
|
-
api: "openai-completions",
|
|
14193
|
-
provider: "ollama-cloud",
|
|
14194
|
-
baseUrl: "https://ollama.com/v1",
|
|
14195
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14196
|
-
reasoning: false,
|
|
14197
|
-
input: ["text"],
|
|
14198
|
-
cost: {
|
|
14199
|
-
input: 0,
|
|
14200
|
-
output: 0,
|
|
14201
|
-
cacheRead: 0,
|
|
14202
|
-
cacheWrite: 0,
|
|
14203
|
-
},
|
|
14204
|
-
contextWindow: 262144,
|
|
14205
|
-
maxTokens: 65536,
|
|
14206
|
-
},
|
|
14207
|
-
"qwen3-coder-next": {
|
|
14208
|
-
id: "qwen3-coder-next",
|
|
14209
|
-
name: "Qwen 3 Coder Next",
|
|
14210
|
-
api: "openai-completions",
|
|
14211
|
-
provider: "ollama-cloud",
|
|
14212
|
-
baseUrl: "https://ollama.com/v1",
|
|
14213
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14214
|
-
reasoning: false,
|
|
14215
|
-
input: ["text"],
|
|
14216
|
-
cost: {
|
|
14217
|
-
input: 0,
|
|
14218
|
-
output: 0,
|
|
14219
|
-
cacheRead: 0,
|
|
14220
|
-
cacheWrite: 0,
|
|
14221
|
-
},
|
|
14222
|
-
contextWindow: 262144,
|
|
14223
|
-
maxTokens: 65536,
|
|
14224
|
-
},
|
|
14225
|
-
"qwen3-next:80b": {
|
|
14226
|
-
id: "qwen3-next:80b",
|
|
14227
|
-
name: "Qwen 3 Next 80B",
|
|
14228
|
-
api: "openai-completions",
|
|
14229
|
-
provider: "ollama-cloud",
|
|
14230
|
-
baseUrl: "https://ollama.com/v1",
|
|
14231
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14232
|
-
reasoning: true,
|
|
14233
|
-
input: ["text"],
|
|
14234
|
-
cost: {
|
|
14235
|
-
input: 0,
|
|
14236
|
-
output: 0,
|
|
14237
|
-
cacheRead: 0,
|
|
14238
|
-
cacheWrite: 0,
|
|
14239
|
-
},
|
|
14240
|
-
contextWindow: 262144,
|
|
14241
|
-
maxTokens: 32768,
|
|
14242
|
-
},
|
|
14243
|
-
"qwen3-vl:235b-instruct": {
|
|
14244
|
-
id: "qwen3-vl:235b-instruct",
|
|
14245
|
-
name: "Qwen 3 VL 235B Instruct",
|
|
14246
|
-
api: "openai-completions",
|
|
14247
|
-
provider: "ollama-cloud",
|
|
14248
|
-
baseUrl: "https://ollama.com/v1",
|
|
14249
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14250
|
-
reasoning: false,
|
|
14251
|
-
input: ["text", "image"],
|
|
14252
|
-
cost: {
|
|
14253
|
-
input: 0,
|
|
14254
|
-
output: 0,
|
|
14255
|
-
cacheRead: 0,
|
|
14256
|
-
cacheWrite: 0,
|
|
14257
|
-
},
|
|
14258
|
-
contextWindow: 262144,
|
|
13841
|
+
contextWindow: 200000,
|
|
14259
13842
|
maxTokens: 131072,
|
|
14260
13843
|
},
|
|
14261
|
-
"qwen3-vl:235b": {
|
|
14262
|
-
id: "qwen3-vl:235b",
|
|
14263
|
-
name: "Qwen 3 VL 235B",
|
|
14264
|
-
api: "openai-completions",
|
|
14265
|
-
provider: "ollama-cloud",
|
|
14266
|
-
baseUrl: "https://ollama.com/v1",
|
|
14267
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": true, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14268
|
-
reasoning: true,
|
|
14269
|
-
input: ["text", "image"],
|
|
14270
|
-
cost: {
|
|
14271
|
-
input: 0,
|
|
14272
|
-
output: 0,
|
|
14273
|
-
cacheRead: 0,
|
|
14274
|
-
cacheWrite: 0,
|
|
14275
|
-
},
|
|
14276
|
-
contextWindow: 262144,
|
|
14277
|
-
maxTokens: 32768,
|
|
14278
|
-
},
|
|
14279
|
-
"rnj-1:8b": {
|
|
14280
|
-
id: "rnj-1:8b",
|
|
14281
|
-
name: "RNJ 1 8B",
|
|
14282
|
-
api: "openai-completions",
|
|
14283
|
-
provider: "ollama-cloud",
|
|
14284
|
-
baseUrl: "https://ollama.com/v1",
|
|
14285
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
|
|
14286
|
-
reasoning: false,
|
|
14287
|
-
input: ["text"],
|
|
14288
|
-
cost: {
|
|
14289
|
-
input: 0,
|
|
14290
|
-
output: 0,
|
|
14291
|
-
cacheRead: 0,
|
|
14292
|
-
cacheWrite: 0,
|
|
14293
|
-
},
|
|
14294
|
-
contextWindow: 32768,
|
|
14295
|
-
maxTokens: 4096,
|
|
14296
|
-
},
|
|
14297
13844
|
},
|
|
14298
13845
|
};
|
|
14299
13846
|
//# sourceMappingURL=models.generated.js.map
|