@draht/ai 2026.3.25 → 2026.4.23
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +107 -12
- package/dist/cli.d.ts.map +1 -1
- package/dist/cli.js +1 -1
- package/dist/cli.js.map +1 -1
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -0
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +1163 -498
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +1392 -756
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +37 -9
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +38 -16
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/faux.d.ts +56 -0
- package/dist/providers/faux.d.ts.map +1 -0
- package/dist/providers/faux.js +367 -0
- package/dist/providers/faux.js.map +1 -0
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +2 -2
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +1 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +29 -4
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +18 -6
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +20 -1
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +2 -2
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/types.d.ts +64 -5
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts +4 -3
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +28 -13
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +5 -8
package/dist/models.generated.js
CHANGED
|
@@ -257,6 +257,23 @@ export const MODELS = {
|
|
|
257
257
|
contextWindow: 1000000,
|
|
258
258
|
maxTokens: 128000,
|
|
259
259
|
},
|
|
260
|
+
"anthropic.claude-opus-4-7": {
|
|
261
|
+
id: "anthropic.claude-opus-4-7",
|
|
262
|
+
name: "Claude Opus 4.7",
|
|
263
|
+
api: "bedrock-converse-stream",
|
|
264
|
+
provider: "amazon-bedrock",
|
|
265
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
266
|
+
reasoning: true,
|
|
267
|
+
input: ["text", "image"],
|
|
268
|
+
cost: {
|
|
269
|
+
input: 5,
|
|
270
|
+
output: 25,
|
|
271
|
+
cacheRead: 0.5,
|
|
272
|
+
cacheWrite: 6.25,
|
|
273
|
+
},
|
|
274
|
+
contextWindow: 1000000,
|
|
275
|
+
maxTokens: 128000,
|
|
276
|
+
},
|
|
260
277
|
"anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
261
278
|
id: "anthropic.claude-sonnet-4-20250514-v1:0",
|
|
262
279
|
name: "Claude Sonnet 4",
|
|
@@ -410,6 +427,23 @@ export const MODELS = {
|
|
|
410
427
|
contextWindow: 1000000,
|
|
411
428
|
maxTokens: 128000,
|
|
412
429
|
},
|
|
430
|
+
"eu.anthropic.claude-opus-4-7": {
|
|
431
|
+
id: "eu.anthropic.claude-opus-4-7",
|
|
432
|
+
name: "Claude Opus 4.7 (EU)",
|
|
433
|
+
api: "bedrock-converse-stream",
|
|
434
|
+
provider: "amazon-bedrock",
|
|
435
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
436
|
+
reasoning: true,
|
|
437
|
+
input: ["text", "image"],
|
|
438
|
+
cost: {
|
|
439
|
+
input: 5,
|
|
440
|
+
output: 25,
|
|
441
|
+
cacheRead: 0.5,
|
|
442
|
+
cacheWrite: 6.25,
|
|
443
|
+
},
|
|
444
|
+
contextWindow: 1000000,
|
|
445
|
+
maxTokens: 128000,
|
|
446
|
+
},
|
|
413
447
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
414
448
|
id: "eu.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
415
449
|
name: "Claude Sonnet 4 (EU)",
|
|
@@ -512,6 +546,23 @@ export const MODELS = {
|
|
|
512
546
|
contextWindow: 1000000,
|
|
513
547
|
maxTokens: 128000,
|
|
514
548
|
},
|
|
549
|
+
"global.anthropic.claude-opus-4-7": {
|
|
550
|
+
id: "global.anthropic.claude-opus-4-7",
|
|
551
|
+
name: "Claude Opus 4.7 (Global)",
|
|
552
|
+
api: "bedrock-converse-stream",
|
|
553
|
+
provider: "amazon-bedrock",
|
|
554
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
555
|
+
reasoning: true,
|
|
556
|
+
input: ["text", "image"],
|
|
557
|
+
cost: {
|
|
558
|
+
input: 5,
|
|
559
|
+
output: 25,
|
|
560
|
+
cacheRead: 0.5,
|
|
561
|
+
cacheWrite: 6.25,
|
|
562
|
+
},
|
|
563
|
+
contextWindow: 1000000,
|
|
564
|
+
maxTokens: 128000,
|
|
565
|
+
},
|
|
515
566
|
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
516
567
|
id: "global.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
517
568
|
name: "Claude Sonnet 4 (Global)",
|
|
@@ -815,8 +866,8 @@ export const MODELS = {
|
|
|
815
866
|
cacheRead: 0,
|
|
816
867
|
cacheWrite: 0,
|
|
817
868
|
},
|
|
818
|
-
contextWindow:
|
|
819
|
-
maxTokens:
|
|
869
|
+
contextWindow: 196608,
|
|
870
|
+
maxTokens: 98304,
|
|
820
871
|
},
|
|
821
872
|
"mistral.devstral-2-123b": {
|
|
822
873
|
id: "mistral.devstral-2-123b",
|
|
@@ -1056,6 +1107,23 @@ export const MODELS = {
|
|
|
1056
1107
|
contextWindow: 128000,
|
|
1057
1108
|
maxTokens: 4096,
|
|
1058
1109
|
},
|
|
1110
|
+
"nvidia.nemotron-super-3-120b": {
|
|
1111
|
+
id: "nvidia.nemotron-super-3-120b",
|
|
1112
|
+
name: "NVIDIA Nemotron 3 Super 120B A12B",
|
|
1113
|
+
api: "bedrock-converse-stream",
|
|
1114
|
+
provider: "amazon-bedrock",
|
|
1115
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1116
|
+
reasoning: true,
|
|
1117
|
+
input: ["text"],
|
|
1118
|
+
cost: {
|
|
1119
|
+
input: 0.15,
|
|
1120
|
+
output: 0.65,
|
|
1121
|
+
cacheRead: 0,
|
|
1122
|
+
cacheWrite: 0,
|
|
1123
|
+
},
|
|
1124
|
+
contextWindow: 262144,
|
|
1125
|
+
maxTokens: 131072,
|
|
1126
|
+
},
|
|
1059
1127
|
"openai.gpt-oss-120b-1:0": {
|
|
1060
1128
|
id: "openai.gpt-oss-120b-1:0",
|
|
1061
1129
|
name: "gpt-oss-120b",
|
|
@@ -1192,6 +1260,23 @@ export const MODELS = {
|
|
|
1192
1260
|
contextWindow: 131072,
|
|
1193
1261
|
maxTokens: 65536,
|
|
1194
1262
|
},
|
|
1263
|
+
"qwen.qwen3-coder-next": {
|
|
1264
|
+
id: "qwen.qwen3-coder-next",
|
|
1265
|
+
name: "Qwen3 Coder Next",
|
|
1266
|
+
api: "bedrock-converse-stream",
|
|
1267
|
+
provider: "amazon-bedrock",
|
|
1268
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1269
|
+
reasoning: true,
|
|
1270
|
+
input: ["text"],
|
|
1271
|
+
cost: {
|
|
1272
|
+
input: 0.22,
|
|
1273
|
+
output: 1.8,
|
|
1274
|
+
cacheRead: 0,
|
|
1275
|
+
cacheWrite: 0,
|
|
1276
|
+
},
|
|
1277
|
+
contextWindow: 131072,
|
|
1278
|
+
maxTokens: 65536,
|
|
1279
|
+
},
|
|
1195
1280
|
"qwen.qwen3-next-80b-a3b": {
|
|
1196
1281
|
id: "qwen.qwen3-next-80b-a3b",
|
|
1197
1282
|
name: "Qwen/Qwen3-Next-80B-A3B-Instruct",
|
|
@@ -1311,6 +1396,23 @@ export const MODELS = {
|
|
|
1311
1396
|
contextWindow: 1000000,
|
|
1312
1397
|
maxTokens: 128000,
|
|
1313
1398
|
},
|
|
1399
|
+
"us.anthropic.claude-opus-4-7": {
|
|
1400
|
+
id: "us.anthropic.claude-opus-4-7",
|
|
1401
|
+
name: "Claude Opus 4.7 (US)",
|
|
1402
|
+
api: "bedrock-converse-stream",
|
|
1403
|
+
provider: "amazon-bedrock",
|
|
1404
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1405
|
+
reasoning: true,
|
|
1406
|
+
input: ["text", "image"],
|
|
1407
|
+
cost: {
|
|
1408
|
+
input: 5,
|
|
1409
|
+
output: 25,
|
|
1410
|
+
cacheRead: 0.5,
|
|
1411
|
+
cacheWrite: 6.25,
|
|
1412
|
+
},
|
|
1413
|
+
contextWindow: 1000000,
|
|
1414
|
+
maxTokens: 128000,
|
|
1415
|
+
},
|
|
1314
1416
|
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
1315
1417
|
id: "us.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
1316
1418
|
name: "Claude Sonnet 4 (US)",
|
|
@@ -1444,8 +1546,8 @@ export const MODELS = {
|
|
|
1444
1546
|
cacheRead: 0,
|
|
1445
1547
|
cacheWrite: 0,
|
|
1446
1548
|
},
|
|
1447
|
-
contextWindow:
|
|
1448
|
-
maxTokens:
|
|
1549
|
+
contextWindow: 202752,
|
|
1550
|
+
maxTokens: 101376,
|
|
1449
1551
|
},
|
|
1450
1552
|
},
|
|
1451
1553
|
"anthropic": {
|
|
@@ -1534,23 +1636,6 @@ export const MODELS = {
|
|
|
1534
1636
|
contextWindow: 200000,
|
|
1535
1637
|
maxTokens: 64000,
|
|
1536
1638
|
},
|
|
1537
|
-
"claude-3-7-sonnet-latest": {
|
|
1538
|
-
id: "claude-3-7-sonnet-latest",
|
|
1539
|
-
name: "Claude Sonnet 3.7 (latest)",
|
|
1540
|
-
api: "anthropic-messages",
|
|
1541
|
-
provider: "anthropic",
|
|
1542
|
-
baseUrl: "https://api.anthropic.com",
|
|
1543
|
-
reasoning: true,
|
|
1544
|
-
input: ["text", "image"],
|
|
1545
|
-
cost: {
|
|
1546
|
-
input: 3,
|
|
1547
|
-
output: 15,
|
|
1548
|
-
cacheRead: 0.3,
|
|
1549
|
-
cacheWrite: 3.75,
|
|
1550
|
-
},
|
|
1551
|
-
contextWindow: 200000,
|
|
1552
|
-
maxTokens: 64000,
|
|
1553
|
-
},
|
|
1554
1639
|
"claude-3-haiku-20240307": {
|
|
1555
1640
|
id: "claude-3-haiku-20240307",
|
|
1556
1641
|
name: "Claude Haiku 3",
|
|
@@ -1755,6 +1840,23 @@ export const MODELS = {
|
|
|
1755
1840
|
contextWindow: 1000000,
|
|
1756
1841
|
maxTokens: 128000,
|
|
1757
1842
|
},
|
|
1843
|
+
"claude-opus-4-7": {
|
|
1844
|
+
id: "claude-opus-4-7",
|
|
1845
|
+
name: "Claude Opus 4.7",
|
|
1846
|
+
api: "anthropic-messages",
|
|
1847
|
+
provider: "anthropic",
|
|
1848
|
+
baseUrl: "https://api.anthropic.com",
|
|
1849
|
+
reasoning: true,
|
|
1850
|
+
input: ["text", "image"],
|
|
1851
|
+
cost: {
|
|
1852
|
+
input: 5,
|
|
1853
|
+
output: 25,
|
|
1854
|
+
cacheRead: 0.5,
|
|
1855
|
+
cacheWrite: 6.25,
|
|
1856
|
+
},
|
|
1857
|
+
contextWindow: 1000000,
|
|
1858
|
+
maxTokens: 128000,
|
|
1859
|
+
},
|
|
1758
1860
|
"claude-sonnet-4-0": {
|
|
1759
1861
|
id: "claude-sonnet-4-0",
|
|
1760
1862
|
name: "Claude Sonnet 4 (latest)",
|
|
@@ -1842,23 +1944,6 @@ export const MODELS = {
|
|
|
1842
1944
|
},
|
|
1843
1945
|
},
|
|
1844
1946
|
"azure-openai-responses": {
|
|
1845
|
-
"codex-mini-latest": {
|
|
1846
|
-
id: "codex-mini-latest",
|
|
1847
|
-
name: "Codex Mini",
|
|
1848
|
-
api: "azure-openai-responses",
|
|
1849
|
-
provider: "azure-openai-responses",
|
|
1850
|
-
baseUrl: "",
|
|
1851
|
-
reasoning: true,
|
|
1852
|
-
input: ["text"],
|
|
1853
|
-
cost: {
|
|
1854
|
-
input: 1.5,
|
|
1855
|
-
output: 6,
|
|
1856
|
-
cacheRead: 0.375,
|
|
1857
|
-
cacheWrite: 0,
|
|
1858
|
-
},
|
|
1859
|
-
contextWindow: 200000,
|
|
1860
|
-
maxTokens: 100000,
|
|
1861
|
-
},
|
|
1862
1947
|
"gpt-4": {
|
|
1863
1948
|
id: "gpt-4",
|
|
1864
1949
|
name: "GPT-4",
|
|
@@ -2284,6 +2369,23 @@ export const MODELS = {
|
|
|
2284
2369
|
contextWindow: 400000,
|
|
2285
2370
|
maxTokens: 128000,
|
|
2286
2371
|
},
|
|
2372
|
+
"gpt-5.3-chat-latest": {
|
|
2373
|
+
id: "gpt-5.3-chat-latest",
|
|
2374
|
+
name: "GPT-5.3 Chat (latest)",
|
|
2375
|
+
api: "azure-openai-responses",
|
|
2376
|
+
provider: "azure-openai-responses",
|
|
2377
|
+
baseUrl: "",
|
|
2378
|
+
reasoning: false,
|
|
2379
|
+
input: ["text", "image"],
|
|
2380
|
+
cost: {
|
|
2381
|
+
input: 1.75,
|
|
2382
|
+
output: 14,
|
|
2383
|
+
cacheRead: 0.175,
|
|
2384
|
+
cacheWrite: 0,
|
|
2385
|
+
},
|
|
2386
|
+
contextWindow: 128000,
|
|
2387
|
+
maxTokens: 16384,
|
|
2388
|
+
},
|
|
2287
2389
|
"gpt-5.3-codex": {
|
|
2288
2390
|
id: "gpt-5.3-codex",
|
|
2289
2391
|
name: "GPT-5.3 Codex",
|
|
@@ -2648,6 +2750,24 @@ export const MODELS = {
|
|
|
2648
2750
|
contextWindow: 1000000,
|
|
2649
2751
|
maxTokens: 64000,
|
|
2650
2752
|
},
|
|
2753
|
+
"claude-opus-4.7": {
|
|
2754
|
+
id: "claude-opus-4.7",
|
|
2755
|
+
name: "Claude Opus 4.7",
|
|
2756
|
+
api: "anthropic-messages",
|
|
2757
|
+
provider: "github-copilot",
|
|
2758
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
2759
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
2760
|
+
reasoning: true,
|
|
2761
|
+
input: ["text", "image"],
|
|
2762
|
+
cost: {
|
|
2763
|
+
input: 0,
|
|
2764
|
+
output: 0,
|
|
2765
|
+
cacheRead: 0,
|
|
2766
|
+
cacheWrite: 0,
|
|
2767
|
+
},
|
|
2768
|
+
contextWindow: 144000,
|
|
2769
|
+
maxTokens: 64000,
|
|
2770
|
+
},
|
|
2651
2771
|
"claude-sonnet-4": {
|
|
2652
2772
|
id: "claude-sonnet-4",
|
|
2653
2773
|
name: "Claude Sonnet 4",
|
|
@@ -2998,7 +3118,7 @@ export const MODELS = {
|
|
|
2998
3118
|
},
|
|
2999
3119
|
"gpt-5.4-mini": {
|
|
3000
3120
|
id: "gpt-5.4-mini",
|
|
3001
|
-
name: "GPT-5.4
|
|
3121
|
+
name: "GPT-5.4 Mini",
|
|
3002
3122
|
api: "openai-responses",
|
|
3003
3123
|
provider: "github-copilot",
|
|
3004
3124
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
@@ -3131,7 +3251,7 @@ export const MODELS = {
|
|
|
3131
3251
|
cost: {
|
|
3132
3252
|
input: 0.3,
|
|
3133
3253
|
output: 2.5,
|
|
3134
|
-
cacheRead: 0.
|
|
3254
|
+
cacheRead: 0.03,
|
|
3135
3255
|
cacheWrite: 0,
|
|
3136
3256
|
},
|
|
3137
3257
|
contextWindow: 1048576,
|
|
@@ -3250,7 +3370,7 @@ export const MODELS = {
|
|
|
3250
3370
|
cost: {
|
|
3251
3371
|
input: 1.25,
|
|
3252
3372
|
output: 10,
|
|
3253
|
-
cacheRead: 0.
|
|
3373
|
+
cacheRead: 0.125,
|
|
3254
3374
|
cacheWrite: 0,
|
|
3255
3375
|
},
|
|
3256
3376
|
contextWindow: 1048576,
|
|
@@ -3443,6 +3563,57 @@ export const MODELS = {
|
|
|
3443
3563
|
contextWindow: 131072,
|
|
3444
3564
|
maxTokens: 65536,
|
|
3445
3565
|
},
|
|
3566
|
+
"gemma-3-27b-it": {
|
|
3567
|
+
id: "gemma-3-27b-it",
|
|
3568
|
+
name: "Gemma 3 27B",
|
|
3569
|
+
api: "google-generative-ai",
|
|
3570
|
+
provider: "google",
|
|
3571
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3572
|
+
reasoning: false,
|
|
3573
|
+
input: ["text", "image"],
|
|
3574
|
+
cost: {
|
|
3575
|
+
input: 0,
|
|
3576
|
+
output: 0,
|
|
3577
|
+
cacheRead: 0,
|
|
3578
|
+
cacheWrite: 0,
|
|
3579
|
+
},
|
|
3580
|
+
contextWindow: 131072,
|
|
3581
|
+
maxTokens: 8192,
|
|
3582
|
+
},
|
|
3583
|
+
"gemma-4-26b-a4b-it": {
|
|
3584
|
+
id: "gemma-4-26b-a4b-it",
|
|
3585
|
+
name: "Gemma 4 26B",
|
|
3586
|
+
api: "google-generative-ai",
|
|
3587
|
+
provider: "google",
|
|
3588
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3589
|
+
reasoning: true,
|
|
3590
|
+
input: ["text", "image"],
|
|
3591
|
+
cost: {
|
|
3592
|
+
input: 0,
|
|
3593
|
+
output: 0,
|
|
3594
|
+
cacheRead: 0,
|
|
3595
|
+
cacheWrite: 0,
|
|
3596
|
+
},
|
|
3597
|
+
contextWindow: 256000,
|
|
3598
|
+
maxTokens: 8192,
|
|
3599
|
+
},
|
|
3600
|
+
"gemma-4-31b-it": {
|
|
3601
|
+
id: "gemma-4-31b-it",
|
|
3602
|
+
name: "Gemma 4 31B",
|
|
3603
|
+
api: "google-generative-ai",
|
|
3604
|
+
provider: "google",
|
|
3605
|
+
baseUrl: "https://generativelanguage.googleapis.com/v1beta",
|
|
3606
|
+
reasoning: true,
|
|
3607
|
+
input: ["text", "image"],
|
|
3608
|
+
cost: {
|
|
3609
|
+
input: 0,
|
|
3610
|
+
output: 0,
|
|
3611
|
+
cacheRead: 0,
|
|
3612
|
+
cacheWrite: 0,
|
|
3613
|
+
},
|
|
3614
|
+
contextWindow: 256000,
|
|
3615
|
+
maxTokens: 8192,
|
|
3616
|
+
},
|
|
3446
3617
|
},
|
|
3447
3618
|
"google-antigravity": {
|
|
3448
3619
|
"claude-opus-4-5-thinking": {
|
|
@@ -3908,6 +4079,23 @@ export const MODELS = {
|
|
|
3908
4079
|
contextWindow: 1048576,
|
|
3909
4080
|
maxTokens: 65536,
|
|
3910
4081
|
},
|
|
4082
|
+
"gemini-3.1-pro-preview-customtools": {
|
|
4083
|
+
id: "gemini-3.1-pro-preview-customtools",
|
|
4084
|
+
name: "Gemini 3.1 Pro Preview Custom Tools (Vertex)",
|
|
4085
|
+
api: "google-vertex",
|
|
4086
|
+
provider: "google-vertex",
|
|
4087
|
+
baseUrl: "https://{location}-aiplatform.googleapis.com",
|
|
4088
|
+
reasoning: true,
|
|
4089
|
+
input: ["text", "image"],
|
|
4090
|
+
cost: {
|
|
4091
|
+
input: 2,
|
|
4092
|
+
output: 12,
|
|
4093
|
+
cacheRead: 0.2,
|
|
4094
|
+
cacheWrite: 0,
|
|
4095
|
+
},
|
|
4096
|
+
contextWindow: 1048576,
|
|
4097
|
+
maxTokens: 65536,
|
|
4098
|
+
},
|
|
3911
4099
|
},
|
|
3912
4100
|
"groq": {
|
|
3913
4101
|
"deepseek-r1-distill-llama-70b": {
|
|
@@ -3944,6 +4132,40 @@ export const MODELS = {
|
|
|
3944
4132
|
contextWindow: 8192,
|
|
3945
4133
|
maxTokens: 8192,
|
|
3946
4134
|
},
|
|
4135
|
+
"groq/compound": {
|
|
4136
|
+
id: "groq/compound",
|
|
4137
|
+
name: "Compound",
|
|
4138
|
+
api: "openai-completions",
|
|
4139
|
+
provider: "groq",
|
|
4140
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
4141
|
+
reasoning: true,
|
|
4142
|
+
input: ["text"],
|
|
4143
|
+
cost: {
|
|
4144
|
+
input: 0,
|
|
4145
|
+
output: 0,
|
|
4146
|
+
cacheRead: 0,
|
|
4147
|
+
cacheWrite: 0,
|
|
4148
|
+
},
|
|
4149
|
+
contextWindow: 131072,
|
|
4150
|
+
maxTokens: 8192,
|
|
4151
|
+
},
|
|
4152
|
+
"groq/compound-mini": {
|
|
4153
|
+
id: "groq/compound-mini",
|
|
4154
|
+
name: "Compound Mini",
|
|
4155
|
+
api: "openai-completions",
|
|
4156
|
+
provider: "groq",
|
|
4157
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
4158
|
+
reasoning: true,
|
|
4159
|
+
input: ["text"],
|
|
4160
|
+
cost: {
|
|
4161
|
+
input: 0,
|
|
4162
|
+
output: 0,
|
|
4163
|
+
cacheRead: 0,
|
|
4164
|
+
cacheWrite: 0,
|
|
4165
|
+
},
|
|
4166
|
+
contextWindow: 131072,
|
|
4167
|
+
maxTokens: 8192,
|
|
4168
|
+
},
|
|
3947
4169
|
"llama-3.1-8b-instant": {
|
|
3948
4170
|
id: "llama-3.1-8b-instant",
|
|
3949
4171
|
name: "Llama 3.1 8B Instant",
|
|
@@ -4131,9 +4353,26 @@ export const MODELS = {
|
|
|
4131
4353
|
contextWindow: 131072,
|
|
4132
4354
|
maxTokens: 65536,
|
|
4133
4355
|
},
|
|
4134
|
-
"
|
|
4135
|
-
id: "
|
|
4136
|
-
name: "
|
|
4356
|
+
"openai/gpt-oss-safeguard-20b": {
|
|
4357
|
+
id: "openai/gpt-oss-safeguard-20b",
|
|
4358
|
+
name: "Safety GPT OSS 20B",
|
|
4359
|
+
api: "openai-completions",
|
|
4360
|
+
provider: "groq",
|
|
4361
|
+
baseUrl: "https://api.groq.com/openai/v1",
|
|
4362
|
+
reasoning: true,
|
|
4363
|
+
input: ["text"],
|
|
4364
|
+
cost: {
|
|
4365
|
+
input: 0.075,
|
|
4366
|
+
output: 0.3,
|
|
4367
|
+
cacheRead: 0.037,
|
|
4368
|
+
cacheWrite: 0,
|
|
4369
|
+
},
|
|
4370
|
+
contextWindow: 131072,
|
|
4371
|
+
maxTokens: 65536,
|
|
4372
|
+
},
|
|
4373
|
+
"qwen-qwq-32b": {
|
|
4374
|
+
id: "qwen-qwq-32b",
|
|
4375
|
+
name: "Qwen QwQ 32B",
|
|
4137
4376
|
api: "openai-completions",
|
|
4138
4377
|
provider: "groq",
|
|
4139
4378
|
baseUrl: "https://api.groq.com/openai/v1",
|
|
@@ -4163,7 +4402,7 @@ export const MODELS = {
|
|
|
4163
4402
|
cacheWrite: 0,
|
|
4164
4403
|
},
|
|
4165
4404
|
contextWindow: 131072,
|
|
4166
|
-
maxTokens:
|
|
4405
|
+
maxTokens: 40960,
|
|
4167
4406
|
},
|
|
4168
4407
|
},
|
|
4169
4408
|
"huggingface": {
|
|
@@ -4203,6 +4442,24 @@ export const MODELS = {
|
|
|
4203
4442
|
contextWindow: 204800,
|
|
4204
4443
|
maxTokens: 131072,
|
|
4205
4444
|
},
|
|
4445
|
+
"MiniMaxAI/MiniMax-M2.7": {
|
|
4446
|
+
id: "MiniMaxAI/MiniMax-M2.7",
|
|
4447
|
+
name: "MiniMax-M2.7",
|
|
4448
|
+
api: "openai-completions",
|
|
4449
|
+
provider: "huggingface",
|
|
4450
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
4451
|
+
compat: { "supportsDeveloperRole": false },
|
|
4452
|
+
reasoning: true,
|
|
4453
|
+
input: ["text"],
|
|
4454
|
+
cost: {
|
|
4455
|
+
input: 0.3,
|
|
4456
|
+
output: 1.2,
|
|
4457
|
+
cacheRead: 0.06,
|
|
4458
|
+
cacheWrite: 0,
|
|
4459
|
+
},
|
|
4460
|
+
contextWindow: 204800,
|
|
4461
|
+
maxTokens: 131072,
|
|
4462
|
+
},
|
|
4206
4463
|
"Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4207
4464
|
id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
|
|
4208
4465
|
name: "Qwen3-235B-A22B-Thinking-2507",
|
|
@@ -4437,6 +4694,24 @@ export const MODELS = {
|
|
|
4437
4694
|
contextWindow: 262144,
|
|
4438
4695
|
maxTokens: 262144,
|
|
4439
4696
|
},
|
|
4697
|
+
"moonshotai/Kimi-K2.6": {
|
|
4698
|
+
id: "moonshotai/Kimi-K2.6",
|
|
4699
|
+
name: "Kimi-K2.6",
|
|
4700
|
+
api: "openai-completions",
|
|
4701
|
+
provider: "huggingface",
|
|
4702
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
4703
|
+
compat: { "supportsDeveloperRole": false },
|
|
4704
|
+
reasoning: true,
|
|
4705
|
+
input: ["text", "image"],
|
|
4706
|
+
cost: {
|
|
4707
|
+
input: 0.95,
|
|
4708
|
+
output: 4,
|
|
4709
|
+
cacheRead: 0.16,
|
|
4710
|
+
cacheWrite: 0,
|
|
4711
|
+
},
|
|
4712
|
+
contextWindow: 262144,
|
|
4713
|
+
maxTokens: 262144,
|
|
4714
|
+
},
|
|
4440
4715
|
"zai-org/GLM-4.7": {
|
|
4441
4716
|
id: "zai-org/GLM-4.7",
|
|
4442
4717
|
name: "GLM-4.7",
|
|
@@ -4491,6 +4766,24 @@ export const MODELS = {
|
|
|
4491
4766
|
contextWindow: 202752,
|
|
4492
4767
|
maxTokens: 131072,
|
|
4493
4768
|
},
|
|
4769
|
+
"zai-org/GLM-5.1": {
|
|
4770
|
+
id: "zai-org/GLM-5.1",
|
|
4771
|
+
name: "GLM-5.1",
|
|
4772
|
+
api: "openai-completions",
|
|
4773
|
+
provider: "huggingface",
|
|
4774
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
4775
|
+
compat: { "supportsDeveloperRole": false },
|
|
4776
|
+
reasoning: true,
|
|
4777
|
+
input: ["text"],
|
|
4778
|
+
cost: {
|
|
4779
|
+
input: 1,
|
|
4780
|
+
output: 3.2,
|
|
4781
|
+
cacheRead: 0.2,
|
|
4782
|
+
cacheWrite: 0,
|
|
4783
|
+
},
|
|
4784
|
+
contextWindow: 202752,
|
|
4785
|
+
maxTokens: 131072,
|
|
4786
|
+
},
|
|
4494
4787
|
},
|
|
4495
4788
|
"kimi-coding": {
|
|
4496
4789
|
"k2p5": {
|
|
@@ -4510,14 +4803,14 @@ export const MODELS = {
|
|
|
4510
4803
|
contextWindow: 262144,
|
|
4511
4804
|
maxTokens: 32768,
|
|
4512
4805
|
},
|
|
4513
|
-
"
|
|
4514
|
-
id: "
|
|
4515
|
-
name: "Kimi K2
|
|
4806
|
+
"k2p6": {
|
|
4807
|
+
id: "k2p6",
|
|
4808
|
+
name: "Kimi K2.6",
|
|
4516
4809
|
api: "anthropic-messages",
|
|
4517
4810
|
provider: "kimi-coding",
|
|
4518
4811
|
baseUrl: "https://api.kimi.com/coding",
|
|
4519
4812
|
reasoning: true,
|
|
4520
|
-
input: ["text"],
|
|
4813
|
+
input: ["text", "image"],
|
|
4521
4814
|
cost: {
|
|
4522
4815
|
input: 0,
|
|
4523
4816
|
output: 0,
|
|
@@ -4527,93 +4820,25 @@ export const MODELS = {
|
|
|
4527
4820
|
contextWindow: 262144,
|
|
4528
4821
|
maxTokens: 32768,
|
|
4529
4822
|
},
|
|
4530
|
-
|
|
4531
|
-
|
|
4532
|
-
|
|
4533
|
-
id: "MiniMax-M2",
|
|
4534
|
-
name: "MiniMax-M2",
|
|
4535
|
-
api: "anthropic-messages",
|
|
4536
|
-
provider: "minimax",
|
|
4537
|
-
baseUrl: "https://api.minimax.io/anthropic",
|
|
4538
|
-
reasoning: true,
|
|
4539
|
-
input: ["text"],
|
|
4540
|
-
cost: {
|
|
4541
|
-
input: 0.3,
|
|
4542
|
-
output: 1.2,
|
|
4543
|
-
cacheRead: 0,
|
|
4544
|
-
cacheWrite: 0,
|
|
4545
|
-
},
|
|
4546
|
-
contextWindow: 204800,
|
|
4547
|
-
maxTokens: 131072,
|
|
4548
|
-
},
|
|
4549
|
-
"MiniMax-M2.1": {
|
|
4550
|
-
id: "MiniMax-M2.1",
|
|
4551
|
-
name: "MiniMax-M2.1",
|
|
4552
|
-
api: "anthropic-messages",
|
|
4553
|
-
provider: "minimax",
|
|
4554
|
-
baseUrl: "https://api.minimax.io/anthropic",
|
|
4555
|
-
reasoning: true,
|
|
4556
|
-
input: ["text"],
|
|
4557
|
-
cost: {
|
|
4558
|
-
input: 0.3,
|
|
4559
|
-
output: 1.2,
|
|
4560
|
-
cacheRead: 0,
|
|
4561
|
-
cacheWrite: 0,
|
|
4562
|
-
},
|
|
4563
|
-
contextWindow: 204800,
|
|
4564
|
-
maxTokens: 131072,
|
|
4565
|
-
},
|
|
4566
|
-
"MiniMax-M2.1-highspeed": {
|
|
4567
|
-
id: "MiniMax-M2.1-highspeed",
|
|
4568
|
-
name: "MiniMax-M2.1-highspeed",
|
|
4823
|
+
"kimi-k2-thinking": {
|
|
4824
|
+
id: "kimi-k2-thinking",
|
|
4825
|
+
name: "Kimi K2 Thinking",
|
|
4569
4826
|
api: "anthropic-messages",
|
|
4570
|
-
provider: "
|
|
4571
|
-
baseUrl: "https://api.
|
|
4827
|
+
provider: "kimi-coding",
|
|
4828
|
+
baseUrl: "https://api.kimi.com/coding",
|
|
4572
4829
|
reasoning: true,
|
|
4573
4830
|
input: ["text"],
|
|
4574
4831
|
cost: {
|
|
4575
|
-
input: 0
|
|
4576
|
-
output:
|
|
4832
|
+
input: 0,
|
|
4833
|
+
output: 0,
|
|
4577
4834
|
cacheRead: 0,
|
|
4578
4835
|
cacheWrite: 0,
|
|
4579
4836
|
},
|
|
4580
|
-
contextWindow:
|
|
4581
|
-
maxTokens:
|
|
4582
|
-
},
|
|
4583
|
-
"MiniMax-M2.5": {
|
|
4584
|
-
id: "MiniMax-M2.5",
|
|
4585
|
-
name: "MiniMax-M2.5",
|
|
4586
|
-
api: "anthropic-messages",
|
|
4587
|
-
provider: "minimax",
|
|
4588
|
-
baseUrl: "https://api.minimax.io/anthropic",
|
|
4589
|
-
reasoning: true,
|
|
4590
|
-
input: ["text"],
|
|
4591
|
-
cost: {
|
|
4592
|
-
input: 0.3,
|
|
4593
|
-
output: 1.2,
|
|
4594
|
-
cacheRead: 0.03,
|
|
4595
|
-
cacheWrite: 0.375,
|
|
4596
|
-
},
|
|
4597
|
-
contextWindow: 204800,
|
|
4598
|
-
maxTokens: 131072,
|
|
4599
|
-
},
|
|
4600
|
-
"MiniMax-M2.5-highspeed": {
|
|
4601
|
-
id: "MiniMax-M2.5-highspeed",
|
|
4602
|
-
name: "MiniMax-M2.5-highspeed",
|
|
4603
|
-
api: "anthropic-messages",
|
|
4604
|
-
provider: "minimax",
|
|
4605
|
-
baseUrl: "https://api.minimax.io/anthropic",
|
|
4606
|
-
reasoning: true,
|
|
4607
|
-
input: ["text"],
|
|
4608
|
-
cost: {
|
|
4609
|
-
input: 0.6,
|
|
4610
|
-
output: 2.4,
|
|
4611
|
-
cacheRead: 0.06,
|
|
4612
|
-
cacheWrite: 0.375,
|
|
4613
|
-
},
|
|
4614
|
-
contextWindow: 204800,
|
|
4615
|
-
maxTokens: 131072,
|
|
4837
|
+
contextWindow: 262144,
|
|
4838
|
+
maxTokens: 32768,
|
|
4616
4839
|
},
|
|
4840
|
+
},
|
|
4841
|
+
"minimax": {
|
|
4617
4842
|
"MiniMax-M2.7": {
|
|
4618
4843
|
id: "MiniMax-M2.7",
|
|
4619
4844
|
name: "MiniMax-M2.7",
|
|
@@ -4650,91 +4875,6 @@ export const MODELS = {
|
|
|
4650
4875
|
},
|
|
4651
4876
|
},
|
|
4652
4877
|
"minimax-cn": {
|
|
4653
|
-
"MiniMax-M2": {
|
|
4654
|
-
id: "MiniMax-M2",
|
|
4655
|
-
name: "MiniMax-M2",
|
|
4656
|
-
api: "anthropic-messages",
|
|
4657
|
-
provider: "minimax-cn",
|
|
4658
|
-
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4659
|
-
reasoning: true,
|
|
4660
|
-
input: ["text"],
|
|
4661
|
-
cost: {
|
|
4662
|
-
input: 0.3,
|
|
4663
|
-
output: 1.2,
|
|
4664
|
-
cacheRead: 0,
|
|
4665
|
-
cacheWrite: 0,
|
|
4666
|
-
},
|
|
4667
|
-
contextWindow: 204800,
|
|
4668
|
-
maxTokens: 131072,
|
|
4669
|
-
},
|
|
4670
|
-
"MiniMax-M2.1": {
|
|
4671
|
-
id: "MiniMax-M2.1",
|
|
4672
|
-
name: "MiniMax-M2.1",
|
|
4673
|
-
api: "anthropic-messages",
|
|
4674
|
-
provider: "minimax-cn",
|
|
4675
|
-
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4676
|
-
reasoning: true,
|
|
4677
|
-
input: ["text"],
|
|
4678
|
-
cost: {
|
|
4679
|
-
input: 0.3,
|
|
4680
|
-
output: 1.2,
|
|
4681
|
-
cacheRead: 0,
|
|
4682
|
-
cacheWrite: 0,
|
|
4683
|
-
},
|
|
4684
|
-
contextWindow: 204800,
|
|
4685
|
-
maxTokens: 131072,
|
|
4686
|
-
},
|
|
4687
|
-
"MiniMax-M2.1-highspeed": {
|
|
4688
|
-
id: "MiniMax-M2.1-highspeed",
|
|
4689
|
-
name: "MiniMax-M2.1-highspeed",
|
|
4690
|
-
api: "anthropic-messages",
|
|
4691
|
-
provider: "minimax-cn",
|
|
4692
|
-
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4693
|
-
reasoning: true,
|
|
4694
|
-
input: ["text"],
|
|
4695
|
-
cost: {
|
|
4696
|
-
input: 0.6,
|
|
4697
|
-
output: 2.4,
|
|
4698
|
-
cacheRead: 0,
|
|
4699
|
-
cacheWrite: 0,
|
|
4700
|
-
},
|
|
4701
|
-
contextWindow: 204800,
|
|
4702
|
-
maxTokens: 131072,
|
|
4703
|
-
},
|
|
4704
|
-
"MiniMax-M2.5": {
|
|
4705
|
-
id: "MiniMax-M2.5",
|
|
4706
|
-
name: "MiniMax-M2.5",
|
|
4707
|
-
api: "anthropic-messages",
|
|
4708
|
-
provider: "minimax-cn",
|
|
4709
|
-
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4710
|
-
reasoning: true,
|
|
4711
|
-
input: ["text"],
|
|
4712
|
-
cost: {
|
|
4713
|
-
input: 0.3,
|
|
4714
|
-
output: 1.2,
|
|
4715
|
-
cacheRead: 0.03,
|
|
4716
|
-
cacheWrite: 0.375,
|
|
4717
|
-
},
|
|
4718
|
-
contextWindow: 204800,
|
|
4719
|
-
maxTokens: 131072,
|
|
4720
|
-
},
|
|
4721
|
-
"MiniMax-M2.5-highspeed": {
|
|
4722
|
-
id: "MiniMax-M2.5-highspeed",
|
|
4723
|
-
name: "MiniMax-M2.5-highspeed",
|
|
4724
|
-
api: "anthropic-messages",
|
|
4725
|
-
provider: "minimax-cn",
|
|
4726
|
-
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
4727
|
-
reasoning: true,
|
|
4728
|
-
input: ["text"],
|
|
4729
|
-
cost: {
|
|
4730
|
-
input: 0.6,
|
|
4731
|
-
output: 2.4,
|
|
4732
|
-
cacheRead: 0.06,
|
|
4733
|
-
cacheWrite: 0.375,
|
|
4734
|
-
},
|
|
4735
|
-
contextWindow: 204800,
|
|
4736
|
-
maxTokens: 131072,
|
|
4737
|
-
},
|
|
4738
4878
|
"MiniMax-M2.7": {
|
|
4739
4879
|
id: "MiniMax-M2.7",
|
|
4740
4880
|
name: "MiniMax-M2.7",
|
|
@@ -4774,9 +4914,9 @@ export const MODELS = {
|
|
|
4774
4914
|
"codestral-latest": {
|
|
4775
4915
|
id: "codestral-latest",
|
|
4776
4916
|
name: "Codestral (latest)",
|
|
4777
|
-
api: "
|
|
4917
|
+
api: "openai-completions",
|
|
4778
4918
|
provider: "mistral",
|
|
4779
|
-
baseUrl: "https://api.mistral.ai",
|
|
4919
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4780
4920
|
reasoning: false,
|
|
4781
4921
|
input: ["text"],
|
|
4782
4922
|
cost: {
|
|
@@ -4791,9 +4931,9 @@ export const MODELS = {
|
|
|
4791
4931
|
"devstral-2512": {
|
|
4792
4932
|
id: "devstral-2512",
|
|
4793
4933
|
name: "Devstral 2",
|
|
4794
|
-
api: "
|
|
4934
|
+
api: "openai-completions",
|
|
4795
4935
|
provider: "mistral",
|
|
4796
|
-
baseUrl: "https://api.mistral.ai",
|
|
4936
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4797
4937
|
reasoning: false,
|
|
4798
4938
|
input: ["text"],
|
|
4799
4939
|
cost: {
|
|
@@ -4808,9 +4948,9 @@ export const MODELS = {
|
|
|
4808
4948
|
"devstral-medium-2507": {
|
|
4809
4949
|
id: "devstral-medium-2507",
|
|
4810
4950
|
name: "Devstral Medium",
|
|
4811
|
-
api: "
|
|
4951
|
+
api: "openai-completions",
|
|
4812
4952
|
provider: "mistral",
|
|
4813
|
-
baseUrl: "https://api.mistral.ai",
|
|
4953
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4814
4954
|
reasoning: false,
|
|
4815
4955
|
input: ["text"],
|
|
4816
4956
|
cost: {
|
|
@@ -4825,9 +4965,9 @@ export const MODELS = {
|
|
|
4825
4965
|
"devstral-medium-latest": {
|
|
4826
4966
|
id: "devstral-medium-latest",
|
|
4827
4967
|
name: "Devstral 2 (latest)",
|
|
4828
|
-
api: "
|
|
4968
|
+
api: "openai-completions",
|
|
4829
4969
|
provider: "mistral",
|
|
4830
|
-
baseUrl: "https://api.mistral.ai",
|
|
4970
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4831
4971
|
reasoning: false,
|
|
4832
4972
|
input: ["text"],
|
|
4833
4973
|
cost: {
|
|
@@ -4842,9 +4982,9 @@ export const MODELS = {
|
|
|
4842
4982
|
"devstral-small-2505": {
|
|
4843
4983
|
id: "devstral-small-2505",
|
|
4844
4984
|
name: "Devstral Small 2505",
|
|
4845
|
-
api: "
|
|
4985
|
+
api: "openai-completions",
|
|
4846
4986
|
provider: "mistral",
|
|
4847
|
-
baseUrl: "https://api.mistral.ai",
|
|
4987
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4848
4988
|
reasoning: false,
|
|
4849
4989
|
input: ["text"],
|
|
4850
4990
|
cost: {
|
|
@@ -4859,9 +4999,9 @@ export const MODELS = {
|
|
|
4859
4999
|
"devstral-small-2507": {
|
|
4860
5000
|
id: "devstral-small-2507",
|
|
4861
5001
|
name: "Devstral Small",
|
|
4862
|
-
api: "
|
|
5002
|
+
api: "openai-completions",
|
|
4863
5003
|
provider: "mistral",
|
|
4864
|
-
baseUrl: "https://api.mistral.ai",
|
|
5004
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4865
5005
|
reasoning: false,
|
|
4866
5006
|
input: ["text"],
|
|
4867
5007
|
cost: {
|
|
@@ -4876,9 +5016,9 @@ export const MODELS = {
|
|
|
4876
5016
|
"labs-devstral-small-2512": {
|
|
4877
5017
|
id: "labs-devstral-small-2512",
|
|
4878
5018
|
name: "Devstral Small 2",
|
|
4879
|
-
api: "
|
|
5019
|
+
api: "openai-completions",
|
|
4880
5020
|
provider: "mistral",
|
|
4881
|
-
baseUrl: "https://api.mistral.ai",
|
|
5021
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4882
5022
|
reasoning: false,
|
|
4883
5023
|
input: ["text", "image"],
|
|
4884
5024
|
cost: {
|
|
@@ -4893,9 +5033,9 @@ export const MODELS = {
|
|
|
4893
5033
|
"magistral-medium-latest": {
|
|
4894
5034
|
id: "magistral-medium-latest",
|
|
4895
5035
|
name: "Magistral Medium (latest)",
|
|
4896
|
-
api: "
|
|
5036
|
+
api: "openai-completions",
|
|
4897
5037
|
provider: "mistral",
|
|
4898
|
-
baseUrl: "https://api.mistral.ai",
|
|
5038
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4899
5039
|
reasoning: true,
|
|
4900
5040
|
input: ["text"],
|
|
4901
5041
|
cost: {
|
|
@@ -4910,9 +5050,9 @@ export const MODELS = {
|
|
|
4910
5050
|
"magistral-small": {
|
|
4911
5051
|
id: "magistral-small",
|
|
4912
5052
|
name: "Magistral Small",
|
|
4913
|
-
api: "
|
|
5053
|
+
api: "openai-completions",
|
|
4914
5054
|
provider: "mistral",
|
|
4915
|
-
baseUrl: "https://api.mistral.ai",
|
|
5055
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4916
5056
|
reasoning: true,
|
|
4917
5057
|
input: ["text"],
|
|
4918
5058
|
cost: {
|
|
@@ -4927,9 +5067,9 @@ export const MODELS = {
|
|
|
4927
5067
|
"ministral-3b-latest": {
|
|
4928
5068
|
id: "ministral-3b-latest",
|
|
4929
5069
|
name: "Ministral 3B (latest)",
|
|
4930
|
-
api: "
|
|
5070
|
+
api: "openai-completions",
|
|
4931
5071
|
provider: "mistral",
|
|
4932
|
-
baseUrl: "https://api.mistral.ai",
|
|
5072
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4933
5073
|
reasoning: false,
|
|
4934
5074
|
input: ["text"],
|
|
4935
5075
|
cost: {
|
|
@@ -4944,9 +5084,9 @@ export const MODELS = {
|
|
|
4944
5084
|
"ministral-8b-latest": {
|
|
4945
5085
|
id: "ministral-8b-latest",
|
|
4946
5086
|
name: "Ministral 8B (latest)",
|
|
4947
|
-
api: "
|
|
5087
|
+
api: "openai-completions",
|
|
4948
5088
|
provider: "mistral",
|
|
4949
|
-
baseUrl: "https://api.mistral.ai",
|
|
5089
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4950
5090
|
reasoning: false,
|
|
4951
5091
|
input: ["text"],
|
|
4952
5092
|
cost: {
|
|
@@ -4961,9 +5101,9 @@ export const MODELS = {
|
|
|
4961
5101
|
"mistral-large-2411": {
|
|
4962
5102
|
id: "mistral-large-2411",
|
|
4963
5103
|
name: "Mistral Large 2.1",
|
|
4964
|
-
api: "
|
|
5104
|
+
api: "openai-completions",
|
|
4965
5105
|
provider: "mistral",
|
|
4966
|
-
baseUrl: "https://api.mistral.ai",
|
|
5106
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4967
5107
|
reasoning: false,
|
|
4968
5108
|
input: ["text"],
|
|
4969
5109
|
cost: {
|
|
@@ -4978,9 +5118,9 @@ export const MODELS = {
|
|
|
4978
5118
|
"mistral-large-2512": {
|
|
4979
5119
|
id: "mistral-large-2512",
|
|
4980
5120
|
name: "Mistral Large 3",
|
|
4981
|
-
api: "
|
|
5121
|
+
api: "openai-completions",
|
|
4982
5122
|
provider: "mistral",
|
|
4983
|
-
baseUrl: "https://api.mistral.ai",
|
|
5123
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
4984
5124
|
reasoning: false,
|
|
4985
5125
|
input: ["text", "image"],
|
|
4986
5126
|
cost: {
|
|
@@ -4995,9 +5135,9 @@ export const MODELS = {
|
|
|
4995
5135
|
"mistral-large-latest": {
|
|
4996
5136
|
id: "mistral-large-latest",
|
|
4997
5137
|
name: "Mistral Large (latest)",
|
|
4998
|
-
api: "
|
|
5138
|
+
api: "openai-completions",
|
|
4999
5139
|
provider: "mistral",
|
|
5000
|
-
baseUrl: "https://api.mistral.ai",
|
|
5140
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5001
5141
|
reasoning: false,
|
|
5002
5142
|
input: ["text", "image"],
|
|
5003
5143
|
cost: {
|
|
@@ -5012,9 +5152,9 @@ export const MODELS = {
|
|
|
5012
5152
|
"mistral-medium-2505": {
|
|
5013
5153
|
id: "mistral-medium-2505",
|
|
5014
5154
|
name: "Mistral Medium 3",
|
|
5015
|
-
api: "
|
|
5155
|
+
api: "openai-completions",
|
|
5016
5156
|
provider: "mistral",
|
|
5017
|
-
baseUrl: "https://api.mistral.ai",
|
|
5157
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5018
5158
|
reasoning: false,
|
|
5019
5159
|
input: ["text", "image"],
|
|
5020
5160
|
cost: {
|
|
@@ -5029,9 +5169,9 @@ export const MODELS = {
|
|
|
5029
5169
|
"mistral-medium-2508": {
|
|
5030
5170
|
id: "mistral-medium-2508",
|
|
5031
5171
|
name: "Mistral Medium 3.1",
|
|
5032
|
-
api: "
|
|
5172
|
+
api: "openai-completions",
|
|
5033
5173
|
provider: "mistral",
|
|
5034
|
-
baseUrl: "https://api.mistral.ai",
|
|
5174
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5035
5175
|
reasoning: false,
|
|
5036
5176
|
input: ["text", "image"],
|
|
5037
5177
|
cost: {
|
|
@@ -5046,9 +5186,9 @@ export const MODELS = {
|
|
|
5046
5186
|
"mistral-medium-latest": {
|
|
5047
5187
|
id: "mistral-medium-latest",
|
|
5048
5188
|
name: "Mistral Medium (latest)",
|
|
5049
|
-
api: "
|
|
5189
|
+
api: "openai-completions",
|
|
5050
5190
|
provider: "mistral",
|
|
5051
|
-
baseUrl: "https://api.mistral.ai",
|
|
5191
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5052
5192
|
reasoning: false,
|
|
5053
5193
|
input: ["text", "image"],
|
|
5054
5194
|
cost: {
|
|
@@ -5063,9 +5203,9 @@ export const MODELS = {
|
|
|
5063
5203
|
"mistral-nemo": {
|
|
5064
5204
|
id: "mistral-nemo",
|
|
5065
5205
|
name: "Mistral Nemo",
|
|
5066
|
-
api: "
|
|
5206
|
+
api: "openai-completions",
|
|
5067
5207
|
provider: "mistral",
|
|
5068
|
-
baseUrl: "https://api.mistral.ai",
|
|
5208
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5069
5209
|
reasoning: false,
|
|
5070
5210
|
input: ["text"],
|
|
5071
5211
|
cost: {
|
|
@@ -5080,9 +5220,9 @@ export const MODELS = {
|
|
|
5080
5220
|
"mistral-small-2506": {
|
|
5081
5221
|
id: "mistral-small-2506",
|
|
5082
5222
|
name: "Mistral Small 3.2",
|
|
5083
|
-
api: "
|
|
5223
|
+
api: "openai-completions",
|
|
5084
5224
|
provider: "mistral",
|
|
5085
|
-
baseUrl: "https://api.mistral.ai",
|
|
5225
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5086
5226
|
reasoning: false,
|
|
5087
5227
|
input: ["text", "image"],
|
|
5088
5228
|
cost: {
|
|
@@ -5094,29 +5234,46 @@ export const MODELS = {
|
|
|
5094
5234
|
contextWindow: 128000,
|
|
5095
5235
|
maxTokens: 16384,
|
|
5096
5236
|
},
|
|
5237
|
+
"mistral-small-2603": {
|
|
5238
|
+
id: "mistral-small-2603",
|
|
5239
|
+
name: "Mistral Small 4",
|
|
5240
|
+
api: "openai-completions",
|
|
5241
|
+
provider: "mistral",
|
|
5242
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5243
|
+
reasoning: true,
|
|
5244
|
+
input: ["text", "image"],
|
|
5245
|
+
cost: {
|
|
5246
|
+
input: 0.15,
|
|
5247
|
+
output: 0.6,
|
|
5248
|
+
cacheRead: 0,
|
|
5249
|
+
cacheWrite: 0,
|
|
5250
|
+
},
|
|
5251
|
+
contextWindow: 256000,
|
|
5252
|
+
maxTokens: 256000,
|
|
5253
|
+
},
|
|
5097
5254
|
"mistral-small-latest": {
|
|
5098
5255
|
id: "mistral-small-latest",
|
|
5099
5256
|
name: "Mistral Small (latest)",
|
|
5100
|
-
api: "
|
|
5257
|
+
api: "openai-completions",
|
|
5101
5258
|
provider: "mistral",
|
|
5102
|
-
baseUrl: "https://api.mistral.ai",
|
|
5103
|
-
reasoning:
|
|
5259
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5260
|
+
reasoning: true,
|
|
5104
5261
|
input: ["text", "image"],
|
|
5105
5262
|
cost: {
|
|
5106
|
-
input: 0.
|
|
5107
|
-
output: 0.
|
|
5263
|
+
input: 0.15,
|
|
5264
|
+
output: 0.6,
|
|
5108
5265
|
cacheRead: 0,
|
|
5109
5266
|
cacheWrite: 0,
|
|
5110
5267
|
},
|
|
5111
|
-
contextWindow:
|
|
5112
|
-
maxTokens:
|
|
5268
|
+
contextWindow: 256000,
|
|
5269
|
+
maxTokens: 256000,
|
|
5113
5270
|
},
|
|
5114
5271
|
"open-mistral-7b": {
|
|
5115
5272
|
id: "open-mistral-7b",
|
|
5116
5273
|
name: "Mistral 7B",
|
|
5117
|
-
api: "
|
|
5274
|
+
api: "openai-completions",
|
|
5118
5275
|
provider: "mistral",
|
|
5119
|
-
baseUrl: "https://api.mistral.ai",
|
|
5276
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5120
5277
|
reasoning: false,
|
|
5121
5278
|
input: ["text"],
|
|
5122
5279
|
cost: {
|
|
@@ -5131,9 +5288,9 @@ export const MODELS = {
|
|
|
5131
5288
|
"open-mixtral-8x22b": {
|
|
5132
5289
|
id: "open-mixtral-8x22b",
|
|
5133
5290
|
name: "Mixtral 8x22B",
|
|
5134
|
-
api: "
|
|
5291
|
+
api: "openai-completions",
|
|
5135
5292
|
provider: "mistral",
|
|
5136
|
-
baseUrl: "https://api.mistral.ai",
|
|
5293
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5137
5294
|
reasoning: false,
|
|
5138
5295
|
input: ["text"],
|
|
5139
5296
|
cost: {
|
|
@@ -5148,9 +5305,9 @@ export const MODELS = {
|
|
|
5148
5305
|
"open-mixtral-8x7b": {
|
|
5149
5306
|
id: "open-mixtral-8x7b",
|
|
5150
5307
|
name: "Mixtral 8x7B",
|
|
5151
|
-
api: "
|
|
5308
|
+
api: "openai-completions",
|
|
5152
5309
|
provider: "mistral",
|
|
5153
|
-
baseUrl: "https://api.mistral.ai",
|
|
5310
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5154
5311
|
reasoning: false,
|
|
5155
5312
|
input: ["text"],
|
|
5156
5313
|
cost: {
|
|
@@ -5165,9 +5322,9 @@ export const MODELS = {
|
|
|
5165
5322
|
"pixtral-12b": {
|
|
5166
5323
|
id: "pixtral-12b",
|
|
5167
5324
|
name: "Pixtral 12B",
|
|
5168
|
-
api: "
|
|
5325
|
+
api: "openai-completions",
|
|
5169
5326
|
provider: "mistral",
|
|
5170
|
-
baseUrl: "https://api.mistral.ai",
|
|
5327
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5171
5328
|
reasoning: false,
|
|
5172
5329
|
input: ["text", "image"],
|
|
5173
5330
|
cost: {
|
|
@@ -5182,9 +5339,9 @@ export const MODELS = {
|
|
|
5182
5339
|
"pixtral-large-latest": {
|
|
5183
5340
|
id: "pixtral-large-latest",
|
|
5184
5341
|
name: "Pixtral Large (latest)",
|
|
5185
|
-
api: "
|
|
5342
|
+
api: "openai-completions",
|
|
5186
5343
|
provider: "mistral",
|
|
5187
|
-
baseUrl: "https://api.mistral.ai",
|
|
5344
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
5188
5345
|
reasoning: false,
|
|
5189
5346
|
input: ["text", "image"],
|
|
5190
5347
|
cost: {
|
|
@@ -5198,23 +5355,6 @@ export const MODELS = {
|
|
|
5198
5355
|
},
|
|
5199
5356
|
},
|
|
5200
5357
|
"openai": {
|
|
5201
|
-
"codex-mini-latest": {
|
|
5202
|
-
id: "codex-mini-latest",
|
|
5203
|
-
name: "Codex Mini",
|
|
5204
|
-
api: "openai-responses",
|
|
5205
|
-
provider: "openai",
|
|
5206
|
-
baseUrl: "https://api.openai.com/v1",
|
|
5207
|
-
reasoning: true,
|
|
5208
|
-
input: ["text"],
|
|
5209
|
-
cost: {
|
|
5210
|
-
input: 1.5,
|
|
5211
|
-
output: 6,
|
|
5212
|
-
cacheRead: 0.375,
|
|
5213
|
-
cacheWrite: 0,
|
|
5214
|
-
},
|
|
5215
|
-
contextWindow: 200000,
|
|
5216
|
-
maxTokens: 100000,
|
|
5217
|
-
},
|
|
5218
5358
|
"gpt-4": {
|
|
5219
5359
|
id: "gpt-4",
|
|
5220
5360
|
name: "GPT-4",
|
|
@@ -5640,6 +5780,23 @@ export const MODELS = {
|
|
|
5640
5780
|
contextWindow: 400000,
|
|
5641
5781
|
maxTokens: 128000,
|
|
5642
5782
|
},
|
|
5783
|
+
"gpt-5.3-chat-latest": {
|
|
5784
|
+
id: "gpt-5.3-chat-latest",
|
|
5785
|
+
name: "GPT-5.3 Chat (latest)",
|
|
5786
|
+
api: "openai-responses",
|
|
5787
|
+
provider: "openai",
|
|
5788
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5789
|
+
reasoning: false,
|
|
5790
|
+
input: ["text", "image"],
|
|
5791
|
+
cost: {
|
|
5792
|
+
input: 1.75,
|
|
5793
|
+
output: 14,
|
|
5794
|
+
cacheRead: 0.175,
|
|
5795
|
+
cacheWrite: 0,
|
|
5796
|
+
},
|
|
5797
|
+
contextWindow: 128000,
|
|
5798
|
+
maxTokens: 16384,
|
|
5799
|
+
},
|
|
5643
5800
|
"gpt-5.3-codex": {
|
|
5644
5801
|
id: "gpt-5.3-codex",
|
|
5645
5802
|
name: "GPT-5.3 Codex",
|
|
@@ -6137,6 +6294,23 @@ export const MODELS = {
|
|
|
6137
6294
|
contextWindow: 1000000,
|
|
6138
6295
|
maxTokens: 128000,
|
|
6139
6296
|
},
|
|
6297
|
+
"claude-opus-4-7": {
|
|
6298
|
+
id: "claude-opus-4-7",
|
|
6299
|
+
name: "Claude Opus 4.7",
|
|
6300
|
+
api: "anthropic-messages",
|
|
6301
|
+
provider: "opencode",
|
|
6302
|
+
baseUrl: "https://opencode.ai/zen",
|
|
6303
|
+
reasoning: true,
|
|
6304
|
+
input: ["text", "image"],
|
|
6305
|
+
cost: {
|
|
6306
|
+
input: 5,
|
|
6307
|
+
output: 25,
|
|
6308
|
+
cacheRead: 0.5,
|
|
6309
|
+
cacheWrite: 6.25,
|
|
6310
|
+
},
|
|
6311
|
+
contextWindow: 1000000,
|
|
6312
|
+
maxTokens: 128000,
|
|
6313
|
+
},
|
|
6140
6314
|
"claude-sonnet-4": {
|
|
6141
6315
|
id: "claude-sonnet-4",
|
|
6142
6316
|
name: "Claude Sonnet 4",
|
|
@@ -6239,6 +6413,23 @@ export const MODELS = {
|
|
|
6239
6413
|
contextWindow: 204800,
|
|
6240
6414
|
maxTokens: 131072,
|
|
6241
6415
|
},
|
|
6416
|
+
"glm-5.1": {
|
|
6417
|
+
id: "glm-5.1",
|
|
6418
|
+
name: "GLM-5.1",
|
|
6419
|
+
api: "openai-completions",
|
|
6420
|
+
provider: "opencode",
|
|
6421
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6422
|
+
reasoning: true,
|
|
6423
|
+
input: ["text"],
|
|
6424
|
+
cost: {
|
|
6425
|
+
input: 1.4,
|
|
6426
|
+
output: 4.4,
|
|
6427
|
+
cacheRead: 0.26,
|
|
6428
|
+
cacheWrite: 0,
|
|
6429
|
+
},
|
|
6430
|
+
contextWindow: 204800,
|
|
6431
|
+
maxTokens: 131072,
|
|
6432
|
+
},
|
|
6242
6433
|
"gpt-5": {
|
|
6243
6434
|
id: "gpt-5",
|
|
6244
6435
|
name: "GPT-5",
|
|
@@ -6477,6 +6668,23 @@ export const MODELS = {
|
|
|
6477
6668
|
contextWindow: 1050000,
|
|
6478
6669
|
maxTokens: 128000,
|
|
6479
6670
|
},
|
|
6671
|
+
"hy3-preview-free": {
|
|
6672
|
+
id: "hy3-preview-free",
|
|
6673
|
+
name: "Hy3 Preview Free",
|
|
6674
|
+
api: "openai-completions",
|
|
6675
|
+
provider: "opencode",
|
|
6676
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6677
|
+
reasoning: true,
|
|
6678
|
+
input: ["text"],
|
|
6679
|
+
cost: {
|
|
6680
|
+
input: 0,
|
|
6681
|
+
output: 0,
|
|
6682
|
+
cacheRead: 0,
|
|
6683
|
+
cacheWrite: 0,
|
|
6684
|
+
},
|
|
6685
|
+
contextWindow: 256000,
|
|
6686
|
+
maxTokens: 64000,
|
|
6687
|
+
},
|
|
6480
6688
|
"kimi-k2.5": {
|
|
6481
6689
|
id: "kimi-k2.5",
|
|
6482
6690
|
name: "Kimi K2.5",
|
|
@@ -6494,30 +6702,30 @@ export const MODELS = {
|
|
|
6494
6702
|
contextWindow: 262144,
|
|
6495
6703
|
maxTokens: 65536,
|
|
6496
6704
|
},
|
|
6497
|
-
"
|
|
6498
|
-
id: "
|
|
6499
|
-
name: "
|
|
6705
|
+
"kimi-k2.6": {
|
|
6706
|
+
id: "kimi-k2.6",
|
|
6707
|
+
name: "Kimi K2.6",
|
|
6500
6708
|
api: "openai-completions",
|
|
6501
6709
|
provider: "opencode",
|
|
6502
6710
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
6503
6711
|
reasoning: true,
|
|
6504
6712
|
input: ["text", "image"],
|
|
6505
6713
|
cost: {
|
|
6506
|
-
input: 0,
|
|
6507
|
-
output:
|
|
6508
|
-
cacheRead: 0,
|
|
6714
|
+
input: 0.95,
|
|
6715
|
+
output: 4,
|
|
6716
|
+
cacheRead: 0.16,
|
|
6509
6717
|
cacheWrite: 0,
|
|
6510
6718
|
},
|
|
6511
6719
|
contextWindow: 262144,
|
|
6512
|
-
maxTokens:
|
|
6720
|
+
maxTokens: 65536,
|
|
6513
6721
|
},
|
|
6514
|
-
"
|
|
6515
|
-
id: "
|
|
6516
|
-
name: "
|
|
6722
|
+
"ling-2.6-flash-free": {
|
|
6723
|
+
id: "ling-2.6-flash-free",
|
|
6724
|
+
name: "Ling 2.6 Flash Free",
|
|
6517
6725
|
api: "openai-completions",
|
|
6518
6726
|
provider: "opencode",
|
|
6519
6727
|
baseUrl: "https://opencode.ai/zen/v1",
|
|
6520
|
-
reasoning:
|
|
6728
|
+
reasoning: false,
|
|
6521
6729
|
input: ["text"],
|
|
6522
6730
|
cost: {
|
|
6523
6731
|
input: 0,
|
|
@@ -6525,8 +6733,8 @@ export const MODELS = {
|
|
|
6525
6733
|
cacheRead: 0,
|
|
6526
6734
|
cacheWrite: 0,
|
|
6527
6735
|
},
|
|
6528
|
-
contextWindow:
|
|
6529
|
-
maxTokens:
|
|
6736
|
+
contextWindow: 262100,
|
|
6737
|
+
maxTokens: 32800,
|
|
6530
6738
|
},
|
|
6531
6739
|
"minimax-m2.5": {
|
|
6532
6740
|
id: "minimax-m2.5",
|
|
@@ -6562,6 +6770,23 @@ export const MODELS = {
|
|
|
6562
6770
|
contextWindow: 204800,
|
|
6563
6771
|
maxTokens: 131072,
|
|
6564
6772
|
},
|
|
6773
|
+
"minimax-m2.7": {
|
|
6774
|
+
id: "minimax-m2.7",
|
|
6775
|
+
name: "MiniMax M2.7",
|
|
6776
|
+
api: "openai-completions",
|
|
6777
|
+
provider: "opencode",
|
|
6778
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6779
|
+
reasoning: true,
|
|
6780
|
+
input: ["text"],
|
|
6781
|
+
cost: {
|
|
6782
|
+
input: 0.3,
|
|
6783
|
+
output: 1.2,
|
|
6784
|
+
cacheRead: 0.06,
|
|
6785
|
+
cacheWrite: 0,
|
|
6786
|
+
},
|
|
6787
|
+
contextWindow: 204800,
|
|
6788
|
+
maxTokens: 131072,
|
|
6789
|
+
},
|
|
6565
6790
|
"nemotron-3-super-free": {
|
|
6566
6791
|
id: "nemotron-3-super-free",
|
|
6567
6792
|
name: "Nemotron 3 Super Free",
|
|
@@ -6576,9 +6801,43 @@ export const MODELS = {
|
|
|
6576
6801
|
cacheRead: 0,
|
|
6577
6802
|
cacheWrite: 0,
|
|
6578
6803
|
},
|
|
6579
|
-
contextWindow:
|
|
6804
|
+
contextWindow: 204800,
|
|
6580
6805
|
maxTokens: 128000,
|
|
6581
6806
|
},
|
|
6807
|
+
"qwen3.5-plus": {
|
|
6808
|
+
id: "qwen3.5-plus",
|
|
6809
|
+
name: "Qwen3.5 Plus",
|
|
6810
|
+
api: "anthropic-messages",
|
|
6811
|
+
provider: "opencode",
|
|
6812
|
+
baseUrl: "https://opencode.ai/zen",
|
|
6813
|
+
reasoning: true,
|
|
6814
|
+
input: ["text", "image"],
|
|
6815
|
+
cost: {
|
|
6816
|
+
input: 0.2,
|
|
6817
|
+
output: 1.2,
|
|
6818
|
+
cacheRead: 0.02,
|
|
6819
|
+
cacheWrite: 0.25,
|
|
6820
|
+
},
|
|
6821
|
+
contextWindow: 262144,
|
|
6822
|
+
maxTokens: 65536,
|
|
6823
|
+
},
|
|
6824
|
+
"qwen3.6-plus": {
|
|
6825
|
+
id: "qwen3.6-plus",
|
|
6826
|
+
name: "Qwen3.6 Plus",
|
|
6827
|
+
api: "anthropic-messages",
|
|
6828
|
+
provider: "opencode",
|
|
6829
|
+
baseUrl: "https://opencode.ai/zen",
|
|
6830
|
+
reasoning: true,
|
|
6831
|
+
input: ["text", "image"],
|
|
6832
|
+
cost: {
|
|
6833
|
+
input: 0.5,
|
|
6834
|
+
output: 3,
|
|
6835
|
+
cacheRead: 0.05,
|
|
6836
|
+
cacheWrite: 0.625,
|
|
6837
|
+
},
|
|
6838
|
+
contextWindow: 262144,
|
|
6839
|
+
maxTokens: 65536,
|
|
6840
|
+
},
|
|
6582
6841
|
},
|
|
6583
6842
|
"opencode-go": {
|
|
6584
6843
|
"glm-5": {
|
|
@@ -6598,6 +6857,23 @@ export const MODELS = {
|
|
|
6598
6857
|
contextWindow: 204800,
|
|
6599
6858
|
maxTokens: 131072,
|
|
6600
6859
|
},
|
|
6860
|
+
"glm-5.1": {
|
|
6861
|
+
id: "glm-5.1",
|
|
6862
|
+
name: "GLM-5.1",
|
|
6863
|
+
api: "openai-completions",
|
|
6864
|
+
provider: "opencode-go",
|
|
6865
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6866
|
+
reasoning: true,
|
|
6867
|
+
input: ["text"],
|
|
6868
|
+
cost: {
|
|
6869
|
+
input: 1.4,
|
|
6870
|
+
output: 4.4,
|
|
6871
|
+
cacheRead: 0.26,
|
|
6872
|
+
cacheWrite: 0,
|
|
6873
|
+
},
|
|
6874
|
+
contextWindow: 204800,
|
|
6875
|
+
maxTokens: 131072,
|
|
6876
|
+
},
|
|
6601
6877
|
"kimi-k2.5": {
|
|
6602
6878
|
id: "kimi-k2.5",
|
|
6603
6879
|
name: "Kimi K2.5",
|
|
@@ -6615,12 +6891,97 @@ export const MODELS = {
|
|
|
6615
6891
|
contextWindow: 262144,
|
|
6616
6892
|
maxTokens: 65536,
|
|
6617
6893
|
},
|
|
6894
|
+
"kimi-k2.6": {
|
|
6895
|
+
id: "kimi-k2.6",
|
|
6896
|
+
name: "Kimi K2.6 (3x limits)",
|
|
6897
|
+
api: "openai-completions",
|
|
6898
|
+
provider: "opencode-go",
|
|
6899
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6900
|
+
reasoning: true,
|
|
6901
|
+
input: ["text", "image"],
|
|
6902
|
+
cost: {
|
|
6903
|
+
input: 0.32,
|
|
6904
|
+
output: 1.34,
|
|
6905
|
+
cacheRead: 0.054,
|
|
6906
|
+
cacheWrite: 0,
|
|
6907
|
+
},
|
|
6908
|
+
contextWindow: 262144,
|
|
6909
|
+
maxTokens: 65536,
|
|
6910
|
+
},
|
|
6911
|
+
"mimo-v2-omni": {
|
|
6912
|
+
id: "mimo-v2-omni",
|
|
6913
|
+
name: "MiMo V2 Omni",
|
|
6914
|
+
api: "openai-completions",
|
|
6915
|
+
provider: "opencode-go",
|
|
6916
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6917
|
+
reasoning: true,
|
|
6918
|
+
input: ["text", "image"],
|
|
6919
|
+
cost: {
|
|
6920
|
+
input: 0.4,
|
|
6921
|
+
output: 2,
|
|
6922
|
+
cacheRead: 0.08,
|
|
6923
|
+
cacheWrite: 0,
|
|
6924
|
+
},
|
|
6925
|
+
contextWindow: 262144,
|
|
6926
|
+
maxTokens: 128000,
|
|
6927
|
+
},
|
|
6928
|
+
"mimo-v2-pro": {
|
|
6929
|
+
id: "mimo-v2-pro",
|
|
6930
|
+
name: "MiMo V2 Pro",
|
|
6931
|
+
api: "openai-completions",
|
|
6932
|
+
provider: "opencode-go",
|
|
6933
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6934
|
+
reasoning: true,
|
|
6935
|
+
input: ["text"],
|
|
6936
|
+
cost: {
|
|
6937
|
+
input: 1,
|
|
6938
|
+
output: 3,
|
|
6939
|
+
cacheRead: 0.2,
|
|
6940
|
+
cacheWrite: 0,
|
|
6941
|
+
},
|
|
6942
|
+
contextWindow: 1048576,
|
|
6943
|
+
maxTokens: 128000,
|
|
6944
|
+
},
|
|
6945
|
+
"mimo-v2.5": {
|
|
6946
|
+
id: "mimo-v2.5",
|
|
6947
|
+
name: "MiMo V2.5",
|
|
6948
|
+
api: "openai-completions",
|
|
6949
|
+
provider: "opencode-go",
|
|
6950
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6951
|
+
reasoning: true,
|
|
6952
|
+
input: ["text", "image"],
|
|
6953
|
+
cost: {
|
|
6954
|
+
input: 0.4,
|
|
6955
|
+
output: 2,
|
|
6956
|
+
cacheRead: 0.08,
|
|
6957
|
+
cacheWrite: 0,
|
|
6958
|
+
},
|
|
6959
|
+
contextWindow: 262144,
|
|
6960
|
+
maxTokens: 128000,
|
|
6961
|
+
},
|
|
6962
|
+
"mimo-v2.5-pro": {
|
|
6963
|
+
id: "mimo-v2.5-pro",
|
|
6964
|
+
name: "MiMo V2.5 Pro",
|
|
6965
|
+
api: "openai-completions",
|
|
6966
|
+
provider: "opencode-go",
|
|
6967
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6968
|
+
reasoning: true,
|
|
6969
|
+
input: ["text"],
|
|
6970
|
+
cost: {
|
|
6971
|
+
input: 1,
|
|
6972
|
+
output: 3,
|
|
6973
|
+
cacheRead: 0.2,
|
|
6974
|
+
cacheWrite: 0,
|
|
6975
|
+
},
|
|
6976
|
+
contextWindow: 1048576,
|
|
6977
|
+
maxTokens: 128000,
|
|
6978
|
+
},
|
|
6618
6979
|
"minimax-m2.5": {
|
|
6619
6980
|
id: "minimax-m2.5",
|
|
6620
6981
|
name: "MiniMax M2.5",
|
|
6621
|
-
api: "
|
|
6982
|
+
api: "openai-completions",
|
|
6622
6983
|
provider: "opencode-go",
|
|
6623
|
-
baseUrl: "https://opencode.ai/zen/go",
|
|
6984
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6624
6985
|
reasoning: true,
|
|
6625
6986
|
input: ["text"],
|
|
6626
6987
|
cost: {
|
|
@@ -6630,7 +6991,7 @@ export const MODELS = {
|
|
|
6630
6991
|
cacheWrite: 0,
|
|
6631
6992
|
},
|
|
6632
6993
|
contextWindow: 204800,
|
|
6633
|
-
maxTokens:
|
|
6994
|
+
maxTokens: 65536,
|
|
6634
6995
|
},
|
|
6635
6996
|
"minimax-m2.7": {
|
|
6636
6997
|
id: "minimax-m2.7",
|
|
@@ -6649,6 +7010,40 @@ export const MODELS = {
|
|
|
6649
7010
|
contextWindow: 204800,
|
|
6650
7011
|
maxTokens: 131072,
|
|
6651
7012
|
},
|
|
7013
|
+
"qwen3.5-plus": {
|
|
7014
|
+
id: "qwen3.5-plus",
|
|
7015
|
+
name: "Qwen3.5 Plus",
|
|
7016
|
+
api: "anthropic-messages",
|
|
7017
|
+
provider: "opencode-go",
|
|
7018
|
+
baseUrl: "https://opencode.ai/zen/go",
|
|
7019
|
+
reasoning: true,
|
|
7020
|
+
input: ["text", "image"],
|
|
7021
|
+
cost: {
|
|
7022
|
+
input: 0.2,
|
|
7023
|
+
output: 1.2,
|
|
7024
|
+
cacheRead: 0.02,
|
|
7025
|
+
cacheWrite: 0.25,
|
|
7026
|
+
},
|
|
7027
|
+
contextWindow: 262144,
|
|
7028
|
+
maxTokens: 65536,
|
|
7029
|
+
},
|
|
7030
|
+
"qwen3.6-plus": {
|
|
7031
|
+
id: "qwen3.6-plus",
|
|
7032
|
+
name: "Qwen3.6 Plus",
|
|
7033
|
+
api: "anthropic-messages",
|
|
7034
|
+
provider: "opencode-go",
|
|
7035
|
+
baseUrl: "https://opencode.ai/zen/go",
|
|
7036
|
+
reasoning: true,
|
|
7037
|
+
input: ["text", "image"],
|
|
7038
|
+
cost: {
|
|
7039
|
+
input: 0.5,
|
|
7040
|
+
output: 3,
|
|
7041
|
+
cacheRead: 0.05,
|
|
7042
|
+
cacheWrite: 0.625,
|
|
7043
|
+
},
|
|
7044
|
+
contextWindow: 262144,
|
|
7045
|
+
maxTokens: 65536,
|
|
7046
|
+
},
|
|
6652
7047
|
},
|
|
6653
7048
|
"openrouter": {
|
|
6654
7049
|
"ai21/jamba-large-1.7": {
|
|
@@ -6821,23 +7216,6 @@ export const MODELS = {
|
|
|
6821
7216
|
contextWindow: 200000,
|
|
6822
7217
|
maxTokens: 8192,
|
|
6823
7218
|
},
|
|
6824
|
-
"anthropic/claude-3.5-sonnet": {
|
|
6825
|
-
id: "anthropic/claude-3.5-sonnet",
|
|
6826
|
-
name: "Anthropic: Claude 3.5 Sonnet",
|
|
6827
|
-
api: "openai-completions",
|
|
6828
|
-
provider: "openrouter",
|
|
6829
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
6830
|
-
reasoning: false,
|
|
6831
|
-
input: ["text", "image"],
|
|
6832
|
-
cost: {
|
|
6833
|
-
input: 6,
|
|
6834
|
-
output: 30,
|
|
6835
|
-
cacheRead: 0.6,
|
|
6836
|
-
cacheWrite: 7.5,
|
|
6837
|
-
},
|
|
6838
|
-
contextWindow: 200000,
|
|
6839
|
-
maxTokens: 8192,
|
|
6840
|
-
},
|
|
6841
7219
|
"anthropic/claude-3.7-sonnet": {
|
|
6842
7220
|
id: "anthropic/claude-3.7-sonnet",
|
|
6843
7221
|
name: "Anthropic: Claude 3.7 Sonnet",
|
|
@@ -6853,7 +7231,7 @@ export const MODELS = {
|
|
|
6853
7231
|
cacheWrite: 3.75,
|
|
6854
7232
|
},
|
|
6855
7233
|
contextWindow: 200000,
|
|
6856
|
-
maxTokens:
|
|
7234
|
+
maxTokens: 128000,
|
|
6857
7235
|
},
|
|
6858
7236
|
"anthropic/claude-3.7-sonnet:thinking": {
|
|
6859
7237
|
id: "anthropic/claude-3.7-sonnet:thinking",
|
|
@@ -6957,6 +7335,40 @@ export const MODELS = {
|
|
|
6957
7335
|
contextWindow: 1000000,
|
|
6958
7336
|
maxTokens: 128000,
|
|
6959
7337
|
},
|
|
7338
|
+
"anthropic/claude-opus-4.6-fast": {
|
|
7339
|
+
id: "anthropic/claude-opus-4.6-fast",
|
|
7340
|
+
name: "Anthropic: Claude Opus 4.6 (Fast)",
|
|
7341
|
+
api: "openai-completions",
|
|
7342
|
+
provider: "openrouter",
|
|
7343
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7344
|
+
reasoning: true,
|
|
7345
|
+
input: ["text", "image"],
|
|
7346
|
+
cost: {
|
|
7347
|
+
input: 30,
|
|
7348
|
+
output: 150,
|
|
7349
|
+
cacheRead: 3,
|
|
7350
|
+
cacheWrite: 37.5,
|
|
7351
|
+
},
|
|
7352
|
+
contextWindow: 1000000,
|
|
7353
|
+
maxTokens: 128000,
|
|
7354
|
+
},
|
|
7355
|
+
"anthropic/claude-opus-4.7": {
|
|
7356
|
+
id: "anthropic/claude-opus-4.7",
|
|
7357
|
+
name: "Anthropic: Claude Opus 4.7",
|
|
7358
|
+
api: "openai-completions",
|
|
7359
|
+
provider: "openrouter",
|
|
7360
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7361
|
+
reasoning: true,
|
|
7362
|
+
input: ["text", "image"],
|
|
7363
|
+
cost: {
|
|
7364
|
+
input: 5,
|
|
7365
|
+
output: 25,
|
|
7366
|
+
cacheRead: 0.5,
|
|
7367
|
+
cacheWrite: 6.25,
|
|
7368
|
+
},
|
|
7369
|
+
contextWindow: 1000000,
|
|
7370
|
+
maxTokens: 128000,
|
|
7371
|
+
},
|
|
6960
7372
|
"anthropic/claude-sonnet-4": {
|
|
6961
7373
|
id: "anthropic/claude-sonnet-4",
|
|
6962
7374
|
name: "Anthropic: Claude Sonnet 4",
|
|
@@ -6971,7 +7383,7 @@ export const MODELS = {
|
|
|
6971
7383
|
cacheRead: 0.3,
|
|
6972
7384
|
cacheWrite: 3.75,
|
|
6973
7385
|
},
|
|
6974
|
-
contextWindow:
|
|
7386
|
+
contextWindow: 1000000,
|
|
6975
7387
|
maxTokens: 64000,
|
|
6976
7388
|
},
|
|
6977
7389
|
"anthropic/claude-sonnet-4.5": {
|
|
@@ -7008,56 +7420,56 @@ export const MODELS = {
|
|
|
7008
7420
|
contextWindow: 1000000,
|
|
7009
7421
|
maxTokens: 128000,
|
|
7010
7422
|
},
|
|
7011
|
-
"arcee-ai/trinity-large-preview
|
|
7012
|
-
id: "arcee-ai/trinity-large-preview
|
|
7013
|
-
name: "Arcee AI: Trinity Large Preview
|
|
7423
|
+
"arcee-ai/trinity-large-preview": {
|
|
7424
|
+
id: "arcee-ai/trinity-large-preview",
|
|
7425
|
+
name: "Arcee AI: Trinity Large Preview",
|
|
7014
7426
|
api: "openai-completions",
|
|
7015
7427
|
provider: "openrouter",
|
|
7016
7428
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7017
7429
|
reasoning: false,
|
|
7018
7430
|
input: ["text"],
|
|
7019
7431
|
cost: {
|
|
7020
|
-
input: 0,
|
|
7021
|
-
output: 0,
|
|
7432
|
+
input: 0.15,
|
|
7433
|
+
output: 0.44999999999999996,
|
|
7022
7434
|
cacheRead: 0,
|
|
7023
7435
|
cacheWrite: 0,
|
|
7024
7436
|
},
|
|
7025
7437
|
contextWindow: 131000,
|
|
7026
7438
|
maxTokens: 4096,
|
|
7027
7439
|
},
|
|
7028
|
-
"arcee-ai/trinity-
|
|
7029
|
-
id: "arcee-ai/trinity-
|
|
7030
|
-
name: "Arcee AI: Trinity
|
|
7440
|
+
"arcee-ai/trinity-large-thinking": {
|
|
7441
|
+
id: "arcee-ai/trinity-large-thinking",
|
|
7442
|
+
name: "Arcee AI: Trinity Large Thinking",
|
|
7031
7443
|
api: "openai-completions",
|
|
7032
7444
|
provider: "openrouter",
|
|
7033
7445
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7034
7446
|
reasoning: true,
|
|
7035
7447
|
input: ["text"],
|
|
7036
7448
|
cost: {
|
|
7037
|
-
input: 0.
|
|
7038
|
-
output: 0.
|
|
7039
|
-
cacheRead: 0,
|
|
7449
|
+
input: 0.22,
|
|
7450
|
+
output: 0.85,
|
|
7451
|
+
cacheRead: 0.06,
|
|
7040
7452
|
cacheWrite: 0,
|
|
7041
7453
|
},
|
|
7042
|
-
contextWindow:
|
|
7043
|
-
maxTokens:
|
|
7454
|
+
contextWindow: 262144,
|
|
7455
|
+
maxTokens: 262144,
|
|
7044
7456
|
},
|
|
7045
|
-
"arcee-ai/trinity-mini
|
|
7046
|
-
id: "arcee-ai/trinity-mini
|
|
7047
|
-
name: "Arcee AI: Trinity Mini
|
|
7457
|
+
"arcee-ai/trinity-mini": {
|
|
7458
|
+
id: "arcee-ai/trinity-mini",
|
|
7459
|
+
name: "Arcee AI: Trinity Mini",
|
|
7048
7460
|
api: "openai-completions",
|
|
7049
7461
|
provider: "openrouter",
|
|
7050
7462
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7051
7463
|
reasoning: true,
|
|
7052
7464
|
input: ["text"],
|
|
7053
7465
|
cost: {
|
|
7054
|
-
input: 0,
|
|
7055
|
-
output: 0,
|
|
7466
|
+
input: 0.045,
|
|
7467
|
+
output: 0.15,
|
|
7056
7468
|
cacheRead: 0,
|
|
7057
7469
|
cacheWrite: 0,
|
|
7058
7470
|
},
|
|
7059
7471
|
contextWindow: 131072,
|
|
7060
|
-
maxTokens:
|
|
7472
|
+
maxTokens: 131072,
|
|
7061
7473
|
},
|
|
7062
7474
|
"arcee-ai/virtuoso-large": {
|
|
7063
7475
|
id: "arcee-ai/virtuoso-large",
|
|
@@ -7252,7 +7664,7 @@ export const MODELS = {
|
|
|
7252
7664
|
api: "openai-completions",
|
|
7253
7665
|
provider: "openrouter",
|
|
7254
7666
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7255
|
-
reasoning:
|
|
7667
|
+
reasoning: false,
|
|
7256
7668
|
input: ["text"],
|
|
7257
7669
|
cost: {
|
|
7258
7670
|
input: 0.19999999999999998,
|
|
@@ -7306,13 +7718,13 @@ export const MODELS = {
|
|
|
7306
7718
|
reasoning: true,
|
|
7307
7719
|
input: ["text"],
|
|
7308
7720
|
cost: {
|
|
7309
|
-
input: 0.
|
|
7721
|
+
input: 0.5,
|
|
7310
7722
|
output: 2.1500000000000004,
|
|
7311
|
-
cacheRead: 0.
|
|
7723
|
+
cacheRead: 0.35,
|
|
7312
7724
|
cacheWrite: 0,
|
|
7313
7725
|
},
|
|
7314
7726
|
contextWindow: 163840,
|
|
7315
|
-
maxTokens:
|
|
7727
|
+
maxTokens: 4096,
|
|
7316
7728
|
},
|
|
7317
7729
|
"deepseek/deepseek-v3.1-terminus": {
|
|
7318
7730
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -7340,13 +7752,13 @@ export const MODELS = {
|
|
|
7340
7752
|
reasoning: true,
|
|
7341
7753
|
input: ["text"],
|
|
7342
7754
|
cost: {
|
|
7343
|
-
input: 0.
|
|
7344
|
-
output: 0.
|
|
7345
|
-
cacheRead: 0.
|
|
7755
|
+
input: 0.252,
|
|
7756
|
+
output: 0.378,
|
|
7757
|
+
cacheRead: 0.0252,
|
|
7346
7758
|
cacheWrite: 0,
|
|
7347
7759
|
},
|
|
7348
|
-
contextWindow:
|
|
7349
|
-
maxTokens:
|
|
7760
|
+
contextWindow: 131072,
|
|
7761
|
+
maxTokens: 65536,
|
|
7350
7762
|
},
|
|
7351
7763
|
"deepseek/deepseek-v3.2-exp": {
|
|
7352
7764
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7396,7 +7808,7 @@ export const MODELS = {
|
|
|
7396
7808
|
cacheRead: 0.024999999999999998,
|
|
7397
7809
|
cacheWrite: 0.08333333333333334,
|
|
7398
7810
|
},
|
|
7399
|
-
contextWindow:
|
|
7811
|
+
contextWindow: 1000000,
|
|
7400
7812
|
maxTokens: 8192,
|
|
7401
7813
|
},
|
|
7402
7814
|
"google/gemini-2.0-flash-lite-001": {
|
|
@@ -7465,7 +7877,7 @@ export const MODELS = {
|
|
|
7465
7877
|
cacheWrite: 0.08333333333333334,
|
|
7466
7878
|
},
|
|
7467
7879
|
contextWindow: 1048576,
|
|
7468
|
-
maxTokens:
|
|
7880
|
+
maxTokens: 65535,
|
|
7469
7881
|
},
|
|
7470
7882
|
"google/gemini-2.5-pro": {
|
|
7471
7883
|
id: "google/gemini-2.5-pro",
|
|
@@ -7535,23 +7947,6 @@ export const MODELS = {
|
|
|
7535
7947
|
contextWindow: 1048576,
|
|
7536
7948
|
maxTokens: 65536,
|
|
7537
7949
|
},
|
|
7538
|
-
"google/gemini-3-pro-preview": {
|
|
7539
|
-
id: "google/gemini-3-pro-preview",
|
|
7540
|
-
name: "Google: Gemini 3 Pro Preview",
|
|
7541
|
-
api: "openai-completions",
|
|
7542
|
-
provider: "openrouter",
|
|
7543
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7544
|
-
reasoning: true,
|
|
7545
|
-
input: ["text", "image"],
|
|
7546
|
-
cost: {
|
|
7547
|
-
input: 2,
|
|
7548
|
-
output: 12,
|
|
7549
|
-
cacheRead: 0.19999999999999998,
|
|
7550
|
-
cacheWrite: 0.375,
|
|
7551
|
-
},
|
|
7552
|
-
contextWindow: 1048576,
|
|
7553
|
-
maxTokens: 65536,
|
|
7554
|
-
},
|
|
7555
7950
|
"google/gemini-3.1-flash-lite-preview": {
|
|
7556
7951
|
id: "google/gemini-3.1-flash-lite-preview",
|
|
7557
7952
|
name: "Google: Gemini 3.1 Flash Lite Preview",
|
|
@@ -7603,22 +7998,73 @@ export const MODELS = {
|
|
|
7603
7998
|
contextWindow: 1048576,
|
|
7604
7999
|
maxTokens: 65536,
|
|
7605
8000
|
},
|
|
7606
|
-
"
|
|
7607
|
-
id: "
|
|
7608
|
-
name: "
|
|
8001
|
+
"google/gemma-4-26b-a4b-it": {
|
|
8002
|
+
id: "google/gemma-4-26b-a4b-it",
|
|
8003
|
+
name: "Google: Gemma 4 26B A4B ",
|
|
8004
|
+
api: "openai-completions",
|
|
8005
|
+
provider: "openrouter",
|
|
8006
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8007
|
+
reasoning: true,
|
|
8008
|
+
input: ["text", "image"],
|
|
8009
|
+
cost: {
|
|
8010
|
+
input: 0.06,
|
|
8011
|
+
output: 0.33,
|
|
8012
|
+
cacheRead: 0,
|
|
8013
|
+
cacheWrite: 0,
|
|
8014
|
+
},
|
|
8015
|
+
contextWindow: 262144,
|
|
8016
|
+
maxTokens: 4096,
|
|
8017
|
+
},
|
|
8018
|
+
"google/gemma-4-26b-a4b-it:free": {
|
|
8019
|
+
id: "google/gemma-4-26b-a4b-it:free",
|
|
8020
|
+
name: "Google: Gemma 4 26B A4B (free)",
|
|
8021
|
+
api: "openai-completions",
|
|
8022
|
+
provider: "openrouter",
|
|
8023
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8024
|
+
reasoning: true,
|
|
8025
|
+
input: ["text", "image"],
|
|
8026
|
+
cost: {
|
|
8027
|
+
input: 0,
|
|
8028
|
+
output: 0,
|
|
8029
|
+
cacheRead: 0,
|
|
8030
|
+
cacheWrite: 0,
|
|
8031
|
+
},
|
|
8032
|
+
contextWindow: 262144,
|
|
8033
|
+
maxTokens: 32768,
|
|
8034
|
+
},
|
|
8035
|
+
"google/gemma-4-31b-it": {
|
|
8036
|
+
id: "google/gemma-4-31b-it",
|
|
8037
|
+
name: "Google: Gemma 4 31B",
|
|
8038
|
+
api: "openai-completions",
|
|
8039
|
+
provider: "openrouter",
|
|
8040
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8041
|
+
reasoning: true,
|
|
8042
|
+
input: ["text", "image"],
|
|
8043
|
+
cost: {
|
|
8044
|
+
input: 0.13,
|
|
8045
|
+
output: 0.38,
|
|
8046
|
+
cacheRead: 0,
|
|
8047
|
+
cacheWrite: 0,
|
|
8048
|
+
},
|
|
8049
|
+
contextWindow: 262144,
|
|
8050
|
+
maxTokens: 4096,
|
|
8051
|
+
},
|
|
8052
|
+
"google/gemma-4-31b-it:free": {
|
|
8053
|
+
id: "google/gemma-4-31b-it:free",
|
|
8054
|
+
name: "Google: Gemma 4 31B (free)",
|
|
7609
8055
|
api: "openai-completions",
|
|
7610
8056
|
provider: "openrouter",
|
|
7611
8057
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7612
|
-
reasoning:
|
|
7613
|
-
input: ["text"],
|
|
8058
|
+
reasoning: true,
|
|
8059
|
+
input: ["text", "image"],
|
|
7614
8060
|
cost: {
|
|
7615
|
-
input: 0
|
|
7616
|
-
output: 0
|
|
7617
|
-
cacheRead: 0
|
|
8061
|
+
input: 0,
|
|
8062
|
+
output: 0,
|
|
8063
|
+
cacheRead: 0,
|
|
7618
8064
|
cacheWrite: 0,
|
|
7619
8065
|
},
|
|
7620
|
-
contextWindow:
|
|
7621
|
-
maxTokens:
|
|
8066
|
+
contextWindow: 262144,
|
|
8067
|
+
maxTokens: 32768,
|
|
7622
8068
|
},
|
|
7623
8069
|
"inception/mercury-2": {
|
|
7624
8070
|
id: "inception/mercury-2",
|
|
@@ -7637,56 +8083,39 @@ export const MODELS = {
|
|
|
7637
8083
|
contextWindow: 128000,
|
|
7638
8084
|
maxTokens: 50000,
|
|
7639
8085
|
},
|
|
7640
|
-
"
|
|
7641
|
-
id: "
|
|
7642
|
-
name: "
|
|
8086
|
+
"inclusionai/ling-2.6-flash:free": {
|
|
8087
|
+
id: "inclusionai/ling-2.6-flash:free",
|
|
8088
|
+
name: "inclusionAI: Ling-2.6-flash (free)",
|
|
7643
8089
|
api: "openai-completions",
|
|
7644
8090
|
provider: "openrouter",
|
|
7645
8091
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7646
8092
|
reasoning: false,
|
|
7647
8093
|
input: ["text"],
|
|
7648
8094
|
cost: {
|
|
7649
|
-
input: 0
|
|
7650
|
-
output: 0
|
|
7651
|
-
cacheRead: 0
|
|
8095
|
+
input: 0,
|
|
8096
|
+
output: 0,
|
|
8097
|
+
cacheRead: 0,
|
|
7652
8098
|
cacheWrite: 0,
|
|
7653
8099
|
},
|
|
7654
|
-
contextWindow:
|
|
7655
|
-
maxTokens:
|
|
8100
|
+
contextWindow: 262144,
|
|
8101
|
+
maxTokens: 32768,
|
|
7656
8102
|
},
|
|
7657
|
-
"kwaipilot/kat-coder-pro": {
|
|
7658
|
-
id: "kwaipilot/kat-coder-pro",
|
|
7659
|
-
name: "Kwaipilot: KAT-Coder-Pro
|
|
8103
|
+
"kwaipilot/kat-coder-pro-v2": {
|
|
8104
|
+
id: "kwaipilot/kat-coder-pro-v2",
|
|
8105
|
+
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
7660
8106
|
api: "openai-completions",
|
|
7661
8107
|
provider: "openrouter",
|
|
7662
8108
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7663
8109
|
reasoning: false,
|
|
7664
8110
|
input: ["text"],
|
|
7665
8111
|
cost: {
|
|
7666
|
-
input: 0.
|
|
7667
|
-
output:
|
|
7668
|
-
cacheRead: 0.
|
|
8112
|
+
input: 0.3,
|
|
8113
|
+
output: 1.2,
|
|
8114
|
+
cacheRead: 0.06,
|
|
7669
8115
|
cacheWrite: 0,
|
|
7670
8116
|
},
|
|
7671
8117
|
contextWindow: 256000,
|
|
7672
|
-
maxTokens:
|
|
7673
|
-
},
|
|
7674
|
-
"meituan/longcat-flash-chat": {
|
|
7675
|
-
id: "meituan/longcat-flash-chat",
|
|
7676
|
-
name: "Meituan: LongCat Flash Chat",
|
|
7677
|
-
api: "openai-completions",
|
|
7678
|
-
provider: "openrouter",
|
|
7679
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7680
|
-
reasoning: false,
|
|
7681
|
-
input: ["text"],
|
|
7682
|
-
cost: {
|
|
7683
|
-
input: 0.19999999999999998,
|
|
7684
|
-
output: 0.7999999999999999,
|
|
7685
|
-
cacheRead: 0.19999999999999998,
|
|
7686
|
-
cacheWrite: 0,
|
|
7687
|
-
},
|
|
7688
|
-
contextWindow: 131072,
|
|
7689
|
-
maxTokens: 131072,
|
|
8118
|
+
maxTokens: 80000,
|
|
7690
8119
|
},
|
|
7691
8120
|
"meta-llama/llama-3-8b-instruct": {
|
|
7692
8121
|
id: "meta-llama/llama-3-8b-instruct",
|
|
@@ -7773,23 +8202,6 @@ export const MODELS = {
|
|
|
7773
8202
|
contextWindow: 65536,
|
|
7774
8203
|
maxTokens: 4096,
|
|
7775
8204
|
},
|
|
7776
|
-
"meta-llama/llama-4-maverick": {
|
|
7777
|
-
id: "meta-llama/llama-4-maverick",
|
|
7778
|
-
name: "Meta: Llama 4 Maverick",
|
|
7779
|
-
api: "openai-completions",
|
|
7780
|
-
provider: "openrouter",
|
|
7781
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7782
|
-
reasoning: false,
|
|
7783
|
-
input: ["text", "image"],
|
|
7784
|
-
cost: {
|
|
7785
|
-
input: 0.15,
|
|
7786
|
-
output: 0.6,
|
|
7787
|
-
cacheRead: 0,
|
|
7788
|
-
cacheWrite: 0,
|
|
7789
|
-
},
|
|
7790
|
-
contextWindow: 1048576,
|
|
7791
|
-
maxTokens: 16384,
|
|
7792
|
-
},
|
|
7793
8205
|
"meta-llama/llama-4-scout": {
|
|
7794
8206
|
id: "meta-llama/llama-4-scout",
|
|
7795
8207
|
name: "Meta: Llama 4 Scout",
|
|
@@ -7850,13 +8262,13 @@ export const MODELS = {
|
|
|
7850
8262
|
reasoning: true,
|
|
7851
8263
|
input: ["text"],
|
|
7852
8264
|
cost: {
|
|
7853
|
-
input: 0.
|
|
8265
|
+
input: 0.29,
|
|
7854
8266
|
output: 0.95,
|
|
7855
|
-
cacheRead: 0.
|
|
8267
|
+
cacheRead: 0.03,
|
|
7856
8268
|
cacheWrite: 0,
|
|
7857
8269
|
},
|
|
7858
8270
|
contextWindow: 196608,
|
|
7859
|
-
maxTokens:
|
|
8271
|
+
maxTokens: 196608,
|
|
7860
8272
|
},
|
|
7861
8273
|
"minimax/minimax-m2.5": {
|
|
7862
8274
|
id: "minimax/minimax-m2.5",
|
|
@@ -7867,9 +8279,9 @@ export const MODELS = {
|
|
|
7867
8279
|
reasoning: true,
|
|
7868
8280
|
input: ["text"],
|
|
7869
8281
|
cost: {
|
|
7870
|
-
input: 0.
|
|
7871
|
-
output: 1.
|
|
7872
|
-
cacheRead: 0.
|
|
8282
|
+
input: 0.15,
|
|
8283
|
+
output: 1.2,
|
|
8284
|
+
cacheRead: 0.075,
|
|
7873
8285
|
cacheWrite: 0,
|
|
7874
8286
|
},
|
|
7875
8287
|
contextWindow: 196608,
|
|
@@ -7890,7 +8302,7 @@ export const MODELS = {
|
|
|
7890
8302
|
cacheWrite: 0,
|
|
7891
8303
|
},
|
|
7892
8304
|
contextWindow: 196608,
|
|
7893
|
-
maxTokens:
|
|
8305
|
+
maxTokens: 8192,
|
|
7894
8306
|
},
|
|
7895
8307
|
"minimax/minimax-m2.7": {
|
|
7896
8308
|
id: "minimax/minimax-m2.7",
|
|
@@ -7903,11 +8315,11 @@ export const MODELS = {
|
|
|
7903
8315
|
cost: {
|
|
7904
8316
|
input: 0.3,
|
|
7905
8317
|
output: 1.2,
|
|
7906
|
-
cacheRead: 0.
|
|
8318
|
+
cacheRead: 0.059,
|
|
7907
8319
|
cacheWrite: 0,
|
|
7908
8320
|
},
|
|
7909
|
-
contextWindow:
|
|
7910
|
-
maxTokens:
|
|
8321
|
+
contextWindow: 196608,
|
|
8322
|
+
maxTokens: 4096,
|
|
7911
8323
|
},
|
|
7912
8324
|
"mistralai/codestral-2508": {
|
|
7913
8325
|
id: "mistralai/codestral-2508",
|
|
@@ -8139,13 +8551,13 @@ export const MODELS = {
|
|
|
8139
8551
|
reasoning: false,
|
|
8140
8552
|
input: ["text"],
|
|
8141
8553
|
cost: {
|
|
8142
|
-
input: 0.
|
|
8143
|
-
output: 0.
|
|
8554
|
+
input: 0.01,
|
|
8555
|
+
output: 0.03,
|
|
8144
8556
|
cacheRead: 0,
|
|
8145
8557
|
cacheWrite: 0,
|
|
8146
8558
|
},
|
|
8147
8559
|
contextWindow: 131072,
|
|
8148
|
-
maxTokens:
|
|
8560
|
+
maxTokens: 4096,
|
|
8149
8561
|
},
|
|
8150
8562
|
"mistralai/mistral-saba": {
|
|
8151
8563
|
id: "mistralai/mistral-saba",
|
|
@@ -8164,23 +8576,6 @@ export const MODELS = {
|
|
|
8164
8576
|
contextWindow: 32768,
|
|
8165
8577
|
maxTokens: 4096,
|
|
8166
8578
|
},
|
|
8167
|
-
"mistralai/mistral-small-24b-instruct-2501": {
|
|
8168
|
-
id: "mistralai/mistral-small-24b-instruct-2501",
|
|
8169
|
-
name: "Mistral: Mistral Small 3",
|
|
8170
|
-
api: "openai-completions",
|
|
8171
|
-
provider: "openrouter",
|
|
8172
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8173
|
-
reasoning: false,
|
|
8174
|
-
input: ["text"],
|
|
8175
|
-
cost: {
|
|
8176
|
-
input: 0.049999999999999996,
|
|
8177
|
-
output: 0.08,
|
|
8178
|
-
cacheRead: 0,
|
|
8179
|
-
cacheWrite: 0,
|
|
8180
|
-
},
|
|
8181
|
-
contextWindow: 32768,
|
|
8182
|
-
maxTokens: 16384,
|
|
8183
|
-
},
|
|
8184
8579
|
"mistralai/mistral-small-2603": {
|
|
8185
8580
|
id: "mistralai/mistral-small-2603",
|
|
8186
8581
|
name: "Mistral: Mistral Small 4",
|
|
@@ -8198,23 +8593,6 @@ export const MODELS = {
|
|
|
8198
8593
|
contextWindow: 262144,
|
|
8199
8594
|
maxTokens: 4096,
|
|
8200
8595
|
},
|
|
8201
|
-
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8202
|
-
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
8203
|
-
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
8204
|
-
api: "openai-completions",
|
|
8205
|
-
provider: "openrouter",
|
|
8206
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8207
|
-
reasoning: false,
|
|
8208
|
-
input: ["text", "image"],
|
|
8209
|
-
cost: {
|
|
8210
|
-
input: 0,
|
|
8211
|
-
output: 0,
|
|
8212
|
-
cacheRead: 0,
|
|
8213
|
-
cacheWrite: 0,
|
|
8214
|
-
},
|
|
8215
|
-
contextWindow: 128000,
|
|
8216
|
-
maxTokens: 4096,
|
|
8217
|
-
},
|
|
8218
8596
|
"mistralai/mistral-small-3.2-24b-instruct": {
|
|
8219
8597
|
id: "mistralai/mistral-small-3.2-24b-instruct",
|
|
8220
8598
|
name: "Mistral: Mistral Small 3.2 24B",
|
|
@@ -8332,7 +8710,7 @@ export const MODELS = {
|
|
|
8332
8710
|
cacheWrite: 0,
|
|
8333
8711
|
},
|
|
8334
8712
|
contextWindow: 131072,
|
|
8335
|
-
maxTokens:
|
|
8713
|
+
maxTokens: 32768,
|
|
8336
8714
|
},
|
|
8337
8715
|
"moonshotai/kimi-k2-0905": {
|
|
8338
8716
|
id: "moonshotai/kimi-k2-0905",
|
|
@@ -8345,11 +8723,11 @@ export const MODELS = {
|
|
|
8345
8723
|
cost: {
|
|
8346
8724
|
input: 0.39999999999999997,
|
|
8347
8725
|
output: 2,
|
|
8348
|
-
cacheRead: 0
|
|
8726
|
+
cacheRead: 0,
|
|
8349
8727
|
cacheWrite: 0,
|
|
8350
8728
|
},
|
|
8351
|
-
contextWindow:
|
|
8352
|
-
maxTokens:
|
|
8729
|
+
contextWindow: 262144,
|
|
8730
|
+
maxTokens: 262144,
|
|
8353
8731
|
},
|
|
8354
8732
|
"moonshotai/kimi-k2-thinking": {
|
|
8355
8733
|
id: "moonshotai/kimi-k2-thinking",
|
|
@@ -8360,13 +8738,13 @@ export const MODELS = {
|
|
|
8360
8738
|
reasoning: true,
|
|
8361
8739
|
input: ["text"],
|
|
8362
8740
|
cost: {
|
|
8363
|
-
input: 0.
|
|
8364
|
-
output: 2,
|
|
8365
|
-
cacheRead: 0.
|
|
8741
|
+
input: 0.6,
|
|
8742
|
+
output: 2.5,
|
|
8743
|
+
cacheRead: 0.15,
|
|
8366
8744
|
cacheWrite: 0,
|
|
8367
8745
|
},
|
|
8368
|
-
contextWindow:
|
|
8369
|
-
maxTokens:
|
|
8746
|
+
contextWindow: 262144,
|
|
8747
|
+
maxTokens: 262144,
|
|
8370
8748
|
},
|
|
8371
8749
|
"moonshotai/kimi-k2.5": {
|
|
8372
8750
|
id: "moonshotai/kimi-k2.5",
|
|
@@ -8385,6 +8763,23 @@ export const MODELS = {
|
|
|
8385
8763
|
contextWindow: 262144,
|
|
8386
8764
|
maxTokens: 4096,
|
|
8387
8765
|
},
|
|
8766
|
+
"moonshotai/kimi-k2.6": {
|
|
8767
|
+
id: "moonshotai/kimi-k2.6",
|
|
8768
|
+
name: "MoonshotAI: Kimi K2.6",
|
|
8769
|
+
api: "openai-completions",
|
|
8770
|
+
provider: "openrouter",
|
|
8771
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8772
|
+
reasoning: true,
|
|
8773
|
+
input: ["text", "image"],
|
|
8774
|
+
cost: {
|
|
8775
|
+
input: 0.56,
|
|
8776
|
+
output: 3.5,
|
|
8777
|
+
cacheRead: 0.11,
|
|
8778
|
+
cacheWrite: 0,
|
|
8779
|
+
},
|
|
8780
|
+
contextWindow: 256000,
|
|
8781
|
+
maxTokens: 65536,
|
|
8782
|
+
},
|
|
8388
8783
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
8389
8784
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
8390
8785
|
name: "Nex AGI: DeepSeek V3.1 Nex N1",
|
|
@@ -8479,9 +8874,9 @@ export const MODELS = {
|
|
|
8479
8874
|
reasoning: true,
|
|
8480
8875
|
input: ["text"],
|
|
8481
8876
|
cost: {
|
|
8482
|
-
input: 0.
|
|
8483
|
-
output: 0.
|
|
8484
|
-
cacheRead: 0
|
|
8877
|
+
input: 0.09,
|
|
8878
|
+
output: 0.44999999999999996,
|
|
8879
|
+
cacheRead: 0,
|
|
8485
8880
|
cacheWrite: 0,
|
|
8486
8881
|
},
|
|
8487
8882
|
contextWindow: 262144,
|
|
@@ -8706,7 +9101,7 @@ export const MODELS = {
|
|
|
8706
9101
|
cacheWrite: 0,
|
|
8707
9102
|
},
|
|
8708
9103
|
contextWindow: 1047576,
|
|
8709
|
-
maxTokens:
|
|
9104
|
+
maxTokens: 4096,
|
|
8710
9105
|
},
|
|
8711
9106
|
"openai/gpt-4.1-mini": {
|
|
8712
9107
|
id: "openai/gpt-4.1-mini",
|
|
@@ -8753,7 +9148,7 @@ export const MODELS = {
|
|
|
8753
9148
|
cost: {
|
|
8754
9149
|
input: 2.5,
|
|
8755
9150
|
output: 10,
|
|
8756
|
-
cacheRead:
|
|
9151
|
+
cacheRead: 0,
|
|
8757
9152
|
cacheWrite: 0,
|
|
8758
9153
|
},
|
|
8759
9154
|
contextWindow: 128000,
|
|
@@ -8861,23 +9256,6 @@ export const MODELS = {
|
|
|
8861
9256
|
contextWindow: 128000,
|
|
8862
9257
|
maxTokens: 16384,
|
|
8863
9258
|
},
|
|
8864
|
-
"openai/gpt-4o:extended": {
|
|
8865
|
-
id: "openai/gpt-4o:extended",
|
|
8866
|
-
name: "OpenAI: GPT-4o (extended)",
|
|
8867
|
-
api: "openai-completions",
|
|
8868
|
-
provider: "openrouter",
|
|
8869
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8870
|
-
reasoning: false,
|
|
8871
|
-
input: ["text", "image"],
|
|
8872
|
-
cost: {
|
|
8873
|
-
input: 6,
|
|
8874
|
-
output: 18,
|
|
8875
|
-
cacheRead: 0,
|
|
8876
|
-
cacheWrite: 0,
|
|
8877
|
-
},
|
|
8878
|
-
contextWindow: 128000,
|
|
8879
|
-
maxTokens: 64000,
|
|
8880
|
-
},
|
|
8881
9259
|
"openai/gpt-5": {
|
|
8882
9260
|
id: "openai/gpt-5",
|
|
8883
9261
|
name: "OpenAI: GPT-5",
|
|
@@ -8912,40 +9290,6 @@ export const MODELS = {
|
|
|
8912
9290
|
contextWindow: 400000,
|
|
8913
9291
|
maxTokens: 128000,
|
|
8914
9292
|
},
|
|
8915
|
-
"openai/gpt-5-image": {
|
|
8916
|
-
id: "openai/gpt-5-image",
|
|
8917
|
-
name: "OpenAI: GPT-5 Image",
|
|
8918
|
-
api: "openai-completions",
|
|
8919
|
-
provider: "openrouter",
|
|
8920
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8921
|
-
reasoning: true,
|
|
8922
|
-
input: ["text", "image"],
|
|
8923
|
-
cost: {
|
|
8924
|
-
input: 10,
|
|
8925
|
-
output: 10,
|
|
8926
|
-
cacheRead: 1.25,
|
|
8927
|
-
cacheWrite: 0,
|
|
8928
|
-
},
|
|
8929
|
-
contextWindow: 400000,
|
|
8930
|
-
maxTokens: 128000,
|
|
8931
|
-
},
|
|
8932
|
-
"openai/gpt-5-image-mini": {
|
|
8933
|
-
id: "openai/gpt-5-image-mini",
|
|
8934
|
-
name: "OpenAI: GPT-5 Image Mini",
|
|
8935
|
-
api: "openai-completions",
|
|
8936
|
-
provider: "openrouter",
|
|
8937
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8938
|
-
reasoning: true,
|
|
8939
|
-
input: ["text", "image"],
|
|
8940
|
-
cost: {
|
|
8941
|
-
input: 2.5,
|
|
8942
|
-
output: 2,
|
|
8943
|
-
cacheRead: 0.25,
|
|
8944
|
-
cacheWrite: 0,
|
|
8945
|
-
},
|
|
8946
|
-
contextWindow: 400000,
|
|
8947
|
-
maxTokens: 128000,
|
|
8948
|
-
},
|
|
8949
9293
|
"openai/gpt-5-mini": {
|
|
8950
9294
|
id: "openai/gpt-5-mini",
|
|
8951
9295
|
name: "OpenAI: GPT-5 Mini",
|
|
@@ -8974,11 +9318,11 @@ export const MODELS = {
|
|
|
8974
9318
|
cost: {
|
|
8975
9319
|
input: 0.049999999999999996,
|
|
8976
9320
|
output: 0.39999999999999997,
|
|
8977
|
-
cacheRead: 0.
|
|
9321
|
+
cacheRead: 0.01,
|
|
8978
9322
|
cacheWrite: 0,
|
|
8979
9323
|
},
|
|
8980
9324
|
contextWindow: 400000,
|
|
8981
|
-
maxTokens:
|
|
9325
|
+
maxTokens: 4096,
|
|
8982
9326
|
},
|
|
8983
9327
|
"openai/gpt-5-pro": {
|
|
8984
9328
|
id: "openai/gpt-5-pro",
|
|
@@ -9008,7 +9352,7 @@ export const MODELS = {
|
|
|
9008
9352
|
cost: {
|
|
9009
9353
|
input: 1.25,
|
|
9010
9354
|
output: 10,
|
|
9011
|
-
cacheRead: 0.
|
|
9355
|
+
cacheRead: 0.13,
|
|
9012
9356
|
cacheWrite: 0,
|
|
9013
9357
|
},
|
|
9014
9358
|
contextWindow: 400000,
|
|
@@ -9076,11 +9420,11 @@ export const MODELS = {
|
|
|
9076
9420
|
cost: {
|
|
9077
9421
|
input: 0.25,
|
|
9078
9422
|
output: 2,
|
|
9079
|
-
cacheRead: 0.
|
|
9423
|
+
cacheRead: 0.03,
|
|
9080
9424
|
cacheWrite: 0,
|
|
9081
9425
|
},
|
|
9082
9426
|
contextWindow: 400000,
|
|
9083
|
-
maxTokens:
|
|
9427
|
+
maxTokens: 128000,
|
|
9084
9428
|
},
|
|
9085
9429
|
"openai/gpt-5.2": {
|
|
9086
9430
|
id: "openai/gpt-5.2",
|
|
@@ -9114,7 +9458,7 @@ export const MODELS = {
|
|
|
9114
9458
|
cacheWrite: 0,
|
|
9115
9459
|
},
|
|
9116
9460
|
contextWindow: 128000,
|
|
9117
|
-
maxTokens:
|
|
9461
|
+
maxTokens: 32000,
|
|
9118
9462
|
},
|
|
9119
9463
|
"openai/gpt-5.2-codex": {
|
|
9120
9464
|
id: "openai/gpt-5.2-codex",
|
|
@@ -9252,6 +9596,40 @@ export const MODELS = {
|
|
|
9252
9596
|
contextWindow: 1050000,
|
|
9253
9597
|
maxTokens: 128000,
|
|
9254
9598
|
},
|
|
9599
|
+
"openai/gpt-audio": {
|
|
9600
|
+
id: "openai/gpt-audio",
|
|
9601
|
+
name: "OpenAI: GPT Audio",
|
|
9602
|
+
api: "openai-completions",
|
|
9603
|
+
provider: "openrouter",
|
|
9604
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9605
|
+
reasoning: false,
|
|
9606
|
+
input: ["text"],
|
|
9607
|
+
cost: {
|
|
9608
|
+
input: 2.5,
|
|
9609
|
+
output: 10,
|
|
9610
|
+
cacheRead: 0,
|
|
9611
|
+
cacheWrite: 0,
|
|
9612
|
+
},
|
|
9613
|
+
contextWindow: 128000,
|
|
9614
|
+
maxTokens: 16384,
|
|
9615
|
+
},
|
|
9616
|
+
"openai/gpt-audio-mini": {
|
|
9617
|
+
id: "openai/gpt-audio-mini",
|
|
9618
|
+
name: "OpenAI: GPT Audio Mini",
|
|
9619
|
+
api: "openai-completions",
|
|
9620
|
+
provider: "openrouter",
|
|
9621
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9622
|
+
reasoning: false,
|
|
9623
|
+
input: ["text"],
|
|
9624
|
+
cost: {
|
|
9625
|
+
input: 0.6,
|
|
9626
|
+
output: 2.4,
|
|
9627
|
+
cacheRead: 0,
|
|
9628
|
+
cacheWrite: 0,
|
|
9629
|
+
},
|
|
9630
|
+
contextWindow: 128000,
|
|
9631
|
+
maxTokens: 16384,
|
|
9632
|
+
},
|
|
9255
9633
|
"openai/gpt-oss-120b": {
|
|
9256
9634
|
id: "openai/gpt-oss-120b",
|
|
9257
9635
|
name: "OpenAI: gpt-oss-120b",
|
|
@@ -9296,12 +9674,12 @@ export const MODELS = {
|
|
|
9296
9674
|
input: ["text"],
|
|
9297
9675
|
cost: {
|
|
9298
9676
|
input: 0.03,
|
|
9299
|
-
output: 0.
|
|
9300
|
-
cacheRead: 0
|
|
9677
|
+
output: 0.14,
|
|
9678
|
+
cacheRead: 0,
|
|
9301
9679
|
cacheWrite: 0,
|
|
9302
9680
|
},
|
|
9303
9681
|
contextWindow: 131072,
|
|
9304
|
-
maxTokens:
|
|
9682
|
+
maxTokens: 4096,
|
|
9305
9683
|
},
|
|
9306
9684
|
"openai/gpt-oss-20b:free": {
|
|
9307
9685
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -9318,7 +9696,7 @@ export const MODELS = {
|
|
|
9318
9696
|
cacheWrite: 0,
|
|
9319
9697
|
},
|
|
9320
9698
|
contextWindow: 131072,
|
|
9321
|
-
maxTokens:
|
|
9699
|
+
maxTokens: 8192,
|
|
9322
9700
|
},
|
|
9323
9701
|
"openai/gpt-oss-safeguard-20b": {
|
|
9324
9702
|
id: "openai/gpt-oss-safeguard-20b",
|
|
@@ -9604,7 +9982,7 @@ export const MODELS = {
|
|
|
9604
9982
|
input: 0.26,
|
|
9605
9983
|
output: 0.78,
|
|
9606
9984
|
cacheRead: 0.052000000000000005,
|
|
9607
|
-
cacheWrite: 0,
|
|
9985
|
+
cacheWrite: 0.325,
|
|
9608
9986
|
},
|
|
9609
9987
|
contextWindow: 1000000,
|
|
9610
9988
|
maxTokens: 32768,
|
|
@@ -9621,7 +9999,7 @@ export const MODELS = {
|
|
|
9621
9999
|
input: 0.26,
|
|
9622
10000
|
output: 0.78,
|
|
9623
10001
|
cacheRead: 0,
|
|
9624
|
-
cacheWrite: 0,
|
|
10002
|
+
cacheWrite: 0.325,
|
|
9625
10003
|
},
|
|
9626
10004
|
contextWindow: 1000000,
|
|
9627
10005
|
maxTokens: 32768,
|
|
@@ -9638,7 +10016,7 @@ export const MODELS = {
|
|
|
9638
10016
|
input: 0.26,
|
|
9639
10017
|
output: 0.78,
|
|
9640
10018
|
cacheRead: 0,
|
|
9641
|
-
cacheWrite: 0,
|
|
10019
|
+
cacheWrite: 0.325,
|
|
9642
10020
|
},
|
|
9643
10021
|
contextWindow: 1000000,
|
|
9644
10022
|
maxTokens: 32768,
|
|
@@ -9717,7 +10095,7 @@ export const MODELS = {
|
|
|
9717
10095
|
api: "openai-completions",
|
|
9718
10096
|
provider: "openrouter",
|
|
9719
10097
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9720
|
-
reasoning:
|
|
10098
|
+
reasoning: false,
|
|
9721
10099
|
input: ["text"],
|
|
9722
10100
|
cost: {
|
|
9723
10101
|
input: 0.071,
|
|
@@ -9793,42 +10171,25 @@ export const MODELS = {
|
|
|
9793
10171
|
cacheRead: 0.08,
|
|
9794
10172
|
cacheWrite: 0,
|
|
9795
10173
|
},
|
|
9796
|
-
contextWindow: 131072,
|
|
9797
|
-
maxTokens: 131072,
|
|
9798
|
-
},
|
|
9799
|
-
"qwen/qwen3-32b": {
|
|
9800
|
-
id: "qwen/qwen3-32b",
|
|
9801
|
-
name: "Qwen: Qwen3 32B",
|
|
9802
|
-
api: "openai-completions",
|
|
9803
|
-
provider: "openrouter",
|
|
9804
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9805
|
-
reasoning: true,
|
|
9806
|
-
input: ["text"],
|
|
9807
|
-
cost: {
|
|
9808
|
-
input: 0.08,
|
|
9809
|
-
output: 0.24,
|
|
9810
|
-
cacheRead: 0.04,
|
|
9811
|
-
cacheWrite: 0,
|
|
9812
|
-
},
|
|
9813
|
-
contextWindow: 40960,
|
|
9814
|
-
maxTokens: 40960,
|
|
10174
|
+
contextWindow: 131072,
|
|
10175
|
+
maxTokens: 131072,
|
|
9815
10176
|
},
|
|
9816
|
-
"qwen/qwen3-
|
|
9817
|
-
id: "qwen/qwen3-
|
|
9818
|
-
name: "Qwen: Qwen3
|
|
10177
|
+
"qwen/qwen3-32b": {
|
|
10178
|
+
id: "qwen/qwen3-32b",
|
|
10179
|
+
name: "Qwen: Qwen3 32B",
|
|
9819
10180
|
api: "openai-completions",
|
|
9820
10181
|
provider: "openrouter",
|
|
9821
10182
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9822
10183
|
reasoning: true,
|
|
9823
10184
|
input: ["text"],
|
|
9824
10185
|
cost: {
|
|
9825
|
-
input: 0,
|
|
9826
|
-
output: 0,
|
|
9827
|
-
cacheRead: 0,
|
|
10186
|
+
input: 0.08,
|
|
10187
|
+
output: 0.24,
|
|
10188
|
+
cacheRead: 0.04,
|
|
9828
10189
|
cacheWrite: 0,
|
|
9829
10190
|
},
|
|
9830
10191
|
contextWindow: 40960,
|
|
9831
|
-
maxTokens:
|
|
10192
|
+
maxTokens: 40960,
|
|
9832
10193
|
},
|
|
9833
10194
|
"qwen/qwen3-8b": {
|
|
9834
10195
|
id: "qwen/qwen3-8b",
|
|
@@ -9893,7 +10254,7 @@ export const MODELS = {
|
|
|
9893
10254
|
input: 0.195,
|
|
9894
10255
|
output: 0.975,
|
|
9895
10256
|
cacheRead: 0.039,
|
|
9896
|
-
cacheWrite: 0,
|
|
10257
|
+
cacheWrite: 0.24375,
|
|
9897
10258
|
},
|
|
9898
10259
|
contextWindow: 1000000,
|
|
9899
10260
|
maxTokens: 65536,
|
|
@@ -9907,13 +10268,13 @@ export const MODELS = {
|
|
|
9907
10268
|
reasoning: false,
|
|
9908
10269
|
input: ["text"],
|
|
9909
10270
|
cost: {
|
|
9910
|
-
input: 0.
|
|
9911
|
-
output: 0.
|
|
9912
|
-
cacheRead: 0.
|
|
10271
|
+
input: 0.15,
|
|
10272
|
+
output: 0.7999999999999999,
|
|
10273
|
+
cacheRead: 0.11,
|
|
9913
10274
|
cacheWrite: 0,
|
|
9914
10275
|
},
|
|
9915
10276
|
contextWindow: 262144,
|
|
9916
|
-
maxTokens:
|
|
10277
|
+
maxTokens: 262144,
|
|
9917
10278
|
},
|
|
9918
10279
|
"qwen/qwen3-coder-plus": {
|
|
9919
10280
|
id: "qwen/qwen3-coder-plus",
|
|
@@ -9927,7 +10288,7 @@ export const MODELS = {
|
|
|
9927
10288
|
input: 0.65,
|
|
9928
10289
|
output: 3.25,
|
|
9929
10290
|
cacheRead: 0.13,
|
|
9930
|
-
cacheWrite: 0,
|
|
10291
|
+
cacheWrite: 0.8125,
|
|
9931
10292
|
},
|
|
9932
10293
|
contextWindow: 1000000,
|
|
9933
10294
|
maxTokens: 65536,
|
|
@@ -9961,7 +10322,7 @@ export const MODELS = {
|
|
|
9961
10322
|
input: 0.78,
|
|
9962
10323
|
output: 3.9,
|
|
9963
10324
|
cacheRead: 0.156,
|
|
9964
|
-
cacheWrite: 0,
|
|
10325
|
+
cacheWrite: 0.975,
|
|
9965
10326
|
},
|
|
9966
10327
|
contextWindow: 262144,
|
|
9967
10328
|
maxTokens: 32768,
|
|
@@ -10215,7 +10576,7 @@ export const MODELS = {
|
|
|
10215
10576
|
cost: {
|
|
10216
10577
|
input: 0.39,
|
|
10217
10578
|
output: 2.34,
|
|
10218
|
-
cacheRead: 0,
|
|
10579
|
+
cacheRead: 0.195,
|
|
10219
10580
|
cacheWrite: 0,
|
|
10220
10581
|
},
|
|
10221
10582
|
contextWindow: 262144,
|
|
@@ -10230,13 +10591,13 @@ export const MODELS = {
|
|
|
10230
10591
|
reasoning: true,
|
|
10231
10592
|
input: ["text", "image"],
|
|
10232
10593
|
cost: {
|
|
10233
|
-
input: 0.
|
|
10594
|
+
input: 0.09999999999999999,
|
|
10234
10595
|
output: 0.15,
|
|
10235
10596
|
cacheRead: 0,
|
|
10236
10597
|
cacheWrite: 0,
|
|
10237
10598
|
},
|
|
10238
|
-
contextWindow:
|
|
10239
|
-
maxTokens:
|
|
10599
|
+
contextWindow: 262144,
|
|
10600
|
+
maxTokens: 4096,
|
|
10240
10601
|
},
|
|
10241
10602
|
"qwen/qwen3.5-flash-02-23": {
|
|
10242
10603
|
id: "qwen/qwen3.5-flash-02-23",
|
|
@@ -10250,7 +10611,7 @@ export const MODELS = {
|
|
|
10250
10611
|
input: 0.065,
|
|
10251
10612
|
output: 0.26,
|
|
10252
10613
|
cacheRead: 0,
|
|
10253
|
-
cacheWrite: 0,
|
|
10614
|
+
cacheWrite: 0.08125,
|
|
10254
10615
|
},
|
|
10255
10616
|
contextWindow: 1000000,
|
|
10256
10617
|
maxTokens: 65536,
|
|
@@ -10267,7 +10628,24 @@ export const MODELS = {
|
|
|
10267
10628
|
input: 0.26,
|
|
10268
10629
|
output: 1.56,
|
|
10269
10630
|
cacheRead: 0,
|
|
10270
|
-
cacheWrite: 0,
|
|
10631
|
+
cacheWrite: 0.325,
|
|
10632
|
+
},
|
|
10633
|
+
contextWindow: 1000000,
|
|
10634
|
+
maxTokens: 65536,
|
|
10635
|
+
},
|
|
10636
|
+
"qwen/qwen3.6-plus": {
|
|
10637
|
+
id: "qwen/qwen3.6-plus",
|
|
10638
|
+
name: "Qwen: Qwen3.6 Plus",
|
|
10639
|
+
api: "openai-completions",
|
|
10640
|
+
provider: "openrouter",
|
|
10641
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10642
|
+
reasoning: true,
|
|
10643
|
+
input: ["text", "image"],
|
|
10644
|
+
cost: {
|
|
10645
|
+
input: 0.325,
|
|
10646
|
+
output: 1.95,
|
|
10647
|
+
cacheRead: 0,
|
|
10648
|
+
cacheWrite: 0.40625,
|
|
10271
10649
|
},
|
|
10272
10650
|
contextWindow: 1000000,
|
|
10273
10651
|
maxTokens: 65536,
|
|
@@ -10289,6 +10667,23 @@ export const MODELS = {
|
|
|
10289
10667
|
contextWindow: 131072,
|
|
10290
10668
|
maxTokens: 131072,
|
|
10291
10669
|
},
|
|
10670
|
+
"rekaai/reka-edge": {
|
|
10671
|
+
id: "rekaai/reka-edge",
|
|
10672
|
+
name: "Reka Edge",
|
|
10673
|
+
api: "openai-completions",
|
|
10674
|
+
provider: "openrouter",
|
|
10675
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10676
|
+
reasoning: false,
|
|
10677
|
+
input: ["text", "image"],
|
|
10678
|
+
cost: {
|
|
10679
|
+
input: 0.09999999999999999,
|
|
10680
|
+
output: 0.09999999999999999,
|
|
10681
|
+
cacheRead: 0,
|
|
10682
|
+
cacheWrite: 0,
|
|
10683
|
+
},
|
|
10684
|
+
contextWindow: 16384,
|
|
10685
|
+
maxTokens: 16384,
|
|
10686
|
+
},
|
|
10292
10687
|
"relace/relace-search": {
|
|
10293
10688
|
id: "relace/relace-search",
|
|
10294
10689
|
name: "Relace: Relace Search",
|
|
@@ -10351,15 +10746,15 @@ export const MODELS = {
|
|
|
10351
10746
|
cost: {
|
|
10352
10747
|
input: 0.09999999999999999,
|
|
10353
10748
|
output: 0.3,
|
|
10354
|
-
cacheRead: 0
|
|
10749
|
+
cacheRead: 0,
|
|
10355
10750
|
cacheWrite: 0,
|
|
10356
10751
|
},
|
|
10357
|
-
contextWindow:
|
|
10358
|
-
maxTokens:
|
|
10752
|
+
contextWindow: 262144,
|
|
10753
|
+
maxTokens: 65536,
|
|
10359
10754
|
},
|
|
10360
|
-
"
|
|
10361
|
-
id: "
|
|
10362
|
-
name: "
|
|
10755
|
+
"tencent/hy3-preview:free": {
|
|
10756
|
+
id: "tencent/hy3-preview:free",
|
|
10757
|
+
name: "Tencent: Hy3 preview (free)",
|
|
10363
10758
|
api: "openai-completions",
|
|
10364
10759
|
provider: "openrouter",
|
|
10365
10760
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -10371,8 +10766,8 @@ export const MODELS = {
|
|
|
10371
10766
|
cacheRead: 0,
|
|
10372
10767
|
cacheWrite: 0,
|
|
10373
10768
|
},
|
|
10374
|
-
contextWindow:
|
|
10375
|
-
maxTokens:
|
|
10769
|
+
contextWindow: 262144,
|
|
10770
|
+
maxTokens: 262144,
|
|
10376
10771
|
},
|
|
10377
10772
|
"thedrummer/rocinante-12b": {
|
|
10378
10773
|
id: "thedrummer/rocinante-12b",
|
|
@@ -10561,9 +10956,9 @@ export const MODELS = {
|
|
|
10561
10956
|
contextWindow: 2000000,
|
|
10562
10957
|
maxTokens: 30000,
|
|
10563
10958
|
},
|
|
10564
|
-
"x-ai/grok-4.20
|
|
10565
|
-
id: "x-ai/grok-4.20
|
|
10566
|
-
name: "xAI: Grok 4.20
|
|
10959
|
+
"x-ai/grok-4.20": {
|
|
10960
|
+
id: "x-ai/grok-4.20",
|
|
10961
|
+
name: "xAI: Grok 4.20",
|
|
10567
10962
|
api: "openai-completions",
|
|
10568
10963
|
provider: "openrouter",
|
|
10569
10964
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -10646,6 +11041,40 @@ export const MODELS = {
|
|
|
10646
11041
|
contextWindow: 1048576,
|
|
10647
11042
|
maxTokens: 131072,
|
|
10648
11043
|
},
|
|
11044
|
+
"xiaomi/mimo-v2.5": {
|
|
11045
|
+
id: "xiaomi/mimo-v2.5",
|
|
11046
|
+
name: "Xiaomi: MiMo-V2.5",
|
|
11047
|
+
api: "openai-completions",
|
|
11048
|
+
provider: "openrouter",
|
|
11049
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11050
|
+
reasoning: true,
|
|
11051
|
+
input: ["text", "image"],
|
|
11052
|
+
cost: {
|
|
11053
|
+
input: 0.39999999999999997,
|
|
11054
|
+
output: 2,
|
|
11055
|
+
cacheRead: 0.08,
|
|
11056
|
+
cacheWrite: 0,
|
|
11057
|
+
},
|
|
11058
|
+
contextWindow: 1048576,
|
|
11059
|
+
maxTokens: 131072,
|
|
11060
|
+
},
|
|
11061
|
+
"xiaomi/mimo-v2.5-pro": {
|
|
11062
|
+
id: "xiaomi/mimo-v2.5-pro",
|
|
11063
|
+
name: "Xiaomi: MiMo-V2.5-Pro",
|
|
11064
|
+
api: "openai-completions",
|
|
11065
|
+
provider: "openrouter",
|
|
11066
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11067
|
+
reasoning: true,
|
|
11068
|
+
input: ["text"],
|
|
11069
|
+
cost: {
|
|
11070
|
+
input: 1,
|
|
11071
|
+
output: 3,
|
|
11072
|
+
cacheRead: 0.19999999999999998,
|
|
11073
|
+
cacheWrite: 0,
|
|
11074
|
+
},
|
|
11075
|
+
contextWindow: 1048576,
|
|
11076
|
+
maxTokens: 131072,
|
|
11077
|
+
},
|
|
10649
11078
|
"z-ai/glm-4-32b": {
|
|
10650
11079
|
id: "z-ai/glm-4-32b",
|
|
10651
11080
|
name: "Z.ai: GLM 4 32B ",
|
|
@@ -10774,13 +11203,13 @@ export const MODELS = {
|
|
|
10774
11203
|
reasoning: true,
|
|
10775
11204
|
input: ["text"],
|
|
10776
11205
|
cost: {
|
|
10777
|
-
input: 0.
|
|
10778
|
-
output: 1.
|
|
10779
|
-
cacheRead: 0
|
|
11206
|
+
input: 0.38,
|
|
11207
|
+
output: 1.74,
|
|
11208
|
+
cacheRead: 0,
|
|
10780
11209
|
cacheWrite: 0,
|
|
10781
11210
|
},
|
|
10782
11211
|
contextWindow: 202752,
|
|
10783
|
-
maxTokens:
|
|
11212
|
+
maxTokens: 4096,
|
|
10784
11213
|
},
|
|
10785
11214
|
"z-ai/glm-4.7-flash": {
|
|
10786
11215
|
id: "z-ai/glm-4.7-flash",
|
|
@@ -10813,8 +11242,8 @@ export const MODELS = {
|
|
|
10813
11242
|
cacheRead: 0.119,
|
|
10814
11243
|
cacheWrite: 0,
|
|
10815
11244
|
},
|
|
10816
|
-
contextWindow:
|
|
10817
|
-
maxTokens:
|
|
11245
|
+
contextWindow: 202752,
|
|
11246
|
+
maxTokens: 4096,
|
|
10818
11247
|
},
|
|
10819
11248
|
"z-ai/glm-5-turbo": {
|
|
10820
11249
|
id: "z-ai/glm-5-turbo",
|
|
@@ -10833,6 +11262,57 @@ export const MODELS = {
|
|
|
10833
11262
|
contextWindow: 202752,
|
|
10834
11263
|
maxTokens: 131072,
|
|
10835
11264
|
},
|
|
11265
|
+
"z-ai/glm-5.1": {
|
|
11266
|
+
id: "z-ai/glm-5.1",
|
|
11267
|
+
name: "Z.ai: GLM 5.1",
|
|
11268
|
+
api: "openai-completions",
|
|
11269
|
+
provider: "openrouter",
|
|
11270
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11271
|
+
reasoning: true,
|
|
11272
|
+
input: ["text"],
|
|
11273
|
+
cost: {
|
|
11274
|
+
input: 1.0499999999999998,
|
|
11275
|
+
output: 3.5,
|
|
11276
|
+
cacheRead: 0.5249999999999999,
|
|
11277
|
+
cacheWrite: 0,
|
|
11278
|
+
},
|
|
11279
|
+
contextWindow: 202752,
|
|
11280
|
+
maxTokens: 65535,
|
|
11281
|
+
},
|
|
11282
|
+
"z-ai/glm-5v-turbo": {
|
|
11283
|
+
id: "z-ai/glm-5v-turbo",
|
|
11284
|
+
name: "Z.ai: GLM 5V Turbo",
|
|
11285
|
+
api: "openai-completions",
|
|
11286
|
+
provider: "openrouter",
|
|
11287
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11288
|
+
reasoning: true,
|
|
11289
|
+
input: ["text", "image"],
|
|
11290
|
+
cost: {
|
|
11291
|
+
input: 1.2,
|
|
11292
|
+
output: 4,
|
|
11293
|
+
cacheRead: 0.24,
|
|
11294
|
+
cacheWrite: 0,
|
|
11295
|
+
},
|
|
11296
|
+
contextWindow: 202752,
|
|
11297
|
+
maxTokens: 131072,
|
|
11298
|
+
},
|
|
11299
|
+
"~anthropic/claude-opus-latest": {
|
|
11300
|
+
id: "~anthropic/claude-opus-latest",
|
|
11301
|
+
name: "Anthropic: Claude Opus Latest",
|
|
11302
|
+
api: "openai-completions",
|
|
11303
|
+
provider: "openrouter",
|
|
11304
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11305
|
+
reasoning: true,
|
|
11306
|
+
input: ["text", "image"],
|
|
11307
|
+
cost: {
|
|
11308
|
+
input: 5,
|
|
11309
|
+
output: 25,
|
|
11310
|
+
cacheRead: 0.5,
|
|
11311
|
+
cacheWrite: 6.25,
|
|
11312
|
+
},
|
|
11313
|
+
contextWindow: 1000000,
|
|
11314
|
+
maxTokens: 128000,
|
|
11315
|
+
},
|
|
10836
11316
|
},
|
|
10837
11317
|
"vercel-ai-gateway": {
|
|
10838
11318
|
"alibaba/qwen-3-14b": {
|
|
@@ -10854,20 +11334,20 @@ export const MODELS = {
|
|
|
10854
11334
|
},
|
|
10855
11335
|
"alibaba/qwen-3-235b": {
|
|
10856
11336
|
id: "alibaba/qwen-3-235b",
|
|
10857
|
-
name: "Qwen3
|
|
11337
|
+
name: "Qwen3 235B A22b Instruct 2507",
|
|
10858
11338
|
api: "anthropic-messages",
|
|
10859
11339
|
provider: "vercel-ai-gateway",
|
|
10860
11340
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10861
11341
|
reasoning: false,
|
|
10862
11342
|
input: ["text"],
|
|
10863
11343
|
cost: {
|
|
10864
|
-
input: 0.
|
|
10865
|
-
output:
|
|
10866
|
-
cacheRead: 0,
|
|
11344
|
+
input: 0.6,
|
|
11345
|
+
output: 1.2,
|
|
11346
|
+
cacheRead: 0.6,
|
|
10867
11347
|
cacheWrite: 0,
|
|
10868
11348
|
},
|
|
10869
|
-
contextWindow:
|
|
10870
|
-
maxTokens:
|
|
11349
|
+
contextWindow: 131000,
|
|
11350
|
+
maxTokens: 40000,
|
|
10871
11351
|
},
|
|
10872
11352
|
"alibaba/qwen-3-30b": {
|
|
10873
11353
|
id: "alibaba/qwen-3-30b",
|
|
@@ -10895,13 +11375,30 @@ export const MODELS = {
|
|
|
10895
11375
|
reasoning: true,
|
|
10896
11376
|
input: ["text"],
|
|
10897
11377
|
cost: {
|
|
10898
|
-
input: 0.
|
|
10899
|
-
output: 0.
|
|
10900
|
-
cacheRead: 0
|
|
11378
|
+
input: 0.16,
|
|
11379
|
+
output: 0.64,
|
|
11380
|
+
cacheRead: 0,
|
|
10901
11381
|
cacheWrite: 0,
|
|
10902
11382
|
},
|
|
10903
|
-
contextWindow:
|
|
10904
|
-
maxTokens:
|
|
11383
|
+
contextWindow: 128000,
|
|
11384
|
+
maxTokens: 8192,
|
|
11385
|
+
},
|
|
11386
|
+
"alibaba/qwen-3.6-max-preview": {
|
|
11387
|
+
id: "alibaba/qwen-3.6-max-preview",
|
|
11388
|
+
name: "Qwen 3.6 Max Preview",
|
|
11389
|
+
api: "anthropic-messages",
|
|
11390
|
+
provider: "vercel-ai-gateway",
|
|
11391
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11392
|
+
reasoning: true,
|
|
11393
|
+
input: ["text", "image"],
|
|
11394
|
+
cost: {
|
|
11395
|
+
input: 1.3,
|
|
11396
|
+
output: 7.8,
|
|
11397
|
+
cacheRead: 0.26,
|
|
11398
|
+
cacheWrite: 1.625,
|
|
11399
|
+
},
|
|
11400
|
+
contextWindow: 240000,
|
|
11401
|
+
maxTokens: 64000,
|
|
10905
11402
|
},
|
|
10906
11403
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10907
11404
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10929,13 +11426,13 @@ export const MODELS = {
|
|
|
10929
11426
|
reasoning: false,
|
|
10930
11427
|
input: ["text"],
|
|
10931
11428
|
cost: {
|
|
10932
|
-
input:
|
|
10933
|
-
output:
|
|
10934
|
-
cacheRead: 0.
|
|
11429
|
+
input: 1.5,
|
|
11430
|
+
output: 7.5,
|
|
11431
|
+
cacheRead: 0.3,
|
|
10935
11432
|
cacheWrite: 0,
|
|
10936
11433
|
},
|
|
10937
11434
|
contextWindow: 262144,
|
|
10938
|
-
maxTokens:
|
|
11435
|
+
maxTokens: 65536,
|
|
10939
11436
|
},
|
|
10940
11437
|
"alibaba/qwen3-coder-30b-a3b": {
|
|
10941
11438
|
id: "alibaba/qwen3-coder-30b-a3b",
|
|
@@ -11048,13 +11545,13 @@ export const MODELS = {
|
|
|
11048
11545
|
reasoning: true,
|
|
11049
11546
|
input: ["text", "image"],
|
|
11050
11547
|
cost: {
|
|
11051
|
-
input: 0.
|
|
11052
|
-
output:
|
|
11548
|
+
input: 0.39999999999999997,
|
|
11549
|
+
output: 4,
|
|
11053
11550
|
cacheRead: 0,
|
|
11054
11551
|
cacheWrite: 0,
|
|
11055
11552
|
},
|
|
11056
|
-
contextWindow:
|
|
11057
|
-
maxTokens:
|
|
11553
|
+
contextWindow: 131072,
|
|
11554
|
+
maxTokens: 32768,
|
|
11058
11555
|
},
|
|
11059
11556
|
"alibaba/qwen3.5-flash": {
|
|
11060
11557
|
id: "alibaba/qwen3.5-flash",
|
|
@@ -11090,6 +11587,23 @@ export const MODELS = {
|
|
|
11090
11587
|
contextWindow: 1000000,
|
|
11091
11588
|
maxTokens: 64000,
|
|
11092
11589
|
},
|
|
11590
|
+
"alibaba/qwen3.6-plus": {
|
|
11591
|
+
id: "alibaba/qwen3.6-plus",
|
|
11592
|
+
name: "Qwen 3.6 Plus",
|
|
11593
|
+
api: "anthropic-messages",
|
|
11594
|
+
provider: "vercel-ai-gateway",
|
|
11595
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11596
|
+
reasoning: true,
|
|
11597
|
+
input: ["text", "image"],
|
|
11598
|
+
cost: {
|
|
11599
|
+
input: 0.5,
|
|
11600
|
+
output: 3,
|
|
11601
|
+
cacheRead: 0.09999999999999999,
|
|
11602
|
+
cacheWrite: 0.625,
|
|
11603
|
+
},
|
|
11604
|
+
contextWindow: 1000000,
|
|
11605
|
+
maxTokens: 64000,
|
|
11606
|
+
},
|
|
11093
11607
|
"anthropic/claude-3-haiku": {
|
|
11094
11608
|
id: "anthropic/claude-3-haiku",
|
|
11095
11609
|
name: "Claude 3 Haiku",
|
|
@@ -11124,40 +11638,6 @@ export const MODELS = {
|
|
|
11124
11638
|
contextWindow: 200000,
|
|
11125
11639
|
maxTokens: 8192,
|
|
11126
11640
|
},
|
|
11127
|
-
"anthropic/claude-3.5-sonnet": {
|
|
11128
|
-
id: "anthropic/claude-3.5-sonnet",
|
|
11129
|
-
name: "Claude 3.5 Sonnet",
|
|
11130
|
-
api: "anthropic-messages",
|
|
11131
|
-
provider: "vercel-ai-gateway",
|
|
11132
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11133
|
-
reasoning: false,
|
|
11134
|
-
input: ["text", "image"],
|
|
11135
|
-
cost: {
|
|
11136
|
-
input: 3,
|
|
11137
|
-
output: 15,
|
|
11138
|
-
cacheRead: 0.3,
|
|
11139
|
-
cacheWrite: 3.75,
|
|
11140
|
-
},
|
|
11141
|
-
contextWindow: 200000,
|
|
11142
|
-
maxTokens: 8192,
|
|
11143
|
-
},
|
|
11144
|
-
"anthropic/claude-3.5-sonnet-20240620": {
|
|
11145
|
-
id: "anthropic/claude-3.5-sonnet-20240620",
|
|
11146
|
-
name: "Claude 3.5 Sonnet (2024-06-20)",
|
|
11147
|
-
api: "anthropic-messages",
|
|
11148
|
-
provider: "vercel-ai-gateway",
|
|
11149
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11150
|
-
reasoning: false,
|
|
11151
|
-
input: ["text", "image"],
|
|
11152
|
-
cost: {
|
|
11153
|
-
input: 3,
|
|
11154
|
-
output: 15,
|
|
11155
|
-
cacheRead: 0.3,
|
|
11156
|
-
cacheWrite: 3.75,
|
|
11157
|
-
},
|
|
11158
|
-
contextWindow: 200000,
|
|
11159
|
-
maxTokens: 8192,
|
|
11160
|
-
},
|
|
11161
11641
|
"anthropic/claude-3.7-sonnet": {
|
|
11162
11642
|
id: "anthropic/claude-3.7-sonnet",
|
|
11163
11643
|
name: "Claude 3.7 Sonnet",
|
|
@@ -11260,6 +11740,23 @@ export const MODELS = {
|
|
|
11260
11740
|
contextWindow: 1000000,
|
|
11261
11741
|
maxTokens: 128000,
|
|
11262
11742
|
},
|
|
11743
|
+
"anthropic/claude-opus-4.7": {
|
|
11744
|
+
id: "anthropic/claude-opus-4.7",
|
|
11745
|
+
name: "Claude Opus 4.7",
|
|
11746
|
+
api: "anthropic-messages",
|
|
11747
|
+
provider: "vercel-ai-gateway",
|
|
11748
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11749
|
+
reasoning: true,
|
|
11750
|
+
input: ["text", "image"],
|
|
11751
|
+
cost: {
|
|
11752
|
+
input: 5,
|
|
11753
|
+
output: 25,
|
|
11754
|
+
cacheRead: 0.5,
|
|
11755
|
+
cacheWrite: 6.25,
|
|
11756
|
+
},
|
|
11757
|
+
contextWindow: 1000000,
|
|
11758
|
+
maxTokens: 128000,
|
|
11759
|
+
},
|
|
11263
11760
|
"anthropic/claude-sonnet-4": {
|
|
11264
11761
|
id: "anthropic/claude-sonnet-4",
|
|
11265
11762
|
name: "Claude Sonnet 4",
|
|
@@ -11328,6 +11825,23 @@ export const MODELS = {
|
|
|
11328
11825
|
contextWindow: 131000,
|
|
11329
11826
|
maxTokens: 131000,
|
|
11330
11827
|
},
|
|
11828
|
+
"arcee-ai/trinity-large-thinking": {
|
|
11829
|
+
id: "arcee-ai/trinity-large-thinking",
|
|
11830
|
+
name: "Trinity Large Thinking",
|
|
11831
|
+
api: "anthropic-messages",
|
|
11832
|
+
provider: "vercel-ai-gateway",
|
|
11833
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11834
|
+
reasoning: true,
|
|
11835
|
+
input: ["text"],
|
|
11836
|
+
cost: {
|
|
11837
|
+
input: 0.25,
|
|
11838
|
+
output: 0.8999999999999999,
|
|
11839
|
+
cacheRead: 0,
|
|
11840
|
+
cacheWrite: 0,
|
|
11841
|
+
},
|
|
11842
|
+
contextWindow: 262100,
|
|
11843
|
+
maxTokens: 80000,
|
|
11844
|
+
},
|
|
11331
11845
|
"bytedance/seed-1.6": {
|
|
11332
11846
|
id: "bytedance/seed-1.6",
|
|
11333
11847
|
name: "Seed 1.6",
|
|
@@ -11405,13 +11919,13 @@ export const MODELS = {
|
|
|
11405
11919
|
reasoning: true,
|
|
11406
11920
|
input: ["text"],
|
|
11407
11921
|
cost: {
|
|
11408
|
-
input: 0.
|
|
11409
|
-
output: 1.
|
|
11410
|
-
cacheRead: 0,
|
|
11922
|
+
input: 0.56,
|
|
11923
|
+
output: 1.68,
|
|
11924
|
+
cacheRead: 0.28,
|
|
11411
11925
|
cacheWrite: 0,
|
|
11412
11926
|
},
|
|
11413
11927
|
contextWindow: 163840,
|
|
11414
|
-
maxTokens:
|
|
11928
|
+
maxTokens: 8192,
|
|
11415
11929
|
},
|
|
11416
11930
|
"deepseek/deepseek-v3.1-terminus": {
|
|
11417
11931
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11594,7 +12108,7 @@ export const MODELS = {
|
|
|
11594
12108
|
cost: {
|
|
11595
12109
|
input: 0.25,
|
|
11596
12110
|
output: 1.5,
|
|
11597
|
-
cacheRead: 0,
|
|
12111
|
+
cacheRead: 0.03,
|
|
11598
12112
|
cacheWrite: 0,
|
|
11599
12113
|
},
|
|
11600
12114
|
contextWindow: 1000000,
|
|
@@ -11617,6 +12131,40 @@ export const MODELS = {
|
|
|
11617
12131
|
contextWindow: 1000000,
|
|
11618
12132
|
maxTokens: 64000,
|
|
11619
12133
|
},
|
|
12134
|
+
"google/gemma-4-26b-a4b-it": {
|
|
12135
|
+
id: "google/gemma-4-26b-a4b-it",
|
|
12136
|
+
name: "Gemma 4 26B A4B IT",
|
|
12137
|
+
api: "anthropic-messages",
|
|
12138
|
+
provider: "vercel-ai-gateway",
|
|
12139
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12140
|
+
reasoning: false,
|
|
12141
|
+
input: ["text", "image"],
|
|
12142
|
+
cost: {
|
|
12143
|
+
input: 0.13,
|
|
12144
|
+
output: 0.39999999999999997,
|
|
12145
|
+
cacheRead: 0,
|
|
12146
|
+
cacheWrite: 0,
|
|
12147
|
+
},
|
|
12148
|
+
contextWindow: 262144,
|
|
12149
|
+
maxTokens: 131072,
|
|
12150
|
+
},
|
|
12151
|
+
"google/gemma-4-31b-it": {
|
|
12152
|
+
id: "google/gemma-4-31b-it",
|
|
12153
|
+
name: "Gemma 4 31B IT",
|
|
12154
|
+
api: "anthropic-messages",
|
|
12155
|
+
provider: "vercel-ai-gateway",
|
|
12156
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12157
|
+
reasoning: false,
|
|
12158
|
+
input: ["text", "image"],
|
|
12159
|
+
cost: {
|
|
12160
|
+
input: 0.14,
|
|
12161
|
+
output: 0.39999999999999997,
|
|
12162
|
+
cacheRead: 0,
|
|
12163
|
+
cacheWrite: 0,
|
|
12164
|
+
},
|
|
12165
|
+
contextWindow: 262144,
|
|
12166
|
+
maxTokens: 131072,
|
|
12167
|
+
},
|
|
11620
12168
|
"inception/mercury-2": {
|
|
11621
12169
|
id: "inception/mercury-2",
|
|
11622
12170
|
name: "Mercury 2",
|
|
@@ -11651,39 +12199,39 @@ export const MODELS = {
|
|
|
11651
12199
|
contextWindow: 32000,
|
|
11652
12200
|
maxTokens: 16384,
|
|
11653
12201
|
},
|
|
11654
|
-
"
|
|
11655
|
-
id: "
|
|
11656
|
-
name: "
|
|
12202
|
+
"kwaipilot/kat-coder-pro-v2": {
|
|
12203
|
+
id: "kwaipilot/kat-coder-pro-v2",
|
|
12204
|
+
name: "Kat Coder Pro V2",
|
|
11657
12205
|
api: "anthropic-messages",
|
|
11658
12206
|
provider: "vercel-ai-gateway",
|
|
11659
12207
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11660
|
-
reasoning:
|
|
12208
|
+
reasoning: true,
|
|
11661
12209
|
input: ["text"],
|
|
11662
12210
|
cost: {
|
|
11663
|
-
input: 0,
|
|
11664
|
-
output:
|
|
11665
|
-
cacheRead: 0,
|
|
12211
|
+
input: 0.3,
|
|
12212
|
+
output: 1.2,
|
|
12213
|
+
cacheRead: 0.06,
|
|
11666
12214
|
cacheWrite: 0,
|
|
11667
12215
|
},
|
|
11668
|
-
contextWindow:
|
|
11669
|
-
maxTokens:
|
|
12216
|
+
contextWindow: 256000,
|
|
12217
|
+
maxTokens: 256000,
|
|
11670
12218
|
},
|
|
11671
|
-
"meituan/longcat-flash-
|
|
11672
|
-
id: "meituan/longcat-flash-
|
|
11673
|
-
name: "LongCat Flash
|
|
12219
|
+
"meituan/longcat-flash-chat": {
|
|
12220
|
+
id: "meituan/longcat-flash-chat",
|
|
12221
|
+
name: "LongCat Flash Chat",
|
|
11674
12222
|
api: "anthropic-messages",
|
|
11675
12223
|
provider: "vercel-ai-gateway",
|
|
11676
12224
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11677
|
-
reasoning:
|
|
12225
|
+
reasoning: false,
|
|
11678
12226
|
input: ["text"],
|
|
11679
12227
|
cost: {
|
|
11680
|
-
input: 0
|
|
11681
|
-
output:
|
|
12228
|
+
input: 0,
|
|
12229
|
+
output: 0,
|
|
11682
12230
|
cacheRead: 0,
|
|
11683
12231
|
cacheWrite: 0,
|
|
11684
12232
|
},
|
|
11685
12233
|
contextWindow: 128000,
|
|
11686
|
-
maxTokens:
|
|
12234
|
+
maxTokens: 100000,
|
|
11687
12235
|
},
|
|
11688
12236
|
"meta/llama-3.1-70b": {
|
|
11689
12237
|
id: "meta/llama-3.1-70b",
|
|
@@ -11711,13 +12259,13 @@ export const MODELS = {
|
|
|
11711
12259
|
reasoning: false,
|
|
11712
12260
|
input: ["text"],
|
|
11713
12261
|
cost: {
|
|
11714
|
-
input: 0.
|
|
11715
|
-
output: 0.
|
|
11716
|
-
cacheRead: 0
|
|
12262
|
+
input: 0.22,
|
|
12263
|
+
output: 0.22,
|
|
12264
|
+
cacheRead: 0,
|
|
11717
12265
|
cacheWrite: 0,
|
|
11718
12266
|
},
|
|
11719
12267
|
contextWindow: 128000,
|
|
11720
|
-
maxTokens:
|
|
12268
|
+
maxTokens: 8192,
|
|
11721
12269
|
},
|
|
11722
12270
|
"meta/llama-3.2-11b": {
|
|
11723
12271
|
id: "meta/llama-3.2-11b",
|
|
@@ -12095,20 +12643,20 @@ export const MODELS = {
|
|
|
12095
12643
|
},
|
|
12096
12644
|
"moonshotai/kimi-k2": {
|
|
12097
12645
|
id: "moonshotai/kimi-k2",
|
|
12098
|
-
name: "Kimi K2",
|
|
12646
|
+
name: "Kimi K2 Instruct",
|
|
12099
12647
|
api: "anthropic-messages",
|
|
12100
12648
|
provider: "vercel-ai-gateway",
|
|
12101
12649
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12102
12650
|
reasoning: false,
|
|
12103
12651
|
input: ["text"],
|
|
12104
12652
|
cost: {
|
|
12105
|
-
input: 0.
|
|
12106
|
-
output: 2.
|
|
12107
|
-
cacheRead: 0
|
|
12653
|
+
input: 0.5700000000000001,
|
|
12654
|
+
output: 2.3,
|
|
12655
|
+
cacheRead: 0,
|
|
12108
12656
|
cacheWrite: 0,
|
|
12109
12657
|
},
|
|
12110
12658
|
contextWindow: 131072,
|
|
12111
|
-
maxTokens:
|
|
12659
|
+
maxTokens: 131072,
|
|
12112
12660
|
},
|
|
12113
12661
|
"moonshotai/kimi-k2-0905": {
|
|
12114
12662
|
id: "moonshotai/kimi-k2-0905",
|
|
@@ -12121,11 +12669,11 @@ export const MODELS = {
|
|
|
12121
12669
|
cost: {
|
|
12122
12670
|
input: 0.6,
|
|
12123
12671
|
output: 2.5,
|
|
12124
|
-
cacheRead: 0.
|
|
12672
|
+
cacheRead: 0.3,
|
|
12125
12673
|
cacheWrite: 0,
|
|
12126
12674
|
},
|
|
12127
12675
|
contextWindow: 256000,
|
|
12128
|
-
maxTokens:
|
|
12676
|
+
maxTokens: 128000,
|
|
12129
12677
|
},
|
|
12130
12678
|
"moonshotai/kimi-k2-thinking": {
|
|
12131
12679
|
id: "moonshotai/kimi-k2-thinking",
|
|
@@ -12195,6 +12743,23 @@ export const MODELS = {
|
|
|
12195
12743
|
contextWindow: 262114,
|
|
12196
12744
|
maxTokens: 262114,
|
|
12197
12745
|
},
|
|
12746
|
+
"moonshotai/kimi-k2.6": {
|
|
12747
|
+
id: "moonshotai/kimi-k2.6",
|
|
12748
|
+
name: "Kimi K2.6",
|
|
12749
|
+
api: "anthropic-messages",
|
|
12750
|
+
provider: "vercel-ai-gateway",
|
|
12751
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12752
|
+
reasoning: true,
|
|
12753
|
+
input: ["text", "image"],
|
|
12754
|
+
cost: {
|
|
12755
|
+
input: 0.95,
|
|
12756
|
+
output: 4,
|
|
12757
|
+
cacheRead: 0.16,
|
|
12758
|
+
cacheWrite: 0,
|
|
12759
|
+
},
|
|
12760
|
+
contextWindow: 262000,
|
|
12761
|
+
maxTokens: 262000,
|
|
12762
|
+
},
|
|
12198
12763
|
"nvidia/nemotron-nano-12b-v2-vl": {
|
|
12199
12764
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
12200
12765
|
name: "Nvidia Nemotron Nano 12B V2 VL",
|
|
@@ -12690,24 +13255,24 @@ export const MODELS = {
|
|
|
12690
13255
|
},
|
|
12691
13256
|
"openai/gpt-oss-20b": {
|
|
12692
13257
|
id: "openai/gpt-oss-20b",
|
|
12693
|
-
name: "
|
|
13258
|
+
name: "GPT OSS 120B",
|
|
12694
13259
|
api: "anthropic-messages",
|
|
12695
13260
|
provider: "vercel-ai-gateway",
|
|
12696
13261
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12697
13262
|
reasoning: true,
|
|
12698
13263
|
input: ["text"],
|
|
12699
13264
|
cost: {
|
|
12700
|
-
input: 0.
|
|
12701
|
-
output: 0.
|
|
13265
|
+
input: 0.049999999999999996,
|
|
13266
|
+
output: 0.19999999999999998,
|
|
12702
13267
|
cacheRead: 0,
|
|
12703
13268
|
cacheWrite: 0,
|
|
12704
13269
|
},
|
|
12705
|
-
contextWindow:
|
|
13270
|
+
contextWindow: 131072,
|
|
12706
13271
|
maxTokens: 8192,
|
|
12707
13272
|
},
|
|
12708
13273
|
"openai/gpt-oss-safeguard-20b": {
|
|
12709
13274
|
id: "openai/gpt-oss-safeguard-20b",
|
|
12710
|
-
name: "
|
|
13275
|
+
name: "GPT OSS Safeguard 20B",
|
|
12711
13276
|
api: "anthropic-messages",
|
|
12712
13277
|
provider: "vercel-ai-gateway",
|
|
12713
13278
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -12875,23 +13440,6 @@ export const MODELS = {
|
|
|
12875
13440
|
contextWindow: 131072,
|
|
12876
13441
|
maxTokens: 131072,
|
|
12877
13442
|
},
|
|
12878
|
-
"xai/grok-2-vision": {
|
|
12879
|
-
id: "xai/grok-2-vision",
|
|
12880
|
-
name: "Grok 2 Vision",
|
|
12881
|
-
api: "anthropic-messages",
|
|
12882
|
-
provider: "vercel-ai-gateway",
|
|
12883
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12884
|
-
reasoning: false,
|
|
12885
|
-
input: ["text", "image"],
|
|
12886
|
-
cost: {
|
|
12887
|
-
input: 2,
|
|
12888
|
-
output: 10,
|
|
12889
|
-
cacheRead: 0,
|
|
12890
|
-
cacheWrite: 0,
|
|
12891
|
-
},
|
|
12892
|
-
contextWindow: 32768,
|
|
12893
|
-
maxTokens: 32768,
|
|
12894
|
-
},
|
|
12895
13443
|
"xai/grok-3": {
|
|
12896
13444
|
id: "xai/grok-3",
|
|
12897
13445
|
name: "Grok 3 Beta",
|
|
@@ -12984,7 +13532,7 @@ export const MODELS = {
|
|
|
12984
13532
|
provider: "vercel-ai-gateway",
|
|
12985
13533
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12986
13534
|
reasoning: false,
|
|
12987
|
-
input: ["text"],
|
|
13535
|
+
input: ["text", "image"],
|
|
12988
13536
|
cost: {
|
|
12989
13537
|
input: 0.19999999999999998,
|
|
12990
13538
|
output: 0.5,
|
|
@@ -13001,7 +13549,7 @@ export const MODELS = {
|
|
|
13001
13549
|
provider: "vercel-ai-gateway",
|
|
13002
13550
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13003
13551
|
reasoning: true,
|
|
13004
|
-
input: ["text"],
|
|
13552
|
+
input: ["text", "image"],
|
|
13005
13553
|
cost: {
|
|
13006
13554
|
input: 0.19999999999999998,
|
|
13007
13555
|
output: 0.5,
|
|
@@ -13018,7 +13566,7 @@ export const MODELS = {
|
|
|
13018
13566
|
provider: "vercel-ai-gateway",
|
|
13019
13567
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13020
13568
|
reasoning: false,
|
|
13021
|
-
input: ["text"],
|
|
13569
|
+
input: ["text", "image"],
|
|
13022
13570
|
cost: {
|
|
13023
13571
|
input: 0.19999999999999998,
|
|
13024
13572
|
output: 0.5,
|
|
@@ -13035,7 +13583,7 @@ export const MODELS = {
|
|
|
13035
13583
|
provider: "vercel-ai-gateway",
|
|
13036
13584
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13037
13585
|
reasoning: true,
|
|
13038
|
-
input: ["text"],
|
|
13586
|
+
input: ["text", "image"],
|
|
13039
13587
|
cost: {
|
|
13040
13588
|
input: 0.19999999999999998,
|
|
13041
13589
|
output: 0.5,
|
|
@@ -13052,7 +13600,7 @@ export const MODELS = {
|
|
|
13052
13600
|
provider: "vercel-ai-gateway",
|
|
13053
13601
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13054
13602
|
reasoning: true,
|
|
13055
|
-
input: ["text"],
|
|
13603
|
+
input: ["text", "image"],
|
|
13056
13604
|
cost: {
|
|
13057
13605
|
input: 2,
|
|
13058
13606
|
output: 6,
|
|
@@ -13069,7 +13617,7 @@ export const MODELS = {
|
|
|
13069
13617
|
provider: "vercel-ai-gateway",
|
|
13070
13618
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13071
13619
|
reasoning: true,
|
|
13072
|
-
input: ["text"],
|
|
13620
|
+
input: ["text", "image"],
|
|
13073
13621
|
cost: {
|
|
13074
13622
|
input: 2,
|
|
13075
13623
|
output: 6,
|
|
@@ -13175,7 +13723,7 @@ export const MODELS = {
|
|
|
13175
13723
|
cost: {
|
|
13176
13724
|
input: 0.09999999999999999,
|
|
13177
13725
|
output: 0.3,
|
|
13178
|
-
cacheRead: 0.
|
|
13726
|
+
cacheRead: 0.01,
|
|
13179
13727
|
cacheWrite: 0,
|
|
13180
13728
|
},
|
|
13181
13729
|
contextWindow: 262144,
|
|
@@ -13309,13 +13857,13 @@ export const MODELS = {
|
|
|
13309
13857
|
reasoning: true,
|
|
13310
13858
|
input: ["text"],
|
|
13311
13859
|
cost: {
|
|
13312
|
-
input:
|
|
13313
|
-
output: 2.
|
|
13314
|
-
cacheRead:
|
|
13860
|
+
input: 2.25,
|
|
13861
|
+
output: 2.75,
|
|
13862
|
+
cacheRead: 2.25,
|
|
13315
13863
|
cacheWrite: 0,
|
|
13316
13864
|
},
|
|
13317
|
-
contextWindow:
|
|
13318
|
-
maxTokens:
|
|
13865
|
+
contextWindow: 131000,
|
|
13866
|
+
maxTokens: 40000,
|
|
13319
13867
|
},
|
|
13320
13868
|
"zai/glm-4.7-flash": {
|
|
13321
13869
|
id: "zai/glm-4.7-flash",
|
|
@@ -13385,6 +13933,40 @@ export const MODELS = {
|
|
|
13385
13933
|
contextWindow: 202800,
|
|
13386
13934
|
maxTokens: 131100,
|
|
13387
13935
|
},
|
|
13936
|
+
"zai/glm-5.1": {
|
|
13937
|
+
id: "zai/glm-5.1",
|
|
13938
|
+
name: "GLM 5.1",
|
|
13939
|
+
api: "anthropic-messages",
|
|
13940
|
+
provider: "vercel-ai-gateway",
|
|
13941
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13942
|
+
reasoning: true,
|
|
13943
|
+
input: ["text"],
|
|
13944
|
+
cost: {
|
|
13945
|
+
input: 1.4,
|
|
13946
|
+
output: 4.4,
|
|
13947
|
+
cacheRead: 0.26,
|
|
13948
|
+
cacheWrite: 0,
|
|
13949
|
+
},
|
|
13950
|
+
contextWindow: 202800,
|
|
13951
|
+
maxTokens: 64000,
|
|
13952
|
+
},
|
|
13953
|
+
"zai/glm-5v-turbo": {
|
|
13954
|
+
id: "zai/glm-5v-turbo",
|
|
13955
|
+
name: "GLM 5V Turbo",
|
|
13956
|
+
api: "anthropic-messages",
|
|
13957
|
+
provider: "vercel-ai-gateway",
|
|
13958
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13959
|
+
reasoning: true,
|
|
13960
|
+
input: ["text", "image"],
|
|
13961
|
+
cost: {
|
|
13962
|
+
input: 1.2,
|
|
13963
|
+
output: 4,
|
|
13964
|
+
cacheRead: 0.24,
|
|
13965
|
+
cacheWrite: 0,
|
|
13966
|
+
},
|
|
13967
|
+
contextWindow: 200000,
|
|
13968
|
+
maxTokens: 128000,
|
|
13969
|
+
},
|
|
13388
13970
|
},
|
|
13389
13971
|
"xai": {
|
|
13390
13972
|
"grok-2": {
|
|
@@ -13807,9 +14389,9 @@ export const MODELS = {
|
|
|
13807
14389
|
reasoning: true,
|
|
13808
14390
|
input: ["text"],
|
|
13809
14391
|
cost: {
|
|
13810
|
-
input: 0
|
|
13811
|
-
output:
|
|
13812
|
-
cacheRead: 0
|
|
14392
|
+
input: 0,
|
|
14393
|
+
output: 0,
|
|
14394
|
+
cacheRead: 0,
|
|
13813
14395
|
cacheWrite: 0,
|
|
13814
14396
|
},
|
|
13815
14397
|
contextWindow: 131072,
|
|
@@ -13825,9 +14407,9 @@ export const MODELS = {
|
|
|
13825
14407
|
reasoning: true,
|
|
13826
14408
|
input: ["text"],
|
|
13827
14409
|
cost: {
|
|
13828
|
-
input: 0
|
|
13829
|
-
output:
|
|
13830
|
-
cacheRead: 0
|
|
14410
|
+
input: 0,
|
|
14411
|
+
output: 0,
|
|
14412
|
+
cacheRead: 0,
|
|
13831
14413
|
cacheWrite: 0,
|
|
13832
14414
|
},
|
|
13833
14415
|
contextWindow: 131072,
|
|
@@ -13861,8 +14443,8 @@ export const MODELS = {
|
|
|
13861
14443
|
reasoning: true,
|
|
13862
14444
|
input: ["text", "image"],
|
|
13863
14445
|
cost: {
|
|
13864
|
-
input: 0
|
|
13865
|
-
output:
|
|
14446
|
+
input: 0,
|
|
14447
|
+
output: 0,
|
|
13866
14448
|
cacheRead: 0,
|
|
13867
14449
|
cacheWrite: 0,
|
|
13868
14450
|
},
|
|
@@ -13875,13 +14457,13 @@ export const MODELS = {
|
|
|
13875
14457
|
api: "openai-completions",
|
|
13876
14458
|
provider: "zai",
|
|
13877
14459
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13878
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14460
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13879
14461
|
reasoning: true,
|
|
13880
14462
|
input: ["text"],
|
|
13881
14463
|
cost: {
|
|
13882
|
-
input: 0
|
|
13883
|
-
output:
|
|
13884
|
-
cacheRead: 0
|
|
14464
|
+
input: 0,
|
|
14465
|
+
output: 0,
|
|
14466
|
+
cacheRead: 0,
|
|
13885
14467
|
cacheWrite: 0,
|
|
13886
14468
|
},
|
|
13887
14469
|
contextWindow: 204800,
|
|
@@ -13893,12 +14475,12 @@ export const MODELS = {
|
|
|
13893
14475
|
api: "openai-completions",
|
|
13894
14476
|
provider: "zai",
|
|
13895
14477
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13896
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14478
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13897
14479
|
reasoning: true,
|
|
13898
14480
|
input: ["text", "image"],
|
|
13899
14481
|
cost: {
|
|
13900
|
-
input: 0
|
|
13901
|
-
output: 0
|
|
14482
|
+
input: 0,
|
|
14483
|
+
output: 0,
|
|
13902
14484
|
cacheRead: 0,
|
|
13903
14485
|
cacheWrite: 0,
|
|
13904
14486
|
},
|
|
@@ -13911,13 +14493,13 @@ export const MODELS = {
|
|
|
13911
14493
|
api: "openai-completions",
|
|
13912
14494
|
provider: "zai",
|
|
13913
14495
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13914
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14496
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13915
14497
|
reasoning: true,
|
|
13916
14498
|
input: ["text"],
|
|
13917
14499
|
cost: {
|
|
13918
|
-
input: 0
|
|
13919
|
-
output:
|
|
13920
|
-
cacheRead: 0
|
|
14500
|
+
input: 0,
|
|
14501
|
+
output: 0,
|
|
14502
|
+
cacheRead: 0,
|
|
13921
14503
|
cacheWrite: 0,
|
|
13922
14504
|
},
|
|
13923
14505
|
contextWindow: 204800,
|
|
@@ -13929,7 +14511,7 @@ export const MODELS = {
|
|
|
13929
14511
|
api: "openai-completions",
|
|
13930
14512
|
provider: "zai",
|
|
13931
14513
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13932
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14514
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13933
14515
|
reasoning: true,
|
|
13934
14516
|
input: ["text"],
|
|
13935
14517
|
cost: {
|
|
@@ -13941,19 +14523,37 @@ export const MODELS = {
|
|
|
13941
14523
|
contextWindow: 200000,
|
|
13942
14524
|
maxTokens: 131072,
|
|
13943
14525
|
},
|
|
14526
|
+
"glm-4.7-flashx": {
|
|
14527
|
+
id: "glm-4.7-flashx",
|
|
14528
|
+
name: "GLM-4.7-FlashX",
|
|
14529
|
+
api: "openai-completions",
|
|
14530
|
+
provider: "zai",
|
|
14531
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14532
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14533
|
+
reasoning: true,
|
|
14534
|
+
input: ["text"],
|
|
14535
|
+
cost: {
|
|
14536
|
+
input: 0.07,
|
|
14537
|
+
output: 0.4,
|
|
14538
|
+
cacheRead: 0.01,
|
|
14539
|
+
cacheWrite: 0,
|
|
14540
|
+
},
|
|
14541
|
+
contextWindow: 200000,
|
|
14542
|
+
maxTokens: 131072,
|
|
14543
|
+
},
|
|
13944
14544
|
"glm-5": {
|
|
13945
14545
|
id: "glm-5",
|
|
13946
14546
|
name: "GLM-5",
|
|
13947
14547
|
api: "openai-completions",
|
|
13948
14548
|
provider: "zai",
|
|
13949
14549
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13950
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14550
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13951
14551
|
reasoning: true,
|
|
13952
14552
|
input: ["text"],
|
|
13953
14553
|
cost: {
|
|
13954
|
-
input:
|
|
13955
|
-
output:
|
|
13956
|
-
cacheRead: 0
|
|
14554
|
+
input: 0,
|
|
14555
|
+
output: 0,
|
|
14556
|
+
cacheRead: 0,
|
|
13957
14557
|
cacheWrite: 0,
|
|
13958
14558
|
},
|
|
13959
14559
|
contextWindow: 204800,
|
|
@@ -13965,13 +14565,49 @@ export const MODELS = {
|
|
|
13965
14565
|
api: "openai-completions",
|
|
13966
14566
|
provider: "zai",
|
|
13967
14567
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13968
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14568
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13969
14569
|
reasoning: true,
|
|
13970
14570
|
input: ["text"],
|
|
13971
14571
|
cost: {
|
|
13972
|
-
input:
|
|
13973
|
-
output:
|
|
13974
|
-
cacheRead: 0
|
|
14572
|
+
input: 0,
|
|
14573
|
+
output: 0,
|
|
14574
|
+
cacheRead: 0,
|
|
14575
|
+
cacheWrite: 0,
|
|
14576
|
+
},
|
|
14577
|
+
contextWindow: 200000,
|
|
14578
|
+
maxTokens: 131072,
|
|
14579
|
+
},
|
|
14580
|
+
"glm-5.1": {
|
|
14581
|
+
id: "glm-5.1",
|
|
14582
|
+
name: "GLM-5.1",
|
|
14583
|
+
api: "openai-completions",
|
|
14584
|
+
provider: "zai",
|
|
14585
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14586
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14587
|
+
reasoning: true,
|
|
14588
|
+
input: ["text"],
|
|
14589
|
+
cost: {
|
|
14590
|
+
input: 0,
|
|
14591
|
+
output: 0,
|
|
14592
|
+
cacheRead: 0,
|
|
14593
|
+
cacheWrite: 0,
|
|
14594
|
+
},
|
|
14595
|
+
contextWindow: 200000,
|
|
14596
|
+
maxTokens: 131072,
|
|
14597
|
+
},
|
|
14598
|
+
"glm-5v-turbo": {
|
|
14599
|
+
id: "glm-5v-turbo",
|
|
14600
|
+
name: "glm-5v-turbo",
|
|
14601
|
+
api: "openai-completions",
|
|
14602
|
+
provider: "zai",
|
|
14603
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14604
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14605
|
+
reasoning: true,
|
|
14606
|
+
input: ["text", "image"],
|
|
14607
|
+
cost: {
|
|
14608
|
+
input: 0,
|
|
14609
|
+
output: 0,
|
|
14610
|
+
cacheRead: 0,
|
|
13975
14611
|
cacheWrite: 0,
|
|
13976
14612
|
},
|
|
13977
14613
|
contextWindow: 200000,
|