@draht/ai 2026.3.25 → 2026.4.23

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (51) hide show
  1. package/README.md +107 -12
  2. package/dist/cli.d.ts.map +1 -1
  3. package/dist/cli.js +1 -1
  4. package/dist/cli.js.map +1 -1
  5. package/dist/index.d.ts +1 -0
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -0
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.generated.d.ts +1163 -498
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +1392 -756
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +37 -9
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +38 -16
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/faux.d.ts +56 -0
  20. package/dist/providers/faux.d.ts.map +1 -0
  21. package/dist/providers/faux.js +367 -0
  22. package/dist/providers/faux.js.map +1 -0
  23. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  24. package/dist/providers/google-gemini-cli.js +2 -2
  25. package/dist/providers/google-gemini-cli.js.map +1 -1
  26. package/dist/providers/google-vertex.d.ts.map +1 -1
  27. package/dist/providers/google-vertex.js +1 -1
  28. package/dist/providers/google-vertex.js.map +1 -1
  29. package/dist/providers/google.d.ts.map +1 -1
  30. package/dist/providers/google.js +29 -4
  31. package/dist/providers/google.js.map +1 -1
  32. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  33. package/dist/providers/openai-codex-responses.js +1 -1
  34. package/dist/providers/openai-codex-responses.js.map +1 -1
  35. package/dist/providers/openai-completions.d.ts.map +1 -1
  36. package/dist/providers/openai-completions.js +18 -6
  37. package/dist/providers/openai-completions.js.map +1 -1
  38. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  39. package/dist/providers/openai-responses-shared.js +20 -1
  40. package/dist/providers/openai-responses-shared.js.map +1 -1
  41. package/dist/providers/openai-responses.d.ts.map +1 -1
  42. package/dist/providers/openai-responses.js +2 -2
  43. package/dist/providers/openai-responses.js.map +1 -1
  44. package/dist/types.d.ts +64 -5
  45. package/dist/types.d.ts.map +1 -1
  46. package/dist/types.js.map +1 -1
  47. package/dist/utils/overflow.d.ts +4 -3
  48. package/dist/utils/overflow.d.ts.map +1 -1
  49. package/dist/utils/overflow.js +28 -13
  50. package/dist/utils/overflow.js.map +1 -1
  51. package/package.json +5 -8
@@ -257,6 +257,23 @@ export const MODELS = {
257
257
  contextWindow: 1000000,
258
258
  maxTokens: 128000,
259
259
  },
260
+ "anthropic.claude-opus-4-7": {
261
+ id: "anthropic.claude-opus-4-7",
262
+ name: "Claude Opus 4.7",
263
+ api: "bedrock-converse-stream",
264
+ provider: "amazon-bedrock",
265
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
266
+ reasoning: true,
267
+ input: ["text", "image"],
268
+ cost: {
269
+ input: 5,
270
+ output: 25,
271
+ cacheRead: 0.5,
272
+ cacheWrite: 6.25,
273
+ },
274
+ contextWindow: 1000000,
275
+ maxTokens: 128000,
276
+ },
260
277
  "anthropic.claude-sonnet-4-20250514-v1:0": {
261
278
  id: "anthropic.claude-sonnet-4-20250514-v1:0",
262
279
  name: "Claude Sonnet 4",
@@ -410,6 +427,23 @@ export const MODELS = {
410
427
  contextWindow: 1000000,
411
428
  maxTokens: 128000,
412
429
  },
430
+ "eu.anthropic.claude-opus-4-7": {
431
+ id: "eu.anthropic.claude-opus-4-7",
432
+ name: "Claude Opus 4.7 (EU)",
433
+ api: "bedrock-converse-stream",
434
+ provider: "amazon-bedrock",
435
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
436
+ reasoning: true,
437
+ input: ["text", "image"],
438
+ cost: {
439
+ input: 5,
440
+ output: 25,
441
+ cacheRead: 0.5,
442
+ cacheWrite: 6.25,
443
+ },
444
+ contextWindow: 1000000,
445
+ maxTokens: 128000,
446
+ },
413
447
  "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
414
448
  id: "eu.anthropic.claude-sonnet-4-20250514-v1:0",
415
449
  name: "Claude Sonnet 4 (EU)",
@@ -512,6 +546,23 @@ export const MODELS = {
512
546
  contextWindow: 1000000,
513
547
  maxTokens: 128000,
514
548
  },
549
+ "global.anthropic.claude-opus-4-7": {
550
+ id: "global.anthropic.claude-opus-4-7",
551
+ name: "Claude Opus 4.7 (Global)",
552
+ api: "bedrock-converse-stream",
553
+ provider: "amazon-bedrock",
554
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
555
+ reasoning: true,
556
+ input: ["text", "image"],
557
+ cost: {
558
+ input: 5,
559
+ output: 25,
560
+ cacheRead: 0.5,
561
+ cacheWrite: 6.25,
562
+ },
563
+ contextWindow: 1000000,
564
+ maxTokens: 128000,
565
+ },
515
566
  "global.anthropic.claude-sonnet-4-20250514-v1:0": {
516
567
  id: "global.anthropic.claude-sonnet-4-20250514-v1:0",
517
568
  name: "Claude Sonnet 4 (Global)",
@@ -815,8 +866,8 @@ export const MODELS = {
815
866
  cacheRead: 0,
816
867
  cacheWrite: 0,
817
868
  },
818
- contextWindow: 1000000,
819
- maxTokens: 131072,
869
+ contextWindow: 196608,
870
+ maxTokens: 98304,
820
871
  },
821
872
  "mistral.devstral-2-123b": {
822
873
  id: "mistral.devstral-2-123b",
@@ -1056,6 +1107,23 @@ export const MODELS = {
1056
1107
  contextWindow: 128000,
1057
1108
  maxTokens: 4096,
1058
1109
  },
1110
+ "nvidia.nemotron-super-3-120b": {
1111
+ id: "nvidia.nemotron-super-3-120b",
1112
+ name: "NVIDIA Nemotron 3 Super 120B A12B",
1113
+ api: "bedrock-converse-stream",
1114
+ provider: "amazon-bedrock",
1115
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1116
+ reasoning: true,
1117
+ input: ["text"],
1118
+ cost: {
1119
+ input: 0.15,
1120
+ output: 0.65,
1121
+ cacheRead: 0,
1122
+ cacheWrite: 0,
1123
+ },
1124
+ contextWindow: 262144,
1125
+ maxTokens: 131072,
1126
+ },
1059
1127
  "openai.gpt-oss-120b-1:0": {
1060
1128
  id: "openai.gpt-oss-120b-1:0",
1061
1129
  name: "gpt-oss-120b",
@@ -1192,6 +1260,23 @@ export const MODELS = {
1192
1260
  contextWindow: 131072,
1193
1261
  maxTokens: 65536,
1194
1262
  },
1263
+ "qwen.qwen3-coder-next": {
1264
+ id: "qwen.qwen3-coder-next",
1265
+ name: "Qwen3 Coder Next",
1266
+ api: "bedrock-converse-stream",
1267
+ provider: "amazon-bedrock",
1268
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1269
+ reasoning: true,
1270
+ input: ["text"],
1271
+ cost: {
1272
+ input: 0.22,
1273
+ output: 1.8,
1274
+ cacheRead: 0,
1275
+ cacheWrite: 0,
1276
+ },
1277
+ contextWindow: 131072,
1278
+ maxTokens: 65536,
1279
+ },
1195
1280
  "qwen.qwen3-next-80b-a3b": {
1196
1281
  id: "qwen.qwen3-next-80b-a3b",
1197
1282
  name: "Qwen/Qwen3-Next-80B-A3B-Instruct",
@@ -1311,6 +1396,23 @@ export const MODELS = {
1311
1396
  contextWindow: 1000000,
1312
1397
  maxTokens: 128000,
1313
1398
  },
1399
+ "us.anthropic.claude-opus-4-7": {
1400
+ id: "us.anthropic.claude-opus-4-7",
1401
+ name: "Claude Opus 4.7 (US)",
1402
+ api: "bedrock-converse-stream",
1403
+ provider: "amazon-bedrock",
1404
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1405
+ reasoning: true,
1406
+ input: ["text", "image"],
1407
+ cost: {
1408
+ input: 5,
1409
+ output: 25,
1410
+ cacheRead: 0.5,
1411
+ cacheWrite: 6.25,
1412
+ },
1413
+ contextWindow: 1000000,
1414
+ maxTokens: 128000,
1415
+ },
1314
1416
  "us.anthropic.claude-sonnet-4-20250514-v1:0": {
1315
1417
  id: "us.anthropic.claude-sonnet-4-20250514-v1:0",
1316
1418
  name: "Claude Sonnet 4 (US)",
@@ -1444,8 +1546,8 @@ export const MODELS = {
1444
1546
  cacheRead: 0,
1445
1547
  cacheWrite: 0,
1446
1548
  },
1447
- contextWindow: 200000,
1448
- maxTokens: 131072,
1549
+ contextWindow: 202752,
1550
+ maxTokens: 101376,
1449
1551
  },
1450
1552
  },
1451
1553
  "anthropic": {
@@ -1534,23 +1636,6 @@ export const MODELS = {
1534
1636
  contextWindow: 200000,
1535
1637
  maxTokens: 64000,
1536
1638
  },
1537
- "claude-3-7-sonnet-latest": {
1538
- id: "claude-3-7-sonnet-latest",
1539
- name: "Claude Sonnet 3.7 (latest)",
1540
- api: "anthropic-messages",
1541
- provider: "anthropic",
1542
- baseUrl: "https://api.anthropic.com",
1543
- reasoning: true,
1544
- input: ["text", "image"],
1545
- cost: {
1546
- input: 3,
1547
- output: 15,
1548
- cacheRead: 0.3,
1549
- cacheWrite: 3.75,
1550
- },
1551
- contextWindow: 200000,
1552
- maxTokens: 64000,
1553
- },
1554
1639
  "claude-3-haiku-20240307": {
1555
1640
  id: "claude-3-haiku-20240307",
1556
1641
  name: "Claude Haiku 3",
@@ -1755,6 +1840,23 @@ export const MODELS = {
1755
1840
  contextWindow: 1000000,
1756
1841
  maxTokens: 128000,
1757
1842
  },
1843
+ "claude-opus-4-7": {
1844
+ id: "claude-opus-4-7",
1845
+ name: "Claude Opus 4.7",
1846
+ api: "anthropic-messages",
1847
+ provider: "anthropic",
1848
+ baseUrl: "https://api.anthropic.com",
1849
+ reasoning: true,
1850
+ input: ["text", "image"],
1851
+ cost: {
1852
+ input: 5,
1853
+ output: 25,
1854
+ cacheRead: 0.5,
1855
+ cacheWrite: 6.25,
1856
+ },
1857
+ contextWindow: 1000000,
1858
+ maxTokens: 128000,
1859
+ },
1758
1860
  "claude-sonnet-4-0": {
1759
1861
  id: "claude-sonnet-4-0",
1760
1862
  name: "Claude Sonnet 4 (latest)",
@@ -1842,23 +1944,6 @@ export const MODELS = {
1842
1944
  },
1843
1945
  },
1844
1946
  "azure-openai-responses": {
1845
- "codex-mini-latest": {
1846
- id: "codex-mini-latest",
1847
- name: "Codex Mini",
1848
- api: "azure-openai-responses",
1849
- provider: "azure-openai-responses",
1850
- baseUrl: "",
1851
- reasoning: true,
1852
- input: ["text"],
1853
- cost: {
1854
- input: 1.5,
1855
- output: 6,
1856
- cacheRead: 0.375,
1857
- cacheWrite: 0,
1858
- },
1859
- contextWindow: 200000,
1860
- maxTokens: 100000,
1861
- },
1862
1947
  "gpt-4": {
1863
1948
  id: "gpt-4",
1864
1949
  name: "GPT-4",
@@ -2284,6 +2369,23 @@ export const MODELS = {
2284
2369
  contextWindow: 400000,
2285
2370
  maxTokens: 128000,
2286
2371
  },
2372
+ "gpt-5.3-chat-latest": {
2373
+ id: "gpt-5.3-chat-latest",
2374
+ name: "GPT-5.3 Chat (latest)",
2375
+ api: "azure-openai-responses",
2376
+ provider: "azure-openai-responses",
2377
+ baseUrl: "",
2378
+ reasoning: false,
2379
+ input: ["text", "image"],
2380
+ cost: {
2381
+ input: 1.75,
2382
+ output: 14,
2383
+ cacheRead: 0.175,
2384
+ cacheWrite: 0,
2385
+ },
2386
+ contextWindow: 128000,
2387
+ maxTokens: 16384,
2388
+ },
2287
2389
  "gpt-5.3-codex": {
2288
2390
  id: "gpt-5.3-codex",
2289
2391
  name: "GPT-5.3 Codex",
@@ -2648,6 +2750,24 @@ export const MODELS = {
2648
2750
  contextWindow: 1000000,
2649
2751
  maxTokens: 64000,
2650
2752
  },
2753
+ "claude-opus-4.7": {
2754
+ id: "claude-opus-4.7",
2755
+ name: "Claude Opus 4.7",
2756
+ api: "anthropic-messages",
2757
+ provider: "github-copilot",
2758
+ baseUrl: "https://api.individual.githubcopilot.com",
2759
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
2760
+ reasoning: true,
2761
+ input: ["text", "image"],
2762
+ cost: {
2763
+ input: 0,
2764
+ output: 0,
2765
+ cacheRead: 0,
2766
+ cacheWrite: 0,
2767
+ },
2768
+ contextWindow: 144000,
2769
+ maxTokens: 64000,
2770
+ },
2651
2771
  "claude-sonnet-4": {
2652
2772
  id: "claude-sonnet-4",
2653
2773
  name: "Claude Sonnet 4",
@@ -2998,7 +3118,7 @@ export const MODELS = {
2998
3118
  },
2999
3119
  "gpt-5.4-mini": {
3000
3120
  id: "gpt-5.4-mini",
3001
- name: "GPT-5.4 mini",
3121
+ name: "GPT-5.4 Mini",
3002
3122
  api: "openai-responses",
3003
3123
  provider: "github-copilot",
3004
3124
  baseUrl: "https://api.individual.githubcopilot.com",
@@ -3131,7 +3251,7 @@ export const MODELS = {
3131
3251
  cost: {
3132
3252
  input: 0.3,
3133
3253
  output: 2.5,
3134
- cacheRead: 0.075,
3254
+ cacheRead: 0.03,
3135
3255
  cacheWrite: 0,
3136
3256
  },
3137
3257
  contextWindow: 1048576,
@@ -3250,7 +3370,7 @@ export const MODELS = {
3250
3370
  cost: {
3251
3371
  input: 1.25,
3252
3372
  output: 10,
3253
- cacheRead: 0.31,
3373
+ cacheRead: 0.125,
3254
3374
  cacheWrite: 0,
3255
3375
  },
3256
3376
  contextWindow: 1048576,
@@ -3443,6 +3563,57 @@ export const MODELS = {
3443
3563
  contextWindow: 131072,
3444
3564
  maxTokens: 65536,
3445
3565
  },
3566
+ "gemma-3-27b-it": {
3567
+ id: "gemma-3-27b-it",
3568
+ name: "Gemma 3 27B",
3569
+ api: "google-generative-ai",
3570
+ provider: "google",
3571
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3572
+ reasoning: false,
3573
+ input: ["text", "image"],
3574
+ cost: {
3575
+ input: 0,
3576
+ output: 0,
3577
+ cacheRead: 0,
3578
+ cacheWrite: 0,
3579
+ },
3580
+ contextWindow: 131072,
3581
+ maxTokens: 8192,
3582
+ },
3583
+ "gemma-4-26b-a4b-it": {
3584
+ id: "gemma-4-26b-a4b-it",
3585
+ name: "Gemma 4 26B",
3586
+ api: "google-generative-ai",
3587
+ provider: "google",
3588
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3589
+ reasoning: true,
3590
+ input: ["text", "image"],
3591
+ cost: {
3592
+ input: 0,
3593
+ output: 0,
3594
+ cacheRead: 0,
3595
+ cacheWrite: 0,
3596
+ },
3597
+ contextWindow: 256000,
3598
+ maxTokens: 8192,
3599
+ },
3600
+ "gemma-4-31b-it": {
3601
+ id: "gemma-4-31b-it",
3602
+ name: "Gemma 4 31B",
3603
+ api: "google-generative-ai",
3604
+ provider: "google",
3605
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3606
+ reasoning: true,
3607
+ input: ["text", "image"],
3608
+ cost: {
3609
+ input: 0,
3610
+ output: 0,
3611
+ cacheRead: 0,
3612
+ cacheWrite: 0,
3613
+ },
3614
+ contextWindow: 256000,
3615
+ maxTokens: 8192,
3616
+ },
3446
3617
  },
3447
3618
  "google-antigravity": {
3448
3619
  "claude-opus-4-5-thinking": {
@@ -3908,6 +4079,23 @@ export const MODELS = {
3908
4079
  contextWindow: 1048576,
3909
4080
  maxTokens: 65536,
3910
4081
  },
4082
+ "gemini-3.1-pro-preview-customtools": {
4083
+ id: "gemini-3.1-pro-preview-customtools",
4084
+ name: "Gemini 3.1 Pro Preview Custom Tools (Vertex)",
4085
+ api: "google-vertex",
4086
+ provider: "google-vertex",
4087
+ baseUrl: "https://{location}-aiplatform.googleapis.com",
4088
+ reasoning: true,
4089
+ input: ["text", "image"],
4090
+ cost: {
4091
+ input: 2,
4092
+ output: 12,
4093
+ cacheRead: 0.2,
4094
+ cacheWrite: 0,
4095
+ },
4096
+ contextWindow: 1048576,
4097
+ maxTokens: 65536,
4098
+ },
3911
4099
  },
3912
4100
  "groq": {
3913
4101
  "deepseek-r1-distill-llama-70b": {
@@ -3944,6 +4132,40 @@ export const MODELS = {
3944
4132
  contextWindow: 8192,
3945
4133
  maxTokens: 8192,
3946
4134
  },
4135
+ "groq/compound": {
4136
+ id: "groq/compound",
4137
+ name: "Compound",
4138
+ api: "openai-completions",
4139
+ provider: "groq",
4140
+ baseUrl: "https://api.groq.com/openai/v1",
4141
+ reasoning: true,
4142
+ input: ["text"],
4143
+ cost: {
4144
+ input: 0,
4145
+ output: 0,
4146
+ cacheRead: 0,
4147
+ cacheWrite: 0,
4148
+ },
4149
+ contextWindow: 131072,
4150
+ maxTokens: 8192,
4151
+ },
4152
+ "groq/compound-mini": {
4153
+ id: "groq/compound-mini",
4154
+ name: "Compound Mini",
4155
+ api: "openai-completions",
4156
+ provider: "groq",
4157
+ baseUrl: "https://api.groq.com/openai/v1",
4158
+ reasoning: true,
4159
+ input: ["text"],
4160
+ cost: {
4161
+ input: 0,
4162
+ output: 0,
4163
+ cacheRead: 0,
4164
+ cacheWrite: 0,
4165
+ },
4166
+ contextWindow: 131072,
4167
+ maxTokens: 8192,
4168
+ },
3947
4169
  "llama-3.1-8b-instant": {
3948
4170
  id: "llama-3.1-8b-instant",
3949
4171
  name: "Llama 3.1 8B Instant",
@@ -4131,9 +4353,26 @@ export const MODELS = {
4131
4353
  contextWindow: 131072,
4132
4354
  maxTokens: 65536,
4133
4355
  },
4134
- "qwen-qwq-32b": {
4135
- id: "qwen-qwq-32b",
4136
- name: "Qwen QwQ 32B",
4356
+ "openai/gpt-oss-safeguard-20b": {
4357
+ id: "openai/gpt-oss-safeguard-20b",
4358
+ name: "Safety GPT OSS 20B",
4359
+ api: "openai-completions",
4360
+ provider: "groq",
4361
+ baseUrl: "https://api.groq.com/openai/v1",
4362
+ reasoning: true,
4363
+ input: ["text"],
4364
+ cost: {
4365
+ input: 0.075,
4366
+ output: 0.3,
4367
+ cacheRead: 0.037,
4368
+ cacheWrite: 0,
4369
+ },
4370
+ contextWindow: 131072,
4371
+ maxTokens: 65536,
4372
+ },
4373
+ "qwen-qwq-32b": {
4374
+ id: "qwen-qwq-32b",
4375
+ name: "Qwen QwQ 32B",
4137
4376
  api: "openai-completions",
4138
4377
  provider: "groq",
4139
4378
  baseUrl: "https://api.groq.com/openai/v1",
@@ -4163,7 +4402,7 @@ export const MODELS = {
4163
4402
  cacheWrite: 0,
4164
4403
  },
4165
4404
  contextWindow: 131072,
4166
- maxTokens: 16384,
4405
+ maxTokens: 40960,
4167
4406
  },
4168
4407
  },
4169
4408
  "huggingface": {
@@ -4203,6 +4442,24 @@ export const MODELS = {
4203
4442
  contextWindow: 204800,
4204
4443
  maxTokens: 131072,
4205
4444
  },
4445
+ "MiniMaxAI/MiniMax-M2.7": {
4446
+ id: "MiniMaxAI/MiniMax-M2.7",
4447
+ name: "MiniMax-M2.7",
4448
+ api: "openai-completions",
4449
+ provider: "huggingface",
4450
+ baseUrl: "https://router.huggingface.co/v1",
4451
+ compat: { "supportsDeveloperRole": false },
4452
+ reasoning: true,
4453
+ input: ["text"],
4454
+ cost: {
4455
+ input: 0.3,
4456
+ output: 1.2,
4457
+ cacheRead: 0.06,
4458
+ cacheWrite: 0,
4459
+ },
4460
+ contextWindow: 204800,
4461
+ maxTokens: 131072,
4462
+ },
4206
4463
  "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4207
4464
  id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
4208
4465
  name: "Qwen3-235B-A22B-Thinking-2507",
@@ -4437,6 +4694,24 @@ export const MODELS = {
4437
4694
  contextWindow: 262144,
4438
4695
  maxTokens: 262144,
4439
4696
  },
4697
+ "moonshotai/Kimi-K2.6": {
4698
+ id: "moonshotai/Kimi-K2.6",
4699
+ name: "Kimi-K2.6",
4700
+ api: "openai-completions",
4701
+ provider: "huggingface",
4702
+ baseUrl: "https://router.huggingface.co/v1",
4703
+ compat: { "supportsDeveloperRole": false },
4704
+ reasoning: true,
4705
+ input: ["text", "image"],
4706
+ cost: {
4707
+ input: 0.95,
4708
+ output: 4,
4709
+ cacheRead: 0.16,
4710
+ cacheWrite: 0,
4711
+ },
4712
+ contextWindow: 262144,
4713
+ maxTokens: 262144,
4714
+ },
4440
4715
  "zai-org/GLM-4.7": {
4441
4716
  id: "zai-org/GLM-4.7",
4442
4717
  name: "GLM-4.7",
@@ -4491,6 +4766,24 @@ export const MODELS = {
4491
4766
  contextWindow: 202752,
4492
4767
  maxTokens: 131072,
4493
4768
  },
4769
+ "zai-org/GLM-5.1": {
4770
+ id: "zai-org/GLM-5.1",
4771
+ name: "GLM-5.1",
4772
+ api: "openai-completions",
4773
+ provider: "huggingface",
4774
+ baseUrl: "https://router.huggingface.co/v1",
4775
+ compat: { "supportsDeveloperRole": false },
4776
+ reasoning: true,
4777
+ input: ["text"],
4778
+ cost: {
4779
+ input: 1,
4780
+ output: 3.2,
4781
+ cacheRead: 0.2,
4782
+ cacheWrite: 0,
4783
+ },
4784
+ contextWindow: 202752,
4785
+ maxTokens: 131072,
4786
+ },
4494
4787
  },
4495
4788
  "kimi-coding": {
4496
4789
  "k2p5": {
@@ -4510,14 +4803,14 @@ export const MODELS = {
4510
4803
  contextWindow: 262144,
4511
4804
  maxTokens: 32768,
4512
4805
  },
4513
- "kimi-k2-thinking": {
4514
- id: "kimi-k2-thinking",
4515
- name: "Kimi K2 Thinking",
4806
+ "k2p6": {
4807
+ id: "k2p6",
4808
+ name: "Kimi K2.6",
4516
4809
  api: "anthropic-messages",
4517
4810
  provider: "kimi-coding",
4518
4811
  baseUrl: "https://api.kimi.com/coding",
4519
4812
  reasoning: true,
4520
- input: ["text"],
4813
+ input: ["text", "image"],
4521
4814
  cost: {
4522
4815
  input: 0,
4523
4816
  output: 0,
@@ -4527,93 +4820,25 @@ export const MODELS = {
4527
4820
  contextWindow: 262144,
4528
4821
  maxTokens: 32768,
4529
4822
  },
4530
- },
4531
- "minimax": {
4532
- "MiniMax-M2": {
4533
- id: "MiniMax-M2",
4534
- name: "MiniMax-M2",
4535
- api: "anthropic-messages",
4536
- provider: "minimax",
4537
- baseUrl: "https://api.minimax.io/anthropic",
4538
- reasoning: true,
4539
- input: ["text"],
4540
- cost: {
4541
- input: 0.3,
4542
- output: 1.2,
4543
- cacheRead: 0,
4544
- cacheWrite: 0,
4545
- },
4546
- contextWindow: 204800,
4547
- maxTokens: 131072,
4548
- },
4549
- "MiniMax-M2.1": {
4550
- id: "MiniMax-M2.1",
4551
- name: "MiniMax-M2.1",
4552
- api: "anthropic-messages",
4553
- provider: "minimax",
4554
- baseUrl: "https://api.minimax.io/anthropic",
4555
- reasoning: true,
4556
- input: ["text"],
4557
- cost: {
4558
- input: 0.3,
4559
- output: 1.2,
4560
- cacheRead: 0,
4561
- cacheWrite: 0,
4562
- },
4563
- contextWindow: 204800,
4564
- maxTokens: 131072,
4565
- },
4566
- "MiniMax-M2.1-highspeed": {
4567
- id: "MiniMax-M2.1-highspeed",
4568
- name: "MiniMax-M2.1-highspeed",
4823
+ "kimi-k2-thinking": {
4824
+ id: "kimi-k2-thinking",
4825
+ name: "Kimi K2 Thinking",
4569
4826
  api: "anthropic-messages",
4570
- provider: "minimax",
4571
- baseUrl: "https://api.minimax.io/anthropic",
4827
+ provider: "kimi-coding",
4828
+ baseUrl: "https://api.kimi.com/coding",
4572
4829
  reasoning: true,
4573
4830
  input: ["text"],
4574
4831
  cost: {
4575
- input: 0.6,
4576
- output: 2.4,
4832
+ input: 0,
4833
+ output: 0,
4577
4834
  cacheRead: 0,
4578
4835
  cacheWrite: 0,
4579
4836
  },
4580
- contextWindow: 204800,
4581
- maxTokens: 131072,
4582
- },
4583
- "MiniMax-M2.5": {
4584
- id: "MiniMax-M2.5",
4585
- name: "MiniMax-M2.5",
4586
- api: "anthropic-messages",
4587
- provider: "minimax",
4588
- baseUrl: "https://api.minimax.io/anthropic",
4589
- reasoning: true,
4590
- input: ["text"],
4591
- cost: {
4592
- input: 0.3,
4593
- output: 1.2,
4594
- cacheRead: 0.03,
4595
- cacheWrite: 0.375,
4596
- },
4597
- contextWindow: 204800,
4598
- maxTokens: 131072,
4599
- },
4600
- "MiniMax-M2.5-highspeed": {
4601
- id: "MiniMax-M2.5-highspeed",
4602
- name: "MiniMax-M2.5-highspeed",
4603
- api: "anthropic-messages",
4604
- provider: "minimax",
4605
- baseUrl: "https://api.minimax.io/anthropic",
4606
- reasoning: true,
4607
- input: ["text"],
4608
- cost: {
4609
- input: 0.6,
4610
- output: 2.4,
4611
- cacheRead: 0.06,
4612
- cacheWrite: 0.375,
4613
- },
4614
- contextWindow: 204800,
4615
- maxTokens: 131072,
4837
+ contextWindow: 262144,
4838
+ maxTokens: 32768,
4616
4839
  },
4840
+ },
4841
+ "minimax": {
4617
4842
  "MiniMax-M2.7": {
4618
4843
  id: "MiniMax-M2.7",
4619
4844
  name: "MiniMax-M2.7",
@@ -4650,91 +4875,6 @@ export const MODELS = {
4650
4875
  },
4651
4876
  },
4652
4877
  "minimax-cn": {
4653
- "MiniMax-M2": {
4654
- id: "MiniMax-M2",
4655
- name: "MiniMax-M2",
4656
- api: "anthropic-messages",
4657
- provider: "minimax-cn",
4658
- baseUrl: "https://api.minimaxi.com/anthropic",
4659
- reasoning: true,
4660
- input: ["text"],
4661
- cost: {
4662
- input: 0.3,
4663
- output: 1.2,
4664
- cacheRead: 0,
4665
- cacheWrite: 0,
4666
- },
4667
- contextWindow: 204800,
4668
- maxTokens: 131072,
4669
- },
4670
- "MiniMax-M2.1": {
4671
- id: "MiniMax-M2.1",
4672
- name: "MiniMax-M2.1",
4673
- api: "anthropic-messages",
4674
- provider: "minimax-cn",
4675
- baseUrl: "https://api.minimaxi.com/anthropic",
4676
- reasoning: true,
4677
- input: ["text"],
4678
- cost: {
4679
- input: 0.3,
4680
- output: 1.2,
4681
- cacheRead: 0,
4682
- cacheWrite: 0,
4683
- },
4684
- contextWindow: 204800,
4685
- maxTokens: 131072,
4686
- },
4687
- "MiniMax-M2.1-highspeed": {
4688
- id: "MiniMax-M2.1-highspeed",
4689
- name: "MiniMax-M2.1-highspeed",
4690
- api: "anthropic-messages",
4691
- provider: "minimax-cn",
4692
- baseUrl: "https://api.minimaxi.com/anthropic",
4693
- reasoning: true,
4694
- input: ["text"],
4695
- cost: {
4696
- input: 0.6,
4697
- output: 2.4,
4698
- cacheRead: 0,
4699
- cacheWrite: 0,
4700
- },
4701
- contextWindow: 204800,
4702
- maxTokens: 131072,
4703
- },
4704
- "MiniMax-M2.5": {
4705
- id: "MiniMax-M2.5",
4706
- name: "MiniMax-M2.5",
4707
- api: "anthropic-messages",
4708
- provider: "minimax-cn",
4709
- baseUrl: "https://api.minimaxi.com/anthropic",
4710
- reasoning: true,
4711
- input: ["text"],
4712
- cost: {
4713
- input: 0.3,
4714
- output: 1.2,
4715
- cacheRead: 0.03,
4716
- cacheWrite: 0.375,
4717
- },
4718
- contextWindow: 204800,
4719
- maxTokens: 131072,
4720
- },
4721
- "MiniMax-M2.5-highspeed": {
4722
- id: "MiniMax-M2.5-highspeed",
4723
- name: "MiniMax-M2.5-highspeed",
4724
- api: "anthropic-messages",
4725
- provider: "minimax-cn",
4726
- baseUrl: "https://api.minimaxi.com/anthropic",
4727
- reasoning: true,
4728
- input: ["text"],
4729
- cost: {
4730
- input: 0.6,
4731
- output: 2.4,
4732
- cacheRead: 0.06,
4733
- cacheWrite: 0.375,
4734
- },
4735
- contextWindow: 204800,
4736
- maxTokens: 131072,
4737
- },
4738
4878
  "MiniMax-M2.7": {
4739
4879
  id: "MiniMax-M2.7",
4740
4880
  name: "MiniMax-M2.7",
@@ -4774,9 +4914,9 @@ export const MODELS = {
4774
4914
  "codestral-latest": {
4775
4915
  id: "codestral-latest",
4776
4916
  name: "Codestral (latest)",
4777
- api: "mistral-conversations",
4917
+ api: "openai-completions",
4778
4918
  provider: "mistral",
4779
- baseUrl: "https://api.mistral.ai",
4919
+ baseUrl: "https://api.mistral.ai/v1",
4780
4920
  reasoning: false,
4781
4921
  input: ["text"],
4782
4922
  cost: {
@@ -4791,9 +4931,9 @@ export const MODELS = {
4791
4931
  "devstral-2512": {
4792
4932
  id: "devstral-2512",
4793
4933
  name: "Devstral 2",
4794
- api: "mistral-conversations",
4934
+ api: "openai-completions",
4795
4935
  provider: "mistral",
4796
- baseUrl: "https://api.mistral.ai",
4936
+ baseUrl: "https://api.mistral.ai/v1",
4797
4937
  reasoning: false,
4798
4938
  input: ["text"],
4799
4939
  cost: {
@@ -4808,9 +4948,9 @@ export const MODELS = {
4808
4948
  "devstral-medium-2507": {
4809
4949
  id: "devstral-medium-2507",
4810
4950
  name: "Devstral Medium",
4811
- api: "mistral-conversations",
4951
+ api: "openai-completions",
4812
4952
  provider: "mistral",
4813
- baseUrl: "https://api.mistral.ai",
4953
+ baseUrl: "https://api.mistral.ai/v1",
4814
4954
  reasoning: false,
4815
4955
  input: ["text"],
4816
4956
  cost: {
@@ -4825,9 +4965,9 @@ export const MODELS = {
4825
4965
  "devstral-medium-latest": {
4826
4966
  id: "devstral-medium-latest",
4827
4967
  name: "Devstral 2 (latest)",
4828
- api: "mistral-conversations",
4968
+ api: "openai-completions",
4829
4969
  provider: "mistral",
4830
- baseUrl: "https://api.mistral.ai",
4970
+ baseUrl: "https://api.mistral.ai/v1",
4831
4971
  reasoning: false,
4832
4972
  input: ["text"],
4833
4973
  cost: {
@@ -4842,9 +4982,9 @@ export const MODELS = {
4842
4982
  "devstral-small-2505": {
4843
4983
  id: "devstral-small-2505",
4844
4984
  name: "Devstral Small 2505",
4845
- api: "mistral-conversations",
4985
+ api: "openai-completions",
4846
4986
  provider: "mistral",
4847
- baseUrl: "https://api.mistral.ai",
4987
+ baseUrl: "https://api.mistral.ai/v1",
4848
4988
  reasoning: false,
4849
4989
  input: ["text"],
4850
4990
  cost: {
@@ -4859,9 +4999,9 @@ export const MODELS = {
4859
4999
  "devstral-small-2507": {
4860
5000
  id: "devstral-small-2507",
4861
5001
  name: "Devstral Small",
4862
- api: "mistral-conversations",
5002
+ api: "openai-completions",
4863
5003
  provider: "mistral",
4864
- baseUrl: "https://api.mistral.ai",
5004
+ baseUrl: "https://api.mistral.ai/v1",
4865
5005
  reasoning: false,
4866
5006
  input: ["text"],
4867
5007
  cost: {
@@ -4876,9 +5016,9 @@ export const MODELS = {
4876
5016
  "labs-devstral-small-2512": {
4877
5017
  id: "labs-devstral-small-2512",
4878
5018
  name: "Devstral Small 2",
4879
- api: "mistral-conversations",
5019
+ api: "openai-completions",
4880
5020
  provider: "mistral",
4881
- baseUrl: "https://api.mistral.ai",
5021
+ baseUrl: "https://api.mistral.ai/v1",
4882
5022
  reasoning: false,
4883
5023
  input: ["text", "image"],
4884
5024
  cost: {
@@ -4893,9 +5033,9 @@ export const MODELS = {
4893
5033
  "magistral-medium-latest": {
4894
5034
  id: "magistral-medium-latest",
4895
5035
  name: "Magistral Medium (latest)",
4896
- api: "mistral-conversations",
5036
+ api: "openai-completions",
4897
5037
  provider: "mistral",
4898
- baseUrl: "https://api.mistral.ai",
5038
+ baseUrl: "https://api.mistral.ai/v1",
4899
5039
  reasoning: true,
4900
5040
  input: ["text"],
4901
5041
  cost: {
@@ -4910,9 +5050,9 @@ export const MODELS = {
4910
5050
  "magistral-small": {
4911
5051
  id: "magistral-small",
4912
5052
  name: "Magistral Small",
4913
- api: "mistral-conversations",
5053
+ api: "openai-completions",
4914
5054
  provider: "mistral",
4915
- baseUrl: "https://api.mistral.ai",
5055
+ baseUrl: "https://api.mistral.ai/v1",
4916
5056
  reasoning: true,
4917
5057
  input: ["text"],
4918
5058
  cost: {
@@ -4927,9 +5067,9 @@ export const MODELS = {
4927
5067
  "ministral-3b-latest": {
4928
5068
  id: "ministral-3b-latest",
4929
5069
  name: "Ministral 3B (latest)",
4930
- api: "mistral-conversations",
5070
+ api: "openai-completions",
4931
5071
  provider: "mistral",
4932
- baseUrl: "https://api.mistral.ai",
5072
+ baseUrl: "https://api.mistral.ai/v1",
4933
5073
  reasoning: false,
4934
5074
  input: ["text"],
4935
5075
  cost: {
@@ -4944,9 +5084,9 @@ export const MODELS = {
4944
5084
  "ministral-8b-latest": {
4945
5085
  id: "ministral-8b-latest",
4946
5086
  name: "Ministral 8B (latest)",
4947
- api: "mistral-conversations",
5087
+ api: "openai-completions",
4948
5088
  provider: "mistral",
4949
- baseUrl: "https://api.mistral.ai",
5089
+ baseUrl: "https://api.mistral.ai/v1",
4950
5090
  reasoning: false,
4951
5091
  input: ["text"],
4952
5092
  cost: {
@@ -4961,9 +5101,9 @@ export const MODELS = {
4961
5101
  "mistral-large-2411": {
4962
5102
  id: "mistral-large-2411",
4963
5103
  name: "Mistral Large 2.1",
4964
- api: "mistral-conversations",
5104
+ api: "openai-completions",
4965
5105
  provider: "mistral",
4966
- baseUrl: "https://api.mistral.ai",
5106
+ baseUrl: "https://api.mistral.ai/v1",
4967
5107
  reasoning: false,
4968
5108
  input: ["text"],
4969
5109
  cost: {
@@ -4978,9 +5118,9 @@ export const MODELS = {
4978
5118
  "mistral-large-2512": {
4979
5119
  id: "mistral-large-2512",
4980
5120
  name: "Mistral Large 3",
4981
- api: "mistral-conversations",
5121
+ api: "openai-completions",
4982
5122
  provider: "mistral",
4983
- baseUrl: "https://api.mistral.ai",
5123
+ baseUrl: "https://api.mistral.ai/v1",
4984
5124
  reasoning: false,
4985
5125
  input: ["text", "image"],
4986
5126
  cost: {
@@ -4995,9 +5135,9 @@ export const MODELS = {
4995
5135
  "mistral-large-latest": {
4996
5136
  id: "mistral-large-latest",
4997
5137
  name: "Mistral Large (latest)",
4998
- api: "mistral-conversations",
5138
+ api: "openai-completions",
4999
5139
  provider: "mistral",
5000
- baseUrl: "https://api.mistral.ai",
5140
+ baseUrl: "https://api.mistral.ai/v1",
5001
5141
  reasoning: false,
5002
5142
  input: ["text", "image"],
5003
5143
  cost: {
@@ -5012,9 +5152,9 @@ export const MODELS = {
5012
5152
  "mistral-medium-2505": {
5013
5153
  id: "mistral-medium-2505",
5014
5154
  name: "Mistral Medium 3",
5015
- api: "mistral-conversations",
5155
+ api: "openai-completions",
5016
5156
  provider: "mistral",
5017
- baseUrl: "https://api.mistral.ai",
5157
+ baseUrl: "https://api.mistral.ai/v1",
5018
5158
  reasoning: false,
5019
5159
  input: ["text", "image"],
5020
5160
  cost: {
@@ -5029,9 +5169,9 @@ export const MODELS = {
5029
5169
  "mistral-medium-2508": {
5030
5170
  id: "mistral-medium-2508",
5031
5171
  name: "Mistral Medium 3.1",
5032
- api: "mistral-conversations",
5172
+ api: "openai-completions",
5033
5173
  provider: "mistral",
5034
- baseUrl: "https://api.mistral.ai",
5174
+ baseUrl: "https://api.mistral.ai/v1",
5035
5175
  reasoning: false,
5036
5176
  input: ["text", "image"],
5037
5177
  cost: {
@@ -5046,9 +5186,9 @@ export const MODELS = {
5046
5186
  "mistral-medium-latest": {
5047
5187
  id: "mistral-medium-latest",
5048
5188
  name: "Mistral Medium (latest)",
5049
- api: "mistral-conversations",
5189
+ api: "openai-completions",
5050
5190
  provider: "mistral",
5051
- baseUrl: "https://api.mistral.ai",
5191
+ baseUrl: "https://api.mistral.ai/v1",
5052
5192
  reasoning: false,
5053
5193
  input: ["text", "image"],
5054
5194
  cost: {
@@ -5063,9 +5203,9 @@ export const MODELS = {
5063
5203
  "mistral-nemo": {
5064
5204
  id: "mistral-nemo",
5065
5205
  name: "Mistral Nemo",
5066
- api: "mistral-conversations",
5206
+ api: "openai-completions",
5067
5207
  provider: "mistral",
5068
- baseUrl: "https://api.mistral.ai",
5208
+ baseUrl: "https://api.mistral.ai/v1",
5069
5209
  reasoning: false,
5070
5210
  input: ["text"],
5071
5211
  cost: {
@@ -5080,9 +5220,9 @@ export const MODELS = {
5080
5220
  "mistral-small-2506": {
5081
5221
  id: "mistral-small-2506",
5082
5222
  name: "Mistral Small 3.2",
5083
- api: "mistral-conversations",
5223
+ api: "openai-completions",
5084
5224
  provider: "mistral",
5085
- baseUrl: "https://api.mistral.ai",
5225
+ baseUrl: "https://api.mistral.ai/v1",
5086
5226
  reasoning: false,
5087
5227
  input: ["text", "image"],
5088
5228
  cost: {
@@ -5094,29 +5234,46 @@ export const MODELS = {
5094
5234
  contextWindow: 128000,
5095
5235
  maxTokens: 16384,
5096
5236
  },
5237
+ "mistral-small-2603": {
5238
+ id: "mistral-small-2603",
5239
+ name: "Mistral Small 4",
5240
+ api: "openai-completions",
5241
+ provider: "mistral",
5242
+ baseUrl: "https://api.mistral.ai/v1",
5243
+ reasoning: true,
5244
+ input: ["text", "image"],
5245
+ cost: {
5246
+ input: 0.15,
5247
+ output: 0.6,
5248
+ cacheRead: 0,
5249
+ cacheWrite: 0,
5250
+ },
5251
+ contextWindow: 256000,
5252
+ maxTokens: 256000,
5253
+ },
5097
5254
  "mistral-small-latest": {
5098
5255
  id: "mistral-small-latest",
5099
5256
  name: "Mistral Small (latest)",
5100
- api: "mistral-conversations",
5257
+ api: "openai-completions",
5101
5258
  provider: "mistral",
5102
- baseUrl: "https://api.mistral.ai",
5103
- reasoning: false,
5259
+ baseUrl: "https://api.mistral.ai/v1",
5260
+ reasoning: true,
5104
5261
  input: ["text", "image"],
5105
5262
  cost: {
5106
- input: 0.1,
5107
- output: 0.3,
5263
+ input: 0.15,
5264
+ output: 0.6,
5108
5265
  cacheRead: 0,
5109
5266
  cacheWrite: 0,
5110
5267
  },
5111
- contextWindow: 128000,
5112
- maxTokens: 16384,
5268
+ contextWindow: 256000,
5269
+ maxTokens: 256000,
5113
5270
  },
5114
5271
  "open-mistral-7b": {
5115
5272
  id: "open-mistral-7b",
5116
5273
  name: "Mistral 7B",
5117
- api: "mistral-conversations",
5274
+ api: "openai-completions",
5118
5275
  provider: "mistral",
5119
- baseUrl: "https://api.mistral.ai",
5276
+ baseUrl: "https://api.mistral.ai/v1",
5120
5277
  reasoning: false,
5121
5278
  input: ["text"],
5122
5279
  cost: {
@@ -5131,9 +5288,9 @@ export const MODELS = {
5131
5288
  "open-mixtral-8x22b": {
5132
5289
  id: "open-mixtral-8x22b",
5133
5290
  name: "Mixtral 8x22B",
5134
- api: "mistral-conversations",
5291
+ api: "openai-completions",
5135
5292
  provider: "mistral",
5136
- baseUrl: "https://api.mistral.ai",
5293
+ baseUrl: "https://api.mistral.ai/v1",
5137
5294
  reasoning: false,
5138
5295
  input: ["text"],
5139
5296
  cost: {
@@ -5148,9 +5305,9 @@ export const MODELS = {
5148
5305
  "open-mixtral-8x7b": {
5149
5306
  id: "open-mixtral-8x7b",
5150
5307
  name: "Mixtral 8x7B",
5151
- api: "mistral-conversations",
5308
+ api: "openai-completions",
5152
5309
  provider: "mistral",
5153
- baseUrl: "https://api.mistral.ai",
5310
+ baseUrl: "https://api.mistral.ai/v1",
5154
5311
  reasoning: false,
5155
5312
  input: ["text"],
5156
5313
  cost: {
@@ -5165,9 +5322,9 @@ export const MODELS = {
5165
5322
  "pixtral-12b": {
5166
5323
  id: "pixtral-12b",
5167
5324
  name: "Pixtral 12B",
5168
- api: "mistral-conversations",
5325
+ api: "openai-completions",
5169
5326
  provider: "mistral",
5170
- baseUrl: "https://api.mistral.ai",
5327
+ baseUrl: "https://api.mistral.ai/v1",
5171
5328
  reasoning: false,
5172
5329
  input: ["text", "image"],
5173
5330
  cost: {
@@ -5182,9 +5339,9 @@ export const MODELS = {
5182
5339
  "pixtral-large-latest": {
5183
5340
  id: "pixtral-large-latest",
5184
5341
  name: "Pixtral Large (latest)",
5185
- api: "mistral-conversations",
5342
+ api: "openai-completions",
5186
5343
  provider: "mistral",
5187
- baseUrl: "https://api.mistral.ai",
5344
+ baseUrl: "https://api.mistral.ai/v1",
5188
5345
  reasoning: false,
5189
5346
  input: ["text", "image"],
5190
5347
  cost: {
@@ -5198,23 +5355,6 @@ export const MODELS = {
5198
5355
  },
5199
5356
  },
5200
5357
  "openai": {
5201
- "codex-mini-latest": {
5202
- id: "codex-mini-latest",
5203
- name: "Codex Mini",
5204
- api: "openai-responses",
5205
- provider: "openai",
5206
- baseUrl: "https://api.openai.com/v1",
5207
- reasoning: true,
5208
- input: ["text"],
5209
- cost: {
5210
- input: 1.5,
5211
- output: 6,
5212
- cacheRead: 0.375,
5213
- cacheWrite: 0,
5214
- },
5215
- contextWindow: 200000,
5216
- maxTokens: 100000,
5217
- },
5218
5358
  "gpt-4": {
5219
5359
  id: "gpt-4",
5220
5360
  name: "GPT-4",
@@ -5640,6 +5780,23 @@ export const MODELS = {
5640
5780
  contextWindow: 400000,
5641
5781
  maxTokens: 128000,
5642
5782
  },
5783
+ "gpt-5.3-chat-latest": {
5784
+ id: "gpt-5.3-chat-latest",
5785
+ name: "GPT-5.3 Chat (latest)",
5786
+ api: "openai-responses",
5787
+ provider: "openai",
5788
+ baseUrl: "https://api.openai.com/v1",
5789
+ reasoning: false,
5790
+ input: ["text", "image"],
5791
+ cost: {
5792
+ input: 1.75,
5793
+ output: 14,
5794
+ cacheRead: 0.175,
5795
+ cacheWrite: 0,
5796
+ },
5797
+ contextWindow: 128000,
5798
+ maxTokens: 16384,
5799
+ },
5643
5800
  "gpt-5.3-codex": {
5644
5801
  id: "gpt-5.3-codex",
5645
5802
  name: "GPT-5.3 Codex",
@@ -6137,6 +6294,23 @@ export const MODELS = {
6137
6294
  contextWindow: 1000000,
6138
6295
  maxTokens: 128000,
6139
6296
  },
6297
+ "claude-opus-4-7": {
6298
+ id: "claude-opus-4-7",
6299
+ name: "Claude Opus 4.7",
6300
+ api: "anthropic-messages",
6301
+ provider: "opencode",
6302
+ baseUrl: "https://opencode.ai/zen",
6303
+ reasoning: true,
6304
+ input: ["text", "image"],
6305
+ cost: {
6306
+ input: 5,
6307
+ output: 25,
6308
+ cacheRead: 0.5,
6309
+ cacheWrite: 6.25,
6310
+ },
6311
+ contextWindow: 1000000,
6312
+ maxTokens: 128000,
6313
+ },
6140
6314
  "claude-sonnet-4": {
6141
6315
  id: "claude-sonnet-4",
6142
6316
  name: "Claude Sonnet 4",
@@ -6239,6 +6413,23 @@ export const MODELS = {
6239
6413
  contextWindow: 204800,
6240
6414
  maxTokens: 131072,
6241
6415
  },
6416
+ "glm-5.1": {
6417
+ id: "glm-5.1",
6418
+ name: "GLM-5.1",
6419
+ api: "openai-completions",
6420
+ provider: "opencode",
6421
+ baseUrl: "https://opencode.ai/zen/v1",
6422
+ reasoning: true,
6423
+ input: ["text"],
6424
+ cost: {
6425
+ input: 1.4,
6426
+ output: 4.4,
6427
+ cacheRead: 0.26,
6428
+ cacheWrite: 0,
6429
+ },
6430
+ contextWindow: 204800,
6431
+ maxTokens: 131072,
6432
+ },
6242
6433
  "gpt-5": {
6243
6434
  id: "gpt-5",
6244
6435
  name: "GPT-5",
@@ -6477,6 +6668,23 @@ export const MODELS = {
6477
6668
  contextWindow: 1050000,
6478
6669
  maxTokens: 128000,
6479
6670
  },
6671
+ "hy3-preview-free": {
6672
+ id: "hy3-preview-free",
6673
+ name: "Hy3 Preview Free",
6674
+ api: "openai-completions",
6675
+ provider: "opencode",
6676
+ baseUrl: "https://opencode.ai/zen/v1",
6677
+ reasoning: true,
6678
+ input: ["text"],
6679
+ cost: {
6680
+ input: 0,
6681
+ output: 0,
6682
+ cacheRead: 0,
6683
+ cacheWrite: 0,
6684
+ },
6685
+ contextWindow: 256000,
6686
+ maxTokens: 64000,
6687
+ },
6480
6688
  "kimi-k2.5": {
6481
6689
  id: "kimi-k2.5",
6482
6690
  name: "Kimi K2.5",
@@ -6494,30 +6702,30 @@ export const MODELS = {
6494
6702
  contextWindow: 262144,
6495
6703
  maxTokens: 65536,
6496
6704
  },
6497
- "mimo-v2-omni-free": {
6498
- id: "mimo-v2-omni-free",
6499
- name: "MiMo V2 Omni Free",
6705
+ "kimi-k2.6": {
6706
+ id: "kimi-k2.6",
6707
+ name: "Kimi K2.6",
6500
6708
  api: "openai-completions",
6501
6709
  provider: "opencode",
6502
6710
  baseUrl: "https://opencode.ai/zen/v1",
6503
6711
  reasoning: true,
6504
6712
  input: ["text", "image"],
6505
6713
  cost: {
6506
- input: 0,
6507
- output: 0,
6508
- cacheRead: 0,
6714
+ input: 0.95,
6715
+ output: 4,
6716
+ cacheRead: 0.16,
6509
6717
  cacheWrite: 0,
6510
6718
  },
6511
6719
  contextWindow: 262144,
6512
- maxTokens: 64000,
6720
+ maxTokens: 65536,
6513
6721
  },
6514
- "mimo-v2-pro-free": {
6515
- id: "mimo-v2-pro-free",
6516
- name: "MiMo V2 Pro Free",
6722
+ "ling-2.6-flash-free": {
6723
+ id: "ling-2.6-flash-free",
6724
+ name: "Ling 2.6 Flash Free",
6517
6725
  api: "openai-completions",
6518
6726
  provider: "opencode",
6519
6727
  baseUrl: "https://opencode.ai/zen/v1",
6520
- reasoning: true,
6728
+ reasoning: false,
6521
6729
  input: ["text"],
6522
6730
  cost: {
6523
6731
  input: 0,
@@ -6525,8 +6733,8 @@ export const MODELS = {
6525
6733
  cacheRead: 0,
6526
6734
  cacheWrite: 0,
6527
6735
  },
6528
- contextWindow: 1048576,
6529
- maxTokens: 64000,
6736
+ contextWindow: 262100,
6737
+ maxTokens: 32800,
6530
6738
  },
6531
6739
  "minimax-m2.5": {
6532
6740
  id: "minimax-m2.5",
@@ -6562,6 +6770,23 @@ export const MODELS = {
6562
6770
  contextWindow: 204800,
6563
6771
  maxTokens: 131072,
6564
6772
  },
6773
+ "minimax-m2.7": {
6774
+ id: "minimax-m2.7",
6775
+ name: "MiniMax M2.7",
6776
+ api: "openai-completions",
6777
+ provider: "opencode",
6778
+ baseUrl: "https://opencode.ai/zen/v1",
6779
+ reasoning: true,
6780
+ input: ["text"],
6781
+ cost: {
6782
+ input: 0.3,
6783
+ output: 1.2,
6784
+ cacheRead: 0.06,
6785
+ cacheWrite: 0,
6786
+ },
6787
+ contextWindow: 204800,
6788
+ maxTokens: 131072,
6789
+ },
6565
6790
  "nemotron-3-super-free": {
6566
6791
  id: "nemotron-3-super-free",
6567
6792
  name: "Nemotron 3 Super Free",
@@ -6576,9 +6801,43 @@ export const MODELS = {
6576
6801
  cacheRead: 0,
6577
6802
  cacheWrite: 0,
6578
6803
  },
6579
- contextWindow: 1000000,
6804
+ contextWindow: 204800,
6580
6805
  maxTokens: 128000,
6581
6806
  },
6807
+ "qwen3.5-plus": {
6808
+ id: "qwen3.5-plus",
6809
+ name: "Qwen3.5 Plus",
6810
+ api: "anthropic-messages",
6811
+ provider: "opencode",
6812
+ baseUrl: "https://opencode.ai/zen",
6813
+ reasoning: true,
6814
+ input: ["text", "image"],
6815
+ cost: {
6816
+ input: 0.2,
6817
+ output: 1.2,
6818
+ cacheRead: 0.02,
6819
+ cacheWrite: 0.25,
6820
+ },
6821
+ contextWindow: 262144,
6822
+ maxTokens: 65536,
6823
+ },
6824
+ "qwen3.6-plus": {
6825
+ id: "qwen3.6-plus",
6826
+ name: "Qwen3.6 Plus",
6827
+ api: "anthropic-messages",
6828
+ provider: "opencode",
6829
+ baseUrl: "https://opencode.ai/zen",
6830
+ reasoning: true,
6831
+ input: ["text", "image"],
6832
+ cost: {
6833
+ input: 0.5,
6834
+ output: 3,
6835
+ cacheRead: 0.05,
6836
+ cacheWrite: 0.625,
6837
+ },
6838
+ contextWindow: 262144,
6839
+ maxTokens: 65536,
6840
+ },
6582
6841
  },
6583
6842
  "opencode-go": {
6584
6843
  "glm-5": {
@@ -6598,6 +6857,23 @@ export const MODELS = {
6598
6857
  contextWindow: 204800,
6599
6858
  maxTokens: 131072,
6600
6859
  },
6860
+ "glm-5.1": {
6861
+ id: "glm-5.1",
6862
+ name: "GLM-5.1",
6863
+ api: "openai-completions",
6864
+ provider: "opencode-go",
6865
+ baseUrl: "https://opencode.ai/zen/go/v1",
6866
+ reasoning: true,
6867
+ input: ["text"],
6868
+ cost: {
6869
+ input: 1.4,
6870
+ output: 4.4,
6871
+ cacheRead: 0.26,
6872
+ cacheWrite: 0,
6873
+ },
6874
+ contextWindow: 204800,
6875
+ maxTokens: 131072,
6876
+ },
6601
6877
  "kimi-k2.5": {
6602
6878
  id: "kimi-k2.5",
6603
6879
  name: "Kimi K2.5",
@@ -6615,12 +6891,97 @@ export const MODELS = {
6615
6891
  contextWindow: 262144,
6616
6892
  maxTokens: 65536,
6617
6893
  },
6894
+ "kimi-k2.6": {
6895
+ id: "kimi-k2.6",
6896
+ name: "Kimi K2.6 (3x limits)",
6897
+ api: "openai-completions",
6898
+ provider: "opencode-go",
6899
+ baseUrl: "https://opencode.ai/zen/go/v1",
6900
+ reasoning: true,
6901
+ input: ["text", "image"],
6902
+ cost: {
6903
+ input: 0.32,
6904
+ output: 1.34,
6905
+ cacheRead: 0.054,
6906
+ cacheWrite: 0,
6907
+ },
6908
+ contextWindow: 262144,
6909
+ maxTokens: 65536,
6910
+ },
6911
+ "mimo-v2-omni": {
6912
+ id: "mimo-v2-omni",
6913
+ name: "MiMo V2 Omni",
6914
+ api: "openai-completions",
6915
+ provider: "opencode-go",
6916
+ baseUrl: "https://opencode.ai/zen/go/v1",
6917
+ reasoning: true,
6918
+ input: ["text", "image"],
6919
+ cost: {
6920
+ input: 0.4,
6921
+ output: 2,
6922
+ cacheRead: 0.08,
6923
+ cacheWrite: 0,
6924
+ },
6925
+ contextWindow: 262144,
6926
+ maxTokens: 128000,
6927
+ },
6928
+ "mimo-v2-pro": {
6929
+ id: "mimo-v2-pro",
6930
+ name: "MiMo V2 Pro",
6931
+ api: "openai-completions",
6932
+ provider: "opencode-go",
6933
+ baseUrl: "https://opencode.ai/zen/go/v1",
6934
+ reasoning: true,
6935
+ input: ["text"],
6936
+ cost: {
6937
+ input: 1,
6938
+ output: 3,
6939
+ cacheRead: 0.2,
6940
+ cacheWrite: 0,
6941
+ },
6942
+ contextWindow: 1048576,
6943
+ maxTokens: 128000,
6944
+ },
6945
+ "mimo-v2.5": {
6946
+ id: "mimo-v2.5",
6947
+ name: "MiMo V2.5",
6948
+ api: "openai-completions",
6949
+ provider: "opencode-go",
6950
+ baseUrl: "https://opencode.ai/zen/go/v1",
6951
+ reasoning: true,
6952
+ input: ["text", "image"],
6953
+ cost: {
6954
+ input: 0.4,
6955
+ output: 2,
6956
+ cacheRead: 0.08,
6957
+ cacheWrite: 0,
6958
+ },
6959
+ contextWindow: 262144,
6960
+ maxTokens: 128000,
6961
+ },
6962
+ "mimo-v2.5-pro": {
6963
+ id: "mimo-v2.5-pro",
6964
+ name: "MiMo V2.5 Pro",
6965
+ api: "openai-completions",
6966
+ provider: "opencode-go",
6967
+ baseUrl: "https://opencode.ai/zen/go/v1",
6968
+ reasoning: true,
6969
+ input: ["text"],
6970
+ cost: {
6971
+ input: 1,
6972
+ output: 3,
6973
+ cacheRead: 0.2,
6974
+ cacheWrite: 0,
6975
+ },
6976
+ contextWindow: 1048576,
6977
+ maxTokens: 128000,
6978
+ },
6618
6979
  "minimax-m2.5": {
6619
6980
  id: "minimax-m2.5",
6620
6981
  name: "MiniMax M2.5",
6621
- api: "anthropic-messages",
6982
+ api: "openai-completions",
6622
6983
  provider: "opencode-go",
6623
- baseUrl: "https://opencode.ai/zen/go",
6984
+ baseUrl: "https://opencode.ai/zen/go/v1",
6624
6985
  reasoning: true,
6625
6986
  input: ["text"],
6626
6987
  cost: {
@@ -6630,7 +6991,7 @@ export const MODELS = {
6630
6991
  cacheWrite: 0,
6631
6992
  },
6632
6993
  contextWindow: 204800,
6633
- maxTokens: 131072,
6994
+ maxTokens: 65536,
6634
6995
  },
6635
6996
  "minimax-m2.7": {
6636
6997
  id: "minimax-m2.7",
@@ -6649,6 +7010,40 @@ export const MODELS = {
6649
7010
  contextWindow: 204800,
6650
7011
  maxTokens: 131072,
6651
7012
  },
7013
+ "qwen3.5-plus": {
7014
+ id: "qwen3.5-plus",
7015
+ name: "Qwen3.5 Plus",
7016
+ api: "anthropic-messages",
7017
+ provider: "opencode-go",
7018
+ baseUrl: "https://opencode.ai/zen/go",
7019
+ reasoning: true,
7020
+ input: ["text", "image"],
7021
+ cost: {
7022
+ input: 0.2,
7023
+ output: 1.2,
7024
+ cacheRead: 0.02,
7025
+ cacheWrite: 0.25,
7026
+ },
7027
+ contextWindow: 262144,
7028
+ maxTokens: 65536,
7029
+ },
7030
+ "qwen3.6-plus": {
7031
+ id: "qwen3.6-plus",
7032
+ name: "Qwen3.6 Plus",
7033
+ api: "anthropic-messages",
7034
+ provider: "opencode-go",
7035
+ baseUrl: "https://opencode.ai/zen/go",
7036
+ reasoning: true,
7037
+ input: ["text", "image"],
7038
+ cost: {
7039
+ input: 0.5,
7040
+ output: 3,
7041
+ cacheRead: 0.05,
7042
+ cacheWrite: 0.625,
7043
+ },
7044
+ contextWindow: 262144,
7045
+ maxTokens: 65536,
7046
+ },
6652
7047
  },
6653
7048
  "openrouter": {
6654
7049
  "ai21/jamba-large-1.7": {
@@ -6821,23 +7216,6 @@ export const MODELS = {
6821
7216
  contextWindow: 200000,
6822
7217
  maxTokens: 8192,
6823
7218
  },
6824
- "anthropic/claude-3.5-sonnet": {
6825
- id: "anthropic/claude-3.5-sonnet",
6826
- name: "Anthropic: Claude 3.5 Sonnet",
6827
- api: "openai-completions",
6828
- provider: "openrouter",
6829
- baseUrl: "https://openrouter.ai/api/v1",
6830
- reasoning: false,
6831
- input: ["text", "image"],
6832
- cost: {
6833
- input: 6,
6834
- output: 30,
6835
- cacheRead: 0.6,
6836
- cacheWrite: 7.5,
6837
- },
6838
- contextWindow: 200000,
6839
- maxTokens: 8192,
6840
- },
6841
7219
  "anthropic/claude-3.7-sonnet": {
6842
7220
  id: "anthropic/claude-3.7-sonnet",
6843
7221
  name: "Anthropic: Claude 3.7 Sonnet",
@@ -6853,7 +7231,7 @@ export const MODELS = {
6853
7231
  cacheWrite: 3.75,
6854
7232
  },
6855
7233
  contextWindow: 200000,
6856
- maxTokens: 64000,
7234
+ maxTokens: 128000,
6857
7235
  },
6858
7236
  "anthropic/claude-3.7-sonnet:thinking": {
6859
7237
  id: "anthropic/claude-3.7-sonnet:thinking",
@@ -6957,6 +7335,40 @@ export const MODELS = {
6957
7335
  contextWindow: 1000000,
6958
7336
  maxTokens: 128000,
6959
7337
  },
7338
+ "anthropic/claude-opus-4.6-fast": {
7339
+ id: "anthropic/claude-opus-4.6-fast",
7340
+ name: "Anthropic: Claude Opus 4.6 (Fast)",
7341
+ api: "openai-completions",
7342
+ provider: "openrouter",
7343
+ baseUrl: "https://openrouter.ai/api/v1",
7344
+ reasoning: true,
7345
+ input: ["text", "image"],
7346
+ cost: {
7347
+ input: 30,
7348
+ output: 150,
7349
+ cacheRead: 3,
7350
+ cacheWrite: 37.5,
7351
+ },
7352
+ contextWindow: 1000000,
7353
+ maxTokens: 128000,
7354
+ },
7355
+ "anthropic/claude-opus-4.7": {
7356
+ id: "anthropic/claude-opus-4.7",
7357
+ name: "Anthropic: Claude Opus 4.7",
7358
+ api: "openai-completions",
7359
+ provider: "openrouter",
7360
+ baseUrl: "https://openrouter.ai/api/v1",
7361
+ reasoning: true,
7362
+ input: ["text", "image"],
7363
+ cost: {
7364
+ input: 5,
7365
+ output: 25,
7366
+ cacheRead: 0.5,
7367
+ cacheWrite: 6.25,
7368
+ },
7369
+ contextWindow: 1000000,
7370
+ maxTokens: 128000,
7371
+ },
6960
7372
  "anthropic/claude-sonnet-4": {
6961
7373
  id: "anthropic/claude-sonnet-4",
6962
7374
  name: "Anthropic: Claude Sonnet 4",
@@ -6971,7 +7383,7 @@ export const MODELS = {
6971
7383
  cacheRead: 0.3,
6972
7384
  cacheWrite: 3.75,
6973
7385
  },
6974
- contextWindow: 200000,
7386
+ contextWindow: 1000000,
6975
7387
  maxTokens: 64000,
6976
7388
  },
6977
7389
  "anthropic/claude-sonnet-4.5": {
@@ -7008,56 +7420,56 @@ export const MODELS = {
7008
7420
  contextWindow: 1000000,
7009
7421
  maxTokens: 128000,
7010
7422
  },
7011
- "arcee-ai/trinity-large-preview:free": {
7012
- id: "arcee-ai/trinity-large-preview:free",
7013
- name: "Arcee AI: Trinity Large Preview (free)",
7423
+ "arcee-ai/trinity-large-preview": {
7424
+ id: "arcee-ai/trinity-large-preview",
7425
+ name: "Arcee AI: Trinity Large Preview",
7014
7426
  api: "openai-completions",
7015
7427
  provider: "openrouter",
7016
7428
  baseUrl: "https://openrouter.ai/api/v1",
7017
7429
  reasoning: false,
7018
7430
  input: ["text"],
7019
7431
  cost: {
7020
- input: 0,
7021
- output: 0,
7432
+ input: 0.15,
7433
+ output: 0.44999999999999996,
7022
7434
  cacheRead: 0,
7023
7435
  cacheWrite: 0,
7024
7436
  },
7025
7437
  contextWindow: 131000,
7026
7438
  maxTokens: 4096,
7027
7439
  },
7028
- "arcee-ai/trinity-mini": {
7029
- id: "arcee-ai/trinity-mini",
7030
- name: "Arcee AI: Trinity Mini",
7440
+ "arcee-ai/trinity-large-thinking": {
7441
+ id: "arcee-ai/trinity-large-thinking",
7442
+ name: "Arcee AI: Trinity Large Thinking",
7031
7443
  api: "openai-completions",
7032
7444
  provider: "openrouter",
7033
7445
  baseUrl: "https://openrouter.ai/api/v1",
7034
7446
  reasoning: true,
7035
7447
  input: ["text"],
7036
7448
  cost: {
7037
- input: 0.045,
7038
- output: 0.15,
7039
- cacheRead: 0,
7449
+ input: 0.22,
7450
+ output: 0.85,
7451
+ cacheRead: 0.06,
7040
7452
  cacheWrite: 0,
7041
7453
  },
7042
- contextWindow: 131072,
7043
- maxTokens: 131072,
7454
+ contextWindow: 262144,
7455
+ maxTokens: 262144,
7044
7456
  },
7045
- "arcee-ai/trinity-mini:free": {
7046
- id: "arcee-ai/trinity-mini:free",
7047
- name: "Arcee AI: Trinity Mini (free)",
7457
+ "arcee-ai/trinity-mini": {
7458
+ id: "arcee-ai/trinity-mini",
7459
+ name: "Arcee AI: Trinity Mini",
7048
7460
  api: "openai-completions",
7049
7461
  provider: "openrouter",
7050
7462
  baseUrl: "https://openrouter.ai/api/v1",
7051
7463
  reasoning: true,
7052
7464
  input: ["text"],
7053
7465
  cost: {
7054
- input: 0,
7055
- output: 0,
7466
+ input: 0.045,
7467
+ output: 0.15,
7056
7468
  cacheRead: 0,
7057
7469
  cacheWrite: 0,
7058
7470
  },
7059
7471
  contextWindow: 131072,
7060
- maxTokens: 4096,
7472
+ maxTokens: 131072,
7061
7473
  },
7062
7474
  "arcee-ai/virtuoso-large": {
7063
7475
  id: "arcee-ai/virtuoso-large",
@@ -7252,7 +7664,7 @@ export const MODELS = {
7252
7664
  api: "openai-completions",
7253
7665
  provider: "openrouter",
7254
7666
  baseUrl: "https://openrouter.ai/api/v1",
7255
- reasoning: true,
7667
+ reasoning: false,
7256
7668
  input: ["text"],
7257
7669
  cost: {
7258
7670
  input: 0.19999999999999998,
@@ -7306,13 +7718,13 @@ export const MODELS = {
7306
7718
  reasoning: true,
7307
7719
  input: ["text"],
7308
7720
  cost: {
7309
- input: 0.44999999999999996,
7721
+ input: 0.5,
7310
7722
  output: 2.1500000000000004,
7311
- cacheRead: 0.22499999999999998,
7723
+ cacheRead: 0.35,
7312
7724
  cacheWrite: 0,
7313
7725
  },
7314
7726
  contextWindow: 163840,
7315
- maxTokens: 65536,
7727
+ maxTokens: 4096,
7316
7728
  },
7317
7729
  "deepseek/deepseek-v3.1-terminus": {
7318
7730
  id: "deepseek/deepseek-v3.1-terminus",
@@ -7340,13 +7752,13 @@ export const MODELS = {
7340
7752
  reasoning: true,
7341
7753
  input: ["text"],
7342
7754
  cost: {
7343
- input: 0.26,
7344
- output: 0.38,
7345
- cacheRead: 0.13,
7755
+ input: 0.252,
7756
+ output: 0.378,
7757
+ cacheRead: 0.0252,
7346
7758
  cacheWrite: 0,
7347
7759
  },
7348
- contextWindow: 163840,
7349
- maxTokens: 4096,
7760
+ contextWindow: 131072,
7761
+ maxTokens: 65536,
7350
7762
  },
7351
7763
  "deepseek/deepseek-v3.2-exp": {
7352
7764
  id: "deepseek/deepseek-v3.2-exp",
@@ -7396,7 +7808,7 @@ export const MODELS = {
7396
7808
  cacheRead: 0.024999999999999998,
7397
7809
  cacheWrite: 0.08333333333333334,
7398
7810
  },
7399
- contextWindow: 1048576,
7811
+ contextWindow: 1000000,
7400
7812
  maxTokens: 8192,
7401
7813
  },
7402
7814
  "google/gemini-2.0-flash-lite-001": {
@@ -7465,7 +7877,7 @@ export const MODELS = {
7465
7877
  cacheWrite: 0.08333333333333334,
7466
7878
  },
7467
7879
  contextWindow: 1048576,
7468
- maxTokens: 65536,
7880
+ maxTokens: 65535,
7469
7881
  },
7470
7882
  "google/gemini-2.5-pro": {
7471
7883
  id: "google/gemini-2.5-pro",
@@ -7535,23 +7947,6 @@ export const MODELS = {
7535
7947
  contextWindow: 1048576,
7536
7948
  maxTokens: 65536,
7537
7949
  },
7538
- "google/gemini-3-pro-preview": {
7539
- id: "google/gemini-3-pro-preview",
7540
- name: "Google: Gemini 3 Pro Preview",
7541
- api: "openai-completions",
7542
- provider: "openrouter",
7543
- baseUrl: "https://openrouter.ai/api/v1",
7544
- reasoning: true,
7545
- input: ["text", "image"],
7546
- cost: {
7547
- input: 2,
7548
- output: 12,
7549
- cacheRead: 0.19999999999999998,
7550
- cacheWrite: 0.375,
7551
- },
7552
- contextWindow: 1048576,
7553
- maxTokens: 65536,
7554
- },
7555
7950
  "google/gemini-3.1-flash-lite-preview": {
7556
7951
  id: "google/gemini-3.1-flash-lite-preview",
7557
7952
  name: "Google: Gemini 3.1 Flash Lite Preview",
@@ -7603,22 +7998,73 @@ export const MODELS = {
7603
7998
  contextWindow: 1048576,
7604
7999
  maxTokens: 65536,
7605
8000
  },
7606
- "inception/mercury": {
7607
- id: "inception/mercury",
7608
- name: "Inception: Mercury",
8001
+ "google/gemma-4-26b-a4b-it": {
8002
+ id: "google/gemma-4-26b-a4b-it",
8003
+ name: "Google: Gemma 4 26B A4B ",
8004
+ api: "openai-completions",
8005
+ provider: "openrouter",
8006
+ baseUrl: "https://openrouter.ai/api/v1",
8007
+ reasoning: true,
8008
+ input: ["text", "image"],
8009
+ cost: {
8010
+ input: 0.06,
8011
+ output: 0.33,
8012
+ cacheRead: 0,
8013
+ cacheWrite: 0,
8014
+ },
8015
+ contextWindow: 262144,
8016
+ maxTokens: 4096,
8017
+ },
8018
+ "google/gemma-4-26b-a4b-it:free": {
8019
+ id: "google/gemma-4-26b-a4b-it:free",
8020
+ name: "Google: Gemma 4 26B A4B (free)",
8021
+ api: "openai-completions",
8022
+ provider: "openrouter",
8023
+ baseUrl: "https://openrouter.ai/api/v1",
8024
+ reasoning: true,
8025
+ input: ["text", "image"],
8026
+ cost: {
8027
+ input: 0,
8028
+ output: 0,
8029
+ cacheRead: 0,
8030
+ cacheWrite: 0,
8031
+ },
8032
+ contextWindow: 262144,
8033
+ maxTokens: 32768,
8034
+ },
8035
+ "google/gemma-4-31b-it": {
8036
+ id: "google/gemma-4-31b-it",
8037
+ name: "Google: Gemma 4 31B",
8038
+ api: "openai-completions",
8039
+ provider: "openrouter",
8040
+ baseUrl: "https://openrouter.ai/api/v1",
8041
+ reasoning: true,
8042
+ input: ["text", "image"],
8043
+ cost: {
8044
+ input: 0.13,
8045
+ output: 0.38,
8046
+ cacheRead: 0,
8047
+ cacheWrite: 0,
8048
+ },
8049
+ contextWindow: 262144,
8050
+ maxTokens: 4096,
8051
+ },
8052
+ "google/gemma-4-31b-it:free": {
8053
+ id: "google/gemma-4-31b-it:free",
8054
+ name: "Google: Gemma 4 31B (free)",
7609
8055
  api: "openai-completions",
7610
8056
  provider: "openrouter",
7611
8057
  baseUrl: "https://openrouter.ai/api/v1",
7612
- reasoning: false,
7613
- input: ["text"],
8058
+ reasoning: true,
8059
+ input: ["text", "image"],
7614
8060
  cost: {
7615
- input: 0.25,
7616
- output: 0.75,
7617
- cacheRead: 0.024999999999999998,
8061
+ input: 0,
8062
+ output: 0,
8063
+ cacheRead: 0,
7618
8064
  cacheWrite: 0,
7619
8065
  },
7620
- contextWindow: 128000,
7621
- maxTokens: 32000,
8066
+ contextWindow: 262144,
8067
+ maxTokens: 32768,
7622
8068
  },
7623
8069
  "inception/mercury-2": {
7624
8070
  id: "inception/mercury-2",
@@ -7637,56 +8083,39 @@ export const MODELS = {
7637
8083
  contextWindow: 128000,
7638
8084
  maxTokens: 50000,
7639
8085
  },
7640
- "inception/mercury-coder": {
7641
- id: "inception/mercury-coder",
7642
- name: "Inception: Mercury Coder",
8086
+ "inclusionai/ling-2.6-flash:free": {
8087
+ id: "inclusionai/ling-2.6-flash:free",
8088
+ name: "inclusionAI: Ling-2.6-flash (free)",
7643
8089
  api: "openai-completions",
7644
8090
  provider: "openrouter",
7645
8091
  baseUrl: "https://openrouter.ai/api/v1",
7646
8092
  reasoning: false,
7647
8093
  input: ["text"],
7648
8094
  cost: {
7649
- input: 0.25,
7650
- output: 0.75,
7651
- cacheRead: 0.024999999999999998,
8095
+ input: 0,
8096
+ output: 0,
8097
+ cacheRead: 0,
7652
8098
  cacheWrite: 0,
7653
8099
  },
7654
- contextWindow: 128000,
7655
- maxTokens: 32000,
8100
+ contextWindow: 262144,
8101
+ maxTokens: 32768,
7656
8102
  },
7657
- "kwaipilot/kat-coder-pro": {
7658
- id: "kwaipilot/kat-coder-pro",
7659
- name: "Kwaipilot: KAT-Coder-Pro V1",
8103
+ "kwaipilot/kat-coder-pro-v2": {
8104
+ id: "kwaipilot/kat-coder-pro-v2",
8105
+ name: "Kwaipilot: KAT-Coder-Pro V2",
7660
8106
  api: "openai-completions",
7661
8107
  provider: "openrouter",
7662
8108
  baseUrl: "https://openrouter.ai/api/v1",
7663
8109
  reasoning: false,
7664
8110
  input: ["text"],
7665
8111
  cost: {
7666
- input: 0.207,
7667
- output: 0.828,
7668
- cacheRead: 0.0414,
8112
+ input: 0.3,
8113
+ output: 1.2,
8114
+ cacheRead: 0.06,
7669
8115
  cacheWrite: 0,
7670
8116
  },
7671
8117
  contextWindow: 256000,
7672
- maxTokens: 128000,
7673
- },
7674
- "meituan/longcat-flash-chat": {
7675
- id: "meituan/longcat-flash-chat",
7676
- name: "Meituan: LongCat Flash Chat",
7677
- api: "openai-completions",
7678
- provider: "openrouter",
7679
- baseUrl: "https://openrouter.ai/api/v1",
7680
- reasoning: false,
7681
- input: ["text"],
7682
- cost: {
7683
- input: 0.19999999999999998,
7684
- output: 0.7999999999999999,
7685
- cacheRead: 0.19999999999999998,
7686
- cacheWrite: 0,
7687
- },
7688
- contextWindow: 131072,
7689
- maxTokens: 131072,
8118
+ maxTokens: 80000,
7690
8119
  },
7691
8120
  "meta-llama/llama-3-8b-instruct": {
7692
8121
  id: "meta-llama/llama-3-8b-instruct",
@@ -7773,23 +8202,6 @@ export const MODELS = {
7773
8202
  contextWindow: 65536,
7774
8203
  maxTokens: 4096,
7775
8204
  },
7776
- "meta-llama/llama-4-maverick": {
7777
- id: "meta-llama/llama-4-maverick",
7778
- name: "Meta: Llama 4 Maverick",
7779
- api: "openai-completions",
7780
- provider: "openrouter",
7781
- baseUrl: "https://openrouter.ai/api/v1",
7782
- reasoning: false,
7783
- input: ["text", "image"],
7784
- cost: {
7785
- input: 0.15,
7786
- output: 0.6,
7787
- cacheRead: 0,
7788
- cacheWrite: 0,
7789
- },
7790
- contextWindow: 1048576,
7791
- maxTokens: 16384,
7792
- },
7793
8205
  "meta-llama/llama-4-scout": {
7794
8206
  id: "meta-llama/llama-4-scout",
7795
8207
  name: "Meta: Llama 4 Scout",
@@ -7850,13 +8262,13 @@ export const MODELS = {
7850
8262
  reasoning: true,
7851
8263
  input: ["text"],
7852
8264
  cost: {
7853
- input: 0.27,
8265
+ input: 0.29,
7854
8266
  output: 0.95,
7855
- cacheRead: 0.0290000007,
8267
+ cacheRead: 0.03,
7856
8268
  cacheWrite: 0,
7857
8269
  },
7858
8270
  contextWindow: 196608,
7859
- maxTokens: 4096,
8271
+ maxTokens: 196608,
7860
8272
  },
7861
8273
  "minimax/minimax-m2.5": {
7862
8274
  id: "minimax/minimax-m2.5",
@@ -7867,9 +8279,9 @@ export const MODELS = {
7867
8279
  reasoning: true,
7868
8280
  input: ["text"],
7869
8281
  cost: {
7870
- input: 0.19999999999999998,
7871
- output: 1.17,
7872
- cacheRead: 0.09999999999999999,
8282
+ input: 0.15,
8283
+ output: 1.2,
8284
+ cacheRead: 0.075,
7873
8285
  cacheWrite: 0,
7874
8286
  },
7875
8287
  contextWindow: 196608,
@@ -7890,7 +8302,7 @@ export const MODELS = {
7890
8302
  cacheWrite: 0,
7891
8303
  },
7892
8304
  contextWindow: 196608,
7893
- maxTokens: 196608,
8305
+ maxTokens: 8192,
7894
8306
  },
7895
8307
  "minimax/minimax-m2.7": {
7896
8308
  id: "minimax/minimax-m2.7",
@@ -7903,11 +8315,11 @@ export const MODELS = {
7903
8315
  cost: {
7904
8316
  input: 0.3,
7905
8317
  output: 1.2,
7906
- cacheRead: 0.06,
8318
+ cacheRead: 0.059,
7907
8319
  cacheWrite: 0,
7908
8320
  },
7909
- contextWindow: 204800,
7910
- maxTokens: 131072,
8321
+ contextWindow: 196608,
8322
+ maxTokens: 4096,
7911
8323
  },
7912
8324
  "mistralai/codestral-2508": {
7913
8325
  id: "mistralai/codestral-2508",
@@ -8139,13 +8551,13 @@ export const MODELS = {
8139
8551
  reasoning: false,
8140
8552
  input: ["text"],
8141
8553
  cost: {
8142
- input: 0.02,
8143
- output: 0.04,
8554
+ input: 0.01,
8555
+ output: 0.03,
8144
8556
  cacheRead: 0,
8145
8557
  cacheWrite: 0,
8146
8558
  },
8147
8559
  contextWindow: 131072,
8148
- maxTokens: 16384,
8560
+ maxTokens: 4096,
8149
8561
  },
8150
8562
  "mistralai/mistral-saba": {
8151
8563
  id: "mistralai/mistral-saba",
@@ -8164,23 +8576,6 @@ export const MODELS = {
8164
8576
  contextWindow: 32768,
8165
8577
  maxTokens: 4096,
8166
8578
  },
8167
- "mistralai/mistral-small-24b-instruct-2501": {
8168
- id: "mistralai/mistral-small-24b-instruct-2501",
8169
- name: "Mistral: Mistral Small 3",
8170
- api: "openai-completions",
8171
- provider: "openrouter",
8172
- baseUrl: "https://openrouter.ai/api/v1",
8173
- reasoning: false,
8174
- input: ["text"],
8175
- cost: {
8176
- input: 0.049999999999999996,
8177
- output: 0.08,
8178
- cacheRead: 0,
8179
- cacheWrite: 0,
8180
- },
8181
- contextWindow: 32768,
8182
- maxTokens: 16384,
8183
- },
8184
8579
  "mistralai/mistral-small-2603": {
8185
8580
  id: "mistralai/mistral-small-2603",
8186
8581
  name: "Mistral: Mistral Small 4",
@@ -8198,23 +8593,6 @@ export const MODELS = {
8198
8593
  contextWindow: 262144,
8199
8594
  maxTokens: 4096,
8200
8595
  },
8201
- "mistralai/mistral-small-3.1-24b-instruct:free": {
8202
- id: "mistralai/mistral-small-3.1-24b-instruct:free",
8203
- name: "Mistral: Mistral Small 3.1 24B (free)",
8204
- api: "openai-completions",
8205
- provider: "openrouter",
8206
- baseUrl: "https://openrouter.ai/api/v1",
8207
- reasoning: false,
8208
- input: ["text", "image"],
8209
- cost: {
8210
- input: 0,
8211
- output: 0,
8212
- cacheRead: 0,
8213
- cacheWrite: 0,
8214
- },
8215
- contextWindow: 128000,
8216
- maxTokens: 4096,
8217
- },
8218
8596
  "mistralai/mistral-small-3.2-24b-instruct": {
8219
8597
  id: "mistralai/mistral-small-3.2-24b-instruct",
8220
8598
  name: "Mistral: Mistral Small 3.2 24B",
@@ -8332,7 +8710,7 @@ export const MODELS = {
8332
8710
  cacheWrite: 0,
8333
8711
  },
8334
8712
  contextWindow: 131072,
8335
- maxTokens: 131072,
8713
+ maxTokens: 32768,
8336
8714
  },
8337
8715
  "moonshotai/kimi-k2-0905": {
8338
8716
  id: "moonshotai/kimi-k2-0905",
@@ -8345,11 +8723,11 @@ export const MODELS = {
8345
8723
  cost: {
8346
8724
  input: 0.39999999999999997,
8347
8725
  output: 2,
8348
- cacheRead: 0.15,
8726
+ cacheRead: 0,
8349
8727
  cacheWrite: 0,
8350
8728
  },
8351
- contextWindow: 131072,
8352
- maxTokens: 4096,
8729
+ contextWindow: 262144,
8730
+ maxTokens: 262144,
8353
8731
  },
8354
8732
  "moonshotai/kimi-k2-thinking": {
8355
8733
  id: "moonshotai/kimi-k2-thinking",
@@ -8360,13 +8738,13 @@ export const MODELS = {
8360
8738
  reasoning: true,
8361
8739
  input: ["text"],
8362
8740
  cost: {
8363
- input: 0.47,
8364
- output: 2,
8365
- cacheRead: 0.14100000000000001,
8741
+ input: 0.6,
8742
+ output: 2.5,
8743
+ cacheRead: 0.15,
8366
8744
  cacheWrite: 0,
8367
8745
  },
8368
- contextWindow: 131072,
8369
- maxTokens: 4096,
8746
+ contextWindow: 262144,
8747
+ maxTokens: 262144,
8370
8748
  },
8371
8749
  "moonshotai/kimi-k2.5": {
8372
8750
  id: "moonshotai/kimi-k2.5",
@@ -8385,6 +8763,23 @@ export const MODELS = {
8385
8763
  contextWindow: 262144,
8386
8764
  maxTokens: 4096,
8387
8765
  },
8766
+ "moonshotai/kimi-k2.6": {
8767
+ id: "moonshotai/kimi-k2.6",
8768
+ name: "MoonshotAI: Kimi K2.6",
8769
+ api: "openai-completions",
8770
+ provider: "openrouter",
8771
+ baseUrl: "https://openrouter.ai/api/v1",
8772
+ reasoning: true,
8773
+ input: ["text", "image"],
8774
+ cost: {
8775
+ input: 0.56,
8776
+ output: 3.5,
8777
+ cacheRead: 0.11,
8778
+ cacheWrite: 0,
8779
+ },
8780
+ contextWindow: 256000,
8781
+ maxTokens: 65536,
8782
+ },
8388
8783
  "nex-agi/deepseek-v3.1-nex-n1": {
8389
8784
  id: "nex-agi/deepseek-v3.1-nex-n1",
8390
8785
  name: "Nex AGI: DeepSeek V3.1 Nex N1",
@@ -8479,9 +8874,9 @@ export const MODELS = {
8479
8874
  reasoning: true,
8480
8875
  input: ["text"],
8481
8876
  cost: {
8482
- input: 0.09999999999999999,
8483
- output: 0.5,
8484
- cacheRead: 0.09999999999999999,
8877
+ input: 0.09,
8878
+ output: 0.44999999999999996,
8879
+ cacheRead: 0,
8485
8880
  cacheWrite: 0,
8486
8881
  },
8487
8882
  contextWindow: 262144,
@@ -8706,7 +9101,7 @@ export const MODELS = {
8706
9101
  cacheWrite: 0,
8707
9102
  },
8708
9103
  contextWindow: 1047576,
8709
- maxTokens: 32768,
9104
+ maxTokens: 4096,
8710
9105
  },
8711
9106
  "openai/gpt-4.1-mini": {
8712
9107
  id: "openai/gpt-4.1-mini",
@@ -8753,7 +9148,7 @@ export const MODELS = {
8753
9148
  cost: {
8754
9149
  input: 2.5,
8755
9150
  output: 10,
8756
- cacheRead: 1.25,
9151
+ cacheRead: 0,
8757
9152
  cacheWrite: 0,
8758
9153
  },
8759
9154
  contextWindow: 128000,
@@ -8861,23 +9256,6 @@ export const MODELS = {
8861
9256
  contextWindow: 128000,
8862
9257
  maxTokens: 16384,
8863
9258
  },
8864
- "openai/gpt-4o:extended": {
8865
- id: "openai/gpt-4o:extended",
8866
- name: "OpenAI: GPT-4o (extended)",
8867
- api: "openai-completions",
8868
- provider: "openrouter",
8869
- baseUrl: "https://openrouter.ai/api/v1",
8870
- reasoning: false,
8871
- input: ["text", "image"],
8872
- cost: {
8873
- input: 6,
8874
- output: 18,
8875
- cacheRead: 0,
8876
- cacheWrite: 0,
8877
- },
8878
- contextWindow: 128000,
8879
- maxTokens: 64000,
8880
- },
8881
9259
  "openai/gpt-5": {
8882
9260
  id: "openai/gpt-5",
8883
9261
  name: "OpenAI: GPT-5",
@@ -8912,40 +9290,6 @@ export const MODELS = {
8912
9290
  contextWindow: 400000,
8913
9291
  maxTokens: 128000,
8914
9292
  },
8915
- "openai/gpt-5-image": {
8916
- id: "openai/gpt-5-image",
8917
- name: "OpenAI: GPT-5 Image",
8918
- api: "openai-completions",
8919
- provider: "openrouter",
8920
- baseUrl: "https://openrouter.ai/api/v1",
8921
- reasoning: true,
8922
- input: ["text", "image"],
8923
- cost: {
8924
- input: 10,
8925
- output: 10,
8926
- cacheRead: 1.25,
8927
- cacheWrite: 0,
8928
- },
8929
- contextWindow: 400000,
8930
- maxTokens: 128000,
8931
- },
8932
- "openai/gpt-5-image-mini": {
8933
- id: "openai/gpt-5-image-mini",
8934
- name: "OpenAI: GPT-5 Image Mini",
8935
- api: "openai-completions",
8936
- provider: "openrouter",
8937
- baseUrl: "https://openrouter.ai/api/v1",
8938
- reasoning: true,
8939
- input: ["text", "image"],
8940
- cost: {
8941
- input: 2.5,
8942
- output: 2,
8943
- cacheRead: 0.25,
8944
- cacheWrite: 0,
8945
- },
8946
- contextWindow: 400000,
8947
- maxTokens: 128000,
8948
- },
8949
9293
  "openai/gpt-5-mini": {
8950
9294
  id: "openai/gpt-5-mini",
8951
9295
  name: "OpenAI: GPT-5 Mini",
@@ -8974,11 +9318,11 @@ export const MODELS = {
8974
9318
  cost: {
8975
9319
  input: 0.049999999999999996,
8976
9320
  output: 0.39999999999999997,
8977
- cacheRead: 0.005,
9321
+ cacheRead: 0.01,
8978
9322
  cacheWrite: 0,
8979
9323
  },
8980
9324
  contextWindow: 400000,
8981
- maxTokens: 128000,
9325
+ maxTokens: 4096,
8982
9326
  },
8983
9327
  "openai/gpt-5-pro": {
8984
9328
  id: "openai/gpt-5-pro",
@@ -9008,7 +9352,7 @@ export const MODELS = {
9008
9352
  cost: {
9009
9353
  input: 1.25,
9010
9354
  output: 10,
9011
- cacheRead: 0.125,
9355
+ cacheRead: 0.13,
9012
9356
  cacheWrite: 0,
9013
9357
  },
9014
9358
  contextWindow: 400000,
@@ -9076,11 +9420,11 @@ export const MODELS = {
9076
9420
  cost: {
9077
9421
  input: 0.25,
9078
9422
  output: 2,
9079
- cacheRead: 0.024999999999999998,
9423
+ cacheRead: 0.03,
9080
9424
  cacheWrite: 0,
9081
9425
  },
9082
9426
  contextWindow: 400000,
9083
- maxTokens: 100000,
9427
+ maxTokens: 128000,
9084
9428
  },
9085
9429
  "openai/gpt-5.2": {
9086
9430
  id: "openai/gpt-5.2",
@@ -9114,7 +9458,7 @@ export const MODELS = {
9114
9458
  cacheWrite: 0,
9115
9459
  },
9116
9460
  contextWindow: 128000,
9117
- maxTokens: 16384,
9461
+ maxTokens: 32000,
9118
9462
  },
9119
9463
  "openai/gpt-5.2-codex": {
9120
9464
  id: "openai/gpt-5.2-codex",
@@ -9252,6 +9596,40 @@ export const MODELS = {
9252
9596
  contextWindow: 1050000,
9253
9597
  maxTokens: 128000,
9254
9598
  },
9599
+ "openai/gpt-audio": {
9600
+ id: "openai/gpt-audio",
9601
+ name: "OpenAI: GPT Audio",
9602
+ api: "openai-completions",
9603
+ provider: "openrouter",
9604
+ baseUrl: "https://openrouter.ai/api/v1",
9605
+ reasoning: false,
9606
+ input: ["text"],
9607
+ cost: {
9608
+ input: 2.5,
9609
+ output: 10,
9610
+ cacheRead: 0,
9611
+ cacheWrite: 0,
9612
+ },
9613
+ contextWindow: 128000,
9614
+ maxTokens: 16384,
9615
+ },
9616
+ "openai/gpt-audio-mini": {
9617
+ id: "openai/gpt-audio-mini",
9618
+ name: "OpenAI: GPT Audio Mini",
9619
+ api: "openai-completions",
9620
+ provider: "openrouter",
9621
+ baseUrl: "https://openrouter.ai/api/v1",
9622
+ reasoning: false,
9623
+ input: ["text"],
9624
+ cost: {
9625
+ input: 0.6,
9626
+ output: 2.4,
9627
+ cacheRead: 0,
9628
+ cacheWrite: 0,
9629
+ },
9630
+ contextWindow: 128000,
9631
+ maxTokens: 16384,
9632
+ },
9255
9633
  "openai/gpt-oss-120b": {
9256
9634
  id: "openai/gpt-oss-120b",
9257
9635
  name: "OpenAI: gpt-oss-120b",
@@ -9296,12 +9674,12 @@ export const MODELS = {
9296
9674
  input: ["text"],
9297
9675
  cost: {
9298
9676
  input: 0.03,
9299
- output: 0.11,
9300
- cacheRead: 0.015,
9677
+ output: 0.14,
9678
+ cacheRead: 0,
9301
9679
  cacheWrite: 0,
9302
9680
  },
9303
9681
  contextWindow: 131072,
9304
- maxTokens: 131072,
9682
+ maxTokens: 4096,
9305
9683
  },
9306
9684
  "openai/gpt-oss-20b:free": {
9307
9685
  id: "openai/gpt-oss-20b:free",
@@ -9318,7 +9696,7 @@ export const MODELS = {
9318
9696
  cacheWrite: 0,
9319
9697
  },
9320
9698
  contextWindow: 131072,
9321
- maxTokens: 131072,
9699
+ maxTokens: 8192,
9322
9700
  },
9323
9701
  "openai/gpt-oss-safeguard-20b": {
9324
9702
  id: "openai/gpt-oss-safeguard-20b",
@@ -9604,7 +9982,7 @@ export const MODELS = {
9604
9982
  input: 0.26,
9605
9983
  output: 0.78,
9606
9984
  cacheRead: 0.052000000000000005,
9607
- cacheWrite: 0,
9985
+ cacheWrite: 0.325,
9608
9986
  },
9609
9987
  contextWindow: 1000000,
9610
9988
  maxTokens: 32768,
@@ -9621,7 +9999,7 @@ export const MODELS = {
9621
9999
  input: 0.26,
9622
10000
  output: 0.78,
9623
10001
  cacheRead: 0,
9624
- cacheWrite: 0,
10002
+ cacheWrite: 0.325,
9625
10003
  },
9626
10004
  contextWindow: 1000000,
9627
10005
  maxTokens: 32768,
@@ -9638,7 +10016,7 @@ export const MODELS = {
9638
10016
  input: 0.26,
9639
10017
  output: 0.78,
9640
10018
  cacheRead: 0,
9641
- cacheWrite: 0,
10019
+ cacheWrite: 0.325,
9642
10020
  },
9643
10021
  contextWindow: 1000000,
9644
10022
  maxTokens: 32768,
@@ -9717,7 +10095,7 @@ export const MODELS = {
9717
10095
  api: "openai-completions",
9718
10096
  provider: "openrouter",
9719
10097
  baseUrl: "https://openrouter.ai/api/v1",
9720
- reasoning: true,
10098
+ reasoning: false,
9721
10099
  input: ["text"],
9722
10100
  cost: {
9723
10101
  input: 0.071,
@@ -9793,42 +10171,25 @@ export const MODELS = {
9793
10171
  cacheRead: 0.08,
9794
10172
  cacheWrite: 0,
9795
10173
  },
9796
- contextWindow: 131072,
9797
- maxTokens: 131072,
9798
- },
9799
- "qwen/qwen3-32b": {
9800
- id: "qwen/qwen3-32b",
9801
- name: "Qwen: Qwen3 32B",
9802
- api: "openai-completions",
9803
- provider: "openrouter",
9804
- baseUrl: "https://openrouter.ai/api/v1",
9805
- reasoning: true,
9806
- input: ["text"],
9807
- cost: {
9808
- input: 0.08,
9809
- output: 0.24,
9810
- cacheRead: 0.04,
9811
- cacheWrite: 0,
9812
- },
9813
- contextWindow: 40960,
9814
- maxTokens: 40960,
10174
+ contextWindow: 131072,
10175
+ maxTokens: 131072,
9815
10176
  },
9816
- "qwen/qwen3-4b:free": {
9817
- id: "qwen/qwen3-4b:free",
9818
- name: "Qwen: Qwen3 4B (free)",
10177
+ "qwen/qwen3-32b": {
10178
+ id: "qwen/qwen3-32b",
10179
+ name: "Qwen: Qwen3 32B",
9819
10180
  api: "openai-completions",
9820
10181
  provider: "openrouter",
9821
10182
  baseUrl: "https://openrouter.ai/api/v1",
9822
10183
  reasoning: true,
9823
10184
  input: ["text"],
9824
10185
  cost: {
9825
- input: 0,
9826
- output: 0,
9827
- cacheRead: 0,
10186
+ input: 0.08,
10187
+ output: 0.24,
10188
+ cacheRead: 0.04,
9828
10189
  cacheWrite: 0,
9829
10190
  },
9830
10191
  contextWindow: 40960,
9831
- maxTokens: 4096,
10192
+ maxTokens: 40960,
9832
10193
  },
9833
10194
  "qwen/qwen3-8b": {
9834
10195
  id: "qwen/qwen3-8b",
@@ -9893,7 +10254,7 @@ export const MODELS = {
9893
10254
  input: 0.195,
9894
10255
  output: 0.975,
9895
10256
  cacheRead: 0.039,
9896
- cacheWrite: 0,
10257
+ cacheWrite: 0.24375,
9897
10258
  },
9898
10259
  contextWindow: 1000000,
9899
10260
  maxTokens: 65536,
@@ -9907,13 +10268,13 @@ export const MODELS = {
9907
10268
  reasoning: false,
9908
10269
  input: ["text"],
9909
10270
  cost: {
9910
- input: 0.12,
9911
- output: 0.75,
9912
- cacheRead: 0.06,
10271
+ input: 0.15,
10272
+ output: 0.7999999999999999,
10273
+ cacheRead: 0.11,
9913
10274
  cacheWrite: 0,
9914
10275
  },
9915
10276
  contextWindow: 262144,
9916
- maxTokens: 65536,
10277
+ maxTokens: 262144,
9917
10278
  },
9918
10279
  "qwen/qwen3-coder-plus": {
9919
10280
  id: "qwen/qwen3-coder-plus",
@@ -9927,7 +10288,7 @@ export const MODELS = {
9927
10288
  input: 0.65,
9928
10289
  output: 3.25,
9929
10290
  cacheRead: 0.13,
9930
- cacheWrite: 0,
10291
+ cacheWrite: 0.8125,
9931
10292
  },
9932
10293
  contextWindow: 1000000,
9933
10294
  maxTokens: 65536,
@@ -9961,7 +10322,7 @@ export const MODELS = {
9961
10322
  input: 0.78,
9962
10323
  output: 3.9,
9963
10324
  cacheRead: 0.156,
9964
- cacheWrite: 0,
10325
+ cacheWrite: 0.975,
9965
10326
  },
9966
10327
  contextWindow: 262144,
9967
10328
  maxTokens: 32768,
@@ -10215,7 +10576,7 @@ export const MODELS = {
10215
10576
  cost: {
10216
10577
  input: 0.39,
10217
10578
  output: 2.34,
10218
- cacheRead: 0,
10579
+ cacheRead: 0.195,
10219
10580
  cacheWrite: 0,
10220
10581
  },
10221
10582
  contextWindow: 262144,
@@ -10230,13 +10591,13 @@ export const MODELS = {
10230
10591
  reasoning: true,
10231
10592
  input: ["text", "image"],
10232
10593
  cost: {
10233
- input: 0.049999999999999996,
10594
+ input: 0.09999999999999999,
10234
10595
  output: 0.15,
10235
10596
  cacheRead: 0,
10236
10597
  cacheWrite: 0,
10237
10598
  },
10238
- contextWindow: 256000,
10239
- maxTokens: 65536,
10599
+ contextWindow: 262144,
10600
+ maxTokens: 4096,
10240
10601
  },
10241
10602
  "qwen/qwen3.5-flash-02-23": {
10242
10603
  id: "qwen/qwen3.5-flash-02-23",
@@ -10250,7 +10611,7 @@ export const MODELS = {
10250
10611
  input: 0.065,
10251
10612
  output: 0.26,
10252
10613
  cacheRead: 0,
10253
- cacheWrite: 0,
10614
+ cacheWrite: 0.08125,
10254
10615
  },
10255
10616
  contextWindow: 1000000,
10256
10617
  maxTokens: 65536,
@@ -10267,7 +10628,24 @@ export const MODELS = {
10267
10628
  input: 0.26,
10268
10629
  output: 1.56,
10269
10630
  cacheRead: 0,
10270
- cacheWrite: 0,
10631
+ cacheWrite: 0.325,
10632
+ },
10633
+ contextWindow: 1000000,
10634
+ maxTokens: 65536,
10635
+ },
10636
+ "qwen/qwen3.6-plus": {
10637
+ id: "qwen/qwen3.6-plus",
10638
+ name: "Qwen: Qwen3.6 Plus",
10639
+ api: "openai-completions",
10640
+ provider: "openrouter",
10641
+ baseUrl: "https://openrouter.ai/api/v1",
10642
+ reasoning: true,
10643
+ input: ["text", "image"],
10644
+ cost: {
10645
+ input: 0.325,
10646
+ output: 1.95,
10647
+ cacheRead: 0,
10648
+ cacheWrite: 0.40625,
10271
10649
  },
10272
10650
  contextWindow: 1000000,
10273
10651
  maxTokens: 65536,
@@ -10289,6 +10667,23 @@ export const MODELS = {
10289
10667
  contextWindow: 131072,
10290
10668
  maxTokens: 131072,
10291
10669
  },
10670
+ "rekaai/reka-edge": {
10671
+ id: "rekaai/reka-edge",
10672
+ name: "Reka Edge",
10673
+ api: "openai-completions",
10674
+ provider: "openrouter",
10675
+ baseUrl: "https://openrouter.ai/api/v1",
10676
+ reasoning: false,
10677
+ input: ["text", "image"],
10678
+ cost: {
10679
+ input: 0.09999999999999999,
10680
+ output: 0.09999999999999999,
10681
+ cacheRead: 0,
10682
+ cacheWrite: 0,
10683
+ },
10684
+ contextWindow: 16384,
10685
+ maxTokens: 16384,
10686
+ },
10292
10687
  "relace/relace-search": {
10293
10688
  id: "relace/relace-search",
10294
10689
  name: "Relace: Relace Search",
@@ -10351,15 +10746,15 @@ export const MODELS = {
10351
10746
  cost: {
10352
10747
  input: 0.09999999999999999,
10353
10748
  output: 0.3,
10354
- cacheRead: 0.02,
10749
+ cacheRead: 0,
10355
10750
  cacheWrite: 0,
10356
10751
  },
10357
- contextWindow: 256000,
10358
- maxTokens: 256000,
10752
+ contextWindow: 262144,
10753
+ maxTokens: 65536,
10359
10754
  },
10360
- "stepfun/step-3.5-flash:free": {
10361
- id: "stepfun/step-3.5-flash:free",
10362
- name: "StepFun: Step 3.5 Flash (free)",
10755
+ "tencent/hy3-preview:free": {
10756
+ id: "tencent/hy3-preview:free",
10757
+ name: "Tencent: Hy3 preview (free)",
10363
10758
  api: "openai-completions",
10364
10759
  provider: "openrouter",
10365
10760
  baseUrl: "https://openrouter.ai/api/v1",
@@ -10371,8 +10766,8 @@ export const MODELS = {
10371
10766
  cacheRead: 0,
10372
10767
  cacheWrite: 0,
10373
10768
  },
10374
- contextWindow: 256000,
10375
- maxTokens: 256000,
10769
+ contextWindow: 262144,
10770
+ maxTokens: 262144,
10376
10771
  },
10377
10772
  "thedrummer/rocinante-12b": {
10378
10773
  id: "thedrummer/rocinante-12b",
@@ -10561,9 +10956,9 @@ export const MODELS = {
10561
10956
  contextWindow: 2000000,
10562
10957
  maxTokens: 30000,
10563
10958
  },
10564
- "x-ai/grok-4.20-beta": {
10565
- id: "x-ai/grok-4.20-beta",
10566
- name: "xAI: Grok 4.20 Beta",
10959
+ "x-ai/grok-4.20": {
10960
+ id: "x-ai/grok-4.20",
10961
+ name: "xAI: Grok 4.20",
10567
10962
  api: "openai-completions",
10568
10963
  provider: "openrouter",
10569
10964
  baseUrl: "https://openrouter.ai/api/v1",
@@ -10646,6 +11041,40 @@ export const MODELS = {
10646
11041
  contextWindow: 1048576,
10647
11042
  maxTokens: 131072,
10648
11043
  },
11044
+ "xiaomi/mimo-v2.5": {
11045
+ id: "xiaomi/mimo-v2.5",
11046
+ name: "Xiaomi: MiMo-V2.5",
11047
+ api: "openai-completions",
11048
+ provider: "openrouter",
11049
+ baseUrl: "https://openrouter.ai/api/v1",
11050
+ reasoning: true,
11051
+ input: ["text", "image"],
11052
+ cost: {
11053
+ input: 0.39999999999999997,
11054
+ output: 2,
11055
+ cacheRead: 0.08,
11056
+ cacheWrite: 0,
11057
+ },
11058
+ contextWindow: 1048576,
11059
+ maxTokens: 131072,
11060
+ },
11061
+ "xiaomi/mimo-v2.5-pro": {
11062
+ id: "xiaomi/mimo-v2.5-pro",
11063
+ name: "Xiaomi: MiMo-V2.5-Pro",
11064
+ api: "openai-completions",
11065
+ provider: "openrouter",
11066
+ baseUrl: "https://openrouter.ai/api/v1",
11067
+ reasoning: true,
11068
+ input: ["text"],
11069
+ cost: {
11070
+ input: 1,
11071
+ output: 3,
11072
+ cacheRead: 0.19999999999999998,
11073
+ cacheWrite: 0,
11074
+ },
11075
+ contextWindow: 1048576,
11076
+ maxTokens: 131072,
11077
+ },
10649
11078
  "z-ai/glm-4-32b": {
10650
11079
  id: "z-ai/glm-4-32b",
10651
11080
  name: "Z.ai: GLM 4 32B ",
@@ -10774,13 +11203,13 @@ export const MODELS = {
10774
11203
  reasoning: true,
10775
11204
  input: ["text"],
10776
11205
  cost: {
10777
- input: 0.39,
10778
- output: 1.75,
10779
- cacheRead: 0.195,
11206
+ input: 0.38,
11207
+ output: 1.74,
11208
+ cacheRead: 0,
10780
11209
  cacheWrite: 0,
10781
11210
  },
10782
11211
  contextWindow: 202752,
10783
- maxTokens: 65535,
11212
+ maxTokens: 4096,
10784
11213
  },
10785
11214
  "z-ai/glm-4.7-flash": {
10786
11215
  id: "z-ai/glm-4.7-flash",
@@ -10813,8 +11242,8 @@ export const MODELS = {
10813
11242
  cacheRead: 0.119,
10814
11243
  cacheWrite: 0,
10815
11244
  },
10816
- contextWindow: 80000,
10817
- maxTokens: 131072,
11245
+ contextWindow: 202752,
11246
+ maxTokens: 4096,
10818
11247
  },
10819
11248
  "z-ai/glm-5-turbo": {
10820
11249
  id: "z-ai/glm-5-turbo",
@@ -10833,6 +11262,57 @@ export const MODELS = {
10833
11262
  contextWindow: 202752,
10834
11263
  maxTokens: 131072,
10835
11264
  },
11265
+ "z-ai/glm-5.1": {
11266
+ id: "z-ai/glm-5.1",
11267
+ name: "Z.ai: GLM 5.1",
11268
+ api: "openai-completions",
11269
+ provider: "openrouter",
11270
+ baseUrl: "https://openrouter.ai/api/v1",
11271
+ reasoning: true,
11272
+ input: ["text"],
11273
+ cost: {
11274
+ input: 1.0499999999999998,
11275
+ output: 3.5,
11276
+ cacheRead: 0.5249999999999999,
11277
+ cacheWrite: 0,
11278
+ },
11279
+ contextWindow: 202752,
11280
+ maxTokens: 65535,
11281
+ },
11282
+ "z-ai/glm-5v-turbo": {
11283
+ id: "z-ai/glm-5v-turbo",
11284
+ name: "Z.ai: GLM 5V Turbo",
11285
+ api: "openai-completions",
11286
+ provider: "openrouter",
11287
+ baseUrl: "https://openrouter.ai/api/v1",
11288
+ reasoning: true,
11289
+ input: ["text", "image"],
11290
+ cost: {
11291
+ input: 1.2,
11292
+ output: 4,
11293
+ cacheRead: 0.24,
11294
+ cacheWrite: 0,
11295
+ },
11296
+ contextWindow: 202752,
11297
+ maxTokens: 131072,
11298
+ },
11299
+ "~anthropic/claude-opus-latest": {
11300
+ id: "~anthropic/claude-opus-latest",
11301
+ name: "Anthropic: Claude Opus Latest",
11302
+ api: "openai-completions",
11303
+ provider: "openrouter",
11304
+ baseUrl: "https://openrouter.ai/api/v1",
11305
+ reasoning: true,
11306
+ input: ["text", "image"],
11307
+ cost: {
11308
+ input: 5,
11309
+ output: 25,
11310
+ cacheRead: 0.5,
11311
+ cacheWrite: 6.25,
11312
+ },
11313
+ contextWindow: 1000000,
11314
+ maxTokens: 128000,
11315
+ },
10836
11316
  },
10837
11317
  "vercel-ai-gateway": {
10838
11318
  "alibaba/qwen-3-14b": {
@@ -10854,20 +11334,20 @@ export const MODELS = {
10854
11334
  },
10855
11335
  "alibaba/qwen-3-235b": {
10856
11336
  id: "alibaba/qwen-3-235b",
10857
- name: "Qwen3-235B-A22B",
11337
+ name: "Qwen3 235B A22b Instruct 2507",
10858
11338
  api: "anthropic-messages",
10859
11339
  provider: "vercel-ai-gateway",
10860
11340
  baseUrl: "https://ai-gateway.vercel.sh",
10861
11341
  reasoning: false,
10862
11342
  input: ["text"],
10863
11343
  cost: {
10864
- input: 0.071,
10865
- output: 0.463,
10866
- cacheRead: 0,
11344
+ input: 0.6,
11345
+ output: 1.2,
11346
+ cacheRead: 0.6,
10867
11347
  cacheWrite: 0,
10868
11348
  },
10869
- contextWindow: 40960,
10870
- maxTokens: 16384,
11349
+ contextWindow: 131000,
11350
+ maxTokens: 40000,
10871
11351
  },
10872
11352
  "alibaba/qwen-3-30b": {
10873
11353
  id: "alibaba/qwen-3-30b",
@@ -10895,13 +11375,30 @@ export const MODELS = {
10895
11375
  reasoning: true,
10896
11376
  input: ["text"],
10897
11377
  cost: {
10898
- input: 0.29,
10899
- output: 0.59,
10900
- cacheRead: 0.145,
11378
+ input: 0.16,
11379
+ output: 0.64,
11380
+ cacheRead: 0,
10901
11381
  cacheWrite: 0,
10902
11382
  },
10903
- contextWindow: 131072,
10904
- maxTokens: 40960,
11383
+ contextWindow: 128000,
11384
+ maxTokens: 8192,
11385
+ },
11386
+ "alibaba/qwen-3.6-max-preview": {
11387
+ id: "alibaba/qwen-3.6-max-preview",
11388
+ name: "Qwen 3.6 Max Preview",
11389
+ api: "anthropic-messages",
11390
+ provider: "vercel-ai-gateway",
11391
+ baseUrl: "https://ai-gateway.vercel.sh",
11392
+ reasoning: true,
11393
+ input: ["text", "image"],
11394
+ cost: {
11395
+ input: 1.3,
11396
+ output: 7.8,
11397
+ cacheRead: 0.26,
11398
+ cacheWrite: 1.625,
11399
+ },
11400
+ contextWindow: 240000,
11401
+ maxTokens: 64000,
10905
11402
  },
10906
11403
  "alibaba/qwen3-235b-a22b-thinking": {
10907
11404
  id: "alibaba/qwen3-235b-a22b-thinking",
@@ -10929,13 +11426,13 @@ export const MODELS = {
10929
11426
  reasoning: false,
10930
11427
  input: ["text"],
10931
11428
  cost: {
10932
- input: 0.39999999999999997,
10933
- output: 1.5999999999999999,
10934
- cacheRead: 0.022,
11429
+ input: 1.5,
11430
+ output: 7.5,
11431
+ cacheRead: 0.3,
10935
11432
  cacheWrite: 0,
10936
11433
  },
10937
11434
  contextWindow: 262144,
10938
- maxTokens: 66536,
11435
+ maxTokens: 65536,
10939
11436
  },
10940
11437
  "alibaba/qwen3-coder-30b-a3b": {
10941
11438
  id: "alibaba/qwen3-coder-30b-a3b",
@@ -11048,13 +11545,13 @@ export const MODELS = {
11048
11545
  reasoning: true,
11049
11546
  input: ["text", "image"],
11050
11547
  cost: {
11051
- input: 0.22,
11052
- output: 0.88,
11548
+ input: 0.39999999999999997,
11549
+ output: 4,
11053
11550
  cacheRead: 0,
11054
11551
  cacheWrite: 0,
11055
11552
  },
11056
- contextWindow: 256000,
11057
- maxTokens: 256000,
11553
+ contextWindow: 131072,
11554
+ maxTokens: 32768,
11058
11555
  },
11059
11556
  "alibaba/qwen3.5-flash": {
11060
11557
  id: "alibaba/qwen3.5-flash",
@@ -11090,6 +11587,23 @@ export const MODELS = {
11090
11587
  contextWindow: 1000000,
11091
11588
  maxTokens: 64000,
11092
11589
  },
11590
+ "alibaba/qwen3.6-plus": {
11591
+ id: "alibaba/qwen3.6-plus",
11592
+ name: "Qwen 3.6 Plus",
11593
+ api: "anthropic-messages",
11594
+ provider: "vercel-ai-gateway",
11595
+ baseUrl: "https://ai-gateway.vercel.sh",
11596
+ reasoning: true,
11597
+ input: ["text", "image"],
11598
+ cost: {
11599
+ input: 0.5,
11600
+ output: 3,
11601
+ cacheRead: 0.09999999999999999,
11602
+ cacheWrite: 0.625,
11603
+ },
11604
+ contextWindow: 1000000,
11605
+ maxTokens: 64000,
11606
+ },
11093
11607
  "anthropic/claude-3-haiku": {
11094
11608
  id: "anthropic/claude-3-haiku",
11095
11609
  name: "Claude 3 Haiku",
@@ -11124,40 +11638,6 @@ export const MODELS = {
11124
11638
  contextWindow: 200000,
11125
11639
  maxTokens: 8192,
11126
11640
  },
11127
- "anthropic/claude-3.5-sonnet": {
11128
- id: "anthropic/claude-3.5-sonnet",
11129
- name: "Claude 3.5 Sonnet",
11130
- api: "anthropic-messages",
11131
- provider: "vercel-ai-gateway",
11132
- baseUrl: "https://ai-gateway.vercel.sh",
11133
- reasoning: false,
11134
- input: ["text", "image"],
11135
- cost: {
11136
- input: 3,
11137
- output: 15,
11138
- cacheRead: 0.3,
11139
- cacheWrite: 3.75,
11140
- },
11141
- contextWindow: 200000,
11142
- maxTokens: 8192,
11143
- },
11144
- "anthropic/claude-3.5-sonnet-20240620": {
11145
- id: "anthropic/claude-3.5-sonnet-20240620",
11146
- name: "Claude 3.5 Sonnet (2024-06-20)",
11147
- api: "anthropic-messages",
11148
- provider: "vercel-ai-gateway",
11149
- baseUrl: "https://ai-gateway.vercel.sh",
11150
- reasoning: false,
11151
- input: ["text", "image"],
11152
- cost: {
11153
- input: 3,
11154
- output: 15,
11155
- cacheRead: 0.3,
11156
- cacheWrite: 3.75,
11157
- },
11158
- contextWindow: 200000,
11159
- maxTokens: 8192,
11160
- },
11161
11641
  "anthropic/claude-3.7-sonnet": {
11162
11642
  id: "anthropic/claude-3.7-sonnet",
11163
11643
  name: "Claude 3.7 Sonnet",
@@ -11260,6 +11740,23 @@ export const MODELS = {
11260
11740
  contextWindow: 1000000,
11261
11741
  maxTokens: 128000,
11262
11742
  },
11743
+ "anthropic/claude-opus-4.7": {
11744
+ id: "anthropic/claude-opus-4.7",
11745
+ name: "Claude Opus 4.7",
11746
+ api: "anthropic-messages",
11747
+ provider: "vercel-ai-gateway",
11748
+ baseUrl: "https://ai-gateway.vercel.sh",
11749
+ reasoning: true,
11750
+ input: ["text", "image"],
11751
+ cost: {
11752
+ input: 5,
11753
+ output: 25,
11754
+ cacheRead: 0.5,
11755
+ cacheWrite: 6.25,
11756
+ },
11757
+ contextWindow: 1000000,
11758
+ maxTokens: 128000,
11759
+ },
11263
11760
  "anthropic/claude-sonnet-4": {
11264
11761
  id: "anthropic/claude-sonnet-4",
11265
11762
  name: "Claude Sonnet 4",
@@ -11328,6 +11825,23 @@ export const MODELS = {
11328
11825
  contextWindow: 131000,
11329
11826
  maxTokens: 131000,
11330
11827
  },
11828
+ "arcee-ai/trinity-large-thinking": {
11829
+ id: "arcee-ai/trinity-large-thinking",
11830
+ name: "Trinity Large Thinking",
11831
+ api: "anthropic-messages",
11832
+ provider: "vercel-ai-gateway",
11833
+ baseUrl: "https://ai-gateway.vercel.sh",
11834
+ reasoning: true,
11835
+ input: ["text"],
11836
+ cost: {
11837
+ input: 0.25,
11838
+ output: 0.8999999999999999,
11839
+ cacheRead: 0,
11840
+ cacheWrite: 0,
11841
+ },
11842
+ contextWindow: 262100,
11843
+ maxTokens: 80000,
11844
+ },
11331
11845
  "bytedance/seed-1.6": {
11332
11846
  id: "bytedance/seed-1.6",
11333
11847
  name: "Seed 1.6",
@@ -11405,13 +11919,13 @@ export const MODELS = {
11405
11919
  reasoning: true,
11406
11920
  input: ["text"],
11407
11921
  cost: {
11408
- input: 0.5,
11409
- output: 1.5,
11410
- cacheRead: 0,
11922
+ input: 0.56,
11923
+ output: 1.68,
11924
+ cacheRead: 0.28,
11411
11925
  cacheWrite: 0,
11412
11926
  },
11413
11927
  contextWindow: 163840,
11414
- maxTokens: 16384,
11928
+ maxTokens: 8192,
11415
11929
  },
11416
11930
  "deepseek/deepseek-v3.1-terminus": {
11417
11931
  id: "deepseek/deepseek-v3.1-terminus",
@@ -11594,7 +12108,7 @@ export const MODELS = {
11594
12108
  cost: {
11595
12109
  input: 0.25,
11596
12110
  output: 1.5,
11597
- cacheRead: 0,
12111
+ cacheRead: 0.03,
11598
12112
  cacheWrite: 0,
11599
12113
  },
11600
12114
  contextWindow: 1000000,
@@ -11617,6 +12131,40 @@ export const MODELS = {
11617
12131
  contextWindow: 1000000,
11618
12132
  maxTokens: 64000,
11619
12133
  },
12134
+ "google/gemma-4-26b-a4b-it": {
12135
+ id: "google/gemma-4-26b-a4b-it",
12136
+ name: "Gemma 4 26B A4B IT",
12137
+ api: "anthropic-messages",
12138
+ provider: "vercel-ai-gateway",
12139
+ baseUrl: "https://ai-gateway.vercel.sh",
12140
+ reasoning: false,
12141
+ input: ["text", "image"],
12142
+ cost: {
12143
+ input: 0.13,
12144
+ output: 0.39999999999999997,
12145
+ cacheRead: 0,
12146
+ cacheWrite: 0,
12147
+ },
12148
+ contextWindow: 262144,
12149
+ maxTokens: 131072,
12150
+ },
12151
+ "google/gemma-4-31b-it": {
12152
+ id: "google/gemma-4-31b-it",
12153
+ name: "Gemma 4 31B IT",
12154
+ api: "anthropic-messages",
12155
+ provider: "vercel-ai-gateway",
12156
+ baseUrl: "https://ai-gateway.vercel.sh",
12157
+ reasoning: false,
12158
+ input: ["text", "image"],
12159
+ cost: {
12160
+ input: 0.14,
12161
+ output: 0.39999999999999997,
12162
+ cacheRead: 0,
12163
+ cacheWrite: 0,
12164
+ },
12165
+ contextWindow: 262144,
12166
+ maxTokens: 131072,
12167
+ },
11620
12168
  "inception/mercury-2": {
11621
12169
  id: "inception/mercury-2",
11622
12170
  name: "Mercury 2",
@@ -11651,39 +12199,39 @@ export const MODELS = {
11651
12199
  contextWindow: 32000,
11652
12200
  maxTokens: 16384,
11653
12201
  },
11654
- "meituan/longcat-flash-chat": {
11655
- id: "meituan/longcat-flash-chat",
11656
- name: "LongCat Flash Chat",
12202
+ "kwaipilot/kat-coder-pro-v2": {
12203
+ id: "kwaipilot/kat-coder-pro-v2",
12204
+ name: "Kat Coder Pro V2",
11657
12205
  api: "anthropic-messages",
11658
12206
  provider: "vercel-ai-gateway",
11659
12207
  baseUrl: "https://ai-gateway.vercel.sh",
11660
- reasoning: false,
12208
+ reasoning: true,
11661
12209
  input: ["text"],
11662
12210
  cost: {
11663
- input: 0,
11664
- output: 0,
11665
- cacheRead: 0,
12211
+ input: 0.3,
12212
+ output: 1.2,
12213
+ cacheRead: 0.06,
11666
12214
  cacheWrite: 0,
11667
12215
  },
11668
- contextWindow: 128000,
11669
- maxTokens: 100000,
12216
+ contextWindow: 256000,
12217
+ maxTokens: 256000,
11670
12218
  },
11671
- "meituan/longcat-flash-thinking": {
11672
- id: "meituan/longcat-flash-thinking",
11673
- name: "LongCat Flash Thinking",
12219
+ "meituan/longcat-flash-chat": {
12220
+ id: "meituan/longcat-flash-chat",
12221
+ name: "LongCat Flash Chat",
11674
12222
  api: "anthropic-messages",
11675
12223
  provider: "vercel-ai-gateway",
11676
12224
  baseUrl: "https://ai-gateway.vercel.sh",
11677
- reasoning: true,
12225
+ reasoning: false,
11678
12226
  input: ["text"],
11679
12227
  cost: {
11680
- input: 0.15,
11681
- output: 1.5,
12228
+ input: 0,
12229
+ output: 0,
11682
12230
  cacheRead: 0,
11683
12231
  cacheWrite: 0,
11684
12232
  },
11685
12233
  contextWindow: 128000,
11686
- maxTokens: 8192,
12234
+ maxTokens: 100000,
11687
12235
  },
11688
12236
  "meta/llama-3.1-70b": {
11689
12237
  id: "meta/llama-3.1-70b",
@@ -11711,13 +12259,13 @@ export const MODELS = {
11711
12259
  reasoning: false,
11712
12260
  input: ["text"],
11713
12261
  cost: {
11714
- input: 0.09999999999999999,
11715
- output: 0.09999999999999999,
11716
- cacheRead: 0.09999999999999999,
12262
+ input: 0.22,
12263
+ output: 0.22,
12264
+ cacheRead: 0,
11717
12265
  cacheWrite: 0,
11718
12266
  },
11719
12267
  contextWindow: 128000,
11720
- maxTokens: 16384,
12268
+ maxTokens: 8192,
11721
12269
  },
11722
12270
  "meta/llama-3.2-11b": {
11723
12271
  id: "meta/llama-3.2-11b",
@@ -12095,20 +12643,20 @@ export const MODELS = {
12095
12643
  },
12096
12644
  "moonshotai/kimi-k2": {
12097
12645
  id: "moonshotai/kimi-k2",
12098
- name: "Kimi K2",
12646
+ name: "Kimi K2 Instruct",
12099
12647
  api: "anthropic-messages",
12100
12648
  provider: "vercel-ai-gateway",
12101
12649
  baseUrl: "https://ai-gateway.vercel.sh",
12102
12650
  reasoning: false,
12103
12651
  input: ["text"],
12104
12652
  cost: {
12105
- input: 0.6,
12106
- output: 2.5,
12107
- cacheRead: 0.15,
12653
+ input: 0.5700000000000001,
12654
+ output: 2.3,
12655
+ cacheRead: 0,
12108
12656
  cacheWrite: 0,
12109
12657
  },
12110
12658
  contextWindow: 131072,
12111
- maxTokens: 16384,
12659
+ maxTokens: 131072,
12112
12660
  },
12113
12661
  "moonshotai/kimi-k2-0905": {
12114
12662
  id: "moonshotai/kimi-k2-0905",
@@ -12121,11 +12669,11 @@ export const MODELS = {
12121
12669
  cost: {
12122
12670
  input: 0.6,
12123
12671
  output: 2.5,
12124
- cacheRead: 0.15,
12672
+ cacheRead: 0.3,
12125
12673
  cacheWrite: 0,
12126
12674
  },
12127
12675
  contextWindow: 256000,
12128
- maxTokens: 16384,
12676
+ maxTokens: 128000,
12129
12677
  },
12130
12678
  "moonshotai/kimi-k2-thinking": {
12131
12679
  id: "moonshotai/kimi-k2-thinking",
@@ -12195,6 +12743,23 @@ export const MODELS = {
12195
12743
  contextWindow: 262114,
12196
12744
  maxTokens: 262114,
12197
12745
  },
12746
+ "moonshotai/kimi-k2.6": {
12747
+ id: "moonshotai/kimi-k2.6",
12748
+ name: "Kimi K2.6",
12749
+ api: "anthropic-messages",
12750
+ provider: "vercel-ai-gateway",
12751
+ baseUrl: "https://ai-gateway.vercel.sh",
12752
+ reasoning: true,
12753
+ input: ["text", "image"],
12754
+ cost: {
12755
+ input: 0.95,
12756
+ output: 4,
12757
+ cacheRead: 0.16,
12758
+ cacheWrite: 0,
12759
+ },
12760
+ contextWindow: 262000,
12761
+ maxTokens: 262000,
12762
+ },
12198
12763
  "nvidia/nemotron-nano-12b-v2-vl": {
12199
12764
  id: "nvidia/nemotron-nano-12b-v2-vl",
12200
12765
  name: "Nvidia Nemotron Nano 12B V2 VL",
@@ -12690,24 +13255,24 @@ export const MODELS = {
12690
13255
  },
12691
13256
  "openai/gpt-oss-20b": {
12692
13257
  id: "openai/gpt-oss-20b",
12693
- name: "gpt-oss-20b",
13258
+ name: "GPT OSS 120B",
12694
13259
  api: "anthropic-messages",
12695
13260
  provider: "vercel-ai-gateway",
12696
13261
  baseUrl: "https://ai-gateway.vercel.sh",
12697
13262
  reasoning: true,
12698
13263
  input: ["text"],
12699
13264
  cost: {
12700
- input: 0.07,
12701
- output: 0.3,
13265
+ input: 0.049999999999999996,
13266
+ output: 0.19999999999999998,
12702
13267
  cacheRead: 0,
12703
13268
  cacheWrite: 0,
12704
13269
  },
12705
- contextWindow: 128000,
13270
+ contextWindow: 131072,
12706
13271
  maxTokens: 8192,
12707
13272
  },
12708
13273
  "openai/gpt-oss-safeguard-20b": {
12709
13274
  id: "openai/gpt-oss-safeguard-20b",
12710
- name: "gpt-oss-safeguard-20b",
13275
+ name: "GPT OSS Safeguard 20B",
12711
13276
  api: "anthropic-messages",
12712
13277
  provider: "vercel-ai-gateway",
12713
13278
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -12875,23 +13440,6 @@ export const MODELS = {
12875
13440
  contextWindow: 131072,
12876
13441
  maxTokens: 131072,
12877
13442
  },
12878
- "xai/grok-2-vision": {
12879
- id: "xai/grok-2-vision",
12880
- name: "Grok 2 Vision",
12881
- api: "anthropic-messages",
12882
- provider: "vercel-ai-gateway",
12883
- baseUrl: "https://ai-gateway.vercel.sh",
12884
- reasoning: false,
12885
- input: ["text", "image"],
12886
- cost: {
12887
- input: 2,
12888
- output: 10,
12889
- cacheRead: 0,
12890
- cacheWrite: 0,
12891
- },
12892
- contextWindow: 32768,
12893
- maxTokens: 32768,
12894
- },
12895
13443
  "xai/grok-3": {
12896
13444
  id: "xai/grok-3",
12897
13445
  name: "Grok 3 Beta",
@@ -12984,7 +13532,7 @@ export const MODELS = {
12984
13532
  provider: "vercel-ai-gateway",
12985
13533
  baseUrl: "https://ai-gateway.vercel.sh",
12986
13534
  reasoning: false,
12987
- input: ["text"],
13535
+ input: ["text", "image"],
12988
13536
  cost: {
12989
13537
  input: 0.19999999999999998,
12990
13538
  output: 0.5,
@@ -13001,7 +13549,7 @@ export const MODELS = {
13001
13549
  provider: "vercel-ai-gateway",
13002
13550
  baseUrl: "https://ai-gateway.vercel.sh",
13003
13551
  reasoning: true,
13004
- input: ["text"],
13552
+ input: ["text", "image"],
13005
13553
  cost: {
13006
13554
  input: 0.19999999999999998,
13007
13555
  output: 0.5,
@@ -13018,7 +13566,7 @@ export const MODELS = {
13018
13566
  provider: "vercel-ai-gateway",
13019
13567
  baseUrl: "https://ai-gateway.vercel.sh",
13020
13568
  reasoning: false,
13021
- input: ["text"],
13569
+ input: ["text", "image"],
13022
13570
  cost: {
13023
13571
  input: 0.19999999999999998,
13024
13572
  output: 0.5,
@@ -13035,7 +13583,7 @@ export const MODELS = {
13035
13583
  provider: "vercel-ai-gateway",
13036
13584
  baseUrl: "https://ai-gateway.vercel.sh",
13037
13585
  reasoning: true,
13038
- input: ["text"],
13586
+ input: ["text", "image"],
13039
13587
  cost: {
13040
13588
  input: 0.19999999999999998,
13041
13589
  output: 0.5,
@@ -13052,7 +13600,7 @@ export const MODELS = {
13052
13600
  provider: "vercel-ai-gateway",
13053
13601
  baseUrl: "https://ai-gateway.vercel.sh",
13054
13602
  reasoning: true,
13055
- input: ["text"],
13603
+ input: ["text", "image"],
13056
13604
  cost: {
13057
13605
  input: 2,
13058
13606
  output: 6,
@@ -13069,7 +13617,7 @@ export const MODELS = {
13069
13617
  provider: "vercel-ai-gateway",
13070
13618
  baseUrl: "https://ai-gateway.vercel.sh",
13071
13619
  reasoning: true,
13072
- input: ["text"],
13620
+ input: ["text", "image"],
13073
13621
  cost: {
13074
13622
  input: 2,
13075
13623
  output: 6,
@@ -13175,7 +13723,7 @@ export const MODELS = {
13175
13723
  cost: {
13176
13724
  input: 0.09999999999999999,
13177
13725
  output: 0.3,
13178
- cacheRead: 0.02,
13726
+ cacheRead: 0.01,
13179
13727
  cacheWrite: 0,
13180
13728
  },
13181
13729
  contextWindow: 262144,
@@ -13309,13 +13857,13 @@ export const MODELS = {
13309
13857
  reasoning: true,
13310
13858
  input: ["text"],
13311
13859
  cost: {
13312
- input: 0.6,
13313
- output: 2.2,
13314
- cacheRead: 0,
13860
+ input: 2.25,
13861
+ output: 2.75,
13862
+ cacheRead: 2.25,
13315
13863
  cacheWrite: 0,
13316
13864
  },
13317
- contextWindow: 200000,
13318
- maxTokens: 120000,
13865
+ contextWindow: 131000,
13866
+ maxTokens: 40000,
13319
13867
  },
13320
13868
  "zai/glm-4.7-flash": {
13321
13869
  id: "zai/glm-4.7-flash",
@@ -13385,6 +13933,40 @@ export const MODELS = {
13385
13933
  contextWindow: 202800,
13386
13934
  maxTokens: 131100,
13387
13935
  },
13936
+ "zai/glm-5.1": {
13937
+ id: "zai/glm-5.1",
13938
+ name: "GLM 5.1",
13939
+ api: "anthropic-messages",
13940
+ provider: "vercel-ai-gateway",
13941
+ baseUrl: "https://ai-gateway.vercel.sh",
13942
+ reasoning: true,
13943
+ input: ["text"],
13944
+ cost: {
13945
+ input: 1.4,
13946
+ output: 4.4,
13947
+ cacheRead: 0.26,
13948
+ cacheWrite: 0,
13949
+ },
13950
+ contextWindow: 202800,
13951
+ maxTokens: 64000,
13952
+ },
13953
+ "zai/glm-5v-turbo": {
13954
+ id: "zai/glm-5v-turbo",
13955
+ name: "GLM 5V Turbo",
13956
+ api: "anthropic-messages",
13957
+ provider: "vercel-ai-gateway",
13958
+ baseUrl: "https://ai-gateway.vercel.sh",
13959
+ reasoning: true,
13960
+ input: ["text", "image"],
13961
+ cost: {
13962
+ input: 1.2,
13963
+ output: 4,
13964
+ cacheRead: 0.24,
13965
+ cacheWrite: 0,
13966
+ },
13967
+ contextWindow: 200000,
13968
+ maxTokens: 128000,
13969
+ },
13388
13970
  },
13389
13971
  "xai": {
13390
13972
  "grok-2": {
@@ -13807,9 +14389,9 @@ export const MODELS = {
13807
14389
  reasoning: true,
13808
14390
  input: ["text"],
13809
14391
  cost: {
13810
- input: 0.6,
13811
- output: 2.2,
13812
- cacheRead: 0.11,
14392
+ input: 0,
14393
+ output: 0,
14394
+ cacheRead: 0,
13813
14395
  cacheWrite: 0,
13814
14396
  },
13815
14397
  contextWindow: 131072,
@@ -13825,9 +14407,9 @@ export const MODELS = {
13825
14407
  reasoning: true,
13826
14408
  input: ["text"],
13827
14409
  cost: {
13828
- input: 0.2,
13829
- output: 1.1,
13830
- cacheRead: 0.03,
14410
+ input: 0,
14411
+ output: 0,
14412
+ cacheRead: 0,
13831
14413
  cacheWrite: 0,
13832
14414
  },
13833
14415
  contextWindow: 131072,
@@ -13861,8 +14443,8 @@ export const MODELS = {
13861
14443
  reasoning: true,
13862
14444
  input: ["text", "image"],
13863
14445
  cost: {
13864
- input: 0.6,
13865
- output: 1.8,
14446
+ input: 0,
14447
+ output: 0,
13866
14448
  cacheRead: 0,
13867
14449
  cacheWrite: 0,
13868
14450
  },
@@ -13875,13 +14457,13 @@ export const MODELS = {
13875
14457
  api: "openai-completions",
13876
14458
  provider: "zai",
13877
14459
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13878
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14460
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13879
14461
  reasoning: true,
13880
14462
  input: ["text"],
13881
14463
  cost: {
13882
- input: 0.6,
13883
- output: 2.2,
13884
- cacheRead: 0.11,
14464
+ input: 0,
14465
+ output: 0,
14466
+ cacheRead: 0,
13885
14467
  cacheWrite: 0,
13886
14468
  },
13887
14469
  contextWindow: 204800,
@@ -13893,12 +14475,12 @@ export const MODELS = {
13893
14475
  api: "openai-completions",
13894
14476
  provider: "zai",
13895
14477
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13896
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14478
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13897
14479
  reasoning: true,
13898
14480
  input: ["text", "image"],
13899
14481
  cost: {
13900
- input: 0.3,
13901
- output: 0.9,
14482
+ input: 0,
14483
+ output: 0,
13902
14484
  cacheRead: 0,
13903
14485
  cacheWrite: 0,
13904
14486
  },
@@ -13911,13 +14493,13 @@ export const MODELS = {
13911
14493
  api: "openai-completions",
13912
14494
  provider: "zai",
13913
14495
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13914
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14496
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13915
14497
  reasoning: true,
13916
14498
  input: ["text"],
13917
14499
  cost: {
13918
- input: 0.6,
13919
- output: 2.2,
13920
- cacheRead: 0.11,
14500
+ input: 0,
14501
+ output: 0,
14502
+ cacheRead: 0,
13921
14503
  cacheWrite: 0,
13922
14504
  },
13923
14505
  contextWindow: 204800,
@@ -13929,7 +14511,7 @@ export const MODELS = {
13929
14511
  api: "openai-completions",
13930
14512
  provider: "zai",
13931
14513
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13932
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14514
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13933
14515
  reasoning: true,
13934
14516
  input: ["text"],
13935
14517
  cost: {
@@ -13941,19 +14523,37 @@ export const MODELS = {
13941
14523
  contextWindow: 200000,
13942
14524
  maxTokens: 131072,
13943
14525
  },
14526
+ "glm-4.7-flashx": {
14527
+ id: "glm-4.7-flashx",
14528
+ name: "GLM-4.7-FlashX",
14529
+ api: "openai-completions",
14530
+ provider: "zai",
14531
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14532
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14533
+ reasoning: true,
14534
+ input: ["text"],
14535
+ cost: {
14536
+ input: 0.07,
14537
+ output: 0.4,
14538
+ cacheRead: 0.01,
14539
+ cacheWrite: 0,
14540
+ },
14541
+ contextWindow: 200000,
14542
+ maxTokens: 131072,
14543
+ },
13944
14544
  "glm-5": {
13945
14545
  id: "glm-5",
13946
14546
  name: "GLM-5",
13947
14547
  api: "openai-completions",
13948
14548
  provider: "zai",
13949
14549
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13950
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14550
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13951
14551
  reasoning: true,
13952
14552
  input: ["text"],
13953
14553
  cost: {
13954
- input: 1,
13955
- output: 3.2,
13956
- cacheRead: 0.2,
14554
+ input: 0,
14555
+ output: 0,
14556
+ cacheRead: 0,
13957
14557
  cacheWrite: 0,
13958
14558
  },
13959
14559
  contextWindow: 204800,
@@ -13965,13 +14565,49 @@ export const MODELS = {
13965
14565
  api: "openai-completions",
13966
14566
  provider: "zai",
13967
14567
  baseUrl: "https://api.z.ai/api/coding/paas/v4",
13968
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14568
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
13969
14569
  reasoning: true,
13970
14570
  input: ["text"],
13971
14571
  cost: {
13972
- input: 1.2,
13973
- output: 4,
13974
- cacheRead: 0.24,
14572
+ input: 0,
14573
+ output: 0,
14574
+ cacheRead: 0,
14575
+ cacheWrite: 0,
14576
+ },
14577
+ contextWindow: 200000,
14578
+ maxTokens: 131072,
14579
+ },
14580
+ "glm-5.1": {
14581
+ id: "glm-5.1",
14582
+ name: "GLM-5.1",
14583
+ api: "openai-completions",
14584
+ provider: "zai",
14585
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14586
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14587
+ reasoning: true,
14588
+ input: ["text"],
14589
+ cost: {
14590
+ input: 0,
14591
+ output: 0,
14592
+ cacheRead: 0,
14593
+ cacheWrite: 0,
14594
+ },
14595
+ contextWindow: 200000,
14596
+ maxTokens: 131072,
14597
+ },
14598
+ "glm-5v-turbo": {
14599
+ id: "glm-5v-turbo",
14600
+ name: "glm-5v-turbo",
14601
+ api: "openai-completions",
14602
+ provider: "zai",
14603
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14604
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14605
+ reasoning: true,
14606
+ input: ["text", "image"],
14607
+ cost: {
14608
+ input: 0,
14609
+ output: 0,
14610
+ cacheRead: 0,
13975
14611
  cacheWrite: 0,
13976
14612
  },
13977
14613
  contextWindow: 200000,