@mariozechner/pi-ai 0.57.1 → 0.58.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -87,40 +87,6 @@ export const MODELS = {
87
87
  contextWindow: 300000,
88
88
  maxTokens: 8192,
89
89
  },
90
- "amazon.titan-text-express-v1": {
91
- id: "amazon.titan-text-express-v1",
92
- name: "Titan Text G1 - Express",
93
- api: "bedrock-converse-stream",
94
- provider: "amazon-bedrock",
95
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
96
- reasoning: false,
97
- input: ["text"],
98
- cost: {
99
- input: 0.2,
100
- output: 0.6,
101
- cacheRead: 0,
102
- cacheWrite: 0,
103
- },
104
- contextWindow: 128000,
105
- maxTokens: 4096,
106
- },
107
- "amazon.titan-text-express-v1:0:8k": {
108
- id: "amazon.titan-text-express-v1:0:8k",
109
- name: "Titan Text G1 - Express",
110
- api: "bedrock-converse-stream",
111
- provider: "amazon-bedrock",
112
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
113
- reasoning: false,
114
- input: ["text"],
115
- cost: {
116
- input: 0.2,
117
- output: 0.6,
118
- cacheRead: 0,
119
- cacheWrite: 0,
120
- },
121
- contextWindow: 128000,
122
- maxTokens: 4096,
123
- },
124
90
  "anthropic.claude-3-5-haiku-20241022-v1:0": {
125
91
  id: "anthropic.claude-3-5-haiku-20241022-v1:0",
126
92
  name: "Claude Haiku 3.5",
@@ -206,40 +172,6 @@ export const MODELS = {
206
172
  contextWindow: 200000,
207
173
  maxTokens: 4096,
208
174
  },
209
- "anthropic.claude-3-opus-20240229-v1:0": {
210
- id: "anthropic.claude-3-opus-20240229-v1:0",
211
- name: "Claude Opus 3",
212
- api: "bedrock-converse-stream",
213
- provider: "amazon-bedrock",
214
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
215
- reasoning: false,
216
- input: ["text", "image"],
217
- cost: {
218
- input: 15,
219
- output: 75,
220
- cacheRead: 0,
221
- cacheWrite: 0,
222
- },
223
- contextWindow: 200000,
224
- maxTokens: 4096,
225
- },
226
- "anthropic.claude-3-sonnet-20240229-v1:0": {
227
- id: "anthropic.claude-3-sonnet-20240229-v1:0",
228
- name: "Claude Sonnet 3",
229
- api: "bedrock-converse-stream",
230
- provider: "amazon-bedrock",
231
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
232
- reasoning: false,
233
- input: ["text", "image"],
234
- cost: {
235
- input: 3,
236
- output: 15,
237
- cacheRead: 0,
238
- cacheWrite: 0,
239
- },
240
- contextWindow: 200000,
241
- maxTokens: 4096,
242
- },
243
175
  "anthropic.claude-haiku-4-5-20251001-v1:0": {
244
176
  id: "anthropic.claude-haiku-4-5-20251001-v1:0",
245
177
  name: "Claude Haiku 4.5",
@@ -322,7 +254,7 @@ export const MODELS = {
322
254
  cacheRead: 0.5,
323
255
  cacheWrite: 6.25,
324
256
  },
325
- contextWindow: 200000,
257
+ contextWindow: 1000000,
326
258
  maxTokens: 128000,
327
259
  },
328
260
  "anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -373,43 +305,9 @@ export const MODELS = {
373
305
  cacheRead: 0.3,
374
306
  cacheWrite: 3.75,
375
307
  },
376
- contextWindow: 200000,
308
+ contextWindow: 1000000,
377
309
  maxTokens: 64000,
378
310
  },
379
- "cohere.command-r-plus-v1:0": {
380
- id: "cohere.command-r-plus-v1:0",
381
- name: "Command R+",
382
- api: "bedrock-converse-stream",
383
- provider: "amazon-bedrock",
384
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
385
- reasoning: false,
386
- input: ["text"],
387
- cost: {
388
- input: 3,
389
- output: 15,
390
- cacheRead: 0,
391
- cacheWrite: 0,
392
- },
393
- contextWindow: 128000,
394
- maxTokens: 4096,
395
- },
396
- "cohere.command-r-v1:0": {
397
- id: "cohere.command-r-v1:0",
398
- name: "Command R",
399
- api: "bedrock-converse-stream",
400
- provider: "amazon-bedrock",
401
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
402
- reasoning: false,
403
- input: ["text"],
404
- cost: {
405
- input: 0.5,
406
- output: 1.5,
407
- cacheRead: 0,
408
- cacheWrite: 0,
409
- },
410
- contextWindow: 128000,
411
- maxTokens: 4096,
412
- },
413
311
  "deepseek.r1-v1:0": {
414
312
  id: "deepseek.r1-v1:0",
415
313
  name: "DeepSeek-R1",
@@ -444,8 +342,8 @@ export const MODELS = {
444
342
  contextWindow: 163840,
445
343
  maxTokens: 81920,
446
344
  },
447
- "deepseek.v3.2-v1:0": {
448
- id: "deepseek.v3.2-v1:0",
345
+ "deepseek.v3.2": {
346
+ id: "deepseek.v3.2",
449
347
  name: "DeepSeek-V3.2",
450
348
  api: "bedrock-converse-stream",
451
349
  provider: "amazon-bedrock",
@@ -509,7 +407,7 @@ export const MODELS = {
509
407
  cacheRead: 0.5,
510
408
  cacheWrite: 6.25,
511
409
  },
512
- contextWindow: 200000,
410
+ contextWindow: 1000000,
513
411
  maxTokens: 128000,
514
412
  },
515
413
  "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -560,7 +458,7 @@ export const MODELS = {
560
458
  cacheRead: 0.3,
561
459
  cacheWrite: 3.75,
562
460
  },
563
- contextWindow: 200000,
461
+ contextWindow: 1000000,
564
462
  maxTokens: 64000,
565
463
  },
566
464
  "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
@@ -611,7 +509,7 @@ export const MODELS = {
611
509
  cacheRead: 0.5,
612
510
  cacheWrite: 6.25,
613
511
  },
614
- contextWindow: 200000,
512
+ contextWindow: 1000000,
615
513
  maxTokens: 128000,
616
514
  },
617
515
  "global.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -662,7 +560,7 @@ export const MODELS = {
662
560
  cacheRead: 0.3,
663
561
  cacheWrite: 3.75,
664
562
  },
665
- contextWindow: 200000,
563
+ contextWindow: 1000000,
666
564
  maxTokens: 64000,
667
565
  },
668
566
  "google.gemma-3-27b-it": {
@@ -699,6 +597,23 @@ export const MODELS = {
699
597
  contextWindow: 128000,
700
598
  maxTokens: 4096,
701
599
  },
600
+ "meta.llama3-1-405b-instruct-v1:0": {
601
+ id: "meta.llama3-1-405b-instruct-v1:0",
602
+ name: "Llama 3.1 405B Instruct",
603
+ api: "bedrock-converse-stream",
604
+ provider: "amazon-bedrock",
605
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
606
+ reasoning: false,
607
+ input: ["text"],
608
+ cost: {
609
+ input: 2.4,
610
+ output: 2.4,
611
+ cacheRead: 0,
612
+ cacheWrite: 0,
613
+ },
614
+ contextWindow: 128000,
615
+ maxTokens: 4096,
616
+ },
702
617
  "meta.llama3-1-70b-instruct-v1:0": {
703
618
  id: "meta.llama3-1-70b-instruct-v1:0",
704
619
  name: "Llama 3.1 70B Instruct",
@@ -886,6 +801,40 @@ export const MODELS = {
886
801
  contextWindow: 204800,
887
802
  maxTokens: 131072,
888
803
  },
804
+ "mistral.devstral-2-123b": {
805
+ id: "mistral.devstral-2-123b",
806
+ name: "Devstral 2 123B",
807
+ api: "bedrock-converse-stream",
808
+ provider: "amazon-bedrock",
809
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
810
+ reasoning: false,
811
+ input: ["text"],
812
+ cost: {
813
+ input: 0.4,
814
+ output: 2,
815
+ cacheRead: 0,
816
+ cacheWrite: 0,
817
+ },
818
+ contextWindow: 256000,
819
+ maxTokens: 8192,
820
+ },
821
+ "mistral.magistral-small-2509": {
822
+ id: "mistral.magistral-small-2509",
823
+ name: "Magistral Small 1.2",
824
+ api: "bedrock-converse-stream",
825
+ provider: "amazon-bedrock",
826
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
827
+ reasoning: true,
828
+ input: ["text", "image"],
829
+ cost: {
830
+ input: 0.5,
831
+ output: 1.5,
832
+ cacheRead: 0,
833
+ cacheWrite: 0,
834
+ },
835
+ contextWindow: 128000,
836
+ maxTokens: 40000,
837
+ },
889
838
  "mistral.ministral-3-14b-instruct": {
890
839
  id: "mistral.ministral-3-14b-instruct",
891
840
  name: "Ministral 14B 3.0",
@@ -903,6 +852,23 @@ export const MODELS = {
903
852
  contextWindow: 128000,
904
853
  maxTokens: 4096,
905
854
  },
855
+ "mistral.ministral-3-3b-instruct": {
856
+ id: "mistral.ministral-3-3b-instruct",
857
+ name: "Ministral 3 3B",
858
+ api: "bedrock-converse-stream",
859
+ provider: "amazon-bedrock",
860
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
861
+ reasoning: false,
862
+ input: ["text", "image"],
863
+ cost: {
864
+ input: 0.1,
865
+ output: 0.1,
866
+ cacheRead: 0,
867
+ cacheWrite: 0,
868
+ },
869
+ contextWindow: 256000,
870
+ maxTokens: 8192,
871
+ },
906
872
  "mistral.ministral-3-8b-instruct": {
907
873
  id: "mistral.ministral-3-8b-instruct",
908
874
  name: "Ministral 3 8B",
@@ -920,22 +886,39 @@ export const MODELS = {
920
886
  contextWindow: 128000,
921
887
  maxTokens: 4096,
922
888
  },
923
- "mistral.mistral-large-2402-v1:0": {
924
- id: "mistral.mistral-large-2402-v1:0",
925
- name: "Mistral Large (24.02)",
889
+ "mistral.mistral-large-3-675b-instruct": {
890
+ id: "mistral.mistral-large-3-675b-instruct",
891
+ name: "Mistral Large 3",
926
892
  api: "bedrock-converse-stream",
927
893
  provider: "amazon-bedrock",
928
894
  baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
929
895
  reasoning: false,
930
- input: ["text"],
896
+ input: ["text", "image"],
931
897
  cost: {
932
898
  input: 0.5,
933
899
  output: 1.5,
934
900
  cacheRead: 0,
935
901
  cacheWrite: 0,
936
902
  },
903
+ contextWindow: 256000,
904
+ maxTokens: 8192,
905
+ },
906
+ "mistral.pixtral-large-2502-v1:0": {
907
+ id: "mistral.pixtral-large-2502-v1:0",
908
+ name: "Pixtral Large (25.02)",
909
+ api: "bedrock-converse-stream",
910
+ provider: "amazon-bedrock",
911
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
912
+ reasoning: false,
913
+ input: ["text", "image"],
914
+ cost: {
915
+ input: 2,
916
+ output: 6,
917
+ cacheRead: 0,
918
+ cacheWrite: 0,
919
+ },
937
920
  contextWindow: 128000,
938
- maxTokens: 4096,
921
+ maxTokens: 8192,
939
922
  },
940
923
  "mistral.voxtral-mini-3b-2507": {
941
924
  id: "mistral.voxtral-mini-3b-2507",
@@ -1022,6 +1005,23 @@ export const MODELS = {
1022
1005
  contextWindow: 128000,
1023
1006
  maxTokens: 4096,
1024
1007
  },
1008
+ "nvidia.nemotron-nano-3-30b": {
1009
+ id: "nvidia.nemotron-nano-3-30b",
1010
+ name: "NVIDIA Nemotron Nano 3 30B",
1011
+ api: "bedrock-converse-stream",
1012
+ provider: "amazon-bedrock",
1013
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
1014
+ reasoning: true,
1015
+ input: ["text"],
1016
+ cost: {
1017
+ input: 0.06,
1018
+ output: 0.24,
1019
+ cacheRead: 0,
1020
+ cacheWrite: 0,
1021
+ },
1022
+ contextWindow: 128000,
1023
+ maxTokens: 4096,
1024
+ },
1025
1025
  "nvidia.nemotron-nano-9b-v2": {
1026
1026
  id: "nvidia.nemotron-nano-9b-v2",
1027
1027
  name: "NVIDIA Nemotron Nano 9B v2",
@@ -1291,7 +1291,7 @@ export const MODELS = {
1291
1291
  cacheRead: 0.5,
1292
1292
  cacheWrite: 6.25,
1293
1293
  },
1294
- contextWindow: 200000,
1294
+ contextWindow: 1000000,
1295
1295
  maxTokens: 128000,
1296
1296
  },
1297
1297
  "us.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -1342,7 +1342,7 @@ export const MODELS = {
1342
1342
  cacheRead: 0.3,
1343
1343
  cacheWrite: 3.75,
1344
1344
  },
1345
- contextWindow: 200000,
1345
+ contextWindow: 1000000,
1346
1346
  maxTokens: 64000,
1347
1347
  },
1348
1348
  "writer.palmyra-x4-v1:0": {
@@ -1718,7 +1718,7 @@ export const MODELS = {
1718
1718
  cacheRead: 0.5,
1719
1719
  cacheWrite: 6.25,
1720
1720
  },
1721
- contextWindow: 200000,
1721
+ contextWindow: 1000000,
1722
1722
  maxTokens: 128000,
1723
1723
  },
1724
1724
  "claude-sonnet-4-0": {
@@ -1803,7 +1803,7 @@ export const MODELS = {
1803
1803
  cacheRead: 0.3,
1804
1804
  cacheWrite: 3.75,
1805
1805
  },
1806
- contextWindow: 200000,
1806
+ contextWindow: 1000000,
1807
1807
  maxTokens: 64000,
1808
1808
  },
1809
1809
  },
@@ -2871,7 +2871,7 @@ export const MODELS = {
2871
2871
  cacheRead: 0,
2872
2872
  cacheWrite: 0,
2873
2873
  },
2874
- contextWindow: 128000,
2874
+ contextWindow: 264000,
2875
2875
  maxTokens: 64000,
2876
2876
  },
2877
2877
  "gpt-5.2-codex": {
@@ -2889,7 +2889,7 @@ export const MODELS = {
2889
2889
  cacheRead: 0,
2890
2890
  cacheWrite: 0,
2891
2891
  },
2892
- contextWindow: 272000,
2892
+ contextWindow: 400000,
2893
2893
  maxTokens: 128000,
2894
2894
  },
2895
2895
  "gpt-5.3-codex": {
@@ -3247,10 +3247,10 @@ export const MODELS = {
3247
3247
  reasoning: true,
3248
3248
  input: ["text", "image"],
3249
3249
  cost: {
3250
- input: 0.5,
3251
- output: 3,
3252
- cacheRead: 0.05,
3253
- cacheWrite: 0,
3250
+ input: 0.25,
3251
+ output: 1.5,
3252
+ cacheRead: 0.025,
3253
+ cacheWrite: 1,
3254
3254
  },
3255
3255
  contextWindow: 1048576,
3256
3256
  maxTokens: 65536,
@@ -5895,7 +5895,7 @@ export const MODELS = {
5895
5895
  cacheRead: 0.5,
5896
5896
  cacheWrite: 6.25,
5897
5897
  },
5898
- contextWindow: 200000,
5898
+ contextWindow: 1000000,
5899
5899
  maxTokens: 128000,
5900
5900
  },
5901
5901
  "claude-sonnet-4": {
@@ -6272,6 +6272,23 @@ export const MODELS = {
6272
6272
  contextWindow: 262144,
6273
6273
  maxTokens: 65536,
6274
6274
  },
6275
+ "mimo-v2-flash-free": {
6276
+ id: "mimo-v2-flash-free",
6277
+ name: "MiMo V2 Flash Free",
6278
+ api: "openai-completions",
6279
+ provider: "opencode",
6280
+ baseUrl: "https://opencode.ai/zen/v1",
6281
+ reasoning: true,
6282
+ input: ["text"],
6283
+ cost: {
6284
+ input: 0,
6285
+ output: 0,
6286
+ cacheRead: 0,
6287
+ cacheWrite: 0,
6288
+ },
6289
+ contextWindow: 262144,
6290
+ maxTokens: 65536,
6291
+ },
6275
6292
  "minimax-m2.1": {
6276
6293
  id: "minimax-m2.1",
6277
6294
  name: "MiniMax M2.1",
@@ -6323,6 +6340,23 @@ export const MODELS = {
6323
6340
  contextWindow: 204800,
6324
6341
  maxTokens: 131072,
6325
6342
  },
6343
+ "nemotron-3-super-free": {
6344
+ id: "nemotron-3-super-free",
6345
+ name: "Nemotron 3 Super Free",
6346
+ api: "openai-completions",
6347
+ provider: "opencode",
6348
+ baseUrl: "https://opencode.ai/zen/v1",
6349
+ reasoning: true,
6350
+ input: ["text"],
6351
+ cost: {
6352
+ input: 0,
6353
+ output: 0,
6354
+ cacheRead: 0,
6355
+ cacheWrite: 0,
6356
+ },
6357
+ contextWindow: 1000000,
6358
+ maxTokens: 128000,
6359
+ },
6326
6360
  },
6327
6361
  "opencode-go": {
6328
6362
  "glm-5": {
@@ -6888,6 +6922,23 @@ export const MODELS = {
6888
6922
  contextWindow: 262144,
6889
6923
  maxTokens: 32768,
6890
6924
  },
6925
+ "bytedance-seed/seed-2.0-lite": {
6926
+ id: "bytedance-seed/seed-2.0-lite",
6927
+ name: "ByteDance Seed: Seed-2.0-Lite",
6928
+ api: "openai-completions",
6929
+ provider: "openrouter",
6930
+ baseUrl: "https://openrouter.ai/api/v1",
6931
+ reasoning: true,
6932
+ input: ["text", "image"],
6933
+ cost: {
6934
+ input: 0.25,
6935
+ output: 2,
6936
+ cacheRead: 0,
6937
+ cacheWrite: 0,
6938
+ },
6939
+ contextWindow: 262144,
6940
+ maxTokens: 131072,
6941
+ },
6891
6942
  "bytedance-seed/seed-2.0-mini": {
6892
6943
  id: "bytedance-seed/seed-2.0-mini",
6893
6944
  name: "ByteDance Seed: Seed-2.0-Mini",
@@ -6967,11 +7018,11 @@ export const MODELS = {
6967
7018
  cost: {
6968
7019
  input: 0.19999999999999998,
6969
7020
  output: 0.77,
6970
- cacheRead: 0.13,
7021
+ cacheRead: 0.135,
6971
7022
  cacheWrite: 0,
6972
7023
  },
6973
7024
  contextWindow: 163840,
6974
- maxTokens: 163840,
7025
+ maxTokens: 4096,
6975
7026
  },
6976
7027
  "deepseek/deepseek-chat-v3.1": {
6977
7028
  id: "deepseek/deepseek-chat-v3.1",
@@ -7041,40 +7092,23 @@ export const MODELS = {
7041
7092
  contextWindow: 163840,
7042
7093
  maxTokens: 4096,
7043
7094
  },
7044
- "deepseek/deepseek-v3.1-terminus:exacto": {
7045
- id: "deepseek/deepseek-v3.1-terminus:exacto",
7046
- name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
7095
+ "deepseek/deepseek-v3.2": {
7096
+ id: "deepseek/deepseek-v3.2",
7097
+ name: "DeepSeek: DeepSeek V3.2",
7047
7098
  api: "openai-completions",
7048
7099
  provider: "openrouter",
7049
7100
  baseUrl: "https://openrouter.ai/api/v1",
7050
7101
  reasoning: true,
7051
7102
  input: ["text"],
7052
7103
  cost: {
7053
- input: 0.21,
7054
- output: 0.7899999999999999,
7055
- cacheRead: 0.16799999999999998,
7104
+ input: 0.26,
7105
+ output: 0.38,
7106
+ cacheRead: 0.13,
7056
7107
  cacheWrite: 0,
7057
7108
  },
7058
7109
  contextWindow: 163840,
7059
7110
  maxTokens: 4096,
7060
7111
  },
7061
- "deepseek/deepseek-v3.2": {
7062
- id: "deepseek/deepseek-v3.2",
7063
- name: "DeepSeek: DeepSeek V3.2",
7064
- api: "openai-completions",
7065
- provider: "openrouter",
7066
- baseUrl: "https://openrouter.ai/api/v1",
7067
- reasoning: true,
7068
- input: ["text"],
7069
- cost: {
7070
- input: 0.25,
7071
- output: 0.39999999999999997,
7072
- cacheRead: 0,
7073
- cacheWrite: 0,
7074
- },
7075
- contextWindow: 163840,
7076
- maxTokens: 65536,
7077
- },
7078
7112
  "deepseek/deepseek-v3.2-exp": {
7079
7113
  id: "deepseek/deepseek-v3.2-exp",
7080
7114
  name: "DeepSeek: DeepSeek V3.2 Exp",
@@ -7339,9 +7373,9 @@ export const MODELS = {
7339
7373
  reasoning: false,
7340
7374
  input: ["text", "image"],
7341
7375
  cost: {
7342
- input: 0.04,
7343
- output: 0.15,
7344
- cacheRead: 0.02,
7376
+ input: 0.03,
7377
+ output: 0.11,
7378
+ cacheRead: 0.015,
7345
7379
  cacheWrite: 0,
7346
7380
  },
7347
7381
  contextWindow: 128000,
@@ -7466,23 +7500,6 @@ export const MODELS = {
7466
7500
  contextWindow: 8192,
7467
7501
  maxTokens: 16384,
7468
7502
  },
7469
- "meta-llama/llama-3.1-405b-instruct": {
7470
- id: "meta-llama/llama-3.1-405b-instruct",
7471
- name: "Meta: Llama 3.1 405B Instruct",
7472
- api: "openai-completions",
7473
- provider: "openrouter",
7474
- baseUrl: "https://openrouter.ai/api/v1",
7475
- reasoning: false,
7476
- input: ["text"],
7477
- cost: {
7478
- input: 4,
7479
- output: 4,
7480
- cacheRead: 0,
7481
- cacheWrite: 0,
7482
- },
7483
- contextWindow: 131000,
7484
- maxTokens: 4096,
7485
- },
7486
7503
  "meta-llama/llama-3.1-70b-instruct": {
7487
7504
  id: "meta-llama/llama-3.1-70b-instruct",
7488
7505
  name: "Meta: Llama 3.1 70B Instruct",
@@ -7645,13 +7662,13 @@ export const MODELS = {
7645
7662
  reasoning: true,
7646
7663
  input: ["text"],
7647
7664
  cost: {
7648
- input: 0.295,
7649
- output: 1.2,
7650
- cacheRead: 0.03,
7665
+ input: 0.27,
7666
+ output: 0.95,
7667
+ cacheRead: 0.0299999997,
7651
7668
  cacheWrite: 0,
7652
7669
  },
7653
7670
  contextWindow: 196608,
7654
- maxTokens: 196608,
7671
+ maxTokens: 4096,
7655
7672
  },
7656
7673
  "mistralai/codestral-2508": {
7657
7674
  id: "mistralai/codestral-2508",
@@ -8078,23 +8095,6 @@ export const MODELS = {
8078
8095
  contextWindow: 131072,
8079
8096
  maxTokens: 4096,
8080
8097
  },
8081
- "moonshotai/kimi-k2-0905:exacto": {
8082
- id: "moonshotai/kimi-k2-0905:exacto",
8083
- name: "MoonshotAI: Kimi K2 0905 (exacto)",
8084
- api: "openai-completions",
8085
- provider: "openrouter",
8086
- baseUrl: "https://openrouter.ai/api/v1",
8087
- reasoning: false,
8088
- input: ["text"],
8089
- cost: {
8090
- input: 0.6,
8091
- output: 2.5,
8092
- cacheRead: 0,
8093
- cacheWrite: 0,
8094
- },
8095
- contextWindow: 262144,
8096
- maxTokens: 4096,
8097
- },
8098
8098
  "moonshotai/kimi-k2-thinking": {
8099
8099
  id: "moonshotai/kimi-k2-thinking",
8100
8100
  name: "MoonshotAI: Kimi K2 Thinking",
@@ -8214,6 +8214,23 @@ export const MODELS = {
8214
8214
  contextWindow: 256000,
8215
8215
  maxTokens: 4096,
8216
8216
  },
8217
+ "nvidia/nemotron-3-super-120b-a12b:free": {
8218
+ id: "nvidia/nemotron-3-super-120b-a12b:free",
8219
+ name: "NVIDIA: Nemotron 3 Super (free)",
8220
+ api: "openai-completions",
8221
+ provider: "openrouter",
8222
+ baseUrl: "https://openrouter.ai/api/v1",
8223
+ reasoning: true,
8224
+ input: ["text"],
8225
+ cost: {
8226
+ input: 0,
8227
+ output: 0,
8228
+ cacheRead: 0,
8229
+ cacheWrite: 0,
8230
+ },
8231
+ contextWindow: 262144,
8232
+ maxTokens: 262144,
8233
+ },
8217
8234
  "nvidia/nemotron-nano-12b-v2-vl:free": {
8218
8235
  id: "nvidia/nemotron-nano-12b-v2-vl:free",
8219
8236
  name: "NVIDIA: Nemotron Nano 12B 2 VL (free)",
@@ -8945,23 +8962,6 @@ export const MODELS = {
8945
8962
  contextWindow: 131072,
8946
8963
  maxTokens: 4096,
8947
8964
  },
8948
- "openai/gpt-oss-120b:exacto": {
8949
- id: "openai/gpt-oss-120b:exacto",
8950
- name: "OpenAI: gpt-oss-120b (exacto)",
8951
- api: "openai-completions",
8952
- provider: "openrouter",
8953
- baseUrl: "https://openrouter.ai/api/v1",
8954
- reasoning: true,
8955
- input: ["text"],
8956
- cost: {
8957
- input: 0.039,
8958
- output: 0.19,
8959
- cacheRead: 0,
8960
- cacheWrite: 0,
8961
- },
8962
- contextWindow: 131072,
8963
- maxTokens: 4096,
8964
- },
8965
8965
  "openai/gpt-oss-120b:free": {
8966
8966
  id: "openai/gpt-oss-120b:free",
8967
8967
  name: "OpenAI: gpt-oss-120b (free)",
@@ -9217,6 +9217,40 @@ export const MODELS = {
9217
9217
  contextWindow: 200000,
9218
9218
  maxTokens: 4096,
9219
9219
  },
9220
+ "openrouter/healer-alpha": {
9221
+ id: "openrouter/healer-alpha",
9222
+ name: "Healer Alpha",
9223
+ api: "openai-completions",
9224
+ provider: "openrouter",
9225
+ baseUrl: "https://openrouter.ai/api/v1",
9226
+ reasoning: true,
9227
+ input: ["text", "image"],
9228
+ cost: {
9229
+ input: 0,
9230
+ output: 0,
9231
+ cacheRead: 0,
9232
+ cacheWrite: 0,
9233
+ },
9234
+ contextWindow: 262144,
9235
+ maxTokens: 32000,
9236
+ },
9237
+ "openrouter/hunter-alpha": {
9238
+ id: "openrouter/hunter-alpha",
9239
+ name: "Hunter Alpha",
9240
+ api: "openai-completions",
9241
+ provider: "openrouter",
9242
+ baseUrl: "https://openrouter.ai/api/v1",
9243
+ reasoning: true,
9244
+ input: ["text"],
9245
+ cost: {
9246
+ input: 0,
9247
+ output: 0,
9248
+ cacheRead: 0,
9249
+ cacheWrite: 0,
9250
+ },
9251
+ contextWindow: 1048576,
9252
+ maxTokens: 32000,
9253
+ },
9220
9254
  "prime-intellect/intellect-3": {
9221
9255
  id: "prime-intellect/intellect-3",
9222
9256
  name: "Prime Intellect: INTELLECT-3",
@@ -9625,23 +9659,6 @@ export const MODELS = {
9625
9659
  contextWindow: 1000000,
9626
9660
  maxTokens: 65536,
9627
9661
  },
9628
- "qwen/qwen3-coder:exacto": {
9629
- id: "qwen/qwen3-coder:exacto",
9630
- name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
9631
- api: "openai-completions",
9632
- provider: "openrouter",
9633
- baseUrl: "https://openrouter.ai/api/v1",
9634
- reasoning: false,
9635
- input: ["text"],
9636
- cost: {
9637
- input: 0.22,
9638
- output: 1.7999999999999998,
9639
- cacheRead: 0.022,
9640
- cacheWrite: 0,
9641
- },
9642
- contextWindow: 262144,
9643
- maxTokens: 65536,
9644
- },
9645
9662
  "qwen/qwen3-coder:free": {
9646
9663
  id: "qwen/qwen3-coder:free",
9647
9664
  name: "Qwen: Qwen3 Coder 480B A35B (free)",
@@ -9707,7 +9724,7 @@ export const MODELS = {
9707
9724
  cacheRead: 0,
9708
9725
  cacheWrite: 0,
9709
9726
  },
9710
- contextWindow: 262144,
9727
+ contextWindow: 131072,
9711
9728
  maxTokens: 4096,
9712
9729
  },
9713
9730
  "qwen/qwen3-next-80b-a3b-instruct:free": {
@@ -9736,13 +9753,13 @@ export const MODELS = {
9736
9753
  reasoning: true,
9737
9754
  input: ["text"],
9738
9755
  cost: {
9739
- input: 0.15,
9740
- output: 1.2,
9756
+ input: 0.0975,
9757
+ output: 0.78,
9741
9758
  cacheRead: 0,
9742
9759
  cacheWrite: 0,
9743
9760
  },
9744
- contextWindow: 128000,
9745
- maxTokens: 4096,
9761
+ contextWindow: 131072,
9762
+ maxTokens: 32768,
9746
9763
  },
9747
9764
  "qwen/qwen3-vl-235b-a22b-instruct": {
9748
9765
  id: "qwen/qwen3-vl-235b-a22b-instruct",
@@ -9770,8 +9787,8 @@ export const MODELS = {
9770
9787
  reasoning: true,
9771
9788
  input: ["text", "image"],
9772
9789
  cost: {
9773
- input: 0,
9774
- output: 0,
9790
+ input: 0.26,
9791
+ output: 2.6,
9775
9792
  cacheRead: 0,
9776
9793
  cacheWrite: 0,
9777
9794
  },
@@ -9804,8 +9821,8 @@ export const MODELS = {
9804
9821
  reasoning: true,
9805
9822
  input: ["text", "image"],
9806
9823
  cost: {
9807
- input: 0,
9808
- output: 0,
9824
+ input: 0.13,
9825
+ output: 1.56,
9809
9826
  cacheRead: 0,
9810
9827
  cacheWrite: 0,
9811
9828
  },
@@ -9931,6 +9948,23 @@ export const MODELS = {
9931
9948
  contextWindow: 262144,
9932
9949
  maxTokens: 65536,
9933
9950
  },
9951
+ "qwen/qwen3.5-9b": {
9952
+ id: "qwen/qwen3.5-9b",
9953
+ name: "Qwen: Qwen3.5-9B",
9954
+ api: "openai-completions",
9955
+ provider: "openrouter",
9956
+ baseUrl: "https://openrouter.ai/api/v1",
9957
+ reasoning: true,
9958
+ input: ["text", "image"],
9959
+ cost: {
9960
+ input: 0.049999999999999996,
9961
+ output: 0.15,
9962
+ cacheRead: 0,
9963
+ cacheWrite: 0,
9964
+ },
9965
+ contextWindow: 256000,
9966
+ maxTokens: 4096,
9967
+ },
9934
9968
  "qwen/qwen3.5-flash-02-23": {
9935
9969
  id: "qwen/qwen3.5-flash-02-23",
9936
9970
  name: "Qwen: Qwen3.5-Flash",
@@ -10025,13 +10059,13 @@ export const MODELS = {
10025
10059
  reasoning: false,
10026
10060
  input: ["text"],
10027
10061
  cost: {
10028
- input: 0.65,
10029
- output: 0.75,
10062
+ input: 0.85,
10063
+ output: 0.85,
10030
10064
  cacheRead: 0,
10031
10065
  cacheWrite: 0,
10032
10066
  },
10033
- contextWindow: 32768,
10034
- maxTokens: 32768,
10067
+ contextWindow: 131072,
10068
+ maxTokens: 16384,
10035
10069
  },
10036
10070
  "stepfun/step-3.5-flash": {
10037
10071
  id: "stepfun/step-3.5-flash",
@@ -10254,6 +10288,23 @@ export const MODELS = {
10254
10288
  contextWindow: 2000000,
10255
10289
  maxTokens: 30000,
10256
10290
  },
10291
+ "x-ai/grok-4.20-beta": {
10292
+ id: "x-ai/grok-4.20-beta",
10293
+ name: "xAI: Grok 4.20 Beta",
10294
+ api: "openai-completions",
10295
+ provider: "openrouter",
10296
+ baseUrl: "https://openrouter.ai/api/v1",
10297
+ reasoning: true,
10298
+ input: ["text", "image"],
10299
+ cost: {
10300
+ input: 2,
10301
+ output: 6,
10302
+ cacheRead: 0.19999999999999998,
10303
+ cacheWrite: 0,
10304
+ },
10305
+ contextWindow: 2000000,
10306
+ maxTokens: 4096,
10307
+ },
10257
10308
  "x-ai/grok-code-fast-1": {
10258
10309
  id: "x-ai/grok-code-fast-1",
10259
10310
  name: "xAI: Grok Code Fast 1",
@@ -10390,23 +10441,6 @@ export const MODELS = {
10390
10441
  contextWindow: 204800,
10391
10442
  maxTokens: 204800,
10392
10443
  },
10393
- "z-ai/glm-4.6:exacto": {
10394
- id: "z-ai/glm-4.6:exacto",
10395
- name: "Z.ai: GLM 4.6 (exacto)",
10396
- api: "openai-completions",
10397
- provider: "openrouter",
10398
- baseUrl: "https://openrouter.ai/api/v1",
10399
- reasoning: true,
10400
- input: ["text"],
10401
- cost: {
10402
- input: 0.44,
10403
- output: 1.76,
10404
- cacheRead: 0.11,
10405
- cacheWrite: 0,
10406
- },
10407
- contextWindow: 204800,
10408
- maxTokens: 131072,
10409
- },
10410
10444
  "z-ai/glm-4.6v": {
10411
10445
  id: "z-ai/glm-4.6v",
10412
10446
  name: "Z.ai: GLM 4.6V",
@@ -10473,7 +10507,7 @@ export const MODELS = {
10473
10507
  cacheWrite: 0,
10474
10508
  },
10475
10509
  contextWindow: 202752,
10476
- maxTokens: 4096,
10510
+ maxTokens: 131072,
10477
10511
  },
10478
10512
  },
10479
10513
  "vercel-ai-gateway": {
@@ -10537,13 +10571,13 @@ export const MODELS = {
10537
10571
  reasoning: true,
10538
10572
  input: ["text"],
10539
10573
  cost: {
10540
- input: 0.09999999999999999,
10541
- output: 0.3,
10574
+ input: 0.29,
10575
+ output: 0.59,
10542
10576
  cacheRead: 0,
10543
10577
  cacheWrite: 0,
10544
10578
  },
10545
- contextWindow: 40960,
10546
- maxTokens: 16384,
10579
+ contextWindow: 131072,
10580
+ maxTokens: 40960,
10547
10581
  },
10548
10582
  "alibaba/qwen3-235b-a22b-thinking": {
10549
10583
  id: "alibaba/qwen3-235b-a22b-thinking",
@@ -10588,13 +10622,13 @@ export const MODELS = {
10588
10622
  reasoning: true,
10589
10623
  input: ["text"],
10590
10624
  cost: {
10591
- input: 0.07,
10592
- output: 0.27,
10625
+ input: 0.15,
10626
+ output: 0.6,
10593
10627
  cacheRead: 0,
10594
10628
  cacheWrite: 0,
10595
10629
  },
10596
- contextWindow: 160000,
10597
- maxTokens: 32768,
10630
+ contextWindow: 262144,
10631
+ maxTokens: 8192,
10598
10632
  },
10599
10633
  "alibaba/qwen3-coder-next": {
10600
10634
  id: "alibaba/qwen3-coder-next",
@@ -10602,7 +10636,7 @@ export const MODELS = {
10602
10636
  api: "anthropic-messages",
10603
10637
  provider: "vercel-ai-gateway",
10604
10638
  baseUrl: "https://ai-gateway.vercel.sh",
10605
- reasoning: true,
10639
+ reasoning: false,
10606
10640
  input: ["text"],
10607
10641
  cost: {
10608
10642
  input: 0.5,
@@ -10630,6 +10664,23 @@ export const MODELS = {
10630
10664
  contextWindow: 1000000,
10631
10665
  maxTokens: 65536,
10632
10666
  },
10667
+ "alibaba/qwen3-max": {
10668
+ id: "alibaba/qwen3-max",
10669
+ name: "Qwen3 Max",
10670
+ api: "anthropic-messages",
10671
+ provider: "vercel-ai-gateway",
10672
+ baseUrl: "https://ai-gateway.vercel.sh",
10673
+ reasoning: false,
10674
+ input: ["text"],
10675
+ cost: {
10676
+ input: 1.2,
10677
+ output: 6,
10678
+ cacheRead: 0.24,
10679
+ cacheWrite: 0,
10680
+ },
10681
+ contextWindow: 262144,
10682
+ maxTokens: 32768,
10683
+ },
10633
10684
  "alibaba/qwen3-max-preview": {
10634
10685
  id: "alibaba/qwen3-max-preview",
10635
10686
  name: "Qwen3 Max Preview",
@@ -10777,8 +10828,8 @@ export const MODELS = {
10777
10828
  cost: {
10778
10829
  input: 3,
10779
10830
  output: 15,
10780
- cacheRead: 0,
10781
- cacheWrite: 0,
10831
+ cacheRead: 0.3,
10832
+ cacheWrite: 3.75,
10782
10833
  },
10783
10834
  contextWindow: 200000,
10784
10835
  maxTokens: 8192,
@@ -10987,6 +11038,23 @@ export const MODELS = {
10987
11038
  contextWindow: 256000,
10988
11039
  maxTokens: 8000,
10989
11040
  },
11041
+ "deepseek/deepseek-r1": {
11042
+ id: "deepseek/deepseek-r1",
11043
+ name: "DeepSeek-R1",
11044
+ api: "anthropic-messages",
11045
+ provider: "vercel-ai-gateway",
11046
+ baseUrl: "https://ai-gateway.vercel.sh",
11047
+ reasoning: true,
11048
+ input: ["text"],
11049
+ cost: {
11050
+ input: 1.35,
11051
+ output: 5.4,
11052
+ cacheRead: 0,
11053
+ cacheWrite: 0,
11054
+ },
11055
+ contextWindow: 128000,
11056
+ maxTokens: 8192,
11057
+ },
10990
11058
  "deepseek/deepseek-v3": {
10991
11059
  id: "deepseek/deepseek-v3",
10992
11060
  name: "DeepSeek V3 0324",
@@ -11013,13 +11081,13 @@ export const MODELS = {
11013
11081
  reasoning: true,
11014
11082
  input: ["text"],
11015
11083
  cost: {
11016
- input: 0.21,
11017
- output: 0.7899999999999999,
11084
+ input: 0.5,
11085
+ output: 1.5,
11018
11086
  cacheRead: 0,
11019
11087
  cacheWrite: 0,
11020
11088
  },
11021
11089
  contextWindow: 163840,
11022
- maxTokens: 128000,
11090
+ maxTokens: 16384,
11023
11091
  },
11024
11092
  "deepseek/deepseek-v3.1-terminus": {
11025
11093
  id: "deepseek/deepseek-v3.1-terminus",
@@ -11072,18 +11140,52 @@ export const MODELS = {
11072
11140
  contextWindow: 128000,
11073
11141
  maxTokens: 64000,
11074
11142
  },
11075
- "google/gemini-2.5-flash": {
11076
- id: "google/gemini-2.5-flash",
11143
+ "google/gemini-2.0-flash": {
11144
+ id: "google/gemini-2.0-flash",
11145
+ name: "Gemini 2.0 Flash",
11146
+ api: "anthropic-messages",
11147
+ provider: "vercel-ai-gateway",
11148
+ baseUrl: "https://ai-gateway.vercel.sh",
11149
+ reasoning: false,
11150
+ input: ["text", "image"],
11151
+ cost: {
11152
+ input: 0.15,
11153
+ output: 0.6,
11154
+ cacheRead: 0,
11155
+ cacheWrite: 0,
11156
+ },
11157
+ contextWindow: 1048576,
11158
+ maxTokens: 8192,
11159
+ },
11160
+ "google/gemini-2.0-flash-lite": {
11161
+ id: "google/gemini-2.0-flash-lite",
11162
+ name: "Gemini 2.0 Flash Lite",
11163
+ api: "anthropic-messages",
11164
+ provider: "vercel-ai-gateway",
11165
+ baseUrl: "https://ai-gateway.vercel.sh",
11166
+ reasoning: false,
11167
+ input: ["text", "image"],
11168
+ cost: {
11169
+ input: 0.075,
11170
+ output: 0.3,
11171
+ cacheRead: 0,
11172
+ cacheWrite: 0,
11173
+ },
11174
+ contextWindow: 1048576,
11175
+ maxTokens: 8192,
11176
+ },
11177
+ "google/gemini-2.5-flash": {
11178
+ id: "google/gemini-2.5-flash",
11077
11179
  name: "Gemini 2.5 Flash",
11078
11180
  api: "anthropic-messages",
11079
11181
  provider: "vercel-ai-gateway",
11080
11182
  baseUrl: "https://ai-gateway.vercel.sh",
11081
11183
  reasoning: true,
11082
- input: ["text"],
11184
+ input: ["text", "image"],
11083
11185
  cost: {
11084
11186
  input: 0.3,
11085
11187
  output: 2.5,
11086
- cacheRead: 0,
11188
+ cacheRead: 0.03,
11087
11189
  cacheWrite: 0,
11088
11190
  },
11089
11191
  contextWindow: 1000000,
@@ -11147,11 +11249,11 @@ export const MODELS = {
11147
11249
  provider: "vercel-ai-gateway",
11148
11250
  baseUrl: "https://ai-gateway.vercel.sh",
11149
11251
  reasoning: true,
11150
- input: ["text"],
11252
+ input: ["text", "image"],
11151
11253
  cost: {
11152
11254
  input: 1.25,
11153
11255
  output: 10,
11154
- cacheRead: 0,
11256
+ cacheRead: 0.125,
11155
11257
  cacheWrite: 0,
11156
11258
  },
11157
11259
  contextWindow: 1048576,
@@ -11172,7 +11274,7 @@ export const MODELS = {
11172
11274
  cacheWrite: 0,
11173
11275
  },
11174
11276
  contextWindow: 1000000,
11175
- maxTokens: 64000,
11277
+ maxTokens: 65000,
11176
11278
  },
11177
11279
  "google/gemini-3-pro-preview": {
11178
11280
  id: "google/gemini-3-pro-preview",
@@ -11274,7 +11376,7 @@ export const MODELS = {
11274
11376
  cacheWrite: 0,
11275
11377
  },
11276
11378
  contextWindow: 128000,
11277
- maxTokens: 8192,
11379
+ maxTokens: 100000,
11278
11380
  },
11279
11381
  "meituan/longcat-flash-thinking": {
11280
11382
  id: "meituan/longcat-flash-thinking",
@@ -11302,13 +11404,13 @@ export const MODELS = {
11302
11404
  reasoning: false,
11303
11405
  input: ["text"],
11304
11406
  cost: {
11305
- input: 0.39999999999999997,
11306
- output: 0.39999999999999997,
11407
+ input: 0.72,
11408
+ output: 0.72,
11307
11409
  cacheRead: 0,
11308
11410
  cacheWrite: 0,
11309
11411
  },
11310
- contextWindow: 131072,
11311
- maxTokens: 16384,
11412
+ contextWindow: 128000,
11413
+ maxTokens: 8192,
11312
11414
  },
11313
11415
  "meta/llama-3.1-8b": {
11314
11416
  id: "meta/llama-3.1-8b",
@@ -11319,12 +11421,12 @@ export const MODELS = {
11319
11421
  reasoning: false,
11320
11422
  input: ["text"],
11321
11423
  cost: {
11322
- input: 0.03,
11323
- output: 0.049999999999999996,
11424
+ input: 0.09999999999999999,
11425
+ output: 0.09999999999999999,
11324
11426
  cacheRead: 0,
11325
11427
  cacheWrite: 0,
11326
11428
  },
11327
- contextWindow: 131072,
11429
+ contextWindow: 128000,
11328
11430
  maxTokens: 16384,
11329
11431
  },
11330
11432
  "meta/llama-3.2-11b": {
@@ -11387,12 +11489,12 @@ export const MODELS = {
11387
11489
  reasoning: false,
11388
11490
  input: ["text", "image"],
11389
11491
  cost: {
11390
- input: 0.15,
11391
- output: 0.6,
11492
+ input: 0.24,
11493
+ output: 0.9700000000000001,
11392
11494
  cacheRead: 0,
11393
11495
  cacheWrite: 0,
11394
11496
  },
11395
- contextWindow: 131072,
11497
+ contextWindow: 128000,
11396
11498
  maxTokens: 8192,
11397
11499
  },
11398
11500
  "meta/llama-4-scout": {
@@ -11404,12 +11506,12 @@ export const MODELS = {
11404
11506
  reasoning: false,
11405
11507
  input: ["text", "image"],
11406
11508
  cost: {
11407
- input: 0.08,
11408
- output: 0.3,
11509
+ input: 0.16999999999999998,
11510
+ output: 0.66,
11409
11511
  cacheRead: 0,
11410
11512
  cacheWrite: 0,
11411
11513
  },
11412
- contextWindow: 131072,
11514
+ contextWindow: 128000,
11413
11515
  maxTokens: 8192,
11414
11516
  },
11415
11517
  "minimax/minimax-m2": {
@@ -11440,8 +11542,8 @@ export const MODELS = {
11440
11542
  cost: {
11441
11543
  input: 0.3,
11442
11544
  output: 1.2,
11443
- cacheRead: 0.15,
11444
- cacheWrite: 0,
11545
+ cacheRead: 0.03,
11546
+ cacheWrite: 0.375,
11445
11547
  },
11446
11548
  contextWindow: 204800,
11447
11549
  maxTokens: 131072,
@@ -11676,14 +11778,31 @@ export const MODELS = {
11676
11778
  reasoning: false,
11677
11779
  input: ["text"],
11678
11780
  cost: {
11679
- input: 0.5,
11680
- output: 2,
11781
+ input: 0.6,
11782
+ output: 2.5,
11681
11783
  cacheRead: 0,
11682
11784
  cacheWrite: 0,
11683
11785
  },
11684
11786
  contextWindow: 131072,
11685
11787
  maxTokens: 16384,
11686
11788
  },
11789
+ "moonshotai/kimi-k2-0905": {
11790
+ id: "moonshotai/kimi-k2-0905",
11791
+ name: "Kimi K2 0905",
11792
+ api: "anthropic-messages",
11793
+ provider: "vercel-ai-gateway",
11794
+ baseUrl: "https://ai-gateway.vercel.sh",
11795
+ reasoning: false,
11796
+ input: ["text"],
11797
+ cost: {
11798
+ input: 0.6,
11799
+ output: 2.5,
11800
+ cacheRead: 0,
11801
+ cacheWrite: 0,
11802
+ },
11803
+ contextWindow: 256000,
11804
+ maxTokens: 16384,
11805
+ },
11687
11806
  "moonshotai/kimi-k2-thinking": {
11688
11807
  id: "moonshotai/kimi-k2-thinking",
11689
11808
  name: "Kimi K2 Thinking",
@@ -11693,13 +11812,13 @@ export const MODELS = {
11693
11812
  reasoning: true,
11694
11813
  input: ["text"],
11695
11814
  cost: {
11696
- input: 0.47,
11697
- output: 2,
11698
- cacheRead: 0.14100000000000001,
11815
+ input: 0.6,
11816
+ output: 2.5,
11817
+ cacheRead: 0.15,
11699
11818
  cacheWrite: 0,
11700
11819
  },
11701
- contextWindow: 216144,
11702
- maxTokens: 216144,
11820
+ contextWindow: 262114,
11821
+ maxTokens: 262114,
11703
11822
  },
11704
11823
  "moonshotai/kimi-k2-thinking-turbo": {
11705
11824
  id: "moonshotai/kimi-k2-thinking-turbo",
@@ -11744,13 +11863,13 @@ export const MODELS = {
11744
11863
  reasoning: true,
11745
11864
  input: ["text", "image"],
11746
11865
  cost: {
11747
- input: 0.5,
11748
- output: 2.8,
11749
- cacheRead: 0,
11866
+ input: 0.6,
11867
+ output: 3,
11868
+ cacheRead: 0.09999999999999999,
11750
11869
  cacheWrite: 0,
11751
11870
  },
11752
- contextWindow: 256000,
11753
- maxTokens: 256000,
11871
+ contextWindow: 262114,
11872
+ maxTokens: 262114,
11754
11873
  },
11755
11874
  "nvidia/nemotron-nano-12b-v2-vl": {
11756
11875
  id: "nvidia/nemotron-nano-12b-v2-vl",
@@ -11778,31 +11897,14 @@ export const MODELS = {
11778
11897
  reasoning: true,
11779
11898
  input: ["text"],
11780
11899
  cost: {
11781
- input: 0.04,
11782
- output: 0.16,
11900
+ input: 0.06,
11901
+ output: 0.22999999999999998,
11783
11902
  cacheRead: 0,
11784
11903
  cacheWrite: 0,
11785
11904
  },
11786
11905
  contextWindow: 131072,
11787
11906
  maxTokens: 131072,
11788
11907
  },
11789
- "openai/codex-mini": {
11790
- id: "openai/codex-mini",
11791
- name: "Codex Mini",
11792
- api: "anthropic-messages",
11793
- provider: "vercel-ai-gateway",
11794
- baseUrl: "https://ai-gateway.vercel.sh",
11795
- reasoning: true,
11796
- input: ["text", "image"],
11797
- cost: {
11798
- input: 1.5,
11799
- output: 6,
11800
- cacheRead: 0.375,
11801
- cacheWrite: 0,
11802
- },
11803
- contextWindow: 200000,
11804
- maxTokens: 100000,
11805
- },
11806
11908
  "openai/gpt-4-turbo": {
11807
11909
  id: "openai/gpt-4-turbo",
11808
11910
  name: "GPT-4 Turbo",
@@ -11865,7 +11967,7 @@ export const MODELS = {
11865
11967
  cost: {
11866
11968
  input: 0.09999999999999999,
11867
11969
  output: 0.39999999999999997,
11868
- cacheRead: 0.03,
11970
+ cacheRead: 0.024999999999999998,
11869
11971
  cacheWrite: 0,
11870
11972
  },
11871
11973
  contextWindow: 1047576,
@@ -11916,7 +12018,7 @@ export const MODELS = {
11916
12018
  cost: {
11917
12019
  input: 1.25,
11918
12020
  output: 10,
11919
- cacheRead: 0.13,
12021
+ cacheRead: 0.125,
11920
12022
  cacheWrite: 0,
11921
12023
  },
11922
12024
  contextWindow: 400000,
@@ -11946,11 +12048,11 @@ export const MODELS = {
11946
12048
  provider: "vercel-ai-gateway",
11947
12049
  baseUrl: "https://ai-gateway.vercel.sh",
11948
12050
  reasoning: true,
11949
- input: ["text", "image"],
12051
+ input: ["text"],
11950
12052
  cost: {
11951
12053
  input: 1.25,
11952
12054
  output: 10,
11953
- cacheRead: 0.13,
12055
+ cacheRead: 0.125,
11954
12056
  cacheWrite: 0,
11955
12057
  },
11956
12058
  contextWindow: 400000,
@@ -11967,7 +12069,7 @@ export const MODELS = {
11967
12069
  cost: {
11968
12070
  input: 0.25,
11969
12071
  output: 2,
11970
- cacheRead: 0.03,
12072
+ cacheRead: 0.024999999999999998,
11971
12073
  cacheWrite: 0,
11972
12074
  },
11973
12075
  contextWindow: 400000,
@@ -11984,7 +12086,7 @@ export const MODELS = {
11984
12086
  cost: {
11985
12087
  input: 0.049999999999999996,
11986
12088
  output: 0.39999999999999997,
11987
- cacheRead: 0.01,
12089
+ cacheRead: 0.005,
11988
12090
  cacheWrite: 0,
11989
12091
  },
11990
12092
  contextWindow: 400000,
@@ -12018,7 +12120,7 @@ export const MODELS = {
12018
12120
  cost: {
12019
12121
  input: 1.25,
12020
12122
  output: 10,
12021
- cacheRead: 0.13,
12123
+ cacheRead: 0.125,
12022
12124
  cacheWrite: 0,
12023
12125
  },
12024
12126
  contextWindow: 400000,
@@ -12069,7 +12171,7 @@ export const MODELS = {
12069
12171
  cost: {
12070
12172
  input: 1.25,
12071
12173
  output: 10,
12072
- cacheRead: 0.13,
12174
+ cacheRead: 0.125,
12073
12175
  cacheWrite: 0,
12074
12176
  },
12075
12177
  contextWindow: 128000,
@@ -12086,7 +12188,7 @@ export const MODELS = {
12086
12188
  cost: {
12087
12189
  input: 1.25,
12088
12190
  output: 10,
12089
- cacheRead: 0.13,
12191
+ cacheRead: 0.125,
12090
12192
  cacheWrite: 0,
12091
12193
  },
12092
12194
  contextWindow: 400000,
@@ -12103,7 +12205,7 @@ export const MODELS = {
12103
12205
  cost: {
12104
12206
  input: 1.75,
12105
12207
  output: 14,
12106
- cacheRead: 0.18,
12208
+ cacheRead: 0.175,
12107
12209
  cacheWrite: 0,
12108
12210
  },
12109
12211
  contextWindow: 400000,
@@ -12208,7 +12310,7 @@ export const MODELS = {
12208
12310
  cacheRead: 0.25,
12209
12311
  cacheWrite: 0,
12210
12312
  },
12211
- contextWindow: 200000,
12313
+ contextWindow: 1050000,
12212
12314
  maxTokens: 128000,
12213
12315
  },
12214
12316
  "openai/gpt-5.4-pro": {
@@ -12225,26 +12327,9 @@ export const MODELS = {
12225
12327
  cacheRead: 0,
12226
12328
  cacheWrite: 0,
12227
12329
  },
12228
- contextWindow: 200000,
12330
+ contextWindow: 1050000,
12229
12331
  maxTokens: 128000,
12230
12332
  },
12231
- "openai/gpt-oss-120b": {
12232
- id: "openai/gpt-oss-120b",
12233
- name: "gpt-oss-120b",
12234
- api: "anthropic-messages",
12235
- provider: "vercel-ai-gateway",
12236
- baseUrl: "https://ai-gateway.vercel.sh",
12237
- reasoning: true,
12238
- input: ["text"],
12239
- cost: {
12240
- input: 0.09999999999999999,
12241
- output: 0.5,
12242
- cacheRead: 0,
12243
- cacheWrite: 0,
12244
- },
12245
- contextWindow: 131072,
12246
- maxTokens: 131072,
12247
- },
12248
12333
  "openai/gpt-oss-20b": {
12249
12334
  id: "openai/gpt-oss-20b",
12250
12335
  name: "gpt-oss-20b",
@@ -12562,7 +12647,7 @@ export const MODELS = {
12562
12647
  cost: {
12563
12648
  input: 3,
12564
12649
  output: 15,
12565
- cacheRead: 0,
12650
+ cacheRead: 0.75,
12566
12651
  cacheWrite: 0,
12567
12652
  },
12568
12653
  contextWindow: 256000,
@@ -12636,6 +12721,57 @@ export const MODELS = {
12636
12721
  contextWindow: 2000000,
12637
12722
  maxTokens: 30000,
12638
12723
  },
12724
+ "xai/grok-4.20-multi-agent-beta": {
12725
+ id: "xai/grok-4.20-multi-agent-beta",
12726
+ name: "Grok 4.20 Multi Agent Beta",
12727
+ api: "anthropic-messages",
12728
+ provider: "vercel-ai-gateway",
12729
+ baseUrl: "https://ai-gateway.vercel.sh",
12730
+ reasoning: true,
12731
+ input: ["text"],
12732
+ cost: {
12733
+ input: 2,
12734
+ output: 6,
12735
+ cacheRead: 0.19999999999999998,
12736
+ cacheWrite: 0,
12737
+ },
12738
+ contextWindow: 2000000,
12739
+ maxTokens: 2000000,
12740
+ },
12741
+ "xai/grok-4.20-non-reasoning-beta": {
12742
+ id: "xai/grok-4.20-non-reasoning-beta",
12743
+ name: "Grok 4.20 Beta Non-Reasoning",
12744
+ api: "anthropic-messages",
12745
+ provider: "vercel-ai-gateway",
12746
+ baseUrl: "https://ai-gateway.vercel.sh",
12747
+ reasoning: false,
12748
+ input: ["text", "image"],
12749
+ cost: {
12750
+ input: 2,
12751
+ output: 6,
12752
+ cacheRead: 0.19999999999999998,
12753
+ cacheWrite: 0,
12754
+ },
12755
+ contextWindow: 2000000,
12756
+ maxTokens: 2000000,
12757
+ },
12758
+ "xai/grok-4.20-reasoning-beta": {
12759
+ id: "xai/grok-4.20-reasoning-beta",
12760
+ name: "Grok 4.20 Beta Reasoning",
12761
+ api: "anthropic-messages",
12762
+ provider: "vercel-ai-gateway",
12763
+ baseUrl: "https://ai-gateway.vercel.sh",
12764
+ reasoning: true,
12765
+ input: ["text", "image"],
12766
+ cost: {
12767
+ input: 2,
12768
+ output: 6,
12769
+ cacheRead: 0.19999999999999998,
12770
+ cacheWrite: 0,
12771
+ },
12772
+ contextWindow: 2000000,
12773
+ maxTokens: 2000000,
12774
+ },
12639
12775
  "xai/grok-code-fast-1": {
12640
12776
  id: "xai/grok-code-fast-1",
12641
12777
  name: "Grok Code Fast 1",
@@ -12662,9 +12798,9 @@ export const MODELS = {
12662
12798
  reasoning: true,
12663
12799
  input: ["text"],
12664
12800
  cost: {
12665
- input: 0.09,
12666
- output: 0.29,
12667
- cacheRead: 0,
12801
+ input: 0.09999999999999999,
12802
+ output: 0.3,
12803
+ cacheRead: 0.02,
12668
12804
  cacheWrite: 0,
12669
12805
  },
12670
12806
  contextWindow: 262144,
@@ -12681,11 +12817,11 @@ export const MODELS = {
12681
12817
  cost: {
12682
12818
  input: 0.6,
12683
12819
  output: 2.2,
12684
- cacheRead: 0,
12820
+ cacheRead: 0.11,
12685
12821
  cacheWrite: 0,
12686
12822
  },
12687
- contextWindow: 131072,
12688
- maxTokens: 131072,
12823
+ contextWindow: 128000,
12824
+ maxTokens: 96000,
12689
12825
  },
12690
12826
  "zai/glm-4.5-air": {
12691
12827
  id: "zai/glm-4.5-air",
@@ -12710,16 +12846,16 @@ export const MODELS = {
12710
12846
  api: "anthropic-messages",
12711
12847
  provider: "vercel-ai-gateway",
12712
12848
  baseUrl: "https://ai-gateway.vercel.sh",
12713
- reasoning: true,
12849
+ reasoning: false,
12714
12850
  input: ["text", "image"],
12715
12851
  cost: {
12716
12852
  input: 0.6,
12717
12853
  output: 1.7999999999999998,
12718
- cacheRead: 0,
12854
+ cacheRead: 0.11,
12719
12855
  cacheWrite: 0,
12720
12856
  },
12721
- contextWindow: 65536,
12722
- maxTokens: 16384,
12857
+ contextWindow: 66000,
12858
+ maxTokens: 16000,
12723
12859
  },
12724
12860
  "zai/glm-4.6": {
12725
12861
  id: "zai/glm-4.6",
@@ -12781,14 +12917,31 @@ export const MODELS = {
12781
12917
  reasoning: true,
12782
12918
  input: ["text"],
12783
12919
  cost: {
12784
- input: 0.43,
12785
- output: 1.75,
12786
- cacheRead: 0.08,
12920
+ input: 0.6,
12921
+ output: 2.2,
12922
+ cacheRead: 0,
12787
12923
  cacheWrite: 0,
12788
12924
  },
12789
- contextWindow: 202752,
12925
+ contextWindow: 200000,
12790
12926
  maxTokens: 120000,
12791
12927
  },
12928
+ "zai/glm-4.7-flash": {
12929
+ id: "zai/glm-4.7-flash",
12930
+ name: "GLM 4.7 Flash",
12931
+ api: "anthropic-messages",
12932
+ provider: "vercel-ai-gateway",
12933
+ baseUrl: "https://ai-gateway.vercel.sh",
12934
+ reasoning: true,
12935
+ input: ["text"],
12936
+ cost: {
12937
+ input: 0.07,
12938
+ output: 0.39999999999999997,
12939
+ cacheRead: 0,
12940
+ cacheWrite: 0,
12941
+ },
12942
+ contextWindow: 200000,
12943
+ maxTokens: 131000,
12944
+ },
12792
12945
  "zai/glm-4.7-flashx": {
12793
12946
  id: "zai/glm-4.7-flashx",
12794
12947
  name: "GLM 4.7 FlashX",
@@ -12815,9 +12968,9 @@ export const MODELS = {
12815
12968
  reasoning: true,
12816
12969
  input: ["text"],
12817
12970
  cost: {
12818
- input: 0.7999999999999999,
12819
- output: 2.56,
12820
- cacheRead: 0.16,
12971
+ input: 1,
12972
+ output: 3.1999999999999997,
12973
+ cacheRead: 0.19999999999999998,
12821
12974
  cacheWrite: 0,
12822
12975
  },
12823
12976
  contextWindow: 202800,
@@ -13148,6 +13301,57 @@ export const MODELS = {
13148
13301
  contextWindow: 2000000,
13149
13302
  maxTokens: 30000,
13150
13303
  },
13304
+ "grok-4.20-beta-latest-non-reasoning": {
13305
+ id: "grok-4.20-beta-latest-non-reasoning",
13306
+ name: "Grok 4.20 Beta (Non-Reasoning)",
13307
+ api: "openai-completions",
13308
+ provider: "xai",
13309
+ baseUrl: "https://api.x.ai/v1",
13310
+ reasoning: false,
13311
+ input: ["text", "image"],
13312
+ cost: {
13313
+ input: 2,
13314
+ output: 6,
13315
+ cacheRead: 0.2,
13316
+ cacheWrite: 0,
13317
+ },
13318
+ contextWindow: 2000000,
13319
+ maxTokens: 30000,
13320
+ },
13321
+ "grok-4.20-beta-latest-reasoning": {
13322
+ id: "grok-4.20-beta-latest-reasoning",
13323
+ name: "Grok 4.20 Beta (Reasoning)",
13324
+ api: "openai-completions",
13325
+ provider: "xai",
13326
+ baseUrl: "https://api.x.ai/v1",
13327
+ reasoning: true,
13328
+ input: ["text", "image"],
13329
+ cost: {
13330
+ input: 2,
13331
+ output: 6,
13332
+ cacheRead: 0.2,
13333
+ cacheWrite: 0,
13334
+ },
13335
+ contextWindow: 2000000,
13336
+ maxTokens: 30000,
13337
+ },
13338
+ "grok-4.20-multi-agent-beta-latest": {
13339
+ id: "grok-4.20-multi-agent-beta-latest",
13340
+ name: "Grok 4.20 Multi-Agent Beta",
13341
+ api: "openai-completions",
13342
+ provider: "xai",
13343
+ baseUrl: "https://api.x.ai/v1",
13344
+ reasoning: true,
13345
+ input: ["text", "image"],
13346
+ cost: {
13347
+ input: 2,
13348
+ output: 6,
13349
+ cacheRead: 0.2,
13350
+ cacheWrite: 0,
13351
+ },
13352
+ contextWindow: 2000000,
13353
+ maxTokens: 30000,
13354
+ },
13151
13355
  "grok-beta": {
13152
13356
  id: "grok-beta",
13153
13357
  name: "Grok Beta",