@mariozechner/pi-ai 0.57.1 → 0.58.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +10 -7
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +5 -2
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.generated.d.ts +429 -225
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +576 -372
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +1 -1
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +21 -3
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +31 -32
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +24 -19
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/types.d.ts +8 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts +14 -6
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +301 -57
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/github-copilot.d.ts.map +1 -1
- package/dist/utils/oauth/github-copilot.js +23 -12
- package/dist/utils/oauth/github-copilot.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -87,40 +87,6 @@ export const MODELS = {
|
|
|
87
87
|
contextWindow: 300000,
|
|
88
88
|
maxTokens: 8192,
|
|
89
89
|
},
|
|
90
|
-
"amazon.titan-text-express-v1": {
|
|
91
|
-
id: "amazon.titan-text-express-v1",
|
|
92
|
-
name: "Titan Text G1 - Express",
|
|
93
|
-
api: "bedrock-converse-stream",
|
|
94
|
-
provider: "amazon-bedrock",
|
|
95
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
96
|
-
reasoning: false,
|
|
97
|
-
input: ["text"],
|
|
98
|
-
cost: {
|
|
99
|
-
input: 0.2,
|
|
100
|
-
output: 0.6,
|
|
101
|
-
cacheRead: 0,
|
|
102
|
-
cacheWrite: 0,
|
|
103
|
-
},
|
|
104
|
-
contextWindow: 128000,
|
|
105
|
-
maxTokens: 4096,
|
|
106
|
-
},
|
|
107
|
-
"amazon.titan-text-express-v1:0:8k": {
|
|
108
|
-
id: "amazon.titan-text-express-v1:0:8k",
|
|
109
|
-
name: "Titan Text G1 - Express",
|
|
110
|
-
api: "bedrock-converse-stream",
|
|
111
|
-
provider: "amazon-bedrock",
|
|
112
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
113
|
-
reasoning: false,
|
|
114
|
-
input: ["text"],
|
|
115
|
-
cost: {
|
|
116
|
-
input: 0.2,
|
|
117
|
-
output: 0.6,
|
|
118
|
-
cacheRead: 0,
|
|
119
|
-
cacheWrite: 0,
|
|
120
|
-
},
|
|
121
|
-
contextWindow: 128000,
|
|
122
|
-
maxTokens: 4096,
|
|
123
|
-
},
|
|
124
90
|
"anthropic.claude-3-5-haiku-20241022-v1:0": {
|
|
125
91
|
id: "anthropic.claude-3-5-haiku-20241022-v1:0",
|
|
126
92
|
name: "Claude Haiku 3.5",
|
|
@@ -206,40 +172,6 @@ export const MODELS = {
|
|
|
206
172
|
contextWindow: 200000,
|
|
207
173
|
maxTokens: 4096,
|
|
208
174
|
},
|
|
209
|
-
"anthropic.claude-3-opus-20240229-v1:0": {
|
|
210
|
-
id: "anthropic.claude-3-opus-20240229-v1:0",
|
|
211
|
-
name: "Claude Opus 3",
|
|
212
|
-
api: "bedrock-converse-stream",
|
|
213
|
-
provider: "amazon-bedrock",
|
|
214
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
215
|
-
reasoning: false,
|
|
216
|
-
input: ["text", "image"],
|
|
217
|
-
cost: {
|
|
218
|
-
input: 15,
|
|
219
|
-
output: 75,
|
|
220
|
-
cacheRead: 0,
|
|
221
|
-
cacheWrite: 0,
|
|
222
|
-
},
|
|
223
|
-
contextWindow: 200000,
|
|
224
|
-
maxTokens: 4096,
|
|
225
|
-
},
|
|
226
|
-
"anthropic.claude-3-sonnet-20240229-v1:0": {
|
|
227
|
-
id: "anthropic.claude-3-sonnet-20240229-v1:0",
|
|
228
|
-
name: "Claude Sonnet 3",
|
|
229
|
-
api: "bedrock-converse-stream",
|
|
230
|
-
provider: "amazon-bedrock",
|
|
231
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
232
|
-
reasoning: false,
|
|
233
|
-
input: ["text", "image"],
|
|
234
|
-
cost: {
|
|
235
|
-
input: 3,
|
|
236
|
-
output: 15,
|
|
237
|
-
cacheRead: 0,
|
|
238
|
-
cacheWrite: 0,
|
|
239
|
-
},
|
|
240
|
-
contextWindow: 200000,
|
|
241
|
-
maxTokens: 4096,
|
|
242
|
-
},
|
|
243
175
|
"anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
244
176
|
id: "anthropic.claude-haiku-4-5-20251001-v1:0",
|
|
245
177
|
name: "Claude Haiku 4.5",
|
|
@@ -322,7 +254,7 @@ export const MODELS = {
|
|
|
322
254
|
cacheRead: 0.5,
|
|
323
255
|
cacheWrite: 6.25,
|
|
324
256
|
},
|
|
325
|
-
contextWindow:
|
|
257
|
+
contextWindow: 1000000,
|
|
326
258
|
maxTokens: 128000,
|
|
327
259
|
},
|
|
328
260
|
"anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -373,43 +305,9 @@ export const MODELS = {
|
|
|
373
305
|
cacheRead: 0.3,
|
|
374
306
|
cacheWrite: 3.75,
|
|
375
307
|
},
|
|
376
|
-
contextWindow:
|
|
308
|
+
contextWindow: 1000000,
|
|
377
309
|
maxTokens: 64000,
|
|
378
310
|
},
|
|
379
|
-
"cohere.command-r-plus-v1:0": {
|
|
380
|
-
id: "cohere.command-r-plus-v1:0",
|
|
381
|
-
name: "Command R+",
|
|
382
|
-
api: "bedrock-converse-stream",
|
|
383
|
-
provider: "amazon-bedrock",
|
|
384
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
385
|
-
reasoning: false,
|
|
386
|
-
input: ["text"],
|
|
387
|
-
cost: {
|
|
388
|
-
input: 3,
|
|
389
|
-
output: 15,
|
|
390
|
-
cacheRead: 0,
|
|
391
|
-
cacheWrite: 0,
|
|
392
|
-
},
|
|
393
|
-
contextWindow: 128000,
|
|
394
|
-
maxTokens: 4096,
|
|
395
|
-
},
|
|
396
|
-
"cohere.command-r-v1:0": {
|
|
397
|
-
id: "cohere.command-r-v1:0",
|
|
398
|
-
name: "Command R",
|
|
399
|
-
api: "bedrock-converse-stream",
|
|
400
|
-
provider: "amazon-bedrock",
|
|
401
|
-
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
402
|
-
reasoning: false,
|
|
403
|
-
input: ["text"],
|
|
404
|
-
cost: {
|
|
405
|
-
input: 0.5,
|
|
406
|
-
output: 1.5,
|
|
407
|
-
cacheRead: 0,
|
|
408
|
-
cacheWrite: 0,
|
|
409
|
-
},
|
|
410
|
-
contextWindow: 128000,
|
|
411
|
-
maxTokens: 4096,
|
|
412
|
-
},
|
|
413
311
|
"deepseek.r1-v1:0": {
|
|
414
312
|
id: "deepseek.r1-v1:0",
|
|
415
313
|
name: "DeepSeek-R1",
|
|
@@ -444,8 +342,8 @@ export const MODELS = {
|
|
|
444
342
|
contextWindow: 163840,
|
|
445
343
|
maxTokens: 81920,
|
|
446
344
|
},
|
|
447
|
-
"deepseek.v3.2
|
|
448
|
-
id: "deepseek.v3.2
|
|
345
|
+
"deepseek.v3.2": {
|
|
346
|
+
id: "deepseek.v3.2",
|
|
449
347
|
name: "DeepSeek-V3.2",
|
|
450
348
|
api: "bedrock-converse-stream",
|
|
451
349
|
provider: "amazon-bedrock",
|
|
@@ -509,7 +407,7 @@ export const MODELS = {
|
|
|
509
407
|
cacheRead: 0.5,
|
|
510
408
|
cacheWrite: 6.25,
|
|
511
409
|
},
|
|
512
|
-
contextWindow:
|
|
410
|
+
contextWindow: 1000000,
|
|
513
411
|
maxTokens: 128000,
|
|
514
412
|
},
|
|
515
413
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -560,7 +458,7 @@ export const MODELS = {
|
|
|
560
458
|
cacheRead: 0.3,
|
|
561
459
|
cacheWrite: 3.75,
|
|
562
460
|
},
|
|
563
|
-
contextWindow:
|
|
461
|
+
contextWindow: 1000000,
|
|
564
462
|
maxTokens: 64000,
|
|
565
463
|
},
|
|
566
464
|
"global.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
@@ -611,7 +509,7 @@ export const MODELS = {
|
|
|
611
509
|
cacheRead: 0.5,
|
|
612
510
|
cacheWrite: 6.25,
|
|
613
511
|
},
|
|
614
|
-
contextWindow:
|
|
512
|
+
contextWindow: 1000000,
|
|
615
513
|
maxTokens: 128000,
|
|
616
514
|
},
|
|
617
515
|
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -662,7 +560,7 @@ export const MODELS = {
|
|
|
662
560
|
cacheRead: 0.3,
|
|
663
561
|
cacheWrite: 3.75,
|
|
664
562
|
},
|
|
665
|
-
contextWindow:
|
|
563
|
+
contextWindow: 1000000,
|
|
666
564
|
maxTokens: 64000,
|
|
667
565
|
},
|
|
668
566
|
"google.gemma-3-27b-it": {
|
|
@@ -699,6 +597,23 @@ export const MODELS = {
|
|
|
699
597
|
contextWindow: 128000,
|
|
700
598
|
maxTokens: 4096,
|
|
701
599
|
},
|
|
600
|
+
"meta.llama3-1-405b-instruct-v1:0": {
|
|
601
|
+
id: "meta.llama3-1-405b-instruct-v1:0",
|
|
602
|
+
name: "Llama 3.1 405B Instruct",
|
|
603
|
+
api: "bedrock-converse-stream",
|
|
604
|
+
provider: "amazon-bedrock",
|
|
605
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
606
|
+
reasoning: false,
|
|
607
|
+
input: ["text"],
|
|
608
|
+
cost: {
|
|
609
|
+
input: 2.4,
|
|
610
|
+
output: 2.4,
|
|
611
|
+
cacheRead: 0,
|
|
612
|
+
cacheWrite: 0,
|
|
613
|
+
},
|
|
614
|
+
contextWindow: 128000,
|
|
615
|
+
maxTokens: 4096,
|
|
616
|
+
},
|
|
702
617
|
"meta.llama3-1-70b-instruct-v1:0": {
|
|
703
618
|
id: "meta.llama3-1-70b-instruct-v1:0",
|
|
704
619
|
name: "Llama 3.1 70B Instruct",
|
|
@@ -886,6 +801,40 @@ export const MODELS = {
|
|
|
886
801
|
contextWindow: 204800,
|
|
887
802
|
maxTokens: 131072,
|
|
888
803
|
},
|
|
804
|
+
"mistral.devstral-2-123b": {
|
|
805
|
+
id: "mistral.devstral-2-123b",
|
|
806
|
+
name: "Devstral 2 123B",
|
|
807
|
+
api: "bedrock-converse-stream",
|
|
808
|
+
provider: "amazon-bedrock",
|
|
809
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
810
|
+
reasoning: false,
|
|
811
|
+
input: ["text"],
|
|
812
|
+
cost: {
|
|
813
|
+
input: 0.4,
|
|
814
|
+
output: 2,
|
|
815
|
+
cacheRead: 0,
|
|
816
|
+
cacheWrite: 0,
|
|
817
|
+
},
|
|
818
|
+
contextWindow: 256000,
|
|
819
|
+
maxTokens: 8192,
|
|
820
|
+
},
|
|
821
|
+
"mistral.magistral-small-2509": {
|
|
822
|
+
id: "mistral.magistral-small-2509",
|
|
823
|
+
name: "Magistral Small 1.2",
|
|
824
|
+
api: "bedrock-converse-stream",
|
|
825
|
+
provider: "amazon-bedrock",
|
|
826
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
827
|
+
reasoning: true,
|
|
828
|
+
input: ["text", "image"],
|
|
829
|
+
cost: {
|
|
830
|
+
input: 0.5,
|
|
831
|
+
output: 1.5,
|
|
832
|
+
cacheRead: 0,
|
|
833
|
+
cacheWrite: 0,
|
|
834
|
+
},
|
|
835
|
+
contextWindow: 128000,
|
|
836
|
+
maxTokens: 40000,
|
|
837
|
+
},
|
|
889
838
|
"mistral.ministral-3-14b-instruct": {
|
|
890
839
|
id: "mistral.ministral-3-14b-instruct",
|
|
891
840
|
name: "Ministral 14B 3.0",
|
|
@@ -903,6 +852,23 @@ export const MODELS = {
|
|
|
903
852
|
contextWindow: 128000,
|
|
904
853
|
maxTokens: 4096,
|
|
905
854
|
},
|
|
855
|
+
"mistral.ministral-3-3b-instruct": {
|
|
856
|
+
id: "mistral.ministral-3-3b-instruct",
|
|
857
|
+
name: "Ministral 3 3B",
|
|
858
|
+
api: "bedrock-converse-stream",
|
|
859
|
+
provider: "amazon-bedrock",
|
|
860
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
861
|
+
reasoning: false,
|
|
862
|
+
input: ["text", "image"],
|
|
863
|
+
cost: {
|
|
864
|
+
input: 0.1,
|
|
865
|
+
output: 0.1,
|
|
866
|
+
cacheRead: 0,
|
|
867
|
+
cacheWrite: 0,
|
|
868
|
+
},
|
|
869
|
+
contextWindow: 256000,
|
|
870
|
+
maxTokens: 8192,
|
|
871
|
+
},
|
|
906
872
|
"mistral.ministral-3-8b-instruct": {
|
|
907
873
|
id: "mistral.ministral-3-8b-instruct",
|
|
908
874
|
name: "Ministral 3 8B",
|
|
@@ -920,22 +886,39 @@ export const MODELS = {
|
|
|
920
886
|
contextWindow: 128000,
|
|
921
887
|
maxTokens: 4096,
|
|
922
888
|
},
|
|
923
|
-
"mistral.mistral-large-
|
|
924
|
-
id: "mistral.mistral-large-
|
|
925
|
-
name: "Mistral Large
|
|
889
|
+
"mistral.mistral-large-3-675b-instruct": {
|
|
890
|
+
id: "mistral.mistral-large-3-675b-instruct",
|
|
891
|
+
name: "Mistral Large 3",
|
|
926
892
|
api: "bedrock-converse-stream",
|
|
927
893
|
provider: "amazon-bedrock",
|
|
928
894
|
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
929
895
|
reasoning: false,
|
|
930
|
-
input: ["text"],
|
|
896
|
+
input: ["text", "image"],
|
|
931
897
|
cost: {
|
|
932
898
|
input: 0.5,
|
|
933
899
|
output: 1.5,
|
|
934
900
|
cacheRead: 0,
|
|
935
901
|
cacheWrite: 0,
|
|
936
902
|
},
|
|
903
|
+
contextWindow: 256000,
|
|
904
|
+
maxTokens: 8192,
|
|
905
|
+
},
|
|
906
|
+
"mistral.pixtral-large-2502-v1:0": {
|
|
907
|
+
id: "mistral.pixtral-large-2502-v1:0",
|
|
908
|
+
name: "Pixtral Large (25.02)",
|
|
909
|
+
api: "bedrock-converse-stream",
|
|
910
|
+
provider: "amazon-bedrock",
|
|
911
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
912
|
+
reasoning: false,
|
|
913
|
+
input: ["text", "image"],
|
|
914
|
+
cost: {
|
|
915
|
+
input: 2,
|
|
916
|
+
output: 6,
|
|
917
|
+
cacheRead: 0,
|
|
918
|
+
cacheWrite: 0,
|
|
919
|
+
},
|
|
937
920
|
contextWindow: 128000,
|
|
938
|
-
maxTokens:
|
|
921
|
+
maxTokens: 8192,
|
|
939
922
|
},
|
|
940
923
|
"mistral.voxtral-mini-3b-2507": {
|
|
941
924
|
id: "mistral.voxtral-mini-3b-2507",
|
|
@@ -1022,6 +1005,23 @@ export const MODELS = {
|
|
|
1022
1005
|
contextWindow: 128000,
|
|
1023
1006
|
maxTokens: 4096,
|
|
1024
1007
|
},
|
|
1008
|
+
"nvidia.nemotron-nano-3-30b": {
|
|
1009
|
+
id: "nvidia.nemotron-nano-3-30b",
|
|
1010
|
+
name: "NVIDIA Nemotron Nano 3 30B",
|
|
1011
|
+
api: "bedrock-converse-stream",
|
|
1012
|
+
provider: "amazon-bedrock",
|
|
1013
|
+
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
|
|
1014
|
+
reasoning: true,
|
|
1015
|
+
input: ["text"],
|
|
1016
|
+
cost: {
|
|
1017
|
+
input: 0.06,
|
|
1018
|
+
output: 0.24,
|
|
1019
|
+
cacheRead: 0,
|
|
1020
|
+
cacheWrite: 0,
|
|
1021
|
+
},
|
|
1022
|
+
contextWindow: 128000,
|
|
1023
|
+
maxTokens: 4096,
|
|
1024
|
+
},
|
|
1025
1025
|
"nvidia.nemotron-nano-9b-v2": {
|
|
1026
1026
|
id: "nvidia.nemotron-nano-9b-v2",
|
|
1027
1027
|
name: "NVIDIA Nemotron Nano 9B v2",
|
|
@@ -1291,7 +1291,7 @@ export const MODELS = {
|
|
|
1291
1291
|
cacheRead: 0.5,
|
|
1292
1292
|
cacheWrite: 6.25,
|
|
1293
1293
|
},
|
|
1294
|
-
contextWindow:
|
|
1294
|
+
contextWindow: 1000000,
|
|
1295
1295
|
maxTokens: 128000,
|
|
1296
1296
|
},
|
|
1297
1297
|
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -1342,7 +1342,7 @@ export const MODELS = {
|
|
|
1342
1342
|
cacheRead: 0.3,
|
|
1343
1343
|
cacheWrite: 3.75,
|
|
1344
1344
|
},
|
|
1345
|
-
contextWindow:
|
|
1345
|
+
contextWindow: 1000000,
|
|
1346
1346
|
maxTokens: 64000,
|
|
1347
1347
|
},
|
|
1348
1348
|
"writer.palmyra-x4-v1:0": {
|
|
@@ -1718,7 +1718,7 @@ export const MODELS = {
|
|
|
1718
1718
|
cacheRead: 0.5,
|
|
1719
1719
|
cacheWrite: 6.25,
|
|
1720
1720
|
},
|
|
1721
|
-
contextWindow:
|
|
1721
|
+
contextWindow: 1000000,
|
|
1722
1722
|
maxTokens: 128000,
|
|
1723
1723
|
},
|
|
1724
1724
|
"claude-sonnet-4-0": {
|
|
@@ -1803,7 +1803,7 @@ export const MODELS = {
|
|
|
1803
1803
|
cacheRead: 0.3,
|
|
1804
1804
|
cacheWrite: 3.75,
|
|
1805
1805
|
},
|
|
1806
|
-
contextWindow:
|
|
1806
|
+
contextWindow: 1000000,
|
|
1807
1807
|
maxTokens: 64000,
|
|
1808
1808
|
},
|
|
1809
1809
|
},
|
|
@@ -2871,7 +2871,7 @@ export const MODELS = {
|
|
|
2871
2871
|
cacheRead: 0,
|
|
2872
2872
|
cacheWrite: 0,
|
|
2873
2873
|
},
|
|
2874
|
-
contextWindow:
|
|
2874
|
+
contextWindow: 264000,
|
|
2875
2875
|
maxTokens: 64000,
|
|
2876
2876
|
},
|
|
2877
2877
|
"gpt-5.2-codex": {
|
|
@@ -2889,7 +2889,7 @@ export const MODELS = {
|
|
|
2889
2889
|
cacheRead: 0,
|
|
2890
2890
|
cacheWrite: 0,
|
|
2891
2891
|
},
|
|
2892
|
-
contextWindow:
|
|
2892
|
+
contextWindow: 400000,
|
|
2893
2893
|
maxTokens: 128000,
|
|
2894
2894
|
},
|
|
2895
2895
|
"gpt-5.3-codex": {
|
|
@@ -3247,10 +3247,10 @@ export const MODELS = {
|
|
|
3247
3247
|
reasoning: true,
|
|
3248
3248
|
input: ["text", "image"],
|
|
3249
3249
|
cost: {
|
|
3250
|
-
input: 0.
|
|
3251
|
-
output:
|
|
3252
|
-
cacheRead: 0.
|
|
3253
|
-
cacheWrite:
|
|
3250
|
+
input: 0.25,
|
|
3251
|
+
output: 1.5,
|
|
3252
|
+
cacheRead: 0.025,
|
|
3253
|
+
cacheWrite: 1,
|
|
3254
3254
|
},
|
|
3255
3255
|
contextWindow: 1048576,
|
|
3256
3256
|
maxTokens: 65536,
|
|
@@ -5895,7 +5895,7 @@ export const MODELS = {
|
|
|
5895
5895
|
cacheRead: 0.5,
|
|
5896
5896
|
cacheWrite: 6.25,
|
|
5897
5897
|
},
|
|
5898
|
-
contextWindow:
|
|
5898
|
+
contextWindow: 1000000,
|
|
5899
5899
|
maxTokens: 128000,
|
|
5900
5900
|
},
|
|
5901
5901
|
"claude-sonnet-4": {
|
|
@@ -6272,6 +6272,23 @@ export const MODELS = {
|
|
|
6272
6272
|
contextWindow: 262144,
|
|
6273
6273
|
maxTokens: 65536,
|
|
6274
6274
|
},
|
|
6275
|
+
"mimo-v2-flash-free": {
|
|
6276
|
+
id: "mimo-v2-flash-free",
|
|
6277
|
+
name: "MiMo V2 Flash Free",
|
|
6278
|
+
api: "openai-completions",
|
|
6279
|
+
provider: "opencode",
|
|
6280
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6281
|
+
reasoning: true,
|
|
6282
|
+
input: ["text"],
|
|
6283
|
+
cost: {
|
|
6284
|
+
input: 0,
|
|
6285
|
+
output: 0,
|
|
6286
|
+
cacheRead: 0,
|
|
6287
|
+
cacheWrite: 0,
|
|
6288
|
+
},
|
|
6289
|
+
contextWindow: 262144,
|
|
6290
|
+
maxTokens: 65536,
|
|
6291
|
+
},
|
|
6275
6292
|
"minimax-m2.1": {
|
|
6276
6293
|
id: "minimax-m2.1",
|
|
6277
6294
|
name: "MiniMax M2.1",
|
|
@@ -6323,6 +6340,23 @@ export const MODELS = {
|
|
|
6323
6340
|
contextWindow: 204800,
|
|
6324
6341
|
maxTokens: 131072,
|
|
6325
6342
|
},
|
|
6343
|
+
"nemotron-3-super-free": {
|
|
6344
|
+
id: "nemotron-3-super-free",
|
|
6345
|
+
name: "Nemotron 3 Super Free",
|
|
6346
|
+
api: "openai-completions",
|
|
6347
|
+
provider: "opencode",
|
|
6348
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6349
|
+
reasoning: true,
|
|
6350
|
+
input: ["text"],
|
|
6351
|
+
cost: {
|
|
6352
|
+
input: 0,
|
|
6353
|
+
output: 0,
|
|
6354
|
+
cacheRead: 0,
|
|
6355
|
+
cacheWrite: 0,
|
|
6356
|
+
},
|
|
6357
|
+
contextWindow: 1000000,
|
|
6358
|
+
maxTokens: 128000,
|
|
6359
|
+
},
|
|
6326
6360
|
},
|
|
6327
6361
|
"opencode-go": {
|
|
6328
6362
|
"glm-5": {
|
|
@@ -6888,6 +6922,23 @@ export const MODELS = {
|
|
|
6888
6922
|
contextWindow: 262144,
|
|
6889
6923
|
maxTokens: 32768,
|
|
6890
6924
|
},
|
|
6925
|
+
"bytedance-seed/seed-2.0-lite": {
|
|
6926
|
+
id: "bytedance-seed/seed-2.0-lite",
|
|
6927
|
+
name: "ByteDance Seed: Seed-2.0-Lite",
|
|
6928
|
+
api: "openai-completions",
|
|
6929
|
+
provider: "openrouter",
|
|
6930
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6931
|
+
reasoning: true,
|
|
6932
|
+
input: ["text", "image"],
|
|
6933
|
+
cost: {
|
|
6934
|
+
input: 0.25,
|
|
6935
|
+
output: 2,
|
|
6936
|
+
cacheRead: 0,
|
|
6937
|
+
cacheWrite: 0,
|
|
6938
|
+
},
|
|
6939
|
+
contextWindow: 262144,
|
|
6940
|
+
maxTokens: 131072,
|
|
6941
|
+
},
|
|
6891
6942
|
"bytedance-seed/seed-2.0-mini": {
|
|
6892
6943
|
id: "bytedance-seed/seed-2.0-mini",
|
|
6893
6944
|
name: "ByteDance Seed: Seed-2.0-Mini",
|
|
@@ -6967,11 +7018,11 @@ export const MODELS = {
|
|
|
6967
7018
|
cost: {
|
|
6968
7019
|
input: 0.19999999999999998,
|
|
6969
7020
|
output: 0.77,
|
|
6970
|
-
cacheRead: 0.
|
|
7021
|
+
cacheRead: 0.135,
|
|
6971
7022
|
cacheWrite: 0,
|
|
6972
7023
|
},
|
|
6973
7024
|
contextWindow: 163840,
|
|
6974
|
-
maxTokens:
|
|
7025
|
+
maxTokens: 4096,
|
|
6975
7026
|
},
|
|
6976
7027
|
"deepseek/deepseek-chat-v3.1": {
|
|
6977
7028
|
id: "deepseek/deepseek-chat-v3.1",
|
|
@@ -7041,40 +7092,23 @@ export const MODELS = {
|
|
|
7041
7092
|
contextWindow: 163840,
|
|
7042
7093
|
maxTokens: 4096,
|
|
7043
7094
|
},
|
|
7044
|
-
"deepseek/deepseek-v3.
|
|
7045
|
-
id: "deepseek/deepseek-v3.
|
|
7046
|
-
name: "DeepSeek: DeepSeek V3.
|
|
7095
|
+
"deepseek/deepseek-v3.2": {
|
|
7096
|
+
id: "deepseek/deepseek-v3.2",
|
|
7097
|
+
name: "DeepSeek: DeepSeek V3.2",
|
|
7047
7098
|
api: "openai-completions",
|
|
7048
7099
|
provider: "openrouter",
|
|
7049
7100
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7050
7101
|
reasoning: true,
|
|
7051
7102
|
input: ["text"],
|
|
7052
7103
|
cost: {
|
|
7053
|
-
input: 0.
|
|
7054
|
-
output: 0.
|
|
7055
|
-
cacheRead: 0.
|
|
7104
|
+
input: 0.26,
|
|
7105
|
+
output: 0.38,
|
|
7106
|
+
cacheRead: 0.13,
|
|
7056
7107
|
cacheWrite: 0,
|
|
7057
7108
|
},
|
|
7058
7109
|
contextWindow: 163840,
|
|
7059
7110
|
maxTokens: 4096,
|
|
7060
7111
|
},
|
|
7061
|
-
"deepseek/deepseek-v3.2": {
|
|
7062
|
-
id: "deepseek/deepseek-v3.2",
|
|
7063
|
-
name: "DeepSeek: DeepSeek V3.2",
|
|
7064
|
-
api: "openai-completions",
|
|
7065
|
-
provider: "openrouter",
|
|
7066
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7067
|
-
reasoning: true,
|
|
7068
|
-
input: ["text"],
|
|
7069
|
-
cost: {
|
|
7070
|
-
input: 0.25,
|
|
7071
|
-
output: 0.39999999999999997,
|
|
7072
|
-
cacheRead: 0,
|
|
7073
|
-
cacheWrite: 0,
|
|
7074
|
-
},
|
|
7075
|
-
contextWindow: 163840,
|
|
7076
|
-
maxTokens: 65536,
|
|
7077
|
-
},
|
|
7078
7112
|
"deepseek/deepseek-v3.2-exp": {
|
|
7079
7113
|
id: "deepseek/deepseek-v3.2-exp",
|
|
7080
7114
|
name: "DeepSeek: DeepSeek V3.2 Exp",
|
|
@@ -7339,9 +7373,9 @@ export const MODELS = {
|
|
|
7339
7373
|
reasoning: false,
|
|
7340
7374
|
input: ["text", "image"],
|
|
7341
7375
|
cost: {
|
|
7342
|
-
input: 0.
|
|
7343
|
-
output: 0.
|
|
7344
|
-
cacheRead: 0.
|
|
7376
|
+
input: 0.03,
|
|
7377
|
+
output: 0.11,
|
|
7378
|
+
cacheRead: 0.015,
|
|
7345
7379
|
cacheWrite: 0,
|
|
7346
7380
|
},
|
|
7347
7381
|
contextWindow: 128000,
|
|
@@ -7466,23 +7500,6 @@ export const MODELS = {
|
|
|
7466
7500
|
contextWindow: 8192,
|
|
7467
7501
|
maxTokens: 16384,
|
|
7468
7502
|
},
|
|
7469
|
-
"meta-llama/llama-3.1-405b-instruct": {
|
|
7470
|
-
id: "meta-llama/llama-3.1-405b-instruct",
|
|
7471
|
-
name: "Meta: Llama 3.1 405B Instruct",
|
|
7472
|
-
api: "openai-completions",
|
|
7473
|
-
provider: "openrouter",
|
|
7474
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7475
|
-
reasoning: false,
|
|
7476
|
-
input: ["text"],
|
|
7477
|
-
cost: {
|
|
7478
|
-
input: 4,
|
|
7479
|
-
output: 4,
|
|
7480
|
-
cacheRead: 0,
|
|
7481
|
-
cacheWrite: 0,
|
|
7482
|
-
},
|
|
7483
|
-
contextWindow: 131000,
|
|
7484
|
-
maxTokens: 4096,
|
|
7485
|
-
},
|
|
7486
7503
|
"meta-llama/llama-3.1-70b-instruct": {
|
|
7487
7504
|
id: "meta-llama/llama-3.1-70b-instruct",
|
|
7488
7505
|
name: "Meta: Llama 3.1 70B Instruct",
|
|
@@ -7645,13 +7662,13 @@ export const MODELS = {
|
|
|
7645
7662
|
reasoning: true,
|
|
7646
7663
|
input: ["text"],
|
|
7647
7664
|
cost: {
|
|
7648
|
-
input: 0.
|
|
7649
|
-
output:
|
|
7650
|
-
cacheRead: 0.
|
|
7665
|
+
input: 0.27,
|
|
7666
|
+
output: 0.95,
|
|
7667
|
+
cacheRead: 0.0299999997,
|
|
7651
7668
|
cacheWrite: 0,
|
|
7652
7669
|
},
|
|
7653
7670
|
contextWindow: 196608,
|
|
7654
|
-
maxTokens:
|
|
7671
|
+
maxTokens: 4096,
|
|
7655
7672
|
},
|
|
7656
7673
|
"mistralai/codestral-2508": {
|
|
7657
7674
|
id: "mistralai/codestral-2508",
|
|
@@ -8078,23 +8095,6 @@ export const MODELS = {
|
|
|
8078
8095
|
contextWindow: 131072,
|
|
8079
8096
|
maxTokens: 4096,
|
|
8080
8097
|
},
|
|
8081
|
-
"moonshotai/kimi-k2-0905:exacto": {
|
|
8082
|
-
id: "moonshotai/kimi-k2-0905:exacto",
|
|
8083
|
-
name: "MoonshotAI: Kimi K2 0905 (exacto)",
|
|
8084
|
-
api: "openai-completions",
|
|
8085
|
-
provider: "openrouter",
|
|
8086
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8087
|
-
reasoning: false,
|
|
8088
|
-
input: ["text"],
|
|
8089
|
-
cost: {
|
|
8090
|
-
input: 0.6,
|
|
8091
|
-
output: 2.5,
|
|
8092
|
-
cacheRead: 0,
|
|
8093
|
-
cacheWrite: 0,
|
|
8094
|
-
},
|
|
8095
|
-
contextWindow: 262144,
|
|
8096
|
-
maxTokens: 4096,
|
|
8097
|
-
},
|
|
8098
8098
|
"moonshotai/kimi-k2-thinking": {
|
|
8099
8099
|
id: "moonshotai/kimi-k2-thinking",
|
|
8100
8100
|
name: "MoonshotAI: Kimi K2 Thinking",
|
|
@@ -8214,6 +8214,23 @@ export const MODELS = {
|
|
|
8214
8214
|
contextWindow: 256000,
|
|
8215
8215
|
maxTokens: 4096,
|
|
8216
8216
|
},
|
|
8217
|
+
"nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8218
|
+
id: "nvidia/nemotron-3-super-120b-a12b:free",
|
|
8219
|
+
name: "NVIDIA: Nemotron 3 Super (free)",
|
|
8220
|
+
api: "openai-completions",
|
|
8221
|
+
provider: "openrouter",
|
|
8222
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8223
|
+
reasoning: true,
|
|
8224
|
+
input: ["text"],
|
|
8225
|
+
cost: {
|
|
8226
|
+
input: 0,
|
|
8227
|
+
output: 0,
|
|
8228
|
+
cacheRead: 0,
|
|
8229
|
+
cacheWrite: 0,
|
|
8230
|
+
},
|
|
8231
|
+
contextWindow: 262144,
|
|
8232
|
+
maxTokens: 262144,
|
|
8233
|
+
},
|
|
8217
8234
|
"nvidia/nemotron-nano-12b-v2-vl:free": {
|
|
8218
8235
|
id: "nvidia/nemotron-nano-12b-v2-vl:free",
|
|
8219
8236
|
name: "NVIDIA: Nemotron Nano 12B 2 VL (free)",
|
|
@@ -8945,23 +8962,6 @@ export const MODELS = {
|
|
|
8945
8962
|
contextWindow: 131072,
|
|
8946
8963
|
maxTokens: 4096,
|
|
8947
8964
|
},
|
|
8948
|
-
"openai/gpt-oss-120b:exacto": {
|
|
8949
|
-
id: "openai/gpt-oss-120b:exacto",
|
|
8950
|
-
name: "OpenAI: gpt-oss-120b (exacto)",
|
|
8951
|
-
api: "openai-completions",
|
|
8952
|
-
provider: "openrouter",
|
|
8953
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8954
|
-
reasoning: true,
|
|
8955
|
-
input: ["text"],
|
|
8956
|
-
cost: {
|
|
8957
|
-
input: 0.039,
|
|
8958
|
-
output: 0.19,
|
|
8959
|
-
cacheRead: 0,
|
|
8960
|
-
cacheWrite: 0,
|
|
8961
|
-
},
|
|
8962
|
-
contextWindow: 131072,
|
|
8963
|
-
maxTokens: 4096,
|
|
8964
|
-
},
|
|
8965
8965
|
"openai/gpt-oss-120b:free": {
|
|
8966
8966
|
id: "openai/gpt-oss-120b:free",
|
|
8967
8967
|
name: "OpenAI: gpt-oss-120b (free)",
|
|
@@ -9217,6 +9217,40 @@ export const MODELS = {
|
|
|
9217
9217
|
contextWindow: 200000,
|
|
9218
9218
|
maxTokens: 4096,
|
|
9219
9219
|
},
|
|
9220
|
+
"openrouter/healer-alpha": {
|
|
9221
|
+
id: "openrouter/healer-alpha",
|
|
9222
|
+
name: "Healer Alpha",
|
|
9223
|
+
api: "openai-completions",
|
|
9224
|
+
provider: "openrouter",
|
|
9225
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9226
|
+
reasoning: true,
|
|
9227
|
+
input: ["text", "image"],
|
|
9228
|
+
cost: {
|
|
9229
|
+
input: 0,
|
|
9230
|
+
output: 0,
|
|
9231
|
+
cacheRead: 0,
|
|
9232
|
+
cacheWrite: 0,
|
|
9233
|
+
},
|
|
9234
|
+
contextWindow: 262144,
|
|
9235
|
+
maxTokens: 32000,
|
|
9236
|
+
},
|
|
9237
|
+
"openrouter/hunter-alpha": {
|
|
9238
|
+
id: "openrouter/hunter-alpha",
|
|
9239
|
+
name: "Hunter Alpha",
|
|
9240
|
+
api: "openai-completions",
|
|
9241
|
+
provider: "openrouter",
|
|
9242
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9243
|
+
reasoning: true,
|
|
9244
|
+
input: ["text"],
|
|
9245
|
+
cost: {
|
|
9246
|
+
input: 0,
|
|
9247
|
+
output: 0,
|
|
9248
|
+
cacheRead: 0,
|
|
9249
|
+
cacheWrite: 0,
|
|
9250
|
+
},
|
|
9251
|
+
contextWindow: 1048576,
|
|
9252
|
+
maxTokens: 32000,
|
|
9253
|
+
},
|
|
9220
9254
|
"prime-intellect/intellect-3": {
|
|
9221
9255
|
id: "prime-intellect/intellect-3",
|
|
9222
9256
|
name: "Prime Intellect: INTELLECT-3",
|
|
@@ -9625,23 +9659,6 @@ export const MODELS = {
|
|
|
9625
9659
|
contextWindow: 1000000,
|
|
9626
9660
|
maxTokens: 65536,
|
|
9627
9661
|
},
|
|
9628
|
-
"qwen/qwen3-coder:exacto": {
|
|
9629
|
-
id: "qwen/qwen3-coder:exacto",
|
|
9630
|
-
name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
|
|
9631
|
-
api: "openai-completions",
|
|
9632
|
-
provider: "openrouter",
|
|
9633
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9634
|
-
reasoning: false,
|
|
9635
|
-
input: ["text"],
|
|
9636
|
-
cost: {
|
|
9637
|
-
input: 0.22,
|
|
9638
|
-
output: 1.7999999999999998,
|
|
9639
|
-
cacheRead: 0.022,
|
|
9640
|
-
cacheWrite: 0,
|
|
9641
|
-
},
|
|
9642
|
-
contextWindow: 262144,
|
|
9643
|
-
maxTokens: 65536,
|
|
9644
|
-
},
|
|
9645
9662
|
"qwen/qwen3-coder:free": {
|
|
9646
9663
|
id: "qwen/qwen3-coder:free",
|
|
9647
9664
|
name: "Qwen: Qwen3 Coder 480B A35B (free)",
|
|
@@ -9707,7 +9724,7 @@ export const MODELS = {
|
|
|
9707
9724
|
cacheRead: 0,
|
|
9708
9725
|
cacheWrite: 0,
|
|
9709
9726
|
},
|
|
9710
|
-
contextWindow:
|
|
9727
|
+
contextWindow: 131072,
|
|
9711
9728
|
maxTokens: 4096,
|
|
9712
9729
|
},
|
|
9713
9730
|
"qwen/qwen3-next-80b-a3b-instruct:free": {
|
|
@@ -9736,13 +9753,13 @@ export const MODELS = {
|
|
|
9736
9753
|
reasoning: true,
|
|
9737
9754
|
input: ["text"],
|
|
9738
9755
|
cost: {
|
|
9739
|
-
input: 0.
|
|
9740
|
-
output:
|
|
9756
|
+
input: 0.0975,
|
|
9757
|
+
output: 0.78,
|
|
9741
9758
|
cacheRead: 0,
|
|
9742
9759
|
cacheWrite: 0,
|
|
9743
9760
|
},
|
|
9744
|
-
contextWindow:
|
|
9745
|
-
maxTokens:
|
|
9761
|
+
contextWindow: 131072,
|
|
9762
|
+
maxTokens: 32768,
|
|
9746
9763
|
},
|
|
9747
9764
|
"qwen/qwen3-vl-235b-a22b-instruct": {
|
|
9748
9765
|
id: "qwen/qwen3-vl-235b-a22b-instruct",
|
|
@@ -9770,8 +9787,8 @@ export const MODELS = {
|
|
|
9770
9787
|
reasoning: true,
|
|
9771
9788
|
input: ["text", "image"],
|
|
9772
9789
|
cost: {
|
|
9773
|
-
input: 0,
|
|
9774
|
-
output:
|
|
9790
|
+
input: 0.26,
|
|
9791
|
+
output: 2.6,
|
|
9775
9792
|
cacheRead: 0,
|
|
9776
9793
|
cacheWrite: 0,
|
|
9777
9794
|
},
|
|
@@ -9804,8 +9821,8 @@ export const MODELS = {
|
|
|
9804
9821
|
reasoning: true,
|
|
9805
9822
|
input: ["text", "image"],
|
|
9806
9823
|
cost: {
|
|
9807
|
-
input: 0,
|
|
9808
|
-
output:
|
|
9824
|
+
input: 0.13,
|
|
9825
|
+
output: 1.56,
|
|
9809
9826
|
cacheRead: 0,
|
|
9810
9827
|
cacheWrite: 0,
|
|
9811
9828
|
},
|
|
@@ -9931,6 +9948,23 @@ export const MODELS = {
|
|
|
9931
9948
|
contextWindow: 262144,
|
|
9932
9949
|
maxTokens: 65536,
|
|
9933
9950
|
},
|
|
9951
|
+
"qwen/qwen3.5-9b": {
|
|
9952
|
+
id: "qwen/qwen3.5-9b",
|
|
9953
|
+
name: "Qwen: Qwen3.5-9B",
|
|
9954
|
+
api: "openai-completions",
|
|
9955
|
+
provider: "openrouter",
|
|
9956
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9957
|
+
reasoning: true,
|
|
9958
|
+
input: ["text", "image"],
|
|
9959
|
+
cost: {
|
|
9960
|
+
input: 0.049999999999999996,
|
|
9961
|
+
output: 0.15,
|
|
9962
|
+
cacheRead: 0,
|
|
9963
|
+
cacheWrite: 0,
|
|
9964
|
+
},
|
|
9965
|
+
contextWindow: 256000,
|
|
9966
|
+
maxTokens: 4096,
|
|
9967
|
+
},
|
|
9934
9968
|
"qwen/qwen3.5-flash-02-23": {
|
|
9935
9969
|
id: "qwen/qwen3.5-flash-02-23",
|
|
9936
9970
|
name: "Qwen: Qwen3.5-Flash",
|
|
@@ -10025,13 +10059,13 @@ export const MODELS = {
|
|
|
10025
10059
|
reasoning: false,
|
|
10026
10060
|
input: ["text"],
|
|
10027
10061
|
cost: {
|
|
10028
|
-
input: 0.
|
|
10029
|
-
output: 0.
|
|
10062
|
+
input: 0.85,
|
|
10063
|
+
output: 0.85,
|
|
10030
10064
|
cacheRead: 0,
|
|
10031
10065
|
cacheWrite: 0,
|
|
10032
10066
|
},
|
|
10033
|
-
contextWindow:
|
|
10034
|
-
maxTokens:
|
|
10067
|
+
contextWindow: 131072,
|
|
10068
|
+
maxTokens: 16384,
|
|
10035
10069
|
},
|
|
10036
10070
|
"stepfun/step-3.5-flash": {
|
|
10037
10071
|
id: "stepfun/step-3.5-flash",
|
|
@@ -10254,6 +10288,23 @@ export const MODELS = {
|
|
|
10254
10288
|
contextWindow: 2000000,
|
|
10255
10289
|
maxTokens: 30000,
|
|
10256
10290
|
},
|
|
10291
|
+
"x-ai/grok-4.20-beta": {
|
|
10292
|
+
id: "x-ai/grok-4.20-beta",
|
|
10293
|
+
name: "xAI: Grok 4.20 Beta",
|
|
10294
|
+
api: "openai-completions",
|
|
10295
|
+
provider: "openrouter",
|
|
10296
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10297
|
+
reasoning: true,
|
|
10298
|
+
input: ["text", "image"],
|
|
10299
|
+
cost: {
|
|
10300
|
+
input: 2,
|
|
10301
|
+
output: 6,
|
|
10302
|
+
cacheRead: 0.19999999999999998,
|
|
10303
|
+
cacheWrite: 0,
|
|
10304
|
+
},
|
|
10305
|
+
contextWindow: 2000000,
|
|
10306
|
+
maxTokens: 4096,
|
|
10307
|
+
},
|
|
10257
10308
|
"x-ai/grok-code-fast-1": {
|
|
10258
10309
|
id: "x-ai/grok-code-fast-1",
|
|
10259
10310
|
name: "xAI: Grok Code Fast 1",
|
|
@@ -10390,23 +10441,6 @@ export const MODELS = {
|
|
|
10390
10441
|
contextWindow: 204800,
|
|
10391
10442
|
maxTokens: 204800,
|
|
10392
10443
|
},
|
|
10393
|
-
"z-ai/glm-4.6:exacto": {
|
|
10394
|
-
id: "z-ai/glm-4.6:exacto",
|
|
10395
|
-
name: "Z.ai: GLM 4.6 (exacto)",
|
|
10396
|
-
api: "openai-completions",
|
|
10397
|
-
provider: "openrouter",
|
|
10398
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
10399
|
-
reasoning: true,
|
|
10400
|
-
input: ["text"],
|
|
10401
|
-
cost: {
|
|
10402
|
-
input: 0.44,
|
|
10403
|
-
output: 1.76,
|
|
10404
|
-
cacheRead: 0.11,
|
|
10405
|
-
cacheWrite: 0,
|
|
10406
|
-
},
|
|
10407
|
-
contextWindow: 204800,
|
|
10408
|
-
maxTokens: 131072,
|
|
10409
|
-
},
|
|
10410
10444
|
"z-ai/glm-4.6v": {
|
|
10411
10445
|
id: "z-ai/glm-4.6v",
|
|
10412
10446
|
name: "Z.ai: GLM 4.6V",
|
|
@@ -10473,7 +10507,7 @@ export const MODELS = {
|
|
|
10473
10507
|
cacheWrite: 0,
|
|
10474
10508
|
},
|
|
10475
10509
|
contextWindow: 202752,
|
|
10476
|
-
maxTokens:
|
|
10510
|
+
maxTokens: 131072,
|
|
10477
10511
|
},
|
|
10478
10512
|
},
|
|
10479
10513
|
"vercel-ai-gateway": {
|
|
@@ -10537,13 +10571,13 @@ export const MODELS = {
|
|
|
10537
10571
|
reasoning: true,
|
|
10538
10572
|
input: ["text"],
|
|
10539
10573
|
cost: {
|
|
10540
|
-
input: 0.
|
|
10541
|
-
output: 0.
|
|
10574
|
+
input: 0.29,
|
|
10575
|
+
output: 0.59,
|
|
10542
10576
|
cacheRead: 0,
|
|
10543
10577
|
cacheWrite: 0,
|
|
10544
10578
|
},
|
|
10545
|
-
contextWindow:
|
|
10546
|
-
maxTokens:
|
|
10579
|
+
contextWindow: 131072,
|
|
10580
|
+
maxTokens: 40960,
|
|
10547
10581
|
},
|
|
10548
10582
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10549
10583
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10588,13 +10622,13 @@ export const MODELS = {
|
|
|
10588
10622
|
reasoning: true,
|
|
10589
10623
|
input: ["text"],
|
|
10590
10624
|
cost: {
|
|
10591
|
-
input: 0.
|
|
10592
|
-
output: 0.
|
|
10625
|
+
input: 0.15,
|
|
10626
|
+
output: 0.6,
|
|
10593
10627
|
cacheRead: 0,
|
|
10594
10628
|
cacheWrite: 0,
|
|
10595
10629
|
},
|
|
10596
|
-
contextWindow:
|
|
10597
|
-
maxTokens:
|
|
10630
|
+
contextWindow: 262144,
|
|
10631
|
+
maxTokens: 8192,
|
|
10598
10632
|
},
|
|
10599
10633
|
"alibaba/qwen3-coder-next": {
|
|
10600
10634
|
id: "alibaba/qwen3-coder-next",
|
|
@@ -10602,7 +10636,7 @@ export const MODELS = {
|
|
|
10602
10636
|
api: "anthropic-messages",
|
|
10603
10637
|
provider: "vercel-ai-gateway",
|
|
10604
10638
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10605
|
-
reasoning:
|
|
10639
|
+
reasoning: false,
|
|
10606
10640
|
input: ["text"],
|
|
10607
10641
|
cost: {
|
|
10608
10642
|
input: 0.5,
|
|
@@ -10630,6 +10664,23 @@ export const MODELS = {
|
|
|
10630
10664
|
contextWindow: 1000000,
|
|
10631
10665
|
maxTokens: 65536,
|
|
10632
10666
|
},
|
|
10667
|
+
"alibaba/qwen3-max": {
|
|
10668
|
+
id: "alibaba/qwen3-max",
|
|
10669
|
+
name: "Qwen3 Max",
|
|
10670
|
+
api: "anthropic-messages",
|
|
10671
|
+
provider: "vercel-ai-gateway",
|
|
10672
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10673
|
+
reasoning: false,
|
|
10674
|
+
input: ["text"],
|
|
10675
|
+
cost: {
|
|
10676
|
+
input: 1.2,
|
|
10677
|
+
output: 6,
|
|
10678
|
+
cacheRead: 0.24,
|
|
10679
|
+
cacheWrite: 0,
|
|
10680
|
+
},
|
|
10681
|
+
contextWindow: 262144,
|
|
10682
|
+
maxTokens: 32768,
|
|
10683
|
+
},
|
|
10633
10684
|
"alibaba/qwen3-max-preview": {
|
|
10634
10685
|
id: "alibaba/qwen3-max-preview",
|
|
10635
10686
|
name: "Qwen3 Max Preview",
|
|
@@ -10777,8 +10828,8 @@ export const MODELS = {
|
|
|
10777
10828
|
cost: {
|
|
10778
10829
|
input: 3,
|
|
10779
10830
|
output: 15,
|
|
10780
|
-
cacheRead: 0,
|
|
10781
|
-
cacheWrite:
|
|
10831
|
+
cacheRead: 0.3,
|
|
10832
|
+
cacheWrite: 3.75,
|
|
10782
10833
|
},
|
|
10783
10834
|
contextWindow: 200000,
|
|
10784
10835
|
maxTokens: 8192,
|
|
@@ -10987,6 +11038,23 @@ export const MODELS = {
|
|
|
10987
11038
|
contextWindow: 256000,
|
|
10988
11039
|
maxTokens: 8000,
|
|
10989
11040
|
},
|
|
11041
|
+
"deepseek/deepseek-r1": {
|
|
11042
|
+
id: "deepseek/deepseek-r1",
|
|
11043
|
+
name: "DeepSeek-R1",
|
|
11044
|
+
api: "anthropic-messages",
|
|
11045
|
+
provider: "vercel-ai-gateway",
|
|
11046
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11047
|
+
reasoning: true,
|
|
11048
|
+
input: ["text"],
|
|
11049
|
+
cost: {
|
|
11050
|
+
input: 1.35,
|
|
11051
|
+
output: 5.4,
|
|
11052
|
+
cacheRead: 0,
|
|
11053
|
+
cacheWrite: 0,
|
|
11054
|
+
},
|
|
11055
|
+
contextWindow: 128000,
|
|
11056
|
+
maxTokens: 8192,
|
|
11057
|
+
},
|
|
10990
11058
|
"deepseek/deepseek-v3": {
|
|
10991
11059
|
id: "deepseek/deepseek-v3",
|
|
10992
11060
|
name: "DeepSeek V3 0324",
|
|
@@ -11013,13 +11081,13 @@ export const MODELS = {
|
|
|
11013
11081
|
reasoning: true,
|
|
11014
11082
|
input: ["text"],
|
|
11015
11083
|
cost: {
|
|
11016
|
-
input: 0.
|
|
11017
|
-
output:
|
|
11084
|
+
input: 0.5,
|
|
11085
|
+
output: 1.5,
|
|
11018
11086
|
cacheRead: 0,
|
|
11019
11087
|
cacheWrite: 0,
|
|
11020
11088
|
},
|
|
11021
11089
|
contextWindow: 163840,
|
|
11022
|
-
maxTokens:
|
|
11090
|
+
maxTokens: 16384,
|
|
11023
11091
|
},
|
|
11024
11092
|
"deepseek/deepseek-v3.1-terminus": {
|
|
11025
11093
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11072,18 +11140,52 @@ export const MODELS = {
|
|
|
11072
11140
|
contextWindow: 128000,
|
|
11073
11141
|
maxTokens: 64000,
|
|
11074
11142
|
},
|
|
11075
|
-
"google/gemini-2.
|
|
11076
|
-
id: "google/gemini-2.
|
|
11143
|
+
"google/gemini-2.0-flash": {
|
|
11144
|
+
id: "google/gemini-2.0-flash",
|
|
11145
|
+
name: "Gemini 2.0 Flash",
|
|
11146
|
+
api: "anthropic-messages",
|
|
11147
|
+
provider: "vercel-ai-gateway",
|
|
11148
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11149
|
+
reasoning: false,
|
|
11150
|
+
input: ["text", "image"],
|
|
11151
|
+
cost: {
|
|
11152
|
+
input: 0.15,
|
|
11153
|
+
output: 0.6,
|
|
11154
|
+
cacheRead: 0,
|
|
11155
|
+
cacheWrite: 0,
|
|
11156
|
+
},
|
|
11157
|
+
contextWindow: 1048576,
|
|
11158
|
+
maxTokens: 8192,
|
|
11159
|
+
},
|
|
11160
|
+
"google/gemini-2.0-flash-lite": {
|
|
11161
|
+
id: "google/gemini-2.0-flash-lite",
|
|
11162
|
+
name: "Gemini 2.0 Flash Lite",
|
|
11163
|
+
api: "anthropic-messages",
|
|
11164
|
+
provider: "vercel-ai-gateway",
|
|
11165
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11166
|
+
reasoning: false,
|
|
11167
|
+
input: ["text", "image"],
|
|
11168
|
+
cost: {
|
|
11169
|
+
input: 0.075,
|
|
11170
|
+
output: 0.3,
|
|
11171
|
+
cacheRead: 0,
|
|
11172
|
+
cacheWrite: 0,
|
|
11173
|
+
},
|
|
11174
|
+
contextWindow: 1048576,
|
|
11175
|
+
maxTokens: 8192,
|
|
11176
|
+
},
|
|
11177
|
+
"google/gemini-2.5-flash": {
|
|
11178
|
+
id: "google/gemini-2.5-flash",
|
|
11077
11179
|
name: "Gemini 2.5 Flash",
|
|
11078
11180
|
api: "anthropic-messages",
|
|
11079
11181
|
provider: "vercel-ai-gateway",
|
|
11080
11182
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11081
11183
|
reasoning: true,
|
|
11082
|
-
input: ["text"],
|
|
11184
|
+
input: ["text", "image"],
|
|
11083
11185
|
cost: {
|
|
11084
11186
|
input: 0.3,
|
|
11085
11187
|
output: 2.5,
|
|
11086
|
-
cacheRead: 0,
|
|
11188
|
+
cacheRead: 0.03,
|
|
11087
11189
|
cacheWrite: 0,
|
|
11088
11190
|
},
|
|
11089
11191
|
contextWindow: 1000000,
|
|
@@ -11147,11 +11249,11 @@ export const MODELS = {
|
|
|
11147
11249
|
provider: "vercel-ai-gateway",
|
|
11148
11250
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11149
11251
|
reasoning: true,
|
|
11150
|
-
input: ["text"],
|
|
11252
|
+
input: ["text", "image"],
|
|
11151
11253
|
cost: {
|
|
11152
11254
|
input: 1.25,
|
|
11153
11255
|
output: 10,
|
|
11154
|
-
cacheRead: 0,
|
|
11256
|
+
cacheRead: 0.125,
|
|
11155
11257
|
cacheWrite: 0,
|
|
11156
11258
|
},
|
|
11157
11259
|
contextWindow: 1048576,
|
|
@@ -11172,7 +11274,7 @@ export const MODELS = {
|
|
|
11172
11274
|
cacheWrite: 0,
|
|
11173
11275
|
},
|
|
11174
11276
|
contextWindow: 1000000,
|
|
11175
|
-
maxTokens:
|
|
11277
|
+
maxTokens: 65000,
|
|
11176
11278
|
},
|
|
11177
11279
|
"google/gemini-3-pro-preview": {
|
|
11178
11280
|
id: "google/gemini-3-pro-preview",
|
|
@@ -11274,7 +11376,7 @@ export const MODELS = {
|
|
|
11274
11376
|
cacheWrite: 0,
|
|
11275
11377
|
},
|
|
11276
11378
|
contextWindow: 128000,
|
|
11277
|
-
maxTokens:
|
|
11379
|
+
maxTokens: 100000,
|
|
11278
11380
|
},
|
|
11279
11381
|
"meituan/longcat-flash-thinking": {
|
|
11280
11382
|
id: "meituan/longcat-flash-thinking",
|
|
@@ -11302,13 +11404,13 @@ export const MODELS = {
|
|
|
11302
11404
|
reasoning: false,
|
|
11303
11405
|
input: ["text"],
|
|
11304
11406
|
cost: {
|
|
11305
|
-
input: 0.
|
|
11306
|
-
output: 0.
|
|
11407
|
+
input: 0.72,
|
|
11408
|
+
output: 0.72,
|
|
11307
11409
|
cacheRead: 0,
|
|
11308
11410
|
cacheWrite: 0,
|
|
11309
11411
|
},
|
|
11310
|
-
contextWindow:
|
|
11311
|
-
maxTokens:
|
|
11412
|
+
contextWindow: 128000,
|
|
11413
|
+
maxTokens: 8192,
|
|
11312
11414
|
},
|
|
11313
11415
|
"meta/llama-3.1-8b": {
|
|
11314
11416
|
id: "meta/llama-3.1-8b",
|
|
@@ -11319,12 +11421,12 @@ export const MODELS = {
|
|
|
11319
11421
|
reasoning: false,
|
|
11320
11422
|
input: ["text"],
|
|
11321
11423
|
cost: {
|
|
11322
|
-
input: 0.
|
|
11323
|
-
output: 0.
|
|
11424
|
+
input: 0.09999999999999999,
|
|
11425
|
+
output: 0.09999999999999999,
|
|
11324
11426
|
cacheRead: 0,
|
|
11325
11427
|
cacheWrite: 0,
|
|
11326
11428
|
},
|
|
11327
|
-
contextWindow:
|
|
11429
|
+
contextWindow: 128000,
|
|
11328
11430
|
maxTokens: 16384,
|
|
11329
11431
|
},
|
|
11330
11432
|
"meta/llama-3.2-11b": {
|
|
@@ -11387,12 +11489,12 @@ export const MODELS = {
|
|
|
11387
11489
|
reasoning: false,
|
|
11388
11490
|
input: ["text", "image"],
|
|
11389
11491
|
cost: {
|
|
11390
|
-
input: 0.
|
|
11391
|
-
output: 0.
|
|
11492
|
+
input: 0.24,
|
|
11493
|
+
output: 0.9700000000000001,
|
|
11392
11494
|
cacheRead: 0,
|
|
11393
11495
|
cacheWrite: 0,
|
|
11394
11496
|
},
|
|
11395
|
-
contextWindow:
|
|
11497
|
+
contextWindow: 128000,
|
|
11396
11498
|
maxTokens: 8192,
|
|
11397
11499
|
},
|
|
11398
11500
|
"meta/llama-4-scout": {
|
|
@@ -11404,12 +11506,12 @@ export const MODELS = {
|
|
|
11404
11506
|
reasoning: false,
|
|
11405
11507
|
input: ["text", "image"],
|
|
11406
11508
|
cost: {
|
|
11407
|
-
input: 0.
|
|
11408
|
-
output: 0.
|
|
11509
|
+
input: 0.16999999999999998,
|
|
11510
|
+
output: 0.66,
|
|
11409
11511
|
cacheRead: 0,
|
|
11410
11512
|
cacheWrite: 0,
|
|
11411
11513
|
},
|
|
11412
|
-
contextWindow:
|
|
11514
|
+
contextWindow: 128000,
|
|
11413
11515
|
maxTokens: 8192,
|
|
11414
11516
|
},
|
|
11415
11517
|
"minimax/minimax-m2": {
|
|
@@ -11440,8 +11542,8 @@ export const MODELS = {
|
|
|
11440
11542
|
cost: {
|
|
11441
11543
|
input: 0.3,
|
|
11442
11544
|
output: 1.2,
|
|
11443
|
-
cacheRead: 0.
|
|
11444
|
-
cacheWrite: 0,
|
|
11545
|
+
cacheRead: 0.03,
|
|
11546
|
+
cacheWrite: 0.375,
|
|
11445
11547
|
},
|
|
11446
11548
|
contextWindow: 204800,
|
|
11447
11549
|
maxTokens: 131072,
|
|
@@ -11676,14 +11778,31 @@ export const MODELS = {
|
|
|
11676
11778
|
reasoning: false,
|
|
11677
11779
|
input: ["text"],
|
|
11678
11780
|
cost: {
|
|
11679
|
-
input: 0.
|
|
11680
|
-
output: 2,
|
|
11781
|
+
input: 0.6,
|
|
11782
|
+
output: 2.5,
|
|
11681
11783
|
cacheRead: 0,
|
|
11682
11784
|
cacheWrite: 0,
|
|
11683
11785
|
},
|
|
11684
11786
|
contextWindow: 131072,
|
|
11685
11787
|
maxTokens: 16384,
|
|
11686
11788
|
},
|
|
11789
|
+
"moonshotai/kimi-k2-0905": {
|
|
11790
|
+
id: "moonshotai/kimi-k2-0905",
|
|
11791
|
+
name: "Kimi K2 0905",
|
|
11792
|
+
api: "anthropic-messages",
|
|
11793
|
+
provider: "vercel-ai-gateway",
|
|
11794
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11795
|
+
reasoning: false,
|
|
11796
|
+
input: ["text"],
|
|
11797
|
+
cost: {
|
|
11798
|
+
input: 0.6,
|
|
11799
|
+
output: 2.5,
|
|
11800
|
+
cacheRead: 0,
|
|
11801
|
+
cacheWrite: 0,
|
|
11802
|
+
},
|
|
11803
|
+
contextWindow: 256000,
|
|
11804
|
+
maxTokens: 16384,
|
|
11805
|
+
},
|
|
11687
11806
|
"moonshotai/kimi-k2-thinking": {
|
|
11688
11807
|
id: "moonshotai/kimi-k2-thinking",
|
|
11689
11808
|
name: "Kimi K2 Thinking",
|
|
@@ -11693,13 +11812,13 @@ export const MODELS = {
|
|
|
11693
11812
|
reasoning: true,
|
|
11694
11813
|
input: ["text"],
|
|
11695
11814
|
cost: {
|
|
11696
|
-
input: 0.
|
|
11697
|
-
output: 2,
|
|
11698
|
-
cacheRead: 0.
|
|
11815
|
+
input: 0.6,
|
|
11816
|
+
output: 2.5,
|
|
11817
|
+
cacheRead: 0.15,
|
|
11699
11818
|
cacheWrite: 0,
|
|
11700
11819
|
},
|
|
11701
|
-
contextWindow:
|
|
11702
|
-
maxTokens:
|
|
11820
|
+
contextWindow: 262114,
|
|
11821
|
+
maxTokens: 262114,
|
|
11703
11822
|
},
|
|
11704
11823
|
"moonshotai/kimi-k2-thinking-turbo": {
|
|
11705
11824
|
id: "moonshotai/kimi-k2-thinking-turbo",
|
|
@@ -11744,13 +11863,13 @@ export const MODELS = {
|
|
|
11744
11863
|
reasoning: true,
|
|
11745
11864
|
input: ["text", "image"],
|
|
11746
11865
|
cost: {
|
|
11747
|
-
input: 0.
|
|
11748
|
-
output:
|
|
11749
|
-
cacheRead: 0,
|
|
11866
|
+
input: 0.6,
|
|
11867
|
+
output: 3,
|
|
11868
|
+
cacheRead: 0.09999999999999999,
|
|
11750
11869
|
cacheWrite: 0,
|
|
11751
11870
|
},
|
|
11752
|
-
contextWindow:
|
|
11753
|
-
maxTokens:
|
|
11871
|
+
contextWindow: 262114,
|
|
11872
|
+
maxTokens: 262114,
|
|
11754
11873
|
},
|
|
11755
11874
|
"nvidia/nemotron-nano-12b-v2-vl": {
|
|
11756
11875
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
@@ -11778,31 +11897,14 @@ export const MODELS = {
|
|
|
11778
11897
|
reasoning: true,
|
|
11779
11898
|
input: ["text"],
|
|
11780
11899
|
cost: {
|
|
11781
|
-
input: 0.
|
|
11782
|
-
output: 0.
|
|
11900
|
+
input: 0.06,
|
|
11901
|
+
output: 0.22999999999999998,
|
|
11783
11902
|
cacheRead: 0,
|
|
11784
11903
|
cacheWrite: 0,
|
|
11785
11904
|
},
|
|
11786
11905
|
contextWindow: 131072,
|
|
11787
11906
|
maxTokens: 131072,
|
|
11788
11907
|
},
|
|
11789
|
-
"openai/codex-mini": {
|
|
11790
|
-
id: "openai/codex-mini",
|
|
11791
|
-
name: "Codex Mini",
|
|
11792
|
-
api: "anthropic-messages",
|
|
11793
|
-
provider: "vercel-ai-gateway",
|
|
11794
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11795
|
-
reasoning: true,
|
|
11796
|
-
input: ["text", "image"],
|
|
11797
|
-
cost: {
|
|
11798
|
-
input: 1.5,
|
|
11799
|
-
output: 6,
|
|
11800
|
-
cacheRead: 0.375,
|
|
11801
|
-
cacheWrite: 0,
|
|
11802
|
-
},
|
|
11803
|
-
contextWindow: 200000,
|
|
11804
|
-
maxTokens: 100000,
|
|
11805
|
-
},
|
|
11806
11908
|
"openai/gpt-4-turbo": {
|
|
11807
11909
|
id: "openai/gpt-4-turbo",
|
|
11808
11910
|
name: "GPT-4 Turbo",
|
|
@@ -11865,7 +11967,7 @@ export const MODELS = {
|
|
|
11865
11967
|
cost: {
|
|
11866
11968
|
input: 0.09999999999999999,
|
|
11867
11969
|
output: 0.39999999999999997,
|
|
11868
|
-
cacheRead: 0.
|
|
11970
|
+
cacheRead: 0.024999999999999998,
|
|
11869
11971
|
cacheWrite: 0,
|
|
11870
11972
|
},
|
|
11871
11973
|
contextWindow: 1047576,
|
|
@@ -11916,7 +12018,7 @@ export const MODELS = {
|
|
|
11916
12018
|
cost: {
|
|
11917
12019
|
input: 1.25,
|
|
11918
12020
|
output: 10,
|
|
11919
|
-
cacheRead: 0.
|
|
12021
|
+
cacheRead: 0.125,
|
|
11920
12022
|
cacheWrite: 0,
|
|
11921
12023
|
},
|
|
11922
12024
|
contextWindow: 400000,
|
|
@@ -11946,11 +12048,11 @@ export const MODELS = {
|
|
|
11946
12048
|
provider: "vercel-ai-gateway",
|
|
11947
12049
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11948
12050
|
reasoning: true,
|
|
11949
|
-
input: ["text"
|
|
12051
|
+
input: ["text"],
|
|
11950
12052
|
cost: {
|
|
11951
12053
|
input: 1.25,
|
|
11952
12054
|
output: 10,
|
|
11953
|
-
cacheRead: 0.
|
|
12055
|
+
cacheRead: 0.125,
|
|
11954
12056
|
cacheWrite: 0,
|
|
11955
12057
|
},
|
|
11956
12058
|
contextWindow: 400000,
|
|
@@ -11967,7 +12069,7 @@ export const MODELS = {
|
|
|
11967
12069
|
cost: {
|
|
11968
12070
|
input: 0.25,
|
|
11969
12071
|
output: 2,
|
|
11970
|
-
cacheRead: 0.
|
|
12072
|
+
cacheRead: 0.024999999999999998,
|
|
11971
12073
|
cacheWrite: 0,
|
|
11972
12074
|
},
|
|
11973
12075
|
contextWindow: 400000,
|
|
@@ -11984,7 +12086,7 @@ export const MODELS = {
|
|
|
11984
12086
|
cost: {
|
|
11985
12087
|
input: 0.049999999999999996,
|
|
11986
12088
|
output: 0.39999999999999997,
|
|
11987
|
-
cacheRead: 0.
|
|
12089
|
+
cacheRead: 0.005,
|
|
11988
12090
|
cacheWrite: 0,
|
|
11989
12091
|
},
|
|
11990
12092
|
contextWindow: 400000,
|
|
@@ -12018,7 +12120,7 @@ export const MODELS = {
|
|
|
12018
12120
|
cost: {
|
|
12019
12121
|
input: 1.25,
|
|
12020
12122
|
output: 10,
|
|
12021
|
-
cacheRead: 0.
|
|
12123
|
+
cacheRead: 0.125,
|
|
12022
12124
|
cacheWrite: 0,
|
|
12023
12125
|
},
|
|
12024
12126
|
contextWindow: 400000,
|
|
@@ -12069,7 +12171,7 @@ export const MODELS = {
|
|
|
12069
12171
|
cost: {
|
|
12070
12172
|
input: 1.25,
|
|
12071
12173
|
output: 10,
|
|
12072
|
-
cacheRead: 0.
|
|
12174
|
+
cacheRead: 0.125,
|
|
12073
12175
|
cacheWrite: 0,
|
|
12074
12176
|
},
|
|
12075
12177
|
contextWindow: 128000,
|
|
@@ -12086,7 +12188,7 @@ export const MODELS = {
|
|
|
12086
12188
|
cost: {
|
|
12087
12189
|
input: 1.25,
|
|
12088
12190
|
output: 10,
|
|
12089
|
-
cacheRead: 0.
|
|
12191
|
+
cacheRead: 0.125,
|
|
12090
12192
|
cacheWrite: 0,
|
|
12091
12193
|
},
|
|
12092
12194
|
contextWindow: 400000,
|
|
@@ -12103,7 +12205,7 @@ export const MODELS = {
|
|
|
12103
12205
|
cost: {
|
|
12104
12206
|
input: 1.75,
|
|
12105
12207
|
output: 14,
|
|
12106
|
-
cacheRead: 0.
|
|
12208
|
+
cacheRead: 0.175,
|
|
12107
12209
|
cacheWrite: 0,
|
|
12108
12210
|
},
|
|
12109
12211
|
contextWindow: 400000,
|
|
@@ -12208,7 +12310,7 @@ export const MODELS = {
|
|
|
12208
12310
|
cacheRead: 0.25,
|
|
12209
12311
|
cacheWrite: 0,
|
|
12210
12312
|
},
|
|
12211
|
-
contextWindow:
|
|
12313
|
+
contextWindow: 1050000,
|
|
12212
12314
|
maxTokens: 128000,
|
|
12213
12315
|
},
|
|
12214
12316
|
"openai/gpt-5.4-pro": {
|
|
@@ -12225,26 +12327,9 @@ export const MODELS = {
|
|
|
12225
12327
|
cacheRead: 0,
|
|
12226
12328
|
cacheWrite: 0,
|
|
12227
12329
|
},
|
|
12228
|
-
contextWindow:
|
|
12330
|
+
contextWindow: 1050000,
|
|
12229
12331
|
maxTokens: 128000,
|
|
12230
12332
|
},
|
|
12231
|
-
"openai/gpt-oss-120b": {
|
|
12232
|
-
id: "openai/gpt-oss-120b",
|
|
12233
|
-
name: "gpt-oss-120b",
|
|
12234
|
-
api: "anthropic-messages",
|
|
12235
|
-
provider: "vercel-ai-gateway",
|
|
12236
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12237
|
-
reasoning: true,
|
|
12238
|
-
input: ["text"],
|
|
12239
|
-
cost: {
|
|
12240
|
-
input: 0.09999999999999999,
|
|
12241
|
-
output: 0.5,
|
|
12242
|
-
cacheRead: 0,
|
|
12243
|
-
cacheWrite: 0,
|
|
12244
|
-
},
|
|
12245
|
-
contextWindow: 131072,
|
|
12246
|
-
maxTokens: 131072,
|
|
12247
|
-
},
|
|
12248
12333
|
"openai/gpt-oss-20b": {
|
|
12249
12334
|
id: "openai/gpt-oss-20b",
|
|
12250
12335
|
name: "gpt-oss-20b",
|
|
@@ -12562,7 +12647,7 @@ export const MODELS = {
|
|
|
12562
12647
|
cost: {
|
|
12563
12648
|
input: 3,
|
|
12564
12649
|
output: 15,
|
|
12565
|
-
cacheRead: 0,
|
|
12650
|
+
cacheRead: 0.75,
|
|
12566
12651
|
cacheWrite: 0,
|
|
12567
12652
|
},
|
|
12568
12653
|
contextWindow: 256000,
|
|
@@ -12636,6 +12721,57 @@ export const MODELS = {
|
|
|
12636
12721
|
contextWindow: 2000000,
|
|
12637
12722
|
maxTokens: 30000,
|
|
12638
12723
|
},
|
|
12724
|
+
"xai/grok-4.20-multi-agent-beta": {
|
|
12725
|
+
id: "xai/grok-4.20-multi-agent-beta",
|
|
12726
|
+
name: "Grok 4.20 Multi Agent Beta",
|
|
12727
|
+
api: "anthropic-messages",
|
|
12728
|
+
provider: "vercel-ai-gateway",
|
|
12729
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12730
|
+
reasoning: true,
|
|
12731
|
+
input: ["text"],
|
|
12732
|
+
cost: {
|
|
12733
|
+
input: 2,
|
|
12734
|
+
output: 6,
|
|
12735
|
+
cacheRead: 0.19999999999999998,
|
|
12736
|
+
cacheWrite: 0,
|
|
12737
|
+
},
|
|
12738
|
+
contextWindow: 2000000,
|
|
12739
|
+
maxTokens: 2000000,
|
|
12740
|
+
},
|
|
12741
|
+
"xai/grok-4.20-non-reasoning-beta": {
|
|
12742
|
+
id: "xai/grok-4.20-non-reasoning-beta",
|
|
12743
|
+
name: "Grok 4.20 Beta Non-Reasoning",
|
|
12744
|
+
api: "anthropic-messages",
|
|
12745
|
+
provider: "vercel-ai-gateway",
|
|
12746
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12747
|
+
reasoning: false,
|
|
12748
|
+
input: ["text", "image"],
|
|
12749
|
+
cost: {
|
|
12750
|
+
input: 2,
|
|
12751
|
+
output: 6,
|
|
12752
|
+
cacheRead: 0.19999999999999998,
|
|
12753
|
+
cacheWrite: 0,
|
|
12754
|
+
},
|
|
12755
|
+
contextWindow: 2000000,
|
|
12756
|
+
maxTokens: 2000000,
|
|
12757
|
+
},
|
|
12758
|
+
"xai/grok-4.20-reasoning-beta": {
|
|
12759
|
+
id: "xai/grok-4.20-reasoning-beta",
|
|
12760
|
+
name: "Grok 4.20 Beta Reasoning",
|
|
12761
|
+
api: "anthropic-messages",
|
|
12762
|
+
provider: "vercel-ai-gateway",
|
|
12763
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12764
|
+
reasoning: true,
|
|
12765
|
+
input: ["text", "image"],
|
|
12766
|
+
cost: {
|
|
12767
|
+
input: 2,
|
|
12768
|
+
output: 6,
|
|
12769
|
+
cacheRead: 0.19999999999999998,
|
|
12770
|
+
cacheWrite: 0,
|
|
12771
|
+
},
|
|
12772
|
+
contextWindow: 2000000,
|
|
12773
|
+
maxTokens: 2000000,
|
|
12774
|
+
},
|
|
12639
12775
|
"xai/grok-code-fast-1": {
|
|
12640
12776
|
id: "xai/grok-code-fast-1",
|
|
12641
12777
|
name: "Grok Code Fast 1",
|
|
@@ -12662,9 +12798,9 @@ export const MODELS = {
|
|
|
12662
12798
|
reasoning: true,
|
|
12663
12799
|
input: ["text"],
|
|
12664
12800
|
cost: {
|
|
12665
|
-
input: 0.
|
|
12666
|
-
output: 0.
|
|
12667
|
-
cacheRead: 0,
|
|
12801
|
+
input: 0.09999999999999999,
|
|
12802
|
+
output: 0.3,
|
|
12803
|
+
cacheRead: 0.02,
|
|
12668
12804
|
cacheWrite: 0,
|
|
12669
12805
|
},
|
|
12670
12806
|
contextWindow: 262144,
|
|
@@ -12681,11 +12817,11 @@ export const MODELS = {
|
|
|
12681
12817
|
cost: {
|
|
12682
12818
|
input: 0.6,
|
|
12683
12819
|
output: 2.2,
|
|
12684
|
-
cacheRead: 0,
|
|
12820
|
+
cacheRead: 0.11,
|
|
12685
12821
|
cacheWrite: 0,
|
|
12686
12822
|
},
|
|
12687
|
-
contextWindow:
|
|
12688
|
-
maxTokens:
|
|
12823
|
+
contextWindow: 128000,
|
|
12824
|
+
maxTokens: 96000,
|
|
12689
12825
|
},
|
|
12690
12826
|
"zai/glm-4.5-air": {
|
|
12691
12827
|
id: "zai/glm-4.5-air",
|
|
@@ -12710,16 +12846,16 @@ export const MODELS = {
|
|
|
12710
12846
|
api: "anthropic-messages",
|
|
12711
12847
|
provider: "vercel-ai-gateway",
|
|
12712
12848
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12713
|
-
reasoning:
|
|
12849
|
+
reasoning: false,
|
|
12714
12850
|
input: ["text", "image"],
|
|
12715
12851
|
cost: {
|
|
12716
12852
|
input: 0.6,
|
|
12717
12853
|
output: 1.7999999999999998,
|
|
12718
|
-
cacheRead: 0,
|
|
12854
|
+
cacheRead: 0.11,
|
|
12719
12855
|
cacheWrite: 0,
|
|
12720
12856
|
},
|
|
12721
|
-
contextWindow:
|
|
12722
|
-
maxTokens:
|
|
12857
|
+
contextWindow: 66000,
|
|
12858
|
+
maxTokens: 16000,
|
|
12723
12859
|
},
|
|
12724
12860
|
"zai/glm-4.6": {
|
|
12725
12861
|
id: "zai/glm-4.6",
|
|
@@ -12781,14 +12917,31 @@ export const MODELS = {
|
|
|
12781
12917
|
reasoning: true,
|
|
12782
12918
|
input: ["text"],
|
|
12783
12919
|
cost: {
|
|
12784
|
-
input: 0.
|
|
12785
|
-
output:
|
|
12786
|
-
cacheRead: 0
|
|
12920
|
+
input: 0.6,
|
|
12921
|
+
output: 2.2,
|
|
12922
|
+
cacheRead: 0,
|
|
12787
12923
|
cacheWrite: 0,
|
|
12788
12924
|
},
|
|
12789
|
-
contextWindow:
|
|
12925
|
+
contextWindow: 200000,
|
|
12790
12926
|
maxTokens: 120000,
|
|
12791
12927
|
},
|
|
12928
|
+
"zai/glm-4.7-flash": {
|
|
12929
|
+
id: "zai/glm-4.7-flash",
|
|
12930
|
+
name: "GLM 4.7 Flash",
|
|
12931
|
+
api: "anthropic-messages",
|
|
12932
|
+
provider: "vercel-ai-gateway",
|
|
12933
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12934
|
+
reasoning: true,
|
|
12935
|
+
input: ["text"],
|
|
12936
|
+
cost: {
|
|
12937
|
+
input: 0.07,
|
|
12938
|
+
output: 0.39999999999999997,
|
|
12939
|
+
cacheRead: 0,
|
|
12940
|
+
cacheWrite: 0,
|
|
12941
|
+
},
|
|
12942
|
+
contextWindow: 200000,
|
|
12943
|
+
maxTokens: 131000,
|
|
12944
|
+
},
|
|
12792
12945
|
"zai/glm-4.7-flashx": {
|
|
12793
12946
|
id: "zai/glm-4.7-flashx",
|
|
12794
12947
|
name: "GLM 4.7 FlashX",
|
|
@@ -12815,9 +12968,9 @@ export const MODELS = {
|
|
|
12815
12968
|
reasoning: true,
|
|
12816
12969
|
input: ["text"],
|
|
12817
12970
|
cost: {
|
|
12818
|
-
input:
|
|
12819
|
-
output:
|
|
12820
|
-
cacheRead: 0.
|
|
12971
|
+
input: 1,
|
|
12972
|
+
output: 3.1999999999999997,
|
|
12973
|
+
cacheRead: 0.19999999999999998,
|
|
12821
12974
|
cacheWrite: 0,
|
|
12822
12975
|
},
|
|
12823
12976
|
contextWindow: 202800,
|
|
@@ -13148,6 +13301,57 @@ export const MODELS = {
|
|
|
13148
13301
|
contextWindow: 2000000,
|
|
13149
13302
|
maxTokens: 30000,
|
|
13150
13303
|
},
|
|
13304
|
+
"grok-4.20-beta-latest-non-reasoning": {
|
|
13305
|
+
id: "grok-4.20-beta-latest-non-reasoning",
|
|
13306
|
+
name: "Grok 4.20 Beta (Non-Reasoning)",
|
|
13307
|
+
api: "openai-completions",
|
|
13308
|
+
provider: "xai",
|
|
13309
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13310
|
+
reasoning: false,
|
|
13311
|
+
input: ["text", "image"],
|
|
13312
|
+
cost: {
|
|
13313
|
+
input: 2,
|
|
13314
|
+
output: 6,
|
|
13315
|
+
cacheRead: 0.2,
|
|
13316
|
+
cacheWrite: 0,
|
|
13317
|
+
},
|
|
13318
|
+
contextWindow: 2000000,
|
|
13319
|
+
maxTokens: 30000,
|
|
13320
|
+
},
|
|
13321
|
+
"grok-4.20-beta-latest-reasoning": {
|
|
13322
|
+
id: "grok-4.20-beta-latest-reasoning",
|
|
13323
|
+
name: "Grok 4.20 Beta (Reasoning)",
|
|
13324
|
+
api: "openai-completions",
|
|
13325
|
+
provider: "xai",
|
|
13326
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13327
|
+
reasoning: true,
|
|
13328
|
+
input: ["text", "image"],
|
|
13329
|
+
cost: {
|
|
13330
|
+
input: 2,
|
|
13331
|
+
output: 6,
|
|
13332
|
+
cacheRead: 0.2,
|
|
13333
|
+
cacheWrite: 0,
|
|
13334
|
+
},
|
|
13335
|
+
contextWindow: 2000000,
|
|
13336
|
+
maxTokens: 30000,
|
|
13337
|
+
},
|
|
13338
|
+
"grok-4.20-multi-agent-beta-latest": {
|
|
13339
|
+
id: "grok-4.20-multi-agent-beta-latest",
|
|
13340
|
+
name: "Grok 4.20 Multi-Agent Beta",
|
|
13341
|
+
api: "openai-completions",
|
|
13342
|
+
provider: "xai",
|
|
13343
|
+
baseUrl: "https://api.x.ai/v1",
|
|
13344
|
+
reasoning: true,
|
|
13345
|
+
input: ["text", "image"],
|
|
13346
|
+
cost: {
|
|
13347
|
+
input: 2,
|
|
13348
|
+
output: 6,
|
|
13349
|
+
cacheRead: 0.2,
|
|
13350
|
+
cacheWrite: 0,
|
|
13351
|
+
},
|
|
13352
|
+
contextWindow: 2000000,
|
|
13353
|
+
maxTokens: 30000,
|
|
13354
|
+
},
|
|
13151
13355
|
"grok-beta": {
|
|
13152
13356
|
id: "grok-beta",
|
|
13153
13357
|
name: "Grok Beta",
|