@mariozechner/pi-ai 0.58.4 → 0.60.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -0
- package/dist/index.d.ts +9 -8
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +0 -8
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +194 -131
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +259 -206
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts +7 -0
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +22 -11
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +3 -0
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +20 -8
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +3 -0
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +3 -0
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/mistral.d.ts.map +1 -1
- package/dist/providers/mistral.js +3 -0
- package/dist/providers/mistral.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +25 -12
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +18 -12
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/register-builtins.d.ts +28 -1
- package/dist/providers/register-builtins.d.ts.map +1 -1
- package/dist/providers/register-builtins.js +170 -47
- package/dist/providers/register-builtins.js.map +1 -1
- package/dist/types.d.ts +1 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +20 -28
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
- package/dist/utils/oauth/google-antigravity.js +22 -19
- package/dist/utils/oauth/google-antigravity.js.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.js +22 -19
- package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
- package/dist/utils/oauth/oauth-page.d.ts +3 -0
- package/dist/utils/oauth/oauth-page.d.ts.map +1 -0
- package/dist/utils/oauth/oauth-page.js +105 -0
- package/dist/utils/oauth/oauth-page.js.map +1 -0
- package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
- package/dist/utils/oauth/openai-codex.js +24 -31
- package/dist/utils/oauth/openai-codex.js.map +1 -1
- package/package.json +39 -5
- package/bedrock-provider.d.ts +0 -1
- package/bedrock-provider.js +0 -1
package/dist/models.generated.js
CHANGED
|
@@ -254,7 +254,7 @@ export const MODELS = {
|
|
|
254
254
|
cacheRead: 0.5,
|
|
255
255
|
cacheWrite: 6.25,
|
|
256
256
|
},
|
|
257
|
-
contextWindow:
|
|
257
|
+
contextWindow: 200000,
|
|
258
258
|
maxTokens: 128000,
|
|
259
259
|
},
|
|
260
260
|
"anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -305,7 +305,7 @@ export const MODELS = {
|
|
|
305
305
|
cacheRead: 0.3,
|
|
306
306
|
cacheWrite: 3.75,
|
|
307
307
|
},
|
|
308
|
-
contextWindow:
|
|
308
|
+
contextWindow: 200000,
|
|
309
309
|
maxTokens: 64000,
|
|
310
310
|
},
|
|
311
311
|
"deepseek.r1-v1:0": {
|
|
@@ -407,7 +407,7 @@ export const MODELS = {
|
|
|
407
407
|
cacheRead: 0.5,
|
|
408
408
|
cacheWrite: 6.25,
|
|
409
409
|
},
|
|
410
|
-
contextWindow:
|
|
410
|
+
contextWindow: 200000,
|
|
411
411
|
maxTokens: 128000,
|
|
412
412
|
},
|
|
413
413
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -458,7 +458,7 @@ export const MODELS = {
|
|
|
458
458
|
cacheRead: 0.3,
|
|
459
459
|
cacheWrite: 3.75,
|
|
460
460
|
},
|
|
461
|
-
contextWindow:
|
|
461
|
+
contextWindow: 200000,
|
|
462
462
|
maxTokens: 64000,
|
|
463
463
|
},
|
|
464
464
|
"global.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
@@ -509,7 +509,7 @@ export const MODELS = {
|
|
|
509
509
|
cacheRead: 0.5,
|
|
510
510
|
cacheWrite: 6.25,
|
|
511
511
|
},
|
|
512
|
-
contextWindow:
|
|
512
|
+
contextWindow: 200000,
|
|
513
513
|
maxTokens: 128000,
|
|
514
514
|
},
|
|
515
515
|
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -560,7 +560,7 @@ export const MODELS = {
|
|
|
560
560
|
cacheRead: 0.3,
|
|
561
561
|
cacheWrite: 3.75,
|
|
562
562
|
},
|
|
563
|
-
contextWindow:
|
|
563
|
+
contextWindow: 200000,
|
|
564
564
|
maxTokens: 64000,
|
|
565
565
|
},
|
|
566
566
|
"google.gemma-3-27b-it": {
|
|
@@ -1291,7 +1291,7 @@ export const MODELS = {
|
|
|
1291
1291
|
cacheRead: 0.5,
|
|
1292
1292
|
cacheWrite: 6.25,
|
|
1293
1293
|
},
|
|
1294
|
-
contextWindow:
|
|
1294
|
+
contextWindow: 200000,
|
|
1295
1295
|
maxTokens: 128000,
|
|
1296
1296
|
},
|
|
1297
1297
|
"us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
@@ -1342,7 +1342,7 @@ export const MODELS = {
|
|
|
1342
1342
|
cacheRead: 0.3,
|
|
1343
1343
|
cacheWrite: 3.75,
|
|
1344
1344
|
},
|
|
1345
|
-
contextWindow:
|
|
1345
|
+
contextWindow: 200000,
|
|
1346
1346
|
maxTokens: 64000,
|
|
1347
1347
|
},
|
|
1348
1348
|
"writer.palmyra-x4-v1:0": {
|
|
@@ -2301,6 +2301,40 @@ export const MODELS = {
|
|
|
2301
2301
|
contextWindow: 272000,
|
|
2302
2302
|
maxTokens: 128000,
|
|
2303
2303
|
},
|
|
2304
|
+
"gpt-5.4-mini": {
|
|
2305
|
+
id: "gpt-5.4-mini",
|
|
2306
|
+
name: "GPT-5.4 mini",
|
|
2307
|
+
api: "azure-openai-responses",
|
|
2308
|
+
provider: "azure-openai-responses",
|
|
2309
|
+
baseUrl: "",
|
|
2310
|
+
reasoning: true,
|
|
2311
|
+
input: ["text", "image"],
|
|
2312
|
+
cost: {
|
|
2313
|
+
input: 0.75,
|
|
2314
|
+
output: 4.5,
|
|
2315
|
+
cacheRead: 0.075,
|
|
2316
|
+
cacheWrite: 0,
|
|
2317
|
+
},
|
|
2318
|
+
contextWindow: 400000,
|
|
2319
|
+
maxTokens: 128000,
|
|
2320
|
+
},
|
|
2321
|
+
"gpt-5.4-nano": {
|
|
2322
|
+
id: "gpt-5.4-nano",
|
|
2323
|
+
name: "GPT-5.4 nano",
|
|
2324
|
+
api: "azure-openai-responses",
|
|
2325
|
+
provider: "azure-openai-responses",
|
|
2326
|
+
baseUrl: "",
|
|
2327
|
+
reasoning: true,
|
|
2328
|
+
input: ["text", "image"],
|
|
2329
|
+
cost: {
|
|
2330
|
+
input: 0.2,
|
|
2331
|
+
output: 1.25,
|
|
2332
|
+
cacheRead: 0.02,
|
|
2333
|
+
cacheWrite: 0,
|
|
2334
|
+
},
|
|
2335
|
+
contextWindow: 400000,
|
|
2336
|
+
maxTokens: 128000,
|
|
2337
|
+
},
|
|
2304
2338
|
"gpt-5.4-pro": {
|
|
2305
2339
|
id: "gpt-5.4-pro",
|
|
2306
2340
|
name: "GPT-5.4 Pro",
|
|
@@ -2577,7 +2611,7 @@ export const MODELS = {
|
|
|
2577
2611
|
cacheRead: 0,
|
|
2578
2612
|
cacheWrite: 0,
|
|
2579
2613
|
},
|
|
2580
|
-
contextWindow:
|
|
2614
|
+
contextWindow: 1000000,
|
|
2581
2615
|
maxTokens: 64000,
|
|
2582
2616
|
},
|
|
2583
2617
|
"claude-sonnet-4": {
|
|
@@ -2631,7 +2665,7 @@ export const MODELS = {
|
|
|
2631
2665
|
cacheRead: 0,
|
|
2632
2666
|
cacheWrite: 0,
|
|
2633
2667
|
},
|
|
2634
|
-
contextWindow:
|
|
2668
|
+
contextWindow: 1000000,
|
|
2635
2669
|
maxTokens: 32000,
|
|
2636
2670
|
},
|
|
2637
2671
|
"gemini-2.5-pro": {
|
|
@@ -2928,6 +2962,42 @@ export const MODELS = {
|
|
|
2928
2962
|
contextWindow: 400000,
|
|
2929
2963
|
maxTokens: 128000,
|
|
2930
2964
|
},
|
|
2965
|
+
"gpt-5.4-mini": {
|
|
2966
|
+
id: "gpt-5.4-mini",
|
|
2967
|
+
name: "GPT-5.4 mini",
|
|
2968
|
+
api: "openai-responses",
|
|
2969
|
+
provider: "github-copilot",
|
|
2970
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
2971
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
2972
|
+
reasoning: true,
|
|
2973
|
+
input: ["text", "image"],
|
|
2974
|
+
cost: {
|
|
2975
|
+
input: 0,
|
|
2976
|
+
output: 0,
|
|
2977
|
+
cacheRead: 0,
|
|
2978
|
+
cacheWrite: 0,
|
|
2979
|
+
},
|
|
2980
|
+
contextWindow: 400000,
|
|
2981
|
+
maxTokens: 128000,
|
|
2982
|
+
},
|
|
2983
|
+
"gpt-5.4-nano": {
|
|
2984
|
+
id: "gpt-5.4-nano",
|
|
2985
|
+
name: "GPT-5.4 nano",
|
|
2986
|
+
api: "openai-responses",
|
|
2987
|
+
provider: "github-copilot",
|
|
2988
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
2989
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
2990
|
+
reasoning: true,
|
|
2991
|
+
input: ["text", "image"],
|
|
2992
|
+
cost: {
|
|
2993
|
+
input: 0,
|
|
2994
|
+
output: 0,
|
|
2995
|
+
cacheRead: 0,
|
|
2996
|
+
cacheWrite: 0,
|
|
2997
|
+
},
|
|
2998
|
+
contextWindow: 400000,
|
|
2999
|
+
maxTokens: 128000,
|
|
3000
|
+
},
|
|
2931
3001
|
"grok-code-fast-1": {
|
|
2932
3002
|
id: "grok-code-fast-1",
|
|
2933
3003
|
name: "Grok Code Fast 1",
|
|
@@ -5503,6 +5573,40 @@ export const MODELS = {
|
|
|
5503
5573
|
contextWindow: 272000,
|
|
5504
5574
|
maxTokens: 128000,
|
|
5505
5575
|
},
|
|
5576
|
+
"gpt-5.4-mini": {
|
|
5577
|
+
id: "gpt-5.4-mini",
|
|
5578
|
+
name: "GPT-5.4 mini",
|
|
5579
|
+
api: "openai-responses",
|
|
5580
|
+
provider: "openai",
|
|
5581
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5582
|
+
reasoning: true,
|
|
5583
|
+
input: ["text", "image"],
|
|
5584
|
+
cost: {
|
|
5585
|
+
input: 0.75,
|
|
5586
|
+
output: 4.5,
|
|
5587
|
+
cacheRead: 0.075,
|
|
5588
|
+
cacheWrite: 0,
|
|
5589
|
+
},
|
|
5590
|
+
contextWindow: 400000,
|
|
5591
|
+
maxTokens: 128000,
|
|
5592
|
+
},
|
|
5593
|
+
"gpt-5.4-nano": {
|
|
5594
|
+
id: "gpt-5.4-nano",
|
|
5595
|
+
name: "GPT-5.4 nano",
|
|
5596
|
+
api: "openai-responses",
|
|
5597
|
+
provider: "openai",
|
|
5598
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5599
|
+
reasoning: true,
|
|
5600
|
+
input: ["text", "image"],
|
|
5601
|
+
cost: {
|
|
5602
|
+
input: 0.2,
|
|
5603
|
+
output: 1.25,
|
|
5604
|
+
cacheRead: 0.02,
|
|
5605
|
+
cacheWrite: 0,
|
|
5606
|
+
},
|
|
5607
|
+
contextWindow: 400000,
|
|
5608
|
+
maxTokens: 128000,
|
|
5609
|
+
},
|
|
5506
5610
|
"gpt-5.4-pro": {
|
|
5507
5611
|
id: "gpt-5.4-pro",
|
|
5508
5612
|
name: "GPT-5.4 Pro",
|
|
@@ -5966,23 +6070,6 @@ export const MODELS = {
|
|
|
5966
6070
|
contextWindow: 1048576,
|
|
5967
6071
|
maxTokens: 65536,
|
|
5968
6072
|
},
|
|
5969
|
-
"gemini-3-pro": {
|
|
5970
|
-
id: "gemini-3-pro",
|
|
5971
|
-
name: "Gemini 3 Pro",
|
|
5972
|
-
api: "google-generative-ai",
|
|
5973
|
-
provider: "opencode",
|
|
5974
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
5975
|
-
reasoning: true,
|
|
5976
|
-
input: ["text", "image"],
|
|
5977
|
-
cost: {
|
|
5978
|
-
input: 2,
|
|
5979
|
-
output: 12,
|
|
5980
|
-
cacheRead: 0.2,
|
|
5981
|
-
cacheWrite: 0,
|
|
5982
|
-
},
|
|
5983
|
-
contextWindow: 1048576,
|
|
5984
|
-
maxTokens: 65536,
|
|
5985
|
-
},
|
|
5986
6073
|
"gemini-3.1-pro": {
|
|
5987
6074
|
id: "gemini-3.1-pro",
|
|
5988
6075
|
name: "Gemini 3.1 Pro Preview",
|
|
@@ -6000,40 +6087,6 @@ export const MODELS = {
|
|
|
6000
6087
|
contextWindow: 1048576,
|
|
6001
6088
|
maxTokens: 65536,
|
|
6002
6089
|
},
|
|
6003
|
-
"glm-4.6": {
|
|
6004
|
-
id: "glm-4.6",
|
|
6005
|
-
name: "GLM-4.6",
|
|
6006
|
-
api: "openai-completions",
|
|
6007
|
-
provider: "opencode",
|
|
6008
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6009
|
-
reasoning: true,
|
|
6010
|
-
input: ["text"],
|
|
6011
|
-
cost: {
|
|
6012
|
-
input: 0.6,
|
|
6013
|
-
output: 2.2,
|
|
6014
|
-
cacheRead: 0.1,
|
|
6015
|
-
cacheWrite: 0,
|
|
6016
|
-
},
|
|
6017
|
-
contextWindow: 204800,
|
|
6018
|
-
maxTokens: 131072,
|
|
6019
|
-
},
|
|
6020
|
-
"glm-4.7": {
|
|
6021
|
-
id: "glm-4.7",
|
|
6022
|
-
name: "GLM-4.7",
|
|
6023
|
-
api: "openai-completions",
|
|
6024
|
-
provider: "opencode",
|
|
6025
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6026
|
-
reasoning: true,
|
|
6027
|
-
input: ["text"],
|
|
6028
|
-
cost: {
|
|
6029
|
-
input: 0.6,
|
|
6030
|
-
output: 2.2,
|
|
6031
|
-
cacheRead: 0.1,
|
|
6032
|
-
cacheWrite: 0,
|
|
6033
|
-
},
|
|
6034
|
-
contextWindow: 204800,
|
|
6035
|
-
maxTokens: 131072,
|
|
6036
|
-
},
|
|
6037
6090
|
"glm-5": {
|
|
6038
6091
|
id: "glm-5",
|
|
6039
6092
|
name: "GLM-5",
|
|
@@ -6289,23 +6342,6 @@ export const MODELS = {
|
|
|
6289
6342
|
contextWindow: 262144,
|
|
6290
6343
|
maxTokens: 65536,
|
|
6291
6344
|
},
|
|
6292
|
-
"minimax-m2.1": {
|
|
6293
|
-
id: "minimax-m2.1",
|
|
6294
|
-
name: "MiniMax M2.1",
|
|
6295
|
-
api: "openai-completions",
|
|
6296
|
-
provider: "opencode",
|
|
6297
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6298
|
-
reasoning: true,
|
|
6299
|
-
input: ["text"],
|
|
6300
|
-
cost: {
|
|
6301
|
-
input: 0.3,
|
|
6302
|
-
output: 1.2,
|
|
6303
|
-
cacheRead: 0.1,
|
|
6304
|
-
cacheWrite: 0,
|
|
6305
|
-
},
|
|
6306
|
-
contextWindow: 204800,
|
|
6307
|
-
maxTokens: 131072,
|
|
6308
|
-
},
|
|
6309
6345
|
"minimax-m2.5": {
|
|
6310
6346
|
id: "minimax-m2.5",
|
|
6311
6347
|
name: "MiniMax M2.5",
|
|
@@ -7101,13 +7137,13 @@ export const MODELS = {
|
|
|
7101
7137
|
reasoning: true,
|
|
7102
7138
|
input: ["text"],
|
|
7103
7139
|
cost: {
|
|
7104
|
-
input: 0.
|
|
7105
|
-
output: 0.
|
|
7106
|
-
cacheRead: 0
|
|
7140
|
+
input: 0.255,
|
|
7141
|
+
output: 0.39999999999999997,
|
|
7142
|
+
cacheRead: 0,
|
|
7107
7143
|
cacheWrite: 0,
|
|
7108
7144
|
},
|
|
7109
7145
|
contextWindow: 163840,
|
|
7110
|
-
maxTokens:
|
|
7146
|
+
maxTokens: 163840,
|
|
7111
7147
|
},
|
|
7112
7148
|
"deepseek/deepseek-v3.2-exp": {
|
|
7113
7149
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -7662,7 +7698,7 @@ export const MODELS = {
|
|
|
7662
7698
|
reasoning: true,
|
|
7663
7699
|
input: ["text"],
|
|
7664
7700
|
cost: {
|
|
7665
|
-
input: 0.
|
|
7701
|
+
input: 0.19999999999999998,
|
|
7666
7702
|
output: 1.2,
|
|
7667
7703
|
cacheRead: 0,
|
|
7668
7704
|
cacheWrite: 0,
|
|
@@ -7698,7 +7734,7 @@ export const MODELS = {
|
|
|
7698
7734
|
cost: {
|
|
7699
7735
|
input: 0.3,
|
|
7700
7736
|
output: 0.8999999999999999,
|
|
7701
|
-
cacheRead: 0,
|
|
7737
|
+
cacheRead: 0.03,
|
|
7702
7738
|
cacheWrite: 0,
|
|
7703
7739
|
},
|
|
7704
7740
|
contextWindow: 256000,
|
|
@@ -7715,7 +7751,7 @@ export const MODELS = {
|
|
|
7715
7751
|
cost: {
|
|
7716
7752
|
input: 0.39999999999999997,
|
|
7717
7753
|
output: 2,
|
|
7718
|
-
cacheRead: 0,
|
|
7754
|
+
cacheRead: 0.04,
|
|
7719
7755
|
cacheWrite: 0,
|
|
7720
7756
|
},
|
|
7721
7757
|
contextWindow: 262144,
|
|
@@ -7732,7 +7768,7 @@ export const MODELS = {
|
|
|
7732
7768
|
cost: {
|
|
7733
7769
|
input: 0.39999999999999997,
|
|
7734
7770
|
output: 2,
|
|
7735
|
-
cacheRead: 0,
|
|
7771
|
+
cacheRead: 0.04,
|
|
7736
7772
|
cacheWrite: 0,
|
|
7737
7773
|
},
|
|
7738
7774
|
contextWindow: 131072,
|
|
@@ -7749,7 +7785,7 @@ export const MODELS = {
|
|
|
7749
7785
|
cost: {
|
|
7750
7786
|
input: 0.09999999999999999,
|
|
7751
7787
|
output: 0.3,
|
|
7752
|
-
cacheRead: 0,
|
|
7788
|
+
cacheRead: 0.01,
|
|
7753
7789
|
cacheWrite: 0,
|
|
7754
7790
|
},
|
|
7755
7791
|
contextWindow: 131072,
|
|
@@ -7766,7 +7802,7 @@ export const MODELS = {
|
|
|
7766
7802
|
cost: {
|
|
7767
7803
|
input: 0.19999999999999998,
|
|
7768
7804
|
output: 0.19999999999999998,
|
|
7769
|
-
cacheRead: 0,
|
|
7805
|
+
cacheRead: 0.02,
|
|
7770
7806
|
cacheWrite: 0,
|
|
7771
7807
|
},
|
|
7772
7808
|
contextWindow: 262144,
|
|
@@ -7783,7 +7819,7 @@ export const MODELS = {
|
|
|
7783
7819
|
cost: {
|
|
7784
7820
|
input: 0.09999999999999999,
|
|
7785
7821
|
output: 0.09999999999999999,
|
|
7786
|
-
cacheRead: 0,
|
|
7822
|
+
cacheRead: 0.01,
|
|
7787
7823
|
cacheWrite: 0,
|
|
7788
7824
|
},
|
|
7789
7825
|
contextWindow: 131072,
|
|
@@ -7800,7 +7836,7 @@ export const MODELS = {
|
|
|
7800
7836
|
cost: {
|
|
7801
7837
|
input: 0.15,
|
|
7802
7838
|
output: 0.15,
|
|
7803
|
-
cacheRead: 0,
|
|
7839
|
+
cacheRead: 0.015,
|
|
7804
7840
|
cacheWrite: 0,
|
|
7805
7841
|
},
|
|
7806
7842
|
contextWindow: 262144,
|
|
@@ -7817,7 +7853,7 @@ export const MODELS = {
|
|
|
7817
7853
|
cost: {
|
|
7818
7854
|
input: 2,
|
|
7819
7855
|
output: 6,
|
|
7820
|
-
cacheRead: 0,
|
|
7856
|
+
cacheRead: 0.19999999999999998,
|
|
7821
7857
|
cacheWrite: 0,
|
|
7822
7858
|
},
|
|
7823
7859
|
contextWindow: 128000,
|
|
@@ -7834,7 +7870,7 @@ export const MODELS = {
|
|
|
7834
7870
|
cost: {
|
|
7835
7871
|
input: 2,
|
|
7836
7872
|
output: 6,
|
|
7837
|
-
cacheRead: 0,
|
|
7873
|
+
cacheRead: 0.19999999999999998,
|
|
7838
7874
|
cacheWrite: 0,
|
|
7839
7875
|
},
|
|
7840
7876
|
contextWindow: 131072,
|
|
@@ -7851,7 +7887,7 @@ export const MODELS = {
|
|
|
7851
7887
|
cost: {
|
|
7852
7888
|
input: 2,
|
|
7853
7889
|
output: 6,
|
|
7854
|
-
cacheRead: 0,
|
|
7890
|
+
cacheRead: 0.19999999999999998,
|
|
7855
7891
|
cacheWrite: 0,
|
|
7856
7892
|
},
|
|
7857
7893
|
contextWindow: 131072,
|
|
@@ -7868,7 +7904,7 @@ export const MODELS = {
|
|
|
7868
7904
|
cost: {
|
|
7869
7905
|
input: 0.5,
|
|
7870
7906
|
output: 1.5,
|
|
7871
|
-
cacheRead: 0,
|
|
7907
|
+
cacheRead: 0.049999999999999996,
|
|
7872
7908
|
cacheWrite: 0,
|
|
7873
7909
|
},
|
|
7874
7910
|
contextWindow: 262144,
|
|
@@ -7885,7 +7921,7 @@ export const MODELS = {
|
|
|
7885
7921
|
cost: {
|
|
7886
7922
|
input: 0.39999999999999997,
|
|
7887
7923
|
output: 2,
|
|
7888
|
-
cacheRead: 0,
|
|
7924
|
+
cacheRead: 0.04,
|
|
7889
7925
|
cacheWrite: 0,
|
|
7890
7926
|
},
|
|
7891
7927
|
contextWindow: 131072,
|
|
@@ -7902,7 +7938,7 @@ export const MODELS = {
|
|
|
7902
7938
|
cost: {
|
|
7903
7939
|
input: 0.39999999999999997,
|
|
7904
7940
|
output: 2,
|
|
7905
|
-
cacheRead: 0,
|
|
7941
|
+
cacheRead: 0.04,
|
|
7906
7942
|
cacheWrite: 0,
|
|
7907
7943
|
},
|
|
7908
7944
|
contextWindow: 131072,
|
|
@@ -7936,7 +7972,7 @@ export const MODELS = {
|
|
|
7936
7972
|
cost: {
|
|
7937
7973
|
input: 0.19999999999999998,
|
|
7938
7974
|
output: 0.6,
|
|
7939
|
-
cacheRead: 0,
|
|
7975
|
+
cacheRead: 0.02,
|
|
7940
7976
|
cacheWrite: 0,
|
|
7941
7977
|
},
|
|
7942
7978
|
contextWindow: 32768,
|
|
@@ -7959,6 +7995,23 @@ export const MODELS = {
|
|
|
7959
7995
|
contextWindow: 32768,
|
|
7960
7996
|
maxTokens: 16384,
|
|
7961
7997
|
},
|
|
7998
|
+
"mistralai/mistral-small-2603": {
|
|
7999
|
+
id: "mistralai/mistral-small-2603",
|
|
8000
|
+
name: "Mistral: Mistral Small 4",
|
|
8001
|
+
api: "openai-completions",
|
|
8002
|
+
provider: "openrouter",
|
|
8003
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8004
|
+
reasoning: true,
|
|
8005
|
+
input: ["text", "image"],
|
|
8006
|
+
cost: {
|
|
8007
|
+
input: 0.15,
|
|
8008
|
+
output: 0.6,
|
|
8009
|
+
cacheRead: 0.015,
|
|
8010
|
+
cacheWrite: 0,
|
|
8011
|
+
},
|
|
8012
|
+
contextWindow: 262144,
|
|
8013
|
+
maxTokens: 4096,
|
|
8014
|
+
},
|
|
7962
8015
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
7963
8016
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
7964
8017
|
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
@@ -8004,7 +8057,7 @@ export const MODELS = {
|
|
|
8004
8057
|
cost: {
|
|
8005
8058
|
input: 0.09999999999999999,
|
|
8006
8059
|
output: 0.3,
|
|
8007
|
-
cacheRead: 0,
|
|
8060
|
+
cacheRead: 0.01,
|
|
8008
8061
|
cacheWrite: 0,
|
|
8009
8062
|
},
|
|
8010
8063
|
contextWindow: 32768,
|
|
@@ -8021,7 +8074,7 @@ export const MODELS = {
|
|
|
8021
8074
|
cost: {
|
|
8022
8075
|
input: 2,
|
|
8023
8076
|
output: 6,
|
|
8024
|
-
cacheRead: 0,
|
|
8077
|
+
cacheRead: 0.19999999999999998,
|
|
8025
8078
|
cacheWrite: 0,
|
|
8026
8079
|
},
|
|
8027
8080
|
contextWindow: 65536,
|
|
@@ -8055,7 +8108,7 @@ export const MODELS = {
|
|
|
8055
8108
|
cost: {
|
|
8056
8109
|
input: 2,
|
|
8057
8110
|
output: 6,
|
|
8058
|
-
cacheRead: 0,
|
|
8111
|
+
cacheRead: 0.19999999999999998,
|
|
8059
8112
|
cacheWrite: 0,
|
|
8060
8113
|
},
|
|
8061
8114
|
contextWindow: 131072,
|
|
@@ -8072,7 +8125,7 @@ export const MODELS = {
|
|
|
8072
8125
|
cost: {
|
|
8073
8126
|
input: 0.09999999999999999,
|
|
8074
8127
|
output: 0.3,
|
|
8075
|
-
cacheRead: 0,
|
|
8128
|
+
cacheRead: 0.01,
|
|
8076
8129
|
cacheWrite: 0,
|
|
8077
8130
|
},
|
|
8078
8131
|
contextWindow: 32000,
|
|
@@ -8945,6 +8998,40 @@ export const MODELS = {
|
|
|
8945
8998
|
contextWindow: 1050000,
|
|
8946
8999
|
maxTokens: 128000,
|
|
8947
9000
|
},
|
|
9001
|
+
"openai/gpt-5.4-mini": {
|
|
9002
|
+
id: "openai/gpt-5.4-mini",
|
|
9003
|
+
name: "OpenAI: GPT-5.4 Mini",
|
|
9004
|
+
api: "openai-completions",
|
|
9005
|
+
provider: "openrouter",
|
|
9006
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9007
|
+
reasoning: true,
|
|
9008
|
+
input: ["text", "image"],
|
|
9009
|
+
cost: {
|
|
9010
|
+
input: 0.75,
|
|
9011
|
+
output: 4.5,
|
|
9012
|
+
cacheRead: 0.075,
|
|
9013
|
+
cacheWrite: 0,
|
|
9014
|
+
},
|
|
9015
|
+
contextWindow: 400000,
|
|
9016
|
+
maxTokens: 128000,
|
|
9017
|
+
},
|
|
9018
|
+
"openai/gpt-5.4-nano": {
|
|
9019
|
+
id: "openai/gpt-5.4-nano",
|
|
9020
|
+
name: "OpenAI: GPT-5.4 Nano",
|
|
9021
|
+
api: "openai-completions",
|
|
9022
|
+
provider: "openrouter",
|
|
9023
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9024
|
+
reasoning: true,
|
|
9025
|
+
input: ["text", "image"],
|
|
9026
|
+
cost: {
|
|
9027
|
+
input: 0.19999999999999998,
|
|
9028
|
+
output: 1.25,
|
|
9029
|
+
cacheRead: 0.02,
|
|
9030
|
+
cacheWrite: 0,
|
|
9031
|
+
},
|
|
9032
|
+
contextWindow: 400000,
|
|
9033
|
+
maxTokens: 128000,
|
|
9034
|
+
},
|
|
8948
9035
|
"openai/gpt-5.4-pro": {
|
|
8949
9036
|
id: "openai/gpt-5.4-pro",
|
|
8950
9037
|
name: "OpenAI: GPT-5.4 Pro",
|
|
@@ -9053,7 +9140,7 @@ export const MODELS = {
|
|
|
9053
9140
|
api: "openai-completions",
|
|
9054
9141
|
provider: "openrouter",
|
|
9055
9142
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9056
|
-
reasoning:
|
|
9143
|
+
reasoning: true,
|
|
9057
9144
|
input: ["text", "image"],
|
|
9058
9145
|
cost: {
|
|
9059
9146
|
input: 15,
|
|
@@ -9104,7 +9191,7 @@ export const MODELS = {
|
|
|
9104
9191
|
api: "openai-completions",
|
|
9105
9192
|
provider: "openrouter",
|
|
9106
9193
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9107
|
-
reasoning:
|
|
9194
|
+
reasoning: true,
|
|
9108
9195
|
input: ["text"],
|
|
9109
9196
|
cost: {
|
|
9110
9197
|
input: 1.1,
|
|
@@ -9121,7 +9208,7 @@ export const MODELS = {
|
|
|
9121
9208
|
api: "openai-completions",
|
|
9122
9209
|
provider: "openrouter",
|
|
9123
9210
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9124
|
-
reasoning:
|
|
9211
|
+
reasoning: true,
|
|
9125
9212
|
input: ["text"],
|
|
9126
9213
|
cost: {
|
|
9127
9214
|
input: 1.1,
|
|
@@ -10523,7 +10610,7 @@ export const MODELS = {
|
|
|
10523
10610
|
cacheRead: 0.119,
|
|
10524
10611
|
cacheWrite: 0,
|
|
10525
10612
|
},
|
|
10526
|
-
contextWindow:
|
|
10613
|
+
contextWindow: 80000,
|
|
10527
10614
|
maxTokens: 131072,
|
|
10528
10615
|
},
|
|
10529
10616
|
"z-ai/glm-5-turbo": {
|
|
@@ -10554,7 +10641,7 @@ export const MODELS = {
|
|
|
10554
10641
|
reasoning: true,
|
|
10555
10642
|
input: ["text"],
|
|
10556
10643
|
cost: {
|
|
10557
|
-
input: 0.
|
|
10644
|
+
input: 0.12,
|
|
10558
10645
|
output: 0.24,
|
|
10559
10646
|
cacheRead: 0,
|
|
10560
10647
|
cacheWrite: 0,
|
|
@@ -10607,7 +10694,7 @@ export const MODELS = {
|
|
|
10607
10694
|
cost: {
|
|
10608
10695
|
input: 0.29,
|
|
10609
10696
|
output: 0.59,
|
|
10610
|
-
cacheRead: 0,
|
|
10697
|
+
cacheRead: 0.145,
|
|
10611
10698
|
cacheWrite: 0,
|
|
10612
10699
|
},
|
|
10613
10700
|
contextWindow: 131072,
|
|
@@ -10622,9 +10709,9 @@ export const MODELS = {
|
|
|
10622
10709
|
reasoning: true,
|
|
10623
10710
|
input: ["text", "image"],
|
|
10624
10711
|
cost: {
|
|
10625
|
-
input: 0.
|
|
10626
|
-
output: 2.
|
|
10627
|
-
cacheRead: 0,
|
|
10712
|
+
input: 0.22999999999999998,
|
|
10713
|
+
output: 2.3,
|
|
10714
|
+
cacheRead: 0.19999999999999998,
|
|
10628
10715
|
cacheWrite: 0,
|
|
10629
10716
|
},
|
|
10630
10717
|
contextWindow: 262114,
|
|
@@ -10641,7 +10728,7 @@ export const MODELS = {
|
|
|
10641
10728
|
cost: {
|
|
10642
10729
|
input: 0.39999999999999997,
|
|
10643
10730
|
output: 1.5999999999999999,
|
|
10644
|
-
cacheRead: 0,
|
|
10731
|
+
cacheRead: 0.022,
|
|
10645
10732
|
cacheWrite: 0,
|
|
10646
10733
|
},
|
|
10647
10734
|
contextWindow: 262144,
|
|
@@ -11134,7 +11221,7 @@ export const MODELS = {
|
|
|
11134
11221
|
cost: {
|
|
11135
11222
|
input: 0.27,
|
|
11136
11223
|
output: 1,
|
|
11137
|
-
cacheRead: 0,
|
|
11224
|
+
cacheRead: 0.135,
|
|
11138
11225
|
cacheWrite: 0,
|
|
11139
11226
|
},
|
|
11140
11227
|
contextWindow: 131072,
|
|
@@ -11149,9 +11236,9 @@ export const MODELS = {
|
|
|
11149
11236
|
reasoning: false,
|
|
11150
11237
|
input: ["text"],
|
|
11151
11238
|
cost: {
|
|
11152
|
-
input: 0.
|
|
11153
|
-
output: 0.
|
|
11154
|
-
cacheRead: 0.
|
|
11239
|
+
input: 0.28,
|
|
11240
|
+
output: 0.42,
|
|
11241
|
+
cacheRead: 0.028,
|
|
11155
11242
|
cacheWrite: 0,
|
|
11156
11243
|
},
|
|
11157
11244
|
contextWindow: 128000,
|
|
@@ -11185,7 +11272,7 @@ export const MODELS = {
|
|
|
11185
11272
|
cost: {
|
|
11186
11273
|
input: 0.15,
|
|
11187
11274
|
output: 0.6,
|
|
11188
|
-
cacheRead: 0,
|
|
11275
|
+
cacheRead: 0.024999999999999998,
|
|
11189
11276
|
cacheWrite: 0,
|
|
11190
11277
|
},
|
|
11191
11278
|
contextWindow: 1048576,
|
|
@@ -11202,7 +11289,7 @@ export const MODELS = {
|
|
|
11202
11289
|
cost: {
|
|
11203
11290
|
input: 0.075,
|
|
11204
11291
|
output: 0.3,
|
|
11205
|
-
cacheRead: 0,
|
|
11292
|
+
cacheRead: 0.02,
|
|
11206
11293
|
cacheWrite: 0,
|
|
11207
11294
|
},
|
|
11208
11295
|
contextWindow: 1048576,
|
|
@@ -11242,40 +11329,6 @@ export const MODELS = {
|
|
|
11242
11329
|
contextWindow: 1048576,
|
|
11243
11330
|
maxTokens: 65536,
|
|
11244
11331
|
},
|
|
11245
|
-
"google/gemini-2.5-flash-lite-preview-09-2025": {
|
|
11246
|
-
id: "google/gemini-2.5-flash-lite-preview-09-2025",
|
|
11247
|
-
name: "Gemini 2.5 Flash Lite Preview 09-2025",
|
|
11248
|
-
api: "anthropic-messages",
|
|
11249
|
-
provider: "vercel-ai-gateway",
|
|
11250
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11251
|
-
reasoning: true,
|
|
11252
|
-
input: ["text", "image"],
|
|
11253
|
-
cost: {
|
|
11254
|
-
input: 0.09999999999999999,
|
|
11255
|
-
output: 0.39999999999999997,
|
|
11256
|
-
cacheRead: 0.01,
|
|
11257
|
-
cacheWrite: 0,
|
|
11258
|
-
},
|
|
11259
|
-
contextWindow: 1048576,
|
|
11260
|
-
maxTokens: 65536,
|
|
11261
|
-
},
|
|
11262
|
-
"google/gemini-2.5-flash-preview-09-2025": {
|
|
11263
|
-
id: "google/gemini-2.5-flash-preview-09-2025",
|
|
11264
|
-
name: "Gemini 2.5 Flash Preview 09-2025",
|
|
11265
|
-
api: "anthropic-messages",
|
|
11266
|
-
provider: "vercel-ai-gateway",
|
|
11267
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11268
|
-
reasoning: true,
|
|
11269
|
-
input: ["text", "image"],
|
|
11270
|
-
cost: {
|
|
11271
|
-
input: 0.3,
|
|
11272
|
-
output: 2.5,
|
|
11273
|
-
cacheRead: 0.03,
|
|
11274
|
-
cacheWrite: 0,
|
|
11275
|
-
},
|
|
11276
|
-
contextWindow: 1000000,
|
|
11277
|
-
maxTokens: 65536,
|
|
11278
|
-
},
|
|
11279
11332
|
"google/gemini-2.5-pro": {
|
|
11280
11333
|
id: "google/gemini-2.5-pro",
|
|
11281
11334
|
name: "Gemini 2.5 Pro",
|
|
@@ -11457,7 +11510,7 @@ export const MODELS = {
|
|
|
11457
11510
|
cost: {
|
|
11458
11511
|
input: 0.09999999999999999,
|
|
11459
11512
|
output: 0.09999999999999999,
|
|
11460
|
-
cacheRead: 0,
|
|
11513
|
+
cacheRead: 0.09999999999999999,
|
|
11461
11514
|
cacheWrite: 0,
|
|
11462
11515
|
},
|
|
11463
11516
|
contextWindow: 128000,
|
|
@@ -11630,8 +11683,8 @@ export const MODELS = {
|
|
|
11630
11683
|
cacheRead: 0.03,
|
|
11631
11684
|
cacheWrite: 0.375,
|
|
11632
11685
|
},
|
|
11633
|
-
contextWindow:
|
|
11634
|
-
maxTokens:
|
|
11686
|
+
contextWindow: 204800,
|
|
11687
|
+
maxTokens: 131000,
|
|
11635
11688
|
},
|
|
11636
11689
|
"mistral/codestral": {
|
|
11637
11690
|
id: "mistral/codestral",
|
|
@@ -11659,8 +11712,8 @@ export const MODELS = {
|
|
|
11659
11712
|
reasoning: false,
|
|
11660
11713
|
input: ["text"],
|
|
11661
11714
|
cost: {
|
|
11662
|
-
input: 0,
|
|
11663
|
-
output:
|
|
11715
|
+
input: 0.39999999999999997,
|
|
11716
|
+
output: 2,
|
|
11664
11717
|
cacheRead: 0,
|
|
11665
11718
|
cacheWrite: 0,
|
|
11666
11719
|
},
|
|
@@ -11693,8 +11746,8 @@ export const MODELS = {
|
|
|
11693
11746
|
reasoning: false,
|
|
11694
11747
|
input: ["text"],
|
|
11695
11748
|
cost: {
|
|
11696
|
-
input: 0,
|
|
11697
|
-
output: 0,
|
|
11749
|
+
input: 0.09999999999999999,
|
|
11750
|
+
output: 0.3,
|
|
11698
11751
|
cacheRead: 0,
|
|
11699
11752
|
cacheWrite: 0,
|
|
11700
11753
|
},
|
|
@@ -11710,8 +11763,8 @@ export const MODELS = {
|
|
|
11710
11763
|
reasoning: false,
|
|
11711
11764
|
input: ["text"],
|
|
11712
11765
|
cost: {
|
|
11713
|
-
input: 0.
|
|
11714
|
-
output: 0.
|
|
11766
|
+
input: 0.09999999999999999,
|
|
11767
|
+
output: 0.09999999999999999,
|
|
11715
11768
|
cacheRead: 0,
|
|
11716
11769
|
cacheWrite: 0,
|
|
11717
11770
|
},
|
|
@@ -11727,8 +11780,8 @@ export const MODELS = {
|
|
|
11727
11780
|
reasoning: false,
|
|
11728
11781
|
input: ["text"],
|
|
11729
11782
|
cost: {
|
|
11730
|
-
input: 0.
|
|
11731
|
-
output: 0.
|
|
11783
|
+
input: 0.15,
|
|
11784
|
+
output: 0.15,
|
|
11732
11785
|
cacheRead: 0,
|
|
11733
11786
|
cacheWrite: 0,
|
|
11734
11787
|
},
|
|
@@ -11814,7 +11867,7 @@ export const MODELS = {
|
|
|
11814
11867
|
cost: {
|
|
11815
11868
|
input: 0.6,
|
|
11816
11869
|
output: 2.5,
|
|
11817
|
-
cacheRead: 0,
|
|
11870
|
+
cacheRead: 0.15,
|
|
11818
11871
|
cacheWrite: 0,
|
|
11819
11872
|
},
|
|
11820
11873
|
contextWindow: 131072,
|
|
@@ -11831,7 +11884,7 @@ export const MODELS = {
|
|
|
11831
11884
|
cost: {
|
|
11832
11885
|
input: 0.6,
|
|
11833
11886
|
output: 2.5,
|
|
11834
|
-
cacheRead: 0,
|
|
11887
|
+
cacheRead: 0.15,
|
|
11835
11888
|
cacheWrite: 0,
|
|
11836
11889
|
},
|
|
11837
11890
|
contextWindow: 256000,
|
|
@@ -11880,9 +11933,9 @@ export const MODELS = {
|
|
|
11880
11933
|
reasoning: false,
|
|
11881
11934
|
input: ["text"],
|
|
11882
11935
|
cost: {
|
|
11883
|
-
input:
|
|
11884
|
-
output:
|
|
11885
|
-
cacheRead: 0,
|
|
11936
|
+
input: 1.15,
|
|
11937
|
+
output: 8,
|
|
11938
|
+
cacheRead: 0.15,
|
|
11886
11939
|
cacheWrite: 0,
|
|
11887
11940
|
},
|
|
11888
11941
|
contextWindow: 256000,
|
|
@@ -12347,6 +12400,40 @@ export const MODELS = {
|
|
|
12347
12400
|
contextWindow: 1050000,
|
|
12348
12401
|
maxTokens: 128000,
|
|
12349
12402
|
},
|
|
12403
|
+
"openai/gpt-5.4-mini": {
|
|
12404
|
+
id: "openai/gpt-5.4-mini",
|
|
12405
|
+
name: "GPT 5.4 Mini",
|
|
12406
|
+
api: "anthropic-messages",
|
|
12407
|
+
provider: "vercel-ai-gateway",
|
|
12408
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12409
|
+
reasoning: true,
|
|
12410
|
+
input: ["text", "image"],
|
|
12411
|
+
cost: {
|
|
12412
|
+
input: 0.75,
|
|
12413
|
+
output: 4.5,
|
|
12414
|
+
cacheRead: 0.075,
|
|
12415
|
+
cacheWrite: 0,
|
|
12416
|
+
},
|
|
12417
|
+
contextWindow: 400000,
|
|
12418
|
+
maxTokens: 128000,
|
|
12419
|
+
},
|
|
12420
|
+
"openai/gpt-5.4-nano": {
|
|
12421
|
+
id: "openai/gpt-5.4-nano",
|
|
12422
|
+
name: "GPT 5.4 Nano",
|
|
12423
|
+
api: "anthropic-messages",
|
|
12424
|
+
provider: "vercel-ai-gateway",
|
|
12425
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12426
|
+
reasoning: true,
|
|
12427
|
+
input: ["text", "image"],
|
|
12428
|
+
cost: {
|
|
12429
|
+
input: 0.19999999999999998,
|
|
12430
|
+
output: 1.25,
|
|
12431
|
+
cacheRead: 0.02,
|
|
12432
|
+
cacheWrite: 0,
|
|
12433
|
+
},
|
|
12434
|
+
contextWindow: 400000,
|
|
12435
|
+
maxTokens: 128000,
|
|
12436
|
+
},
|
|
12350
12437
|
"openai/gpt-5.4-pro": {
|
|
12351
12438
|
id: "openai/gpt-5.4-pro",
|
|
12352
12439
|
name: "GPT 5.4 Pro",
|
|
@@ -12551,40 +12638,6 @@ export const MODELS = {
|
|
|
12551
12638
|
contextWindow: 131072,
|
|
12552
12639
|
maxTokens: 131072,
|
|
12553
12640
|
},
|
|
12554
|
-
"vercel/v0-1.0-md": {
|
|
12555
|
-
id: "vercel/v0-1.0-md",
|
|
12556
|
-
name: "v0-1.0-md",
|
|
12557
|
-
api: "anthropic-messages",
|
|
12558
|
-
provider: "vercel-ai-gateway",
|
|
12559
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12560
|
-
reasoning: false,
|
|
12561
|
-
input: ["text", "image"],
|
|
12562
|
-
cost: {
|
|
12563
|
-
input: 3,
|
|
12564
|
-
output: 15,
|
|
12565
|
-
cacheRead: 0,
|
|
12566
|
-
cacheWrite: 0,
|
|
12567
|
-
},
|
|
12568
|
-
contextWindow: 128000,
|
|
12569
|
-
maxTokens: 32000,
|
|
12570
|
-
},
|
|
12571
|
-
"vercel/v0-1.5-md": {
|
|
12572
|
-
id: "vercel/v0-1.5-md",
|
|
12573
|
-
name: "v0-1.5-md",
|
|
12574
|
-
api: "anthropic-messages",
|
|
12575
|
-
provider: "vercel-ai-gateway",
|
|
12576
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12577
|
-
reasoning: false,
|
|
12578
|
-
input: ["text", "image"],
|
|
12579
|
-
cost: {
|
|
12580
|
-
input: 3,
|
|
12581
|
-
output: 15,
|
|
12582
|
-
cacheRead: 0,
|
|
12583
|
-
cacheWrite: 0,
|
|
12584
|
-
},
|
|
12585
|
-
contextWindow: 128000,
|
|
12586
|
-
maxTokens: 32768,
|
|
12587
|
-
},
|
|
12588
12641
|
"xai/grok-2-vision": {
|
|
12589
12642
|
id: "xai/grok-2-vision",
|
|
12590
12643
|
name: "Grok 2 Vision",
|
|
@@ -12613,7 +12666,7 @@ export const MODELS = {
|
|
|
12613
12666
|
cost: {
|
|
12614
12667
|
input: 3,
|
|
12615
12668
|
output: 15,
|
|
12616
|
-
cacheRead: 0,
|
|
12669
|
+
cacheRead: 0.75,
|
|
12617
12670
|
cacheWrite: 0,
|
|
12618
12671
|
},
|
|
12619
12672
|
contextWindow: 131072,
|
|
@@ -12630,7 +12683,7 @@ export const MODELS = {
|
|
|
12630
12683
|
cost: {
|
|
12631
12684
|
input: 5,
|
|
12632
12685
|
output: 25,
|
|
12633
|
-
cacheRead:
|
|
12686
|
+
cacheRead: 1.25,
|
|
12634
12687
|
cacheWrite: 0,
|
|
12635
12688
|
},
|
|
12636
12689
|
contextWindow: 131072,
|
|
@@ -12647,7 +12700,7 @@ export const MODELS = {
|
|
|
12647
12700
|
cost: {
|
|
12648
12701
|
input: 0.3,
|
|
12649
12702
|
output: 0.5,
|
|
12650
|
-
cacheRead: 0,
|
|
12703
|
+
cacheRead: 0.075,
|
|
12651
12704
|
cacheWrite: 0,
|
|
12652
12705
|
},
|
|
12653
12706
|
contextWindow: 131072,
|
|
@@ -12900,8 +12953,8 @@ export const MODELS = {
|
|
|
12900
12953
|
reasoning: true,
|
|
12901
12954
|
input: ["text"],
|
|
12902
12955
|
cost: {
|
|
12903
|
-
input: 0.
|
|
12904
|
-
output:
|
|
12956
|
+
input: 0.6,
|
|
12957
|
+
output: 2.2,
|
|
12905
12958
|
cacheRead: 0.11,
|
|
12906
12959
|
cacheWrite: 0,
|
|
12907
12960
|
},
|