@mariozechner/pi-ai 0.58.3 → 0.59.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -0
- package/dist/index.d.ts +9 -8
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +0 -8
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +92 -139
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +135 -185
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts +7 -0
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +22 -11
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +3 -0
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +3 -0
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +3 -0
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/mistral.d.ts.map +1 -1
- package/dist/providers/mistral.js +3 -0
- package/dist/providers/mistral.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +3 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +7 -1
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/register-builtins.d.ts +28 -1
- package/dist/providers/register-builtins.d.ts.map +1 -1
- package/dist/providers/register-builtins.js +178 -45
- package/dist/providers/register-builtins.js.map +1 -1
- package/dist/types.d.ts +1 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +39 -5
- package/bedrock-provider.d.ts +0 -1
- package/bedrock-provider.js +0 -1
package/dist/models.generated.js
CHANGED
|
@@ -2577,7 +2577,7 @@ export const MODELS = {
|
|
|
2577
2577
|
cacheRead: 0,
|
|
2578
2578
|
cacheWrite: 0,
|
|
2579
2579
|
},
|
|
2580
|
-
contextWindow:
|
|
2580
|
+
contextWindow: 1000000,
|
|
2581
2581
|
maxTokens: 64000,
|
|
2582
2582
|
},
|
|
2583
2583
|
"claude-sonnet-4": {
|
|
@@ -2631,7 +2631,7 @@ export const MODELS = {
|
|
|
2631
2631
|
cacheRead: 0,
|
|
2632
2632
|
cacheWrite: 0,
|
|
2633
2633
|
},
|
|
2634
|
-
contextWindow:
|
|
2634
|
+
contextWindow: 1000000,
|
|
2635
2635
|
maxTokens: 32000,
|
|
2636
2636
|
},
|
|
2637
2637
|
"gemini-2.5-pro": {
|
|
@@ -5966,23 +5966,6 @@ export const MODELS = {
|
|
|
5966
5966
|
contextWindow: 1048576,
|
|
5967
5967
|
maxTokens: 65536,
|
|
5968
5968
|
},
|
|
5969
|
-
"gemini-3-pro": {
|
|
5970
|
-
id: "gemini-3-pro",
|
|
5971
|
-
name: "Gemini 3 Pro",
|
|
5972
|
-
api: "google-generative-ai",
|
|
5973
|
-
provider: "opencode",
|
|
5974
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
5975
|
-
reasoning: true,
|
|
5976
|
-
input: ["text", "image"],
|
|
5977
|
-
cost: {
|
|
5978
|
-
input: 2,
|
|
5979
|
-
output: 12,
|
|
5980
|
-
cacheRead: 0.2,
|
|
5981
|
-
cacheWrite: 0,
|
|
5982
|
-
},
|
|
5983
|
-
contextWindow: 1048576,
|
|
5984
|
-
maxTokens: 65536,
|
|
5985
|
-
},
|
|
5986
5969
|
"gemini-3.1-pro": {
|
|
5987
5970
|
id: "gemini-3.1-pro",
|
|
5988
5971
|
name: "Gemini 3.1 Pro Preview",
|
|
@@ -6000,40 +5983,6 @@ export const MODELS = {
|
|
|
6000
5983
|
contextWindow: 1048576,
|
|
6001
5984
|
maxTokens: 65536,
|
|
6002
5985
|
},
|
|
6003
|
-
"glm-4.6": {
|
|
6004
|
-
id: "glm-4.6",
|
|
6005
|
-
name: "GLM-4.6",
|
|
6006
|
-
api: "openai-completions",
|
|
6007
|
-
provider: "opencode",
|
|
6008
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6009
|
-
reasoning: true,
|
|
6010
|
-
input: ["text"],
|
|
6011
|
-
cost: {
|
|
6012
|
-
input: 0.6,
|
|
6013
|
-
output: 2.2,
|
|
6014
|
-
cacheRead: 0.1,
|
|
6015
|
-
cacheWrite: 0,
|
|
6016
|
-
},
|
|
6017
|
-
contextWindow: 204800,
|
|
6018
|
-
maxTokens: 131072,
|
|
6019
|
-
},
|
|
6020
|
-
"glm-4.7": {
|
|
6021
|
-
id: "glm-4.7",
|
|
6022
|
-
name: "GLM-4.7",
|
|
6023
|
-
api: "openai-completions",
|
|
6024
|
-
provider: "opencode",
|
|
6025
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6026
|
-
reasoning: true,
|
|
6027
|
-
input: ["text"],
|
|
6028
|
-
cost: {
|
|
6029
|
-
input: 0.6,
|
|
6030
|
-
output: 2.2,
|
|
6031
|
-
cacheRead: 0.1,
|
|
6032
|
-
cacheWrite: 0,
|
|
6033
|
-
},
|
|
6034
|
-
contextWindow: 204800,
|
|
6035
|
-
maxTokens: 131072,
|
|
6036
|
-
},
|
|
6037
5986
|
"glm-5": {
|
|
6038
5987
|
id: "glm-5",
|
|
6039
5988
|
name: "GLM-5",
|
|
@@ -6289,23 +6238,6 @@ export const MODELS = {
|
|
|
6289
6238
|
contextWindow: 262144,
|
|
6290
6239
|
maxTokens: 65536,
|
|
6291
6240
|
},
|
|
6292
|
-
"minimax-m2.1": {
|
|
6293
|
-
id: "minimax-m2.1",
|
|
6294
|
-
name: "MiniMax M2.1",
|
|
6295
|
-
api: "openai-completions",
|
|
6296
|
-
provider: "opencode",
|
|
6297
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6298
|
-
reasoning: true,
|
|
6299
|
-
input: ["text"],
|
|
6300
|
-
cost: {
|
|
6301
|
-
input: 0.3,
|
|
6302
|
-
output: 1.2,
|
|
6303
|
-
cacheRead: 0.1,
|
|
6304
|
-
cacheWrite: 0,
|
|
6305
|
-
},
|
|
6306
|
-
contextWindow: 204800,
|
|
6307
|
-
maxTokens: 131072,
|
|
6308
|
-
},
|
|
6309
6241
|
"minimax-m2.5": {
|
|
6310
6242
|
id: "minimax-m2.5",
|
|
6311
6243
|
name: "MiniMax M2.5",
|
|
@@ -7662,7 +7594,7 @@ export const MODELS = {
|
|
|
7662
7594
|
reasoning: true,
|
|
7663
7595
|
input: ["text"],
|
|
7664
7596
|
cost: {
|
|
7665
|
-
input: 0.
|
|
7597
|
+
input: 0.19999999999999998,
|
|
7666
7598
|
output: 1.2,
|
|
7667
7599
|
cacheRead: 0,
|
|
7668
7600
|
cacheWrite: 0,
|
|
@@ -7959,6 +7891,23 @@ export const MODELS = {
|
|
|
7959
7891
|
contextWindow: 32768,
|
|
7960
7892
|
maxTokens: 16384,
|
|
7961
7893
|
},
|
|
7894
|
+
"mistralai/mistral-small-2603": {
|
|
7895
|
+
id: "mistralai/mistral-small-2603",
|
|
7896
|
+
name: "Mistral: Mistral Small 4",
|
|
7897
|
+
api: "openai-completions",
|
|
7898
|
+
provider: "openrouter",
|
|
7899
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7900
|
+
reasoning: true,
|
|
7901
|
+
input: ["text", "image"],
|
|
7902
|
+
cost: {
|
|
7903
|
+
input: 0.15,
|
|
7904
|
+
output: 0.6,
|
|
7905
|
+
cacheRead: 0,
|
|
7906
|
+
cacheWrite: 0,
|
|
7907
|
+
},
|
|
7908
|
+
contextWindow: 262144,
|
|
7909
|
+
maxTokens: 4096,
|
|
7910
|
+
},
|
|
7962
7911
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
7963
7912
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
7964
7913
|
name: "Mistral: Mistral Small 3.1 24B (free)",
|
|
@@ -8945,6 +8894,40 @@ export const MODELS = {
|
|
|
8945
8894
|
contextWindow: 1050000,
|
|
8946
8895
|
maxTokens: 128000,
|
|
8947
8896
|
},
|
|
8897
|
+
"openai/gpt-5.4-mini": {
|
|
8898
|
+
id: "openai/gpt-5.4-mini",
|
|
8899
|
+
name: "OpenAI: GPT-5.4 Mini",
|
|
8900
|
+
api: "openai-completions",
|
|
8901
|
+
provider: "openrouter",
|
|
8902
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8903
|
+
reasoning: true,
|
|
8904
|
+
input: ["text", "image"],
|
|
8905
|
+
cost: {
|
|
8906
|
+
input: 0.75,
|
|
8907
|
+
output: 4.5,
|
|
8908
|
+
cacheRead: 0.075,
|
|
8909
|
+
cacheWrite: 0,
|
|
8910
|
+
},
|
|
8911
|
+
contextWindow: 400000,
|
|
8912
|
+
maxTokens: 128000,
|
|
8913
|
+
},
|
|
8914
|
+
"openai/gpt-5.4-nano": {
|
|
8915
|
+
id: "openai/gpt-5.4-nano",
|
|
8916
|
+
name: "OpenAI: GPT-5.4 Nano",
|
|
8917
|
+
api: "openai-completions",
|
|
8918
|
+
provider: "openrouter",
|
|
8919
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8920
|
+
reasoning: true,
|
|
8921
|
+
input: ["text", "image"],
|
|
8922
|
+
cost: {
|
|
8923
|
+
input: 0.19999999999999998,
|
|
8924
|
+
output: 1.25,
|
|
8925
|
+
cacheRead: 0.02,
|
|
8926
|
+
cacheWrite: 0,
|
|
8927
|
+
},
|
|
8928
|
+
contextWindow: 400000,
|
|
8929
|
+
maxTokens: 128000,
|
|
8930
|
+
},
|
|
8948
8931
|
"openai/gpt-5.4-pro": {
|
|
8949
8932
|
id: "openai/gpt-5.4-pro",
|
|
8950
8933
|
name: "OpenAI: GPT-5.4 Pro",
|
|
@@ -9053,7 +9036,7 @@ export const MODELS = {
|
|
|
9053
9036
|
api: "openai-completions",
|
|
9054
9037
|
provider: "openrouter",
|
|
9055
9038
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9056
|
-
reasoning:
|
|
9039
|
+
reasoning: true,
|
|
9057
9040
|
input: ["text", "image"],
|
|
9058
9041
|
cost: {
|
|
9059
9042
|
input: 15,
|
|
@@ -9104,7 +9087,7 @@ export const MODELS = {
|
|
|
9104
9087
|
api: "openai-completions",
|
|
9105
9088
|
provider: "openrouter",
|
|
9106
9089
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9107
|
-
reasoning:
|
|
9090
|
+
reasoning: true,
|
|
9108
9091
|
input: ["text"],
|
|
9109
9092
|
cost: {
|
|
9110
9093
|
input: 1.1,
|
|
@@ -9121,7 +9104,7 @@ export const MODELS = {
|
|
|
9121
9104
|
api: "openai-completions",
|
|
9122
9105
|
provider: "openrouter",
|
|
9123
9106
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
9124
|
-
reasoning:
|
|
9107
|
+
reasoning: true,
|
|
9125
9108
|
input: ["text"],
|
|
9126
9109
|
cost: {
|
|
9127
9110
|
input: 1.1,
|
|
@@ -9345,9 +9328,9 @@ export const MODELS = {
|
|
|
9345
9328
|
reasoning: false,
|
|
9346
9329
|
input: ["text"],
|
|
9347
9330
|
cost: {
|
|
9348
|
-
input: 0.
|
|
9349
|
-
output:
|
|
9350
|
-
cacheRead: 0.
|
|
9331
|
+
input: 0.26,
|
|
9332
|
+
output: 0.78,
|
|
9333
|
+
cacheRead: 0.052000000000000005,
|
|
9351
9334
|
cacheWrite: 0,
|
|
9352
9335
|
},
|
|
9353
9336
|
contextWindow: 1000000,
|
|
@@ -9413,8 +9396,8 @@ export const MODELS = {
|
|
|
9413
9396
|
reasoning: false,
|
|
9414
9397
|
input: ["text", "image"],
|
|
9415
9398
|
cost: {
|
|
9416
|
-
input: 0.
|
|
9417
|
-
output:
|
|
9399
|
+
input: 0.52,
|
|
9400
|
+
output: 2.08,
|
|
9418
9401
|
cacheRead: 0,
|
|
9419
9402
|
cacheWrite: 0,
|
|
9420
9403
|
},
|
|
@@ -9702,9 +9685,9 @@ export const MODELS = {
|
|
|
9702
9685
|
reasoning: false,
|
|
9703
9686
|
input: ["text"],
|
|
9704
9687
|
cost: {
|
|
9705
|
-
input:
|
|
9706
|
-
output:
|
|
9707
|
-
cacheRead: 0.
|
|
9688
|
+
input: 0.78,
|
|
9689
|
+
output: 3.9,
|
|
9690
|
+
cacheRead: 0.156,
|
|
9708
9691
|
cacheWrite: 0,
|
|
9709
9692
|
},
|
|
9710
9693
|
contextWindow: 262144,
|
|
@@ -9991,8 +9974,8 @@ export const MODELS = {
|
|
|
9991
9974
|
reasoning: true,
|
|
9992
9975
|
input: ["text", "image"],
|
|
9993
9976
|
cost: {
|
|
9994
|
-
input: 0.
|
|
9995
|
-
output: 0.
|
|
9977
|
+
input: 0.065,
|
|
9978
|
+
output: 0.26,
|
|
9996
9979
|
cacheRead: 0,
|
|
9997
9980
|
cacheWrite: 0,
|
|
9998
9981
|
},
|
|
@@ -10554,7 +10537,7 @@ export const MODELS = {
|
|
|
10554
10537
|
reasoning: true,
|
|
10555
10538
|
input: ["text"],
|
|
10556
10539
|
cost: {
|
|
10557
|
-
input: 0.
|
|
10540
|
+
input: 0.12,
|
|
10558
10541
|
output: 0.24,
|
|
10559
10542
|
cacheRead: 0,
|
|
10560
10543
|
cacheWrite: 0,
|
|
@@ -10607,7 +10590,7 @@ export const MODELS = {
|
|
|
10607
10590
|
cost: {
|
|
10608
10591
|
input: 0.29,
|
|
10609
10592
|
output: 0.59,
|
|
10610
|
-
cacheRead: 0,
|
|
10593
|
+
cacheRead: 0.145,
|
|
10611
10594
|
cacheWrite: 0,
|
|
10612
10595
|
},
|
|
10613
10596
|
contextWindow: 131072,
|
|
@@ -10622,9 +10605,9 @@ export const MODELS = {
|
|
|
10622
10605
|
reasoning: true,
|
|
10623
10606
|
input: ["text", "image"],
|
|
10624
10607
|
cost: {
|
|
10625
|
-
input: 0.
|
|
10626
|
-
output: 2.
|
|
10627
|
-
cacheRead: 0,
|
|
10608
|
+
input: 0.22999999999999998,
|
|
10609
|
+
output: 2.3,
|
|
10610
|
+
cacheRead: 0.19999999999999998,
|
|
10628
10611
|
cacheWrite: 0,
|
|
10629
10612
|
},
|
|
10630
10613
|
contextWindow: 262114,
|
|
@@ -10641,7 +10624,7 @@ export const MODELS = {
|
|
|
10641
10624
|
cost: {
|
|
10642
10625
|
input: 0.39999999999999997,
|
|
10643
10626
|
output: 1.5999999999999999,
|
|
10644
|
-
cacheRead: 0,
|
|
10627
|
+
cacheRead: 0.022,
|
|
10645
10628
|
cacheWrite: 0,
|
|
10646
10629
|
},
|
|
10647
10630
|
contextWindow: 262144,
|
|
@@ -11134,7 +11117,7 @@ export const MODELS = {
|
|
|
11134
11117
|
cost: {
|
|
11135
11118
|
input: 0.27,
|
|
11136
11119
|
output: 1,
|
|
11137
|
-
cacheRead: 0,
|
|
11120
|
+
cacheRead: 0.135,
|
|
11138
11121
|
cacheWrite: 0,
|
|
11139
11122
|
},
|
|
11140
11123
|
contextWindow: 131072,
|
|
@@ -11149,9 +11132,9 @@ export const MODELS = {
|
|
|
11149
11132
|
reasoning: false,
|
|
11150
11133
|
input: ["text"],
|
|
11151
11134
|
cost: {
|
|
11152
|
-
input: 0.
|
|
11153
|
-
output: 0.
|
|
11154
|
-
cacheRead: 0.
|
|
11135
|
+
input: 0.28,
|
|
11136
|
+
output: 0.42,
|
|
11137
|
+
cacheRead: 0.028,
|
|
11155
11138
|
cacheWrite: 0,
|
|
11156
11139
|
},
|
|
11157
11140
|
contextWindow: 128000,
|
|
@@ -11185,7 +11168,7 @@ export const MODELS = {
|
|
|
11185
11168
|
cost: {
|
|
11186
11169
|
input: 0.15,
|
|
11187
11170
|
output: 0.6,
|
|
11188
|
-
cacheRead: 0,
|
|
11171
|
+
cacheRead: 0.024999999999999998,
|
|
11189
11172
|
cacheWrite: 0,
|
|
11190
11173
|
},
|
|
11191
11174
|
contextWindow: 1048576,
|
|
@@ -11202,7 +11185,7 @@ export const MODELS = {
|
|
|
11202
11185
|
cost: {
|
|
11203
11186
|
input: 0.075,
|
|
11204
11187
|
output: 0.3,
|
|
11205
|
-
cacheRead: 0,
|
|
11188
|
+
cacheRead: 0.02,
|
|
11206
11189
|
cacheWrite: 0,
|
|
11207
11190
|
},
|
|
11208
11191
|
contextWindow: 1048576,
|
|
@@ -11242,40 +11225,6 @@ export const MODELS = {
|
|
|
11242
11225
|
contextWindow: 1048576,
|
|
11243
11226
|
maxTokens: 65536,
|
|
11244
11227
|
},
|
|
11245
|
-
"google/gemini-2.5-flash-lite-preview-09-2025": {
|
|
11246
|
-
id: "google/gemini-2.5-flash-lite-preview-09-2025",
|
|
11247
|
-
name: "Gemini 2.5 Flash Lite Preview 09-2025",
|
|
11248
|
-
api: "anthropic-messages",
|
|
11249
|
-
provider: "vercel-ai-gateway",
|
|
11250
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11251
|
-
reasoning: true,
|
|
11252
|
-
input: ["text", "image"],
|
|
11253
|
-
cost: {
|
|
11254
|
-
input: 0.09999999999999999,
|
|
11255
|
-
output: 0.39999999999999997,
|
|
11256
|
-
cacheRead: 0.01,
|
|
11257
|
-
cacheWrite: 0,
|
|
11258
|
-
},
|
|
11259
|
-
contextWindow: 1048576,
|
|
11260
|
-
maxTokens: 65536,
|
|
11261
|
-
},
|
|
11262
|
-
"google/gemini-2.5-flash-preview-09-2025": {
|
|
11263
|
-
id: "google/gemini-2.5-flash-preview-09-2025",
|
|
11264
|
-
name: "Gemini 2.5 Flash Preview 09-2025",
|
|
11265
|
-
api: "anthropic-messages",
|
|
11266
|
-
provider: "vercel-ai-gateway",
|
|
11267
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11268
|
-
reasoning: true,
|
|
11269
|
-
input: ["text", "image"],
|
|
11270
|
-
cost: {
|
|
11271
|
-
input: 0.3,
|
|
11272
|
-
output: 2.5,
|
|
11273
|
-
cacheRead: 0.03,
|
|
11274
|
-
cacheWrite: 0,
|
|
11275
|
-
},
|
|
11276
|
-
contextWindow: 1000000,
|
|
11277
|
-
maxTokens: 65536,
|
|
11278
|
-
},
|
|
11279
11228
|
"google/gemini-2.5-pro": {
|
|
11280
11229
|
id: "google/gemini-2.5-pro",
|
|
11281
11230
|
name: "Gemini 2.5 Pro",
|
|
@@ -11457,7 +11406,7 @@ export const MODELS = {
|
|
|
11457
11406
|
cost: {
|
|
11458
11407
|
input: 0.09999999999999999,
|
|
11459
11408
|
output: 0.09999999999999999,
|
|
11460
|
-
cacheRead: 0,
|
|
11409
|
+
cacheRead: 0.09999999999999999,
|
|
11461
11410
|
cacheWrite: 0,
|
|
11462
11411
|
},
|
|
11463
11412
|
contextWindow: 128000,
|
|
@@ -11630,8 +11579,8 @@ export const MODELS = {
|
|
|
11630
11579
|
cacheRead: 0.03,
|
|
11631
11580
|
cacheWrite: 0.375,
|
|
11632
11581
|
},
|
|
11633
|
-
contextWindow:
|
|
11634
|
-
maxTokens:
|
|
11582
|
+
contextWindow: 204800,
|
|
11583
|
+
maxTokens: 131000,
|
|
11635
11584
|
},
|
|
11636
11585
|
"mistral/codestral": {
|
|
11637
11586
|
id: "mistral/codestral",
|
|
@@ -11659,8 +11608,8 @@ export const MODELS = {
|
|
|
11659
11608
|
reasoning: false,
|
|
11660
11609
|
input: ["text"],
|
|
11661
11610
|
cost: {
|
|
11662
|
-
input: 0,
|
|
11663
|
-
output:
|
|
11611
|
+
input: 0.39999999999999997,
|
|
11612
|
+
output: 2,
|
|
11664
11613
|
cacheRead: 0,
|
|
11665
11614
|
cacheWrite: 0,
|
|
11666
11615
|
},
|
|
@@ -11693,8 +11642,8 @@ export const MODELS = {
|
|
|
11693
11642
|
reasoning: false,
|
|
11694
11643
|
input: ["text"],
|
|
11695
11644
|
cost: {
|
|
11696
|
-
input: 0,
|
|
11697
|
-
output: 0,
|
|
11645
|
+
input: 0.09999999999999999,
|
|
11646
|
+
output: 0.3,
|
|
11698
11647
|
cacheRead: 0,
|
|
11699
11648
|
cacheWrite: 0,
|
|
11700
11649
|
},
|
|
@@ -11710,8 +11659,8 @@ export const MODELS = {
|
|
|
11710
11659
|
reasoning: false,
|
|
11711
11660
|
input: ["text"],
|
|
11712
11661
|
cost: {
|
|
11713
|
-
input: 0.
|
|
11714
|
-
output: 0.
|
|
11662
|
+
input: 0.09999999999999999,
|
|
11663
|
+
output: 0.09999999999999999,
|
|
11715
11664
|
cacheRead: 0,
|
|
11716
11665
|
cacheWrite: 0,
|
|
11717
11666
|
},
|
|
@@ -11727,8 +11676,8 @@ export const MODELS = {
|
|
|
11727
11676
|
reasoning: false,
|
|
11728
11677
|
input: ["text"],
|
|
11729
11678
|
cost: {
|
|
11730
|
-
input: 0.
|
|
11731
|
-
output: 0.
|
|
11679
|
+
input: 0.15,
|
|
11680
|
+
output: 0.15,
|
|
11732
11681
|
cacheRead: 0,
|
|
11733
11682
|
cacheWrite: 0,
|
|
11734
11683
|
},
|
|
@@ -11814,7 +11763,7 @@ export const MODELS = {
|
|
|
11814
11763
|
cost: {
|
|
11815
11764
|
input: 0.6,
|
|
11816
11765
|
output: 2.5,
|
|
11817
|
-
cacheRead: 0,
|
|
11766
|
+
cacheRead: 0.15,
|
|
11818
11767
|
cacheWrite: 0,
|
|
11819
11768
|
},
|
|
11820
11769
|
contextWindow: 131072,
|
|
@@ -11831,7 +11780,7 @@ export const MODELS = {
|
|
|
11831
11780
|
cost: {
|
|
11832
11781
|
input: 0.6,
|
|
11833
11782
|
output: 2.5,
|
|
11834
|
-
cacheRead: 0,
|
|
11783
|
+
cacheRead: 0.15,
|
|
11835
11784
|
cacheWrite: 0,
|
|
11836
11785
|
},
|
|
11837
11786
|
contextWindow: 256000,
|
|
@@ -11880,9 +11829,9 @@ export const MODELS = {
|
|
|
11880
11829
|
reasoning: false,
|
|
11881
11830
|
input: ["text"],
|
|
11882
11831
|
cost: {
|
|
11883
|
-
input:
|
|
11884
|
-
output:
|
|
11885
|
-
cacheRead: 0,
|
|
11832
|
+
input: 1.15,
|
|
11833
|
+
output: 8,
|
|
11834
|
+
cacheRead: 0.15,
|
|
11886
11835
|
cacheWrite: 0,
|
|
11887
11836
|
},
|
|
11888
11837
|
contextWindow: 256000,
|
|
@@ -12551,40 +12500,6 @@ export const MODELS = {
|
|
|
12551
12500
|
contextWindow: 131072,
|
|
12552
12501
|
maxTokens: 131072,
|
|
12553
12502
|
},
|
|
12554
|
-
"vercel/v0-1.0-md": {
|
|
12555
|
-
id: "vercel/v0-1.0-md",
|
|
12556
|
-
name: "v0-1.0-md",
|
|
12557
|
-
api: "anthropic-messages",
|
|
12558
|
-
provider: "vercel-ai-gateway",
|
|
12559
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12560
|
-
reasoning: false,
|
|
12561
|
-
input: ["text", "image"],
|
|
12562
|
-
cost: {
|
|
12563
|
-
input: 3,
|
|
12564
|
-
output: 15,
|
|
12565
|
-
cacheRead: 0,
|
|
12566
|
-
cacheWrite: 0,
|
|
12567
|
-
},
|
|
12568
|
-
contextWindow: 128000,
|
|
12569
|
-
maxTokens: 32000,
|
|
12570
|
-
},
|
|
12571
|
-
"vercel/v0-1.5-md": {
|
|
12572
|
-
id: "vercel/v0-1.5-md",
|
|
12573
|
-
name: "v0-1.5-md",
|
|
12574
|
-
api: "anthropic-messages",
|
|
12575
|
-
provider: "vercel-ai-gateway",
|
|
12576
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12577
|
-
reasoning: false,
|
|
12578
|
-
input: ["text", "image"],
|
|
12579
|
-
cost: {
|
|
12580
|
-
input: 3,
|
|
12581
|
-
output: 15,
|
|
12582
|
-
cacheRead: 0,
|
|
12583
|
-
cacheWrite: 0,
|
|
12584
|
-
},
|
|
12585
|
-
contextWindow: 128000,
|
|
12586
|
-
maxTokens: 32768,
|
|
12587
|
-
},
|
|
12588
12503
|
"xai/grok-2-vision": {
|
|
12589
12504
|
id: "xai/grok-2-vision",
|
|
12590
12505
|
name: "Grok 2 Vision",
|
|
@@ -12613,7 +12528,7 @@ export const MODELS = {
|
|
|
12613
12528
|
cost: {
|
|
12614
12529
|
input: 3,
|
|
12615
12530
|
output: 15,
|
|
12616
|
-
cacheRead: 0,
|
|
12531
|
+
cacheRead: 0.75,
|
|
12617
12532
|
cacheWrite: 0,
|
|
12618
12533
|
},
|
|
12619
12534
|
contextWindow: 131072,
|
|
@@ -12630,7 +12545,7 @@ export const MODELS = {
|
|
|
12630
12545
|
cost: {
|
|
12631
12546
|
input: 5,
|
|
12632
12547
|
output: 25,
|
|
12633
|
-
cacheRead:
|
|
12548
|
+
cacheRead: 1.25,
|
|
12634
12549
|
cacheWrite: 0,
|
|
12635
12550
|
},
|
|
12636
12551
|
contextWindow: 131072,
|
|
@@ -12647,7 +12562,7 @@ export const MODELS = {
|
|
|
12647
12562
|
cost: {
|
|
12648
12563
|
input: 0.3,
|
|
12649
12564
|
output: 0.5,
|
|
12650
|
-
cacheRead: 0,
|
|
12565
|
+
cacheRead: 0.075,
|
|
12651
12566
|
cacheWrite: 0,
|
|
12652
12567
|
},
|
|
12653
12568
|
contextWindow: 131072,
|
|
@@ -12900,8 +12815,8 @@ export const MODELS = {
|
|
|
12900
12815
|
reasoning: true,
|
|
12901
12816
|
input: ["text"],
|
|
12902
12817
|
cost: {
|
|
12903
|
-
input: 0.
|
|
12904
|
-
output:
|
|
12818
|
+
input: 0.6,
|
|
12819
|
+
output: 2.2,
|
|
12905
12820
|
cacheRead: 0.11,
|
|
12906
12821
|
cacheWrite: 0,
|
|
12907
12822
|
},
|
|
@@ -13010,6 +12925,23 @@ export const MODELS = {
|
|
|
13010
12925
|
contextWindow: 202800,
|
|
13011
12926
|
maxTokens: 131100,
|
|
13012
12927
|
},
|
|
12928
|
+
"zai/glm-5-turbo": {
|
|
12929
|
+
id: "zai/glm-5-turbo",
|
|
12930
|
+
name: "GLM 5 Turbo",
|
|
12931
|
+
api: "anthropic-messages",
|
|
12932
|
+
provider: "vercel-ai-gateway",
|
|
12933
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12934
|
+
reasoning: true,
|
|
12935
|
+
input: ["text"],
|
|
12936
|
+
cost: {
|
|
12937
|
+
input: 1.2,
|
|
12938
|
+
output: 4,
|
|
12939
|
+
cacheRead: 0.24,
|
|
12940
|
+
cacheWrite: 0,
|
|
12941
|
+
},
|
|
12942
|
+
contextWindow: 202800,
|
|
12943
|
+
maxTokens: 131100,
|
|
12944
|
+
},
|
|
13013
12945
|
},
|
|
13014
12946
|
"xai": {
|
|
13015
12947
|
"grok-2": {
|
|
@@ -13601,6 +13533,24 @@ export const MODELS = {
|
|
|
13601
13533
|
contextWindow: 204800,
|
|
13602
13534
|
maxTokens: 131072,
|
|
13603
13535
|
},
|
|
13536
|
+
"glm-5-turbo": {
|
|
13537
|
+
id: "glm-5-turbo",
|
|
13538
|
+
name: "GLM-5-Turbo",
|
|
13539
|
+
api: "openai-completions",
|
|
13540
|
+
provider: "zai",
|
|
13541
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13542
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
13543
|
+
reasoning: true,
|
|
13544
|
+
input: ["text"],
|
|
13545
|
+
cost: {
|
|
13546
|
+
input: 1.2,
|
|
13547
|
+
output: 4,
|
|
13548
|
+
cacheRead: 0.24,
|
|
13549
|
+
cacheWrite: 0,
|
|
13550
|
+
},
|
|
13551
|
+
contextWindow: 200000,
|
|
13552
|
+
maxTokens: 131072,
|
|
13553
|
+
},
|
|
13604
13554
|
},
|
|
13605
13555
|
};
|
|
13606
13556
|
//# sourceMappingURL=models.generated.js.map
|