@mariozechner/pi-ai 0.50.3 → 0.50.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +68 -51
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +84 -67
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +2 -2
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +1 -1
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +2 -2
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +14 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/types.d.ts +13 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -3668,8 +3668,8 @@ export const MODELS = {
|
|
|
3668
3668
|
reasoning: false,
|
|
3669
3669
|
input: ["text"],
|
|
3670
3670
|
cost: {
|
|
3671
|
-
input: 0,
|
|
3672
|
-
output:
|
|
3671
|
+
input: 0.4,
|
|
3672
|
+
output: 2,
|
|
3673
3673
|
cacheRead: 0,
|
|
3674
3674
|
cacheWrite: 0,
|
|
3675
3675
|
},
|
|
@@ -4923,6 +4923,23 @@ export const MODELS = {
|
|
|
4923
4923
|
contextWindow: 204800,
|
|
4924
4924
|
maxTokens: 131072,
|
|
4925
4925
|
},
|
|
4926
|
+
"glm-4.7-free": {
|
|
4927
|
+
id: "glm-4.7-free",
|
|
4928
|
+
name: "GLM-4.7 Free",
|
|
4929
|
+
api: "openai-completions",
|
|
4930
|
+
provider: "opencode",
|
|
4931
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
4932
|
+
reasoning: true,
|
|
4933
|
+
input: ["text"],
|
|
4934
|
+
cost: {
|
|
4935
|
+
input: 0,
|
|
4936
|
+
output: 0,
|
|
4937
|
+
cacheRead: 0,
|
|
4938
|
+
cacheWrite: 0,
|
|
4939
|
+
},
|
|
4940
|
+
contextWindow: 204800,
|
|
4941
|
+
maxTokens: 131072,
|
|
4942
|
+
},
|
|
4926
4943
|
"gpt-5": {
|
|
4927
4944
|
id: "gpt-5",
|
|
4928
4945
|
name: "GPT-5",
|
|
@@ -5121,7 +5138,24 @@ export const MODELS = {
|
|
|
5121
5138
|
cost: {
|
|
5122
5139
|
input: 0.6,
|
|
5123
5140
|
output: 3,
|
|
5124
|
-
cacheRead: 0.
|
|
5141
|
+
cacheRead: 0.08,
|
|
5142
|
+
cacheWrite: 0,
|
|
5143
|
+
},
|
|
5144
|
+
contextWindow: 262144,
|
|
5145
|
+
maxTokens: 262144,
|
|
5146
|
+
},
|
|
5147
|
+
"kimi-k2.5-free": {
|
|
5148
|
+
id: "kimi-k2.5-free",
|
|
5149
|
+
name: "Kimi K2.5 Free",
|
|
5150
|
+
api: "openai-completions",
|
|
5151
|
+
provider: "opencode",
|
|
5152
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
5153
|
+
reasoning: true,
|
|
5154
|
+
input: ["text", "image"],
|
|
5155
|
+
cost: {
|
|
5156
|
+
input: 0,
|
|
5157
|
+
output: 0,
|
|
5158
|
+
cacheRead: 0,
|
|
5125
5159
|
cacheWrite: 0,
|
|
5126
5160
|
},
|
|
5127
5161
|
contextWindow: 262144,
|
|
@@ -5144,6 +5178,23 @@ export const MODELS = {
|
|
|
5144
5178
|
contextWindow: 204800,
|
|
5145
5179
|
maxTokens: 131072,
|
|
5146
5180
|
},
|
|
5181
|
+
"minimax-m2.1-free": {
|
|
5182
|
+
id: "minimax-m2.1-free",
|
|
5183
|
+
name: "MiniMax M2.1 Free",
|
|
5184
|
+
api: "anthropic-messages",
|
|
5185
|
+
provider: "opencode",
|
|
5186
|
+
baseUrl: "https://opencode.ai/zen",
|
|
5187
|
+
reasoning: true,
|
|
5188
|
+
input: ["text"],
|
|
5189
|
+
cost: {
|
|
5190
|
+
input: 0,
|
|
5191
|
+
output: 0,
|
|
5192
|
+
cacheRead: 0,
|
|
5193
|
+
cacheWrite: 0,
|
|
5194
|
+
},
|
|
5195
|
+
contextWindow: 204800,
|
|
5196
|
+
maxTokens: 131072,
|
|
5197
|
+
},
|
|
5147
5198
|
"qwen3-coder": {
|
|
5148
5199
|
id: "qwen3-coder",
|
|
5149
5200
|
name: "Qwen3 Coder",
|
|
@@ -5161,6 +5212,23 @@ export const MODELS = {
|
|
|
5161
5212
|
contextWindow: 262144,
|
|
5162
5213
|
maxTokens: 65536,
|
|
5163
5214
|
},
|
|
5215
|
+
"trinity-large-preview-free": {
|
|
5216
|
+
id: "trinity-large-preview-free",
|
|
5217
|
+
name: "Trinity Large Preview",
|
|
5218
|
+
api: "openai-completions",
|
|
5219
|
+
provider: "opencode",
|
|
5220
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
5221
|
+
reasoning: false,
|
|
5222
|
+
input: ["text"],
|
|
5223
|
+
cost: {
|
|
5224
|
+
input: 0,
|
|
5225
|
+
output: 0,
|
|
5226
|
+
cacheRead: 0,
|
|
5227
|
+
cacheWrite: 0,
|
|
5228
|
+
},
|
|
5229
|
+
contextWindow: 131072,
|
|
5230
|
+
maxTokens: 131072,
|
|
5231
|
+
},
|
|
5164
5232
|
},
|
|
5165
5233
|
"openrouter": {
|
|
5166
5234
|
"ai21/jamba-large-1.7": {
|
|
@@ -5911,23 +5979,6 @@ export const MODELS = {
|
|
|
5911
5979
|
contextWindow: 1048576,
|
|
5912
5980
|
maxTokens: 8192,
|
|
5913
5981
|
},
|
|
5914
|
-
"google/gemini-2.0-flash-exp:free": {
|
|
5915
|
-
id: "google/gemini-2.0-flash-exp:free",
|
|
5916
|
-
name: "Google: Gemini 2.0 Flash Experimental (free)",
|
|
5917
|
-
api: "openai-completions",
|
|
5918
|
-
provider: "openrouter",
|
|
5919
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5920
|
-
reasoning: false,
|
|
5921
|
-
input: ["text", "image"],
|
|
5922
|
-
cost: {
|
|
5923
|
-
input: 0,
|
|
5924
|
-
output: 0,
|
|
5925
|
-
cacheRead: 0,
|
|
5926
|
-
cacheWrite: 0,
|
|
5927
|
-
},
|
|
5928
|
-
contextWindow: 1048576,
|
|
5929
|
-
maxTokens: 8192,
|
|
5930
|
-
},
|
|
5931
5982
|
"google/gemini-2.0-flash-lite-001": {
|
|
5932
5983
|
id: "google/gemini-2.0-flash-lite-001",
|
|
5933
5984
|
name: "Google: Gemini 2.0 Flash Lite",
|
|
@@ -6923,13 +6974,13 @@ export const MODELS = {
|
|
|
6923
6974
|
reasoning: true,
|
|
6924
6975
|
input: ["text", "image"],
|
|
6925
6976
|
cost: {
|
|
6926
|
-
input: 0.
|
|
6927
|
-
output: 2.
|
|
6977
|
+
input: 0.5,
|
|
6978
|
+
output: 2.8,
|
|
6928
6979
|
cacheRead: 0,
|
|
6929
6980
|
cacheWrite: 0,
|
|
6930
6981
|
},
|
|
6931
6982
|
contextWindow: 262144,
|
|
6932
|
-
maxTokens:
|
|
6983
|
+
maxTokens: 4096,
|
|
6933
6984
|
},
|
|
6934
6985
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
6935
6986
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
@@ -9260,7 +9311,7 @@ export const MODELS = {
|
|
|
9260
9311
|
cacheWrite: 0,
|
|
9261
9312
|
},
|
|
9262
9313
|
contextWindow: 256000,
|
|
9263
|
-
maxTokens:
|
|
9314
|
+
maxTokens: 65536,
|
|
9264
9315
|
},
|
|
9265
9316
|
"anthropic/claude-3-haiku": {
|
|
9266
9317
|
id: "anthropic/claude-3-haiku",
|
|
@@ -10172,9 +10223,9 @@ export const MODELS = {
|
|
|
10172
10223
|
reasoning: true,
|
|
10173
10224
|
input: ["text", "image"],
|
|
10174
10225
|
cost: {
|
|
10175
|
-
input:
|
|
10176
|
-
output:
|
|
10177
|
-
cacheRead: 0.
|
|
10226
|
+
input: 0.6,
|
|
10227
|
+
output: 3,
|
|
10228
|
+
cacheRead: 0.09999999999999999,
|
|
10178
10229
|
cacheWrite: 0,
|
|
10179
10230
|
},
|
|
10180
10231
|
contextWindow: 256000,
|
|
@@ -10293,7 +10344,7 @@ export const MODELS = {
|
|
|
10293
10344
|
cost: {
|
|
10294
10345
|
input: 0.09999999999999999,
|
|
10295
10346
|
output: 0.39999999999999997,
|
|
10296
|
-
cacheRead: 0.
|
|
10347
|
+
cacheRead: 0.03,
|
|
10297
10348
|
cacheWrite: 0,
|
|
10298
10349
|
},
|
|
10299
10350
|
contextWindow: 1047576,
|
|
@@ -10497,7 +10548,7 @@ export const MODELS = {
|
|
|
10497
10548
|
cost: {
|
|
10498
10549
|
input: 1.25,
|
|
10499
10550
|
output: 10,
|
|
10500
|
-
cacheRead: 0.
|
|
10551
|
+
cacheRead: 0.13,
|
|
10501
10552
|
cacheWrite: 0,
|
|
10502
10553
|
},
|
|
10503
10554
|
contextWindow: 128000,
|
|
@@ -10514,7 +10565,7 @@ export const MODELS = {
|
|
|
10514
10565
|
cost: {
|
|
10515
10566
|
input: 1.25,
|
|
10516
10567
|
output: 10,
|
|
10517
|
-
cacheRead: 0.
|
|
10568
|
+
cacheRead: 0.13,
|
|
10518
10569
|
cacheWrite: 0,
|
|
10519
10570
|
},
|
|
10520
10571
|
contextWindow: 400000,
|
|
@@ -10522,7 +10573,7 @@ export const MODELS = {
|
|
|
10522
10573
|
},
|
|
10523
10574
|
"openai/gpt-5.2": {
|
|
10524
10575
|
id: "openai/gpt-5.2",
|
|
10525
|
-
name: "GPT
|
|
10576
|
+
name: "GPT 5.2",
|
|
10526
10577
|
api: "anthropic-messages",
|
|
10527
10578
|
provider: "vercel-ai-gateway",
|
|
10528
10579
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -10531,7 +10582,7 @@ export const MODELS = {
|
|
|
10531
10582
|
cost: {
|
|
10532
10583
|
input: 1.75,
|
|
10533
10584
|
output: 14,
|
|
10534
|
-
cacheRead: 0.
|
|
10585
|
+
cacheRead: 0.18,
|
|
10535
10586
|
cacheWrite: 0,
|
|
10536
10587
|
},
|
|
10537
10588
|
contextWindow: 400000,
|
|
@@ -10792,40 +10843,6 @@ export const MODELS = {
|
|
|
10792
10843
|
contextWindow: 131072,
|
|
10793
10844
|
maxTokens: 131072,
|
|
10794
10845
|
},
|
|
10795
|
-
"stealth/sonoma-dusk-alpha": {
|
|
10796
|
-
id: "stealth/sonoma-dusk-alpha",
|
|
10797
|
-
name: "Sonoma Dusk Alpha",
|
|
10798
|
-
api: "anthropic-messages",
|
|
10799
|
-
provider: "vercel-ai-gateway",
|
|
10800
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10801
|
-
reasoning: false,
|
|
10802
|
-
input: ["text", "image"],
|
|
10803
|
-
cost: {
|
|
10804
|
-
input: 0.19999999999999998,
|
|
10805
|
-
output: 0.5,
|
|
10806
|
-
cacheRead: 0.049999999999999996,
|
|
10807
|
-
cacheWrite: 0,
|
|
10808
|
-
},
|
|
10809
|
-
contextWindow: 2000000,
|
|
10810
|
-
maxTokens: 131072,
|
|
10811
|
-
},
|
|
10812
|
-
"stealth/sonoma-sky-alpha": {
|
|
10813
|
-
id: "stealth/sonoma-sky-alpha",
|
|
10814
|
-
name: "Sonoma Sky Alpha",
|
|
10815
|
-
api: "anthropic-messages",
|
|
10816
|
-
provider: "vercel-ai-gateway",
|
|
10817
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10818
|
-
reasoning: false,
|
|
10819
|
-
input: ["text", "image"],
|
|
10820
|
-
cost: {
|
|
10821
|
-
input: 0.19999999999999998,
|
|
10822
|
-
output: 0.5,
|
|
10823
|
-
cacheRead: 0.049999999999999996,
|
|
10824
|
-
cacheWrite: 0,
|
|
10825
|
-
},
|
|
10826
|
-
contextWindow: 2000000,
|
|
10827
|
-
maxTokens: 131072,
|
|
10828
|
-
},
|
|
10829
10846
|
"vercel/v0-1.0-md": {
|
|
10830
10847
|
id: "vercel/v0-1.0-md",
|
|
10831
10848
|
name: "v0-1.0-md",
|
|
@@ -11113,7 +11130,7 @@ export const MODELS = {
|
|
|
11113
11130
|
cacheWrite: 0,
|
|
11114
11131
|
},
|
|
11115
11132
|
contextWindow: 65536,
|
|
11116
|
-
maxTokens:
|
|
11133
|
+
maxTokens: 16384,
|
|
11117
11134
|
},
|
|
11118
11135
|
"zai/glm-4.6": {
|
|
11119
11136
|
id: "zai/glm-4.6",
|