@mariozechner/pi-ai 0.50.3 → 0.50.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +51 -51
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +67 -67
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +2 -2
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +1 -1
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +2 -2
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +14 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/types.d.ts +13 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -3668,8 +3668,8 @@ export const MODELS = {
|
|
|
3668
3668
|
reasoning: false,
|
|
3669
3669
|
input: ["text"],
|
|
3670
3670
|
cost: {
|
|
3671
|
-
input: 0,
|
|
3672
|
-
output:
|
|
3671
|
+
input: 0.4,
|
|
3672
|
+
output: 2,
|
|
3673
3673
|
cacheRead: 0,
|
|
3674
3674
|
cacheWrite: 0,
|
|
3675
3675
|
},
|
|
@@ -4923,6 +4923,23 @@ export const MODELS = {
|
|
|
4923
4923
|
contextWindow: 204800,
|
|
4924
4924
|
maxTokens: 131072,
|
|
4925
4925
|
},
|
|
4926
|
+
"glm-4.7-free": {
|
|
4927
|
+
id: "glm-4.7-free",
|
|
4928
|
+
name: "GLM-4.7 Free",
|
|
4929
|
+
api: "openai-completions",
|
|
4930
|
+
provider: "opencode",
|
|
4931
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
4932
|
+
reasoning: true,
|
|
4933
|
+
input: ["text"],
|
|
4934
|
+
cost: {
|
|
4935
|
+
input: 0,
|
|
4936
|
+
output: 0,
|
|
4937
|
+
cacheRead: 0,
|
|
4938
|
+
cacheWrite: 0,
|
|
4939
|
+
},
|
|
4940
|
+
contextWindow: 204800,
|
|
4941
|
+
maxTokens: 131072,
|
|
4942
|
+
},
|
|
4926
4943
|
"gpt-5": {
|
|
4927
4944
|
id: "gpt-5",
|
|
4928
4945
|
name: "GPT-5",
|
|
@@ -5121,7 +5138,24 @@ export const MODELS = {
|
|
|
5121
5138
|
cost: {
|
|
5122
5139
|
input: 0.6,
|
|
5123
5140
|
output: 3,
|
|
5124
|
-
cacheRead: 0.
|
|
5141
|
+
cacheRead: 0.08,
|
|
5142
|
+
cacheWrite: 0,
|
|
5143
|
+
},
|
|
5144
|
+
contextWindow: 262144,
|
|
5145
|
+
maxTokens: 262144,
|
|
5146
|
+
},
|
|
5147
|
+
"kimi-k2.5-free": {
|
|
5148
|
+
id: "kimi-k2.5-free",
|
|
5149
|
+
name: "Kimi K2.5 Free",
|
|
5150
|
+
api: "openai-completions",
|
|
5151
|
+
provider: "opencode",
|
|
5152
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
5153
|
+
reasoning: true,
|
|
5154
|
+
input: ["text", "image"],
|
|
5155
|
+
cost: {
|
|
5156
|
+
input: 0,
|
|
5157
|
+
output: 0,
|
|
5158
|
+
cacheRead: 0,
|
|
5125
5159
|
cacheWrite: 0,
|
|
5126
5160
|
},
|
|
5127
5161
|
contextWindow: 262144,
|
|
@@ -5144,6 +5178,23 @@ export const MODELS = {
|
|
|
5144
5178
|
contextWindow: 204800,
|
|
5145
5179
|
maxTokens: 131072,
|
|
5146
5180
|
},
|
|
5181
|
+
"minimax-m2.1-free": {
|
|
5182
|
+
id: "minimax-m2.1-free",
|
|
5183
|
+
name: "MiniMax M2.1 Free",
|
|
5184
|
+
api: "anthropic-messages",
|
|
5185
|
+
provider: "opencode",
|
|
5186
|
+
baseUrl: "https://opencode.ai/zen",
|
|
5187
|
+
reasoning: true,
|
|
5188
|
+
input: ["text"],
|
|
5189
|
+
cost: {
|
|
5190
|
+
input: 0,
|
|
5191
|
+
output: 0,
|
|
5192
|
+
cacheRead: 0,
|
|
5193
|
+
cacheWrite: 0,
|
|
5194
|
+
},
|
|
5195
|
+
contextWindow: 204800,
|
|
5196
|
+
maxTokens: 131072,
|
|
5197
|
+
},
|
|
5147
5198
|
"qwen3-coder": {
|
|
5148
5199
|
id: "qwen3-coder",
|
|
5149
5200
|
name: "Qwen3 Coder",
|
|
@@ -5911,23 +5962,6 @@ export const MODELS = {
|
|
|
5911
5962
|
contextWindow: 1048576,
|
|
5912
5963
|
maxTokens: 8192,
|
|
5913
5964
|
},
|
|
5914
|
-
"google/gemini-2.0-flash-exp:free": {
|
|
5915
|
-
id: "google/gemini-2.0-flash-exp:free",
|
|
5916
|
-
name: "Google: Gemini 2.0 Flash Experimental (free)",
|
|
5917
|
-
api: "openai-completions",
|
|
5918
|
-
provider: "openrouter",
|
|
5919
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5920
|
-
reasoning: false,
|
|
5921
|
-
input: ["text", "image"],
|
|
5922
|
-
cost: {
|
|
5923
|
-
input: 0,
|
|
5924
|
-
output: 0,
|
|
5925
|
-
cacheRead: 0,
|
|
5926
|
-
cacheWrite: 0,
|
|
5927
|
-
},
|
|
5928
|
-
contextWindow: 1048576,
|
|
5929
|
-
maxTokens: 8192,
|
|
5930
|
-
},
|
|
5931
5965
|
"google/gemini-2.0-flash-lite-001": {
|
|
5932
5966
|
id: "google/gemini-2.0-flash-lite-001",
|
|
5933
5967
|
name: "Google: Gemini 2.0 Flash Lite",
|
|
@@ -6923,13 +6957,13 @@ export const MODELS = {
|
|
|
6923
6957
|
reasoning: true,
|
|
6924
6958
|
input: ["text", "image"],
|
|
6925
6959
|
cost: {
|
|
6926
|
-
input: 0.
|
|
6927
|
-
output: 2.
|
|
6960
|
+
input: 0.5,
|
|
6961
|
+
output: 2.8,
|
|
6928
6962
|
cacheRead: 0,
|
|
6929
6963
|
cacheWrite: 0,
|
|
6930
6964
|
},
|
|
6931
6965
|
contextWindow: 262144,
|
|
6932
|
-
maxTokens:
|
|
6966
|
+
maxTokens: 4096,
|
|
6933
6967
|
},
|
|
6934
6968
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
6935
6969
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
@@ -9260,7 +9294,7 @@ export const MODELS = {
|
|
|
9260
9294
|
cacheWrite: 0,
|
|
9261
9295
|
},
|
|
9262
9296
|
contextWindow: 256000,
|
|
9263
|
-
maxTokens:
|
|
9297
|
+
maxTokens: 65536,
|
|
9264
9298
|
},
|
|
9265
9299
|
"anthropic/claude-3-haiku": {
|
|
9266
9300
|
id: "anthropic/claude-3-haiku",
|
|
@@ -10172,9 +10206,9 @@ export const MODELS = {
|
|
|
10172
10206
|
reasoning: true,
|
|
10173
10207
|
input: ["text", "image"],
|
|
10174
10208
|
cost: {
|
|
10175
|
-
input:
|
|
10176
|
-
output:
|
|
10177
|
-
cacheRead: 0.
|
|
10209
|
+
input: 0.6,
|
|
10210
|
+
output: 3,
|
|
10211
|
+
cacheRead: 0.09999999999999999,
|
|
10178
10212
|
cacheWrite: 0,
|
|
10179
10213
|
},
|
|
10180
10214
|
contextWindow: 256000,
|
|
@@ -10293,7 +10327,7 @@ export const MODELS = {
|
|
|
10293
10327
|
cost: {
|
|
10294
10328
|
input: 0.09999999999999999,
|
|
10295
10329
|
output: 0.39999999999999997,
|
|
10296
|
-
cacheRead: 0.
|
|
10330
|
+
cacheRead: 0.03,
|
|
10297
10331
|
cacheWrite: 0,
|
|
10298
10332
|
},
|
|
10299
10333
|
contextWindow: 1047576,
|
|
@@ -10497,7 +10531,7 @@ export const MODELS = {
|
|
|
10497
10531
|
cost: {
|
|
10498
10532
|
input: 1.25,
|
|
10499
10533
|
output: 10,
|
|
10500
|
-
cacheRead: 0.
|
|
10534
|
+
cacheRead: 0.13,
|
|
10501
10535
|
cacheWrite: 0,
|
|
10502
10536
|
},
|
|
10503
10537
|
contextWindow: 128000,
|
|
@@ -10514,7 +10548,7 @@ export const MODELS = {
|
|
|
10514
10548
|
cost: {
|
|
10515
10549
|
input: 1.25,
|
|
10516
10550
|
output: 10,
|
|
10517
|
-
cacheRead: 0.
|
|
10551
|
+
cacheRead: 0.13,
|
|
10518
10552
|
cacheWrite: 0,
|
|
10519
10553
|
},
|
|
10520
10554
|
contextWindow: 400000,
|
|
@@ -10522,7 +10556,7 @@ export const MODELS = {
|
|
|
10522
10556
|
},
|
|
10523
10557
|
"openai/gpt-5.2": {
|
|
10524
10558
|
id: "openai/gpt-5.2",
|
|
10525
|
-
name: "GPT
|
|
10559
|
+
name: "GPT 5.2",
|
|
10526
10560
|
api: "anthropic-messages",
|
|
10527
10561
|
provider: "vercel-ai-gateway",
|
|
10528
10562
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
@@ -10531,7 +10565,7 @@ export const MODELS = {
|
|
|
10531
10565
|
cost: {
|
|
10532
10566
|
input: 1.75,
|
|
10533
10567
|
output: 14,
|
|
10534
|
-
cacheRead: 0.
|
|
10568
|
+
cacheRead: 0.18,
|
|
10535
10569
|
cacheWrite: 0,
|
|
10536
10570
|
},
|
|
10537
10571
|
contextWindow: 400000,
|
|
@@ -10792,40 +10826,6 @@ export const MODELS = {
|
|
|
10792
10826
|
contextWindow: 131072,
|
|
10793
10827
|
maxTokens: 131072,
|
|
10794
10828
|
},
|
|
10795
|
-
"stealth/sonoma-dusk-alpha": {
|
|
10796
|
-
id: "stealth/sonoma-dusk-alpha",
|
|
10797
|
-
name: "Sonoma Dusk Alpha",
|
|
10798
|
-
api: "anthropic-messages",
|
|
10799
|
-
provider: "vercel-ai-gateway",
|
|
10800
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10801
|
-
reasoning: false,
|
|
10802
|
-
input: ["text", "image"],
|
|
10803
|
-
cost: {
|
|
10804
|
-
input: 0.19999999999999998,
|
|
10805
|
-
output: 0.5,
|
|
10806
|
-
cacheRead: 0.049999999999999996,
|
|
10807
|
-
cacheWrite: 0,
|
|
10808
|
-
},
|
|
10809
|
-
contextWindow: 2000000,
|
|
10810
|
-
maxTokens: 131072,
|
|
10811
|
-
},
|
|
10812
|
-
"stealth/sonoma-sky-alpha": {
|
|
10813
|
-
id: "stealth/sonoma-sky-alpha",
|
|
10814
|
-
name: "Sonoma Sky Alpha",
|
|
10815
|
-
api: "anthropic-messages",
|
|
10816
|
-
provider: "vercel-ai-gateway",
|
|
10817
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
10818
|
-
reasoning: false,
|
|
10819
|
-
input: ["text", "image"],
|
|
10820
|
-
cost: {
|
|
10821
|
-
input: 0.19999999999999998,
|
|
10822
|
-
output: 0.5,
|
|
10823
|
-
cacheRead: 0.049999999999999996,
|
|
10824
|
-
cacheWrite: 0,
|
|
10825
|
-
},
|
|
10826
|
-
contextWindow: 2000000,
|
|
10827
|
-
maxTokens: 131072,
|
|
10828
|
-
},
|
|
10829
10829
|
"vercel/v0-1.0-md": {
|
|
10830
10830
|
id: "vercel/v0-1.0-md",
|
|
10831
10831
|
name: "v0-1.0-md",
|
|
@@ -11113,7 +11113,7 @@ export const MODELS = {
|
|
|
11113
11113
|
cacheWrite: 0,
|
|
11114
11114
|
},
|
|
11115
11115
|
contextWindow: 65536,
|
|
11116
|
-
maxTokens:
|
|
11116
|
+
maxTokens: 16384,
|
|
11117
11117
|
},
|
|
11118
11118
|
"zai/glm-4.6": {
|
|
11119
11119
|
id: "zai/glm-4.6",
|