@mariozechner/pi-ai 0.69.0 → 0.70.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/env-api-keys.d.ts +9 -0
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +41 -31
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.d.ts +1 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +122 -19
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +138 -47
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +5 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +28 -11
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +34 -13
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +8 -7
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +53 -31
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +19 -19
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/types.d.ts +20 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -2488,6 +2488,23 @@ export const MODELS = {
|
|
|
2488
2488
|
contextWindow: 1050000,
|
|
2489
2489
|
maxTokens: 128000,
|
|
2490
2490
|
},
|
|
2491
|
+
"gpt-5.5": {
|
|
2492
|
+
id: "gpt-5.5",
|
|
2493
|
+
name: "GPT-5.5",
|
|
2494
|
+
api: "azure-openai-responses",
|
|
2495
|
+
provider: "azure-openai-responses",
|
|
2496
|
+
baseUrl: "",
|
|
2497
|
+
reasoning: true,
|
|
2498
|
+
input: ["text", "image"],
|
|
2499
|
+
cost: {
|
|
2500
|
+
input: 5,
|
|
2501
|
+
output: 30,
|
|
2502
|
+
cacheRead: 0.5,
|
|
2503
|
+
cacheWrite: 0,
|
|
2504
|
+
},
|
|
2505
|
+
contextWindow: 1050000,
|
|
2506
|
+
maxTokens: 130000,
|
|
2507
|
+
},
|
|
2491
2508
|
"o1": {
|
|
2492
2509
|
id: "o1",
|
|
2493
2510
|
name: "o1",
|
|
@@ -3011,6 +3028,7 @@ export const MODELS = {
|
|
|
3011
3028
|
provider: "github-copilot",
|
|
3012
3029
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3013
3030
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3031
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3014
3032
|
reasoning: true,
|
|
3015
3033
|
input: ["text", "image"],
|
|
3016
3034
|
cost: {
|
|
@@ -3083,6 +3101,7 @@ export const MODELS = {
|
|
|
3083
3101
|
provider: "github-copilot",
|
|
3084
3102
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3085
3103
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3104
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3086
3105
|
reasoning: true,
|
|
3087
3106
|
input: ["text", "image"],
|
|
3088
3107
|
cost: {
|
|
@@ -3101,6 +3120,7 @@ export const MODELS = {
|
|
|
3101
3120
|
provider: "github-copilot",
|
|
3102
3121
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3103
3122
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3123
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3104
3124
|
reasoning: true,
|
|
3105
3125
|
input: ["text", "image"],
|
|
3106
3126
|
cost: {
|
|
@@ -3888,8 +3908,8 @@ export const MODELS = {
|
|
|
3888
3908
|
contextWindow: 131072,
|
|
3889
3909
|
maxTokens: 8192,
|
|
3890
3910
|
},
|
|
3891
|
-
"gemma-4-26b-it": {
|
|
3892
|
-
id: "gemma-4-26b-it",
|
|
3911
|
+
"gemma-4-26b-a4b-it": {
|
|
3912
|
+
id: "gemma-4-26b-a4b-it",
|
|
3893
3913
|
name: "Gemma 4 26B",
|
|
3894
3914
|
api: "google-generative-ai",
|
|
3895
3915
|
provider: "google",
|
|
@@ -5117,6 +5137,7 @@ export const MODELS = {
|
|
|
5117
5137
|
api: "anthropic-messages",
|
|
5118
5138
|
provider: "kimi-coding",
|
|
5119
5139
|
baseUrl: "https://api.kimi.com/coding",
|
|
5140
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5120
5141
|
reasoning: true,
|
|
5121
5142
|
input: ["text", "image"],
|
|
5122
5143
|
cost: {
|
|
@@ -5134,6 +5155,7 @@ export const MODELS = {
|
|
|
5134
5155
|
api: "anthropic-messages",
|
|
5135
5156
|
provider: "kimi-coding",
|
|
5136
5157
|
baseUrl: "https://api.kimi.com/coding",
|
|
5158
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5137
5159
|
reasoning: true,
|
|
5138
5160
|
input: ["text", "image"],
|
|
5139
5161
|
cost: {
|
|
@@ -5151,6 +5173,7 @@ export const MODELS = {
|
|
|
5151
5173
|
api: "anthropic-messages",
|
|
5152
5174
|
provider: "kimi-coding",
|
|
5153
5175
|
baseUrl: "https://api.kimi.com/coding",
|
|
5176
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5154
5177
|
reasoning: true,
|
|
5155
5178
|
input: ["text"],
|
|
5156
5179
|
cost: {
|
|
@@ -6224,6 +6247,23 @@ export const MODELS = {
|
|
|
6224
6247
|
contextWindow: 1050000,
|
|
6225
6248
|
maxTokens: 128000,
|
|
6226
6249
|
},
|
|
6250
|
+
"gpt-5.5": {
|
|
6251
|
+
id: "gpt-5.5",
|
|
6252
|
+
name: "GPT-5.5",
|
|
6253
|
+
api: "openai-responses",
|
|
6254
|
+
provider: "openai",
|
|
6255
|
+
baseUrl: "https://api.openai.com/v1",
|
|
6256
|
+
reasoning: true,
|
|
6257
|
+
input: ["text", "image"],
|
|
6258
|
+
cost: {
|
|
6259
|
+
input: 5,
|
|
6260
|
+
output: 30,
|
|
6261
|
+
cacheRead: 0.5,
|
|
6262
|
+
cacheWrite: 0,
|
|
6263
|
+
},
|
|
6264
|
+
contextWindow: 1050000,
|
|
6265
|
+
maxTokens: 130000,
|
|
6266
|
+
},
|
|
6227
6267
|
"o1": {
|
|
6228
6268
|
id: "o1",
|
|
6229
6269
|
name: "o1",
|
|
@@ -6515,6 +6555,23 @@ export const MODELS = {
|
|
|
6515
6555
|
contextWindow: 272000,
|
|
6516
6556
|
maxTokens: 128000,
|
|
6517
6557
|
},
|
|
6558
|
+
"gpt-5.5": {
|
|
6559
|
+
id: "gpt-5.5",
|
|
6560
|
+
name: "GPT-5.5",
|
|
6561
|
+
api: "openai-codex-responses",
|
|
6562
|
+
provider: "openai-codex",
|
|
6563
|
+
baseUrl: "https://chatgpt.com/backend-api",
|
|
6564
|
+
reasoning: true,
|
|
6565
|
+
input: ["text", "image"],
|
|
6566
|
+
cost: {
|
|
6567
|
+
input: 5,
|
|
6568
|
+
output: 30,
|
|
6569
|
+
cacheRead: 0.5,
|
|
6570
|
+
cacheWrite: 0,
|
|
6571
|
+
},
|
|
6572
|
+
contextWindow: 400000,
|
|
6573
|
+
maxTokens: 128000,
|
|
6574
|
+
},
|
|
6518
6575
|
},
|
|
6519
6576
|
"opencode": {
|
|
6520
6577
|
"big-pickle": {
|
|
@@ -6993,6 +7050,23 @@ export const MODELS = {
|
|
|
6993
7050
|
contextWindow: 1050000,
|
|
6994
7051
|
maxTokens: 128000,
|
|
6995
7052
|
},
|
|
7053
|
+
"hy3-preview-free": {
|
|
7054
|
+
id: "hy3-preview-free",
|
|
7055
|
+
name: "Hy3 preview Free",
|
|
7056
|
+
api: "openai-completions",
|
|
7057
|
+
provider: "opencode",
|
|
7058
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7059
|
+
reasoning: true,
|
|
7060
|
+
input: ["text"],
|
|
7061
|
+
cost: {
|
|
7062
|
+
input: 0,
|
|
7063
|
+
output: 0,
|
|
7064
|
+
cacheRead: 0,
|
|
7065
|
+
cacheWrite: 0,
|
|
7066
|
+
},
|
|
7067
|
+
contextWindow: 256000,
|
|
7068
|
+
maxTokens: 64000,
|
|
7069
|
+
},
|
|
6996
7070
|
"kimi-k2.5": {
|
|
6997
7071
|
id: "kimi-k2.5",
|
|
6998
7072
|
name: "Kimi K2.5",
|
|
@@ -7745,23 +7819,6 @@ export const MODELS = {
|
|
|
7745
7819
|
contextWindow: 131000,
|
|
7746
7820
|
maxTokens: 4096,
|
|
7747
7821
|
},
|
|
7748
|
-
"arcee-ai/trinity-large-preview:free": {
|
|
7749
|
-
id: "arcee-ai/trinity-large-preview:free",
|
|
7750
|
-
name: "Arcee AI: Trinity Large Preview (free)",
|
|
7751
|
-
api: "openai-completions",
|
|
7752
|
-
provider: "openrouter",
|
|
7753
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7754
|
-
reasoning: false,
|
|
7755
|
-
input: ["text"],
|
|
7756
|
-
cost: {
|
|
7757
|
-
input: 0,
|
|
7758
|
-
output: 0,
|
|
7759
|
-
cacheRead: 0,
|
|
7760
|
-
cacheWrite: 0,
|
|
7761
|
-
},
|
|
7762
|
-
contextWindow: 131000,
|
|
7763
|
-
maxTokens: 4096,
|
|
7764
|
-
},
|
|
7765
7822
|
"arcee-ai/trinity-large-thinking": {
|
|
7766
7823
|
id: "arcee-ai/trinity-large-thinking",
|
|
7767
7824
|
name: "Arcee AI: Trinity Large Thinking",
|
|
@@ -8083,7 +8140,7 @@ export const MODELS = {
|
|
|
8083
8140
|
cacheWrite: 0,
|
|
8084
8141
|
},
|
|
8085
8142
|
contextWindow: 131072,
|
|
8086
|
-
maxTokens:
|
|
8143
|
+
maxTokens: 65536,
|
|
8087
8144
|
},
|
|
8088
8145
|
"deepseek/deepseek-v3.2-exp": {
|
|
8089
8146
|
id: "deepseek/deepseek-v3.2-exp",
|
|
@@ -8332,13 +8389,13 @@ export const MODELS = {
|
|
|
8332
8389
|
reasoning: true,
|
|
8333
8390
|
input: ["text", "image"],
|
|
8334
8391
|
cost: {
|
|
8335
|
-
input: 0.
|
|
8336
|
-
output: 0.
|
|
8337
|
-
cacheRead: 0
|
|
8392
|
+
input: 0.06,
|
|
8393
|
+
output: 0.33,
|
|
8394
|
+
cacheRead: 0,
|
|
8338
8395
|
cacheWrite: 0,
|
|
8339
8396
|
},
|
|
8340
8397
|
contextWindow: 262144,
|
|
8341
|
-
maxTokens:
|
|
8398
|
+
maxTokens: 4096,
|
|
8342
8399
|
},
|
|
8343
8400
|
"google/gemma-4-26b-a4b-it:free": {
|
|
8344
8401
|
id: "google/gemma-4-26b-a4b-it:free",
|
|
@@ -8368,7 +8425,7 @@ export const MODELS = {
|
|
|
8368
8425
|
cost: {
|
|
8369
8426
|
input: 0.13,
|
|
8370
8427
|
output: 0.38,
|
|
8371
|
-
cacheRead: 0
|
|
8428
|
+
cacheRead: 0,
|
|
8372
8429
|
cacheWrite: 0,
|
|
8373
8430
|
},
|
|
8374
8431
|
contextWindow: 262144,
|
|
@@ -8408,6 +8465,23 @@ export const MODELS = {
|
|
|
8408
8465
|
contextWindow: 128000,
|
|
8409
8466
|
maxTokens: 50000,
|
|
8410
8467
|
},
|
|
8468
|
+
"inclusionai/ling-2.6-1t:free": {
|
|
8469
|
+
id: "inclusionai/ling-2.6-1t:free",
|
|
8470
|
+
name: "inclusionAI: Ling-2.6-1T (free)",
|
|
8471
|
+
api: "openai-completions",
|
|
8472
|
+
provider: "openrouter",
|
|
8473
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8474
|
+
reasoning: false,
|
|
8475
|
+
input: ["text"],
|
|
8476
|
+
cost: {
|
|
8477
|
+
input: 0,
|
|
8478
|
+
output: 0,
|
|
8479
|
+
cacheRead: 0,
|
|
8480
|
+
cacheWrite: 0,
|
|
8481
|
+
},
|
|
8482
|
+
contextWindow: 262144,
|
|
8483
|
+
maxTokens: 32768,
|
|
8484
|
+
},
|
|
8411
8485
|
"inclusionai/ling-2.6-flash:free": {
|
|
8412
8486
|
id: "inclusionai/ling-2.6-flash:free",
|
|
8413
8487
|
name: "inclusionAI: Ling-2.6-flash (free)",
|
|
@@ -8605,12 +8679,12 @@ export const MODELS = {
|
|
|
8605
8679
|
input: ["text"],
|
|
8606
8680
|
cost: {
|
|
8607
8681
|
input: 0.15,
|
|
8608
|
-
output: 1.
|
|
8609
|
-
cacheRead: 0.
|
|
8682
|
+
output: 1.15,
|
|
8683
|
+
cacheRead: 0.03,
|
|
8610
8684
|
cacheWrite: 0,
|
|
8611
8685
|
},
|
|
8612
8686
|
contextWindow: 196608,
|
|
8613
|
-
maxTokens:
|
|
8687
|
+
maxTokens: 4096,
|
|
8614
8688
|
},
|
|
8615
8689
|
"minimax/minimax-m2.5:free": {
|
|
8616
8690
|
id: "minimax/minimax-m2.5:free",
|
|
@@ -8876,13 +8950,13 @@ export const MODELS = {
|
|
|
8876
8950
|
reasoning: false,
|
|
8877
8951
|
input: ["text"],
|
|
8878
8952
|
cost: {
|
|
8879
|
-
input: 0.
|
|
8880
|
-
output: 0.
|
|
8953
|
+
input: 0.01,
|
|
8954
|
+
output: 0.03,
|
|
8881
8955
|
cacheRead: 0,
|
|
8882
8956
|
cacheWrite: 0,
|
|
8883
8957
|
},
|
|
8884
8958
|
contextWindow: 131072,
|
|
8885
|
-
maxTokens:
|
|
8959
|
+
maxTokens: 4096,
|
|
8886
8960
|
},
|
|
8887
8961
|
"mistralai/mistral-saba": {
|
|
8888
8962
|
id: "mistralai/mistral-saba",
|
|
@@ -9097,13 +9171,13 @@ export const MODELS = {
|
|
|
9097
9171
|
reasoning: true,
|
|
9098
9172
|
input: ["text", "image"],
|
|
9099
9173
|
cost: {
|
|
9100
|
-
input: 0.
|
|
9101
|
-
output:
|
|
9102
|
-
cacheRead: 0.
|
|
9174
|
+
input: 0.7448,
|
|
9175
|
+
output: 4.655,
|
|
9176
|
+
cacheRead: 0.1463,
|
|
9103
9177
|
cacheWrite: 0,
|
|
9104
9178
|
},
|
|
9105
|
-
contextWindow:
|
|
9106
|
-
maxTokens:
|
|
9179
|
+
contextWindow: 256000,
|
|
9180
|
+
maxTokens: 65536,
|
|
9107
9181
|
},
|
|
9108
9182
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
9109
9183
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
@@ -10440,13 +10514,13 @@ export const MODELS = {
|
|
|
10440
10514
|
reasoning: true,
|
|
10441
10515
|
input: ["text"],
|
|
10442
10516
|
cost: {
|
|
10443
|
-
input: 0.
|
|
10444
|
-
output:
|
|
10517
|
+
input: 0.14950000000000002,
|
|
10518
|
+
output: 1.495,
|
|
10445
10519
|
cacheRead: 0,
|
|
10446
10520
|
cacheWrite: 0,
|
|
10447
10521
|
},
|
|
10448
|
-
contextWindow:
|
|
10449
|
-
maxTokens:
|
|
10522
|
+
contextWindow: 131072,
|
|
10523
|
+
maxTokens: 4096,
|
|
10450
10524
|
},
|
|
10451
10525
|
"qwen/qwen3-30b-a3b": {
|
|
10452
10526
|
id: "qwen/qwen3-30b-a3b",
|
|
@@ -11077,6 +11151,23 @@ export const MODELS = {
|
|
|
11077
11151
|
contextWindow: 262144,
|
|
11078
11152
|
maxTokens: 65536,
|
|
11079
11153
|
},
|
|
11154
|
+
"tencent/hy3-preview:free": {
|
|
11155
|
+
id: "tencent/hy3-preview:free",
|
|
11156
|
+
name: "Tencent: Hy3 preview (free)",
|
|
11157
|
+
api: "openai-completions",
|
|
11158
|
+
provider: "openrouter",
|
|
11159
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
11160
|
+
reasoning: true,
|
|
11161
|
+
input: ["text"],
|
|
11162
|
+
cost: {
|
|
11163
|
+
input: 0,
|
|
11164
|
+
output: 0,
|
|
11165
|
+
cacheRead: 0,
|
|
11166
|
+
cacheWrite: 0,
|
|
11167
|
+
},
|
|
11168
|
+
contextWindow: 262144,
|
|
11169
|
+
maxTokens: 262144,
|
|
11170
|
+
},
|
|
11080
11171
|
"thedrummer/rocinante-12b": {
|
|
11081
11172
|
id: "thedrummer/rocinante-12b",
|
|
11082
11173
|
name: "TheDrummer: Rocinante 12B",
|
|
@@ -11907,7 +11998,7 @@ export const MODELS = {
|
|
|
11907
11998
|
input: 0.5,
|
|
11908
11999
|
output: 3,
|
|
11909
12000
|
cacheRead: 0.09999999999999999,
|
|
11910
|
-
cacheWrite: 0,
|
|
12001
|
+
cacheWrite: 0.625,
|
|
11911
12002
|
},
|
|
11912
12003
|
contextWindow: 1000000,
|
|
11913
12004
|
maxTokens: 64000,
|
|
@@ -14029,9 +14120,9 @@ export const MODELS = {
|
|
|
14029
14120
|
reasoning: true,
|
|
14030
14121
|
input: ["text"],
|
|
14031
14122
|
cost: {
|
|
14032
|
-
input: 0.
|
|
14033
|
-
output: 0.
|
|
14034
|
-
cacheRead: 0.
|
|
14123
|
+
input: 0.09999999999999999,
|
|
14124
|
+
output: 0.3,
|
|
14125
|
+
cacheRead: 0.01,
|
|
14035
14126
|
cacheWrite: 0,
|
|
14036
14127
|
},
|
|
14037
14128
|
contextWindow: 262144,
|
|
@@ -14248,15 +14339,15 @@ export const MODELS = {
|
|
|
14248
14339
|
provider: "vercel-ai-gateway",
|
|
14249
14340
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
14250
14341
|
reasoning: true,
|
|
14251
|
-
input: ["text"
|
|
14342
|
+
input: ["text"],
|
|
14252
14343
|
cost: {
|
|
14253
14344
|
input: 1.4,
|
|
14254
14345
|
output: 4.4,
|
|
14255
14346
|
cacheRead: 0.26,
|
|
14256
14347
|
cacheWrite: 0,
|
|
14257
14348
|
},
|
|
14258
|
-
contextWindow:
|
|
14259
|
-
maxTokens:
|
|
14349
|
+
contextWindow: 202800,
|
|
14350
|
+
maxTokens: 64000,
|
|
14260
14351
|
},
|
|
14261
14352
|
"zai/glm-5v-turbo": {
|
|
14262
14353
|
id: "zai/glm-5v-turbo",
|