@mariozechner/pi-ai 0.69.0 → 0.70.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2488,6 +2488,23 @@ export const MODELS = {
2488
2488
  contextWindow: 1050000,
2489
2489
  maxTokens: 128000,
2490
2490
  },
2491
+ "gpt-5.5": {
2492
+ id: "gpt-5.5",
2493
+ name: "GPT-5.5",
2494
+ api: "azure-openai-responses",
2495
+ provider: "azure-openai-responses",
2496
+ baseUrl: "",
2497
+ reasoning: true,
2498
+ input: ["text", "image"],
2499
+ cost: {
2500
+ input: 5,
2501
+ output: 30,
2502
+ cacheRead: 0.5,
2503
+ cacheWrite: 0,
2504
+ },
2505
+ contextWindow: 1050000,
2506
+ maxTokens: 130000,
2507
+ },
2491
2508
  "o1": {
2492
2509
  id: "o1",
2493
2510
  name: "o1",
@@ -3011,6 +3028,7 @@ export const MODELS = {
3011
3028
  provider: "github-copilot",
3012
3029
  baseUrl: "https://api.individual.githubcopilot.com",
3013
3030
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3031
+ compat: { "supportsEagerToolInputStreaming": false },
3014
3032
  reasoning: true,
3015
3033
  input: ["text", "image"],
3016
3034
  cost: {
@@ -3083,6 +3101,7 @@ export const MODELS = {
3083
3101
  provider: "github-copilot",
3084
3102
  baseUrl: "https://api.individual.githubcopilot.com",
3085
3103
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3104
+ compat: { "supportsEagerToolInputStreaming": false },
3086
3105
  reasoning: true,
3087
3106
  input: ["text", "image"],
3088
3107
  cost: {
@@ -3101,6 +3120,7 @@ export const MODELS = {
3101
3120
  provider: "github-copilot",
3102
3121
  baseUrl: "https://api.individual.githubcopilot.com",
3103
3122
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3123
+ compat: { "supportsEagerToolInputStreaming": false },
3104
3124
  reasoning: true,
3105
3125
  input: ["text", "image"],
3106
3126
  cost: {
@@ -3888,8 +3908,8 @@ export const MODELS = {
3888
3908
  contextWindow: 131072,
3889
3909
  maxTokens: 8192,
3890
3910
  },
3891
- "gemma-4-26b-it": {
3892
- id: "gemma-4-26b-it",
3911
+ "gemma-4-26b-a4b-it": {
3912
+ id: "gemma-4-26b-a4b-it",
3893
3913
  name: "Gemma 4 26B",
3894
3914
  api: "google-generative-ai",
3895
3915
  provider: "google",
@@ -5117,6 +5137,7 @@ export const MODELS = {
5117
5137
  api: "anthropic-messages",
5118
5138
  provider: "kimi-coding",
5119
5139
  baseUrl: "https://api.kimi.com/coding",
5140
+ headers: { "User-Agent": "KimiCLI/1.5" },
5120
5141
  reasoning: true,
5121
5142
  input: ["text", "image"],
5122
5143
  cost: {
@@ -5134,6 +5155,7 @@ export const MODELS = {
5134
5155
  api: "anthropic-messages",
5135
5156
  provider: "kimi-coding",
5136
5157
  baseUrl: "https://api.kimi.com/coding",
5158
+ headers: { "User-Agent": "KimiCLI/1.5" },
5137
5159
  reasoning: true,
5138
5160
  input: ["text", "image"],
5139
5161
  cost: {
@@ -5151,6 +5173,7 @@ export const MODELS = {
5151
5173
  api: "anthropic-messages",
5152
5174
  provider: "kimi-coding",
5153
5175
  baseUrl: "https://api.kimi.com/coding",
5176
+ headers: { "User-Agent": "KimiCLI/1.5" },
5154
5177
  reasoning: true,
5155
5178
  input: ["text"],
5156
5179
  cost: {
@@ -6224,6 +6247,23 @@ export const MODELS = {
6224
6247
  contextWindow: 1050000,
6225
6248
  maxTokens: 128000,
6226
6249
  },
6250
+ "gpt-5.5": {
6251
+ id: "gpt-5.5",
6252
+ name: "GPT-5.5",
6253
+ api: "openai-responses",
6254
+ provider: "openai",
6255
+ baseUrl: "https://api.openai.com/v1",
6256
+ reasoning: true,
6257
+ input: ["text", "image"],
6258
+ cost: {
6259
+ input: 5,
6260
+ output: 30,
6261
+ cacheRead: 0.5,
6262
+ cacheWrite: 0,
6263
+ },
6264
+ contextWindow: 1050000,
6265
+ maxTokens: 130000,
6266
+ },
6227
6267
  "o1": {
6228
6268
  id: "o1",
6229
6269
  name: "o1",
@@ -6515,6 +6555,23 @@ export const MODELS = {
6515
6555
  contextWindow: 272000,
6516
6556
  maxTokens: 128000,
6517
6557
  },
6558
+ "gpt-5.5": {
6559
+ id: "gpt-5.5",
6560
+ name: "GPT-5.5",
6561
+ api: "openai-codex-responses",
6562
+ provider: "openai-codex",
6563
+ baseUrl: "https://chatgpt.com/backend-api",
6564
+ reasoning: true,
6565
+ input: ["text", "image"],
6566
+ cost: {
6567
+ input: 5,
6568
+ output: 30,
6569
+ cacheRead: 0.5,
6570
+ cacheWrite: 0,
6571
+ },
6572
+ contextWindow: 400000,
6573
+ maxTokens: 128000,
6574
+ },
6518
6575
  },
6519
6576
  "opencode": {
6520
6577
  "big-pickle": {
@@ -6993,6 +7050,23 @@ export const MODELS = {
6993
7050
  contextWindow: 1050000,
6994
7051
  maxTokens: 128000,
6995
7052
  },
7053
+ "hy3-preview-free": {
7054
+ id: "hy3-preview-free",
7055
+ name: "Hy3 preview Free",
7056
+ api: "openai-completions",
7057
+ provider: "opencode",
7058
+ baseUrl: "https://opencode.ai/zen/v1",
7059
+ reasoning: true,
7060
+ input: ["text"],
7061
+ cost: {
7062
+ input: 0,
7063
+ output: 0,
7064
+ cacheRead: 0,
7065
+ cacheWrite: 0,
7066
+ },
7067
+ contextWindow: 256000,
7068
+ maxTokens: 64000,
7069
+ },
6996
7070
  "kimi-k2.5": {
6997
7071
  id: "kimi-k2.5",
6998
7072
  name: "Kimi K2.5",
@@ -7745,23 +7819,6 @@ export const MODELS = {
7745
7819
  contextWindow: 131000,
7746
7820
  maxTokens: 4096,
7747
7821
  },
7748
- "arcee-ai/trinity-large-preview:free": {
7749
- id: "arcee-ai/trinity-large-preview:free",
7750
- name: "Arcee AI: Trinity Large Preview (free)",
7751
- api: "openai-completions",
7752
- provider: "openrouter",
7753
- baseUrl: "https://openrouter.ai/api/v1",
7754
- reasoning: false,
7755
- input: ["text"],
7756
- cost: {
7757
- input: 0,
7758
- output: 0,
7759
- cacheRead: 0,
7760
- cacheWrite: 0,
7761
- },
7762
- contextWindow: 131000,
7763
- maxTokens: 4096,
7764
- },
7765
7822
  "arcee-ai/trinity-large-thinking": {
7766
7823
  id: "arcee-ai/trinity-large-thinking",
7767
7824
  name: "Arcee AI: Trinity Large Thinking",
@@ -8083,7 +8140,7 @@ export const MODELS = {
8083
8140
  cacheWrite: 0,
8084
8141
  },
8085
8142
  contextWindow: 131072,
8086
- maxTokens: 32768,
8143
+ maxTokens: 65536,
8087
8144
  },
8088
8145
  "deepseek/deepseek-v3.2-exp": {
8089
8146
  id: "deepseek/deepseek-v3.2-exp",
@@ -8332,13 +8389,13 @@ export const MODELS = {
8332
8389
  reasoning: true,
8333
8390
  input: ["text", "image"],
8334
8391
  cost: {
8335
- input: 0.07,
8336
- output: 0.35,
8337
- cacheRead: 0.04,
8392
+ input: 0.06,
8393
+ output: 0.33,
8394
+ cacheRead: 0,
8338
8395
  cacheWrite: 0,
8339
8396
  },
8340
8397
  contextWindow: 262144,
8341
- maxTokens: 262144,
8398
+ maxTokens: 4096,
8342
8399
  },
8343
8400
  "google/gemma-4-26b-a4b-it:free": {
8344
8401
  id: "google/gemma-4-26b-a4b-it:free",
@@ -8368,7 +8425,7 @@ export const MODELS = {
8368
8425
  cost: {
8369
8426
  input: 0.13,
8370
8427
  output: 0.38,
8371
- cacheRead: 0.019999999499999997,
8428
+ cacheRead: 0,
8372
8429
  cacheWrite: 0,
8373
8430
  },
8374
8431
  contextWindow: 262144,
@@ -8408,6 +8465,23 @@ export const MODELS = {
8408
8465
  contextWindow: 128000,
8409
8466
  maxTokens: 50000,
8410
8467
  },
8468
+ "inclusionai/ling-2.6-1t:free": {
8469
+ id: "inclusionai/ling-2.6-1t:free",
8470
+ name: "inclusionAI: Ling-2.6-1T (free)",
8471
+ api: "openai-completions",
8472
+ provider: "openrouter",
8473
+ baseUrl: "https://openrouter.ai/api/v1",
8474
+ reasoning: false,
8475
+ input: ["text"],
8476
+ cost: {
8477
+ input: 0,
8478
+ output: 0,
8479
+ cacheRead: 0,
8480
+ cacheWrite: 0,
8481
+ },
8482
+ contextWindow: 262144,
8483
+ maxTokens: 32768,
8484
+ },
8411
8485
  "inclusionai/ling-2.6-flash:free": {
8412
8486
  id: "inclusionai/ling-2.6-flash:free",
8413
8487
  name: "inclusionAI: Ling-2.6-flash (free)",
@@ -8605,12 +8679,12 @@ export const MODELS = {
8605
8679
  input: ["text"],
8606
8680
  cost: {
8607
8681
  input: 0.15,
8608
- output: 1.2,
8609
- cacheRead: 0.075,
8682
+ output: 1.15,
8683
+ cacheRead: 0.03,
8610
8684
  cacheWrite: 0,
8611
8685
  },
8612
8686
  contextWindow: 196608,
8613
- maxTokens: 65536,
8687
+ maxTokens: 4096,
8614
8688
  },
8615
8689
  "minimax/minimax-m2.5:free": {
8616
8690
  id: "minimax/minimax-m2.5:free",
@@ -8876,13 +8950,13 @@ export const MODELS = {
8876
8950
  reasoning: false,
8877
8951
  input: ["text"],
8878
8952
  cost: {
8879
- input: 0.02,
8880
- output: 0.04,
8953
+ input: 0.01,
8954
+ output: 0.03,
8881
8955
  cacheRead: 0,
8882
8956
  cacheWrite: 0,
8883
8957
  },
8884
8958
  contextWindow: 131072,
8885
- maxTokens: 16384,
8959
+ maxTokens: 4096,
8886
8960
  },
8887
8961
  "mistralai/mistral-saba": {
8888
8962
  id: "mistralai/mistral-saba",
@@ -9097,13 +9171,13 @@ export const MODELS = {
9097
9171
  reasoning: true,
9098
9172
  input: ["text", "image"],
9099
9173
  cost: {
9100
- input: 0.7999999999999999,
9101
- output: 3.5,
9102
- cacheRead: 0.19999999999999998,
9174
+ input: 0.7448,
9175
+ output: 4.655,
9176
+ cacheRead: 0.1463,
9103
9177
  cacheWrite: 0,
9104
9178
  },
9105
- contextWindow: 262144,
9106
- maxTokens: 262144,
9179
+ contextWindow: 256000,
9180
+ maxTokens: 65536,
9107
9181
  },
9108
9182
  "nex-agi/deepseek-v3.1-nex-n1": {
9109
9183
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -10440,13 +10514,13 @@ export const MODELS = {
10440
10514
  reasoning: true,
10441
10515
  input: ["text"],
10442
10516
  cost: {
10443
- input: 0.13,
10444
- output: 0.6,
10517
+ input: 0.14950000000000002,
10518
+ output: 1.495,
10445
10519
  cacheRead: 0,
10446
10520
  cacheWrite: 0,
10447
10521
  },
10448
- contextWindow: 262144,
10449
- maxTokens: 262144,
10522
+ contextWindow: 131072,
10523
+ maxTokens: 4096,
10450
10524
  },
10451
10525
  "qwen/qwen3-30b-a3b": {
10452
10526
  id: "qwen/qwen3-30b-a3b",
@@ -11077,6 +11151,23 @@ export const MODELS = {
11077
11151
  contextWindow: 262144,
11078
11152
  maxTokens: 65536,
11079
11153
  },
11154
+ "tencent/hy3-preview:free": {
11155
+ id: "tencent/hy3-preview:free",
11156
+ name: "Tencent: Hy3 preview (free)",
11157
+ api: "openai-completions",
11158
+ provider: "openrouter",
11159
+ baseUrl: "https://openrouter.ai/api/v1",
11160
+ reasoning: true,
11161
+ input: ["text"],
11162
+ cost: {
11163
+ input: 0,
11164
+ output: 0,
11165
+ cacheRead: 0,
11166
+ cacheWrite: 0,
11167
+ },
11168
+ contextWindow: 262144,
11169
+ maxTokens: 262144,
11170
+ },
11080
11171
  "thedrummer/rocinante-12b": {
11081
11172
  id: "thedrummer/rocinante-12b",
11082
11173
  name: "TheDrummer: Rocinante 12B",
@@ -11907,7 +11998,7 @@ export const MODELS = {
11907
11998
  input: 0.5,
11908
11999
  output: 3,
11909
12000
  cacheRead: 0.09999999999999999,
11910
- cacheWrite: 0,
12001
+ cacheWrite: 0.625,
11911
12002
  },
11912
12003
  contextWindow: 1000000,
11913
12004
  maxTokens: 64000,
@@ -14029,9 +14120,9 @@ export const MODELS = {
14029
14120
  reasoning: true,
14030
14121
  input: ["text"],
14031
14122
  cost: {
14032
- input: 0.09,
14033
- output: 0.29,
14034
- cacheRead: 0.045,
14123
+ input: 0.09999999999999999,
14124
+ output: 0.3,
14125
+ cacheRead: 0.01,
14035
14126
  cacheWrite: 0,
14036
14127
  },
14037
14128
  contextWindow: 262144,
@@ -14248,15 +14339,15 @@ export const MODELS = {
14248
14339
  provider: "vercel-ai-gateway",
14249
14340
  baseUrl: "https://ai-gateway.vercel.sh",
14250
14341
  reasoning: true,
14251
- input: ["text", "image"],
14342
+ input: ["text"],
14252
14343
  cost: {
14253
14344
  input: 1.4,
14254
14345
  output: 4.4,
14255
14346
  cacheRead: 0.26,
14256
14347
  cacheWrite: 0,
14257
14348
  },
14258
- contextWindow: 202752,
14259
- maxTokens: 202752,
14349
+ contextWindow: 202800,
14350
+ maxTokens: 64000,
14260
14351
  },
14261
14352
  "zai/glm-5v-turbo": {
14262
14353
  id: "zai/glm-5v-turbo",