@mariozechner/pi-ai 0.50.3 → 0.50.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3668,8 +3668,8 @@ export const MODELS = {
3668
3668
  reasoning: false,
3669
3669
  input: ["text"],
3670
3670
  cost: {
3671
- input: 0,
3672
- output: 0,
3671
+ input: 0.4,
3672
+ output: 2,
3673
3673
  cacheRead: 0,
3674
3674
  cacheWrite: 0,
3675
3675
  },
@@ -4923,6 +4923,23 @@ export const MODELS = {
4923
4923
  contextWindow: 204800,
4924
4924
  maxTokens: 131072,
4925
4925
  },
4926
+ "glm-4.7-free": {
4927
+ id: "glm-4.7-free",
4928
+ name: "GLM-4.7 Free",
4929
+ api: "openai-completions",
4930
+ provider: "opencode",
4931
+ baseUrl: "https://opencode.ai/zen/v1",
4932
+ reasoning: true,
4933
+ input: ["text"],
4934
+ cost: {
4935
+ input: 0,
4936
+ output: 0,
4937
+ cacheRead: 0,
4938
+ cacheWrite: 0,
4939
+ },
4940
+ contextWindow: 204800,
4941
+ maxTokens: 131072,
4942
+ },
4926
4943
  "gpt-5": {
4927
4944
  id: "gpt-5",
4928
4945
  name: "GPT-5",
@@ -5121,7 +5138,24 @@ export const MODELS = {
5121
5138
  cost: {
5122
5139
  input: 0.6,
5123
5140
  output: 3,
5124
- cacheRead: 0.1,
5141
+ cacheRead: 0.08,
5142
+ cacheWrite: 0,
5143
+ },
5144
+ contextWindow: 262144,
5145
+ maxTokens: 262144,
5146
+ },
5147
+ "kimi-k2.5-free": {
5148
+ id: "kimi-k2.5-free",
5149
+ name: "Kimi K2.5 Free",
5150
+ api: "openai-completions",
5151
+ provider: "opencode",
5152
+ baseUrl: "https://opencode.ai/zen/v1",
5153
+ reasoning: true,
5154
+ input: ["text", "image"],
5155
+ cost: {
5156
+ input: 0,
5157
+ output: 0,
5158
+ cacheRead: 0,
5125
5159
  cacheWrite: 0,
5126
5160
  },
5127
5161
  contextWindow: 262144,
@@ -5144,6 +5178,23 @@ export const MODELS = {
5144
5178
  contextWindow: 204800,
5145
5179
  maxTokens: 131072,
5146
5180
  },
5181
+ "minimax-m2.1-free": {
5182
+ id: "minimax-m2.1-free",
5183
+ name: "MiniMax M2.1 Free",
5184
+ api: "anthropic-messages",
5185
+ provider: "opencode",
5186
+ baseUrl: "https://opencode.ai/zen",
5187
+ reasoning: true,
5188
+ input: ["text"],
5189
+ cost: {
5190
+ input: 0,
5191
+ output: 0,
5192
+ cacheRead: 0,
5193
+ cacheWrite: 0,
5194
+ },
5195
+ contextWindow: 204800,
5196
+ maxTokens: 131072,
5197
+ },
5147
5198
  "qwen3-coder": {
5148
5199
  id: "qwen3-coder",
5149
5200
  name: "Qwen3 Coder",
@@ -5161,6 +5212,23 @@ export const MODELS = {
5161
5212
  contextWindow: 262144,
5162
5213
  maxTokens: 65536,
5163
5214
  },
5215
+ "trinity-large-preview-free": {
5216
+ id: "trinity-large-preview-free",
5217
+ name: "Trinity Large Preview",
5218
+ api: "openai-completions",
5219
+ provider: "opencode",
5220
+ baseUrl: "https://opencode.ai/zen/v1",
5221
+ reasoning: false,
5222
+ input: ["text"],
5223
+ cost: {
5224
+ input: 0,
5225
+ output: 0,
5226
+ cacheRead: 0,
5227
+ cacheWrite: 0,
5228
+ },
5229
+ contextWindow: 131072,
5230
+ maxTokens: 131072,
5231
+ },
5164
5232
  },
5165
5233
  "openrouter": {
5166
5234
  "ai21/jamba-large-1.7": {
@@ -5911,23 +5979,6 @@ export const MODELS = {
5911
5979
  contextWindow: 1048576,
5912
5980
  maxTokens: 8192,
5913
5981
  },
5914
- "google/gemini-2.0-flash-exp:free": {
5915
- id: "google/gemini-2.0-flash-exp:free",
5916
- name: "Google: Gemini 2.0 Flash Experimental (free)",
5917
- api: "openai-completions",
5918
- provider: "openrouter",
5919
- baseUrl: "https://openrouter.ai/api/v1",
5920
- reasoning: false,
5921
- input: ["text", "image"],
5922
- cost: {
5923
- input: 0,
5924
- output: 0,
5925
- cacheRead: 0,
5926
- cacheWrite: 0,
5927
- },
5928
- contextWindow: 1048576,
5929
- maxTokens: 8192,
5930
- },
5931
5982
  "google/gemini-2.0-flash-lite-001": {
5932
5983
  id: "google/gemini-2.0-flash-lite-001",
5933
5984
  name: "Google: Gemini 2.0 Flash Lite",
@@ -6923,13 +6974,13 @@ export const MODELS = {
6923
6974
  reasoning: true,
6924
6975
  input: ["text", "image"],
6925
6976
  cost: {
6926
- input: 0.5700000000000001,
6927
- output: 2.8499999999999996,
6977
+ input: 0.5,
6978
+ output: 2.8,
6928
6979
  cacheRead: 0,
6929
6980
  cacheWrite: 0,
6930
6981
  },
6931
6982
  contextWindow: 262144,
6932
- maxTokens: 262144,
6983
+ maxTokens: 4096,
6933
6984
  },
6934
6985
  "nex-agi/deepseek-v3.1-nex-n1": {
6935
6986
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -9260,7 +9311,7 @@ export const MODELS = {
9260
9311
  cacheWrite: 0,
9261
9312
  },
9262
9313
  contextWindow: 256000,
9263
- maxTokens: 256000,
9314
+ maxTokens: 65536,
9264
9315
  },
9265
9316
  "anthropic/claude-3-haiku": {
9266
9317
  id: "anthropic/claude-3-haiku",
@@ -10172,9 +10223,9 @@ export const MODELS = {
10172
10223
  reasoning: true,
10173
10224
  input: ["text", "image"],
10174
10225
  cost: {
10175
- input: 1.2,
10176
- output: 1.2,
10177
- cacheRead: 0.6,
10226
+ input: 0.6,
10227
+ output: 3,
10228
+ cacheRead: 0.09999999999999999,
10178
10229
  cacheWrite: 0,
10179
10230
  },
10180
10231
  contextWindow: 256000,
@@ -10293,7 +10344,7 @@ export const MODELS = {
10293
10344
  cost: {
10294
10345
  input: 0.09999999999999999,
10295
10346
  output: 0.39999999999999997,
10296
- cacheRead: 0.024999999999999998,
10347
+ cacheRead: 0.03,
10297
10348
  cacheWrite: 0,
10298
10349
  },
10299
10350
  contextWindow: 1047576,
@@ -10497,7 +10548,7 @@ export const MODELS = {
10497
10548
  cost: {
10498
10549
  input: 1.25,
10499
10550
  output: 10,
10500
- cacheRead: 0.125,
10551
+ cacheRead: 0.13,
10501
10552
  cacheWrite: 0,
10502
10553
  },
10503
10554
  contextWindow: 128000,
@@ -10514,7 +10565,7 @@ export const MODELS = {
10514
10565
  cost: {
10515
10566
  input: 1.25,
10516
10567
  output: 10,
10517
- cacheRead: 0.125,
10568
+ cacheRead: 0.13,
10518
10569
  cacheWrite: 0,
10519
10570
  },
10520
10571
  contextWindow: 400000,
@@ -10522,7 +10573,7 @@ export const MODELS = {
10522
10573
  },
10523
10574
  "openai/gpt-5.2": {
10524
10575
  id: "openai/gpt-5.2",
10525
- name: "GPT-5.2",
10576
+ name: "GPT 5.2",
10526
10577
  api: "anthropic-messages",
10527
10578
  provider: "vercel-ai-gateway",
10528
10579
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -10531,7 +10582,7 @@ export const MODELS = {
10531
10582
  cost: {
10532
10583
  input: 1.75,
10533
10584
  output: 14,
10534
- cacheRead: 0.175,
10585
+ cacheRead: 0.18,
10535
10586
  cacheWrite: 0,
10536
10587
  },
10537
10588
  contextWindow: 400000,
@@ -10792,40 +10843,6 @@ export const MODELS = {
10792
10843
  contextWindow: 131072,
10793
10844
  maxTokens: 131072,
10794
10845
  },
10795
- "stealth/sonoma-dusk-alpha": {
10796
- id: "stealth/sonoma-dusk-alpha",
10797
- name: "Sonoma Dusk Alpha",
10798
- api: "anthropic-messages",
10799
- provider: "vercel-ai-gateway",
10800
- baseUrl: "https://ai-gateway.vercel.sh",
10801
- reasoning: false,
10802
- input: ["text", "image"],
10803
- cost: {
10804
- input: 0.19999999999999998,
10805
- output: 0.5,
10806
- cacheRead: 0.049999999999999996,
10807
- cacheWrite: 0,
10808
- },
10809
- contextWindow: 2000000,
10810
- maxTokens: 131072,
10811
- },
10812
- "stealth/sonoma-sky-alpha": {
10813
- id: "stealth/sonoma-sky-alpha",
10814
- name: "Sonoma Sky Alpha",
10815
- api: "anthropic-messages",
10816
- provider: "vercel-ai-gateway",
10817
- baseUrl: "https://ai-gateway.vercel.sh",
10818
- reasoning: false,
10819
- input: ["text", "image"],
10820
- cost: {
10821
- input: 0.19999999999999998,
10822
- output: 0.5,
10823
- cacheRead: 0.049999999999999996,
10824
- cacheWrite: 0,
10825
- },
10826
- contextWindow: 2000000,
10827
- maxTokens: 131072,
10828
- },
10829
10846
  "vercel/v0-1.0-md": {
10830
10847
  id: "vercel/v0-1.0-md",
10831
10848
  name: "v0-1.0-md",
@@ -11113,7 +11130,7 @@ export const MODELS = {
11113
11130
  cacheWrite: 0,
11114
11131
  },
11115
11132
  contextWindow: 65536,
11116
- maxTokens: 66000,
11133
+ maxTokens: 16384,
11117
11134
  },
11118
11135
  "zai/glm-4.6": {
11119
11136
  id: "zai/glm-4.6",