@mariozechner/pi-ai 0.50.3 → 0.50.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3668,8 +3668,8 @@ export const MODELS = {
3668
3668
  reasoning: false,
3669
3669
  input: ["text"],
3670
3670
  cost: {
3671
- input: 0,
3672
- output: 0,
3671
+ input: 0.4,
3672
+ output: 2,
3673
3673
  cacheRead: 0,
3674
3674
  cacheWrite: 0,
3675
3675
  },
@@ -4923,6 +4923,23 @@ export const MODELS = {
4923
4923
  contextWindow: 204800,
4924
4924
  maxTokens: 131072,
4925
4925
  },
4926
+ "glm-4.7-free": {
4927
+ id: "glm-4.7-free",
4928
+ name: "GLM-4.7 Free",
4929
+ api: "openai-completions",
4930
+ provider: "opencode",
4931
+ baseUrl: "https://opencode.ai/zen/v1",
4932
+ reasoning: true,
4933
+ input: ["text"],
4934
+ cost: {
4935
+ input: 0,
4936
+ output: 0,
4937
+ cacheRead: 0,
4938
+ cacheWrite: 0,
4939
+ },
4940
+ contextWindow: 204800,
4941
+ maxTokens: 131072,
4942
+ },
4926
4943
  "gpt-5": {
4927
4944
  id: "gpt-5",
4928
4945
  name: "GPT-5",
@@ -5121,7 +5138,24 @@ export const MODELS = {
5121
5138
  cost: {
5122
5139
  input: 0.6,
5123
5140
  output: 3,
5124
- cacheRead: 0.1,
5141
+ cacheRead: 0.08,
5142
+ cacheWrite: 0,
5143
+ },
5144
+ contextWindow: 262144,
5145
+ maxTokens: 262144,
5146
+ },
5147
+ "kimi-k2.5-free": {
5148
+ id: "kimi-k2.5-free",
5149
+ name: "Kimi K2.5 Free",
5150
+ api: "openai-completions",
5151
+ provider: "opencode",
5152
+ baseUrl: "https://opencode.ai/zen/v1",
5153
+ reasoning: true,
5154
+ input: ["text", "image"],
5155
+ cost: {
5156
+ input: 0,
5157
+ output: 0,
5158
+ cacheRead: 0,
5125
5159
  cacheWrite: 0,
5126
5160
  },
5127
5161
  contextWindow: 262144,
@@ -5144,6 +5178,23 @@ export const MODELS = {
5144
5178
  contextWindow: 204800,
5145
5179
  maxTokens: 131072,
5146
5180
  },
5181
+ "minimax-m2.1-free": {
5182
+ id: "minimax-m2.1-free",
5183
+ name: "MiniMax M2.1 Free",
5184
+ api: "anthropic-messages",
5185
+ provider: "opencode",
5186
+ baseUrl: "https://opencode.ai/zen",
5187
+ reasoning: true,
5188
+ input: ["text"],
5189
+ cost: {
5190
+ input: 0,
5191
+ output: 0,
5192
+ cacheRead: 0,
5193
+ cacheWrite: 0,
5194
+ },
5195
+ contextWindow: 204800,
5196
+ maxTokens: 131072,
5197
+ },
5147
5198
  "qwen3-coder": {
5148
5199
  id: "qwen3-coder",
5149
5200
  name: "Qwen3 Coder",
@@ -5911,23 +5962,6 @@ export const MODELS = {
5911
5962
  contextWindow: 1048576,
5912
5963
  maxTokens: 8192,
5913
5964
  },
5914
- "google/gemini-2.0-flash-exp:free": {
5915
- id: "google/gemini-2.0-flash-exp:free",
5916
- name: "Google: Gemini 2.0 Flash Experimental (free)",
5917
- api: "openai-completions",
5918
- provider: "openrouter",
5919
- baseUrl: "https://openrouter.ai/api/v1",
5920
- reasoning: false,
5921
- input: ["text", "image"],
5922
- cost: {
5923
- input: 0,
5924
- output: 0,
5925
- cacheRead: 0,
5926
- cacheWrite: 0,
5927
- },
5928
- contextWindow: 1048576,
5929
- maxTokens: 8192,
5930
- },
5931
5965
  "google/gemini-2.0-flash-lite-001": {
5932
5966
  id: "google/gemini-2.0-flash-lite-001",
5933
5967
  name: "Google: Gemini 2.0 Flash Lite",
@@ -6923,13 +6957,13 @@ export const MODELS = {
6923
6957
  reasoning: true,
6924
6958
  input: ["text", "image"],
6925
6959
  cost: {
6926
- input: 0.5700000000000001,
6927
- output: 2.8499999999999996,
6960
+ input: 0.5,
6961
+ output: 2.8,
6928
6962
  cacheRead: 0,
6929
6963
  cacheWrite: 0,
6930
6964
  },
6931
6965
  contextWindow: 262144,
6932
- maxTokens: 262144,
6966
+ maxTokens: 4096,
6933
6967
  },
6934
6968
  "nex-agi/deepseek-v3.1-nex-n1": {
6935
6969
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -9260,7 +9294,7 @@ export const MODELS = {
9260
9294
  cacheWrite: 0,
9261
9295
  },
9262
9296
  contextWindow: 256000,
9263
- maxTokens: 256000,
9297
+ maxTokens: 65536,
9264
9298
  },
9265
9299
  "anthropic/claude-3-haiku": {
9266
9300
  id: "anthropic/claude-3-haiku",
@@ -10172,9 +10206,9 @@ export const MODELS = {
10172
10206
  reasoning: true,
10173
10207
  input: ["text", "image"],
10174
10208
  cost: {
10175
- input: 1.2,
10176
- output: 1.2,
10177
- cacheRead: 0.6,
10209
+ input: 0.6,
10210
+ output: 3,
10211
+ cacheRead: 0.09999999999999999,
10178
10212
  cacheWrite: 0,
10179
10213
  },
10180
10214
  contextWindow: 256000,
@@ -10293,7 +10327,7 @@ export const MODELS = {
10293
10327
  cost: {
10294
10328
  input: 0.09999999999999999,
10295
10329
  output: 0.39999999999999997,
10296
- cacheRead: 0.024999999999999998,
10330
+ cacheRead: 0.03,
10297
10331
  cacheWrite: 0,
10298
10332
  },
10299
10333
  contextWindow: 1047576,
@@ -10497,7 +10531,7 @@ export const MODELS = {
10497
10531
  cost: {
10498
10532
  input: 1.25,
10499
10533
  output: 10,
10500
- cacheRead: 0.125,
10534
+ cacheRead: 0.13,
10501
10535
  cacheWrite: 0,
10502
10536
  },
10503
10537
  contextWindow: 128000,
@@ -10514,7 +10548,7 @@ export const MODELS = {
10514
10548
  cost: {
10515
10549
  input: 1.25,
10516
10550
  output: 10,
10517
- cacheRead: 0.125,
10551
+ cacheRead: 0.13,
10518
10552
  cacheWrite: 0,
10519
10553
  },
10520
10554
  contextWindow: 400000,
@@ -10522,7 +10556,7 @@ export const MODELS = {
10522
10556
  },
10523
10557
  "openai/gpt-5.2": {
10524
10558
  id: "openai/gpt-5.2",
10525
- name: "GPT-5.2",
10559
+ name: "GPT 5.2",
10526
10560
  api: "anthropic-messages",
10527
10561
  provider: "vercel-ai-gateway",
10528
10562
  baseUrl: "https://ai-gateway.vercel.sh",
@@ -10531,7 +10565,7 @@ export const MODELS = {
10531
10565
  cost: {
10532
10566
  input: 1.75,
10533
10567
  output: 14,
10534
- cacheRead: 0.175,
10568
+ cacheRead: 0.18,
10535
10569
  cacheWrite: 0,
10536
10570
  },
10537
10571
  contextWindow: 400000,
@@ -10792,40 +10826,6 @@ export const MODELS = {
10792
10826
  contextWindow: 131072,
10793
10827
  maxTokens: 131072,
10794
10828
  },
10795
- "stealth/sonoma-dusk-alpha": {
10796
- id: "stealth/sonoma-dusk-alpha",
10797
- name: "Sonoma Dusk Alpha",
10798
- api: "anthropic-messages",
10799
- provider: "vercel-ai-gateway",
10800
- baseUrl: "https://ai-gateway.vercel.sh",
10801
- reasoning: false,
10802
- input: ["text", "image"],
10803
- cost: {
10804
- input: 0.19999999999999998,
10805
- output: 0.5,
10806
- cacheRead: 0.049999999999999996,
10807
- cacheWrite: 0,
10808
- },
10809
- contextWindow: 2000000,
10810
- maxTokens: 131072,
10811
- },
10812
- "stealth/sonoma-sky-alpha": {
10813
- id: "stealth/sonoma-sky-alpha",
10814
- name: "Sonoma Sky Alpha",
10815
- api: "anthropic-messages",
10816
- provider: "vercel-ai-gateway",
10817
- baseUrl: "https://ai-gateway.vercel.sh",
10818
- reasoning: false,
10819
- input: ["text", "image"],
10820
- cost: {
10821
- input: 0.19999999999999998,
10822
- output: 0.5,
10823
- cacheRead: 0.049999999999999996,
10824
- cacheWrite: 0,
10825
- },
10826
- contextWindow: 2000000,
10827
- maxTokens: 131072,
10828
- },
10829
10829
  "vercel/v0-1.0-md": {
10830
10830
  id: "vercel/v0-1.0-md",
10831
10831
  name: "v0-1.0-md",
@@ -11113,7 +11113,7 @@ export const MODELS = {
11113
11113
  cacheWrite: 0,
11114
11114
  },
11115
11115
  contextWindow: 65536,
11116
- maxTokens: 66000,
11116
+ maxTokens: 16384,
11117
11117
  },
11118
11118
  "zai/glm-4.6": {
11119
11119
  id: "zai/glm-4.6",