@mariozechner/pi-ai 0.55.2 → 0.55.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -6066,40 +6066,6 @@ export const MODELS = {
6066
6066
  contextWindow: 400000,
6067
6067
  maxTokens: 128000,
6068
6068
  },
6069
- "kimi-k2": {
6070
- id: "kimi-k2",
6071
- name: "Kimi K2",
6072
- api: "openai-completions",
6073
- provider: "opencode",
6074
- baseUrl: "https://opencode.ai/zen/v1",
6075
- reasoning: false,
6076
- input: ["text"],
6077
- cost: {
6078
- input: 0.4,
6079
- output: 2.5,
6080
- cacheRead: 0.4,
6081
- cacheWrite: 0,
6082
- },
6083
- contextWindow: 262144,
6084
- maxTokens: 262144,
6085
- },
6086
- "kimi-k2-thinking": {
6087
- id: "kimi-k2-thinking",
6088
- name: "Kimi K2 Thinking",
6089
- api: "openai-completions",
6090
- provider: "opencode",
6091
- baseUrl: "https://opencode.ai/zen/v1",
6092
- reasoning: true,
6093
- input: ["text"],
6094
- cost: {
6095
- input: 0.4,
6096
- output: 2.5,
6097
- cacheRead: 0.4,
6098
- cacheWrite: 0,
6099
- },
6100
- contextWindow: 262144,
6101
- maxTokens: 262144,
6102
- },
6103
6069
  "kimi-k2.5": {
6104
6070
  id: "kimi-k2.5",
6105
6071
  name: "Kimi K2.5",
@@ -6561,23 +6527,6 @@ export const MODELS = {
6561
6527
  contextWindow: 131000,
6562
6528
  maxTokens: 4096,
6563
6529
  },
6564
- "arcee-ai/trinity-mini": {
6565
- id: "arcee-ai/trinity-mini",
6566
- name: "Arcee AI: Trinity Mini",
6567
- api: "openai-completions",
6568
- provider: "openrouter",
6569
- baseUrl: "https://openrouter.ai/api/v1",
6570
- reasoning: true,
6571
- input: ["text"],
6572
- cost: {
6573
- input: 0.045,
6574
- output: 0.15,
6575
- cacheRead: 0,
6576
- cacheWrite: 0,
6577
- },
6578
- contextWindow: 131072,
6579
- maxTokens: 131072,
6580
- },
6581
6530
  "arcee-ai/trinity-mini:free": {
6582
6531
  id: "arcee-ai/trinity-mini:free",
6583
6532
  name: "Arcee AI: Trinity Mini (free)",
@@ -6774,13 +6723,13 @@ export const MODELS = {
6774
6723
  reasoning: true,
6775
6724
  input: ["text"],
6776
6725
  cost: {
6777
- input: 0.19,
6778
- output: 0.87,
6779
- cacheRead: 0.095,
6726
+ input: 0.19999999999999998,
6727
+ output: 0.77,
6728
+ cacheRead: 0.135,
6780
6729
  cacheWrite: 0,
6781
6730
  },
6782
6731
  contextWindow: 163840,
6783
- maxTokens: 65536,
6732
+ maxTokens: 4096,
6784
6733
  },
6785
6734
  "deepseek/deepseek-chat-v3.1": {
6786
6735
  id: "deepseek/deepseek-chat-v3.1",
@@ -6825,9 +6774,9 @@ export const MODELS = {
6825
6774
  reasoning: true,
6826
6775
  input: ["text"],
6827
6776
  cost: {
6828
- input: 0.39999999999999997,
6829
- output: 1.75,
6830
- cacheRead: 0.19999999999999998,
6777
+ input: 0.44999999999999996,
6778
+ output: 2.1500000000000004,
6779
+ cacheRead: 0.22499999999999998,
6831
6780
  cacheWrite: 0,
6832
6781
  },
6833
6782
  contextWindow: 163840,
@@ -6882,7 +6831,7 @@ export const MODELS = {
6882
6831
  cacheWrite: 0,
6883
6832
  },
6884
6833
  contextWindow: 163840,
6885
- maxTokens: 163840,
6834
+ maxTokens: 65536,
6886
6835
  },
6887
6836
  "deepseek/deepseek-v3.2-exp": {
6888
6837
  id: "deepseek/deepseek-v3.2-exp",
@@ -6901,6 +6850,23 @@ export const MODELS = {
6901
6850
  contextWindow: 163840,
6902
6851
  maxTokens: 65536,
6903
6852
  },
6853
+ "essentialai/rnj-1-instruct": {
6854
+ id: "essentialai/rnj-1-instruct",
6855
+ name: "EssentialAI: Rnj 1 Instruct",
6856
+ api: "openai-completions",
6857
+ provider: "openrouter",
6858
+ baseUrl: "https://openrouter.ai/api/v1",
6859
+ reasoning: false,
6860
+ input: ["text"],
6861
+ cost: {
6862
+ input: 0.15,
6863
+ output: 0.15,
6864
+ cacheRead: 0,
6865
+ cacheWrite: 0,
6866
+ },
6867
+ contextWindow: 32768,
6868
+ maxTokens: 4096,
6869
+ },
6904
6870
  "google/gemini-2.0-flash-001": {
6905
6871
  id: "google/gemini-2.0-flash-001",
6906
6872
  name: "Google: Gemini 2.0 Flash",
@@ -6984,7 +6950,7 @@ export const MODELS = {
6984
6950
  cacheWrite: 0.08333333333333334,
6985
6951
  },
6986
6952
  contextWindow: 1048576,
6987
- maxTokens: 65535,
6953
+ maxTokens: 65536,
6988
6954
  },
6989
6955
  "google/gemini-2.5-pro": {
6990
6956
  id: "google/gemini-2.5-pro",
@@ -7052,7 +7018,7 @@ export const MODELS = {
7052
7018
  cacheWrite: 0.08333333333333334,
7053
7019
  },
7054
7020
  contextWindow: 1048576,
7055
- maxTokens: 65535,
7021
+ maxTokens: 65536,
7056
7022
  },
7057
7023
  "google/gemini-3-pro-preview": {
7058
7024
  id: "google/gemini-3-pro-preview",
@@ -7205,7 +7171,7 @@ export const MODELS = {
7205
7171
  cacheWrite: 0,
7206
7172
  },
7207
7173
  contextWindow: 131072,
7208
- maxTokens: 32768,
7174
+ maxTokens: 131072,
7209
7175
  },
7210
7176
  "meta-llama/llama-3-8b-instruct": {
7211
7177
  id: "meta-llama/llama-3-8b-instruct",
@@ -7375,7 +7341,7 @@ export const MODELS = {
7375
7341
  cacheWrite: 0,
7376
7342
  },
7377
7343
  contextWindow: 196608,
7378
- maxTokens: 65536,
7344
+ maxTokens: 196608,
7379
7345
  },
7380
7346
  "minimax/minimax-m2.1": {
7381
7347
  id: "minimax/minimax-m2.1",
@@ -9614,12 +9580,12 @@ export const MODELS = {
9614
9580
  input: ["text", "image"],
9615
9581
  cost: {
9616
9582
  input: 0.25,
9617
- output: 2,
9583
+ output: 1,
9618
9584
  cacheRead: 0,
9619
9585
  cacheWrite: 0,
9620
9586
  },
9621
9587
  contextWindow: 262144,
9622
- maxTokens: 65536,
9588
+ maxTokens: 262144,
9623
9589
  },
9624
9590
  "qwen/qwen3.5-397b-a17b": {
9625
9591
  id: "qwen/qwen3.5-397b-a17b",
@@ -10174,13 +10140,13 @@ export const MODELS = {
10174
10140
  reasoning: true,
10175
10141
  input: ["text"],
10176
10142
  cost: {
10177
- input: 0.95,
10178
- output: 2.5500000000000003,
10179
- cacheRead: 0.19999999999999998,
10143
+ input: 0.7999999999999999,
10144
+ output: 2.56,
10145
+ cacheRead: 0.16,
10180
10146
  cacheWrite: 0,
10181
10147
  },
10182
- contextWindow: 204800,
10183
- maxTokens: 131072,
10148
+ contextWindow: 202752,
10149
+ maxTokens: 4096,
10184
10150
  },
10185
10151
  },
10186
10152
  "vercel-ai-gateway": {