@mariozechner/pi-ai 0.18.1 → 0.18.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2415,6 +2415,23 @@ export const MODELS = {
2415
2415
  contextWindow: 262144,
2416
2416
  maxTokens: 4096,
2417
2417
  },
2418
+ "mistralai/devstral-2512": {
2419
+ id: "mistralai/devstral-2512",
2420
+ name: "Mistral: Devstral 2 2512",
2421
+ api: "openai-completions",
2422
+ provider: "openrouter",
2423
+ baseUrl: "https://openrouter.ai/api/v1",
2424
+ reasoning: false,
2425
+ input: ["text"],
2426
+ cost: {
2427
+ input: 0.15,
2428
+ output: 0.6,
2429
+ cacheRead: 0,
2430
+ cacheWrite: 0,
2431
+ },
2432
+ contextWindow: 262144,
2433
+ maxTokens: 65536,
2434
+ },
2418
2435
  "relace/relace-search": {
2419
2436
  id: "relace/relace-search",
2420
2437
  name: "Relace: Relace Search",
@@ -2449,6 +2466,23 @@ export const MODELS = {
2449
2466
  contextWindow: 131072,
2450
2467
  maxTokens: 24000,
2451
2468
  },
2469
+ "nex-agi/deepseek-v3.1-nex-n1:free": {
2470
+ id: "nex-agi/deepseek-v3.1-nex-n1:free",
2471
+ name: "Nex AGI: DeepSeek V3.1 Nex N1 (free)",
2472
+ api: "openai-completions",
2473
+ provider: "openrouter",
2474
+ baseUrl: "https://openrouter.ai/api/v1",
2475
+ reasoning: true,
2476
+ input: ["text"],
2477
+ cost: {
2478
+ input: 0,
2479
+ output: 0,
2480
+ cacheRead: 0,
2481
+ cacheWrite: 0,
2482
+ },
2483
+ contextWindow: 131072,
2484
+ maxTokens: 163840,
2485
+ },
2452
2486
  "openai/gpt-5.1-codex-max": {
2453
2487
  id: "openai/gpt-5.1-codex-max",
2454
2488
  name: "OpenAI: GPT-5.1-Codex-Max",
@@ -5475,34 +5509,34 @@ export const MODELS = {
5475
5509
  contextWindow: 200000,
5476
5510
  maxTokens: 8192,
5477
5511
  },
5478
- "mistralai/ministral-3b": {
5479
- id: "mistralai/ministral-3b",
5480
- name: "Mistral: Ministral 3B",
5512
+ "mistralai/ministral-8b": {
5513
+ id: "mistralai/ministral-8b",
5514
+ name: "Mistral: Ministral 8B",
5481
5515
  api: "openai-completions",
5482
5516
  provider: "openrouter",
5483
5517
  baseUrl: "https://openrouter.ai/api/v1",
5484
5518
  reasoning: false,
5485
5519
  input: ["text"],
5486
5520
  cost: {
5487
- input: 0.04,
5488
- output: 0.04,
5521
+ input: 0.09999999999999999,
5522
+ output: 0.09999999999999999,
5489
5523
  cacheRead: 0,
5490
5524
  cacheWrite: 0,
5491
5525
  },
5492
5526
  contextWindow: 131072,
5493
5527
  maxTokens: 4096,
5494
5528
  },
5495
- "mistralai/ministral-8b": {
5496
- id: "mistralai/ministral-8b",
5497
- name: "Mistral: Ministral 8B",
5529
+ "mistralai/ministral-3b": {
5530
+ id: "mistralai/ministral-3b",
5531
+ name: "Mistral: Ministral 3B",
5498
5532
  api: "openai-completions",
5499
5533
  provider: "openrouter",
5500
5534
  baseUrl: "https://openrouter.ai/api/v1",
5501
5535
  reasoning: false,
5502
5536
  input: ["text"],
5503
5537
  cost: {
5504
- input: 0.09999999999999999,
5505
- output: 0.09999999999999999,
5538
+ input: 0.04,
5539
+ output: 0.04,
5506
5540
  cacheRead: 0,
5507
5541
  cacheWrite: 0,
5508
5542
  },
@@ -5679,6 +5713,23 @@ export const MODELS = {
5679
5713
  contextWindow: 128000,
5680
5714
  maxTokens: 16384,
5681
5715
  },
5716
+ "meta-llama/llama-3.1-8b-instruct": {
5717
+ id: "meta-llama/llama-3.1-8b-instruct",
5718
+ name: "Meta: Llama 3.1 8B Instruct",
5719
+ api: "openai-completions",
5720
+ provider: "openrouter",
5721
+ baseUrl: "https://openrouter.ai/api/v1",
5722
+ reasoning: false,
5723
+ input: ["text"],
5724
+ cost: {
5725
+ input: 0.02,
5726
+ output: 0.03,
5727
+ cacheRead: 0,
5728
+ cacheWrite: 0,
5729
+ },
5730
+ contextWindow: 131072,
5731
+ maxTokens: 16384,
5732
+ },
5682
5733
  "meta-llama/llama-3.1-405b-instruct": {
5683
5734
  id: "meta-llama/llama-3.1-405b-instruct",
5684
5735
  name: "Meta: Llama 3.1 405B Instruct",
@@ -5713,23 +5764,6 @@ export const MODELS = {
5713
5764
  contextWindow: 131072,
5714
5765
  maxTokens: 4096,
5715
5766
  },
5716
- "meta-llama/llama-3.1-8b-instruct": {
5717
- id: "meta-llama/llama-3.1-8b-instruct",
5718
- name: "Meta: Llama 3.1 8B Instruct",
5719
- api: "openai-completions",
5720
- provider: "openrouter",
5721
- baseUrl: "https://openrouter.ai/api/v1",
5722
- reasoning: false,
5723
- input: ["text"],
5724
- cost: {
5725
- input: 0.02,
5726
- output: 0.03,
5727
- cacheRead: 0,
5728
- cacheWrite: 0,
5729
- },
5730
- contextWindow: 131072,
5731
- maxTokens: 16384,
5732
- },
5733
5767
  "mistralai/mistral-nemo": {
5734
5768
  id: "mistralai/mistral-nemo",
5735
5769
  name: "Mistral: Mistral Nemo",
@@ -5866,6 +5900,23 @@ export const MODELS = {
5866
5900
  contextWindow: 128000,
5867
5901
  maxTokens: 4096,
5868
5902
  },
5903
+ "openai/gpt-4o-2024-05-13": {
5904
+ id: "openai/gpt-4o-2024-05-13",
5905
+ name: "OpenAI: GPT-4o (2024-05-13)",
5906
+ api: "openai-completions",
5907
+ provider: "openrouter",
5908
+ baseUrl: "https://openrouter.ai/api/v1",
5909
+ reasoning: false,
5910
+ input: ["text", "image"],
5911
+ cost: {
5912
+ input: 5,
5913
+ output: 15,
5914
+ cacheRead: 0,
5915
+ cacheWrite: 0,
5916
+ },
5917
+ contextWindow: 128000,
5918
+ maxTokens: 4096,
5919
+ },
5869
5920
  "openai/gpt-4o": {
5870
5921
  id: "openai/gpt-4o",
5871
5922
  name: "OpenAI: GPT-4o",
@@ -5900,22 +5951,22 @@ export const MODELS = {
5900
5951
  contextWindow: 128000,
5901
5952
  maxTokens: 64000,
5902
5953
  },
5903
- "openai/gpt-4o-2024-05-13": {
5904
- id: "openai/gpt-4o-2024-05-13",
5905
- name: "OpenAI: GPT-4o (2024-05-13)",
5954
+ "meta-llama/llama-3-70b-instruct": {
5955
+ id: "meta-llama/llama-3-70b-instruct",
5956
+ name: "Meta: Llama 3 70B Instruct",
5906
5957
  api: "openai-completions",
5907
5958
  provider: "openrouter",
5908
5959
  baseUrl: "https://openrouter.ai/api/v1",
5909
5960
  reasoning: false,
5910
- input: ["text", "image"],
5961
+ input: ["text"],
5911
5962
  cost: {
5912
- input: 5,
5913
- output: 15,
5963
+ input: 0.3,
5964
+ output: 0.39999999999999997,
5914
5965
  cacheRead: 0,
5915
5966
  cacheWrite: 0,
5916
5967
  },
5917
- contextWindow: 128000,
5918
- maxTokens: 4096,
5968
+ contextWindow: 8192,
5969
+ maxTokens: 16384,
5919
5970
  },
5920
5971
  "meta-llama/llama-3-8b-instruct": {
5921
5972
  id: "meta-llama/llama-3-8b-instruct",
@@ -5934,23 +5985,6 @@ export const MODELS = {
5934
5985
  contextWindow: 8192,
5935
5986
  maxTokens: 16384,
5936
5987
  },
5937
- "meta-llama/llama-3-70b-instruct": {
5938
- id: "meta-llama/llama-3-70b-instruct",
5939
- name: "Meta: Llama 3 70B Instruct",
5940
- api: "openai-completions",
5941
- provider: "openrouter",
5942
- baseUrl: "https://openrouter.ai/api/v1",
5943
- reasoning: false,
5944
- input: ["text"],
5945
- cost: {
5946
- input: 0.3,
5947
- output: 0.39999999999999997,
5948
- cacheRead: 0,
5949
- cacheWrite: 0,
5950
- },
5951
- contextWindow: 8192,
5952
- maxTokens: 16384,
5953
- },
5954
5988
  "mistralai/mixtral-8x22b-instruct": {
5955
5989
  id: "mistralai/mixtral-8x22b-instruct",
5956
5990
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -6036,38 +6070,38 @@ export const MODELS = {
6036
6070
  contextWindow: 128000,
6037
6071
  maxTokens: 4096,
6038
6072
  },
6039
- "openai/gpt-4-turbo-preview": {
6040
- id: "openai/gpt-4-turbo-preview",
6041
- name: "OpenAI: GPT-4 Turbo Preview",
6073
+ "openai/gpt-3.5-turbo-0613": {
6074
+ id: "openai/gpt-3.5-turbo-0613",
6075
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6042
6076
  api: "openai-completions",
6043
6077
  provider: "openrouter",
6044
6078
  baseUrl: "https://openrouter.ai/api/v1",
6045
6079
  reasoning: false,
6046
6080
  input: ["text"],
6047
6081
  cost: {
6048
- input: 10,
6049
- output: 30,
6082
+ input: 1,
6083
+ output: 2,
6050
6084
  cacheRead: 0,
6051
6085
  cacheWrite: 0,
6052
6086
  },
6053
- contextWindow: 128000,
6087
+ contextWindow: 4095,
6054
6088
  maxTokens: 4096,
6055
6089
  },
6056
- "openai/gpt-3.5-turbo-0613": {
6057
- id: "openai/gpt-3.5-turbo-0613",
6058
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6090
+ "openai/gpt-4-turbo-preview": {
6091
+ id: "openai/gpt-4-turbo-preview",
6092
+ name: "OpenAI: GPT-4 Turbo Preview",
6059
6093
  api: "openai-completions",
6060
6094
  provider: "openrouter",
6061
6095
  baseUrl: "https://openrouter.ai/api/v1",
6062
6096
  reasoning: false,
6063
6097
  input: ["text"],
6064
6098
  cost: {
6065
- input: 1,
6066
- output: 2,
6099
+ input: 10,
6100
+ output: 30,
6067
6101
  cacheRead: 0,
6068
6102
  cacheWrite: 0,
6069
6103
  },
6070
- contextWindow: 4095,
6104
+ contextWindow: 128000,
6071
6105
  maxTokens: 4096,
6072
6106
  },
6073
6107
  "mistralai/mistral-tiny": {
@@ -6138,9 +6172,9 @@ export const MODELS = {
6138
6172
  contextWindow: 16385,
6139
6173
  maxTokens: 4096,
6140
6174
  },
6141
- "openai/gpt-4": {
6142
- id: "openai/gpt-4",
6143
- name: "OpenAI: GPT-4",
6175
+ "openai/gpt-4-0314": {
6176
+ id: "openai/gpt-4-0314",
6177
+ name: "OpenAI: GPT-4 (older v0314)",
6144
6178
  api: "openai-completions",
6145
6179
  provider: "openrouter",
6146
6180
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6155,38 +6189,38 @@ export const MODELS = {
6155
6189
  contextWindow: 8191,
6156
6190
  maxTokens: 4096,
6157
6191
  },
6158
- "openai/gpt-3.5-turbo": {
6159
- id: "openai/gpt-3.5-turbo",
6160
- name: "OpenAI: GPT-3.5 Turbo",
6192
+ "openai/gpt-4": {
6193
+ id: "openai/gpt-4",
6194
+ name: "OpenAI: GPT-4",
6161
6195
  api: "openai-completions",
6162
6196
  provider: "openrouter",
6163
6197
  baseUrl: "https://openrouter.ai/api/v1",
6164
6198
  reasoning: false,
6165
6199
  input: ["text"],
6166
6200
  cost: {
6167
- input: 0.5,
6168
- output: 1.5,
6201
+ input: 30,
6202
+ output: 60,
6169
6203
  cacheRead: 0,
6170
6204
  cacheWrite: 0,
6171
6205
  },
6172
- contextWindow: 16385,
6206
+ contextWindow: 8191,
6173
6207
  maxTokens: 4096,
6174
6208
  },
6175
- "openai/gpt-4-0314": {
6176
- id: "openai/gpt-4-0314",
6177
- name: "OpenAI: GPT-4 (older v0314)",
6209
+ "openai/gpt-3.5-turbo": {
6210
+ id: "openai/gpt-3.5-turbo",
6211
+ name: "OpenAI: GPT-3.5 Turbo",
6178
6212
  api: "openai-completions",
6179
6213
  provider: "openrouter",
6180
6214
  baseUrl: "https://openrouter.ai/api/v1",
6181
6215
  reasoning: false,
6182
6216
  input: ["text"],
6183
6217
  cost: {
6184
- input: 30,
6185
- output: 60,
6218
+ input: 0.5,
6219
+ output: 1.5,
6186
6220
  cacheRead: 0,
6187
6221
  cacheWrite: 0,
6188
6222
  },
6189
- contextWindow: 8191,
6223
+ contextWindow: 16385,
6190
6224
  maxTokens: 4096,
6191
6225
  },
6192
6226
  "openrouter/auto": {