@mariozechner/pi-ai 0.16.0 → 0.17.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1988,6 +1988,23 @@ export const MODELS = {
1988
1988
  },
1989
1989
  },
1990
1990
  openrouter: {
1991
+ "mistralai/devstral-2512:free": {
1992
+ id: "mistralai/devstral-2512:free",
1993
+ name: "Mistral: Devstral 2 2512 (free)",
1994
+ api: "openai-completions",
1995
+ provider: "openrouter",
1996
+ baseUrl: "https://openrouter.ai/api/v1",
1997
+ reasoning: false,
1998
+ input: ["text"],
1999
+ cost: {
2000
+ input: 0,
2001
+ output: 0,
2002
+ cacheRead: 0,
2003
+ cacheWrite: 0,
2004
+ },
2005
+ contextWindow: 262144,
2006
+ maxTokens: 4096,
2007
+ },
1991
2008
  "relace/relace-search": {
1992
2009
  id: "relace/relace-search",
1993
2010
  name: "Relace: Relace Search",
@@ -2173,7 +2190,7 @@ export const MODELS = {
2173
2190
  cacheWrite: 0,
2174
2191
  },
2175
2192
  contextWindow: 131072,
2176
- maxTokens: 4096,
2193
+ maxTokens: 131072,
2177
2194
  },
2178
2195
  "deepseek/deepseek-v3.2": {
2179
2196
  id: "deepseek/deepseek-v3.2",
@@ -2490,13 +2507,13 @@ export const MODELS = {
2490
2507
  reasoning: true,
2491
2508
  input: ["text"],
2492
2509
  cost: {
2493
- input: 0.255,
2510
+ input: 0.254,
2494
2511
  output: 1.02,
2495
- cacheRead: 0,
2512
+ cacheRead: 0.127,
2496
2513
  cacheWrite: 0,
2497
2514
  },
2498
- contextWindow: 204800,
2499
- maxTokens: 131072,
2515
+ contextWindow: 262144,
2516
+ maxTokens: 4096,
2500
2517
  },
2501
2518
  "deepcogito/cogito-v2-preview-llama-405b": {
2502
2519
  id: "deepcogito/cogito-v2-preview-llama-405b",
@@ -3272,13 +3289,13 @@ export const MODELS = {
3272
3289
  reasoning: true,
3273
3290
  input: ["text"],
3274
3291
  cost: {
3275
- input: 0.19999999999999998,
3276
- output: 0.7999999999999999,
3292
+ input: 0.15,
3293
+ output: 0.75,
3277
3294
  cacheRead: 0,
3278
3295
  cacheWrite: 0,
3279
3296
  },
3280
- contextWindow: 163840,
3281
- maxTokens: 163840,
3297
+ contextWindow: 8192,
3298
+ maxTokens: 7168,
3282
3299
  },
3283
3300
  "openai/gpt-4o-audio-preview": {
3284
3301
  id: "openai/gpt-4o-audio-preview",
@@ -4513,8 +4530,8 @@ export const MODELS = {
4513
4530
  reasoning: false,
4514
4531
  input: ["text", "image"],
4515
4532
  cost: {
4516
- input: 0.049999999999999996,
4517
- output: 0.22,
4533
+ input: 0.04,
4534
+ output: 0.15,
4518
4535
  cacheRead: 0,
4519
4536
  cacheWrite: 0,
4520
4537
  },
@@ -5048,34 +5065,34 @@ export const MODELS = {
5048
5065
  contextWindow: 200000,
5049
5066
  maxTokens: 8192,
5050
5067
  },
5051
- "mistralai/ministral-3b": {
5052
- id: "mistralai/ministral-3b",
5053
- name: "Mistral: Ministral 3B",
5068
+ "mistralai/ministral-8b": {
5069
+ id: "mistralai/ministral-8b",
5070
+ name: "Mistral: Ministral 8B",
5054
5071
  api: "openai-completions",
5055
5072
  provider: "openrouter",
5056
5073
  baseUrl: "https://openrouter.ai/api/v1",
5057
5074
  reasoning: false,
5058
5075
  input: ["text"],
5059
5076
  cost: {
5060
- input: 0.04,
5061
- output: 0.04,
5077
+ input: 0.09999999999999999,
5078
+ output: 0.09999999999999999,
5062
5079
  cacheRead: 0,
5063
5080
  cacheWrite: 0,
5064
5081
  },
5065
5082
  contextWindow: 131072,
5066
5083
  maxTokens: 4096,
5067
5084
  },
5068
- "mistralai/ministral-8b": {
5069
- id: "mistralai/ministral-8b",
5070
- name: "Mistral: Ministral 8B",
5085
+ "mistralai/ministral-3b": {
5086
+ id: "mistralai/ministral-3b",
5087
+ name: "Mistral: Ministral 3B",
5071
5088
  api: "openai-completions",
5072
5089
  provider: "openrouter",
5073
5090
  baseUrl: "https://openrouter.ai/api/v1",
5074
5091
  reasoning: false,
5075
5092
  input: ["text"],
5076
5093
  cost: {
5077
- input: 0.09999999999999999,
5078
- output: 0.09999999999999999,
5094
+ input: 0.04,
5095
+ output: 0.04,
5079
5096
  cacheRead: 0,
5080
5097
  cacheWrite: 0,
5081
5098
  },
@@ -5269,38 +5286,38 @@ export const MODELS = {
5269
5286
  contextWindow: 131072,
5270
5287
  maxTokens: 16384,
5271
5288
  },
5272
- "meta-llama/llama-3.1-70b-instruct": {
5273
- id: "meta-llama/llama-3.1-70b-instruct",
5274
- name: "Meta: Llama 3.1 70B Instruct",
5289
+ "meta-llama/llama-3.1-405b-instruct": {
5290
+ id: "meta-llama/llama-3.1-405b-instruct",
5291
+ name: "Meta: Llama 3.1 405B Instruct",
5275
5292
  api: "openai-completions",
5276
5293
  provider: "openrouter",
5277
5294
  baseUrl: "https://openrouter.ai/api/v1",
5278
5295
  reasoning: false,
5279
5296
  input: ["text"],
5280
5297
  cost: {
5281
- input: 0.39999999999999997,
5282
- output: 0.39999999999999997,
5298
+ input: 3.5,
5299
+ output: 3.5,
5283
5300
  cacheRead: 0,
5284
5301
  cacheWrite: 0,
5285
5302
  },
5286
- contextWindow: 131072,
5303
+ contextWindow: 130815,
5287
5304
  maxTokens: 4096,
5288
5305
  },
5289
- "meta-llama/llama-3.1-405b-instruct": {
5290
- id: "meta-llama/llama-3.1-405b-instruct",
5291
- name: "Meta: Llama 3.1 405B Instruct",
5306
+ "meta-llama/llama-3.1-70b-instruct": {
5307
+ id: "meta-llama/llama-3.1-70b-instruct",
5308
+ name: "Meta: Llama 3.1 70B Instruct",
5292
5309
  api: "openai-completions",
5293
5310
  provider: "openrouter",
5294
5311
  baseUrl: "https://openrouter.ai/api/v1",
5295
5312
  reasoning: false,
5296
5313
  input: ["text"],
5297
5314
  cost: {
5298
- input: 3.5,
5299
- output: 3.5,
5315
+ input: 0.39999999999999997,
5316
+ output: 0.39999999999999997,
5300
5317
  cacheRead: 0,
5301
5318
  cacheWrite: 0,
5302
5319
  },
5303
- contextWindow: 130815,
5320
+ contextWindow: 131072,
5304
5321
  maxTokens: 4096,
5305
5322
  },
5306
5323
  "mistralai/mistral-nemo": {
@@ -5320,9 +5337,9 @@ export const MODELS = {
5320
5337
  contextWindow: 131072,
5321
5338
  maxTokens: 16384,
5322
5339
  },
5323
- "openai/gpt-4o-mini": {
5324
- id: "openai/gpt-4o-mini",
5325
- name: "OpenAI: GPT-4o-mini",
5340
+ "openai/gpt-4o-mini-2024-07-18": {
5341
+ id: "openai/gpt-4o-mini-2024-07-18",
5342
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5326
5343
  api: "openai-completions",
5327
5344
  provider: "openrouter",
5328
5345
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5337,9 +5354,9 @@ export const MODELS = {
5337
5354
  contextWindow: 128000,
5338
5355
  maxTokens: 16384,
5339
5356
  },
5340
- "openai/gpt-4o-mini-2024-07-18": {
5341
- id: "openai/gpt-4o-mini-2024-07-18",
5342
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5357
+ "openai/gpt-4o-mini": {
5358
+ id: "openai/gpt-4o-mini",
5359
+ name: "OpenAI: GPT-4o-mini",
5343
5360
  api: "openai-completions",
5344
5361
  provider: "openrouter",
5345
5362
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5439,6 +5456,23 @@ export const MODELS = {
5439
5456
  contextWindow: 128000,
5440
5457
  maxTokens: 4096,
5441
5458
  },
5459
+ "openai/gpt-4o-2024-05-13": {
5460
+ id: "openai/gpt-4o-2024-05-13",
5461
+ name: "OpenAI: GPT-4o (2024-05-13)",
5462
+ api: "openai-completions",
5463
+ provider: "openrouter",
5464
+ baseUrl: "https://openrouter.ai/api/v1",
5465
+ reasoning: false,
5466
+ input: ["text", "image"],
5467
+ cost: {
5468
+ input: 5,
5469
+ output: 15,
5470
+ cacheRead: 0,
5471
+ cacheWrite: 0,
5472
+ },
5473
+ contextWindow: 128000,
5474
+ maxTokens: 4096,
5475
+ },
5442
5476
  "openai/gpt-4o": {
5443
5477
  id: "openai/gpt-4o",
5444
5478
  name: "OpenAI: GPT-4o",
@@ -5473,23 +5507,6 @@ export const MODELS = {
5473
5507
  contextWindow: 128000,
5474
5508
  maxTokens: 64000,
5475
5509
  },
5476
- "openai/gpt-4o-2024-05-13": {
5477
- id: "openai/gpt-4o-2024-05-13",
5478
- name: "OpenAI: GPT-4o (2024-05-13)",
5479
- api: "openai-completions",
5480
- provider: "openrouter",
5481
- baseUrl: "https://openrouter.ai/api/v1",
5482
- reasoning: false,
5483
- input: ["text", "image"],
5484
- cost: {
5485
- input: 5,
5486
- output: 15,
5487
- cacheRead: 0,
5488
- cacheWrite: 0,
5489
- },
5490
- contextWindow: 128000,
5491
- maxTokens: 4096,
5492
- },
5493
5510
  "meta-llama/llama-3-70b-instruct": {
5494
5511
  id: "meta-llama/llama-3-70b-instruct",
5495
5512
  name: "Meta: Llama 3 70B Instruct",
@@ -5609,38 +5626,38 @@ export const MODELS = {
5609
5626
  contextWindow: 128000,
5610
5627
  maxTokens: 4096,
5611
5628
  },
5612
- "openai/gpt-4-turbo-preview": {
5613
- id: "openai/gpt-4-turbo-preview",
5614
- name: "OpenAI: GPT-4 Turbo Preview",
5629
+ "openai/gpt-3.5-turbo-0613": {
5630
+ id: "openai/gpt-3.5-turbo-0613",
5631
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5615
5632
  api: "openai-completions",
5616
5633
  provider: "openrouter",
5617
5634
  baseUrl: "https://openrouter.ai/api/v1",
5618
5635
  reasoning: false,
5619
5636
  input: ["text"],
5620
5637
  cost: {
5621
- input: 10,
5622
- output: 30,
5638
+ input: 1,
5639
+ output: 2,
5623
5640
  cacheRead: 0,
5624
5641
  cacheWrite: 0,
5625
5642
  },
5626
- contextWindow: 128000,
5643
+ contextWindow: 4095,
5627
5644
  maxTokens: 4096,
5628
5645
  },
5629
- "openai/gpt-3.5-turbo-0613": {
5630
- id: "openai/gpt-3.5-turbo-0613",
5631
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5646
+ "openai/gpt-4-turbo-preview": {
5647
+ id: "openai/gpt-4-turbo-preview",
5648
+ name: "OpenAI: GPT-4 Turbo Preview",
5632
5649
  api: "openai-completions",
5633
5650
  provider: "openrouter",
5634
5651
  baseUrl: "https://openrouter.ai/api/v1",
5635
5652
  reasoning: false,
5636
5653
  input: ["text"],
5637
5654
  cost: {
5638
- input: 1,
5639
- output: 2,
5655
+ input: 10,
5656
+ output: 30,
5640
5657
  cacheRead: 0,
5641
5658
  cacheWrite: 0,
5642
5659
  },
5643
- contextWindow: 4095,
5660
+ contextWindow: 128000,
5644
5661
  maxTokens: 4096,
5645
5662
  },
5646
5663
  "mistralai/mistral-tiny": {