@mariozechner/pi-ai 0.15.0 → 0.17.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1988,6 +1988,23 @@ export const MODELS = {
1988
1988
  },
1989
1989
  },
1990
1990
  openrouter: {
1991
+ "mistralai/devstral-2512:free": {
1992
+ id: "mistralai/devstral-2512:free",
1993
+ name: "Mistral: Devstral 2 2512 (free)",
1994
+ api: "openai-completions",
1995
+ provider: "openrouter",
1996
+ baseUrl: "https://openrouter.ai/api/v1",
1997
+ reasoning: false,
1998
+ input: ["text"],
1999
+ cost: {
2000
+ input: 0,
2001
+ output: 0,
2002
+ cacheRead: 0,
2003
+ cacheWrite: 0,
2004
+ },
2005
+ contextWindow: 262144,
2006
+ maxTokens: 4096,
2007
+ },
1991
2008
  "relace/relace-search": {
1992
2009
  id: "relace/relace-search",
1993
2010
  name: "Relace: Relace Search",
@@ -2173,7 +2190,7 @@ export const MODELS = {
2173
2190
  cacheWrite: 0,
2174
2191
  },
2175
2192
  contextWindow: 131072,
2176
- maxTokens: 4096,
2193
+ maxTokens: 131072,
2177
2194
  },
2178
2195
  "deepseek/deepseek-v3.2": {
2179
2196
  id: "deepseek/deepseek-v3.2",
@@ -2490,13 +2507,13 @@ export const MODELS = {
2490
2507
  reasoning: true,
2491
2508
  input: ["text"],
2492
2509
  cost: {
2493
- input: 0.255,
2510
+ input: 0.254,
2494
2511
  output: 1.02,
2495
- cacheRead: 0,
2512
+ cacheRead: 0.127,
2496
2513
  cacheWrite: 0,
2497
2514
  },
2498
- contextWindow: 204800,
2499
- maxTokens: 131072,
2515
+ contextWindow: 262144,
2516
+ maxTokens: 4096,
2500
2517
  },
2501
2518
  "deepcogito/cogito-v2-preview-llama-405b": {
2502
2519
  id: "deepcogito/cogito-v2-preview-llama-405b",
@@ -3272,13 +3289,13 @@ export const MODELS = {
3272
3289
  reasoning: true,
3273
3290
  input: ["text"],
3274
3291
  cost: {
3275
- input: 0.19999999999999998,
3276
- output: 0.7999999999999999,
3292
+ input: 0.15,
3293
+ output: 0.75,
3277
3294
  cacheRead: 0,
3278
3295
  cacheWrite: 0,
3279
3296
  },
3280
- contextWindow: 163840,
3281
- maxTokens: 163840,
3297
+ contextWindow: 8192,
3298
+ maxTokens: 7168,
3282
3299
  },
3283
3300
  "openai/gpt-4o-audio-preview": {
3284
3301
  id: "openai/gpt-4o-audio-preview",
@@ -4462,13 +4479,13 @@ export const MODELS = {
4462
4479
  reasoning: true,
4463
4480
  input: ["text"],
4464
4481
  cost: {
4465
- input: 0.15,
4466
- output: 0.7,
4467
- cacheRead: 0,
4482
+ input: 0.19999999999999998,
4483
+ output: 0.88,
4484
+ cacheRead: 0.106,
4468
4485
  cacheWrite: 0,
4469
4486
  },
4470
- contextWindow: 8192,
4471
- maxTokens: 7168,
4487
+ contextWindow: 163840,
4488
+ maxTokens: 4096,
4472
4489
  },
4473
4490
  "mistralai/mistral-small-3.1-24b-instruct:free": {
4474
4491
  id: "mistralai/mistral-small-3.1-24b-instruct:free",
@@ -4513,8 +4530,8 @@ export const MODELS = {
4513
4530
  reasoning: false,
4514
4531
  input: ["text", "image"],
4515
4532
  cost: {
4516
- input: 0.049999999999999996,
4517
- output: 0.22,
4533
+ input: 0.04,
4534
+ output: 0.15,
4518
4535
  cacheRead: 0,
4519
4536
  cacheWrite: 0,
4520
4537
  },