@mariozechner/pi-ai 0.12.10 → 0.12.12

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1077,6 +1077,23 @@ export const MODELS = {
1077
1077
  contextWindow: 128000,
1078
1078
  maxTokens: 16384,
1079
1079
  },
1080
+ "gpt-5.1-codex-max": {
1081
+ id: "gpt-5.1-codex-max",
1082
+ name: "GPT-5.1 Codex Max",
1083
+ api: "openai-responses",
1084
+ provider: "openai",
1085
+ baseUrl: "https://api.openai.com/v1",
1086
+ reasoning: true,
1087
+ input: ["text", "image"],
1088
+ cost: {
1089
+ input: 1.25,
1090
+ output: 10,
1091
+ cacheRead: 0.125,
1092
+ cacheWrite: 0,
1093
+ },
1094
+ contextWindow: 400000,
1095
+ maxTokens: 128000,
1096
+ },
1080
1097
  "o3": {
1081
1098
  id: "o3",
1082
1099
  name: "o3",
@@ -1196,23 +1213,6 @@ export const MODELS = {
1196
1213
  contextWindow: 128000,
1197
1214
  maxTokens: 16384,
1198
1215
  },
1199
- "gpt-5.1-codex-max": {
1200
- id: "gpt-5.1-codex-max",
1201
- name: "GPT-5.1 Codex Max",
1202
- api: "openai-responses",
1203
- provider: "openai",
1204
- baseUrl: "https://api.openai.com/v1",
1205
- reasoning: true,
1206
- input: ["text", "image"],
1207
- cost: {
1208
- input: 1.25,
1209
- output: 10,
1210
- cacheRead: 0.125,
1211
- cacheWrite: 0,
1212
- },
1213
- contextWindow: 400000,
1214
- maxTokens: 128000,
1215
- },
1216
1216
  },
1217
1217
  groq: {
1218
1218
  "llama-3.1-8b-instant": {
@@ -1988,6 +1988,23 @@ export const MODELS = {
1988
1988
  },
1989
1989
  },
1990
1990
  openrouter: {
1991
+ "openai/gpt-5.1-codex-max": {
1992
+ id: "openai/gpt-5.1-codex-max",
1993
+ name: "OpenAI: GPT-5.1-Codex-Max",
1994
+ api: "openai-completions",
1995
+ provider: "openrouter",
1996
+ baseUrl: "https://openrouter.ai/api/v1",
1997
+ reasoning: true,
1998
+ input: ["text", "image"],
1999
+ cost: {
2000
+ input: 1.25,
2001
+ output: 10,
2002
+ cacheRead: 0.125,
2003
+ cacheWrite: 0,
2004
+ },
2005
+ contextWindow: 400000,
2006
+ maxTokens: 128000,
2007
+ },
1991
2008
  "amazon/nova-2-lite-v1:free": {
1992
2009
  id: "amazon/nova-2-lite-v1:free",
1993
2010
  name: "Amazon: Nova 2 Lite (free)",
@@ -2022,6 +2039,57 @@ export const MODELS = {
2022
2039
  contextWindow: 1000000,
2023
2040
  maxTokens: 65535,
2024
2041
  },
2042
+ "mistralai/ministral-14b-2512": {
2043
+ id: "mistralai/ministral-14b-2512",
2044
+ name: "Mistral: Ministral 3 14B 2512",
2045
+ api: "openai-completions",
2046
+ provider: "openrouter",
2047
+ baseUrl: "https://openrouter.ai/api/v1",
2048
+ reasoning: false,
2049
+ input: ["text", "image"],
2050
+ cost: {
2051
+ input: 0.19999999999999998,
2052
+ output: 0.19999999999999998,
2053
+ cacheRead: 0,
2054
+ cacheWrite: 0,
2055
+ },
2056
+ contextWindow: 262144,
2057
+ maxTokens: 4096,
2058
+ },
2059
+ "mistralai/ministral-8b-2512": {
2060
+ id: "mistralai/ministral-8b-2512",
2061
+ name: "Mistral: Ministral 3 8B 2512",
2062
+ api: "openai-completions",
2063
+ provider: "openrouter",
2064
+ baseUrl: "https://openrouter.ai/api/v1",
2065
+ reasoning: false,
2066
+ input: ["text", "image"],
2067
+ cost: {
2068
+ input: 0.15,
2069
+ output: 0.15,
2070
+ cacheRead: 0,
2071
+ cacheWrite: 0,
2072
+ },
2073
+ contextWindow: 262144,
2074
+ maxTokens: 4096,
2075
+ },
2076
+ "mistralai/ministral-3b-2512": {
2077
+ id: "mistralai/ministral-3b-2512",
2078
+ name: "Mistral: Ministral 3 3B 2512",
2079
+ api: "openai-completions",
2080
+ provider: "openrouter",
2081
+ baseUrl: "https://openrouter.ai/api/v1",
2082
+ reasoning: false,
2083
+ input: ["text", "image"],
2084
+ cost: {
2085
+ input: 0.09999999999999999,
2086
+ output: 0.09999999999999999,
2087
+ cacheRead: 0,
2088
+ cacheWrite: 0,
2089
+ },
2090
+ contextWindow: 131072,
2091
+ maxTokens: 4096,
2092
+ },
2025
2093
  "mistralai/mistral-large-2512": {
2026
2094
  id: "mistralai/mistral-large-2512",
2027
2095
  name: "Mistral: Mistral Large 3 2512",
@@ -3348,22 +3416,22 @@ export const MODELS = {
3348
3416
  contextWindow: 400000,
3349
3417
  maxTokens: 128000,
3350
3418
  },
3351
- "openai/gpt-oss-120b:exacto": {
3352
- id: "openai/gpt-oss-120b:exacto",
3353
- name: "OpenAI: gpt-oss-120b (exacto)",
3419
+ "openai/gpt-oss-120b:free": {
3420
+ id: "openai/gpt-oss-120b:free",
3421
+ name: "OpenAI: gpt-oss-120b (free)",
3354
3422
  api: "openai-completions",
3355
3423
  provider: "openrouter",
3356
3424
  baseUrl: "https://openrouter.ai/api/v1",
3357
3425
  reasoning: true,
3358
3426
  input: ["text"],
3359
3427
  cost: {
3360
- input: 0.04,
3361
- output: 0.19999999999999998,
3428
+ input: 0,
3429
+ output: 0,
3362
3430
  cacheRead: 0,
3363
3431
  cacheWrite: 0,
3364
3432
  },
3365
3433
  contextWindow: 131072,
3366
- maxTokens: 32768,
3434
+ maxTokens: 4096,
3367
3435
  },
3368
3436
  "openai/gpt-oss-120b": {
3369
3437
  id: "openai/gpt-oss-120b",
@@ -3374,13 +3442,30 @@ export const MODELS = {
3374
3442
  reasoning: true,
3375
3443
  input: ["text"],
3376
3444
  cost: {
3377
- input: 0.04,
3378
- output: 0.19999999999999998,
3445
+ input: 0.039,
3446
+ output: 0.19,
3379
3447
  cacheRead: 0,
3380
3448
  cacheWrite: 0,
3381
3449
  },
3382
3450
  contextWindow: 131072,
3383
- maxTokens: 32768,
3451
+ maxTokens: 4096,
3452
+ },
3453
+ "openai/gpt-oss-120b:exacto": {
3454
+ id: "openai/gpt-oss-120b:exacto",
3455
+ name: "OpenAI: gpt-oss-120b (exacto)",
3456
+ api: "openai-completions",
3457
+ provider: "openrouter",
3458
+ baseUrl: "https://openrouter.ai/api/v1",
3459
+ reasoning: true,
3460
+ input: ["text"],
3461
+ cost: {
3462
+ input: 0.039,
3463
+ output: 0.19,
3464
+ cacheRead: 0,
3465
+ cacheWrite: 0,
3466
+ },
3467
+ contextWindow: 131072,
3468
+ maxTokens: 4096,
3384
3469
  },
3385
3470
  "openai/gpt-oss-20b:free": {
3386
3471
  id: "openai/gpt-oss-20b:free",
@@ -4632,13 +4717,13 @@ export const MODELS = {
4632
4717
  reasoning: false,
4633
4718
  input: ["text"],
4634
4719
  cost: {
4635
- input: 0.049999999999999996,
4636
- output: 0.08,
4720
+ input: 0.03,
4721
+ output: 0.11,
4637
4722
  cacheRead: 0,
4638
4723
  cacheWrite: 0,
4639
4724
  },
4640
4725
  contextWindow: 32768,
4641
- maxTokens: 16384,
4726
+ maxTokens: 32768,
4642
4727
  },
4643
4728
  "deepseek/deepseek-r1-distill-llama-70b": {
4644
4729
  id: "deepseek/deepseek-r1-distill-llama-70b",