@mariozechner/pi-ai 0.12.10 → 0.12.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -1077,6 +1077,23 @@ export const MODELS = {
|
|
|
1077
1077
|
contextWindow: 128000,
|
|
1078
1078
|
maxTokens: 16384,
|
|
1079
1079
|
},
|
|
1080
|
+
"gpt-5.1-codex-max": {
|
|
1081
|
+
id: "gpt-5.1-codex-max",
|
|
1082
|
+
name: "GPT-5.1 Codex Max",
|
|
1083
|
+
api: "openai-responses",
|
|
1084
|
+
provider: "openai",
|
|
1085
|
+
baseUrl: "https://api.openai.com/v1",
|
|
1086
|
+
reasoning: true,
|
|
1087
|
+
input: ["text", "image"],
|
|
1088
|
+
cost: {
|
|
1089
|
+
input: 1.25,
|
|
1090
|
+
output: 10,
|
|
1091
|
+
cacheRead: 0.125,
|
|
1092
|
+
cacheWrite: 0,
|
|
1093
|
+
},
|
|
1094
|
+
contextWindow: 400000,
|
|
1095
|
+
maxTokens: 128000,
|
|
1096
|
+
},
|
|
1080
1097
|
"o3": {
|
|
1081
1098
|
id: "o3",
|
|
1082
1099
|
name: "o3",
|
|
@@ -1196,23 +1213,6 @@ export const MODELS = {
|
|
|
1196
1213
|
contextWindow: 128000,
|
|
1197
1214
|
maxTokens: 16384,
|
|
1198
1215
|
},
|
|
1199
|
-
"gpt-5.1-codex-max": {
|
|
1200
|
-
id: "gpt-5.1-codex-max",
|
|
1201
|
-
name: "GPT-5.1 Codex Max",
|
|
1202
|
-
api: "openai-responses",
|
|
1203
|
-
provider: "openai",
|
|
1204
|
-
baseUrl: "https://api.openai.com/v1",
|
|
1205
|
-
reasoning: true,
|
|
1206
|
-
input: ["text", "image"],
|
|
1207
|
-
cost: {
|
|
1208
|
-
input: 1.25,
|
|
1209
|
-
output: 10,
|
|
1210
|
-
cacheRead: 0.125,
|
|
1211
|
-
cacheWrite: 0,
|
|
1212
|
-
},
|
|
1213
|
-
contextWindow: 400000,
|
|
1214
|
-
maxTokens: 128000,
|
|
1215
|
-
},
|
|
1216
1216
|
},
|
|
1217
1217
|
groq: {
|
|
1218
1218
|
"llama-3.1-8b-instant": {
|
|
@@ -1988,6 +1988,23 @@ export const MODELS = {
|
|
|
1988
1988
|
},
|
|
1989
1989
|
},
|
|
1990
1990
|
openrouter: {
|
|
1991
|
+
"openai/gpt-5.1-codex-max": {
|
|
1992
|
+
id: "openai/gpt-5.1-codex-max",
|
|
1993
|
+
name: "OpenAI: GPT-5.1-Codex-Max",
|
|
1994
|
+
api: "openai-completions",
|
|
1995
|
+
provider: "openrouter",
|
|
1996
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1997
|
+
reasoning: true,
|
|
1998
|
+
input: ["text", "image"],
|
|
1999
|
+
cost: {
|
|
2000
|
+
input: 1.25,
|
|
2001
|
+
output: 10,
|
|
2002
|
+
cacheRead: 0.125,
|
|
2003
|
+
cacheWrite: 0,
|
|
2004
|
+
},
|
|
2005
|
+
contextWindow: 400000,
|
|
2006
|
+
maxTokens: 128000,
|
|
2007
|
+
},
|
|
1991
2008
|
"amazon/nova-2-lite-v1:free": {
|
|
1992
2009
|
id: "amazon/nova-2-lite-v1:free",
|
|
1993
2010
|
name: "Amazon: Nova 2 Lite (free)",
|
|
@@ -2022,6 +2039,57 @@ export const MODELS = {
|
|
|
2022
2039
|
contextWindow: 1000000,
|
|
2023
2040
|
maxTokens: 65535,
|
|
2024
2041
|
},
|
|
2042
|
+
"mistralai/ministral-14b-2512": {
|
|
2043
|
+
id: "mistralai/ministral-14b-2512",
|
|
2044
|
+
name: "Mistral: Ministral 3 14B 2512",
|
|
2045
|
+
api: "openai-completions",
|
|
2046
|
+
provider: "openrouter",
|
|
2047
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2048
|
+
reasoning: false,
|
|
2049
|
+
input: ["text", "image"],
|
|
2050
|
+
cost: {
|
|
2051
|
+
input: 0.19999999999999998,
|
|
2052
|
+
output: 0.19999999999999998,
|
|
2053
|
+
cacheRead: 0,
|
|
2054
|
+
cacheWrite: 0,
|
|
2055
|
+
},
|
|
2056
|
+
contextWindow: 262144,
|
|
2057
|
+
maxTokens: 4096,
|
|
2058
|
+
},
|
|
2059
|
+
"mistralai/ministral-8b-2512": {
|
|
2060
|
+
id: "mistralai/ministral-8b-2512",
|
|
2061
|
+
name: "Mistral: Ministral 3 8B 2512",
|
|
2062
|
+
api: "openai-completions",
|
|
2063
|
+
provider: "openrouter",
|
|
2064
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2065
|
+
reasoning: false,
|
|
2066
|
+
input: ["text", "image"],
|
|
2067
|
+
cost: {
|
|
2068
|
+
input: 0.15,
|
|
2069
|
+
output: 0.15,
|
|
2070
|
+
cacheRead: 0,
|
|
2071
|
+
cacheWrite: 0,
|
|
2072
|
+
},
|
|
2073
|
+
contextWindow: 262144,
|
|
2074
|
+
maxTokens: 4096,
|
|
2075
|
+
},
|
|
2076
|
+
"mistralai/ministral-3b-2512": {
|
|
2077
|
+
id: "mistralai/ministral-3b-2512",
|
|
2078
|
+
name: "Mistral: Ministral 3 3B 2512",
|
|
2079
|
+
api: "openai-completions",
|
|
2080
|
+
provider: "openrouter",
|
|
2081
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2082
|
+
reasoning: false,
|
|
2083
|
+
input: ["text", "image"],
|
|
2084
|
+
cost: {
|
|
2085
|
+
input: 0.09999999999999999,
|
|
2086
|
+
output: 0.09999999999999999,
|
|
2087
|
+
cacheRead: 0,
|
|
2088
|
+
cacheWrite: 0,
|
|
2089
|
+
},
|
|
2090
|
+
contextWindow: 131072,
|
|
2091
|
+
maxTokens: 4096,
|
|
2092
|
+
},
|
|
2025
2093
|
"mistralai/mistral-large-2512": {
|
|
2026
2094
|
id: "mistralai/mistral-large-2512",
|
|
2027
2095
|
name: "Mistral: Mistral Large 3 2512",
|
|
@@ -3348,22 +3416,22 @@ export const MODELS = {
|
|
|
3348
3416
|
contextWindow: 400000,
|
|
3349
3417
|
maxTokens: 128000,
|
|
3350
3418
|
},
|
|
3351
|
-
"openai/gpt-oss-120b:
|
|
3352
|
-
id: "openai/gpt-oss-120b:
|
|
3353
|
-
name: "OpenAI: gpt-oss-120b (
|
|
3419
|
+
"openai/gpt-oss-120b:free": {
|
|
3420
|
+
id: "openai/gpt-oss-120b:free",
|
|
3421
|
+
name: "OpenAI: gpt-oss-120b (free)",
|
|
3354
3422
|
api: "openai-completions",
|
|
3355
3423
|
provider: "openrouter",
|
|
3356
3424
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
3357
3425
|
reasoning: true,
|
|
3358
3426
|
input: ["text"],
|
|
3359
3427
|
cost: {
|
|
3360
|
-
input: 0
|
|
3361
|
-
output: 0
|
|
3428
|
+
input: 0,
|
|
3429
|
+
output: 0,
|
|
3362
3430
|
cacheRead: 0,
|
|
3363
3431
|
cacheWrite: 0,
|
|
3364
3432
|
},
|
|
3365
3433
|
contextWindow: 131072,
|
|
3366
|
-
maxTokens:
|
|
3434
|
+
maxTokens: 4096,
|
|
3367
3435
|
},
|
|
3368
3436
|
"openai/gpt-oss-120b": {
|
|
3369
3437
|
id: "openai/gpt-oss-120b",
|
|
@@ -3374,13 +3442,30 @@ export const MODELS = {
|
|
|
3374
3442
|
reasoning: true,
|
|
3375
3443
|
input: ["text"],
|
|
3376
3444
|
cost: {
|
|
3377
|
-
input: 0.
|
|
3378
|
-
output: 0.
|
|
3445
|
+
input: 0.039,
|
|
3446
|
+
output: 0.19,
|
|
3379
3447
|
cacheRead: 0,
|
|
3380
3448
|
cacheWrite: 0,
|
|
3381
3449
|
},
|
|
3382
3450
|
contextWindow: 131072,
|
|
3383
|
-
maxTokens:
|
|
3451
|
+
maxTokens: 4096,
|
|
3452
|
+
},
|
|
3453
|
+
"openai/gpt-oss-120b:exacto": {
|
|
3454
|
+
id: "openai/gpt-oss-120b:exacto",
|
|
3455
|
+
name: "OpenAI: gpt-oss-120b (exacto)",
|
|
3456
|
+
api: "openai-completions",
|
|
3457
|
+
provider: "openrouter",
|
|
3458
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
3459
|
+
reasoning: true,
|
|
3460
|
+
input: ["text"],
|
|
3461
|
+
cost: {
|
|
3462
|
+
input: 0.039,
|
|
3463
|
+
output: 0.19,
|
|
3464
|
+
cacheRead: 0,
|
|
3465
|
+
cacheWrite: 0,
|
|
3466
|
+
},
|
|
3467
|
+
contextWindow: 131072,
|
|
3468
|
+
maxTokens: 4096,
|
|
3384
3469
|
},
|
|
3385
3470
|
"openai/gpt-oss-20b:free": {
|
|
3386
3471
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -4632,13 +4717,13 @@ export const MODELS = {
|
|
|
4632
4717
|
reasoning: false,
|
|
4633
4718
|
input: ["text"],
|
|
4634
4719
|
cost: {
|
|
4635
|
-
input: 0.
|
|
4636
|
-
output: 0.
|
|
4720
|
+
input: 0.03,
|
|
4721
|
+
output: 0.11,
|
|
4637
4722
|
cacheRead: 0,
|
|
4638
4723
|
cacheWrite: 0,
|
|
4639
4724
|
},
|
|
4640
4725
|
contextWindow: 32768,
|
|
4641
|
-
maxTokens:
|
|
4726
|
+
maxTokens: 32768,
|
|
4642
4727
|
},
|
|
4643
4728
|
"deepseek/deepseek-r1-distill-llama-70b": {
|
|
4644
4729
|
id: "deepseek/deepseek-r1-distill-llama-70b",
|