@mariozechner/pi-ai 0.9.4 → 0.10.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -115,8 +115,8 @@ export const MODELS = {
|
|
|
115
115
|
cost: {
|
|
116
116
|
input: 5,
|
|
117
117
|
output: 25,
|
|
118
|
-
cacheRead:
|
|
119
|
-
cacheWrite:
|
|
118
|
+
cacheRead: 0.5,
|
|
119
|
+
cacheWrite: 6.25,
|
|
120
120
|
},
|
|
121
121
|
contextWindow: 200000,
|
|
122
122
|
maxTokens: 64000,
|
|
@@ -1954,6 +1954,40 @@ export const MODELS = {
|
|
|
1954
1954
|
},
|
|
1955
1955
|
},
|
|
1956
1956
|
openrouter: {
|
|
1957
|
+
"tngtech/tng-r1t-chimera:free": {
|
|
1958
|
+
id: "tngtech/tng-r1t-chimera:free",
|
|
1959
|
+
name: "TNG: R1T Chimera (free)",
|
|
1960
|
+
api: "openai-completions",
|
|
1961
|
+
provider: "openrouter",
|
|
1962
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1963
|
+
reasoning: true,
|
|
1964
|
+
input: ["text"],
|
|
1965
|
+
cost: {
|
|
1966
|
+
input: 0,
|
|
1967
|
+
output: 0,
|
|
1968
|
+
cacheRead: 0,
|
|
1969
|
+
cacheWrite: 0,
|
|
1970
|
+
},
|
|
1971
|
+
contextWindow: 163840,
|
|
1972
|
+
maxTokens: 163840,
|
|
1973
|
+
},
|
|
1974
|
+
"tngtech/tng-r1t-chimera": {
|
|
1975
|
+
id: "tngtech/tng-r1t-chimera",
|
|
1976
|
+
name: "TNG: R1T Chimera",
|
|
1977
|
+
api: "openai-completions",
|
|
1978
|
+
provider: "openrouter",
|
|
1979
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1980
|
+
reasoning: true,
|
|
1981
|
+
input: ["text"],
|
|
1982
|
+
cost: {
|
|
1983
|
+
input: 0.3,
|
|
1984
|
+
output: 1.2,
|
|
1985
|
+
cacheRead: 0,
|
|
1986
|
+
cacheWrite: 0,
|
|
1987
|
+
},
|
|
1988
|
+
contextWindow: 163840,
|
|
1989
|
+
maxTokens: 163840,
|
|
1990
|
+
},
|
|
1957
1991
|
"anthropic/claude-opus-4.5": {
|
|
1958
1992
|
id: "anthropic/claude-opus-4.5",
|
|
1959
1993
|
name: "Anthropic: Claude Opus 4.5",
|
|
@@ -3569,23 +3603,6 @@ export const MODELS = {
|
|
|
3569
3603
|
contextWindow: 128000,
|
|
3570
3604
|
maxTokens: 16384,
|
|
3571
3605
|
},
|
|
3572
|
-
"mistralai/mistral-small-3.2-24b-instruct:free": {
|
|
3573
|
-
id: "mistralai/mistral-small-3.2-24b-instruct:free",
|
|
3574
|
-
name: "Mistral: Mistral Small 3.2 24B (free)",
|
|
3575
|
-
api: "openai-completions",
|
|
3576
|
-
provider: "openrouter",
|
|
3577
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
3578
|
-
reasoning: false,
|
|
3579
|
-
input: ["text", "image"],
|
|
3580
|
-
cost: {
|
|
3581
|
-
input: 0,
|
|
3582
|
-
output: 0,
|
|
3583
|
-
cacheRead: 0,
|
|
3584
|
-
cacheWrite: 0,
|
|
3585
|
-
},
|
|
3586
|
-
contextWindow: 131072,
|
|
3587
|
-
maxTokens: 4096,
|
|
3588
|
-
},
|
|
3589
3606
|
"mistralai/mistral-small-3.2-24b-instruct": {
|
|
3590
3607
|
id: "mistralai/mistral-small-3.2-24b-instruct",
|
|
3591
3608
|
name: "Mistral: Mistral Small 3.2 24B",
|
|
@@ -4042,7 +4059,7 @@ export const MODELS = {
|
|
|
4042
4059
|
cacheRead: 0,
|
|
4043
4060
|
cacheWrite: 0,
|
|
4044
4061
|
},
|
|
4045
|
-
contextWindow:
|
|
4062
|
+
contextWindow: 131072,
|
|
4046
4063
|
maxTokens: 4096,
|
|
4047
4064
|
},
|
|
4048
4065
|
"qwen/qwen3-235b-a22b": {
|
|
@@ -4232,23 +4249,6 @@ export const MODELS = {
|
|
|
4232
4249
|
contextWindow: 327680,
|
|
4233
4250
|
maxTokens: 16384,
|
|
4234
4251
|
},
|
|
4235
|
-
"deepseek/deepseek-chat-v3-0324:free": {
|
|
4236
|
-
id: "deepseek/deepseek-chat-v3-0324:free",
|
|
4237
|
-
name: "DeepSeek: DeepSeek V3 0324 (free)",
|
|
4238
|
-
api: "openai-completions",
|
|
4239
|
-
provider: "openrouter",
|
|
4240
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4241
|
-
reasoning: false,
|
|
4242
|
-
input: ["text"],
|
|
4243
|
-
cost: {
|
|
4244
|
-
input: 0,
|
|
4245
|
-
output: 0,
|
|
4246
|
-
cacheRead: 0,
|
|
4247
|
-
cacheWrite: 0,
|
|
4248
|
-
},
|
|
4249
|
-
contextWindow: 163840,
|
|
4250
|
-
maxTokens: 4096,
|
|
4251
|
-
},
|
|
4252
4252
|
"deepseek/deepseek-chat-v3-0324": {
|
|
4253
4253
|
id: "deepseek/deepseek-chat-v3-0324",
|
|
4254
4254
|
name: "DeepSeek: DeepSeek V3 0324",
|
|
@@ -4258,13 +4258,13 @@ export const MODELS = {
|
|
|
4258
4258
|
reasoning: true,
|
|
4259
4259
|
input: ["text"],
|
|
4260
4260
|
cost: {
|
|
4261
|
-
input: 0.
|
|
4262
|
-
output: 0.
|
|
4263
|
-
cacheRead: 0
|
|
4261
|
+
input: 0.19999999999999998,
|
|
4262
|
+
output: 0.88,
|
|
4263
|
+
cacheRead: 0,
|
|
4264
4264
|
cacheWrite: 0,
|
|
4265
4265
|
},
|
|
4266
4266
|
contextWindow: 163840,
|
|
4267
|
-
maxTokens:
|
|
4267
|
+
maxTokens: 4096,
|
|
4268
4268
|
},
|
|
4269
4269
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
4270
4270
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
|
@@ -4280,8 +4280,8 @@ export const MODELS = {
|
|
|
4280
4280
|
cacheRead: 0,
|
|
4281
4281
|
cacheWrite: 0,
|
|
4282
4282
|
},
|
|
4283
|
-
contextWindow:
|
|
4284
|
-
maxTokens:
|
|
4283
|
+
contextWindow: 128000,
|
|
4284
|
+
maxTokens: 4096,
|
|
4285
4285
|
},
|
|
4286
4286
|
"mistralai/mistral-small-3.1-24b-instruct": {
|
|
4287
4287
|
id: "mistralai/mistral-small-3.1-24b-instruct",
|
|
@@ -4810,9 +4810,9 @@ export const MODELS = {
|
|
|
4810
4810
|
contextWindow: 32768,
|
|
4811
4811
|
maxTokens: 4096,
|
|
4812
4812
|
},
|
|
4813
|
-
"anthropic/claude-3.5-haiku": {
|
|
4814
|
-
id: "anthropic/claude-3.5-haiku",
|
|
4815
|
-
name: "Anthropic: Claude 3.5 Haiku",
|
|
4813
|
+
"anthropic/claude-3.5-haiku-20241022": {
|
|
4814
|
+
id: "anthropic/claude-3.5-haiku-20241022",
|
|
4815
|
+
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
4816
4816
|
api: "openai-completions",
|
|
4817
4817
|
provider: "openrouter",
|
|
4818
4818
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -4827,9 +4827,9 @@ export const MODELS = {
|
|
|
4827
4827
|
contextWindow: 200000,
|
|
4828
4828
|
maxTokens: 8192,
|
|
4829
4829
|
},
|
|
4830
|
-
"anthropic/claude-3.5-haiku
|
|
4831
|
-
id: "anthropic/claude-3.5-haiku
|
|
4832
|
-
name: "Anthropic: Claude 3.5 Haiku
|
|
4830
|
+
"anthropic/claude-3.5-haiku": {
|
|
4831
|
+
id: "anthropic/claude-3.5-haiku",
|
|
4832
|
+
name: "Anthropic: Claude 3.5 Haiku",
|
|
4833
4833
|
api: "openai-completions",
|
|
4834
4834
|
provider: "openrouter",
|
|
4835
4835
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5150,9 +5150,9 @@ export const MODELS = {
|
|
|
5150
5150
|
contextWindow: 131072,
|
|
5151
5151
|
maxTokens: 16384,
|
|
5152
5152
|
},
|
|
5153
|
-
"openai/gpt-4o-mini": {
|
|
5154
|
-
id: "openai/gpt-4o-mini",
|
|
5155
|
-
name: "OpenAI: GPT-4o-mini",
|
|
5153
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
5154
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
5155
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
5156
5156
|
api: "openai-completions",
|
|
5157
5157
|
provider: "openrouter",
|
|
5158
5158
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5167,9 +5167,9 @@ export const MODELS = {
|
|
|
5167
5167
|
contextWindow: 128000,
|
|
5168
5168
|
maxTokens: 16384,
|
|
5169
5169
|
},
|
|
5170
|
-
"openai/gpt-4o-mini
|
|
5171
|
-
id: "openai/gpt-4o-mini
|
|
5172
|
-
name: "OpenAI: GPT-4o-mini
|
|
5170
|
+
"openai/gpt-4o-mini": {
|
|
5171
|
+
id: "openai/gpt-4o-mini",
|
|
5172
|
+
name: "OpenAI: GPT-4o-mini",
|
|
5173
5173
|
api: "openai-completions",
|
|
5174
5174
|
provider: "openrouter",
|
|
5175
5175
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5269,6 +5269,23 @@ export const MODELS = {
|
|
|
5269
5269
|
contextWindow: 128000,
|
|
5270
5270
|
maxTokens: 4096,
|
|
5271
5271
|
},
|
|
5272
|
+
"openai/gpt-4o-2024-05-13": {
|
|
5273
|
+
id: "openai/gpt-4o-2024-05-13",
|
|
5274
|
+
name: "OpenAI: GPT-4o (2024-05-13)",
|
|
5275
|
+
api: "openai-completions",
|
|
5276
|
+
provider: "openrouter",
|
|
5277
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5278
|
+
reasoning: false,
|
|
5279
|
+
input: ["text", "image"],
|
|
5280
|
+
cost: {
|
|
5281
|
+
input: 5,
|
|
5282
|
+
output: 15,
|
|
5283
|
+
cacheRead: 0,
|
|
5284
|
+
cacheWrite: 0,
|
|
5285
|
+
},
|
|
5286
|
+
contextWindow: 128000,
|
|
5287
|
+
maxTokens: 4096,
|
|
5288
|
+
},
|
|
5272
5289
|
"openai/gpt-4o": {
|
|
5273
5290
|
id: "openai/gpt-4o",
|
|
5274
5291
|
name: "OpenAI: GPT-4o",
|
|
@@ -5303,22 +5320,22 @@ export const MODELS = {
|
|
|
5303
5320
|
contextWindow: 128000,
|
|
5304
5321
|
maxTokens: 64000,
|
|
5305
5322
|
},
|
|
5306
|
-
"
|
|
5307
|
-
id: "
|
|
5308
|
-
name: "
|
|
5323
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
5324
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
5325
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
5309
5326
|
api: "openai-completions",
|
|
5310
5327
|
provider: "openrouter",
|
|
5311
5328
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5312
5329
|
reasoning: false,
|
|
5313
|
-
input: ["text"
|
|
5330
|
+
input: ["text"],
|
|
5314
5331
|
cost: {
|
|
5315
|
-
input:
|
|
5316
|
-
output:
|
|
5332
|
+
input: 0.3,
|
|
5333
|
+
output: 0.39999999999999997,
|
|
5317
5334
|
cacheRead: 0,
|
|
5318
5335
|
cacheWrite: 0,
|
|
5319
5336
|
},
|
|
5320
|
-
contextWindow:
|
|
5321
|
-
maxTokens:
|
|
5337
|
+
contextWindow: 8192,
|
|
5338
|
+
maxTokens: 16384,
|
|
5322
5339
|
},
|
|
5323
5340
|
"meta-llama/llama-3-8b-instruct": {
|
|
5324
5341
|
id: "meta-llama/llama-3-8b-instruct",
|
|
@@ -5337,23 +5354,6 @@ export const MODELS = {
|
|
|
5337
5354
|
contextWindow: 8192,
|
|
5338
5355
|
maxTokens: 16384,
|
|
5339
5356
|
},
|
|
5340
|
-
"meta-llama/llama-3-70b-instruct": {
|
|
5341
|
-
id: "meta-llama/llama-3-70b-instruct",
|
|
5342
|
-
name: "Meta: Llama 3 70B Instruct",
|
|
5343
|
-
api: "openai-completions",
|
|
5344
|
-
provider: "openrouter",
|
|
5345
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5346
|
-
reasoning: false,
|
|
5347
|
-
input: ["text"],
|
|
5348
|
-
cost: {
|
|
5349
|
-
input: 0.3,
|
|
5350
|
-
output: 0.39999999999999997,
|
|
5351
|
-
cacheRead: 0,
|
|
5352
|
-
cacheWrite: 0,
|
|
5353
|
-
},
|
|
5354
|
-
contextWindow: 8192,
|
|
5355
|
-
maxTokens: 16384,
|
|
5356
|
-
},
|
|
5357
5357
|
"mistralai/mixtral-8x22b-instruct": {
|
|
5358
5358
|
id: "mistralai/mixtral-8x22b-instruct",
|
|
5359
5359
|
name: "Mistral: Mixtral 8x22B Instruct",
|
|
@@ -5439,38 +5439,38 @@ export const MODELS = {
|
|
|
5439
5439
|
contextWindow: 128000,
|
|
5440
5440
|
maxTokens: 4096,
|
|
5441
5441
|
},
|
|
5442
|
-
"openai/gpt-
|
|
5443
|
-
id: "openai/gpt-
|
|
5444
|
-
name: "OpenAI: GPT-
|
|
5442
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
5443
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
5444
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
5445
5445
|
api: "openai-completions",
|
|
5446
5446
|
provider: "openrouter",
|
|
5447
5447
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5448
5448
|
reasoning: false,
|
|
5449
5449
|
input: ["text"],
|
|
5450
5450
|
cost: {
|
|
5451
|
-
input:
|
|
5452
|
-
output:
|
|
5451
|
+
input: 1,
|
|
5452
|
+
output: 2,
|
|
5453
5453
|
cacheRead: 0,
|
|
5454
5454
|
cacheWrite: 0,
|
|
5455
5455
|
},
|
|
5456
|
-
contextWindow:
|
|
5456
|
+
contextWindow: 4095,
|
|
5457
5457
|
maxTokens: 4096,
|
|
5458
5458
|
},
|
|
5459
|
-
"openai/gpt-
|
|
5460
|
-
id: "openai/gpt-
|
|
5461
|
-
name: "OpenAI: GPT-
|
|
5459
|
+
"openai/gpt-4-turbo-preview": {
|
|
5460
|
+
id: "openai/gpt-4-turbo-preview",
|
|
5461
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
5462
5462
|
api: "openai-completions",
|
|
5463
5463
|
provider: "openrouter",
|
|
5464
5464
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5465
5465
|
reasoning: false,
|
|
5466
5466
|
input: ["text"],
|
|
5467
5467
|
cost: {
|
|
5468
|
-
input:
|
|
5469
|
-
output:
|
|
5468
|
+
input: 10,
|
|
5469
|
+
output: 30,
|
|
5470
5470
|
cacheRead: 0,
|
|
5471
5471
|
cacheWrite: 0,
|
|
5472
5472
|
},
|
|
5473
|
-
contextWindow:
|
|
5473
|
+
contextWindow: 128000,
|
|
5474
5474
|
maxTokens: 4096,
|
|
5475
5475
|
},
|
|
5476
5476
|
"mistralai/mistral-small": {
|