@mariozechner/pi-ai 0.9.4 → 0.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -115,8 +115,8 @@ export const MODELS = {
115
115
  cost: {
116
116
  input: 5,
117
117
  output: 25,
118
- cacheRead: 1.5,
119
- cacheWrite: 18.75,
118
+ cacheRead: 0.5,
119
+ cacheWrite: 6.25,
120
120
  },
121
121
  contextWindow: 200000,
122
122
  maxTokens: 64000,
@@ -1954,6 +1954,40 @@ export const MODELS = {
1954
1954
  },
1955
1955
  },
1956
1956
  openrouter: {
1957
+ "tngtech/tng-r1t-chimera:free": {
1958
+ id: "tngtech/tng-r1t-chimera:free",
1959
+ name: "TNG: R1T Chimera (free)",
1960
+ api: "openai-completions",
1961
+ provider: "openrouter",
1962
+ baseUrl: "https://openrouter.ai/api/v1",
1963
+ reasoning: true,
1964
+ input: ["text"],
1965
+ cost: {
1966
+ input: 0,
1967
+ output: 0,
1968
+ cacheRead: 0,
1969
+ cacheWrite: 0,
1970
+ },
1971
+ contextWindow: 163840,
1972
+ maxTokens: 163840,
1973
+ },
1974
+ "tngtech/tng-r1t-chimera": {
1975
+ id: "tngtech/tng-r1t-chimera",
1976
+ name: "TNG: R1T Chimera",
1977
+ api: "openai-completions",
1978
+ provider: "openrouter",
1979
+ baseUrl: "https://openrouter.ai/api/v1",
1980
+ reasoning: true,
1981
+ input: ["text"],
1982
+ cost: {
1983
+ input: 0.3,
1984
+ output: 1.2,
1985
+ cacheRead: 0,
1986
+ cacheWrite: 0,
1987
+ },
1988
+ contextWindow: 163840,
1989
+ maxTokens: 163840,
1990
+ },
1957
1991
  "anthropic/claude-opus-4.5": {
1958
1992
  id: "anthropic/claude-opus-4.5",
1959
1993
  name: "Anthropic: Claude Opus 4.5",
@@ -3569,23 +3603,6 @@ export const MODELS = {
3569
3603
  contextWindow: 128000,
3570
3604
  maxTokens: 16384,
3571
3605
  },
3572
- "mistralai/mistral-small-3.2-24b-instruct:free": {
3573
- id: "mistralai/mistral-small-3.2-24b-instruct:free",
3574
- name: "Mistral: Mistral Small 3.2 24B (free)",
3575
- api: "openai-completions",
3576
- provider: "openrouter",
3577
- baseUrl: "https://openrouter.ai/api/v1",
3578
- reasoning: false,
3579
- input: ["text", "image"],
3580
- cost: {
3581
- input: 0,
3582
- output: 0,
3583
- cacheRead: 0,
3584
- cacheWrite: 0,
3585
- },
3586
- contextWindow: 131072,
3587
- maxTokens: 4096,
3588
- },
3589
3606
  "mistralai/mistral-small-3.2-24b-instruct": {
3590
3607
  id: "mistralai/mistral-small-3.2-24b-instruct",
3591
3608
  name: "Mistral: Mistral Small 3.2 24B",
@@ -4042,7 +4059,7 @@ export const MODELS = {
4042
4059
  cacheRead: 0,
4043
4060
  cacheWrite: 0,
4044
4061
  },
4045
- contextWindow: 40960,
4062
+ contextWindow: 131072,
4046
4063
  maxTokens: 4096,
4047
4064
  },
4048
4065
  "qwen/qwen3-235b-a22b": {
@@ -4232,23 +4249,6 @@ export const MODELS = {
4232
4249
  contextWindow: 327680,
4233
4250
  maxTokens: 16384,
4234
4251
  },
4235
- "deepseek/deepseek-chat-v3-0324:free": {
4236
- id: "deepseek/deepseek-chat-v3-0324:free",
4237
- name: "DeepSeek: DeepSeek V3 0324 (free)",
4238
- api: "openai-completions",
4239
- provider: "openrouter",
4240
- baseUrl: "https://openrouter.ai/api/v1",
4241
- reasoning: false,
4242
- input: ["text"],
4243
- cost: {
4244
- input: 0,
4245
- output: 0,
4246
- cacheRead: 0,
4247
- cacheWrite: 0,
4248
- },
4249
- contextWindow: 163840,
4250
- maxTokens: 4096,
4251
- },
4252
4252
  "deepseek/deepseek-chat-v3-0324": {
4253
4253
  id: "deepseek/deepseek-chat-v3-0324",
4254
4254
  name: "DeepSeek: DeepSeek V3 0324",
@@ -4258,13 +4258,13 @@ export const MODELS = {
4258
4258
  reasoning: true,
4259
4259
  input: ["text"],
4260
4260
  cost: {
4261
- input: 0.216,
4262
- output: 0.896,
4263
- cacheRead: 0.135,
4261
+ input: 0.19999999999999998,
4262
+ output: 0.88,
4263
+ cacheRead: 0,
4264
4264
  cacheWrite: 0,
4265
4265
  },
4266
4266
  contextWindow: 163840,
4267
- maxTokens: 163840,
4267
+ maxTokens: 4096,
4268
4268
  },
4269
4269
  "mistralai/mistral-small-3.1-24b-instruct:free": {
4270
4270
  id: "mistralai/mistral-small-3.1-24b-instruct:free",
@@ -4280,8 +4280,8 @@ export const MODELS = {
4280
4280
  cacheRead: 0,
4281
4281
  cacheWrite: 0,
4282
4282
  },
4283
- contextWindow: 96000,
4284
- maxTokens: 96000,
4283
+ contextWindow: 128000,
4284
+ maxTokens: 4096,
4285
4285
  },
4286
4286
  "mistralai/mistral-small-3.1-24b-instruct": {
4287
4287
  id: "mistralai/mistral-small-3.1-24b-instruct",
@@ -4810,9 +4810,9 @@ export const MODELS = {
4810
4810
  contextWindow: 32768,
4811
4811
  maxTokens: 4096,
4812
4812
  },
4813
- "anthropic/claude-3.5-haiku": {
4814
- id: "anthropic/claude-3.5-haiku",
4815
- name: "Anthropic: Claude 3.5 Haiku",
4813
+ "anthropic/claude-3.5-haiku-20241022": {
4814
+ id: "anthropic/claude-3.5-haiku-20241022",
4815
+ name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4816
4816
  api: "openai-completions",
4817
4817
  provider: "openrouter",
4818
4818
  baseUrl: "https://openrouter.ai/api/v1",
@@ -4827,9 +4827,9 @@ export const MODELS = {
4827
4827
  contextWindow: 200000,
4828
4828
  maxTokens: 8192,
4829
4829
  },
4830
- "anthropic/claude-3.5-haiku-20241022": {
4831
- id: "anthropic/claude-3.5-haiku-20241022",
4832
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4830
+ "anthropic/claude-3.5-haiku": {
4831
+ id: "anthropic/claude-3.5-haiku",
4832
+ name: "Anthropic: Claude 3.5 Haiku",
4833
4833
  api: "openai-completions",
4834
4834
  provider: "openrouter",
4835
4835
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5150,9 +5150,9 @@ export const MODELS = {
5150
5150
  contextWindow: 131072,
5151
5151
  maxTokens: 16384,
5152
5152
  },
5153
- "openai/gpt-4o-mini": {
5154
- id: "openai/gpt-4o-mini",
5155
- name: "OpenAI: GPT-4o-mini",
5153
+ "openai/gpt-4o-mini-2024-07-18": {
5154
+ id: "openai/gpt-4o-mini-2024-07-18",
5155
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5156
5156
  api: "openai-completions",
5157
5157
  provider: "openrouter",
5158
5158
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5167,9 +5167,9 @@ export const MODELS = {
5167
5167
  contextWindow: 128000,
5168
5168
  maxTokens: 16384,
5169
5169
  },
5170
- "openai/gpt-4o-mini-2024-07-18": {
5171
- id: "openai/gpt-4o-mini-2024-07-18",
5172
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5170
+ "openai/gpt-4o-mini": {
5171
+ id: "openai/gpt-4o-mini",
5172
+ name: "OpenAI: GPT-4o-mini",
5173
5173
  api: "openai-completions",
5174
5174
  provider: "openrouter",
5175
5175
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5269,6 +5269,23 @@ export const MODELS = {
5269
5269
  contextWindow: 128000,
5270
5270
  maxTokens: 4096,
5271
5271
  },
5272
+ "openai/gpt-4o-2024-05-13": {
5273
+ id: "openai/gpt-4o-2024-05-13",
5274
+ name: "OpenAI: GPT-4o (2024-05-13)",
5275
+ api: "openai-completions",
5276
+ provider: "openrouter",
5277
+ baseUrl: "https://openrouter.ai/api/v1",
5278
+ reasoning: false,
5279
+ input: ["text", "image"],
5280
+ cost: {
5281
+ input: 5,
5282
+ output: 15,
5283
+ cacheRead: 0,
5284
+ cacheWrite: 0,
5285
+ },
5286
+ contextWindow: 128000,
5287
+ maxTokens: 4096,
5288
+ },
5272
5289
  "openai/gpt-4o": {
5273
5290
  id: "openai/gpt-4o",
5274
5291
  name: "OpenAI: GPT-4o",
@@ -5303,22 +5320,22 @@ export const MODELS = {
5303
5320
  contextWindow: 128000,
5304
5321
  maxTokens: 64000,
5305
5322
  },
5306
- "openai/gpt-4o-2024-05-13": {
5307
- id: "openai/gpt-4o-2024-05-13",
5308
- name: "OpenAI: GPT-4o (2024-05-13)",
5323
+ "meta-llama/llama-3-70b-instruct": {
5324
+ id: "meta-llama/llama-3-70b-instruct",
5325
+ name: "Meta: Llama 3 70B Instruct",
5309
5326
  api: "openai-completions",
5310
5327
  provider: "openrouter",
5311
5328
  baseUrl: "https://openrouter.ai/api/v1",
5312
5329
  reasoning: false,
5313
- input: ["text", "image"],
5330
+ input: ["text"],
5314
5331
  cost: {
5315
- input: 5,
5316
- output: 15,
5332
+ input: 0.3,
5333
+ output: 0.39999999999999997,
5317
5334
  cacheRead: 0,
5318
5335
  cacheWrite: 0,
5319
5336
  },
5320
- contextWindow: 128000,
5321
- maxTokens: 4096,
5337
+ contextWindow: 8192,
5338
+ maxTokens: 16384,
5322
5339
  },
5323
5340
  "meta-llama/llama-3-8b-instruct": {
5324
5341
  id: "meta-llama/llama-3-8b-instruct",
@@ -5337,23 +5354,6 @@ export const MODELS = {
5337
5354
  contextWindow: 8192,
5338
5355
  maxTokens: 16384,
5339
5356
  },
5340
- "meta-llama/llama-3-70b-instruct": {
5341
- id: "meta-llama/llama-3-70b-instruct",
5342
- name: "Meta: Llama 3 70B Instruct",
5343
- api: "openai-completions",
5344
- provider: "openrouter",
5345
- baseUrl: "https://openrouter.ai/api/v1",
5346
- reasoning: false,
5347
- input: ["text"],
5348
- cost: {
5349
- input: 0.3,
5350
- output: 0.39999999999999997,
5351
- cacheRead: 0,
5352
- cacheWrite: 0,
5353
- },
5354
- contextWindow: 8192,
5355
- maxTokens: 16384,
5356
- },
5357
5357
  "mistralai/mixtral-8x22b-instruct": {
5358
5358
  id: "mistralai/mixtral-8x22b-instruct",
5359
5359
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -5439,38 +5439,38 @@ export const MODELS = {
5439
5439
  contextWindow: 128000,
5440
5440
  maxTokens: 4096,
5441
5441
  },
5442
- "openai/gpt-4-turbo-preview": {
5443
- id: "openai/gpt-4-turbo-preview",
5444
- name: "OpenAI: GPT-4 Turbo Preview",
5442
+ "openai/gpt-3.5-turbo-0613": {
5443
+ id: "openai/gpt-3.5-turbo-0613",
5444
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5445
5445
  api: "openai-completions",
5446
5446
  provider: "openrouter",
5447
5447
  baseUrl: "https://openrouter.ai/api/v1",
5448
5448
  reasoning: false,
5449
5449
  input: ["text"],
5450
5450
  cost: {
5451
- input: 10,
5452
- output: 30,
5451
+ input: 1,
5452
+ output: 2,
5453
5453
  cacheRead: 0,
5454
5454
  cacheWrite: 0,
5455
5455
  },
5456
- contextWindow: 128000,
5456
+ contextWindow: 4095,
5457
5457
  maxTokens: 4096,
5458
5458
  },
5459
- "openai/gpt-3.5-turbo-0613": {
5460
- id: "openai/gpt-3.5-turbo-0613",
5461
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5459
+ "openai/gpt-4-turbo-preview": {
5460
+ id: "openai/gpt-4-turbo-preview",
5461
+ name: "OpenAI: GPT-4 Turbo Preview",
5462
5462
  api: "openai-completions",
5463
5463
  provider: "openrouter",
5464
5464
  baseUrl: "https://openrouter.ai/api/v1",
5465
5465
  reasoning: false,
5466
5466
  input: ["text"],
5467
5467
  cost: {
5468
- input: 1,
5469
- output: 2,
5468
+ input: 10,
5469
+ output: 30,
5470
5470
  cacheRead: 0,
5471
5471
  cacheWrite: 0,
5472
5472
  },
5473
- contextWindow: 4095,
5473
+ contextWindow: 128000,
5474
5474
  maxTokens: 4096,
5475
5475
  },
5476
5476
  "mistralai/mistral-small": {