@mariozechner/pi-ai 0.12.14 → 0.12.15

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2224,7 +2224,7 @@ export const MODELS = {
2224
2224
  cacheWrite: 6.25,
2225
2225
  },
2226
2226
  contextWindow: 200000,
2227
- maxTokens: 64000,
2227
+ maxTokens: 32000,
2228
2228
  },
2229
2229
  "allenai/olmo-3-7b-instruct": {
2230
2230
  id: "allenai/olmo-3-7b-instruct",
@@ -4980,9 +4980,9 @@ export const MODELS = {
4980
4980
  contextWindow: 32768,
4981
4981
  maxTokens: 4096,
4982
4982
  },
4983
- "anthropic/claude-3.5-haiku": {
4984
- id: "anthropic/claude-3.5-haiku",
4985
- name: "Anthropic: Claude 3.5 Haiku",
4983
+ "anthropic/claude-3.5-haiku-20241022": {
4984
+ id: "anthropic/claude-3.5-haiku-20241022",
4985
+ name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4986
4986
  api: "openai-completions",
4987
4987
  provider: "openrouter",
4988
4988
  baseUrl: "https://openrouter.ai/api/v1",
@@ -4997,9 +4997,9 @@ export const MODELS = {
4997
4997
  contextWindow: 200000,
4998
4998
  maxTokens: 8192,
4999
4999
  },
5000
- "anthropic/claude-3.5-haiku-20241022": {
5001
- id: "anthropic/claude-3.5-haiku-20241022",
5002
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
5000
+ "anthropic/claude-3.5-haiku": {
5001
+ id: "anthropic/claude-3.5-haiku",
5002
+ name: "Anthropic: Claude 3.5 Haiku",
5003
5003
  api: "openai-completions",
5004
5004
  provider: "openrouter",
5005
5005
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5235,23 +5235,6 @@ export const MODELS = {
5235
5235
  contextWindow: 128000,
5236
5236
  maxTokens: 16384,
5237
5237
  },
5238
- "meta-llama/llama-3.1-405b-instruct": {
5239
- id: "meta-llama/llama-3.1-405b-instruct",
5240
- name: "Meta: Llama 3.1 405B Instruct",
5241
- api: "openai-completions",
5242
- provider: "openrouter",
5243
- baseUrl: "https://openrouter.ai/api/v1",
5244
- reasoning: false,
5245
- input: ["text"],
5246
- cost: {
5247
- input: 3.5,
5248
- output: 3.5,
5249
- cacheRead: 0,
5250
- cacheWrite: 0,
5251
- },
5252
- contextWindow: 130815,
5253
- maxTokens: 4096,
5254
- },
5255
5238
  "meta-llama/llama-3.1-8b-instruct": {
5256
5239
  id: "meta-llama/llama-3.1-8b-instruct",
5257
5240
  name: "Meta: Llama 3.1 8B Instruct",
@@ -5286,6 +5269,23 @@ export const MODELS = {
5286
5269
  contextWindow: 131072,
5287
5270
  maxTokens: 4096,
5288
5271
  },
5272
+ "meta-llama/llama-3.1-405b-instruct": {
5273
+ id: "meta-llama/llama-3.1-405b-instruct",
5274
+ name: "Meta: Llama 3.1 405B Instruct",
5275
+ api: "openai-completions",
5276
+ provider: "openrouter",
5277
+ baseUrl: "https://openrouter.ai/api/v1",
5278
+ reasoning: false,
5279
+ input: ["text"],
5280
+ cost: {
5281
+ input: 3.5,
5282
+ output: 3.5,
5283
+ cacheRead: 0,
5284
+ cacheWrite: 0,
5285
+ },
5286
+ contextWindow: 130815,
5287
+ maxTokens: 4096,
5288
+ },
5289
5289
  "mistralai/mistral-nemo": {
5290
5290
  id: "mistralai/mistral-nemo",
5291
5291
  name: "Mistral: Mistral Nemo",
@@ -5303,9 +5303,9 @@ export const MODELS = {
5303
5303
  contextWindow: 131072,
5304
5304
  maxTokens: 16384,
5305
5305
  },
5306
- "openai/gpt-4o-mini-2024-07-18": {
5307
- id: "openai/gpt-4o-mini-2024-07-18",
5308
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5306
+ "openai/gpt-4o-mini": {
5307
+ id: "openai/gpt-4o-mini",
5308
+ name: "OpenAI: GPT-4o-mini",
5309
5309
  api: "openai-completions",
5310
5310
  provider: "openrouter",
5311
5311
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5320,9 +5320,9 @@ export const MODELS = {
5320
5320
  contextWindow: 128000,
5321
5321
  maxTokens: 16384,
5322
5322
  },
5323
- "openai/gpt-4o-mini": {
5324
- id: "openai/gpt-4o-mini",
5325
- name: "OpenAI: GPT-4o-mini",
5323
+ "openai/gpt-4o-mini-2024-07-18": {
5324
+ id: "openai/gpt-4o-mini-2024-07-18",
5325
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5326
5326
  api: "openai-completions",
5327
5327
  provider: "openrouter",
5328
5328
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5422,23 +5422,6 @@ export const MODELS = {
5422
5422
  contextWindow: 128000,
5423
5423
  maxTokens: 4096,
5424
5424
  },
5425
- "openai/gpt-4o-2024-05-13": {
5426
- id: "openai/gpt-4o-2024-05-13",
5427
- name: "OpenAI: GPT-4o (2024-05-13)",
5428
- api: "openai-completions",
5429
- provider: "openrouter",
5430
- baseUrl: "https://openrouter.ai/api/v1",
5431
- reasoning: false,
5432
- input: ["text", "image"],
5433
- cost: {
5434
- input: 5,
5435
- output: 15,
5436
- cacheRead: 0,
5437
- cacheWrite: 0,
5438
- },
5439
- contextWindow: 128000,
5440
- maxTokens: 4096,
5441
- },
5442
5425
  "openai/gpt-4o": {
5443
5426
  id: "openai/gpt-4o",
5444
5427
  name: "OpenAI: GPT-4o",
@@ -5473,22 +5456,22 @@ export const MODELS = {
5473
5456
  contextWindow: 128000,
5474
5457
  maxTokens: 64000,
5475
5458
  },
5476
- "meta-llama/llama-3-70b-instruct": {
5477
- id: "meta-llama/llama-3-70b-instruct",
5478
- name: "Meta: Llama 3 70B Instruct",
5459
+ "openai/gpt-4o-2024-05-13": {
5460
+ id: "openai/gpt-4o-2024-05-13",
5461
+ name: "OpenAI: GPT-4o (2024-05-13)",
5479
5462
  api: "openai-completions",
5480
5463
  provider: "openrouter",
5481
5464
  baseUrl: "https://openrouter.ai/api/v1",
5482
5465
  reasoning: false,
5483
- input: ["text"],
5466
+ input: ["text", "image"],
5484
5467
  cost: {
5485
- input: 0.3,
5486
- output: 0.39999999999999997,
5468
+ input: 5,
5469
+ output: 15,
5487
5470
  cacheRead: 0,
5488
5471
  cacheWrite: 0,
5489
5472
  },
5490
- contextWindow: 8192,
5491
- maxTokens: 16384,
5473
+ contextWindow: 128000,
5474
+ maxTokens: 4096,
5492
5475
  },
5493
5476
  "meta-llama/llama-3-8b-instruct": {
5494
5477
  id: "meta-llama/llama-3-8b-instruct",
@@ -5507,6 +5490,23 @@ export const MODELS = {
5507
5490
  contextWindow: 8192,
5508
5491
  maxTokens: 16384,
5509
5492
  },
5493
+ "meta-llama/llama-3-70b-instruct": {
5494
+ id: "meta-llama/llama-3-70b-instruct",
5495
+ name: "Meta: Llama 3 70B Instruct",
5496
+ api: "openai-completions",
5497
+ provider: "openrouter",
5498
+ baseUrl: "https://openrouter.ai/api/v1",
5499
+ reasoning: false,
5500
+ input: ["text"],
5501
+ cost: {
5502
+ input: 0.3,
5503
+ output: 0.39999999999999997,
5504
+ cacheRead: 0,
5505
+ cacheWrite: 0,
5506
+ },
5507
+ contextWindow: 8192,
5508
+ maxTokens: 16384,
5509
+ },
5510
5510
  "mistralai/mixtral-8x22b-instruct": {
5511
5511
  id: "mistralai/mixtral-8x22b-instruct",
5512
5512
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -5592,38 +5592,38 @@ export const MODELS = {
5592
5592
  contextWindow: 128000,
5593
5593
  maxTokens: 4096,
5594
5594
  },
5595
- "openai/gpt-3.5-turbo-0613": {
5596
- id: "openai/gpt-3.5-turbo-0613",
5597
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5595
+ "openai/gpt-4-turbo-preview": {
5596
+ id: "openai/gpt-4-turbo-preview",
5597
+ name: "OpenAI: GPT-4 Turbo Preview",
5598
5598
  api: "openai-completions",
5599
5599
  provider: "openrouter",
5600
5600
  baseUrl: "https://openrouter.ai/api/v1",
5601
5601
  reasoning: false,
5602
5602
  input: ["text"],
5603
5603
  cost: {
5604
- input: 1,
5605
- output: 2,
5604
+ input: 10,
5605
+ output: 30,
5606
5606
  cacheRead: 0,
5607
5607
  cacheWrite: 0,
5608
5608
  },
5609
- contextWindow: 4095,
5609
+ contextWindow: 128000,
5610
5610
  maxTokens: 4096,
5611
5611
  },
5612
- "openai/gpt-4-turbo-preview": {
5613
- id: "openai/gpt-4-turbo-preview",
5614
- name: "OpenAI: GPT-4 Turbo Preview",
5612
+ "openai/gpt-3.5-turbo-0613": {
5613
+ id: "openai/gpt-3.5-turbo-0613",
5614
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5615
5615
  api: "openai-completions",
5616
5616
  provider: "openrouter",
5617
5617
  baseUrl: "https://openrouter.ai/api/v1",
5618
5618
  reasoning: false,
5619
5619
  input: ["text"],
5620
5620
  cost: {
5621
- input: 10,
5622
- output: 30,
5621
+ input: 1,
5622
+ output: 2,
5623
5623
  cacheRead: 0,
5624
5624
  cacheWrite: 0,
5625
5625
  },
5626
- contextWindow: 128000,
5626
+ contextWindow: 4095,
5627
5627
  maxTokens: 4096,
5628
5628
  },
5629
5629
  "mistralai/mistral-tiny": {