@mariozechner/pi-ai 0.8.5 → 0.9.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4275,8 +4275,8 @@ export const MODELS = {
4275
4275
  reasoning: false,
4276
4276
  input: ["text", "image"],
4277
4277
  cost: {
4278
- input: 0.049999999999999996,
4279
- output: 0.22,
4278
+ input: 0.03,
4279
+ output: 0.11,
4280
4280
  cacheRead: 0,
4281
4281
  cacheWrite: 0,
4282
4282
  },
@@ -5099,38 +5099,38 @@ export const MODELS = {
5099
5099
  contextWindow: 131072,
5100
5100
  maxTokens: 16384,
5101
5101
  },
5102
- "meta-llama/llama-3.1-70b-instruct": {
5103
- id: "meta-llama/llama-3.1-70b-instruct",
5104
- name: "Meta: Llama 3.1 70B Instruct",
5102
+ "meta-llama/llama-3.1-405b-instruct": {
5103
+ id: "meta-llama/llama-3.1-405b-instruct",
5104
+ name: "Meta: Llama 3.1 405B Instruct",
5105
5105
  api: "openai-completions",
5106
5106
  provider: "openrouter",
5107
5107
  baseUrl: "https://openrouter.ai/api/v1",
5108
5108
  reasoning: false,
5109
5109
  input: ["text"],
5110
5110
  cost: {
5111
- input: 0.39999999999999997,
5112
- output: 0.39999999999999997,
5111
+ input: 3.5,
5112
+ output: 3.5,
5113
5113
  cacheRead: 0,
5114
5114
  cacheWrite: 0,
5115
5115
  },
5116
- contextWindow: 131072,
5116
+ contextWindow: 130815,
5117
5117
  maxTokens: 4096,
5118
5118
  },
5119
- "meta-llama/llama-3.1-405b-instruct": {
5120
- id: "meta-llama/llama-3.1-405b-instruct",
5121
- name: "Meta: Llama 3.1 405B Instruct",
5119
+ "meta-llama/llama-3.1-70b-instruct": {
5120
+ id: "meta-llama/llama-3.1-70b-instruct",
5121
+ name: "Meta: Llama 3.1 70B Instruct",
5122
5122
  api: "openai-completions",
5123
5123
  provider: "openrouter",
5124
5124
  baseUrl: "https://openrouter.ai/api/v1",
5125
5125
  reasoning: false,
5126
5126
  input: ["text"],
5127
5127
  cost: {
5128
- input: 3.5,
5129
- output: 3.5,
5128
+ input: 0.39999999999999997,
5129
+ output: 0.39999999999999997,
5130
5130
  cacheRead: 0,
5131
5131
  cacheWrite: 0,
5132
5132
  },
5133
- contextWindow: 130815,
5133
+ contextWindow: 131072,
5134
5134
  maxTokens: 4096,
5135
5135
  },
5136
5136
  "mistralai/mistral-nemo": {
@@ -5150,9 +5150,9 @@ export const MODELS = {
5150
5150
  contextWindow: 131072,
5151
5151
  maxTokens: 16384,
5152
5152
  },
5153
- "openai/gpt-4o-mini": {
5154
- id: "openai/gpt-4o-mini",
5155
- name: "OpenAI: GPT-4o-mini",
5153
+ "openai/gpt-4o-mini-2024-07-18": {
5154
+ id: "openai/gpt-4o-mini-2024-07-18",
5155
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5156
5156
  api: "openai-completions",
5157
5157
  provider: "openrouter",
5158
5158
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5167,9 +5167,9 @@ export const MODELS = {
5167
5167
  contextWindow: 128000,
5168
5168
  maxTokens: 16384,
5169
5169
  },
5170
- "openai/gpt-4o-mini-2024-07-18": {
5171
- id: "openai/gpt-4o-mini-2024-07-18",
5172
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5170
+ "openai/gpt-4o-mini": {
5171
+ id: "openai/gpt-4o-mini",
5172
+ name: "OpenAI: GPT-4o-mini",
5173
5173
  api: "openai-completions",
5174
5174
  provider: "openrouter",
5175
5175
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5269,6 +5269,23 @@ export const MODELS = {
5269
5269
  contextWindow: 128000,
5270
5270
  maxTokens: 4096,
5271
5271
  },
5272
+ "openai/gpt-4o-2024-05-13": {
5273
+ id: "openai/gpt-4o-2024-05-13",
5274
+ name: "OpenAI: GPT-4o (2024-05-13)",
5275
+ api: "openai-completions",
5276
+ provider: "openrouter",
5277
+ baseUrl: "https://openrouter.ai/api/v1",
5278
+ reasoning: false,
5279
+ input: ["text", "image"],
5280
+ cost: {
5281
+ input: 5,
5282
+ output: 15,
5283
+ cacheRead: 0,
5284
+ cacheWrite: 0,
5285
+ },
5286
+ contextWindow: 128000,
5287
+ maxTokens: 4096,
5288
+ },
5272
5289
  "openai/gpt-4o": {
5273
5290
  id: "openai/gpt-4o",
5274
5291
  name: "OpenAI: GPT-4o",
@@ -5303,23 +5320,6 @@ export const MODELS = {
5303
5320
  contextWindow: 128000,
5304
5321
  maxTokens: 64000,
5305
5322
  },
5306
- "openai/gpt-4o-2024-05-13": {
5307
- id: "openai/gpt-4o-2024-05-13",
5308
- name: "OpenAI: GPT-4o (2024-05-13)",
5309
- api: "openai-completions",
5310
- provider: "openrouter",
5311
- baseUrl: "https://openrouter.ai/api/v1",
5312
- reasoning: false,
5313
- input: ["text", "image"],
5314
- cost: {
5315
- input: 5,
5316
- output: 15,
5317
- cacheRead: 0,
5318
- cacheWrite: 0,
5319
- },
5320
- contextWindow: 128000,
5321
- maxTokens: 4096,
5322
- },
5323
5323
  "meta-llama/llama-3-70b-instruct": {
5324
5324
  id: "meta-llama/llama-3-70b-instruct",
5325
5325
  name: "Meta: Llama 3 70B Instruct",
@@ -5439,38 +5439,38 @@ export const MODELS = {
5439
5439
  contextWindow: 128000,
5440
5440
  maxTokens: 4096,
5441
5441
  },
5442
- "openai/gpt-4-turbo-preview": {
5443
- id: "openai/gpt-4-turbo-preview",
5444
- name: "OpenAI: GPT-4 Turbo Preview",
5442
+ "openai/gpt-3.5-turbo-0613": {
5443
+ id: "openai/gpt-3.5-turbo-0613",
5444
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5445
5445
  api: "openai-completions",
5446
5446
  provider: "openrouter",
5447
5447
  baseUrl: "https://openrouter.ai/api/v1",
5448
5448
  reasoning: false,
5449
5449
  input: ["text"],
5450
5450
  cost: {
5451
- input: 10,
5452
- output: 30,
5451
+ input: 1,
5452
+ output: 2,
5453
5453
  cacheRead: 0,
5454
5454
  cacheWrite: 0,
5455
5455
  },
5456
- contextWindow: 128000,
5456
+ contextWindow: 4095,
5457
5457
  maxTokens: 4096,
5458
5458
  },
5459
- "openai/gpt-3.5-turbo-0613": {
5460
- id: "openai/gpt-3.5-turbo-0613",
5461
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5459
+ "openai/gpt-4-turbo-preview": {
5460
+ id: "openai/gpt-4-turbo-preview",
5461
+ name: "OpenAI: GPT-4 Turbo Preview",
5462
5462
  api: "openai-completions",
5463
5463
  provider: "openrouter",
5464
5464
  baseUrl: "https://openrouter.ai/api/v1",
5465
5465
  reasoning: false,
5466
5466
  input: ["text"],
5467
5467
  cost: {
5468
- input: 1,
5469
- output: 2,
5468
+ input: 10,
5469
+ output: 30,
5470
5470
  cacheRead: 0,
5471
5471
  cacheWrite: 0,
5472
5472
  },
5473
- contextWindow: 4095,
5473
+ contextWindow: 128000,
5474
5474
  maxTokens: 4096,
5475
5475
  },
5476
5476
  "mistralai/mistral-small": {
@@ -5575,38 +5575,38 @@ export const MODELS = {
5575
5575
  contextWindow: 8191,
5576
5576
  maxTokens: 4096,
5577
5577
  },
5578
- "openai/gpt-3.5-turbo": {
5579
- id: "openai/gpt-3.5-turbo",
5580
- name: "OpenAI: GPT-3.5 Turbo",
5578
+ "openai/gpt-4": {
5579
+ id: "openai/gpt-4",
5580
+ name: "OpenAI: GPT-4",
5581
5581
  api: "openai-completions",
5582
5582
  provider: "openrouter",
5583
5583
  baseUrl: "https://openrouter.ai/api/v1",
5584
5584
  reasoning: false,
5585
5585
  input: ["text"],
5586
5586
  cost: {
5587
- input: 0.5,
5588
- output: 1.5,
5587
+ input: 30,
5588
+ output: 60,
5589
5589
  cacheRead: 0,
5590
5590
  cacheWrite: 0,
5591
5591
  },
5592
- contextWindow: 16385,
5592
+ contextWindow: 8191,
5593
5593
  maxTokens: 4096,
5594
5594
  },
5595
- "openai/gpt-4": {
5596
- id: "openai/gpt-4",
5597
- name: "OpenAI: GPT-4",
5595
+ "openai/gpt-3.5-turbo": {
5596
+ id: "openai/gpt-3.5-turbo",
5597
+ name: "OpenAI: GPT-3.5 Turbo",
5598
5598
  api: "openai-completions",
5599
5599
  provider: "openrouter",
5600
5600
  baseUrl: "https://openrouter.ai/api/v1",
5601
5601
  reasoning: false,
5602
5602
  input: ["text"],
5603
5603
  cost: {
5604
- input: 30,
5605
- output: 60,
5604
+ input: 0.5,
5605
+ output: 1.5,
5606
5606
  cacheRead: 0,
5607
5607
  cacheWrite: 0,
5608
5608
  },
5609
- contextWindow: 8191,
5609
+ contextWindow: 16385,
5610
5610
  maxTokens: 4096,
5611
5611
  },
5612
5612
  "openrouter/auto": {