@mariozechner/pi-ai 0.10.0 → 0.10.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -138,6 +138,23 @@ export const MODELS = {
138
138
  contextWindow: 200000,
139
139
  maxTokens: 4096,
140
140
  },
141
+ "claude-opus-4-5-20251101": {
142
+ id: "claude-opus-4-5-20251101",
143
+ name: "Claude Opus 4.5",
144
+ api: "anthropic-messages",
145
+ provider: "anthropic",
146
+ baseUrl: "https://api.anthropic.com",
147
+ reasoning: true,
148
+ input: ["text", "image"],
149
+ cost: {
150
+ input: 5,
151
+ output: 25,
152
+ cacheRead: 0.5,
153
+ cacheWrite: 6.25,
154
+ },
155
+ contextWindow: 200000,
156
+ maxTokens: 64000,
157
+ },
141
158
  "claude-sonnet-4-5": {
142
159
  id: "claude-sonnet-4-5",
143
160
  name: "Claude Sonnet 4.5 (latest)",
@@ -1954,6 +1971,23 @@ export const MODELS = {
1954
1971
  },
1955
1972
  },
1956
1973
  openrouter: {
1974
+ "prime-intellect/intellect-3": {
1975
+ id: "prime-intellect/intellect-3",
1976
+ name: "Prime Intellect: INTELLECT-3",
1977
+ api: "openai-completions",
1978
+ provider: "openrouter",
1979
+ baseUrl: "https://openrouter.ai/api/v1",
1980
+ reasoning: true,
1981
+ input: ["text"],
1982
+ cost: {
1983
+ input: 0.19999999999999998,
1984
+ output: 1.1,
1985
+ cacheRead: 0,
1986
+ cacheWrite: 0,
1987
+ },
1988
+ contextWindow: 131072,
1989
+ maxTokens: 131072,
1990
+ },
1957
1991
  "tngtech/tng-r1t-chimera:free": {
1958
1992
  id: "tngtech/tng-r1t-chimera:free",
1959
1993
  name: "TNG: R1T Chimera (free)",
@@ -2003,7 +2037,7 @@ export const MODELS = {
2003
2037
  cacheWrite: 6.25,
2004
2038
  },
2005
2039
  contextWindow: 200000,
2006
- maxTokens: 32000,
2040
+ maxTokens: 64000,
2007
2041
  },
2008
2042
  "allenai/olmo-3-7b-instruct": {
2009
2043
  id: "allenai/olmo-3-7b-instruct",
@@ -2235,8 +2269,8 @@ export const MODELS = {
2235
2269
  reasoning: true,
2236
2270
  input: ["text"],
2237
2271
  cost: {
2238
- input: 0.24,
2239
- output: 0.96,
2272
+ input: 0.255,
2273
+ output: 1.02,
2240
2274
  cacheRead: 0,
2241
2275
  cacheWrite: 0,
2242
2276
  },
@@ -4980,34 +5014,34 @@ export const MODELS = {
4980
5014
  contextWindow: 32768,
4981
5015
  maxTokens: 4096,
4982
5016
  },
4983
- "cohere/command-r-08-2024": {
4984
- id: "cohere/command-r-08-2024",
4985
- name: "Cohere: Command R (08-2024)",
5017
+ "cohere/command-r-plus-08-2024": {
5018
+ id: "cohere/command-r-plus-08-2024",
5019
+ name: "Cohere: Command R+ (08-2024)",
4986
5020
  api: "openai-completions",
4987
5021
  provider: "openrouter",
4988
5022
  baseUrl: "https://openrouter.ai/api/v1",
4989
5023
  reasoning: false,
4990
5024
  input: ["text"],
4991
5025
  cost: {
4992
- input: 0.15,
4993
- output: 0.6,
5026
+ input: 2.5,
5027
+ output: 10,
4994
5028
  cacheRead: 0,
4995
5029
  cacheWrite: 0,
4996
5030
  },
4997
5031
  contextWindow: 128000,
4998
5032
  maxTokens: 4000,
4999
5033
  },
5000
- "cohere/command-r-plus-08-2024": {
5001
- id: "cohere/command-r-plus-08-2024",
5002
- name: "Cohere: Command R+ (08-2024)",
5034
+ "cohere/command-r-08-2024": {
5035
+ id: "cohere/command-r-08-2024",
5036
+ name: "Cohere: Command R (08-2024)",
5003
5037
  api: "openai-completions",
5004
5038
  provider: "openrouter",
5005
5039
  baseUrl: "https://openrouter.ai/api/v1",
5006
5040
  reasoning: false,
5007
5041
  input: ["text"],
5008
5042
  cost: {
5009
- input: 2.5,
5010
- output: 10,
5043
+ input: 0.15,
5044
+ output: 0.6,
5011
5045
  cacheRead: 0,
5012
5046
  cacheWrite: 0,
5013
5047
  },
@@ -5150,9 +5184,9 @@ export const MODELS = {
5150
5184
  contextWindow: 131072,
5151
5185
  maxTokens: 16384,
5152
5186
  },
5153
- "openai/gpt-4o-mini-2024-07-18": {
5154
- id: "openai/gpt-4o-mini-2024-07-18",
5155
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5187
+ "openai/gpt-4o-mini": {
5188
+ id: "openai/gpt-4o-mini",
5189
+ name: "OpenAI: GPT-4o-mini",
5156
5190
  api: "openai-completions",
5157
5191
  provider: "openrouter",
5158
5192
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5167,9 +5201,9 @@ export const MODELS = {
5167
5201
  contextWindow: 128000,
5168
5202
  maxTokens: 16384,
5169
5203
  },
5170
- "openai/gpt-4o-mini": {
5171
- id: "openai/gpt-4o-mini",
5172
- name: "OpenAI: GPT-4o-mini",
5204
+ "openai/gpt-4o-mini-2024-07-18": {
5205
+ id: "openai/gpt-4o-mini-2024-07-18",
5206
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5173
5207
  api: "openai-completions",
5174
5208
  provider: "openrouter",
5175
5209
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5269,23 +5303,6 @@ export const MODELS = {
5269
5303
  contextWindow: 128000,
5270
5304
  maxTokens: 4096,
5271
5305
  },
5272
- "openai/gpt-4o-2024-05-13": {
5273
- id: "openai/gpt-4o-2024-05-13",
5274
- name: "OpenAI: GPT-4o (2024-05-13)",
5275
- api: "openai-completions",
5276
- provider: "openrouter",
5277
- baseUrl: "https://openrouter.ai/api/v1",
5278
- reasoning: false,
5279
- input: ["text", "image"],
5280
- cost: {
5281
- input: 5,
5282
- output: 15,
5283
- cacheRead: 0,
5284
- cacheWrite: 0,
5285
- },
5286
- contextWindow: 128000,
5287
- maxTokens: 4096,
5288
- },
5289
5306
  "openai/gpt-4o": {
5290
5307
  id: "openai/gpt-4o",
5291
5308
  name: "OpenAI: GPT-4o",
@@ -5320,22 +5337,22 @@ export const MODELS = {
5320
5337
  contextWindow: 128000,
5321
5338
  maxTokens: 64000,
5322
5339
  },
5323
- "meta-llama/llama-3-70b-instruct": {
5324
- id: "meta-llama/llama-3-70b-instruct",
5325
- name: "Meta: Llama 3 70B Instruct",
5340
+ "openai/gpt-4o-2024-05-13": {
5341
+ id: "openai/gpt-4o-2024-05-13",
5342
+ name: "OpenAI: GPT-4o (2024-05-13)",
5326
5343
  api: "openai-completions",
5327
5344
  provider: "openrouter",
5328
5345
  baseUrl: "https://openrouter.ai/api/v1",
5329
5346
  reasoning: false,
5330
- input: ["text"],
5347
+ input: ["text", "image"],
5331
5348
  cost: {
5332
- input: 0.3,
5333
- output: 0.39999999999999997,
5349
+ input: 5,
5350
+ output: 15,
5334
5351
  cacheRead: 0,
5335
5352
  cacheWrite: 0,
5336
5353
  },
5337
- contextWindow: 8192,
5338
- maxTokens: 16384,
5354
+ contextWindow: 128000,
5355
+ maxTokens: 4096,
5339
5356
  },
5340
5357
  "meta-llama/llama-3-8b-instruct": {
5341
5358
  id: "meta-llama/llama-3-8b-instruct",
@@ -5354,6 +5371,23 @@ export const MODELS = {
5354
5371
  contextWindow: 8192,
5355
5372
  maxTokens: 16384,
5356
5373
  },
5374
+ "meta-llama/llama-3-70b-instruct": {
5375
+ id: "meta-llama/llama-3-70b-instruct",
5376
+ name: "Meta: Llama 3 70B Instruct",
5377
+ api: "openai-completions",
5378
+ provider: "openrouter",
5379
+ baseUrl: "https://openrouter.ai/api/v1",
5380
+ reasoning: false,
5381
+ input: ["text"],
5382
+ cost: {
5383
+ input: 0.3,
5384
+ output: 0.39999999999999997,
5385
+ cacheRead: 0,
5386
+ cacheWrite: 0,
5387
+ },
5388
+ contextWindow: 8192,
5389
+ maxTokens: 16384,
5390
+ },
5357
5391
  "mistralai/mixtral-8x22b-instruct": {
5358
5392
  id: "mistralai/mixtral-8x22b-instruct",
5359
5393
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -5439,38 +5473,38 @@ export const MODELS = {
5439
5473
  contextWindow: 128000,
5440
5474
  maxTokens: 4096,
5441
5475
  },
5442
- "openai/gpt-3.5-turbo-0613": {
5443
- id: "openai/gpt-3.5-turbo-0613",
5444
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5476
+ "openai/gpt-4-turbo-preview": {
5477
+ id: "openai/gpt-4-turbo-preview",
5478
+ name: "OpenAI: GPT-4 Turbo Preview",
5445
5479
  api: "openai-completions",
5446
5480
  provider: "openrouter",
5447
5481
  baseUrl: "https://openrouter.ai/api/v1",
5448
5482
  reasoning: false,
5449
5483
  input: ["text"],
5450
5484
  cost: {
5451
- input: 1,
5452
- output: 2,
5485
+ input: 10,
5486
+ output: 30,
5453
5487
  cacheRead: 0,
5454
5488
  cacheWrite: 0,
5455
5489
  },
5456
- contextWindow: 4095,
5490
+ contextWindow: 128000,
5457
5491
  maxTokens: 4096,
5458
5492
  },
5459
- "openai/gpt-4-turbo-preview": {
5460
- id: "openai/gpt-4-turbo-preview",
5461
- name: "OpenAI: GPT-4 Turbo Preview",
5493
+ "openai/gpt-3.5-turbo-0613": {
5494
+ id: "openai/gpt-3.5-turbo-0613",
5495
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5462
5496
  api: "openai-completions",
5463
5497
  provider: "openrouter",
5464
5498
  baseUrl: "https://openrouter.ai/api/v1",
5465
5499
  reasoning: false,
5466
5500
  input: ["text"],
5467
5501
  cost: {
5468
- input: 10,
5469
- output: 30,
5502
+ input: 1,
5503
+ output: 2,
5470
5504
  cacheRead: 0,
5471
5505
  cacheWrite: 0,
5472
5506
  },
5473
- contextWindow: 128000,
5507
+ contextWindow: 4095,
5474
5508
  maxTokens: 4096,
5475
5509
  },
5476
5510
  "mistralai/mistral-small": {
@@ -5575,38 +5609,38 @@ export const MODELS = {
5575
5609
  contextWindow: 8191,
5576
5610
  maxTokens: 4096,
5577
5611
  },
5578
- "openai/gpt-4": {
5579
- id: "openai/gpt-4",
5580
- name: "OpenAI: GPT-4",
5612
+ "openai/gpt-3.5-turbo": {
5613
+ id: "openai/gpt-3.5-turbo",
5614
+ name: "OpenAI: GPT-3.5 Turbo",
5581
5615
  api: "openai-completions",
5582
5616
  provider: "openrouter",
5583
5617
  baseUrl: "https://openrouter.ai/api/v1",
5584
5618
  reasoning: false,
5585
5619
  input: ["text"],
5586
5620
  cost: {
5587
- input: 30,
5588
- output: 60,
5621
+ input: 0.5,
5622
+ output: 1.5,
5589
5623
  cacheRead: 0,
5590
5624
  cacheWrite: 0,
5591
5625
  },
5592
- contextWindow: 8191,
5626
+ contextWindow: 16385,
5593
5627
  maxTokens: 4096,
5594
5628
  },
5595
- "openai/gpt-3.5-turbo": {
5596
- id: "openai/gpt-3.5-turbo",
5597
- name: "OpenAI: GPT-3.5 Turbo",
5629
+ "openai/gpt-4": {
5630
+ id: "openai/gpt-4",
5631
+ name: "OpenAI: GPT-4",
5598
5632
  api: "openai-completions",
5599
5633
  provider: "openrouter",
5600
5634
  baseUrl: "https://openrouter.ai/api/v1",
5601
5635
  reasoning: false,
5602
5636
  input: ["text"],
5603
5637
  cost: {
5604
- input: 0.5,
5605
- output: 1.5,
5638
+ input: 30,
5639
+ output: 60,
5606
5640
  cacheRead: 0,
5607
5641
  cacheWrite: 0,
5608
5642
  },
5609
- contextWindow: 16385,
5643
+ contextWindow: 8191,
5610
5644
  maxTokens: 4096,
5611
5645
  },
5612
5646
  "openrouter/auto": {