@mariozechner/pi-ai 0.11.4 → 0.11.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1971,6 +1971,57 @@ export const MODELS = {
1971
1971
  },
1972
1972
  },
1973
1973
  openrouter: {
1974
+ "arcee-ai/trinity-mini:free": {
1975
+ id: "arcee-ai/trinity-mini:free",
1976
+ name: "Arcee AI: Trinity Mini (free)",
1977
+ api: "openai-completions",
1978
+ provider: "openrouter",
1979
+ baseUrl: "https://openrouter.ai/api/v1",
1980
+ reasoning: true,
1981
+ input: ["text"],
1982
+ cost: {
1983
+ input: 0,
1984
+ output: 0,
1985
+ cacheRead: 0,
1986
+ cacheWrite: 0,
1987
+ },
1988
+ contextWindow: 131072,
1989
+ maxTokens: 4096,
1990
+ },
1991
+ "arcee-ai/trinity-mini": {
1992
+ id: "arcee-ai/trinity-mini",
1993
+ name: "Arcee AI: Trinity Mini",
1994
+ api: "openai-completions",
1995
+ provider: "openrouter",
1996
+ baseUrl: "https://openrouter.ai/api/v1",
1997
+ reasoning: true,
1998
+ input: ["text"],
1999
+ cost: {
2000
+ input: 0.045,
2001
+ output: 0.15,
2002
+ cacheRead: 0,
2003
+ cacheWrite: 0,
2004
+ },
2005
+ contextWindow: 131072,
2006
+ maxTokens: 4096,
2007
+ },
2008
+ "deepseek/deepseek-v3.2": {
2009
+ id: "deepseek/deepseek-v3.2",
2010
+ name: "DeepSeek: DeepSeek V3.2",
2011
+ api: "openai-completions",
2012
+ provider: "openrouter",
2013
+ baseUrl: "https://openrouter.ai/api/v1",
2014
+ reasoning: true,
2015
+ input: ["text"],
2016
+ cost: {
2017
+ input: 0.28,
2018
+ output: 0.39999999999999997,
2019
+ cacheRead: 0,
2020
+ cacheWrite: 0,
2021
+ },
2022
+ contextWindow: 163840,
2023
+ maxTokens: 65536,
2024
+ },
1974
2025
  "prime-intellect/intellect-3": {
1975
2026
  id: "prime-intellect/intellect-3",
1976
2027
  name: "Prime Intellect: INTELLECT-3",
@@ -2558,13 +2609,13 @@ export const MODELS = {
2558
2609
  reasoning: true,
2559
2610
  input: ["text"],
2560
2611
  cost: {
2561
- input: 0.216,
2562
- output: 0.328,
2563
- cacheRead: 0,
2612
+ input: 0.21,
2613
+ output: 0.32,
2614
+ cacheRead: 0.16799999999999998,
2564
2615
  cacheWrite: 0,
2565
2616
  },
2566
2617
  contextWindow: 163840,
2567
- maxTokens: 65536,
2618
+ maxTokens: 4096,
2568
2619
  },
2569
2620
  "google/gemini-2.5-flash-preview-09-2025": {
2570
2621
  id: "google/gemini-2.5-flash-preview-09-2025",
@@ -2685,26 +2736,26 @@ export const MODELS = {
2685
2736
  contextWindow: 400000,
2686
2737
  maxTokens: 128000,
2687
2738
  },
2688
- "deepseek/deepseek-v3.1-terminus:exacto": {
2689
- id: "deepseek/deepseek-v3.1-terminus:exacto",
2690
- name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
2739
+ "deepseek/deepseek-v3.1-terminus": {
2740
+ id: "deepseek/deepseek-v3.1-terminus",
2741
+ name: "DeepSeek: DeepSeek V3.1 Terminus",
2691
2742
  api: "openai-completions",
2692
2743
  provider: "openrouter",
2693
2744
  baseUrl: "https://openrouter.ai/api/v1",
2694
2745
  reasoning: true,
2695
2746
  input: ["text"],
2696
2747
  cost: {
2697
- input: 0.216,
2698
- output: 0.7999999999999999,
2699
- cacheRead: 0,
2748
+ input: 0.21,
2749
+ output: 0.7899999999999999,
2750
+ cacheRead: 0.16799999999999998,
2700
2751
  cacheWrite: 0,
2701
2752
  },
2702
- contextWindow: 131072,
2703
- maxTokens: 65536,
2753
+ contextWindow: 163840,
2754
+ maxTokens: 4096,
2704
2755
  },
2705
- "deepseek/deepseek-v3.1-terminus": {
2706
- id: "deepseek/deepseek-v3.1-terminus",
2707
- name: "DeepSeek: DeepSeek V3.1 Terminus",
2756
+ "deepseek/deepseek-v3.1-terminus:exacto": {
2757
+ id: "deepseek/deepseek-v3.1-terminus:exacto",
2758
+ name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
2708
2759
  api: "openai-completions",
2709
2760
  provider: "openrouter",
2710
2761
  baseUrl: "https://openrouter.ai/api/v1",
@@ -4311,7 +4362,7 @@ export const MODELS = {
4311
4362
  cost: {
4312
4363
  input: 0.19999999999999998,
4313
4364
  output: 0.88,
4314
- cacheRead: 0.135,
4365
+ cacheRead: 0.106,
4315
4366
  cacheWrite: 0,
4316
4367
  },
4317
4368
  contextWindow: 163840,
@@ -4904,10 +4955,10 @@ export const MODELS = {
4904
4955
  reasoning: false,
4905
4956
  input: ["text", "image"],
4906
4957
  cost: {
4907
- input: 3,
4908
- output: 15,
4909
- cacheRead: 0.3,
4910
- cacheWrite: 3.75,
4958
+ input: 6,
4959
+ output: 30,
4960
+ cacheRead: 0,
4961
+ cacheWrite: 0,
4911
4962
  },
4912
4963
  contextWindow: 200000,
4913
4964
  maxTokens: 8192,
@@ -5031,34 +5082,34 @@ export const MODELS = {
5031
5082
  contextWindow: 32768,
5032
5083
  maxTokens: 4096,
5033
5084
  },
5034
- "cohere/command-r-plus-08-2024": {
5035
- id: "cohere/command-r-plus-08-2024",
5036
- name: "Cohere: Command R+ (08-2024)",
5085
+ "cohere/command-r-08-2024": {
5086
+ id: "cohere/command-r-08-2024",
5087
+ name: "Cohere: Command R (08-2024)",
5037
5088
  api: "openai-completions",
5038
5089
  provider: "openrouter",
5039
5090
  baseUrl: "https://openrouter.ai/api/v1",
5040
5091
  reasoning: false,
5041
5092
  input: ["text"],
5042
5093
  cost: {
5043
- input: 2.5,
5044
- output: 10,
5094
+ input: 0.15,
5095
+ output: 0.6,
5045
5096
  cacheRead: 0,
5046
5097
  cacheWrite: 0,
5047
5098
  },
5048
5099
  contextWindow: 128000,
5049
5100
  maxTokens: 4000,
5050
5101
  },
5051
- "cohere/command-r-08-2024": {
5052
- id: "cohere/command-r-08-2024",
5053
- name: "Cohere: Command R (08-2024)",
5102
+ "cohere/command-r-plus-08-2024": {
5103
+ id: "cohere/command-r-plus-08-2024",
5104
+ name: "Cohere: Command R+ (08-2024)",
5054
5105
  api: "openai-completions",
5055
5106
  provider: "openrouter",
5056
5107
  baseUrl: "https://openrouter.ai/api/v1",
5057
5108
  reasoning: false,
5058
5109
  input: ["text"],
5059
5110
  cost: {
5060
- input: 0.15,
5061
- output: 0.6,
5111
+ input: 2.5,
5112
+ output: 10,
5062
5113
  cacheRead: 0,
5063
5114
  cacheWrite: 0,
5064
5115
  },
@@ -5184,9 +5235,9 @@ export const MODELS = {
5184
5235
  contextWindow: 131072,
5185
5236
  maxTokens: 16384,
5186
5237
  },
5187
- "openai/gpt-4o-mini": {
5188
- id: "openai/gpt-4o-mini",
5189
- name: "OpenAI: GPT-4o-mini",
5238
+ "openai/gpt-4o-mini-2024-07-18": {
5239
+ id: "openai/gpt-4o-mini-2024-07-18",
5240
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5190
5241
  api: "openai-completions",
5191
5242
  provider: "openrouter",
5192
5243
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5201,9 +5252,9 @@ export const MODELS = {
5201
5252
  contextWindow: 128000,
5202
5253
  maxTokens: 16384,
5203
5254
  },
5204
- "openai/gpt-4o-mini-2024-07-18": {
5205
- id: "openai/gpt-4o-mini-2024-07-18",
5206
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5255
+ "openai/gpt-4o-mini": {
5256
+ id: "openai/gpt-4o-mini",
5257
+ name: "OpenAI: GPT-4o-mini",
5207
5258
  api: "openai-completions",
5208
5259
  provider: "openrouter",
5209
5260
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5303,6 +5354,23 @@ export const MODELS = {
5303
5354
  contextWindow: 128000,
5304
5355
  maxTokens: 4096,
5305
5356
  },
5357
+ "openai/gpt-4o-2024-05-13": {
5358
+ id: "openai/gpt-4o-2024-05-13",
5359
+ name: "OpenAI: GPT-4o (2024-05-13)",
5360
+ api: "openai-completions",
5361
+ provider: "openrouter",
5362
+ baseUrl: "https://openrouter.ai/api/v1",
5363
+ reasoning: false,
5364
+ input: ["text", "image"],
5365
+ cost: {
5366
+ input: 5,
5367
+ output: 15,
5368
+ cacheRead: 0,
5369
+ cacheWrite: 0,
5370
+ },
5371
+ contextWindow: 128000,
5372
+ maxTokens: 4096,
5373
+ },
5306
5374
  "openai/gpt-4o": {
5307
5375
  id: "openai/gpt-4o",
5308
5376
  name: "OpenAI: GPT-4o",
@@ -5337,22 +5405,22 @@ export const MODELS = {
5337
5405
  contextWindow: 128000,
5338
5406
  maxTokens: 64000,
5339
5407
  },
5340
- "openai/gpt-4o-2024-05-13": {
5341
- id: "openai/gpt-4o-2024-05-13",
5342
- name: "OpenAI: GPT-4o (2024-05-13)",
5408
+ "meta-llama/llama-3-70b-instruct": {
5409
+ id: "meta-llama/llama-3-70b-instruct",
5410
+ name: "Meta: Llama 3 70B Instruct",
5343
5411
  api: "openai-completions",
5344
5412
  provider: "openrouter",
5345
5413
  baseUrl: "https://openrouter.ai/api/v1",
5346
5414
  reasoning: false,
5347
- input: ["text", "image"],
5415
+ input: ["text"],
5348
5416
  cost: {
5349
- input: 5,
5350
- output: 15,
5417
+ input: 0.3,
5418
+ output: 0.39999999999999997,
5351
5419
  cacheRead: 0,
5352
5420
  cacheWrite: 0,
5353
5421
  },
5354
- contextWindow: 128000,
5355
- maxTokens: 4096,
5422
+ contextWindow: 8192,
5423
+ maxTokens: 16384,
5356
5424
  },
5357
5425
  "meta-llama/llama-3-8b-instruct": {
5358
5426
  id: "meta-llama/llama-3-8b-instruct",
@@ -5371,23 +5439,6 @@ export const MODELS = {
5371
5439
  contextWindow: 8192,
5372
5440
  maxTokens: 16384,
5373
5441
  },
5374
- "meta-llama/llama-3-70b-instruct": {
5375
- id: "meta-llama/llama-3-70b-instruct",
5376
- name: "Meta: Llama 3 70B Instruct",
5377
- api: "openai-completions",
5378
- provider: "openrouter",
5379
- baseUrl: "https://openrouter.ai/api/v1",
5380
- reasoning: false,
5381
- input: ["text"],
5382
- cost: {
5383
- input: 0.3,
5384
- output: 0.39999999999999997,
5385
- cacheRead: 0,
5386
- cacheWrite: 0,
5387
- },
5388
- contextWindow: 8192,
5389
- maxTokens: 16384,
5390
- },
5391
5442
  "mistralai/mixtral-8x22b-instruct": {
5392
5443
  id: "mistralai/mixtral-8x22b-instruct",
5393
5444
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -5473,38 +5524,38 @@ export const MODELS = {
5473
5524
  contextWindow: 128000,
5474
5525
  maxTokens: 4096,
5475
5526
  },
5476
- "openai/gpt-4-turbo-preview": {
5477
- id: "openai/gpt-4-turbo-preview",
5478
- name: "OpenAI: GPT-4 Turbo Preview",
5527
+ "openai/gpt-3.5-turbo-0613": {
5528
+ id: "openai/gpt-3.5-turbo-0613",
5529
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5479
5530
  api: "openai-completions",
5480
5531
  provider: "openrouter",
5481
5532
  baseUrl: "https://openrouter.ai/api/v1",
5482
5533
  reasoning: false,
5483
5534
  input: ["text"],
5484
5535
  cost: {
5485
- input: 10,
5486
- output: 30,
5536
+ input: 1,
5537
+ output: 2,
5487
5538
  cacheRead: 0,
5488
5539
  cacheWrite: 0,
5489
5540
  },
5490
- contextWindow: 128000,
5541
+ contextWindow: 4095,
5491
5542
  maxTokens: 4096,
5492
5543
  },
5493
- "openai/gpt-3.5-turbo-0613": {
5494
- id: "openai/gpt-3.5-turbo-0613",
5495
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
5544
+ "openai/gpt-4-turbo-preview": {
5545
+ id: "openai/gpt-4-turbo-preview",
5546
+ name: "OpenAI: GPT-4 Turbo Preview",
5496
5547
  api: "openai-completions",
5497
5548
  provider: "openrouter",
5498
5549
  baseUrl: "https://openrouter.ai/api/v1",
5499
5550
  reasoning: false,
5500
5551
  input: ["text"],
5501
5552
  cost: {
5502
- input: 1,
5503
- output: 2,
5553
+ input: 10,
5554
+ output: 30,
5504
5555
  cacheRead: 0,
5505
5556
  cacheWrite: 0,
5506
5557
  },
5507
- contextWindow: 4095,
5558
+ contextWindow: 128000,
5508
5559
  maxTokens: 4096,
5509
5560
  },
5510
5561
  "mistralai/mistral-small": {
@@ -5609,38 +5660,38 @@ export const MODELS = {
5609
5660
  contextWindow: 8191,
5610
5661
  maxTokens: 4096,
5611
5662
  },
5612
- "openai/gpt-3.5-turbo": {
5613
- id: "openai/gpt-3.5-turbo",
5614
- name: "OpenAI: GPT-3.5 Turbo",
5663
+ "openai/gpt-4": {
5664
+ id: "openai/gpt-4",
5665
+ name: "OpenAI: GPT-4",
5615
5666
  api: "openai-completions",
5616
5667
  provider: "openrouter",
5617
5668
  baseUrl: "https://openrouter.ai/api/v1",
5618
5669
  reasoning: false,
5619
5670
  input: ["text"],
5620
5671
  cost: {
5621
- input: 0.5,
5622
- output: 1.5,
5672
+ input: 30,
5673
+ output: 60,
5623
5674
  cacheRead: 0,
5624
5675
  cacheWrite: 0,
5625
5676
  },
5626
- contextWindow: 16385,
5677
+ contextWindow: 8191,
5627
5678
  maxTokens: 4096,
5628
5679
  },
5629
- "openai/gpt-4": {
5630
- id: "openai/gpt-4",
5631
- name: "OpenAI: GPT-4",
5680
+ "openai/gpt-3.5-turbo": {
5681
+ id: "openai/gpt-3.5-turbo",
5682
+ name: "OpenAI: GPT-3.5 Turbo",
5632
5683
  api: "openai-completions",
5633
5684
  provider: "openrouter",
5634
5685
  baseUrl: "https://openrouter.ai/api/v1",
5635
5686
  reasoning: false,
5636
5687
  input: ["text"],
5637
5688
  cost: {
5638
- input: 30,
5639
- output: 60,
5689
+ input: 0.5,
5690
+ output: 1.5,
5640
5691
  cacheRead: 0,
5641
5692
  cacheWrite: 0,
5642
5693
  },
5643
- contextWindow: 8191,
5694
+ contextWindow: 16385,
5644
5695
  maxTokens: 4096,
5645
5696
  },
5646
5697
  "openrouter/auto": {