@mariozechner/pi-ai 0.12.3 → 0.12.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1971,6 +1971,57 @@ export const MODELS = {
1971
1971
  },
1972
1972
  },
1973
1973
  openrouter: {
1974
+ "amazon/nova-2-lite-v1:free": {
1975
+ id: "amazon/nova-2-lite-v1:free",
1976
+ name: "Amazon: Nova 2 Lite (free)",
1977
+ api: "openai-completions",
1978
+ provider: "openrouter",
1979
+ baseUrl: "https://openrouter.ai/api/v1",
1980
+ reasoning: true,
1981
+ input: ["text", "image"],
1982
+ cost: {
1983
+ input: 0,
1984
+ output: 0,
1985
+ cacheRead: 0,
1986
+ cacheWrite: 0,
1987
+ },
1988
+ contextWindow: 1000000,
1989
+ maxTokens: 4096,
1990
+ },
1991
+ "amazon/nova-2-lite-v1": {
1992
+ id: "amazon/nova-2-lite-v1",
1993
+ name: "Amazon: Nova 2 Lite",
1994
+ api: "openai-completions",
1995
+ provider: "openrouter",
1996
+ baseUrl: "https://openrouter.ai/api/v1",
1997
+ reasoning: true,
1998
+ input: ["text", "image"],
1999
+ cost: {
2000
+ input: 0.3,
2001
+ output: 2.5,
2002
+ cacheRead: 0,
2003
+ cacheWrite: 0,
2004
+ },
2005
+ contextWindow: 1000000,
2006
+ maxTokens: 4096,
2007
+ },
2008
+ "mistralai/mistral-large-2512": {
2009
+ id: "mistralai/mistral-large-2512",
2010
+ name: "Mistral: Mistral Large 3 2512",
2011
+ api: "openai-completions",
2012
+ provider: "openrouter",
2013
+ baseUrl: "https://openrouter.ai/api/v1",
2014
+ reasoning: false,
2015
+ input: ["text", "image"],
2016
+ cost: {
2017
+ input: 0.5,
2018
+ output: 1.5,
2019
+ cacheRead: 0,
2020
+ cacheWrite: 0,
2021
+ },
2022
+ contextWindow: 262144,
2023
+ maxTokens: 4096,
2024
+ },
1974
2025
  "arcee-ai/trinity-mini:free": {
1975
2026
  id: "arcee-ai/trinity-mini:free",
1976
2027
  name: "Arcee AI: Trinity Mini (free)",
@@ -2014,13 +2065,13 @@ export const MODELS = {
2014
2065
  reasoning: true,
2015
2066
  input: ["text"],
2016
2067
  cost: {
2017
- input: 0.28,
2068
+ input: 0.27,
2018
2069
  output: 0.39999999999999997,
2019
- cacheRead: 0,
2070
+ cacheRead: 0.216,
2020
2071
  cacheWrite: 0,
2021
2072
  },
2022
2073
  contextWindow: 163840,
2023
- maxTokens: 65536,
2074
+ maxTokens: 4096,
2024
2075
  },
2025
2076
  "prime-intellect/intellect-3": {
2026
2077
  id: "prime-intellect/intellect-3",
@@ -2090,23 +2141,6 @@ export const MODELS = {
2090
2141
  contextWindow: 200000,
2091
2142
  maxTokens: 64000,
2092
2143
  },
2093
- "openrouter/bert-nebulon-alpha": {
2094
- id: "openrouter/bert-nebulon-alpha",
2095
- name: "Bert-Nebulon Alpha",
2096
- api: "openai-completions",
2097
- provider: "openrouter",
2098
- baseUrl: "https://openrouter.ai/api/v1",
2099
- reasoning: false,
2100
- input: ["text", "image"],
2101
- cost: {
2102
- input: 0,
2103
- output: 0,
2104
- cacheRead: 0,
2105
- cacheWrite: 0,
2106
- },
2107
- contextWindow: 256000,
2108
- maxTokens: 4096,
2109
- },
2110
2144
  "allenai/olmo-3-7b-instruct": {
2111
2145
  id: "allenai/olmo-3-7b-instruct",
2112
2146
  name: "AllenAI: Olmo 3 7B Instruct",
@@ -2575,13 +2609,13 @@ export const MODELS = {
2575
2609
  reasoning: true,
2576
2610
  input: ["text"],
2577
2611
  cost: {
2578
- input: 0.44,
2579
- output: 1.76,
2580
- cacheRead: 0,
2612
+ input: 0.43,
2613
+ output: 1.75,
2614
+ cacheRead: 0.0799999993,
2581
2615
  cacheWrite: 0,
2582
2616
  },
2583
- contextWindow: 204800,
2584
- maxTokens: 131072,
2617
+ contextWindow: 202752,
2618
+ maxTokens: 4096,
2585
2619
  },
2586
2620
  "anthropic/claude-sonnet-4.5": {
2587
2621
  id: "anthropic/claude-sonnet-4.5",
@@ -2677,13 +2711,13 @@ export const MODELS = {
2677
2711
  reasoning: false,
2678
2712
  input: ["text", "image"],
2679
2713
  cost: {
2680
- input: 0.21,
2681
- output: 1.9,
2714
+ input: 0.19999999999999998,
2715
+ output: 1.2,
2682
2716
  cacheRead: 0,
2683
2717
  cacheWrite: 0,
2684
2718
  },
2685
- contextWindow: 131072,
2686
- maxTokens: 32768,
2719
+ contextWindow: 262144,
2720
+ maxTokens: 4096,
2687
2721
  },
2688
2722
  "qwen/qwen3-max": {
2689
2723
  id: "qwen/qwen3-max",
@@ -2736,9 +2770,9 @@ export const MODELS = {
2736
2770
  contextWindow: 400000,
2737
2771
  maxTokens: 128000,
2738
2772
  },
2739
- "deepseek/deepseek-v3.1-terminus": {
2740
- id: "deepseek/deepseek-v3.1-terminus",
2741
- name: "DeepSeek: DeepSeek V3.1 Terminus",
2773
+ "deepseek/deepseek-v3.1-terminus:exacto": {
2774
+ id: "deepseek/deepseek-v3.1-terminus:exacto",
2775
+ name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
2742
2776
  api: "openai-completions",
2743
2777
  provider: "openrouter",
2744
2778
  baseUrl: "https://openrouter.ai/api/v1",
@@ -2753,22 +2787,22 @@ export const MODELS = {
2753
2787
  contextWindow: 163840,
2754
2788
  maxTokens: 4096,
2755
2789
  },
2756
- "deepseek/deepseek-v3.1-terminus:exacto": {
2757
- id: "deepseek/deepseek-v3.1-terminus:exacto",
2758
- name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
2790
+ "deepseek/deepseek-v3.1-terminus": {
2791
+ id: "deepseek/deepseek-v3.1-terminus",
2792
+ name: "DeepSeek: DeepSeek V3.1 Terminus",
2759
2793
  api: "openai-completions",
2760
2794
  provider: "openrouter",
2761
2795
  baseUrl: "https://openrouter.ai/api/v1",
2762
2796
  reasoning: true,
2763
2797
  input: ["text"],
2764
2798
  cost: {
2765
- input: 0.216,
2766
- output: 0.7999999999999999,
2767
- cacheRead: 0,
2799
+ input: 0.21,
2800
+ output: 0.7899999999999999,
2801
+ cacheRead: 0.16799999999999998,
2768
2802
  cacheWrite: 0,
2769
2803
  },
2770
- contextWindow: 131072,
2771
- maxTokens: 65536,
2804
+ contextWindow: 163840,
2805
+ maxTokens: 4096,
2772
2806
  },
2773
2807
  "x-ai/grok-4-fast": {
2774
2808
  id: "x-ai/grok-4-fast",
@@ -3595,13 +3629,13 @@ export const MODELS = {
3595
3629
  reasoning: true,
3596
3630
  input: ["text"],
3597
3631
  cost: {
3598
- input: 0.072,
3599
- output: 0.464,
3632
+ input: 0.071,
3633
+ output: 0.463,
3600
3634
  cacheRead: 0,
3601
3635
  cacheWrite: 0,
3602
3636
  },
3603
- contextWindow: 131072,
3604
- maxTokens: 16384,
3637
+ contextWindow: 262144,
3638
+ maxTokens: 4096,
3605
3639
  },
3606
3640
  "moonshotai/kimi-k2": {
3607
3641
  id: "moonshotai/kimi-k2",
@@ -4768,8 +4802,8 @@ export const MODELS = {
4768
4802
  reasoning: false,
4769
4803
  input: ["text"],
4770
4804
  cost: {
4771
- input: 0.10400000000000001,
4772
- output: 0.312,
4805
+ input: 0.108,
4806
+ output: 0.32,
4773
4807
  cacheRead: 0,
4774
4808
  cacheWrite: 0,
4775
4809
  },
@@ -5082,34 +5116,34 @@ export const MODELS = {
5082
5116
  contextWindow: 32768,
5083
5117
  maxTokens: 4096,
5084
5118
  },
5085
- "cohere/command-r-plus-08-2024": {
5086
- id: "cohere/command-r-plus-08-2024",
5087
- name: "Cohere: Command R+ (08-2024)",
5119
+ "cohere/command-r-08-2024": {
5120
+ id: "cohere/command-r-08-2024",
5121
+ name: "Cohere: Command R (08-2024)",
5088
5122
  api: "openai-completions",
5089
5123
  provider: "openrouter",
5090
5124
  baseUrl: "https://openrouter.ai/api/v1",
5091
5125
  reasoning: false,
5092
5126
  input: ["text"],
5093
5127
  cost: {
5094
- input: 2.5,
5095
- output: 10,
5128
+ input: 0.15,
5129
+ output: 0.6,
5096
5130
  cacheRead: 0,
5097
5131
  cacheWrite: 0,
5098
5132
  },
5099
5133
  contextWindow: 128000,
5100
5134
  maxTokens: 4000,
5101
5135
  },
5102
- "cohere/command-r-08-2024": {
5103
- id: "cohere/command-r-08-2024",
5104
- name: "Cohere: Command R (08-2024)",
5136
+ "cohere/command-r-plus-08-2024": {
5137
+ id: "cohere/command-r-plus-08-2024",
5138
+ name: "Cohere: Command R+ (08-2024)",
5105
5139
  api: "openai-completions",
5106
5140
  provider: "openrouter",
5107
5141
  baseUrl: "https://openrouter.ai/api/v1",
5108
5142
  reasoning: false,
5109
5143
  input: ["text"],
5110
5144
  cost: {
5111
- input: 0.15,
5112
- output: 0.6,
5145
+ input: 2.5,
5146
+ output: 10,
5113
5147
  cacheRead: 0,
5114
5148
  cacheWrite: 0,
5115
5149
  },
@@ -5167,23 +5201,6 @@ export const MODELS = {
5167
5201
  contextWindow: 128000,
5168
5202
  maxTokens: 16384,
5169
5203
  },
5170
- "meta-llama/llama-3.1-70b-instruct": {
5171
- id: "meta-llama/llama-3.1-70b-instruct",
5172
- name: "Meta: Llama 3.1 70B Instruct",
5173
- api: "openai-completions",
5174
- provider: "openrouter",
5175
- baseUrl: "https://openrouter.ai/api/v1",
5176
- reasoning: false,
5177
- input: ["text"],
5178
- cost: {
5179
- input: 0.39999999999999997,
5180
- output: 0.39999999999999997,
5181
- cacheRead: 0,
5182
- cacheWrite: 0,
5183
- },
5184
- contextWindow: 131072,
5185
- maxTokens: 4096,
5186
- },
5187
5204
  "meta-llama/llama-3.1-8b-instruct": {
5188
5205
  id: "meta-llama/llama-3.1-8b-instruct",
5189
5206
  name: "Meta: Llama 3.1 8B Instruct",
@@ -5218,6 +5235,23 @@ export const MODELS = {
5218
5235
  contextWindow: 130815,
5219
5236
  maxTokens: 4096,
5220
5237
  },
5238
+ "meta-llama/llama-3.1-70b-instruct": {
5239
+ id: "meta-llama/llama-3.1-70b-instruct",
5240
+ name: "Meta: Llama 3.1 70B Instruct",
5241
+ api: "openai-completions",
5242
+ provider: "openrouter",
5243
+ baseUrl: "https://openrouter.ai/api/v1",
5244
+ reasoning: false,
5245
+ input: ["text"],
5246
+ cost: {
5247
+ input: 0.39999999999999997,
5248
+ output: 0.39999999999999997,
5249
+ cacheRead: 0,
5250
+ cacheWrite: 0,
5251
+ },
5252
+ contextWindow: 131072,
5253
+ maxTokens: 4096,
5254
+ },
5221
5255
  "mistralai/mistral-nemo": {
5222
5256
  id: "mistralai/mistral-nemo",
5223
5257
  name: "Mistral: Mistral Nemo",
@@ -5235,9 +5269,9 @@ export const MODELS = {
5235
5269
  contextWindow: 131072,
5236
5270
  maxTokens: 16384,
5237
5271
  },
5238
- "openai/gpt-4o-mini": {
5239
- id: "openai/gpt-4o-mini",
5240
- name: "OpenAI: GPT-4o-mini",
5272
+ "openai/gpt-4o-mini-2024-07-18": {
5273
+ id: "openai/gpt-4o-mini-2024-07-18",
5274
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5241
5275
  api: "openai-completions",
5242
5276
  provider: "openrouter",
5243
5277
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5252,9 +5286,9 @@ export const MODELS = {
5252
5286
  contextWindow: 128000,
5253
5287
  maxTokens: 16384,
5254
5288
  },
5255
- "openai/gpt-4o-mini-2024-07-18": {
5256
- id: "openai/gpt-4o-mini-2024-07-18",
5257
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5289
+ "openai/gpt-4o-mini": {
5290
+ id: "openai/gpt-4o-mini",
5291
+ name: "OpenAI: GPT-4o-mini",
5258
5292
  api: "openai-completions",
5259
5293
  provider: "openrouter",
5260
5294
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5643,26 +5677,26 @@ export const MODELS = {
5643
5677
  contextWindow: 16385,
5644
5678
  maxTokens: 4096,
5645
5679
  },
5646
- "openai/gpt-3.5-turbo": {
5647
- id: "openai/gpt-3.5-turbo",
5648
- name: "OpenAI: GPT-3.5 Turbo",
5680
+ "openai/gpt-4-0314": {
5681
+ id: "openai/gpt-4-0314",
5682
+ name: "OpenAI: GPT-4 (older v0314)",
5649
5683
  api: "openai-completions",
5650
5684
  provider: "openrouter",
5651
5685
  baseUrl: "https://openrouter.ai/api/v1",
5652
5686
  reasoning: false,
5653
5687
  input: ["text"],
5654
5688
  cost: {
5655
- input: 0.5,
5656
- output: 1.5,
5689
+ input: 30,
5690
+ output: 60,
5657
5691
  cacheRead: 0,
5658
5692
  cacheWrite: 0,
5659
5693
  },
5660
- contextWindow: 16385,
5694
+ contextWindow: 8191,
5661
5695
  maxTokens: 4096,
5662
5696
  },
5663
- "openai/gpt-4-0314": {
5664
- id: "openai/gpt-4-0314",
5665
- name: "OpenAI: GPT-4 (older v0314)",
5697
+ "openai/gpt-4": {
5698
+ id: "openai/gpt-4",
5699
+ name: "OpenAI: GPT-4",
5666
5700
  api: "openai-completions",
5667
5701
  provider: "openrouter",
5668
5702
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5677,21 +5711,21 @@ export const MODELS = {
5677
5711
  contextWindow: 8191,
5678
5712
  maxTokens: 4096,
5679
5713
  },
5680
- "openai/gpt-4": {
5681
- id: "openai/gpt-4",
5682
- name: "OpenAI: GPT-4",
5714
+ "openai/gpt-3.5-turbo": {
5715
+ id: "openai/gpt-3.5-turbo",
5716
+ name: "OpenAI: GPT-3.5 Turbo",
5683
5717
  api: "openai-completions",
5684
5718
  provider: "openrouter",
5685
5719
  baseUrl: "https://openrouter.ai/api/v1",
5686
5720
  reasoning: false,
5687
5721
  input: ["text"],
5688
5722
  cost: {
5689
- input: 30,
5690
- output: 60,
5723
+ input: 0.5,
5724
+ output: 1.5,
5691
5725
  cacheRead: 0,
5692
5726
  cacheWrite: 0,
5693
5727
  },
5694
- contextWindow: 8191,
5728
+ contextWindow: 16385,
5695
5729
  maxTokens: 4096,
5696
5730
  },
5697
5731
  "openrouter/auto": {