@mariozechner/pi-ai 0.12.3 → 0.12.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -1971,6 +1971,57 @@ export const MODELS = {
|
|
|
1971
1971
|
},
|
|
1972
1972
|
},
|
|
1973
1973
|
openrouter: {
|
|
1974
|
+
"amazon/nova-2-lite-v1:free": {
|
|
1975
|
+
id: "amazon/nova-2-lite-v1:free",
|
|
1976
|
+
name: "Amazon: Nova 2 Lite (free)",
|
|
1977
|
+
api: "openai-completions",
|
|
1978
|
+
provider: "openrouter",
|
|
1979
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1980
|
+
reasoning: true,
|
|
1981
|
+
input: ["text", "image"],
|
|
1982
|
+
cost: {
|
|
1983
|
+
input: 0,
|
|
1984
|
+
output: 0,
|
|
1985
|
+
cacheRead: 0,
|
|
1986
|
+
cacheWrite: 0,
|
|
1987
|
+
},
|
|
1988
|
+
contextWindow: 1000000,
|
|
1989
|
+
maxTokens: 4096,
|
|
1990
|
+
},
|
|
1991
|
+
"amazon/nova-2-lite-v1": {
|
|
1992
|
+
id: "amazon/nova-2-lite-v1",
|
|
1993
|
+
name: "Amazon: Nova 2 Lite",
|
|
1994
|
+
api: "openai-completions",
|
|
1995
|
+
provider: "openrouter",
|
|
1996
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1997
|
+
reasoning: true,
|
|
1998
|
+
input: ["text", "image"],
|
|
1999
|
+
cost: {
|
|
2000
|
+
input: 0.3,
|
|
2001
|
+
output: 2.5,
|
|
2002
|
+
cacheRead: 0,
|
|
2003
|
+
cacheWrite: 0,
|
|
2004
|
+
},
|
|
2005
|
+
contextWindow: 1000000,
|
|
2006
|
+
maxTokens: 4096,
|
|
2007
|
+
},
|
|
2008
|
+
"mistralai/mistral-large-2512": {
|
|
2009
|
+
id: "mistralai/mistral-large-2512",
|
|
2010
|
+
name: "Mistral: Mistral Large 3 2512",
|
|
2011
|
+
api: "openai-completions",
|
|
2012
|
+
provider: "openrouter",
|
|
2013
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2014
|
+
reasoning: false,
|
|
2015
|
+
input: ["text", "image"],
|
|
2016
|
+
cost: {
|
|
2017
|
+
input: 0.5,
|
|
2018
|
+
output: 1.5,
|
|
2019
|
+
cacheRead: 0,
|
|
2020
|
+
cacheWrite: 0,
|
|
2021
|
+
},
|
|
2022
|
+
contextWindow: 262144,
|
|
2023
|
+
maxTokens: 4096,
|
|
2024
|
+
},
|
|
1974
2025
|
"arcee-ai/trinity-mini:free": {
|
|
1975
2026
|
id: "arcee-ai/trinity-mini:free",
|
|
1976
2027
|
name: "Arcee AI: Trinity Mini (free)",
|
|
@@ -2014,13 +2065,13 @@ export const MODELS = {
|
|
|
2014
2065
|
reasoning: true,
|
|
2015
2066
|
input: ["text"],
|
|
2016
2067
|
cost: {
|
|
2017
|
-
input: 0.
|
|
2068
|
+
input: 0.27,
|
|
2018
2069
|
output: 0.39999999999999997,
|
|
2019
|
-
cacheRead: 0,
|
|
2070
|
+
cacheRead: 0.216,
|
|
2020
2071
|
cacheWrite: 0,
|
|
2021
2072
|
},
|
|
2022
2073
|
contextWindow: 163840,
|
|
2023
|
-
maxTokens:
|
|
2074
|
+
maxTokens: 4096,
|
|
2024
2075
|
},
|
|
2025
2076
|
"prime-intellect/intellect-3": {
|
|
2026
2077
|
id: "prime-intellect/intellect-3",
|
|
@@ -2090,23 +2141,6 @@ export const MODELS = {
|
|
|
2090
2141
|
contextWindow: 200000,
|
|
2091
2142
|
maxTokens: 64000,
|
|
2092
2143
|
},
|
|
2093
|
-
"openrouter/bert-nebulon-alpha": {
|
|
2094
|
-
id: "openrouter/bert-nebulon-alpha",
|
|
2095
|
-
name: "Bert-Nebulon Alpha",
|
|
2096
|
-
api: "openai-completions",
|
|
2097
|
-
provider: "openrouter",
|
|
2098
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
2099
|
-
reasoning: false,
|
|
2100
|
-
input: ["text", "image"],
|
|
2101
|
-
cost: {
|
|
2102
|
-
input: 0,
|
|
2103
|
-
output: 0,
|
|
2104
|
-
cacheRead: 0,
|
|
2105
|
-
cacheWrite: 0,
|
|
2106
|
-
},
|
|
2107
|
-
contextWindow: 256000,
|
|
2108
|
-
maxTokens: 4096,
|
|
2109
|
-
},
|
|
2110
2144
|
"allenai/olmo-3-7b-instruct": {
|
|
2111
2145
|
id: "allenai/olmo-3-7b-instruct",
|
|
2112
2146
|
name: "AllenAI: Olmo 3 7B Instruct",
|
|
@@ -2575,13 +2609,13 @@ export const MODELS = {
|
|
|
2575
2609
|
reasoning: true,
|
|
2576
2610
|
input: ["text"],
|
|
2577
2611
|
cost: {
|
|
2578
|
-
input: 0.
|
|
2579
|
-
output: 1.
|
|
2580
|
-
cacheRead: 0,
|
|
2612
|
+
input: 0.43,
|
|
2613
|
+
output: 1.75,
|
|
2614
|
+
cacheRead: 0.0799999993,
|
|
2581
2615
|
cacheWrite: 0,
|
|
2582
2616
|
},
|
|
2583
|
-
contextWindow:
|
|
2584
|
-
maxTokens:
|
|
2617
|
+
contextWindow: 202752,
|
|
2618
|
+
maxTokens: 4096,
|
|
2585
2619
|
},
|
|
2586
2620
|
"anthropic/claude-sonnet-4.5": {
|
|
2587
2621
|
id: "anthropic/claude-sonnet-4.5",
|
|
@@ -2677,13 +2711,13 @@ export const MODELS = {
|
|
|
2677
2711
|
reasoning: false,
|
|
2678
2712
|
input: ["text", "image"],
|
|
2679
2713
|
cost: {
|
|
2680
|
-
input: 0.
|
|
2681
|
-
output: 1.
|
|
2714
|
+
input: 0.19999999999999998,
|
|
2715
|
+
output: 1.2,
|
|
2682
2716
|
cacheRead: 0,
|
|
2683
2717
|
cacheWrite: 0,
|
|
2684
2718
|
},
|
|
2685
|
-
contextWindow:
|
|
2686
|
-
maxTokens:
|
|
2719
|
+
contextWindow: 262144,
|
|
2720
|
+
maxTokens: 4096,
|
|
2687
2721
|
},
|
|
2688
2722
|
"qwen/qwen3-max": {
|
|
2689
2723
|
id: "qwen/qwen3-max",
|
|
@@ -2736,9 +2770,9 @@ export const MODELS = {
|
|
|
2736
2770
|
contextWindow: 400000,
|
|
2737
2771
|
maxTokens: 128000,
|
|
2738
2772
|
},
|
|
2739
|
-
"deepseek/deepseek-v3.1-terminus": {
|
|
2740
|
-
id: "deepseek/deepseek-v3.1-terminus",
|
|
2741
|
-
name: "DeepSeek: DeepSeek V3.1 Terminus",
|
|
2773
|
+
"deepseek/deepseek-v3.1-terminus:exacto": {
|
|
2774
|
+
id: "deepseek/deepseek-v3.1-terminus:exacto",
|
|
2775
|
+
name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)",
|
|
2742
2776
|
api: "openai-completions",
|
|
2743
2777
|
provider: "openrouter",
|
|
2744
2778
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -2753,22 +2787,22 @@ export const MODELS = {
|
|
|
2753
2787
|
contextWindow: 163840,
|
|
2754
2788
|
maxTokens: 4096,
|
|
2755
2789
|
},
|
|
2756
|
-
"deepseek/deepseek-v3.1-terminus
|
|
2757
|
-
id: "deepseek/deepseek-v3.1-terminus
|
|
2758
|
-
name: "DeepSeek: DeepSeek V3.1 Terminus
|
|
2790
|
+
"deepseek/deepseek-v3.1-terminus": {
|
|
2791
|
+
id: "deepseek/deepseek-v3.1-terminus",
|
|
2792
|
+
name: "DeepSeek: DeepSeek V3.1 Terminus",
|
|
2759
2793
|
api: "openai-completions",
|
|
2760
2794
|
provider: "openrouter",
|
|
2761
2795
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
2762
2796
|
reasoning: true,
|
|
2763
2797
|
input: ["text"],
|
|
2764
2798
|
cost: {
|
|
2765
|
-
input: 0.
|
|
2766
|
-
output: 0.
|
|
2767
|
-
cacheRead: 0,
|
|
2799
|
+
input: 0.21,
|
|
2800
|
+
output: 0.7899999999999999,
|
|
2801
|
+
cacheRead: 0.16799999999999998,
|
|
2768
2802
|
cacheWrite: 0,
|
|
2769
2803
|
},
|
|
2770
|
-
contextWindow:
|
|
2771
|
-
maxTokens:
|
|
2804
|
+
contextWindow: 163840,
|
|
2805
|
+
maxTokens: 4096,
|
|
2772
2806
|
},
|
|
2773
2807
|
"x-ai/grok-4-fast": {
|
|
2774
2808
|
id: "x-ai/grok-4-fast",
|
|
@@ -3595,13 +3629,13 @@ export const MODELS = {
|
|
|
3595
3629
|
reasoning: true,
|
|
3596
3630
|
input: ["text"],
|
|
3597
3631
|
cost: {
|
|
3598
|
-
input: 0.
|
|
3599
|
-
output: 0.
|
|
3632
|
+
input: 0.071,
|
|
3633
|
+
output: 0.463,
|
|
3600
3634
|
cacheRead: 0,
|
|
3601
3635
|
cacheWrite: 0,
|
|
3602
3636
|
},
|
|
3603
|
-
contextWindow:
|
|
3604
|
-
maxTokens:
|
|
3637
|
+
contextWindow: 262144,
|
|
3638
|
+
maxTokens: 4096,
|
|
3605
3639
|
},
|
|
3606
3640
|
"moonshotai/kimi-k2": {
|
|
3607
3641
|
id: "moonshotai/kimi-k2",
|
|
@@ -4768,8 +4802,8 @@ export const MODELS = {
|
|
|
4768
4802
|
reasoning: false,
|
|
4769
4803
|
input: ["text"],
|
|
4770
4804
|
cost: {
|
|
4771
|
-
input: 0.
|
|
4772
|
-
output: 0.
|
|
4805
|
+
input: 0.108,
|
|
4806
|
+
output: 0.32,
|
|
4773
4807
|
cacheRead: 0,
|
|
4774
4808
|
cacheWrite: 0,
|
|
4775
4809
|
},
|
|
@@ -5082,34 +5116,34 @@ export const MODELS = {
|
|
|
5082
5116
|
contextWindow: 32768,
|
|
5083
5117
|
maxTokens: 4096,
|
|
5084
5118
|
},
|
|
5085
|
-
"cohere/command-r-
|
|
5086
|
-
id: "cohere/command-r-
|
|
5087
|
-
name: "Cohere: Command R
|
|
5119
|
+
"cohere/command-r-08-2024": {
|
|
5120
|
+
id: "cohere/command-r-08-2024",
|
|
5121
|
+
name: "Cohere: Command R (08-2024)",
|
|
5088
5122
|
api: "openai-completions",
|
|
5089
5123
|
provider: "openrouter",
|
|
5090
5124
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5091
5125
|
reasoning: false,
|
|
5092
5126
|
input: ["text"],
|
|
5093
5127
|
cost: {
|
|
5094
|
-
input:
|
|
5095
|
-
output:
|
|
5128
|
+
input: 0.15,
|
|
5129
|
+
output: 0.6,
|
|
5096
5130
|
cacheRead: 0,
|
|
5097
5131
|
cacheWrite: 0,
|
|
5098
5132
|
},
|
|
5099
5133
|
contextWindow: 128000,
|
|
5100
5134
|
maxTokens: 4000,
|
|
5101
5135
|
},
|
|
5102
|
-
"cohere/command-r-08-2024": {
|
|
5103
|
-
id: "cohere/command-r-08-2024",
|
|
5104
|
-
name: "Cohere: Command R (08-2024)",
|
|
5136
|
+
"cohere/command-r-plus-08-2024": {
|
|
5137
|
+
id: "cohere/command-r-plus-08-2024",
|
|
5138
|
+
name: "Cohere: Command R+ (08-2024)",
|
|
5105
5139
|
api: "openai-completions",
|
|
5106
5140
|
provider: "openrouter",
|
|
5107
5141
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5108
5142
|
reasoning: false,
|
|
5109
5143
|
input: ["text"],
|
|
5110
5144
|
cost: {
|
|
5111
|
-
input:
|
|
5112
|
-
output:
|
|
5145
|
+
input: 2.5,
|
|
5146
|
+
output: 10,
|
|
5113
5147
|
cacheRead: 0,
|
|
5114
5148
|
cacheWrite: 0,
|
|
5115
5149
|
},
|
|
@@ -5167,23 +5201,6 @@ export const MODELS = {
|
|
|
5167
5201
|
contextWindow: 128000,
|
|
5168
5202
|
maxTokens: 16384,
|
|
5169
5203
|
},
|
|
5170
|
-
"meta-llama/llama-3.1-70b-instruct": {
|
|
5171
|
-
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5172
|
-
name: "Meta: Llama 3.1 70B Instruct",
|
|
5173
|
-
api: "openai-completions",
|
|
5174
|
-
provider: "openrouter",
|
|
5175
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5176
|
-
reasoning: false,
|
|
5177
|
-
input: ["text"],
|
|
5178
|
-
cost: {
|
|
5179
|
-
input: 0.39999999999999997,
|
|
5180
|
-
output: 0.39999999999999997,
|
|
5181
|
-
cacheRead: 0,
|
|
5182
|
-
cacheWrite: 0,
|
|
5183
|
-
},
|
|
5184
|
-
contextWindow: 131072,
|
|
5185
|
-
maxTokens: 4096,
|
|
5186
|
-
},
|
|
5187
5204
|
"meta-llama/llama-3.1-8b-instruct": {
|
|
5188
5205
|
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5189
5206
|
name: "Meta: Llama 3.1 8B Instruct",
|
|
@@ -5218,6 +5235,23 @@ export const MODELS = {
|
|
|
5218
5235
|
contextWindow: 130815,
|
|
5219
5236
|
maxTokens: 4096,
|
|
5220
5237
|
},
|
|
5238
|
+
"meta-llama/llama-3.1-70b-instruct": {
|
|
5239
|
+
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5240
|
+
name: "Meta: Llama 3.1 70B Instruct",
|
|
5241
|
+
api: "openai-completions",
|
|
5242
|
+
provider: "openrouter",
|
|
5243
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5244
|
+
reasoning: false,
|
|
5245
|
+
input: ["text"],
|
|
5246
|
+
cost: {
|
|
5247
|
+
input: 0.39999999999999997,
|
|
5248
|
+
output: 0.39999999999999997,
|
|
5249
|
+
cacheRead: 0,
|
|
5250
|
+
cacheWrite: 0,
|
|
5251
|
+
},
|
|
5252
|
+
contextWindow: 131072,
|
|
5253
|
+
maxTokens: 4096,
|
|
5254
|
+
},
|
|
5221
5255
|
"mistralai/mistral-nemo": {
|
|
5222
5256
|
id: "mistralai/mistral-nemo",
|
|
5223
5257
|
name: "Mistral: Mistral Nemo",
|
|
@@ -5235,9 +5269,9 @@ export const MODELS = {
|
|
|
5235
5269
|
contextWindow: 131072,
|
|
5236
5270
|
maxTokens: 16384,
|
|
5237
5271
|
},
|
|
5238
|
-
"openai/gpt-4o-mini": {
|
|
5239
|
-
id: "openai/gpt-4o-mini",
|
|
5240
|
-
name: "OpenAI: GPT-4o-mini",
|
|
5272
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
5273
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
5274
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
5241
5275
|
api: "openai-completions",
|
|
5242
5276
|
provider: "openrouter",
|
|
5243
5277
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5252,9 +5286,9 @@ export const MODELS = {
|
|
|
5252
5286
|
contextWindow: 128000,
|
|
5253
5287
|
maxTokens: 16384,
|
|
5254
5288
|
},
|
|
5255
|
-
"openai/gpt-4o-mini
|
|
5256
|
-
id: "openai/gpt-4o-mini
|
|
5257
|
-
name: "OpenAI: GPT-4o-mini
|
|
5289
|
+
"openai/gpt-4o-mini": {
|
|
5290
|
+
id: "openai/gpt-4o-mini",
|
|
5291
|
+
name: "OpenAI: GPT-4o-mini",
|
|
5258
5292
|
api: "openai-completions",
|
|
5259
5293
|
provider: "openrouter",
|
|
5260
5294
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5643,26 +5677,26 @@ export const MODELS = {
|
|
|
5643
5677
|
contextWindow: 16385,
|
|
5644
5678
|
maxTokens: 4096,
|
|
5645
5679
|
},
|
|
5646
|
-
"openai/gpt-
|
|
5647
|
-
id: "openai/gpt-
|
|
5648
|
-
name: "OpenAI: GPT-
|
|
5680
|
+
"openai/gpt-4-0314": {
|
|
5681
|
+
id: "openai/gpt-4-0314",
|
|
5682
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
5649
5683
|
api: "openai-completions",
|
|
5650
5684
|
provider: "openrouter",
|
|
5651
5685
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5652
5686
|
reasoning: false,
|
|
5653
5687
|
input: ["text"],
|
|
5654
5688
|
cost: {
|
|
5655
|
-
input:
|
|
5656
|
-
output:
|
|
5689
|
+
input: 30,
|
|
5690
|
+
output: 60,
|
|
5657
5691
|
cacheRead: 0,
|
|
5658
5692
|
cacheWrite: 0,
|
|
5659
5693
|
},
|
|
5660
|
-
contextWindow:
|
|
5694
|
+
contextWindow: 8191,
|
|
5661
5695
|
maxTokens: 4096,
|
|
5662
5696
|
},
|
|
5663
|
-
"openai/gpt-4
|
|
5664
|
-
id: "openai/gpt-4
|
|
5665
|
-
name: "OpenAI: GPT-4
|
|
5697
|
+
"openai/gpt-4": {
|
|
5698
|
+
id: "openai/gpt-4",
|
|
5699
|
+
name: "OpenAI: GPT-4",
|
|
5666
5700
|
api: "openai-completions",
|
|
5667
5701
|
provider: "openrouter",
|
|
5668
5702
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5677,21 +5711,21 @@ export const MODELS = {
|
|
|
5677
5711
|
contextWindow: 8191,
|
|
5678
5712
|
maxTokens: 4096,
|
|
5679
5713
|
},
|
|
5680
|
-
"openai/gpt-
|
|
5681
|
-
id: "openai/gpt-
|
|
5682
|
-
name: "OpenAI: GPT-
|
|
5714
|
+
"openai/gpt-3.5-turbo": {
|
|
5715
|
+
id: "openai/gpt-3.5-turbo",
|
|
5716
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
5683
5717
|
api: "openai-completions",
|
|
5684
5718
|
provider: "openrouter",
|
|
5685
5719
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5686
5720
|
reasoning: false,
|
|
5687
5721
|
input: ["text"],
|
|
5688
5722
|
cost: {
|
|
5689
|
-
input:
|
|
5690
|
-
output:
|
|
5723
|
+
input: 0.5,
|
|
5724
|
+
output: 1.5,
|
|
5691
5725
|
cacheRead: 0,
|
|
5692
5726
|
cacheWrite: 0,
|
|
5693
5727
|
},
|
|
5694
|
-
contextWindow:
|
|
5728
|
+
contextWindow: 16385,
|
|
5695
5729
|
maxTokens: 4096,
|
|
5696
5730
|
},
|
|
5697
5731
|
"openrouter/auto": {
|