@mariozechner/pi-ai 0.18.7 → 0.19.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -2770,7 +2770,7 @@ export const MODELS = {
|
|
|
2770
2770
|
cacheWrite: 0,
|
|
2771
2771
|
},
|
|
2772
2772
|
contextWindow: 163840,
|
|
2773
|
-
maxTokens:
|
|
2773
|
+
maxTokens: 163840,
|
|
2774
2774
|
},
|
|
2775
2775
|
"prime-intellect/intellect-3": {
|
|
2776
2776
|
id: "prime-intellect/intellect-3",
|
|
@@ -3257,13 +3257,13 @@ export const MODELS = {
|
|
|
3257
3257
|
reasoning: false,
|
|
3258
3258
|
input: ["text", "image"],
|
|
3259
3259
|
cost: {
|
|
3260
|
-
input: 0.
|
|
3261
|
-
output:
|
|
3260
|
+
input: 0.15,
|
|
3261
|
+
output: 0.6,
|
|
3262
3262
|
cacheRead: 0,
|
|
3263
3263
|
cacheWrite: 0,
|
|
3264
3264
|
},
|
|
3265
|
-
contextWindow:
|
|
3266
|
-
maxTokens:
|
|
3265
|
+
contextWindow: 262144,
|
|
3266
|
+
maxTokens: 4096,
|
|
3267
3267
|
},
|
|
3268
3268
|
"openai/gpt-5-pro": {
|
|
3269
3269
|
id: "openai/gpt-5-pro",
|
|
@@ -5577,9 +5577,9 @@ export const MODELS = {
|
|
|
5577
5577
|
contextWindow: 32768,
|
|
5578
5578
|
maxTokens: 4096,
|
|
5579
5579
|
},
|
|
5580
|
-
"anthropic/claude-3.5-haiku": {
|
|
5581
|
-
id: "anthropic/claude-3.5-haiku",
|
|
5582
|
-
name: "Anthropic: Claude 3.5 Haiku",
|
|
5580
|
+
"anthropic/claude-3.5-haiku-20241022": {
|
|
5581
|
+
id: "anthropic/claude-3.5-haiku-20241022",
|
|
5582
|
+
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
5583
5583
|
api: "openai-completions",
|
|
5584
5584
|
provider: "openrouter",
|
|
5585
5585
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5594,9 +5594,9 @@ export const MODELS = {
|
|
|
5594
5594
|
contextWindow: 200000,
|
|
5595
5595
|
maxTokens: 8192,
|
|
5596
5596
|
},
|
|
5597
|
-
"anthropic/claude-3.5-haiku
|
|
5598
|
-
id: "anthropic/claude-3.5-haiku
|
|
5599
|
-
name: "Anthropic: Claude 3.5 Haiku
|
|
5597
|
+
"anthropic/claude-3.5-haiku": {
|
|
5598
|
+
id: "anthropic/claude-3.5-haiku",
|
|
5599
|
+
name: "Anthropic: Claude 3.5 Haiku",
|
|
5600
5600
|
api: "openai-completions",
|
|
5601
5601
|
provider: "openrouter",
|
|
5602
5602
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5628,34 +5628,34 @@ export const MODELS = {
|
|
|
5628
5628
|
contextWindow: 200000,
|
|
5629
5629
|
maxTokens: 8192,
|
|
5630
5630
|
},
|
|
5631
|
-
"mistralai/ministral-
|
|
5632
|
-
id: "mistralai/ministral-
|
|
5633
|
-
name: "Mistral: Ministral
|
|
5631
|
+
"mistralai/ministral-8b": {
|
|
5632
|
+
id: "mistralai/ministral-8b",
|
|
5633
|
+
name: "Mistral: Ministral 8B",
|
|
5634
5634
|
api: "openai-completions",
|
|
5635
5635
|
provider: "openrouter",
|
|
5636
5636
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5637
5637
|
reasoning: false,
|
|
5638
5638
|
input: ["text"],
|
|
5639
5639
|
cost: {
|
|
5640
|
-
input: 0.
|
|
5641
|
-
output: 0.
|
|
5640
|
+
input: 0.09999999999999999,
|
|
5641
|
+
output: 0.09999999999999999,
|
|
5642
5642
|
cacheRead: 0,
|
|
5643
5643
|
cacheWrite: 0,
|
|
5644
5644
|
},
|
|
5645
5645
|
contextWindow: 131072,
|
|
5646
5646
|
maxTokens: 4096,
|
|
5647
5647
|
},
|
|
5648
|
-
"mistralai/ministral-
|
|
5649
|
-
id: "mistralai/ministral-
|
|
5650
|
-
name: "Mistral: Ministral
|
|
5648
|
+
"mistralai/ministral-3b": {
|
|
5649
|
+
id: "mistralai/ministral-3b",
|
|
5650
|
+
name: "Mistral: Ministral 3B",
|
|
5651
5651
|
api: "openai-completions",
|
|
5652
5652
|
provider: "openrouter",
|
|
5653
5653
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5654
5654
|
reasoning: false,
|
|
5655
5655
|
input: ["text"],
|
|
5656
5656
|
cost: {
|
|
5657
|
-
input: 0.
|
|
5658
|
-
output: 0.
|
|
5657
|
+
input: 0.04,
|
|
5658
|
+
output: 0.04,
|
|
5659
5659
|
cacheRead: 0,
|
|
5660
5660
|
cacheWrite: 0,
|
|
5661
5661
|
},
|
|
@@ -5747,34 +5747,34 @@ export const MODELS = {
|
|
|
5747
5747
|
contextWindow: 32768,
|
|
5748
5748
|
maxTokens: 4096,
|
|
5749
5749
|
},
|
|
5750
|
-
"cohere/command-r-08-2024": {
|
|
5751
|
-
id: "cohere/command-r-08-2024",
|
|
5752
|
-
name: "Cohere: Command R (08-2024)",
|
|
5750
|
+
"cohere/command-r-plus-08-2024": {
|
|
5751
|
+
id: "cohere/command-r-plus-08-2024",
|
|
5752
|
+
name: "Cohere: Command R+ (08-2024)",
|
|
5753
5753
|
api: "openai-completions",
|
|
5754
5754
|
provider: "openrouter",
|
|
5755
5755
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5756
5756
|
reasoning: false,
|
|
5757
5757
|
input: ["text"],
|
|
5758
5758
|
cost: {
|
|
5759
|
-
input:
|
|
5760
|
-
output:
|
|
5759
|
+
input: 2.5,
|
|
5760
|
+
output: 10,
|
|
5761
5761
|
cacheRead: 0,
|
|
5762
5762
|
cacheWrite: 0,
|
|
5763
5763
|
},
|
|
5764
5764
|
contextWindow: 128000,
|
|
5765
5765
|
maxTokens: 4000,
|
|
5766
5766
|
},
|
|
5767
|
-
"cohere/command-r-
|
|
5768
|
-
id: "cohere/command-r-
|
|
5769
|
-
name: "Cohere: Command R
|
|
5767
|
+
"cohere/command-r-08-2024": {
|
|
5768
|
+
id: "cohere/command-r-08-2024",
|
|
5769
|
+
name: "Cohere: Command R (08-2024)",
|
|
5770
5770
|
api: "openai-completions",
|
|
5771
5771
|
provider: "openrouter",
|
|
5772
5772
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5773
5773
|
reasoning: false,
|
|
5774
5774
|
input: ["text"],
|
|
5775
5775
|
cost: {
|
|
5776
|
-
input:
|
|
5777
|
-
output:
|
|
5776
|
+
input: 0.15,
|
|
5777
|
+
output: 0.6,
|
|
5778
5778
|
cacheRead: 0,
|
|
5779
5779
|
cacheWrite: 0,
|
|
5780
5780
|
},
|
|
@@ -5832,23 +5832,6 @@ export const MODELS = {
|
|
|
5832
5832
|
contextWindow: 128000,
|
|
5833
5833
|
maxTokens: 16384,
|
|
5834
5834
|
},
|
|
5835
|
-
"meta-llama/llama-3.1-405b-instruct": {
|
|
5836
|
-
id: "meta-llama/llama-3.1-405b-instruct",
|
|
5837
|
-
name: "Meta: Llama 3.1 405B Instruct",
|
|
5838
|
-
api: "openai-completions",
|
|
5839
|
-
provider: "openrouter",
|
|
5840
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5841
|
-
reasoning: false,
|
|
5842
|
-
input: ["text"],
|
|
5843
|
-
cost: {
|
|
5844
|
-
input: 3.5,
|
|
5845
|
-
output: 3.5,
|
|
5846
|
-
cacheRead: 0,
|
|
5847
|
-
cacheWrite: 0,
|
|
5848
|
-
},
|
|
5849
|
-
contextWindow: 130815,
|
|
5850
|
-
maxTokens: 4096,
|
|
5851
|
-
},
|
|
5852
5835
|
"meta-llama/llama-3.1-8b-instruct": {
|
|
5853
5836
|
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5854
5837
|
name: "Meta: Llama 3.1 8B Instruct",
|
|
@@ -5883,6 +5866,23 @@ export const MODELS = {
|
|
|
5883
5866
|
contextWindow: 131072,
|
|
5884
5867
|
maxTokens: 4096,
|
|
5885
5868
|
},
|
|
5869
|
+
"meta-llama/llama-3.1-405b-instruct": {
|
|
5870
|
+
id: "meta-llama/llama-3.1-405b-instruct",
|
|
5871
|
+
name: "Meta: Llama 3.1 405B Instruct",
|
|
5872
|
+
api: "openai-completions",
|
|
5873
|
+
provider: "openrouter",
|
|
5874
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5875
|
+
reasoning: false,
|
|
5876
|
+
input: ["text"],
|
|
5877
|
+
cost: {
|
|
5878
|
+
input: 3.5,
|
|
5879
|
+
output: 3.5,
|
|
5880
|
+
cacheRead: 0,
|
|
5881
|
+
cacheWrite: 0,
|
|
5882
|
+
},
|
|
5883
|
+
contextWindow: 130815,
|
|
5884
|
+
maxTokens: 4096,
|
|
5885
|
+
},
|
|
5886
5886
|
"mistralai/mistral-nemo": {
|
|
5887
5887
|
id: "mistralai/mistral-nemo",
|
|
5888
5888
|
name: "Mistral: Mistral Nemo",
|
|
@@ -6070,34 +6070,34 @@ export const MODELS = {
|
|
|
6070
6070
|
contextWindow: 128000,
|
|
6071
6071
|
maxTokens: 64000,
|
|
6072
6072
|
},
|
|
6073
|
-
"meta-llama/llama-3-
|
|
6074
|
-
id: "meta-llama/llama-3-
|
|
6075
|
-
name: "Meta: Llama 3
|
|
6073
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
6074
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
6075
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
6076
6076
|
api: "openai-completions",
|
|
6077
6077
|
provider: "openrouter",
|
|
6078
6078
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6079
6079
|
reasoning: false,
|
|
6080
6080
|
input: ["text"],
|
|
6081
6081
|
cost: {
|
|
6082
|
-
input: 0.
|
|
6083
|
-
output: 0.
|
|
6082
|
+
input: 0.3,
|
|
6083
|
+
output: 0.39999999999999997,
|
|
6084
6084
|
cacheRead: 0,
|
|
6085
6085
|
cacheWrite: 0,
|
|
6086
6086
|
},
|
|
6087
6087
|
contextWindow: 8192,
|
|
6088
6088
|
maxTokens: 16384,
|
|
6089
6089
|
},
|
|
6090
|
-
"meta-llama/llama-3-
|
|
6091
|
-
id: "meta-llama/llama-3-
|
|
6092
|
-
name: "Meta: Llama 3
|
|
6090
|
+
"meta-llama/llama-3-8b-instruct": {
|
|
6091
|
+
id: "meta-llama/llama-3-8b-instruct",
|
|
6092
|
+
name: "Meta: Llama 3 8B Instruct",
|
|
6093
6093
|
api: "openai-completions",
|
|
6094
6094
|
provider: "openrouter",
|
|
6095
6095
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6096
6096
|
reasoning: false,
|
|
6097
6097
|
input: ["text"],
|
|
6098
6098
|
cost: {
|
|
6099
|
-
input: 0.
|
|
6100
|
-
output: 0.
|
|
6099
|
+
input: 0.03,
|
|
6100
|
+
output: 0.06,
|
|
6101
6101
|
cacheRead: 0,
|
|
6102
6102
|
cacheWrite: 0,
|
|
6103
6103
|
},
|
|
@@ -6291,43 +6291,43 @@ export const MODELS = {
|
|
|
6291
6291
|
contextWindow: 16385,
|
|
6292
6292
|
maxTokens: 4096,
|
|
6293
6293
|
},
|
|
6294
|
-
"openai/gpt-
|
|
6295
|
-
id: "openai/gpt-
|
|
6296
|
-
name: "OpenAI: GPT-
|
|
6294
|
+
"openai/gpt-4": {
|
|
6295
|
+
id: "openai/gpt-4",
|
|
6296
|
+
name: "OpenAI: GPT-4",
|
|
6297
6297
|
api: "openai-completions",
|
|
6298
6298
|
provider: "openrouter",
|
|
6299
6299
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6300
6300
|
reasoning: false,
|
|
6301
6301
|
input: ["text"],
|
|
6302
6302
|
cost: {
|
|
6303
|
-
input:
|
|
6304
|
-
output:
|
|
6303
|
+
input: 30,
|
|
6304
|
+
output: 60,
|
|
6305
6305
|
cacheRead: 0,
|
|
6306
6306
|
cacheWrite: 0,
|
|
6307
6307
|
},
|
|
6308
|
-
contextWindow:
|
|
6308
|
+
contextWindow: 8191,
|
|
6309
6309
|
maxTokens: 4096,
|
|
6310
6310
|
},
|
|
6311
|
-
"openai/gpt-
|
|
6312
|
-
id: "openai/gpt-
|
|
6313
|
-
name: "OpenAI: GPT-
|
|
6311
|
+
"openai/gpt-3.5-turbo": {
|
|
6312
|
+
id: "openai/gpt-3.5-turbo",
|
|
6313
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
6314
6314
|
api: "openai-completions",
|
|
6315
6315
|
provider: "openrouter",
|
|
6316
6316
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6317
6317
|
reasoning: false,
|
|
6318
6318
|
input: ["text"],
|
|
6319
6319
|
cost: {
|
|
6320
|
-
input:
|
|
6321
|
-
output:
|
|
6320
|
+
input: 0.5,
|
|
6321
|
+
output: 1.5,
|
|
6322
6322
|
cacheRead: 0,
|
|
6323
6323
|
cacheWrite: 0,
|
|
6324
6324
|
},
|
|
6325
|
-
contextWindow:
|
|
6325
|
+
contextWindow: 16385,
|
|
6326
6326
|
maxTokens: 4096,
|
|
6327
6327
|
},
|
|
6328
|
-
"openai/gpt-4": {
|
|
6329
|
-
id: "openai/gpt-4",
|
|
6330
|
-
name: "OpenAI: GPT-4",
|
|
6328
|
+
"openai/gpt-4-0314": {
|
|
6329
|
+
id: "openai/gpt-4-0314",
|
|
6330
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
6331
6331
|
api: "openai-completions",
|
|
6332
6332
|
provider: "openrouter",
|
|
6333
6333
|
baseUrl: "https://openrouter.ai/api/v1",
|