@mariozechner/pi-ai 0.18.8 → 0.19.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -2770,7 +2770,7 @@ export const MODELS = {
|
|
|
2770
2770
|
cacheWrite: 0,
|
|
2771
2771
|
},
|
|
2772
2772
|
contextWindow: 163840,
|
|
2773
|
-
maxTokens:
|
|
2773
|
+
maxTokens: 163840,
|
|
2774
2774
|
},
|
|
2775
2775
|
"prime-intellect/intellect-3": {
|
|
2776
2776
|
id: "prime-intellect/intellect-3",
|
|
@@ -3257,13 +3257,13 @@ export const MODELS = {
|
|
|
3257
3257
|
reasoning: false,
|
|
3258
3258
|
input: ["text", "image"],
|
|
3259
3259
|
cost: {
|
|
3260
|
-
input: 0.
|
|
3261
|
-
output:
|
|
3260
|
+
input: 0.15,
|
|
3261
|
+
output: 0.6,
|
|
3262
3262
|
cacheRead: 0,
|
|
3263
3263
|
cacheWrite: 0,
|
|
3264
3264
|
},
|
|
3265
|
-
contextWindow:
|
|
3266
|
-
maxTokens:
|
|
3265
|
+
contextWindow: 262144,
|
|
3266
|
+
maxTokens: 4096,
|
|
3267
3267
|
},
|
|
3268
3268
|
"openai/gpt-5-pro": {
|
|
3269
3269
|
id: "openai/gpt-5-pro",
|
|
@@ -5577,9 +5577,9 @@ export const MODELS = {
|
|
|
5577
5577
|
contextWindow: 32768,
|
|
5578
5578
|
maxTokens: 4096,
|
|
5579
5579
|
},
|
|
5580
|
-
"anthropic/claude-3.5-haiku": {
|
|
5581
|
-
id: "anthropic/claude-3.5-haiku",
|
|
5582
|
-
name: "Anthropic: Claude 3.5 Haiku",
|
|
5580
|
+
"anthropic/claude-3.5-haiku-20241022": {
|
|
5581
|
+
id: "anthropic/claude-3.5-haiku-20241022",
|
|
5582
|
+
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
5583
5583
|
api: "openai-completions",
|
|
5584
5584
|
provider: "openrouter",
|
|
5585
5585
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5594,9 +5594,9 @@ export const MODELS = {
|
|
|
5594
5594
|
contextWindow: 200000,
|
|
5595
5595
|
maxTokens: 8192,
|
|
5596
5596
|
},
|
|
5597
|
-
"anthropic/claude-3.5-haiku
|
|
5598
|
-
id: "anthropic/claude-3.5-haiku
|
|
5599
|
-
name: "Anthropic: Claude 3.5 Haiku
|
|
5597
|
+
"anthropic/claude-3.5-haiku": {
|
|
5598
|
+
id: "anthropic/claude-3.5-haiku",
|
|
5599
|
+
name: "Anthropic: Claude 3.5 Haiku",
|
|
5600
5600
|
api: "openai-completions",
|
|
5601
5601
|
provider: "openrouter",
|
|
5602
5602
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5628,34 +5628,34 @@ export const MODELS = {
|
|
|
5628
5628
|
contextWindow: 200000,
|
|
5629
5629
|
maxTokens: 8192,
|
|
5630
5630
|
},
|
|
5631
|
-
"mistralai/ministral-
|
|
5632
|
-
id: "mistralai/ministral-
|
|
5633
|
-
name: "Mistral: Ministral
|
|
5631
|
+
"mistralai/ministral-8b": {
|
|
5632
|
+
id: "mistralai/ministral-8b",
|
|
5633
|
+
name: "Mistral: Ministral 8B",
|
|
5634
5634
|
api: "openai-completions",
|
|
5635
5635
|
provider: "openrouter",
|
|
5636
5636
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5637
5637
|
reasoning: false,
|
|
5638
5638
|
input: ["text"],
|
|
5639
5639
|
cost: {
|
|
5640
|
-
input: 0.
|
|
5641
|
-
output: 0.
|
|
5640
|
+
input: 0.09999999999999999,
|
|
5641
|
+
output: 0.09999999999999999,
|
|
5642
5642
|
cacheRead: 0,
|
|
5643
5643
|
cacheWrite: 0,
|
|
5644
5644
|
},
|
|
5645
5645
|
contextWindow: 131072,
|
|
5646
5646
|
maxTokens: 4096,
|
|
5647
5647
|
},
|
|
5648
|
-
"mistralai/ministral-
|
|
5649
|
-
id: "mistralai/ministral-
|
|
5650
|
-
name: "Mistral: Ministral
|
|
5648
|
+
"mistralai/ministral-3b": {
|
|
5649
|
+
id: "mistralai/ministral-3b",
|
|
5650
|
+
name: "Mistral: Ministral 3B",
|
|
5651
5651
|
api: "openai-completions",
|
|
5652
5652
|
provider: "openrouter",
|
|
5653
5653
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5654
5654
|
reasoning: false,
|
|
5655
5655
|
input: ["text"],
|
|
5656
5656
|
cost: {
|
|
5657
|
-
input: 0.
|
|
5658
|
-
output: 0.
|
|
5657
|
+
input: 0.04,
|
|
5658
|
+
output: 0.04,
|
|
5659
5659
|
cacheRead: 0,
|
|
5660
5660
|
cacheWrite: 0,
|
|
5661
5661
|
},
|
|
@@ -5832,39 +5832,39 @@ export const MODELS = {
|
|
|
5832
5832
|
contextWindow: 128000,
|
|
5833
5833
|
maxTokens: 16384,
|
|
5834
5834
|
},
|
|
5835
|
-
"meta-llama/llama-3.1-
|
|
5836
|
-
id: "meta-llama/llama-3.1-
|
|
5837
|
-
name: "Meta: Llama 3.1
|
|
5835
|
+
"meta-llama/llama-3.1-8b-instruct": {
|
|
5836
|
+
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5837
|
+
name: "Meta: Llama 3.1 8B Instruct",
|
|
5838
5838
|
api: "openai-completions",
|
|
5839
5839
|
provider: "openrouter",
|
|
5840
5840
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5841
5841
|
reasoning: false,
|
|
5842
5842
|
input: ["text"],
|
|
5843
5843
|
cost: {
|
|
5844
|
-
input:
|
|
5845
|
-
output:
|
|
5844
|
+
input: 0.02,
|
|
5845
|
+
output: 0.03,
|
|
5846
5846
|
cacheRead: 0,
|
|
5847
5847
|
cacheWrite: 0,
|
|
5848
5848
|
},
|
|
5849
|
-
contextWindow:
|
|
5850
|
-
maxTokens:
|
|
5849
|
+
contextWindow: 131072,
|
|
5850
|
+
maxTokens: 16384,
|
|
5851
5851
|
},
|
|
5852
|
-
"meta-llama/llama-3.1-
|
|
5853
|
-
id: "meta-llama/llama-3.1-
|
|
5854
|
-
name: "Meta: Llama 3.1
|
|
5852
|
+
"meta-llama/llama-3.1-405b-instruct": {
|
|
5853
|
+
id: "meta-llama/llama-3.1-405b-instruct",
|
|
5854
|
+
name: "Meta: Llama 3.1 405B Instruct",
|
|
5855
5855
|
api: "openai-completions",
|
|
5856
5856
|
provider: "openrouter",
|
|
5857
5857
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5858
5858
|
reasoning: false,
|
|
5859
5859
|
input: ["text"],
|
|
5860
5860
|
cost: {
|
|
5861
|
-
input:
|
|
5862
|
-
output:
|
|
5861
|
+
input: 3.5,
|
|
5862
|
+
output: 3.5,
|
|
5863
5863
|
cacheRead: 0,
|
|
5864
5864
|
cacheWrite: 0,
|
|
5865
5865
|
},
|
|
5866
|
-
contextWindow:
|
|
5867
|
-
maxTokens:
|
|
5866
|
+
contextWindow: 130815,
|
|
5867
|
+
maxTokens: 4096,
|
|
5868
5868
|
},
|
|
5869
5869
|
"meta-llama/llama-3.1-70b-instruct": {
|
|
5870
5870
|
id: "meta-llama/llama-3.1-70b-instruct",
|
|
@@ -5900,9 +5900,9 @@ export const MODELS = {
|
|
|
5900
5900
|
contextWindow: 131072,
|
|
5901
5901
|
maxTokens: 16384,
|
|
5902
5902
|
},
|
|
5903
|
-
"openai/gpt-4o-mini": {
|
|
5904
|
-
id: "openai/gpt-4o-mini",
|
|
5905
|
-
name: "OpenAI: GPT-4o-mini",
|
|
5903
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
5904
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
5905
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
5906
5906
|
api: "openai-completions",
|
|
5907
5907
|
provider: "openrouter",
|
|
5908
5908
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5917,9 +5917,9 @@ export const MODELS = {
|
|
|
5917
5917
|
contextWindow: 128000,
|
|
5918
5918
|
maxTokens: 16384,
|
|
5919
5919
|
},
|
|
5920
|
-
"openai/gpt-4o-mini
|
|
5921
|
-
id: "openai/gpt-4o-mini
|
|
5922
|
-
name: "OpenAI: GPT-4o-mini
|
|
5920
|
+
"openai/gpt-4o-mini": {
|
|
5921
|
+
id: "openai/gpt-4o-mini",
|
|
5922
|
+
name: "OpenAI: GPT-4o-mini",
|
|
5923
5923
|
api: "openai-completions",
|
|
5924
5924
|
provider: "openrouter",
|
|
5925
5925
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6070,34 +6070,34 @@ export const MODELS = {
|
|
|
6070
6070
|
contextWindow: 128000,
|
|
6071
6071
|
maxTokens: 64000,
|
|
6072
6072
|
},
|
|
6073
|
-
"meta-llama/llama-3-
|
|
6074
|
-
id: "meta-llama/llama-3-
|
|
6075
|
-
name: "Meta: Llama 3
|
|
6073
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
6074
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
6075
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
6076
6076
|
api: "openai-completions",
|
|
6077
6077
|
provider: "openrouter",
|
|
6078
6078
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6079
6079
|
reasoning: false,
|
|
6080
6080
|
input: ["text"],
|
|
6081
6081
|
cost: {
|
|
6082
|
-
input: 0.
|
|
6083
|
-
output: 0.
|
|
6082
|
+
input: 0.3,
|
|
6083
|
+
output: 0.39999999999999997,
|
|
6084
6084
|
cacheRead: 0,
|
|
6085
6085
|
cacheWrite: 0,
|
|
6086
6086
|
},
|
|
6087
6087
|
contextWindow: 8192,
|
|
6088
6088
|
maxTokens: 16384,
|
|
6089
6089
|
},
|
|
6090
|
-
"meta-llama/llama-3-
|
|
6091
|
-
id: "meta-llama/llama-3-
|
|
6092
|
-
name: "Meta: Llama 3
|
|
6090
|
+
"meta-llama/llama-3-8b-instruct": {
|
|
6091
|
+
id: "meta-llama/llama-3-8b-instruct",
|
|
6092
|
+
name: "Meta: Llama 3 8B Instruct",
|
|
6093
6093
|
api: "openai-completions",
|
|
6094
6094
|
provider: "openrouter",
|
|
6095
6095
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6096
6096
|
reasoning: false,
|
|
6097
6097
|
input: ["text"],
|
|
6098
6098
|
cost: {
|
|
6099
|
-
input: 0.
|
|
6100
|
-
output: 0.
|
|
6099
|
+
input: 0.03,
|
|
6100
|
+
output: 0.06,
|
|
6101
6101
|
cacheRead: 0,
|
|
6102
6102
|
cacheWrite: 0,
|
|
6103
6103
|
},
|
|
@@ -6291,26 +6291,26 @@ export const MODELS = {
|
|
|
6291
6291
|
contextWindow: 16385,
|
|
6292
6292
|
maxTokens: 4096,
|
|
6293
6293
|
},
|
|
6294
|
-
"openai/gpt-
|
|
6295
|
-
id: "openai/gpt-
|
|
6296
|
-
name: "OpenAI: GPT-
|
|
6294
|
+
"openai/gpt-4-0314": {
|
|
6295
|
+
id: "openai/gpt-4-0314",
|
|
6296
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
6297
6297
|
api: "openai-completions",
|
|
6298
6298
|
provider: "openrouter",
|
|
6299
6299
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6300
6300
|
reasoning: false,
|
|
6301
6301
|
input: ["text"],
|
|
6302
6302
|
cost: {
|
|
6303
|
-
input:
|
|
6304
|
-
output:
|
|
6303
|
+
input: 30,
|
|
6304
|
+
output: 60,
|
|
6305
6305
|
cacheRead: 0,
|
|
6306
6306
|
cacheWrite: 0,
|
|
6307
6307
|
},
|
|
6308
|
-
contextWindow:
|
|
6308
|
+
contextWindow: 8191,
|
|
6309
6309
|
maxTokens: 4096,
|
|
6310
6310
|
},
|
|
6311
|
-
"openai/gpt-4
|
|
6312
|
-
id: "openai/gpt-4
|
|
6313
|
-
name: "OpenAI: GPT-4
|
|
6311
|
+
"openai/gpt-4": {
|
|
6312
|
+
id: "openai/gpt-4",
|
|
6313
|
+
name: "OpenAI: GPT-4",
|
|
6314
6314
|
api: "openai-completions",
|
|
6315
6315
|
provider: "openrouter",
|
|
6316
6316
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6325,21 +6325,21 @@ export const MODELS = {
|
|
|
6325
6325
|
contextWindow: 8191,
|
|
6326
6326
|
maxTokens: 4096,
|
|
6327
6327
|
},
|
|
6328
|
-
"openai/gpt-
|
|
6329
|
-
id: "openai/gpt-
|
|
6330
|
-
name: "OpenAI: GPT-
|
|
6328
|
+
"openai/gpt-3.5-turbo": {
|
|
6329
|
+
id: "openai/gpt-3.5-turbo",
|
|
6330
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
6331
6331
|
api: "openai-completions",
|
|
6332
6332
|
provider: "openrouter",
|
|
6333
6333
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6334
6334
|
reasoning: false,
|
|
6335
6335
|
input: ["text"],
|
|
6336
6336
|
cost: {
|
|
6337
|
-
input:
|
|
6338
|
-
output:
|
|
6337
|
+
input: 0.5,
|
|
6338
|
+
output: 1.5,
|
|
6339
6339
|
cacheRead: 0,
|
|
6340
6340
|
cacheWrite: 0,
|
|
6341
6341
|
},
|
|
6342
|
-
contextWindow:
|
|
6342
|
+
contextWindow: 16385,
|
|
6343
6343
|
maxTokens: 4096,
|
|
6344
6344
|
},
|
|
6345
6345
|
"openrouter/auto": {
|