@mariozechner/pi-ai 0.18.4 → 0.18.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -2005,7 +2005,7 @@ export const MODELS = {
|
|
|
2005
2005
|
},
|
|
2006
2006
|
"glm-4.5v": {
|
|
2007
2007
|
id: "glm-4.5v",
|
|
2008
|
-
name: "GLM
|
|
2008
|
+
name: "GLM-4.5V",
|
|
2009
2009
|
api: "anthropic-messages",
|
|
2010
2010
|
provider: "zai",
|
|
2011
2011
|
baseUrl: "https://api.z.ai/api/anthropic",
|
|
@@ -4311,13 +4311,13 @@ export const MODELS = {
|
|
|
4311
4311
|
reasoning: true,
|
|
4312
4312
|
input: ["text"],
|
|
4313
4313
|
cost: {
|
|
4314
|
-
input: 0.
|
|
4315
|
-
output: 1.
|
|
4314
|
+
input: 0.22,
|
|
4315
|
+
output: 1.7999999999999998,
|
|
4316
4316
|
cacheRead: 0,
|
|
4317
4317
|
cacheWrite: 0,
|
|
4318
4318
|
},
|
|
4319
4319
|
contextWindow: 262144,
|
|
4320
|
-
maxTokens:
|
|
4320
|
+
maxTokens: 65536,
|
|
4321
4321
|
},
|
|
4322
4322
|
"google/gemini-2.5-flash-lite": {
|
|
4323
4323
|
id: "google/gemini-2.5-flash-lite",
|
|
@@ -5577,9 +5577,9 @@ export const MODELS = {
|
|
|
5577
5577
|
contextWindow: 32768,
|
|
5578
5578
|
maxTokens: 4096,
|
|
5579
5579
|
},
|
|
5580
|
-
"anthropic/claude-3.5-haiku
|
|
5581
|
-
id: "anthropic/claude-3.5-haiku
|
|
5582
|
-
name: "Anthropic: Claude 3.5 Haiku
|
|
5580
|
+
"anthropic/claude-3.5-haiku": {
|
|
5581
|
+
id: "anthropic/claude-3.5-haiku",
|
|
5582
|
+
name: "Anthropic: Claude 3.5 Haiku",
|
|
5583
5583
|
api: "openai-completions",
|
|
5584
5584
|
provider: "openrouter",
|
|
5585
5585
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5594,9 +5594,9 @@ export const MODELS = {
|
|
|
5594
5594
|
contextWindow: 200000,
|
|
5595
5595
|
maxTokens: 8192,
|
|
5596
5596
|
},
|
|
5597
|
-
"anthropic/claude-3.5-haiku": {
|
|
5598
|
-
id: "anthropic/claude-3.5-haiku",
|
|
5599
|
-
name: "Anthropic: Claude 3.5 Haiku",
|
|
5597
|
+
"anthropic/claude-3.5-haiku-20241022": {
|
|
5598
|
+
id: "anthropic/claude-3.5-haiku-20241022",
|
|
5599
|
+
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
5600
5600
|
api: "openai-completions",
|
|
5601
5601
|
provider: "openrouter",
|
|
5602
5602
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5628,34 +5628,34 @@ export const MODELS = {
|
|
|
5628
5628
|
contextWindow: 200000,
|
|
5629
5629
|
maxTokens: 8192,
|
|
5630
5630
|
},
|
|
5631
|
-
"mistralai/ministral-
|
|
5632
|
-
id: "mistralai/ministral-
|
|
5633
|
-
name: "Mistral: Ministral
|
|
5631
|
+
"mistralai/ministral-3b": {
|
|
5632
|
+
id: "mistralai/ministral-3b",
|
|
5633
|
+
name: "Mistral: Ministral 3B",
|
|
5634
5634
|
api: "openai-completions",
|
|
5635
5635
|
provider: "openrouter",
|
|
5636
5636
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5637
5637
|
reasoning: false,
|
|
5638
5638
|
input: ["text"],
|
|
5639
5639
|
cost: {
|
|
5640
|
-
input: 0.
|
|
5641
|
-
output: 0.
|
|
5640
|
+
input: 0.04,
|
|
5641
|
+
output: 0.04,
|
|
5642
5642
|
cacheRead: 0,
|
|
5643
5643
|
cacheWrite: 0,
|
|
5644
5644
|
},
|
|
5645
5645
|
contextWindow: 131072,
|
|
5646
5646
|
maxTokens: 4096,
|
|
5647
5647
|
},
|
|
5648
|
-
"mistralai/ministral-
|
|
5649
|
-
id: "mistralai/ministral-
|
|
5650
|
-
name: "Mistral: Ministral
|
|
5648
|
+
"mistralai/ministral-8b": {
|
|
5649
|
+
id: "mistralai/ministral-8b",
|
|
5650
|
+
name: "Mistral: Ministral 8B",
|
|
5651
5651
|
api: "openai-completions",
|
|
5652
5652
|
provider: "openrouter",
|
|
5653
5653
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5654
5654
|
reasoning: false,
|
|
5655
5655
|
input: ["text"],
|
|
5656
5656
|
cost: {
|
|
5657
|
-
input: 0.
|
|
5658
|
-
output: 0.
|
|
5657
|
+
input: 0.09999999999999999,
|
|
5658
|
+
output: 0.09999999999999999,
|
|
5659
5659
|
cacheRead: 0,
|
|
5660
5660
|
cacheWrite: 0,
|
|
5661
5661
|
},
|
|
@@ -5832,39 +5832,39 @@ export const MODELS = {
|
|
|
5832
5832
|
contextWindow: 128000,
|
|
5833
5833
|
maxTokens: 16384,
|
|
5834
5834
|
},
|
|
5835
|
-
"meta-llama/llama-3.1-
|
|
5836
|
-
id: "meta-llama/llama-3.1-
|
|
5837
|
-
name: "Meta: Llama 3.1
|
|
5835
|
+
"meta-llama/llama-3.1-405b-instruct": {
|
|
5836
|
+
id: "meta-llama/llama-3.1-405b-instruct",
|
|
5837
|
+
name: "Meta: Llama 3.1 405B Instruct",
|
|
5838
5838
|
api: "openai-completions",
|
|
5839
5839
|
provider: "openrouter",
|
|
5840
5840
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5841
5841
|
reasoning: false,
|
|
5842
5842
|
input: ["text"],
|
|
5843
5843
|
cost: {
|
|
5844
|
-
input:
|
|
5845
|
-
output:
|
|
5844
|
+
input: 3.5,
|
|
5845
|
+
output: 3.5,
|
|
5846
5846
|
cacheRead: 0,
|
|
5847
5847
|
cacheWrite: 0,
|
|
5848
5848
|
},
|
|
5849
|
-
contextWindow:
|
|
5850
|
-
maxTokens:
|
|
5849
|
+
contextWindow: 130815,
|
|
5850
|
+
maxTokens: 4096,
|
|
5851
5851
|
},
|
|
5852
|
-
"meta-llama/llama-3.1-
|
|
5853
|
-
id: "meta-llama/llama-3.1-
|
|
5854
|
-
name: "Meta: Llama 3.1
|
|
5852
|
+
"meta-llama/llama-3.1-8b-instruct": {
|
|
5853
|
+
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5854
|
+
name: "Meta: Llama 3.1 8B Instruct",
|
|
5855
5855
|
api: "openai-completions",
|
|
5856
5856
|
provider: "openrouter",
|
|
5857
5857
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5858
5858
|
reasoning: false,
|
|
5859
5859
|
input: ["text"],
|
|
5860
5860
|
cost: {
|
|
5861
|
-
input:
|
|
5862
|
-
output:
|
|
5861
|
+
input: 0.02,
|
|
5862
|
+
output: 0.03,
|
|
5863
5863
|
cacheRead: 0,
|
|
5864
5864
|
cacheWrite: 0,
|
|
5865
5865
|
},
|
|
5866
|
-
contextWindow:
|
|
5867
|
-
maxTokens:
|
|
5866
|
+
contextWindow: 131072,
|
|
5867
|
+
maxTokens: 16384,
|
|
5868
5868
|
},
|
|
5869
5869
|
"meta-llama/llama-3.1-70b-instruct": {
|
|
5870
5870
|
id: "meta-llama/llama-3.1-70b-instruct",
|
|
@@ -5900,9 +5900,9 @@ export const MODELS = {
|
|
|
5900
5900
|
contextWindow: 131072,
|
|
5901
5901
|
maxTokens: 16384,
|
|
5902
5902
|
},
|
|
5903
|
-
"openai/gpt-4o-mini
|
|
5904
|
-
id: "openai/gpt-4o-mini
|
|
5905
|
-
name: "OpenAI: GPT-4o-mini
|
|
5903
|
+
"openai/gpt-4o-mini": {
|
|
5904
|
+
id: "openai/gpt-4o-mini",
|
|
5905
|
+
name: "OpenAI: GPT-4o-mini",
|
|
5906
5906
|
api: "openai-completions",
|
|
5907
5907
|
provider: "openrouter",
|
|
5908
5908
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5917,9 +5917,9 @@ export const MODELS = {
|
|
|
5917
5917
|
contextWindow: 128000,
|
|
5918
5918
|
maxTokens: 16384,
|
|
5919
5919
|
},
|
|
5920
|
-
"openai/gpt-4o-mini": {
|
|
5921
|
-
id: "openai/gpt-4o-mini",
|
|
5922
|
-
name: "OpenAI: GPT-4o-mini",
|
|
5920
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
5921
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
5922
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
5923
5923
|
api: "openai-completions",
|
|
5924
5924
|
provider: "openrouter",
|
|
5925
5925
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6070,34 +6070,34 @@ export const MODELS = {
|
|
|
6070
6070
|
contextWindow: 128000,
|
|
6071
6071
|
maxTokens: 64000,
|
|
6072
6072
|
},
|
|
6073
|
-
"meta-llama/llama-3-
|
|
6074
|
-
id: "meta-llama/llama-3-
|
|
6075
|
-
name: "Meta: Llama 3
|
|
6073
|
+
"meta-llama/llama-3-8b-instruct": {
|
|
6074
|
+
id: "meta-llama/llama-3-8b-instruct",
|
|
6075
|
+
name: "Meta: Llama 3 8B Instruct",
|
|
6076
6076
|
api: "openai-completions",
|
|
6077
6077
|
provider: "openrouter",
|
|
6078
6078
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6079
6079
|
reasoning: false,
|
|
6080
6080
|
input: ["text"],
|
|
6081
6081
|
cost: {
|
|
6082
|
-
input: 0.
|
|
6083
|
-
output: 0.
|
|
6082
|
+
input: 0.03,
|
|
6083
|
+
output: 0.06,
|
|
6084
6084
|
cacheRead: 0,
|
|
6085
6085
|
cacheWrite: 0,
|
|
6086
6086
|
},
|
|
6087
6087
|
contextWindow: 8192,
|
|
6088
6088
|
maxTokens: 16384,
|
|
6089
6089
|
},
|
|
6090
|
-
"meta-llama/llama-3-
|
|
6091
|
-
id: "meta-llama/llama-3-
|
|
6092
|
-
name: "Meta: Llama 3
|
|
6090
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
6091
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
6092
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
6093
6093
|
api: "openai-completions",
|
|
6094
6094
|
provider: "openrouter",
|
|
6095
6095
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6096
6096
|
reasoning: false,
|
|
6097
6097
|
input: ["text"],
|
|
6098
6098
|
cost: {
|
|
6099
|
-
input: 0.
|
|
6100
|
-
output: 0.
|
|
6099
|
+
input: 0.3,
|
|
6100
|
+
output: 0.39999999999999997,
|
|
6101
6101
|
cacheRead: 0,
|
|
6102
6102
|
cacheWrite: 0,
|
|
6103
6103
|
},
|
|
@@ -6291,26 +6291,26 @@ export const MODELS = {
|
|
|
6291
6291
|
contextWindow: 16385,
|
|
6292
6292
|
maxTokens: 4096,
|
|
6293
6293
|
},
|
|
6294
|
-
"openai/gpt-
|
|
6295
|
-
id: "openai/gpt-
|
|
6296
|
-
name: "OpenAI: GPT-
|
|
6294
|
+
"openai/gpt-3.5-turbo": {
|
|
6295
|
+
id: "openai/gpt-3.5-turbo",
|
|
6296
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
6297
6297
|
api: "openai-completions",
|
|
6298
6298
|
provider: "openrouter",
|
|
6299
6299
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6300
6300
|
reasoning: false,
|
|
6301
6301
|
input: ["text"],
|
|
6302
6302
|
cost: {
|
|
6303
|
-
input:
|
|
6304
|
-
output:
|
|
6303
|
+
input: 0.5,
|
|
6304
|
+
output: 1.5,
|
|
6305
6305
|
cacheRead: 0,
|
|
6306
6306
|
cacheWrite: 0,
|
|
6307
6307
|
},
|
|
6308
|
-
contextWindow:
|
|
6308
|
+
contextWindow: 16385,
|
|
6309
6309
|
maxTokens: 4096,
|
|
6310
6310
|
},
|
|
6311
|
-
"openai/gpt-4": {
|
|
6312
|
-
id: "openai/gpt-4",
|
|
6313
|
-
name: "OpenAI: GPT-4",
|
|
6311
|
+
"openai/gpt-4-0314": {
|
|
6312
|
+
id: "openai/gpt-4-0314",
|
|
6313
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
6314
6314
|
api: "openai-completions",
|
|
6315
6315
|
provider: "openrouter",
|
|
6316
6316
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6325,21 +6325,21 @@ export const MODELS = {
|
|
|
6325
6325
|
contextWindow: 8191,
|
|
6326
6326
|
maxTokens: 4096,
|
|
6327
6327
|
},
|
|
6328
|
-
"openai/gpt-
|
|
6329
|
-
id: "openai/gpt-
|
|
6330
|
-
name: "OpenAI: GPT-
|
|
6328
|
+
"openai/gpt-4": {
|
|
6329
|
+
id: "openai/gpt-4",
|
|
6330
|
+
name: "OpenAI: GPT-4",
|
|
6331
6331
|
api: "openai-completions",
|
|
6332
6332
|
provider: "openrouter",
|
|
6333
6333
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6334
6334
|
reasoning: false,
|
|
6335
6335
|
input: ["text"],
|
|
6336
6336
|
cost: {
|
|
6337
|
-
input:
|
|
6338
|
-
output:
|
|
6337
|
+
input: 30,
|
|
6338
|
+
output: 60,
|
|
6339
6339
|
cacheRead: 0,
|
|
6340
6340
|
cacheWrite: 0,
|
|
6341
6341
|
},
|
|
6342
|
-
contextWindow:
|
|
6342
|
+
contextWindow: 8191,
|
|
6343
6343
|
maxTokens: 4096,
|
|
6344
6344
|
},
|
|
6345
6345
|
"openrouter/auto": {
|