@mariozechner/pi-ai 0.18.1 → 0.18.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +49 -15
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +113 -79
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +2 -0
- package/dist/providers/anthropic.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -2415,6 +2415,23 @@ export const MODELS = {
|
|
|
2415
2415
|
contextWindow: 262144,
|
|
2416
2416
|
maxTokens: 4096,
|
|
2417
2417
|
},
|
|
2418
|
+
"mistralai/devstral-2512": {
|
|
2419
|
+
id: "mistralai/devstral-2512",
|
|
2420
|
+
name: "Mistral: Devstral 2 2512",
|
|
2421
|
+
api: "openai-completions",
|
|
2422
|
+
provider: "openrouter",
|
|
2423
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2424
|
+
reasoning: false,
|
|
2425
|
+
input: ["text"],
|
|
2426
|
+
cost: {
|
|
2427
|
+
input: 0.15,
|
|
2428
|
+
output: 0.6,
|
|
2429
|
+
cacheRead: 0,
|
|
2430
|
+
cacheWrite: 0,
|
|
2431
|
+
},
|
|
2432
|
+
contextWindow: 262144,
|
|
2433
|
+
maxTokens: 65536,
|
|
2434
|
+
},
|
|
2418
2435
|
"relace/relace-search": {
|
|
2419
2436
|
id: "relace/relace-search",
|
|
2420
2437
|
name: "Relace: Relace Search",
|
|
@@ -2449,6 +2466,23 @@ export const MODELS = {
|
|
|
2449
2466
|
contextWindow: 131072,
|
|
2450
2467
|
maxTokens: 24000,
|
|
2451
2468
|
},
|
|
2469
|
+
"nex-agi/deepseek-v3.1-nex-n1:free": {
|
|
2470
|
+
id: "nex-agi/deepseek-v3.1-nex-n1:free",
|
|
2471
|
+
name: "Nex AGI: DeepSeek V3.1 Nex N1 (free)",
|
|
2472
|
+
api: "openai-completions",
|
|
2473
|
+
provider: "openrouter",
|
|
2474
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2475
|
+
reasoning: true,
|
|
2476
|
+
input: ["text"],
|
|
2477
|
+
cost: {
|
|
2478
|
+
input: 0,
|
|
2479
|
+
output: 0,
|
|
2480
|
+
cacheRead: 0,
|
|
2481
|
+
cacheWrite: 0,
|
|
2482
|
+
},
|
|
2483
|
+
contextWindow: 131072,
|
|
2484
|
+
maxTokens: 163840,
|
|
2485
|
+
},
|
|
2452
2486
|
"openai/gpt-5.1-codex-max": {
|
|
2453
2487
|
id: "openai/gpt-5.1-codex-max",
|
|
2454
2488
|
name: "OpenAI: GPT-5.1-Codex-Max",
|
|
@@ -5475,34 +5509,34 @@ export const MODELS = {
|
|
|
5475
5509
|
contextWindow: 200000,
|
|
5476
5510
|
maxTokens: 8192,
|
|
5477
5511
|
},
|
|
5478
|
-
"mistralai/ministral-
|
|
5479
|
-
id: "mistralai/ministral-
|
|
5480
|
-
name: "Mistral: Ministral
|
|
5512
|
+
"mistralai/ministral-8b": {
|
|
5513
|
+
id: "mistralai/ministral-8b",
|
|
5514
|
+
name: "Mistral: Ministral 8B",
|
|
5481
5515
|
api: "openai-completions",
|
|
5482
5516
|
provider: "openrouter",
|
|
5483
5517
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5484
5518
|
reasoning: false,
|
|
5485
5519
|
input: ["text"],
|
|
5486
5520
|
cost: {
|
|
5487
|
-
input: 0.
|
|
5488
|
-
output: 0.
|
|
5521
|
+
input: 0.09999999999999999,
|
|
5522
|
+
output: 0.09999999999999999,
|
|
5489
5523
|
cacheRead: 0,
|
|
5490
5524
|
cacheWrite: 0,
|
|
5491
5525
|
},
|
|
5492
5526
|
contextWindow: 131072,
|
|
5493
5527
|
maxTokens: 4096,
|
|
5494
5528
|
},
|
|
5495
|
-
"mistralai/ministral-
|
|
5496
|
-
id: "mistralai/ministral-
|
|
5497
|
-
name: "Mistral: Ministral
|
|
5529
|
+
"mistralai/ministral-3b": {
|
|
5530
|
+
id: "mistralai/ministral-3b",
|
|
5531
|
+
name: "Mistral: Ministral 3B",
|
|
5498
5532
|
api: "openai-completions",
|
|
5499
5533
|
provider: "openrouter",
|
|
5500
5534
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5501
5535
|
reasoning: false,
|
|
5502
5536
|
input: ["text"],
|
|
5503
5537
|
cost: {
|
|
5504
|
-
input: 0.
|
|
5505
|
-
output: 0.
|
|
5538
|
+
input: 0.04,
|
|
5539
|
+
output: 0.04,
|
|
5506
5540
|
cacheRead: 0,
|
|
5507
5541
|
cacheWrite: 0,
|
|
5508
5542
|
},
|
|
@@ -5679,6 +5713,23 @@ export const MODELS = {
|
|
|
5679
5713
|
contextWindow: 128000,
|
|
5680
5714
|
maxTokens: 16384,
|
|
5681
5715
|
},
|
|
5716
|
+
"meta-llama/llama-3.1-8b-instruct": {
|
|
5717
|
+
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5718
|
+
name: "Meta: Llama 3.1 8B Instruct",
|
|
5719
|
+
api: "openai-completions",
|
|
5720
|
+
provider: "openrouter",
|
|
5721
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5722
|
+
reasoning: false,
|
|
5723
|
+
input: ["text"],
|
|
5724
|
+
cost: {
|
|
5725
|
+
input: 0.02,
|
|
5726
|
+
output: 0.03,
|
|
5727
|
+
cacheRead: 0,
|
|
5728
|
+
cacheWrite: 0,
|
|
5729
|
+
},
|
|
5730
|
+
contextWindow: 131072,
|
|
5731
|
+
maxTokens: 16384,
|
|
5732
|
+
},
|
|
5682
5733
|
"meta-llama/llama-3.1-405b-instruct": {
|
|
5683
5734
|
id: "meta-llama/llama-3.1-405b-instruct",
|
|
5684
5735
|
name: "Meta: Llama 3.1 405B Instruct",
|
|
@@ -5713,23 +5764,6 @@ export const MODELS = {
|
|
|
5713
5764
|
contextWindow: 131072,
|
|
5714
5765
|
maxTokens: 4096,
|
|
5715
5766
|
},
|
|
5716
|
-
"meta-llama/llama-3.1-8b-instruct": {
|
|
5717
|
-
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5718
|
-
name: "Meta: Llama 3.1 8B Instruct",
|
|
5719
|
-
api: "openai-completions",
|
|
5720
|
-
provider: "openrouter",
|
|
5721
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5722
|
-
reasoning: false,
|
|
5723
|
-
input: ["text"],
|
|
5724
|
-
cost: {
|
|
5725
|
-
input: 0.02,
|
|
5726
|
-
output: 0.03,
|
|
5727
|
-
cacheRead: 0,
|
|
5728
|
-
cacheWrite: 0,
|
|
5729
|
-
},
|
|
5730
|
-
contextWindow: 131072,
|
|
5731
|
-
maxTokens: 16384,
|
|
5732
|
-
},
|
|
5733
5767
|
"mistralai/mistral-nemo": {
|
|
5734
5768
|
id: "mistralai/mistral-nemo",
|
|
5735
5769
|
name: "Mistral: Mistral Nemo",
|
|
@@ -5866,6 +5900,23 @@ export const MODELS = {
|
|
|
5866
5900
|
contextWindow: 128000,
|
|
5867
5901
|
maxTokens: 4096,
|
|
5868
5902
|
},
|
|
5903
|
+
"openai/gpt-4o-2024-05-13": {
|
|
5904
|
+
id: "openai/gpt-4o-2024-05-13",
|
|
5905
|
+
name: "OpenAI: GPT-4o (2024-05-13)",
|
|
5906
|
+
api: "openai-completions",
|
|
5907
|
+
provider: "openrouter",
|
|
5908
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5909
|
+
reasoning: false,
|
|
5910
|
+
input: ["text", "image"],
|
|
5911
|
+
cost: {
|
|
5912
|
+
input: 5,
|
|
5913
|
+
output: 15,
|
|
5914
|
+
cacheRead: 0,
|
|
5915
|
+
cacheWrite: 0,
|
|
5916
|
+
},
|
|
5917
|
+
contextWindow: 128000,
|
|
5918
|
+
maxTokens: 4096,
|
|
5919
|
+
},
|
|
5869
5920
|
"openai/gpt-4o": {
|
|
5870
5921
|
id: "openai/gpt-4o",
|
|
5871
5922
|
name: "OpenAI: GPT-4o",
|
|
@@ -5900,22 +5951,22 @@ export const MODELS = {
|
|
|
5900
5951
|
contextWindow: 128000,
|
|
5901
5952
|
maxTokens: 64000,
|
|
5902
5953
|
},
|
|
5903
|
-
"
|
|
5904
|
-
id: "
|
|
5905
|
-
name: "
|
|
5954
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
5955
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
5956
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
5906
5957
|
api: "openai-completions",
|
|
5907
5958
|
provider: "openrouter",
|
|
5908
5959
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5909
5960
|
reasoning: false,
|
|
5910
|
-
input: ["text"
|
|
5961
|
+
input: ["text"],
|
|
5911
5962
|
cost: {
|
|
5912
|
-
input:
|
|
5913
|
-
output:
|
|
5963
|
+
input: 0.3,
|
|
5964
|
+
output: 0.39999999999999997,
|
|
5914
5965
|
cacheRead: 0,
|
|
5915
5966
|
cacheWrite: 0,
|
|
5916
5967
|
},
|
|
5917
|
-
contextWindow:
|
|
5918
|
-
maxTokens:
|
|
5968
|
+
contextWindow: 8192,
|
|
5969
|
+
maxTokens: 16384,
|
|
5919
5970
|
},
|
|
5920
5971
|
"meta-llama/llama-3-8b-instruct": {
|
|
5921
5972
|
id: "meta-llama/llama-3-8b-instruct",
|
|
@@ -5934,23 +5985,6 @@ export const MODELS = {
|
|
|
5934
5985
|
contextWindow: 8192,
|
|
5935
5986
|
maxTokens: 16384,
|
|
5936
5987
|
},
|
|
5937
|
-
"meta-llama/llama-3-70b-instruct": {
|
|
5938
|
-
id: "meta-llama/llama-3-70b-instruct",
|
|
5939
|
-
name: "Meta: Llama 3 70B Instruct",
|
|
5940
|
-
api: "openai-completions",
|
|
5941
|
-
provider: "openrouter",
|
|
5942
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5943
|
-
reasoning: false,
|
|
5944
|
-
input: ["text"],
|
|
5945
|
-
cost: {
|
|
5946
|
-
input: 0.3,
|
|
5947
|
-
output: 0.39999999999999997,
|
|
5948
|
-
cacheRead: 0,
|
|
5949
|
-
cacheWrite: 0,
|
|
5950
|
-
},
|
|
5951
|
-
contextWindow: 8192,
|
|
5952
|
-
maxTokens: 16384,
|
|
5953
|
-
},
|
|
5954
5988
|
"mistralai/mixtral-8x22b-instruct": {
|
|
5955
5989
|
id: "mistralai/mixtral-8x22b-instruct",
|
|
5956
5990
|
name: "Mistral: Mixtral 8x22B Instruct",
|
|
@@ -6036,38 +6070,38 @@ export const MODELS = {
|
|
|
6036
6070
|
contextWindow: 128000,
|
|
6037
6071
|
maxTokens: 4096,
|
|
6038
6072
|
},
|
|
6039
|
-
"openai/gpt-
|
|
6040
|
-
id: "openai/gpt-
|
|
6041
|
-
name: "OpenAI: GPT-
|
|
6073
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
6074
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
6075
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
6042
6076
|
api: "openai-completions",
|
|
6043
6077
|
provider: "openrouter",
|
|
6044
6078
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6045
6079
|
reasoning: false,
|
|
6046
6080
|
input: ["text"],
|
|
6047
6081
|
cost: {
|
|
6048
|
-
input:
|
|
6049
|
-
output:
|
|
6082
|
+
input: 1,
|
|
6083
|
+
output: 2,
|
|
6050
6084
|
cacheRead: 0,
|
|
6051
6085
|
cacheWrite: 0,
|
|
6052
6086
|
},
|
|
6053
|
-
contextWindow:
|
|
6087
|
+
contextWindow: 4095,
|
|
6054
6088
|
maxTokens: 4096,
|
|
6055
6089
|
},
|
|
6056
|
-
"openai/gpt-
|
|
6057
|
-
id: "openai/gpt-
|
|
6058
|
-
name: "OpenAI: GPT-
|
|
6090
|
+
"openai/gpt-4-turbo-preview": {
|
|
6091
|
+
id: "openai/gpt-4-turbo-preview",
|
|
6092
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
6059
6093
|
api: "openai-completions",
|
|
6060
6094
|
provider: "openrouter",
|
|
6061
6095
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6062
6096
|
reasoning: false,
|
|
6063
6097
|
input: ["text"],
|
|
6064
6098
|
cost: {
|
|
6065
|
-
input:
|
|
6066
|
-
output:
|
|
6099
|
+
input: 10,
|
|
6100
|
+
output: 30,
|
|
6067
6101
|
cacheRead: 0,
|
|
6068
6102
|
cacheWrite: 0,
|
|
6069
6103
|
},
|
|
6070
|
-
contextWindow:
|
|
6104
|
+
contextWindow: 128000,
|
|
6071
6105
|
maxTokens: 4096,
|
|
6072
6106
|
},
|
|
6073
6107
|
"mistralai/mistral-tiny": {
|
|
@@ -6138,9 +6172,9 @@ export const MODELS = {
|
|
|
6138
6172
|
contextWindow: 16385,
|
|
6139
6173
|
maxTokens: 4096,
|
|
6140
6174
|
},
|
|
6141
|
-
"openai/gpt-4": {
|
|
6142
|
-
id: "openai/gpt-4",
|
|
6143
|
-
name: "OpenAI: GPT-4",
|
|
6175
|
+
"openai/gpt-4-0314": {
|
|
6176
|
+
id: "openai/gpt-4-0314",
|
|
6177
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
6144
6178
|
api: "openai-completions",
|
|
6145
6179
|
provider: "openrouter",
|
|
6146
6180
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6155,38 +6189,38 @@ export const MODELS = {
|
|
|
6155
6189
|
contextWindow: 8191,
|
|
6156
6190
|
maxTokens: 4096,
|
|
6157
6191
|
},
|
|
6158
|
-
"openai/gpt-
|
|
6159
|
-
id: "openai/gpt-
|
|
6160
|
-
name: "OpenAI: GPT-
|
|
6192
|
+
"openai/gpt-4": {
|
|
6193
|
+
id: "openai/gpt-4",
|
|
6194
|
+
name: "OpenAI: GPT-4",
|
|
6161
6195
|
api: "openai-completions",
|
|
6162
6196
|
provider: "openrouter",
|
|
6163
6197
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6164
6198
|
reasoning: false,
|
|
6165
6199
|
input: ["text"],
|
|
6166
6200
|
cost: {
|
|
6167
|
-
input:
|
|
6168
|
-
output:
|
|
6201
|
+
input: 30,
|
|
6202
|
+
output: 60,
|
|
6169
6203
|
cacheRead: 0,
|
|
6170
6204
|
cacheWrite: 0,
|
|
6171
6205
|
},
|
|
6172
|
-
contextWindow:
|
|
6206
|
+
contextWindow: 8191,
|
|
6173
6207
|
maxTokens: 4096,
|
|
6174
6208
|
},
|
|
6175
|
-
"openai/gpt-
|
|
6176
|
-
id: "openai/gpt-
|
|
6177
|
-
name: "OpenAI: GPT-
|
|
6209
|
+
"openai/gpt-3.5-turbo": {
|
|
6210
|
+
id: "openai/gpt-3.5-turbo",
|
|
6211
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
6178
6212
|
api: "openai-completions",
|
|
6179
6213
|
provider: "openrouter",
|
|
6180
6214
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6181
6215
|
reasoning: false,
|
|
6182
6216
|
input: ["text"],
|
|
6183
6217
|
cost: {
|
|
6184
|
-
input:
|
|
6185
|
-
output:
|
|
6218
|
+
input: 0.5,
|
|
6219
|
+
output: 1.5,
|
|
6186
6220
|
cacheRead: 0,
|
|
6187
6221
|
cacheWrite: 0,
|
|
6188
6222
|
},
|
|
6189
|
-
contextWindow:
|
|
6223
|
+
contextWindow: 16385,
|
|
6190
6224
|
maxTokens: 4096,
|
|
6191
6225
|
},
|
|
6192
6226
|
"openrouter/auto": {
|