@mariozechner/pi-ai 0.27.1 → 0.27.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -2871,6 +2871,23 @@ export const MODELS = {
|
|
|
2871
2871
|
},
|
|
2872
2872
|
},
|
|
2873
2873
|
"openrouter": {
|
|
2874
|
+
"minimax/minimax-m2.1": {
|
|
2875
|
+
id: "minimax/minimax-m2.1",
|
|
2876
|
+
name: "MiniMax: MiniMax M2.1",
|
|
2877
|
+
api: "openai-completions",
|
|
2878
|
+
provider: "openrouter",
|
|
2879
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2880
|
+
reasoning: true,
|
|
2881
|
+
input: ["text"],
|
|
2882
|
+
cost: {
|
|
2883
|
+
input: 0.3,
|
|
2884
|
+
output: 1.2,
|
|
2885
|
+
cacheRead: 0,
|
|
2886
|
+
cacheWrite: 0,
|
|
2887
|
+
},
|
|
2888
|
+
contextWindow: 204800,
|
|
2889
|
+
maxTokens: 131072,
|
|
2890
|
+
},
|
|
2874
2891
|
"z-ai/glm-4.7": {
|
|
2875
2892
|
id: "z-ai/glm-4.7",
|
|
2876
2893
|
name: "Z.AI: GLM 4.7",
|
|
@@ -2880,12 +2897,12 @@ export const MODELS = {
|
|
|
2880
2897
|
reasoning: true,
|
|
2881
2898
|
input: ["text"],
|
|
2882
2899
|
cost: {
|
|
2883
|
-
input: 0.
|
|
2884
|
-
output:
|
|
2885
|
-
cacheRead: 0
|
|
2900
|
+
input: 0.44,
|
|
2901
|
+
output: 1.74,
|
|
2902
|
+
cacheRead: 0,
|
|
2886
2903
|
cacheWrite: 0,
|
|
2887
2904
|
},
|
|
2888
|
-
contextWindow:
|
|
2905
|
+
contextWindow: 202752,
|
|
2889
2906
|
maxTokens: 131072,
|
|
2890
2907
|
},
|
|
2891
2908
|
"google/gemini-3-flash-preview": {
|
|
@@ -2971,7 +2988,7 @@ export const MODELS = {
|
|
|
2971
2988
|
cacheWrite: 0,
|
|
2972
2989
|
},
|
|
2973
2990
|
contextWindow: 262144,
|
|
2974
|
-
maxTokens:
|
|
2991
|
+
maxTokens: 262144,
|
|
2975
2992
|
},
|
|
2976
2993
|
"openai/gpt-5.2-chat": {
|
|
2977
2994
|
id: "openai/gpt-5.2-chat",
|
|
@@ -3254,13 +3271,13 @@ export const MODELS = {
|
|
|
3254
3271
|
reasoning: true,
|
|
3255
3272
|
input: ["text"],
|
|
3256
3273
|
cost: {
|
|
3257
|
-
input: 0.
|
|
3258
|
-
output: 0.
|
|
3259
|
-
cacheRead: 0
|
|
3274
|
+
input: 0.26,
|
|
3275
|
+
output: 0.38,
|
|
3276
|
+
cacheRead: 0,
|
|
3260
3277
|
cacheWrite: 0,
|
|
3261
3278
|
},
|
|
3262
3279
|
contextWindow: 163840,
|
|
3263
|
-
maxTokens:
|
|
3280
|
+
maxTokens: 65536,
|
|
3264
3281
|
},
|
|
3265
3282
|
"prime-intellect/intellect-3": {
|
|
3266
3283
|
id: "prime-intellect/intellect-3",
|
|
@@ -6101,34 +6118,34 @@ export const MODELS = {
|
|
|
6101
6118
|
contextWindow: 200000,
|
|
6102
6119
|
maxTokens: 8192,
|
|
6103
6120
|
},
|
|
6104
|
-
"mistralai/ministral-
|
|
6105
|
-
id: "mistralai/ministral-
|
|
6106
|
-
name: "Mistral: Ministral
|
|
6121
|
+
"mistralai/ministral-8b": {
|
|
6122
|
+
id: "mistralai/ministral-8b",
|
|
6123
|
+
name: "Mistral: Ministral 8B",
|
|
6107
6124
|
api: "openai-completions",
|
|
6108
6125
|
provider: "openrouter",
|
|
6109
6126
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6110
6127
|
reasoning: false,
|
|
6111
6128
|
input: ["text"],
|
|
6112
6129
|
cost: {
|
|
6113
|
-
input: 0.
|
|
6114
|
-
output: 0.
|
|
6130
|
+
input: 0.09999999999999999,
|
|
6131
|
+
output: 0.09999999999999999,
|
|
6115
6132
|
cacheRead: 0,
|
|
6116
6133
|
cacheWrite: 0,
|
|
6117
6134
|
},
|
|
6118
6135
|
contextWindow: 131072,
|
|
6119
6136
|
maxTokens: 4096,
|
|
6120
6137
|
},
|
|
6121
|
-
"mistralai/ministral-
|
|
6122
|
-
id: "mistralai/ministral-
|
|
6123
|
-
name: "Mistral: Ministral
|
|
6138
|
+
"mistralai/ministral-3b": {
|
|
6139
|
+
id: "mistralai/ministral-3b",
|
|
6140
|
+
name: "Mistral: Ministral 3B",
|
|
6124
6141
|
api: "openai-completions",
|
|
6125
6142
|
provider: "openrouter",
|
|
6126
6143
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6127
6144
|
reasoning: false,
|
|
6128
6145
|
input: ["text"],
|
|
6129
6146
|
cost: {
|
|
6130
|
-
input: 0.
|
|
6131
|
-
output: 0.
|
|
6147
|
+
input: 0.04,
|
|
6148
|
+
output: 0.04,
|
|
6132
6149
|
cacheRead: 0,
|
|
6133
6150
|
cacheWrite: 0,
|
|
6134
6151
|
},
|
|
@@ -6195,13 +6212,13 @@ export const MODELS = {
|
|
|
6195
6212
|
reasoning: false,
|
|
6196
6213
|
input: ["text"],
|
|
6197
6214
|
cost: {
|
|
6198
|
-
input: 0.
|
|
6199
|
-
output: 0.
|
|
6215
|
+
input: 0.12,
|
|
6216
|
+
output: 0.39,
|
|
6200
6217
|
cacheRead: 0,
|
|
6201
6218
|
cacheWrite: 0,
|
|
6202
6219
|
},
|
|
6203
6220
|
contextWindow: 32768,
|
|
6204
|
-
maxTokens:
|
|
6221
|
+
maxTokens: 16384,
|
|
6205
6222
|
},
|
|
6206
6223
|
"mistralai/pixtral-12b": {
|
|
6207
6224
|
id: "mistralai/pixtral-12b",
|
|
@@ -6305,23 +6322,6 @@ export const MODELS = {
|
|
|
6305
6322
|
contextWindow: 128000,
|
|
6306
6323
|
maxTokens: 16384,
|
|
6307
6324
|
},
|
|
6308
|
-
"meta-llama/llama-3.1-70b-instruct": {
|
|
6309
|
-
id: "meta-llama/llama-3.1-70b-instruct",
|
|
6310
|
-
name: "Meta: Llama 3.1 70B Instruct",
|
|
6311
|
-
api: "openai-completions",
|
|
6312
|
-
provider: "openrouter",
|
|
6313
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
6314
|
-
reasoning: false,
|
|
6315
|
-
input: ["text"],
|
|
6316
|
-
cost: {
|
|
6317
|
-
input: 0.39999999999999997,
|
|
6318
|
-
output: 0.39999999999999997,
|
|
6319
|
-
cacheRead: 0,
|
|
6320
|
-
cacheWrite: 0,
|
|
6321
|
-
},
|
|
6322
|
-
contextWindow: 131072,
|
|
6323
|
-
maxTokens: 4096,
|
|
6324
|
-
},
|
|
6325
6325
|
"meta-llama/llama-3.1-405b-instruct": {
|
|
6326
6326
|
id: "meta-llama/llama-3.1-405b-instruct",
|
|
6327
6327
|
name: "Meta: Llama 3.1 405B Instruct",
|
|
@@ -6356,6 +6356,23 @@ export const MODELS = {
|
|
|
6356
6356
|
contextWindow: 131072,
|
|
6357
6357
|
maxTokens: 16384,
|
|
6358
6358
|
},
|
|
6359
|
+
"meta-llama/llama-3.1-70b-instruct": {
|
|
6360
|
+
id: "meta-llama/llama-3.1-70b-instruct",
|
|
6361
|
+
name: "Meta: Llama 3.1 70B Instruct",
|
|
6362
|
+
api: "openai-completions",
|
|
6363
|
+
provider: "openrouter",
|
|
6364
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6365
|
+
reasoning: false,
|
|
6366
|
+
input: ["text"],
|
|
6367
|
+
cost: {
|
|
6368
|
+
input: 0.39999999999999997,
|
|
6369
|
+
output: 0.39999999999999997,
|
|
6370
|
+
cacheRead: 0,
|
|
6371
|
+
cacheWrite: 0,
|
|
6372
|
+
},
|
|
6373
|
+
contextWindow: 131072,
|
|
6374
|
+
maxTokens: 4096,
|
|
6375
|
+
},
|
|
6359
6376
|
"mistralai/mistral-nemo": {
|
|
6360
6377
|
id: "mistralai/mistral-nemo",
|
|
6361
6378
|
name: "Mistral: Mistral Nemo",
|
|
@@ -6373,9 +6390,9 @@ export const MODELS = {
|
|
|
6373
6390
|
contextWindow: 131072,
|
|
6374
6391
|
maxTokens: 16384,
|
|
6375
6392
|
},
|
|
6376
|
-
"openai/gpt-4o-mini": {
|
|
6377
|
-
id: "openai/gpt-4o-mini",
|
|
6378
|
-
name: "OpenAI: GPT-4o-mini",
|
|
6393
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
6394
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
6395
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
6379
6396
|
api: "openai-completions",
|
|
6380
6397
|
provider: "openrouter",
|
|
6381
6398
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6390,9 +6407,9 @@ export const MODELS = {
|
|
|
6390
6407
|
contextWindow: 128000,
|
|
6391
6408
|
maxTokens: 16384,
|
|
6392
6409
|
},
|
|
6393
|
-
"openai/gpt-4o-mini
|
|
6394
|
-
id: "openai/gpt-4o-mini
|
|
6395
|
-
name: "OpenAI: GPT-4o-mini
|
|
6410
|
+
"openai/gpt-4o-mini": {
|
|
6411
|
+
id: "openai/gpt-4o-mini",
|
|
6412
|
+
name: "OpenAI: GPT-4o-mini",
|
|
6396
6413
|
api: "openai-completions",
|
|
6397
6414
|
provider: "openrouter",
|
|
6398
6415
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6492,6 +6509,23 @@ export const MODELS = {
|
|
|
6492
6509
|
contextWindow: 128000,
|
|
6493
6510
|
maxTokens: 4096,
|
|
6494
6511
|
},
|
|
6512
|
+
"openai/gpt-4o-2024-05-13": {
|
|
6513
|
+
id: "openai/gpt-4o-2024-05-13",
|
|
6514
|
+
name: "OpenAI: GPT-4o (2024-05-13)",
|
|
6515
|
+
api: "openai-completions",
|
|
6516
|
+
provider: "openrouter",
|
|
6517
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6518
|
+
reasoning: false,
|
|
6519
|
+
input: ["text", "image"],
|
|
6520
|
+
cost: {
|
|
6521
|
+
input: 5,
|
|
6522
|
+
output: 15,
|
|
6523
|
+
cacheRead: 0,
|
|
6524
|
+
cacheWrite: 0,
|
|
6525
|
+
},
|
|
6526
|
+
contextWindow: 128000,
|
|
6527
|
+
maxTokens: 4096,
|
|
6528
|
+
},
|
|
6495
6529
|
"openai/gpt-4o": {
|
|
6496
6530
|
id: "openai/gpt-4o",
|
|
6497
6531
|
name: "OpenAI: GPT-4o",
|
|
@@ -6526,22 +6560,22 @@ export const MODELS = {
|
|
|
6526
6560
|
contextWindow: 128000,
|
|
6527
6561
|
maxTokens: 64000,
|
|
6528
6562
|
},
|
|
6529
|
-
"
|
|
6530
|
-
id: "
|
|
6531
|
-
name: "
|
|
6563
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
6564
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
6565
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
6532
6566
|
api: "openai-completions",
|
|
6533
6567
|
provider: "openrouter",
|
|
6534
6568
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6535
6569
|
reasoning: false,
|
|
6536
|
-
input: ["text"
|
|
6570
|
+
input: ["text"],
|
|
6537
6571
|
cost: {
|
|
6538
|
-
input:
|
|
6539
|
-
output:
|
|
6572
|
+
input: 0.3,
|
|
6573
|
+
output: 0.39999999999999997,
|
|
6540
6574
|
cacheRead: 0,
|
|
6541
6575
|
cacheWrite: 0,
|
|
6542
6576
|
},
|
|
6543
|
-
contextWindow:
|
|
6544
|
-
maxTokens:
|
|
6577
|
+
contextWindow: 8192,
|
|
6578
|
+
maxTokens: 16384,
|
|
6545
6579
|
},
|
|
6546
6580
|
"meta-llama/llama-3-8b-instruct": {
|
|
6547
6581
|
id: "meta-llama/llama-3-8b-instruct",
|
|
@@ -6560,23 +6594,6 @@ export const MODELS = {
|
|
|
6560
6594
|
contextWindow: 8192,
|
|
6561
6595
|
maxTokens: 16384,
|
|
6562
6596
|
},
|
|
6563
|
-
"meta-llama/llama-3-70b-instruct": {
|
|
6564
|
-
id: "meta-llama/llama-3-70b-instruct",
|
|
6565
|
-
name: "Meta: Llama 3 70B Instruct",
|
|
6566
|
-
api: "openai-completions",
|
|
6567
|
-
provider: "openrouter",
|
|
6568
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
6569
|
-
reasoning: false,
|
|
6570
|
-
input: ["text"],
|
|
6571
|
-
cost: {
|
|
6572
|
-
input: 0.3,
|
|
6573
|
-
output: 0.39999999999999997,
|
|
6574
|
-
cacheRead: 0,
|
|
6575
|
-
cacheWrite: 0,
|
|
6576
|
-
},
|
|
6577
|
-
contextWindow: 8192,
|
|
6578
|
-
maxTokens: 16384,
|
|
6579
|
-
},
|
|
6580
6597
|
"mistralai/mixtral-8x22b-instruct": {
|
|
6581
6598
|
id: "mistralai/mixtral-8x22b-instruct",
|
|
6582
6599
|
name: "Mistral: Mixtral 8x22B Instruct",
|
|
@@ -6662,38 +6679,38 @@ export const MODELS = {
|
|
|
6662
6679
|
contextWindow: 128000,
|
|
6663
6680
|
maxTokens: 4096,
|
|
6664
6681
|
},
|
|
6665
|
-
"openai/gpt-
|
|
6666
|
-
id: "openai/gpt-
|
|
6667
|
-
name: "OpenAI: GPT-
|
|
6682
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
6683
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
6684
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
6668
6685
|
api: "openai-completions",
|
|
6669
6686
|
provider: "openrouter",
|
|
6670
6687
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6671
6688
|
reasoning: false,
|
|
6672
6689
|
input: ["text"],
|
|
6673
6690
|
cost: {
|
|
6674
|
-
input:
|
|
6675
|
-
output:
|
|
6691
|
+
input: 1,
|
|
6692
|
+
output: 2,
|
|
6676
6693
|
cacheRead: 0,
|
|
6677
6694
|
cacheWrite: 0,
|
|
6678
6695
|
},
|
|
6679
|
-
contextWindow:
|
|
6696
|
+
contextWindow: 4095,
|
|
6680
6697
|
maxTokens: 4096,
|
|
6681
6698
|
},
|
|
6682
|
-
"openai/gpt-
|
|
6683
|
-
id: "openai/gpt-
|
|
6684
|
-
name: "OpenAI: GPT-
|
|
6699
|
+
"openai/gpt-4-turbo-preview": {
|
|
6700
|
+
id: "openai/gpt-4-turbo-preview",
|
|
6701
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
6685
6702
|
api: "openai-completions",
|
|
6686
6703
|
provider: "openrouter",
|
|
6687
6704
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6688
6705
|
reasoning: false,
|
|
6689
6706
|
input: ["text"],
|
|
6690
6707
|
cost: {
|
|
6691
|
-
input:
|
|
6692
|
-
output:
|
|
6708
|
+
input: 10,
|
|
6709
|
+
output: 30,
|
|
6693
6710
|
cacheRead: 0,
|
|
6694
6711
|
cacheWrite: 0,
|
|
6695
6712
|
},
|
|
6696
|
-
contextWindow:
|
|
6713
|
+
contextWindow: 128000,
|
|
6697
6714
|
maxTokens: 4096,
|
|
6698
6715
|
},
|
|
6699
6716
|
"mistralai/mistral-tiny": {
|
|
@@ -6764,9 +6781,9 @@ export const MODELS = {
|
|
|
6764
6781
|
contextWindow: 16385,
|
|
6765
6782
|
maxTokens: 4096,
|
|
6766
6783
|
},
|
|
6767
|
-
"openai/gpt-4
|
|
6768
|
-
id: "openai/gpt-4
|
|
6769
|
-
name: "OpenAI: GPT-4
|
|
6784
|
+
"openai/gpt-4": {
|
|
6785
|
+
id: "openai/gpt-4",
|
|
6786
|
+
name: "OpenAI: GPT-4",
|
|
6770
6787
|
api: "openai-completions",
|
|
6771
6788
|
provider: "openrouter",
|
|
6772
6789
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6781,38 +6798,38 @@ export const MODELS = {
|
|
|
6781
6798
|
contextWindow: 8191,
|
|
6782
6799
|
maxTokens: 4096,
|
|
6783
6800
|
},
|
|
6784
|
-
"openai/gpt-
|
|
6785
|
-
id: "openai/gpt-
|
|
6786
|
-
name: "OpenAI: GPT-
|
|
6801
|
+
"openai/gpt-4-0314": {
|
|
6802
|
+
id: "openai/gpt-4-0314",
|
|
6803
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
6787
6804
|
api: "openai-completions",
|
|
6788
6805
|
provider: "openrouter",
|
|
6789
6806
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6790
6807
|
reasoning: false,
|
|
6791
6808
|
input: ["text"],
|
|
6792
6809
|
cost: {
|
|
6793
|
-
input:
|
|
6794
|
-
output:
|
|
6810
|
+
input: 30,
|
|
6811
|
+
output: 60,
|
|
6795
6812
|
cacheRead: 0,
|
|
6796
6813
|
cacheWrite: 0,
|
|
6797
6814
|
},
|
|
6798
|
-
contextWindow:
|
|
6815
|
+
contextWindow: 8191,
|
|
6799
6816
|
maxTokens: 4096,
|
|
6800
6817
|
},
|
|
6801
|
-
"openai/gpt-
|
|
6802
|
-
id: "openai/gpt-
|
|
6803
|
-
name: "OpenAI: GPT-
|
|
6818
|
+
"openai/gpt-3.5-turbo": {
|
|
6819
|
+
id: "openai/gpt-3.5-turbo",
|
|
6820
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
6804
6821
|
api: "openai-completions",
|
|
6805
6822
|
provider: "openrouter",
|
|
6806
6823
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6807
6824
|
reasoning: false,
|
|
6808
6825
|
input: ["text"],
|
|
6809
6826
|
cost: {
|
|
6810
|
-
input:
|
|
6811
|
-
output:
|
|
6827
|
+
input: 0.5,
|
|
6828
|
+
output: 1.5,
|
|
6812
6829
|
cacheRead: 0,
|
|
6813
6830
|
cacheWrite: 0,
|
|
6814
6831
|
},
|
|
6815
|
-
contextWindow:
|
|
6832
|
+
contextWindow: 16385,
|
|
6816
6833
|
maxTokens: 4096,
|
|
6817
6834
|
},
|
|
6818
6835
|
"openrouter/auto": {
|