@mariozechner/pi-ai 0.18.1 → 0.18.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +169 -16
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +246 -93
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +2 -0
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +8 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -839,6 +839,23 @@ export const MODELS = {
|
|
|
839
839
|
contextWindow: 200000,
|
|
840
840
|
maxTokens: 100000,
|
|
841
841
|
},
|
|
842
|
+
"gpt-5.2-pro": {
|
|
843
|
+
id: "gpt-5.2-pro",
|
|
844
|
+
name: "GPT-5.2 Pro",
|
|
845
|
+
api: "openai-responses",
|
|
846
|
+
provider: "openai",
|
|
847
|
+
baseUrl: "https://api.openai.com/v1",
|
|
848
|
+
reasoning: true,
|
|
849
|
+
input: ["text", "image"],
|
|
850
|
+
cost: {
|
|
851
|
+
input: 21,
|
|
852
|
+
output: 168,
|
|
853
|
+
cacheRead: 0,
|
|
854
|
+
cacheWrite: 0,
|
|
855
|
+
},
|
|
856
|
+
contextWindow: 400000,
|
|
857
|
+
maxTokens: 128000,
|
|
858
|
+
},
|
|
842
859
|
"gpt-4-turbo": {
|
|
843
860
|
id: "gpt-4-turbo",
|
|
844
861
|
name: "GPT-4 Turbo",
|
|
@@ -890,6 +907,23 @@ export const MODELS = {
|
|
|
890
907
|
contextWindow: 200000,
|
|
891
908
|
maxTokens: 100000,
|
|
892
909
|
},
|
|
910
|
+
"gpt-5.2-chat-latest": {
|
|
911
|
+
id: "gpt-5.2-chat-latest",
|
|
912
|
+
name: "GPT-5.2 Chat",
|
|
913
|
+
api: "openai-responses",
|
|
914
|
+
provider: "openai",
|
|
915
|
+
baseUrl: "https://api.openai.com/v1",
|
|
916
|
+
reasoning: true,
|
|
917
|
+
input: ["text", "image"],
|
|
918
|
+
cost: {
|
|
919
|
+
input: 1.75,
|
|
920
|
+
output: 14,
|
|
921
|
+
cacheRead: 0.175,
|
|
922
|
+
cacheWrite: 0,
|
|
923
|
+
},
|
|
924
|
+
contextWindow: 128000,
|
|
925
|
+
maxTokens: 16384,
|
|
926
|
+
},
|
|
893
927
|
"gpt-5.1": {
|
|
894
928
|
id: "gpt-5.1",
|
|
895
929
|
name: "GPT-5.1",
|
|
@@ -1179,6 +1213,23 @@ export const MODELS = {
|
|
|
1179
1213
|
contextWindow: 400000,
|
|
1180
1214
|
maxTokens: 272000,
|
|
1181
1215
|
},
|
|
1216
|
+
"gpt-5.2": {
|
|
1217
|
+
id: "gpt-5.2",
|
|
1218
|
+
name: "GPT-5.2",
|
|
1219
|
+
api: "openai-responses",
|
|
1220
|
+
provider: "openai",
|
|
1221
|
+
baseUrl: "https://api.openai.com/v1",
|
|
1222
|
+
reasoning: true,
|
|
1223
|
+
input: ["text", "image"],
|
|
1224
|
+
cost: {
|
|
1225
|
+
input: 1.75,
|
|
1226
|
+
output: 14,
|
|
1227
|
+
cacheRead: 0.175,
|
|
1228
|
+
cacheWrite: 0,
|
|
1229
|
+
},
|
|
1230
|
+
contextWindow: 400000,
|
|
1231
|
+
maxTokens: 128000,
|
|
1232
|
+
},
|
|
1182
1233
|
"gpt-5.1-chat-latest": {
|
|
1183
1234
|
id: "gpt-5.1-chat-latest",
|
|
1184
1235
|
name: "GPT-5.1 Chat",
|
|
@@ -2090,6 +2141,23 @@ export const MODELS = {
|
|
|
2090
2141
|
contextWindow: 128000,
|
|
2091
2142
|
maxTokens: 128000,
|
|
2092
2143
|
},
|
|
2144
|
+
"mistral-small-2506": {
|
|
2145
|
+
id: "mistral-small-2506",
|
|
2146
|
+
name: "Mistral Small 3.2",
|
|
2147
|
+
api: "openai-completions",
|
|
2148
|
+
provider: "mistral",
|
|
2149
|
+
baseUrl: "https://api.mistral.ai/v1",
|
|
2150
|
+
reasoning: false,
|
|
2151
|
+
input: ["text", "image"],
|
|
2152
|
+
cost: {
|
|
2153
|
+
input: 0.1,
|
|
2154
|
+
output: 0.3,
|
|
2155
|
+
cacheRead: 0,
|
|
2156
|
+
cacheWrite: 0,
|
|
2157
|
+
},
|
|
2158
|
+
contextWindow: 128000,
|
|
2159
|
+
maxTokens: 16384,
|
|
2160
|
+
},
|
|
2093
2161
|
"ministral-3b-latest": {
|
|
2094
2162
|
id: "ministral-3b-latest",
|
|
2095
2163
|
name: "Ministral 3B",
|
|
@@ -2398,6 +2466,57 @@ export const MODELS = {
|
|
|
2398
2466
|
},
|
|
2399
2467
|
},
|
|
2400
2468
|
openrouter: {
|
|
2469
|
+
"openai/gpt-5.2-chat": {
|
|
2470
|
+
id: "openai/gpt-5.2-chat",
|
|
2471
|
+
name: "OpenAI: GPT-5.2 Chat",
|
|
2472
|
+
api: "openai-completions",
|
|
2473
|
+
provider: "openrouter",
|
|
2474
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2475
|
+
reasoning: false,
|
|
2476
|
+
input: ["text", "image"],
|
|
2477
|
+
cost: {
|
|
2478
|
+
input: 1.75,
|
|
2479
|
+
output: 14,
|
|
2480
|
+
cacheRead: 0.175,
|
|
2481
|
+
cacheWrite: 0,
|
|
2482
|
+
},
|
|
2483
|
+
contextWindow: 128000,
|
|
2484
|
+
maxTokens: 16384,
|
|
2485
|
+
},
|
|
2486
|
+
"openai/gpt-5.2-pro": {
|
|
2487
|
+
id: "openai/gpt-5.2-pro",
|
|
2488
|
+
name: "OpenAI: GPT-5.2 Pro",
|
|
2489
|
+
api: "openai-completions",
|
|
2490
|
+
provider: "openrouter",
|
|
2491
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2492
|
+
reasoning: true,
|
|
2493
|
+
input: ["text", "image"],
|
|
2494
|
+
cost: {
|
|
2495
|
+
input: 21,
|
|
2496
|
+
output: 168,
|
|
2497
|
+
cacheRead: 0,
|
|
2498
|
+
cacheWrite: 0,
|
|
2499
|
+
},
|
|
2500
|
+
contextWindow: 400000,
|
|
2501
|
+
maxTokens: 128000,
|
|
2502
|
+
},
|
|
2503
|
+
"openai/gpt-5.2": {
|
|
2504
|
+
id: "openai/gpt-5.2",
|
|
2505
|
+
name: "OpenAI: GPT-5.2",
|
|
2506
|
+
api: "openai-completions",
|
|
2507
|
+
provider: "openrouter",
|
|
2508
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2509
|
+
reasoning: true,
|
|
2510
|
+
input: ["text", "image"],
|
|
2511
|
+
cost: {
|
|
2512
|
+
input: 1.75,
|
|
2513
|
+
output: 14,
|
|
2514
|
+
cacheRead: 0.175,
|
|
2515
|
+
cacheWrite: 0,
|
|
2516
|
+
},
|
|
2517
|
+
contextWindow: 400000,
|
|
2518
|
+
maxTokens: 128000,
|
|
2519
|
+
},
|
|
2401
2520
|
"mistralai/devstral-2512:free": {
|
|
2402
2521
|
id: "mistralai/devstral-2512:free",
|
|
2403
2522
|
name: "Mistral: Devstral 2 2512 (free)",
|
|
@@ -2415,6 +2534,23 @@ export const MODELS = {
|
|
|
2415
2534
|
contextWindow: 262144,
|
|
2416
2535
|
maxTokens: 4096,
|
|
2417
2536
|
},
|
|
2537
|
+
"mistralai/devstral-2512": {
|
|
2538
|
+
id: "mistralai/devstral-2512",
|
|
2539
|
+
name: "Mistral: Devstral 2 2512",
|
|
2540
|
+
api: "openai-completions",
|
|
2541
|
+
provider: "openrouter",
|
|
2542
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2543
|
+
reasoning: false,
|
|
2544
|
+
input: ["text"],
|
|
2545
|
+
cost: {
|
|
2546
|
+
input: 0.15,
|
|
2547
|
+
output: 0.6,
|
|
2548
|
+
cacheRead: 0,
|
|
2549
|
+
cacheWrite: 0,
|
|
2550
|
+
},
|
|
2551
|
+
contextWindow: 262144,
|
|
2552
|
+
maxTokens: 65536,
|
|
2553
|
+
},
|
|
2418
2554
|
"relace/relace-search": {
|
|
2419
2555
|
id: "relace/relace-search",
|
|
2420
2556
|
name: "Relace: Relace Search",
|
|
@@ -2449,6 +2585,23 @@ export const MODELS = {
|
|
|
2449
2585
|
contextWindow: 131072,
|
|
2450
2586
|
maxTokens: 24000,
|
|
2451
2587
|
},
|
|
2588
|
+
"nex-agi/deepseek-v3.1-nex-n1:free": {
|
|
2589
|
+
id: "nex-agi/deepseek-v3.1-nex-n1:free",
|
|
2590
|
+
name: "Nex AGI: DeepSeek V3.1 Nex N1 (free)",
|
|
2591
|
+
api: "openai-completions",
|
|
2592
|
+
provider: "openrouter",
|
|
2593
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
2594
|
+
reasoning: false,
|
|
2595
|
+
input: ["text"],
|
|
2596
|
+
cost: {
|
|
2597
|
+
input: 0,
|
|
2598
|
+
output: 0,
|
|
2599
|
+
cacheRead: 0,
|
|
2600
|
+
cacheWrite: 0,
|
|
2601
|
+
},
|
|
2602
|
+
contextWindow: 131072,
|
|
2603
|
+
maxTokens: 163840,
|
|
2604
|
+
},
|
|
2452
2605
|
"openai/gpt-5.1-codex-max": {
|
|
2453
2606
|
id: "openai/gpt-5.1-codex-max",
|
|
2454
2607
|
name: "OpenAI: GPT-5.1-Codex-Max",
|
|
@@ -2611,9 +2764,9 @@ export const MODELS = {
|
|
|
2611
2764
|
reasoning: true,
|
|
2612
2765
|
input: ["text"],
|
|
2613
2766
|
cost: {
|
|
2614
|
-
input: 0.
|
|
2615
|
-
output: 0.
|
|
2616
|
-
cacheRead: 0.
|
|
2767
|
+
input: 0.25,
|
|
2768
|
+
output: 0.38,
|
|
2769
|
+
cacheRead: 0.19,
|
|
2617
2770
|
cacheWrite: 0,
|
|
2618
2771
|
},
|
|
2619
2772
|
contextWindow: 163840,
|
|
@@ -3155,13 +3308,13 @@ export const MODELS = {
|
|
|
3155
3308
|
reasoning: true,
|
|
3156
3309
|
input: ["text"],
|
|
3157
3310
|
cost: {
|
|
3158
|
-
input: 0.
|
|
3159
|
-
output: 1.
|
|
3160
|
-
cacheRead: 0
|
|
3311
|
+
input: 0.44,
|
|
3312
|
+
output: 1.76,
|
|
3313
|
+
cacheRead: 0,
|
|
3161
3314
|
cacheWrite: 0,
|
|
3162
3315
|
},
|
|
3163
|
-
contextWindow:
|
|
3164
|
-
maxTokens:
|
|
3316
|
+
contextWindow: 204800,
|
|
3317
|
+
maxTokens: 131072,
|
|
3165
3318
|
},
|
|
3166
3319
|
"anthropic/claude-sonnet-4.5": {
|
|
3167
3320
|
id: "anthropic/claude-sonnet-4.5",
|
|
@@ -3444,13 +3597,13 @@ export const MODELS = {
|
|
|
3444
3597
|
reasoning: false,
|
|
3445
3598
|
input: ["text"],
|
|
3446
3599
|
cost: {
|
|
3447
|
-
input: 0.
|
|
3448
|
-
output:
|
|
3600
|
+
input: 0.09,
|
|
3601
|
+
output: 1.1,
|
|
3449
3602
|
cacheRead: 0,
|
|
3450
3603
|
cacheWrite: 0,
|
|
3451
3604
|
},
|
|
3452
3605
|
contextWindow: 262144,
|
|
3453
|
-
maxTokens:
|
|
3606
|
+
maxTokens: 4096,
|
|
3454
3607
|
},
|
|
3455
3608
|
"meituan/longcat-flash-chat:free": {
|
|
3456
3609
|
id: "meituan/longcat-flash-chat:free",
|
|
@@ -4189,7 +4342,7 @@ export const MODELS = {
|
|
|
4189
4342
|
api: "openai-completions",
|
|
4190
4343
|
provider: "openrouter",
|
|
4191
4344
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
4192
|
-
reasoning:
|
|
4345
|
+
reasoning: false,
|
|
4193
4346
|
input: ["text"],
|
|
4194
4347
|
cost: {
|
|
4195
4348
|
input: 0.071,
|
|
@@ -5280,13 +5433,13 @@ export const MODELS = {
|
|
|
5280
5433
|
reasoning: false,
|
|
5281
5434
|
input: ["text"],
|
|
5282
5435
|
cost: {
|
|
5283
|
-
input: 0.
|
|
5436
|
+
input: 0.09999999999999999,
|
|
5284
5437
|
output: 0.32,
|
|
5285
5438
|
cacheRead: 0,
|
|
5286
5439
|
cacheWrite: 0,
|
|
5287
5440
|
},
|
|
5288
5441
|
contextWindow: 131072,
|
|
5289
|
-
maxTokens:
|
|
5442
|
+
maxTokens: 16384,
|
|
5290
5443
|
},
|
|
5291
5444
|
"amazon/nova-lite-v1": {
|
|
5292
5445
|
id: "amazon/nova-lite-v1",
|
|
@@ -5475,34 +5628,34 @@ export const MODELS = {
|
|
|
5475
5628
|
contextWindow: 200000,
|
|
5476
5629
|
maxTokens: 8192,
|
|
5477
5630
|
},
|
|
5478
|
-
"mistralai/ministral-
|
|
5479
|
-
id: "mistralai/ministral-
|
|
5480
|
-
name: "Mistral: Ministral
|
|
5631
|
+
"mistralai/ministral-8b": {
|
|
5632
|
+
id: "mistralai/ministral-8b",
|
|
5633
|
+
name: "Mistral: Ministral 8B",
|
|
5481
5634
|
api: "openai-completions",
|
|
5482
5635
|
provider: "openrouter",
|
|
5483
5636
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5484
5637
|
reasoning: false,
|
|
5485
5638
|
input: ["text"],
|
|
5486
5639
|
cost: {
|
|
5487
|
-
input: 0.
|
|
5488
|
-
output: 0.
|
|
5640
|
+
input: 0.09999999999999999,
|
|
5641
|
+
output: 0.09999999999999999,
|
|
5489
5642
|
cacheRead: 0,
|
|
5490
5643
|
cacheWrite: 0,
|
|
5491
5644
|
},
|
|
5492
5645
|
contextWindow: 131072,
|
|
5493
5646
|
maxTokens: 4096,
|
|
5494
5647
|
},
|
|
5495
|
-
"mistralai/ministral-
|
|
5496
|
-
id: "mistralai/ministral-
|
|
5497
|
-
name: "Mistral: Ministral
|
|
5648
|
+
"mistralai/ministral-3b": {
|
|
5649
|
+
id: "mistralai/ministral-3b",
|
|
5650
|
+
name: "Mistral: Ministral 3B",
|
|
5498
5651
|
api: "openai-completions",
|
|
5499
5652
|
provider: "openrouter",
|
|
5500
5653
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5501
5654
|
reasoning: false,
|
|
5502
5655
|
input: ["text"],
|
|
5503
5656
|
cost: {
|
|
5504
|
-
input: 0.
|
|
5505
|
-
output: 0.
|
|
5657
|
+
input: 0.04,
|
|
5658
|
+
output: 0.04,
|
|
5506
5659
|
cacheRead: 0,
|
|
5507
5660
|
cacheWrite: 0,
|
|
5508
5661
|
},
|
|
@@ -5679,6 +5832,23 @@ export const MODELS = {
|
|
|
5679
5832
|
contextWindow: 128000,
|
|
5680
5833
|
maxTokens: 16384,
|
|
5681
5834
|
},
|
|
5835
|
+
"meta-llama/llama-3.1-8b-instruct": {
|
|
5836
|
+
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5837
|
+
name: "Meta: Llama 3.1 8B Instruct",
|
|
5838
|
+
api: "openai-completions",
|
|
5839
|
+
provider: "openrouter",
|
|
5840
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5841
|
+
reasoning: false,
|
|
5842
|
+
input: ["text"],
|
|
5843
|
+
cost: {
|
|
5844
|
+
input: 0.02,
|
|
5845
|
+
output: 0.03,
|
|
5846
|
+
cacheRead: 0,
|
|
5847
|
+
cacheWrite: 0,
|
|
5848
|
+
},
|
|
5849
|
+
contextWindow: 131072,
|
|
5850
|
+
maxTokens: 16384,
|
|
5851
|
+
},
|
|
5682
5852
|
"meta-llama/llama-3.1-405b-instruct": {
|
|
5683
5853
|
id: "meta-llama/llama-3.1-405b-instruct",
|
|
5684
5854
|
name: "Meta: Llama 3.1 405B Instruct",
|
|
@@ -5713,23 +5883,6 @@ export const MODELS = {
|
|
|
5713
5883
|
contextWindow: 131072,
|
|
5714
5884
|
maxTokens: 4096,
|
|
5715
5885
|
},
|
|
5716
|
-
"meta-llama/llama-3.1-8b-instruct": {
|
|
5717
|
-
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5718
|
-
name: "Meta: Llama 3.1 8B Instruct",
|
|
5719
|
-
api: "openai-completions",
|
|
5720
|
-
provider: "openrouter",
|
|
5721
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5722
|
-
reasoning: false,
|
|
5723
|
-
input: ["text"],
|
|
5724
|
-
cost: {
|
|
5725
|
-
input: 0.02,
|
|
5726
|
-
output: 0.03,
|
|
5727
|
-
cacheRead: 0,
|
|
5728
|
-
cacheWrite: 0,
|
|
5729
|
-
},
|
|
5730
|
-
contextWindow: 131072,
|
|
5731
|
-
maxTokens: 16384,
|
|
5732
|
-
},
|
|
5733
5886
|
"mistralai/mistral-nemo": {
|
|
5734
5887
|
id: "mistralai/mistral-nemo",
|
|
5735
5888
|
name: "Mistral: Mistral Nemo",
|
|
@@ -5866,6 +6019,23 @@ export const MODELS = {
|
|
|
5866
6019
|
contextWindow: 128000,
|
|
5867
6020
|
maxTokens: 4096,
|
|
5868
6021
|
},
|
|
6022
|
+
"openai/gpt-4o-2024-05-13": {
|
|
6023
|
+
id: "openai/gpt-4o-2024-05-13",
|
|
6024
|
+
name: "OpenAI: GPT-4o (2024-05-13)",
|
|
6025
|
+
api: "openai-completions",
|
|
6026
|
+
provider: "openrouter",
|
|
6027
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6028
|
+
reasoning: false,
|
|
6029
|
+
input: ["text", "image"],
|
|
6030
|
+
cost: {
|
|
6031
|
+
input: 5,
|
|
6032
|
+
output: 15,
|
|
6033
|
+
cacheRead: 0,
|
|
6034
|
+
cacheWrite: 0,
|
|
6035
|
+
},
|
|
6036
|
+
contextWindow: 128000,
|
|
6037
|
+
maxTokens: 4096,
|
|
6038
|
+
},
|
|
5869
6039
|
"openai/gpt-4o": {
|
|
5870
6040
|
id: "openai/gpt-4o",
|
|
5871
6041
|
name: "OpenAI: GPT-4o",
|
|
@@ -5900,22 +6070,22 @@ export const MODELS = {
|
|
|
5900
6070
|
contextWindow: 128000,
|
|
5901
6071
|
maxTokens: 64000,
|
|
5902
6072
|
},
|
|
5903
|
-
"
|
|
5904
|
-
id: "
|
|
5905
|
-
name: "
|
|
6073
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
6074
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
6075
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
5906
6076
|
api: "openai-completions",
|
|
5907
6077
|
provider: "openrouter",
|
|
5908
6078
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5909
6079
|
reasoning: false,
|
|
5910
|
-
input: ["text"
|
|
6080
|
+
input: ["text"],
|
|
5911
6081
|
cost: {
|
|
5912
|
-
input:
|
|
5913
|
-
output:
|
|
6082
|
+
input: 0.3,
|
|
6083
|
+
output: 0.39999999999999997,
|
|
5914
6084
|
cacheRead: 0,
|
|
5915
6085
|
cacheWrite: 0,
|
|
5916
6086
|
},
|
|
5917
|
-
contextWindow:
|
|
5918
|
-
maxTokens:
|
|
6087
|
+
contextWindow: 8192,
|
|
6088
|
+
maxTokens: 16384,
|
|
5919
6089
|
},
|
|
5920
6090
|
"meta-llama/llama-3-8b-instruct": {
|
|
5921
6091
|
id: "meta-llama/llama-3-8b-instruct",
|
|
@@ -5934,23 +6104,6 @@ export const MODELS = {
|
|
|
5934
6104
|
contextWindow: 8192,
|
|
5935
6105
|
maxTokens: 16384,
|
|
5936
6106
|
},
|
|
5937
|
-
"meta-llama/llama-3-70b-instruct": {
|
|
5938
|
-
id: "meta-llama/llama-3-70b-instruct",
|
|
5939
|
-
name: "Meta: Llama 3 70B Instruct",
|
|
5940
|
-
api: "openai-completions",
|
|
5941
|
-
provider: "openrouter",
|
|
5942
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5943
|
-
reasoning: false,
|
|
5944
|
-
input: ["text"],
|
|
5945
|
-
cost: {
|
|
5946
|
-
input: 0.3,
|
|
5947
|
-
output: 0.39999999999999997,
|
|
5948
|
-
cacheRead: 0,
|
|
5949
|
-
cacheWrite: 0,
|
|
5950
|
-
},
|
|
5951
|
-
contextWindow: 8192,
|
|
5952
|
-
maxTokens: 16384,
|
|
5953
|
-
},
|
|
5954
6107
|
"mistralai/mixtral-8x22b-instruct": {
|
|
5955
6108
|
id: "mistralai/mixtral-8x22b-instruct",
|
|
5956
6109
|
name: "Mistral: Mixtral 8x22B Instruct",
|
|
@@ -6036,38 +6189,38 @@ export const MODELS = {
|
|
|
6036
6189
|
contextWindow: 128000,
|
|
6037
6190
|
maxTokens: 4096,
|
|
6038
6191
|
},
|
|
6039
|
-
"openai/gpt-
|
|
6040
|
-
id: "openai/gpt-
|
|
6041
|
-
name: "OpenAI: GPT-
|
|
6192
|
+
"openai/gpt-3.5-turbo-0613": {
|
|
6193
|
+
id: "openai/gpt-3.5-turbo-0613",
|
|
6194
|
+
name: "OpenAI: GPT-3.5 Turbo (older v0613)",
|
|
6042
6195
|
api: "openai-completions",
|
|
6043
6196
|
provider: "openrouter",
|
|
6044
6197
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6045
6198
|
reasoning: false,
|
|
6046
6199
|
input: ["text"],
|
|
6047
6200
|
cost: {
|
|
6048
|
-
input:
|
|
6049
|
-
output:
|
|
6201
|
+
input: 1,
|
|
6202
|
+
output: 2,
|
|
6050
6203
|
cacheRead: 0,
|
|
6051
6204
|
cacheWrite: 0,
|
|
6052
6205
|
},
|
|
6053
|
-
contextWindow:
|
|
6206
|
+
contextWindow: 4095,
|
|
6054
6207
|
maxTokens: 4096,
|
|
6055
6208
|
},
|
|
6056
|
-
"openai/gpt-
|
|
6057
|
-
id: "openai/gpt-
|
|
6058
|
-
name: "OpenAI: GPT-
|
|
6209
|
+
"openai/gpt-4-turbo-preview": {
|
|
6210
|
+
id: "openai/gpt-4-turbo-preview",
|
|
6211
|
+
name: "OpenAI: GPT-4 Turbo Preview",
|
|
6059
6212
|
api: "openai-completions",
|
|
6060
6213
|
provider: "openrouter",
|
|
6061
6214
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6062
6215
|
reasoning: false,
|
|
6063
6216
|
input: ["text"],
|
|
6064
6217
|
cost: {
|
|
6065
|
-
input:
|
|
6066
|
-
output:
|
|
6218
|
+
input: 10,
|
|
6219
|
+
output: 30,
|
|
6067
6220
|
cacheRead: 0,
|
|
6068
6221
|
cacheWrite: 0,
|
|
6069
6222
|
},
|
|
6070
|
-
contextWindow:
|
|
6223
|
+
contextWindow: 128000,
|
|
6071
6224
|
maxTokens: 4096,
|
|
6072
6225
|
},
|
|
6073
6226
|
"mistralai/mistral-tiny": {
|
|
@@ -6138,9 +6291,9 @@ export const MODELS = {
|
|
|
6138
6291
|
contextWindow: 16385,
|
|
6139
6292
|
maxTokens: 4096,
|
|
6140
6293
|
},
|
|
6141
|
-
"openai/gpt-4": {
|
|
6142
|
-
id: "openai/gpt-4",
|
|
6143
|
-
name: "OpenAI: GPT-4",
|
|
6294
|
+
"openai/gpt-4-0314": {
|
|
6295
|
+
id: "openai/gpt-4-0314",
|
|
6296
|
+
name: "OpenAI: GPT-4 (older v0314)",
|
|
6144
6297
|
api: "openai-completions",
|
|
6145
6298
|
provider: "openrouter",
|
|
6146
6299
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -6155,38 +6308,38 @@ export const MODELS = {
|
|
|
6155
6308
|
contextWindow: 8191,
|
|
6156
6309
|
maxTokens: 4096,
|
|
6157
6310
|
},
|
|
6158
|
-
"openai/gpt-
|
|
6159
|
-
id: "openai/gpt-
|
|
6160
|
-
name: "OpenAI: GPT-
|
|
6311
|
+
"openai/gpt-4": {
|
|
6312
|
+
id: "openai/gpt-4",
|
|
6313
|
+
name: "OpenAI: GPT-4",
|
|
6161
6314
|
api: "openai-completions",
|
|
6162
6315
|
provider: "openrouter",
|
|
6163
6316
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6164
6317
|
reasoning: false,
|
|
6165
6318
|
input: ["text"],
|
|
6166
6319
|
cost: {
|
|
6167
|
-
input:
|
|
6168
|
-
output:
|
|
6320
|
+
input: 30,
|
|
6321
|
+
output: 60,
|
|
6169
6322
|
cacheRead: 0,
|
|
6170
6323
|
cacheWrite: 0,
|
|
6171
6324
|
},
|
|
6172
|
-
contextWindow:
|
|
6325
|
+
contextWindow: 8191,
|
|
6173
6326
|
maxTokens: 4096,
|
|
6174
6327
|
},
|
|
6175
|
-
"openai/gpt-
|
|
6176
|
-
id: "openai/gpt-
|
|
6177
|
-
name: "OpenAI: GPT-
|
|
6328
|
+
"openai/gpt-3.5-turbo": {
|
|
6329
|
+
id: "openai/gpt-3.5-turbo",
|
|
6330
|
+
name: "OpenAI: GPT-3.5 Turbo",
|
|
6178
6331
|
api: "openai-completions",
|
|
6179
6332
|
provider: "openrouter",
|
|
6180
6333
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
6181
6334
|
reasoning: false,
|
|
6182
6335
|
input: ["text"],
|
|
6183
6336
|
cost: {
|
|
6184
|
-
input:
|
|
6185
|
-
output:
|
|
6337
|
+
input: 0.5,
|
|
6338
|
+
output: 1.5,
|
|
6186
6339
|
cacheRead: 0,
|
|
6187
6340
|
cacheWrite: 0,
|
|
6188
6341
|
},
|
|
6189
|
-
contextWindow:
|
|
6342
|
+
contextWindow: 16385,
|
|
6190
6343
|
maxTokens: 4096,
|
|
6191
6344
|
},
|
|
6192
6345
|
"openrouter/auto": {
|